Skip to content

Use Spark to count the number of transactions from files containing Ethereum Blockchain data

Jörn Franke edited this page Nov 12, 2017 · 1 revision

This is a Spark application written in Scala demonstrating some of the capabilities of the hadoopcryptoledger library. It takes as input a set of files on HDFS containing Ethereum Blockchain data. As output it returns the total number of inputs of transactions found in the blockchain data. It has successfully been tested with the Cloudera Quickstart VM 5.5 and HDP Sandbox 2.5, but other Hadoop distributions should work equally well. Spark 1.5 was used for testing.

Getting blockchain data

See here how to fetch Ethereum blockchain data.

After it has been copied you are ready to use the example.

Building the example

Execute

git clone https://github.com/ZuInnoTe/hadoopcryptoledger.git hadoopcryptoledger

You can build the application by changing to the directory hadoopcryptoledger/examples/spark-scala-ethereumblock and using the following command:

sbt clean assembly test it:test

This will also execute the integration tests

You will find the jar "example-hcl-spark-scala-ethereumblock.jar" in ./target/scala-2.10

Running the example

Make sure that the output directory is clean:

hadoop fs -rm -R /user/ethereum/output

Execute the following command (to execute it using a local master)

spark-submit --class org.zuinnote.spark.ethereum.example.SparkScalaEthereumBlockCounter --master local[8] ./target/scala-2.10/example-hcl-spark-scala-ethereumblock.jar /user/ethereum/input /user/ethereum/output

After the Spark job has completed, you find the result in /user/ethereum/output. You can display it using the following command:

hadoop fs -cat /user/ethereum/output/part-00000

More Information

Understanding the structure of Ethereum data:

Ethereum Yellow paper: http://yellowpaper.io/

Clone this wiki locally