Open jornfranke opened 7 years ago
Ripple (https://wiki.ripple.com/NodeBackEnd) stores the ledger in Rocksdb format (https://github.com/facebook/rocksdb/wiki/A-Tutorial-of-RocksDB-SST-formats) . Hence, one must be able to read rocksdb files from HDFS. See also: https://github.com/facebook/rocksdb/wiki/Rocksdb-BlockBasedTable-Format
Ripple "is a network of computers which use the Ripple consensus algorithm to atomically settle and record transactions on a secure distributed database, the Ripple Consensus Ledger (RCL). Because of its distributed nature, the RCL offers transaction immutability without a central operator. The RCL contains a built-in currency exchange and its path-finding algorithm finds competitive exchange rates across order books and currency pairs": https://github.com/ripple/rippled
This implies to add additional HadoopInputFormats to process the Ripple blockchain. Furthermore a HiveSerde and Spark datasource should be created. Unit tests must be included.
Finally, an example for MapReduce, Flink, Hive and Spark should be provided. The examples should include unit and integration tests.