The Spark-Riak connector enables you to connect Spark applications to Riak TS with the Spark RDD and Spark DataFrames APIs. You can write your app in Scala, Python, and Java. The connector makes it easy to partition the data you get from Riak so multiple Spark workers can process the data in parallel, and it has support for failover if a Riak node goes down while your Spark job is running.
- Riak TS 1.2+
- Apache Spark 1.6+
- Scala 2.10
- Java 8
- Construct a Spark Dataframe from a Riak TS table using range queries and schema discovery
- Save a Spark Dataframe into a Riak TS table