The root package of the DataStax Connector for Apache Spark to Apache Cassandra.
The root package of the DataStax Connector for Apache Spark to Apache Cassandra. Offers handy implicit conversions that add Cassandra-specific methods to SparkContext and RDD.
Call cassandraTable method on the SparkContext object to create a CassandraRDD exposing Cassandra tables as Spark RDDs.
Call RDDFunctions saveToCassandra
function on any RDD
to save distributed collection to a Cassandra table.
Example:
CREATE KEYSPACE test WITH replication = {'class': 'SimpleStrategy', 'replication_factor': 1 }; CREATE TABLE test.words (word text PRIMARY KEY, count int); INSERT INTO test.words(word, count) VALUES ("and", 50);
import com.datastax.spark.connector._ val sparkMasterHost = "127.0.0.1" val cassandraHost = "127.0.0.1" val keyspace = "test" val table = "words" // Tell Spark the address of one Cassandra node: val conf = new SparkConf(true).set("spark.cassandra.connection.host", cassandraHost) // Connect to the Spark cluster: val sc = new SparkContext("spark://" + sparkMasterHost + ":7077", "example", conf) // Read the table and print its contents: val rdd = sc.cassandraTable(keyspace, table) rdd.toArray().foreach(println) // Write two rows to the table: val col = sc.parallelize(Seq(("of", 1200), ("the", "863"))) col.saveToCassandra(keyspace, table) sc.stop()