Packages

  • package root
    Definition Classes
    root
  • package com
    Definition Classes
    root
  • package datastax
    Definition Classes
    com
  • package spark
    Definition Classes
    datastax
  • package connector

    The root package of Cassandra connector for Apache Spark.

    The root package of Cassandra connector for Apache Spark. Offers handy implicit conversions that add Cassandra-specific methods to SparkContext and RDD.

    Call cassandraTable method on the SparkContext object to create a CassandraRDD exposing Cassandra tables as Spark RDDs.

    Call RDDFunctions saveToCassandra function on any RDD to save distributed collection to a Cassandra table.

    Example:

    CREATE KEYSPACE test WITH replication = {'class': 'SimpleStrategy', 'replication_factor': 1 };
    CREATE TABLE test.words (word text PRIMARY KEY, count int);
    INSERT INTO test.words(word, count) VALUES ("and", 50);
    import com.datastax.spark.connector._
    
    val sparkMasterHost = "127.0.0.1"
    val cassandraHost = "127.0.0.1"
    val keyspace = "test"
    val table = "words"
    
    // Tell Spark the address of one Cassandra node:
    val conf = new SparkConf(true).set("spark.cassandra.connection.host", cassandraHost)
    
    // Connect to the Spark cluster:
    val sc = new SparkContext("spark://" + sparkMasterHost + ":7077", "example", conf)
    
    // Read the table and print its contents:
    val rdd = sc.cassandraTable(keyspace, table)
    rdd.toArray().foreach(println)
    
    // Write two rows to the table:
    val col = sc.parallelize(Seq(("of", 1200), ("the", "863")))
    col.saveToCassandra(keyspace, table)
    
    sc.stop()
    Definition Classes
    spark
  • package streaming
    Definition Classes
    connector
  • CassandraStreamingRDD
  • DStreamFunctions
  • StreamingContextFunctions
c

com.datastax.spark.connector.streaming

StreamingContextFunctions

class StreamingContextFunctions extends SparkContextFunctions

Provides Cassandra-specific methods on org.apache.spark.streaming.StreamingContext.

Linear Supertypes
SparkContextFunctions, Serializable, Serializable, AnyRef, Any
Ordering
  1. Alphabetic
  2. By Inheritance
Inherited
  1. StreamingContextFunctions
  2. SparkContextFunctions
  3. Serializable
  4. Serializable
  5. AnyRef
  6. Any
  1. Hide All
  2. Show All
Visibility
  1. Public
  2. All

Instance Constructors

  1. new StreamingContextFunctions(ssc: StreamingContext)

    ssc

    the Spark Streaming context

Value Members

  1. final def !=(arg0: Any): Boolean
    Definition Classes
    AnyRef → Any
  2. final def ##(): Int
    Definition Classes
    AnyRef → Any
  3. final def ==(arg0: Any): Boolean
    Definition Classes
    AnyRef → Any
  4. final def asInstanceOf[T0]: T0
    Definition Classes
    Any
  5. def cassandraTable[T](keyspace: String, table: String)(implicit connector: CassandraConnector = ..., readConf: ReadConf = ReadConf.fromSparkConf(sc.getConf), ct: ClassTag[T], rrf: RowReaderFactory[T], ev: ValidRDDType[T]): CassandraStreamingRDD[T]

    Returns a view of a Cassandra table as CassandraRDD.

    Returns a view of a Cassandra table as CassandraRDD. This method is made available on SparkContext by importing com.datastax.spark.connector._

    Depending on the type parameter passed to cassandraTable, every row is converted to one of the following:

    • an CassandraRow object (default, if no type given)
    • a tuple containing column values in the same order as columns selected by CassandraRDD#select
    • object of a user defined class, populated by appropriate ColumnMapper

    Example:

    CREATE KEYSPACE test WITH replication = {'class': 'SimpleStrategy', 'replication_factor': 1 };
    CREATE TABLE test.words (word text PRIMARY KEY, count int);
    INSERT INTO test.words (word, count) VALUES ('foo', 20);
    INSERT INTO test.words (word, count) VALUES ('bar', 20);
    ...
    // Obtaining RDD of CassandraRow objects:
    val rdd1 = sc.cassandraTable("test", "words")
    rdd1.first.getString("word")  // foo
    rdd1.first.getInt("count")    // 20
    
    // Obtaining RDD of tuples:
    val rdd2 = sc.cassandraTable[(String, Int)]("test", "words").select("word", "count")
    rdd2.first._1  // foo
    rdd2.first._2  // 20
    
    // Obtaining RDD of user defined objects:
    case class WordCount(word: String, count: Int)
    val rdd3 = sc.cassandraTable[WordCount]("test", "words")
    rdd3.first.word  // foo
    rdd3.first.count // 20
    Definition Classes
    StreamingContextFunctionsSparkContextFunctions
  6. def clone(): AnyRef
    Attributes
    protected[lang]
    Definition Classes
    AnyRef
    Annotations
    @throws( ... ) @native() @HotSpotIntrinsicCandidate()
  7. def emptyCassandraTable[T](keyspace: String, table: String)(implicit connector: CassandraConnector = CassandraConnector(sc), readConf: ReadConf = ReadConf.fromSparkConf(sc.getConf), ct: ClassTag[T], rrf: RowReaderFactory[T], ev: ValidRDDType[T]): EmptyCassandraRDD[T]

    Produces the empty CassandraRDD which does not perform any validation and it does not even try to return any rows.

    Produces the empty CassandraRDD which does not perform any validation and it does not even try to return any rows.

    Definition Classes
    SparkContextFunctions
  8. final def eq(arg0: AnyRef): Boolean
    Definition Classes
    AnyRef
  9. def equals(arg0: Any): Boolean
    Definition Classes
    AnyRef → Any
  10. final def getClass(): Class[_]
    Definition Classes
    AnyRef → Any
    Annotations
    @native() @HotSpotIntrinsicCandidate()
  11. def hashCode(): Int
    Definition Classes
    AnyRef → Any
    Annotations
    @native() @HotSpotIntrinsicCandidate()
  12. final def isInstanceOf[T0]: Boolean
    Definition Classes
    Any
  13. final def ne(arg0: AnyRef): Boolean
    Definition Classes
    AnyRef
  14. final def notify(): Unit
    Definition Classes
    AnyRef
    Annotations
    @native() @HotSpotIntrinsicCandidate()
  15. final def notifyAll(): Unit
    Definition Classes
    AnyRef
    Annotations
    @native() @HotSpotIntrinsicCandidate()
  16. val sc: SparkContext
    Definition Classes
    SparkContextFunctions
  17. final def synchronized[T0](arg0: ⇒ T0): T0
    Definition Classes
    AnyRef
  18. def toString(): String
    Definition Classes
    AnyRef → Any
  19. final def wait(arg0: Long, arg1: Int): Unit
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  20. final def wait(arg0: Long): Unit
    Definition Classes
    AnyRef
    Annotations
    @throws( ... ) @native()
  21. final def wait(): Unit
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )

Deprecated Value Members

  1. def finalize(): Unit
    Attributes
    protected[lang]
    Definition Classes
    AnyRef
    Annotations
    @throws( classOf[java.lang.Throwable] ) @Deprecated @deprecated
    Deprecated

    (Since version ) see corresponding Javadoc for more information.

Inherited from SparkContextFunctions

Inherited from Serializable

Inherited from Serializable

Inherited from AnyRef

Inherited from Any

Ungrouped