Packages

  • package root
    Definition Classes
    root
  • package com
    Definition Classes
    root
  • package datastax
    Definition Classes
    com
  • package spark
    Definition Classes
    datastax
  • package connector

    The root package of Cassandra connector for Apache Spark.

    The root package of Cassandra connector for Apache Spark. Offers handy implicit conversions that add Cassandra-specific methods to SparkContext and RDD.

    Call cassandraTable method on the SparkContext object to create a CassandraRDD exposing Cassandra tables as Spark RDDs.

    Call RDDFunctions saveToCassandra function on any RDD to save distributed collection to a Cassandra table.

    Example:

    CREATE KEYSPACE test WITH replication = {'class': 'SimpleStrategy', 'replication_factor': 1 };
    CREATE TABLE test.words (word text PRIMARY KEY, count int);
    INSERT INTO test.words(word, count) VALUES ("and", 50);
    import com.datastax.spark.connector._
    
    val sparkMasterHost = "127.0.0.1"
    val cassandraHost = "127.0.0.1"
    val keyspace = "test"
    val table = "words"
    
    // Tell Spark the address of one Cassandra node:
    val conf = new SparkConf(true).set("spark.cassandra.connection.host", cassandraHost)
    
    // Connect to the Spark cluster:
    val sc = new SparkContext("spark://" + sparkMasterHost + ":7077", "example", conf)
    
    // Read the table and print its contents:
    val rdd = sc.cassandraTable(keyspace, table)
    rdd.toArray().foreach(println)
    
    // Write two rows to the table:
    val col = sc.parallelize(Seq(("of", 1200), ("the", "863")))
    col.saveToCassandra(keyspace, table)
    
    sc.stop()
    Definition Classes
    spark
  • package streaming
    Definition Classes
    connector
  • CassandraStreamingRDD
  • DStreamFunctions
  • StreamingContextFunctions

class DStreamFunctions[T] extends WritableToCassandra[T] with Serializable with Logging

Linear Supertypes
Logging, Serializable, Serializable, WritableToCassandra[T], AnyRef, Any
Ordering
  1. Alphabetic
  2. By Inheritance
Inherited
  1. DStreamFunctions
  2. Logging
  3. Serializable
  4. Serializable
  5. WritableToCassandra
  6. AnyRef
  7. Any
  1. Hide All
  2. Show All
Visibility
  1. Public
  2. All

Instance Constructors

  1. new DStreamFunctions(dstream: DStream[T])

Value Members

  1. final def !=(arg0: Any): Boolean
    Definition Classes
    AnyRef → Any
  2. final def ##(): Int
    Definition Classes
    AnyRef → Any
  3. final def ==(arg0: Any): Boolean
    Definition Classes
    AnyRef → Any
  4. final def asInstanceOf[T0]: T0
    Definition Classes
    Any
  5. def clone(): AnyRef
    Attributes
    protected[lang]
    Definition Classes
    AnyRef
    Annotations
    @throws( ... ) @native() @HotSpotIntrinsicCandidate()
  6. def conf: SparkConf
  7. def deleteFromCassandra(keyspaceName: String, tableName: String, deleteColumns: ColumnSelector = SomeColumns(), keyColumns: ColumnSelector = PrimaryKeyColumns, writeConf: WriteConf = ...)(implicit connector: CassandraConnector = CassandraConnector(sparkContext), rwf: RowWriterFactory[T]): Unit

    Delete data from Cassandra table, using data from the stream as a list of primary keys.

    Delete data from Cassandra table, using data from the stream as a list of primary keys. Uses the specified column names.

    keyspaceName

    the name of the Keyspace to use

    tableName

    the name of the Table to use

    deleteColumns

    The list of column names to delete, empty ColumnSelector means full row.

    keyColumns

    Primary key columns selector, Optional. All RDD primary columns columns will be checked by default

    writeConf

    additional configuration object allowing to set consistency level, batch size, etc.

    Definition Classes
    DStreamFunctionsWritableToCassandra
    See also

    com.datastax.spark.connector.writer.WritableToCassandra

  8. final def eq(arg0: AnyRef): Boolean
    Definition Classes
    AnyRef
  9. def equals(arg0: Any): Boolean
    Definition Classes
    AnyRef → Any
  10. final def getClass(): Class[_]
    Definition Classes
    AnyRef → Any
    Annotations
    @native() @HotSpotIntrinsicCandidate()
  11. def hashCode(): Int
    Definition Classes
    AnyRef → Any
    Annotations
    @native() @HotSpotIntrinsicCandidate()
  12. final def isInstanceOf[T0]: Boolean
    Definition Classes
    Any
  13. def isTraceEnabled(): Boolean
    Attributes
    protected
    Definition Classes
    Logging
  14. def joinWithCassandraTable[R](keyspaceName: String, tableName: String, selectedColumns: ColumnSelector = AllColumns, joinColumns: ColumnSelector = PartitionKeyColumns)(implicit connector: CassandraConnector = CassandraConnector(sparkContext), newType: ClassTag[R], rrf: RowReaderFactory[R], ev: ValidRDDType[R], currentType: ClassTag[T], rwf: RowWriterFactory[T]): DStream[(T, R)]

    Transforms RDDs with com.datastax.spark.connector.RDDFunctions.joinWithCassandraTable for each produced RDD

  15. def log: Logger
    Attributes
    protected
    Definition Classes
    Logging
  16. def logDebug(msg: ⇒ String, throwable: Throwable): Unit
    Attributes
    protected
    Definition Classes
    Logging
  17. def logDebug(msg: ⇒ String): Unit
    Attributes
    protected
    Definition Classes
    Logging
  18. def logError(msg: ⇒ String, throwable: Throwable): Unit
    Attributes
    protected
    Definition Classes
    Logging
  19. def logError(msg: ⇒ String): Unit
    Attributes
    protected
    Definition Classes
    Logging
  20. def logInfo(msg: ⇒ String, throwable: Throwable): Unit
    Attributes
    protected
    Definition Classes
    Logging
  21. def logInfo(msg: ⇒ String): Unit
    Attributes
    protected
    Definition Classes
    Logging
  22. def logName: String
    Attributes
    protected
    Definition Classes
    Logging
  23. def logTrace(msg: ⇒ String, throwable: Throwable): Unit
    Attributes
    protected
    Definition Classes
    Logging
  24. def logTrace(msg: ⇒ String): Unit
    Attributes
    protected
    Definition Classes
    Logging
  25. def logWarning(msg: ⇒ String, throwable: Throwable): Unit
    Attributes
    protected
    Definition Classes
    Logging
  26. def logWarning(msg: ⇒ String): Unit
    Attributes
    protected
    Definition Classes
    Logging
  27. final def ne(arg0: AnyRef): Boolean
    Definition Classes
    AnyRef
  28. final def notify(): Unit
    Definition Classes
    AnyRef
    Annotations
    @native() @HotSpotIntrinsicCandidate()
  29. final def notifyAll(): Unit
    Definition Classes
    AnyRef
    Annotations
    @native() @HotSpotIntrinsicCandidate()
  30. def repartitionByCassandraReplica(keyspaceName: String, tableName: String, partitionsPerHost: Int = 10, partitionKeyMapper: ColumnSelector = PartitionKeyColumns)(implicit connector: CassandraConnector = CassandraConnector(conf), currentType: ClassTag[T], rwf: RowWriterFactory[T]): DStream[T]

    Transforms RDDs with com.datastax.spark.connector.RDDFunctions.repartitionByCassandraReplica for each produced RDD.

  31. def saveToCassandra(keyspaceName: String, tableName: String, columnNames: ColumnSelector = AllColumns, writeConf: WriteConf = WriteConf.fromSparkConf(conf))(implicit connector: CassandraConnector = CassandraConnector(conf), rwf: RowWriterFactory[T]): Unit

    Performs com.datastax.spark.connector.writer.WritableToCassandra for each produced RDD.

    Performs com.datastax.spark.connector.writer.WritableToCassandra for each produced RDD. Uses specific column names with an additional batch size.

    keyspaceName

    the name of the Keyspace to use

    tableName

    the name of the Table to use

    columnNames

    The list of column names to save data to. Uses only the unique column names, and you must select at least all primary key columns. All other fields are discarded. Non-selected property/column names are left unchanged.

    writeConf

    additional configuration object allowing to set consistency level, batch size, etc.

    Definition Classes
    DStreamFunctionsWritableToCassandra
  32. def sparkContext: SparkContext
    Definition Classes
    DStreamFunctionsWritableToCassandra
  33. final def synchronized[T0](arg0: ⇒ T0): T0
    Definition Classes
    AnyRef
  34. def toString(): String
    Definition Classes
    AnyRef → Any
  35. final def wait(arg0: Long, arg1: Int): Unit
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  36. final def wait(arg0: Long): Unit
    Definition Classes
    AnyRef
    Annotations
    @throws( ... ) @native()
  37. final def wait(): Unit
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  38. def warnIfKeepAliveIsShort(): Unit

Deprecated Value Members

  1. def finalize(): Unit
    Attributes
    protected[lang]
    Definition Classes
    AnyRef
    Annotations
    @throws( classOf[java.lang.Throwable] ) @Deprecated @deprecated
    Deprecated

    (Since version ) see corresponding Javadoc for more information.

Inherited from Logging

Inherited from Serializable

Inherited from Serializable

Inherited from WritableToCassandra[T]

Inherited from AnyRef

Inherited from Any

Ungrouped