Class/Object

org.apache.spark.sql.cassandra

CassandraSQLContext

Related Docs: object CassandraSQLContext | package cassandra

Permalink

class CassandraSQLContext extends SQLContext

Allows to execute SQL queries against Cassandra and access results as SchemaRDD collections. Predicate pushdown to Cassandra is supported.

Example:

import com.datastax.spark.connector._

val sparkMasterHost = "127.0.0.1"
val cassandraHost = "127.0.0.1"

// Tell Spark the address of one Cassandra node:
val conf = new SparkConf(true).set("spark.cassandra.connection.host", cassandraHost)

// Connect to the Spark cluster:
val sc = new SparkContext("spark://" + sparkMasterHost + ":7077", "example", conf)

// Create CassandraSQLContext:
val cc = new CassandraSQLContext(sc)

// Execute SQL query:
val rdd = cc.sql("SELECT * FROM keyspace.table ...")
Linear Supertypes
SQLContext, Serializable, Serializable, Logging, AnyRef, Any
Ordering
  1. Alphabetic
  2. By inheritance
Inherited
  1. CassandraSQLContext
  2. SQLContext
  3. Serializable
  4. Serializable
  5. Logging
  6. AnyRef
  7. Any
Implicitly
  1. by any2stringadd
  2. by StringFormat
  3. by Ensuring
  4. by ArrowAssoc
  1. Hide All
  2. Show all
Visibility
  1. Public
  2. All

Instance Constructors

  1. new CassandraSQLContext(sc: SparkContext)

    Permalink

Type Members

  1. class QueryExecution extends execution.QueryExecution

    Permalink
    Attributes
    protected[org.apache.spark.sql]
    Definition Classes
    SQLContext
    Annotations
    @deprecated
    Deprecated

    (Since version 1.6.0) use org.apache.spark.sql.QueryExecution

  2. class SparkPlanner extends execution.SparkPlanner

    Permalink
    Attributes
    protected[org.apache.spark.sql]
    Definition Classes
    SQLContext
    Annotations
    @deprecated
    Deprecated

    (Since version 1.6.0) use org.apache.spark.sql.SparkPlanner

Value Members

  1. final def !=(arg0: Any): Boolean

    Permalink
    Definition Classes
    AnyRef → Any
  2. final def ##(): Int

    Permalink
    Definition Classes
    AnyRef → Any
  3. def +(other: String): String

    Permalink
    Implicit information
    This member is added by an implicit conversion from CassandraSQLContext to any2stringadd[CassandraSQLContext] performed by method any2stringadd in scala.Predef.
    Definition Classes
    any2stringadd
  4. def ->[B](y: B): (CassandraSQLContext, B)

    Permalink
    Implicit information
    This member is added by an implicit conversion from CassandraSQLContext to ArrowAssoc[CassandraSQLContext] performed by method ArrowAssoc in scala.Predef.
    Definition Classes
    ArrowAssoc
    Annotations
    @inline()
  5. final def ==(arg0: Any): Boolean

    Permalink
    Definition Classes
    AnyRef → Any
  6. def addJar(path: String): Unit

    Permalink
    Attributes
    protected[org.apache.spark.sql]
    Definition Classes
    SQLContext
  7. lazy val analyzer: Analyzer

    Permalink
    Attributes
    protected[org.apache.spark.sql]
    Definition Classes
    SQLContext
  8. def applySchemaToPythonRDD(rdd: RDD[Array[Any]], schema: StructType): DataFrame

    Permalink
    Attributes
    protected[org.apache.spark.sql]
    Definition Classes
    SQLContext
  9. def applySchemaToPythonRDD(rdd: RDD[Array[Any]], schemaString: String): DataFrame

    Permalink
    Attributes
    protected[org.apache.spark.sql]
    Definition Classes
    SQLContext
  10. final def asInstanceOf[T0]: T0

    Permalink
    Definition Classes
    Any
  11. def baseRelationToDataFrame(baseRelation: BaseRelation): DataFrame

    Permalink
    Definition Classes
    SQLContext
  12. val cacheManager: execution.CacheManager

    Permalink
    Attributes
    protected[org.apache.spark.sql]
    Definition Classes
    SQLContext
  13. def cacheTable(tableName: String): Unit

    Permalink
    Definition Classes
    SQLContext
  14. def cassandraSql(cassandraQuery: String): DataFrame

    Permalink

    Executes SQL query against Cassandra and returns DataFrame representing the result.

  15. lazy val catalog: CassandraCatalog

    Permalink

    A catalyst metadata catalog that points to Cassandra.

    A catalyst metadata catalog that points to Cassandra.

    Attributes
    protected[org.apache.spark.sql]
    Definition Classes
    CassandraSQLContext → SQLContext
  16. def clearCache(): Unit

    Permalink
    Definition Classes
    SQLContext
  17. def clone(): AnyRef

    Permalink
    Attributes
    protected[java.lang]
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  18. lazy val conf: SQLConf

    Permalink
    Attributes
    protected[org.apache.spark.sql]
    Definition Classes
    SQLContext
  19. def createDataFrame(data: List[_], beanClass: Class[_]): DataFrame

    Permalink
    Definition Classes
    SQLContext
  20. def createDataFrame(rdd: JavaRDD[_], beanClass: Class[_]): DataFrame

    Permalink
    Definition Classes
    SQLContext
  21. def createDataFrame(rdd: RDD[_], beanClass: Class[_]): DataFrame

    Permalink
    Definition Classes
    SQLContext
  22. def createDataFrame(rows: List[Row], schema: StructType): DataFrame

    Permalink
    Definition Classes
    SQLContext
    Annotations
    @DeveloperApi()
  23. def createDataFrame(rowRDD: JavaRDD[Row], schema: StructType): DataFrame

    Permalink
    Definition Classes
    SQLContext
    Annotations
    @DeveloperApi()
  24. def createDataFrame(rowRDD: RDD[Row], schema: StructType): DataFrame

    Permalink
    Definition Classes
    SQLContext
    Annotations
    @DeveloperApi()
  25. def createDataFrame[A <: Product](data: Seq[A])(implicit arg0: scala.reflect.api.JavaUniverse.TypeTag[A]): DataFrame

    Permalink
    Definition Classes
    SQLContext
    Annotations
    @Experimental()
  26. def createDataFrame[A <: Product](rdd: RDD[A])(implicit arg0: scala.reflect.api.JavaUniverse.TypeTag[A]): DataFrame

    Permalink
    Definition Classes
    SQLContext
    Annotations
    @Experimental()
  27. def createDataset[T](data: List[T])(implicit arg0: Encoder[T]): Dataset[T]

    Permalink
    Definition Classes
    SQLContext
  28. def createDataset[T](data: RDD[T])(implicit arg0: Encoder[T]): Dataset[T]

    Permalink
    Definition Classes
    SQLContext
  29. def createDataset[T](data: Seq[T])(implicit arg0: Encoder[T]): Dataset[T]

    Permalink
    Definition Classes
    SQLContext
  30. def createExternalTable(tableName: String, source: String, schema: StructType, options: Map[String, String]): DataFrame

    Permalink
    Definition Classes
    SQLContext
    Annotations
    @Experimental()
  31. def createExternalTable(tableName: String, source: String, schema: StructType, options: Map[String, String]): DataFrame

    Permalink
    Definition Classes
    SQLContext
    Annotations
    @Experimental()
  32. def createExternalTable(tableName: String, source: String, options: Map[String, String]): DataFrame

    Permalink
    Definition Classes
    SQLContext
    Annotations
    @Experimental()
  33. def createExternalTable(tableName: String, source: String, options: Map[String, String]): DataFrame

    Permalink
    Definition Classes
    SQLContext
    Annotations
    @Experimental()
  34. def createExternalTable(tableName: String, path: String, source: String): DataFrame

    Permalink
    Definition Classes
    SQLContext
    Annotations
    @Experimental()
  35. def createExternalTable(tableName: String, path: String): DataFrame

    Permalink
    Definition Classes
    SQLContext
    Annotations
    @Experimental()
  36. val ddlParser: DDLParser

    Permalink
    Attributes
    protected[org.apache.spark.sql]
    Definition Classes
    SQLContext
  37. def dialectClassName: String

    Permalink
    Attributes
    protected[org.apache.spark.sql]
    Definition Classes
    SQLContext
  38. def dropTempTable(tableName: String): Unit

    Permalink
    Definition Classes
    SQLContext
  39. lazy val emptyDataFrame: DataFrame

    Permalink
    Definition Classes
    SQLContext
  40. lazy val emptyResult: RDD[InternalRow]

    Permalink
    Attributes
    protected[org.apache.spark.sql]
    Definition Classes
    SQLContext
  41. def ensuring(cond: (CassandraSQLContext) ⇒ Boolean, msg: ⇒ Any): CassandraSQLContext

    Permalink
    Implicit information
    This member is added by an implicit conversion from CassandraSQLContext to Ensuring[CassandraSQLContext] performed by method Ensuring in scala.Predef.
    Definition Classes
    Ensuring
  42. def ensuring(cond: (CassandraSQLContext) ⇒ Boolean): CassandraSQLContext

    Permalink
    Implicit information
    This member is added by an implicit conversion from CassandraSQLContext to Ensuring[CassandraSQLContext] performed by method Ensuring in scala.Predef.
    Definition Classes
    Ensuring
  43. def ensuring(cond: Boolean, msg: ⇒ Any): CassandraSQLContext

    Permalink
    Implicit information
    This member is added by an implicit conversion from CassandraSQLContext to Ensuring[CassandraSQLContext] performed by method Ensuring in scala.Predef.
    Definition Classes
    Ensuring
  44. def ensuring(cond: Boolean): CassandraSQLContext

    Permalink
    Implicit information
    This member is added by an implicit conversion from CassandraSQLContext to Ensuring[CassandraSQLContext] performed by method Ensuring in scala.Predef.
    Definition Classes
    Ensuring
  45. final def eq(arg0: AnyRef): Boolean

    Permalink
    Definition Classes
    AnyRef
  46. def equals(arg0: Any): Boolean

    Permalink
    Definition Classes
    AnyRef → Any
  47. def executePlan(plan: LogicalPlan): execution.QueryExecution

    Permalink
    Attributes
    protected[org.apache.spark.sql]
    Definition Classes
    CassandraSQLContext → SQLContext
  48. def executeSql(sql: String): execution.QueryExecution

    Permalink
    Attributes
    protected[org.apache.spark.sql]
    Definition Classes
    SQLContext
  49. val experimental: ExperimentalMethods

    Permalink
    Definition Classes
    SQLContext
  50. def finalize(): Unit

    Permalink
    Attributes
    protected[java.lang]
    Definition Classes
    AnyRef
    Annotations
    @throws( classOf[java.lang.Throwable] )
  51. def formatted(fmtstr: String): String

    Permalink
    Implicit information
    This member is added by an implicit conversion from CassandraSQLContext to StringFormat[CassandraSQLContext] performed by method StringFormat in scala.Predef.
    Definition Classes
    StringFormat
    Annotations
    @inline()
  52. lazy val functionRegistry: FunctionRegistry

    Permalink
    Attributes
    protected[org.apache.spark.sql]
    Definition Classes
    SQLContext
  53. def getAllConfs: Map[String, String]

    Permalink
    Definition Classes
    SQLContext
  54. final def getClass(): Class[_]

    Permalink
    Definition Classes
    AnyRef → Any
  55. def getCluster: String

    Permalink

    Get current used cluster name

  56. def getConf(key: String, defaultValue: String): String

    Permalink
    Definition Classes
    SQLContext
  57. def getConf(key: String): String

    Permalink
    Definition Classes
    SQLContext
  58. def getKeyspace: String

    Permalink

    Returns keyspace/database set previously by setKeyspace or throws IllegalStateException if keyspace has not been set yet.

  59. def getSQLDialect(): ParserDialect

    Permalink
    Attributes
    protected[org.apache.spark.sql]
    Definition Classes
    SQLContext
  60. def getSchema(beanClass: Class[_]): Seq[AttributeReference]

    Permalink
    Attributes
    protected
    Definition Classes
    SQLContext
  61. def hashCode(): Int

    Permalink
    Definition Classes
    AnyRef → Any
  62. def isCached(tableName: String): Boolean

    Permalink
    Definition Classes
    SQLContext
  63. final def isInstanceOf[T0]: Boolean

    Permalink
    Definition Classes
    Any
  64. val isRootContext: Boolean

    Permalink
    Definition Classes
    SQLContext
  65. def isTraceEnabled(): Boolean

    Permalink
    Attributes
    protected
    Definition Classes
    Logging
  66. lazy val listenerManager: ExecutionListenerManager

    Permalink
    Definition Classes
    SQLContext
  67. def log: Logger

    Permalink
    Attributes
    protected
    Definition Classes
    Logging
  68. def logDebug(msg: ⇒ String, throwable: Throwable): Unit

    Permalink
    Attributes
    protected
    Definition Classes
    Logging
  69. def logDebug(msg: ⇒ String): Unit

    Permalink
    Attributes
    protected
    Definition Classes
    Logging
  70. def logError(msg: ⇒ String, throwable: Throwable): Unit

    Permalink
    Attributes
    protected
    Definition Classes
    Logging
  71. def logError(msg: ⇒ String): Unit

    Permalink
    Attributes
    protected
    Definition Classes
    Logging
  72. def logInfo(msg: ⇒ String, throwable: Throwable): Unit

    Permalink
    Attributes
    protected
    Definition Classes
    Logging
  73. def logInfo(msg: ⇒ String): Unit

    Permalink
    Attributes
    protected
    Definition Classes
    Logging
  74. def logName: String

    Permalink
    Attributes
    protected
    Definition Classes
    Logging
  75. def logTrace(msg: ⇒ String, throwable: Throwable): Unit

    Permalink
    Attributes
    protected
    Definition Classes
    Logging
  76. def logTrace(msg: ⇒ String): Unit

    Permalink
    Attributes
    protected
    Definition Classes
    Logging
  77. def logWarning(msg: ⇒ String, throwable: Throwable): Unit

    Permalink
    Attributes
    protected
    Definition Classes
    Logging
  78. def logWarning(msg: ⇒ String): Unit

    Permalink
    Attributes
    protected
    Definition Classes
    Logging
  79. final def ne(arg0: AnyRef): Boolean

    Permalink
    Definition Classes
    AnyRef
  80. def newSession(): SQLContext

    Permalink
    Definition Classes
    SQLContext
  81. final def notify(): Unit

    Permalink
    Definition Classes
    AnyRef
  82. final def notifyAll(): Unit

    Permalink
    Definition Classes
    AnyRef
  83. lazy val optimizer: Optimizer

    Permalink
    Attributes
    protected[org.apache.spark.sql]
    Definition Classes
    SQLContext
  84. def parseDataType(dataTypeString: String): DataType

    Permalink
    Attributes
    protected[org.apache.spark.sql]
    Definition Classes
    SQLContext
  85. def parseSql(sql: String): LogicalPlan

    Permalink
    Attributes
    protected[org.apache.spark.sql]
    Definition Classes
    SQLContext
  86. val planner: execution.SparkPlanner

    Permalink
    Attributes
    protected[org.apache.spark.sql]
    Definition Classes
    SQLContext
  87. val prepareForExecution: RuleExecutor[SparkPlan]

    Permalink
    Attributes
    protected[org.apache.spark.sql]
    Definition Classes
    SQLContext
  88. def range(start: Long, end: Long, step: Long, numPartitions: Int): DataFrame

    Permalink
    Definition Classes
    SQLContext
    Annotations
    @Experimental()
  89. def range(start: Long, end: Long): DataFrame

    Permalink
    Definition Classes
    SQLContext
    Annotations
    @Experimental()
  90. def range(end: Long): DataFrame

    Permalink
    Definition Classes
    SQLContext
    Annotations
    @Experimental()
  91. def read: DataFrameReader

    Permalink
    Definition Classes
    SQLContext
    Annotations
    @Experimental()
  92. def setCluster(cluster: String): Unit

    Permalink

    Set current used cluster name

  93. def setConf(cluster: String, keyspace: String, options: Map[String, String]): CassandraSQLContext

    Permalink

    Set the Spark Cassandra Connector configuration parameters which will be used when accessing a given keyspace in a given cluster

  94. def setConf(cluster: String, options: Map[String, String]): CassandraSQLContext

    Permalink

    Set the Spark Cassandra Connector configuration parameters which will be used when accessing a given cluster

  95. def setConf(options: Map[String, String]): CassandraSQLContext

    Permalink

    Set the Spark Cassandra Connector configuration parameters

  96. def setConf(key: String, value: String): Unit

    Permalink
    Definition Classes
    SQLContext
  97. def setConf(props: Properties): Unit

    Permalink
    Definition Classes
    SQLContext
  98. def setDatabase(db: String): Unit

    Permalink

    Set current used database name.

    Set current used database name. Database is equivalent to keyspace

  99. def setKeyspace(ks: String): Unit

    Permalink

    Set default Cassandra keyspace to be used when accessing tables with unqualified names.

  100. val sparkContext: SparkContext

    Permalink
    Definition Classes
    SQLContext
  101. def sql(cassandraQuery: String): DataFrame

    Permalink

    Delegates to cassandraSql

    Delegates to cassandraSql

    Definition Classes
    CassandraSQLContext → SQLContext
  102. val sqlParser: SparkSQLParser

    Permalink
    Attributes
    protected[org.apache.spark.sql]
    Definition Classes
    SQLContext
  103. final def synchronized[T0](arg0: ⇒ T0): T0

    Permalink
    Definition Classes
    AnyRef
  104. def table(tableName: String): DataFrame

    Permalink
    Definition Classes
    SQLContext
  105. def tableNames(databaseName: String): Array[String]

    Permalink
    Definition Classes
    SQLContext
  106. def tableNames(): Array[String]

    Permalink
    Definition Classes
    SQLContext
  107. def tables(databaseName: String): DataFrame

    Permalink
    Definition Classes
    SQLContext
  108. def tables(): DataFrame

    Permalink
    Definition Classes
    SQLContext
  109. def toString(): String

    Permalink
    Definition Classes
    AnyRef → Any
  110. val udf: UDFRegistration

    Permalink
    Definition Classes
    SQLContext
  111. def uncacheTable(tableName: String): Unit

    Permalink
    Definition Classes
    SQLContext
  112. final def wait(): Unit

    Permalink
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  113. final def wait(arg0: Long, arg1: Int): Unit

    Permalink
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  114. final def wait(arg0: Long): Unit

    Permalink
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  115. def [B](y: B): (CassandraSQLContext, B)

    Permalink
    Implicit information
    This member is added by an implicit conversion from CassandraSQLContext to ArrowAssoc[CassandraSQLContext] performed by method ArrowAssoc in scala.Predef.
    Definition Classes
    ArrowAssoc

Deprecated Value Members

  1. def applySchema(rdd: JavaRDD[_], beanClass: Class[_]): DataFrame

    Permalink
    Definition Classes
    SQLContext
    Annotations
    @deprecated
    Deprecated

    (Since version 1.3.0) Use createDataFrame. This will be removed in Spark 2.0.

  2. def applySchema(rdd: RDD[_], beanClass: Class[_]): DataFrame

    Permalink
    Definition Classes
    SQLContext
    Annotations
    @deprecated
    Deprecated

    (Since version 1.3.0) Use createDataFrame. This will be removed in Spark 2.0.

  3. def applySchema(rowRDD: JavaRDD[Row], schema: StructType): DataFrame

    Permalink
    Definition Classes
    SQLContext
    Annotations
    @deprecated
    Deprecated

    (Since version 1.3.0) Use createDataFrame. This will be removed in Spark 2.0.

  4. def applySchema(rowRDD: RDD[Row], schema: StructType): DataFrame

    Permalink
    Definition Classes
    SQLContext
    Annotations
    @deprecated
    Deprecated

    (Since version 1.3.0) Use createDataFrame. This will be removed in Spark 2.0.

  5. def jdbc(url: String, table: String, theParts: Array[String]): DataFrame

    Permalink
    Definition Classes
    SQLContext
    Annotations
    @deprecated
    Deprecated

    (Since version 1.4.0) Use read.jdbc(). This will be removed in Spark 2.0.

  6. def jdbc(url: String, table: String, columnName: String, lowerBound: Long, upperBound: Long, numPartitions: Int): DataFrame

    Permalink
    Definition Classes
    SQLContext
    Annotations
    @deprecated
    Deprecated

    (Since version 1.4.0) Use read.jdbc(). This will be removed in Spark 2.0.

  7. def jdbc(url: String, table: String): DataFrame

    Permalink
    Definition Classes
    SQLContext
    Annotations
    @deprecated
    Deprecated

    (Since version 1.4.0) Use read.jdbc(). This will be removed in Spark 2.0.

  8. def jsonFile(path: String, samplingRatio: Double): DataFrame

    Permalink
    Definition Classes
    SQLContext
    Annotations
    @deprecated
    Deprecated

    (Since version 1.4.0) Use read.json(). This will be removed in Spark 2.0.

  9. def jsonFile(path: String, schema: StructType): DataFrame

    Permalink
    Definition Classes
    SQLContext
    Annotations
    @deprecated
    Deprecated

    (Since version 1.4.0) Use read.json(). This will be removed in Spark 2.0.

  10. def jsonFile(path: String): DataFrame

    Permalink
    Definition Classes
    SQLContext
    Annotations
    @deprecated
    Deprecated

    (Since version 1.4.0) Use read.json(). This will be removed in Spark 2.0.

  11. def jsonRDD(json: JavaRDD[String], samplingRatio: Double): DataFrame

    Permalink
    Definition Classes
    SQLContext
    Annotations
    @deprecated
    Deprecated

    (Since version 1.4.0) Use read.json(). This will be removed in Spark 2.0.

  12. def jsonRDD(json: RDD[String], samplingRatio: Double): DataFrame

    Permalink
    Definition Classes
    SQLContext
    Annotations
    @deprecated
    Deprecated

    (Since version 1.4.0) Use read.json(). This will be removed in Spark 2.0.

  13. def jsonRDD(json: JavaRDD[String], schema: StructType): DataFrame

    Permalink
    Definition Classes
    SQLContext
    Annotations
    @deprecated
    Deprecated

    (Since version 1.4.0) Use read.json(). This will be removed in Spark 2.0.

  14. def jsonRDD(json: RDD[String], schema: StructType): DataFrame

    Permalink
    Definition Classes
    SQLContext
    Annotations
    @deprecated
    Deprecated

    (Since version 1.4.0) Use read.json(). This will be removed in Spark 2.0.

  15. def jsonRDD(json: JavaRDD[String]): DataFrame

    Permalink
    Definition Classes
    SQLContext
    Annotations
    @deprecated
    Deprecated

    (Since version 1.4.0) Use read.json(). This will be removed in Spark 2.0.

  16. def jsonRDD(json: RDD[String]): DataFrame

    Permalink
    Definition Classes
    SQLContext
    Annotations
    @deprecated
    Deprecated

    (Since version 1.4.0) Use read.json(). This will be removed in Spark 2.0.

  17. def load(source: String, schema: StructType, options: Map[String, String]): DataFrame

    Permalink
    Definition Classes
    SQLContext
    Annotations
    @deprecated
    Deprecated

    (Since version 1.4.0) Use read.format(source).schema(schema).options(options).load(). This will be removed in Spark 2.0.

  18. def load(source: String, schema: StructType, options: Map[String, String]): DataFrame

    Permalink
    Definition Classes
    SQLContext
    Annotations
    @deprecated
    Deprecated

    (Since version 1.4.0) Use read.format(source).schema(schema).options(options).load(). This will be removed in Spark 2.0.

  19. def load(source: String, options: Map[String, String]): DataFrame

    Permalink
    Definition Classes
    SQLContext
    Annotations
    @deprecated
    Deprecated

    (Since version 1.4.0) Use read.format(source).options(options).load(). This will be removed in Spark 2.0.

  20. def load(source: String, options: Map[String, String]): DataFrame

    Permalink
    Definition Classes
    SQLContext
    Annotations
    @deprecated
    Deprecated

    (Since version 1.4.0) Use read.format(source).options(options).load(). This will be removed in Spark 2.0.

  21. def load(path: String, source: String): DataFrame

    Permalink
    Definition Classes
    SQLContext
    Annotations
    @deprecated
    Deprecated

    (Since version 1.4.0) Use read.format(source).load(path). This will be removed in Spark 2.0.

  22. def load(path: String): DataFrame

    Permalink
    Definition Classes
    SQLContext
    Annotations
    @deprecated
    Deprecated

    (Since version 1.4.0) Use read.load(path). This will be removed in Spark 2.0.

  23. def parquetFile(paths: String*): DataFrame

    Permalink
    Definition Classes
    SQLContext
    Annotations
    @deprecated @varargs()
    Deprecated

    (Since version 1.4.0) Use read.parquet(). This will be removed in Spark 2.0.

Inherited from SQLContext

Inherited from Serializable

Inherited from Serializable

Inherited from Logging

Inherited from AnyRef

Inherited from Any

Inherited by implicit conversion any2stringadd from CassandraSQLContext to any2stringadd[CassandraSQLContext]

Inherited by implicit conversion StringFormat from CassandraSQLContext to StringFormat[CassandraSQLContext]

Inherited by implicit conversion Ensuring from CassandraSQLContext to Ensuring[CassandraSQLContext]

Inherited by implicit conversion ArrowAssoc from CassandraSQLContext to ArrowAssoc[CassandraSQLContext]

Ungrouped