Packages

abstract class CassandraRDD[R] extends RDD[R]

Linear Supertypes
RDD[R], Logging, Serializable, Serializable, AnyRef, Any
Ordering
  1. Alphabetic
  2. By Inheritance
Inherited
  1. CassandraRDD
  2. RDD
  3. Logging
  4. Serializable
  5. Serializable
  6. AnyRef
  7. Any
  1. Hide All
  2. Show All
Visibility
  1. Public
  2. All

Instance Constructors

  1. new CassandraRDD(sc: SparkContext, dep: Seq[Dependency[_]])(implicit arg0: ClassTag[R])

Type Members

  1. abstract type Self <: CassandraRDD[R]

    This is slightly different than Scala this.type.

    This is slightly different than Scala this.type. this.type is the unique singleton type of an object which is not compatible with other instances of the same type, so returning anything other than this is not really possible without lying to the compiler by explicit casts. Here SelfType is used to return a copy of the object - a different instance of the same type

Abstract Value Members

  1. abstract def cassandraCount(): Long

    Counts the number of items in this RDD by selecting count(*) on Cassandra table

  2. abstract def clusteringOrder: Option[ClusteringOrder]
    Attributes
    protected
  3. abstract def columnNames: ColumnSelector
    Attributes
    protected
  4. abstract def compute(split: Partition, context: TaskContext): Iterator[R]
    Definition Classes
    RDD
    Annotations
    @DeveloperApi()
  5. abstract def connector: CassandraConnector
    Attributes
    protected
  6. abstract def copy(columnNames: ColumnSelector = columnNames, where: CqlWhereClause = where, limit: Option[CassandraLimit] = limit, clusteringOrder: Option[ClusteringOrder] = None, readConf: ReadConf = readConf, connector: CassandraConnector = connector): Self

    Allows to copy this RDD with changing some of the properties

    Allows to copy this RDD with changing some of the properties

    Attributes
    protected
  7. abstract def getPartitions: Array[Partition]
    Attributes
    protected
    Definition Classes
    RDD
  8. abstract def keyspaceName: String
    Attributes
    protected[spark.connector]
  9. abstract def limit: Option[CassandraLimit]
    Attributes
    protected
  10. abstract def narrowColumnSelection(columns: Seq[ColumnRef]): Seq[ColumnRef]
    Attributes
    protected
  11. abstract def readConf: ReadConf
    Attributes
    protected
  12. abstract val selectedColumnRefs: Seq[ColumnRef]
  13. abstract def tableName: String
    Attributes
    protected[spark.connector]
  14. abstract def toEmptyCassandraRDD: EmptyCassandraRDD[R]
  15. abstract def where: CqlWhereClause
    Attributes
    protected

Concrete Value Members

  1. final def !=(arg0: Any): Boolean
    Definition Classes
    AnyRef → Any
  2. final def ##(): Int
    Definition Classes
    AnyRef → Any
  3. def ++(other: RDD[R]): RDD[R]
    Definition Classes
    RDD
  4. final def ==(arg0: Any): Boolean
    Definition Classes
    AnyRef → Any
  5. def aggregate[U](zeroValue: U)(seqOp: (U, R) ⇒ U, combOp: (U, U) ⇒ U)(implicit arg0: ClassTag[U]): U
    Definition Classes
    RDD
  6. def as[B, A0, A1, A2, A3, A4, A5, A6, A7, A8, A9, A10, A11](f: (A0, A1, A2, A3, A4, A5, A6, A7, A8, A9, A10, A11) ⇒ B)(implicit arg0: ClassTag[B], arg1: TypeConverter[A0], arg2: TypeConverter[A1], arg3: TypeConverter[A2], arg4: TypeConverter[A3], arg5: TypeConverter[A4], arg6: TypeConverter[A5], arg7: TypeConverter[A6], arg8: TypeConverter[A7], arg9: TypeConverter[A8], arg10: TypeConverter[A9], arg11: TypeConverter[A10], arg12: TypeConverter[A11]): CassandraRDD[B]
  7. def as[B, A0, A1, A2, A3, A4, A5, A6, A7, A8, A9, A10](f: (A0, A1, A2, A3, A4, A5, A6, A7, A8, A9, A10) ⇒ B)(implicit arg0: ClassTag[B], arg1: TypeConverter[A0], arg2: TypeConverter[A1], arg3: TypeConverter[A2], arg4: TypeConverter[A3], arg5: TypeConverter[A4], arg6: TypeConverter[A5], arg7: TypeConverter[A6], arg8: TypeConverter[A7], arg9: TypeConverter[A8], arg10: TypeConverter[A9], arg11: TypeConverter[A10]): CassandraRDD[B]
  8. def as[B, A0, A1, A2, A3, A4, A5, A6, A7, A8, A9](f: (A0, A1, A2, A3, A4, A5, A6, A7, A8, A9) ⇒ B)(implicit arg0: ClassTag[B], arg1: TypeConverter[A0], arg2: TypeConverter[A1], arg3: TypeConverter[A2], arg4: TypeConverter[A3], arg5: TypeConverter[A4], arg6: TypeConverter[A5], arg7: TypeConverter[A6], arg8: TypeConverter[A7], arg9: TypeConverter[A8], arg10: TypeConverter[A9]): CassandraRDD[B]
  9. def as[B, A0, A1, A2, A3, A4, A5, A6, A7, A8](f: (A0, A1, A2, A3, A4, A5, A6, A7, A8) ⇒ B)(implicit arg0: ClassTag[B], arg1: TypeConverter[A0], arg2: TypeConverter[A1], arg3: TypeConverter[A2], arg4: TypeConverter[A3], arg5: TypeConverter[A4], arg6: TypeConverter[A5], arg7: TypeConverter[A6], arg8: TypeConverter[A7], arg9: TypeConverter[A8]): CassandraRDD[B]
  10. def as[B, A0, A1, A2, A3, A4, A5, A6, A7](f: (A0, A1, A2, A3, A4, A5, A6, A7) ⇒ B)(implicit arg0: ClassTag[B], arg1: TypeConverter[A0], arg2: TypeConverter[A1], arg3: TypeConverter[A2], arg4: TypeConverter[A3], arg5: TypeConverter[A4], arg6: TypeConverter[A5], arg7: TypeConverter[A6], arg8: TypeConverter[A7]): CassandraRDD[B]
  11. def as[B, A0, A1, A2, A3, A4, A5, A6](f: (A0, A1, A2, A3, A4, A5, A6) ⇒ B)(implicit arg0: ClassTag[B], arg1: TypeConverter[A0], arg2: TypeConverter[A1], arg3: TypeConverter[A2], arg4: TypeConverter[A3], arg5: TypeConverter[A4], arg6: TypeConverter[A5], arg7: TypeConverter[A6]): CassandraRDD[B]
  12. def as[B, A0, A1, A2, A3, A4, A5](f: (A0, A1, A2, A3, A4, A5) ⇒ B)(implicit arg0: ClassTag[B], arg1: TypeConverter[A0], arg2: TypeConverter[A1], arg3: TypeConverter[A2], arg4: TypeConverter[A3], arg5: TypeConverter[A4], arg6: TypeConverter[A5]): CassandraRDD[B]
  13. def as[B, A0, A1, A2, A3, A4](f: (A0, A1, A2, A3, A4) ⇒ B)(implicit arg0: ClassTag[B], arg1: TypeConverter[A0], arg2: TypeConverter[A1], arg3: TypeConverter[A2], arg4: TypeConverter[A3], arg5: TypeConverter[A4]): CassandraRDD[B]
  14. def as[B, A0, A1, A2, A3](f: (A0, A1, A2, A3) ⇒ B)(implicit arg0: ClassTag[B], arg1: TypeConverter[A0], arg2: TypeConverter[A1], arg3: TypeConverter[A2], arg4: TypeConverter[A3]): CassandraRDD[B]
  15. def as[B, A0, A1, A2](f: (A0, A1, A2) ⇒ B)(implicit arg0: ClassTag[B], arg1: TypeConverter[A0], arg2: TypeConverter[A1], arg3: TypeConverter[A2]): CassandraRDD[B]
  16. def as[B, A0, A1](f: (A0, A1) ⇒ B)(implicit arg0: ClassTag[B], arg1: TypeConverter[A0], arg2: TypeConverter[A1]): CassandraRDD[B]
  17. def as[B, A0](f: (A0) ⇒ B)(implicit arg0: ClassTag[B], arg1: TypeConverter[A0]): CassandraRDD[B]

    Maps each row into object of a different type using provided function taking column value(s) as argument(s).

    Maps each row into object of a different type using provided function taking column value(s) as argument(s). Can be used to convert each row to a tuple or a case class object:

    sc.cassandraTable("ks", "table")
      .select("column1")
      .as((s: String) => s)                 // yields CassandraRDD[String]
    
    sc.cassandraTable("ks", "table")
      .select("column1", "column2")
      .as((_: String, _: Long))             // yields CassandraRDD[(String, Long)]
    
    case class MyRow(key: String, value: Long)
    sc.cassandraTable("ks", "table")
      .select("column1", "column2")
      .as(MyRow)                            // yields CassandraRDD[MyRow]
  18. final def asInstanceOf[T0]: T0
    Definition Classes
    Any
  19. def barrier(): RDDBarrier[R]
    Definition Classes
    RDD
    Annotations
    @Experimental() @Since( "2.4.0" )
  20. def cache(): CassandraRDD.this.type
    Definition Classes
    RDD
  21. def cartesian[U](other: RDD[U])(implicit arg0: ClassTag[U]): RDD[(R, U)]
    Definition Classes
    RDD
  22. def checkpoint(): Unit
    Definition Classes
    RDD
  23. def clearDependencies(): Unit
    Attributes
    protected
    Definition Classes
    RDD
  24. def clone(): AnyRef
    Attributes
    protected[lang]
    Definition Classes
    AnyRef
    Annotations
    @throws( ... ) @native()
  25. def clusteringOrder(order: ClusteringOrder): Self

    Adds a CQL ORDER BY clause to the query.

    Adds a CQL ORDER BY clause to the query. It can be applied only in case there are clustering columns and primary key predicate is pushed down in where. It is useful when the default direction of ordering rows within a single Cassandra partition needs to be changed.

  26. def coalesce(numPartitions: Int, shuffle: Boolean, partitionCoalescer: Option[PartitionCoalescer])(implicit ord: Ordering[R]): RDD[R]
    Definition Classes
    RDD
  27. def collect[U](f: PartialFunction[R, U])(implicit arg0: ClassTag[U]): RDD[U]
    Definition Classes
    RDD
  28. def collect(): Array[R]
    Definition Classes
    RDD
  29. def context: SparkContext
    Definition Classes
    RDD
  30. def convertTo[B](implicit arg0: ClassTag[B], arg1: RowReaderFactory[B]): CassandraRDD[B]
    Attributes
    protected
  31. def count(): Long
    Definition Classes
    RDD
  32. def countApprox(timeout: Long, confidence: Double): PartialResult[BoundedDouble]
    Definition Classes
    RDD
  33. def countApproxDistinct(relativeSD: Double): Long
    Definition Classes
    RDD
  34. def countApproxDistinct(p: Int, sp: Int): Long
    Definition Classes
    RDD
  35. def countByValue()(implicit ord: Ordering[R]): Map[R, Long]
    Definition Classes
    RDD
  36. def countByValueApprox(timeout: Long, confidence: Double)(implicit ord: Ordering[R]): PartialResult[Map[R, BoundedDouble]]
    Definition Classes
    RDD
  37. final def dependencies: Seq[Dependency[_]]
    Definition Classes
    RDD
  38. def distinct(): RDD[R]
    Definition Classes
    RDD
  39. def distinct(numPartitions: Int)(implicit ord: Ordering[R]): RDD[R]
    Definition Classes
    RDD
  40. final def eq(arg0: AnyRef): Boolean
    Definition Classes
    AnyRef
  41. def equals(arg0: Any): Boolean
    Definition Classes
    AnyRef → Any
  42. def filter(f: (R) ⇒ Boolean): RDD[R]
    Definition Classes
    RDD
  43. def finalize(): Unit
    Attributes
    protected[lang]
    Definition Classes
    AnyRef
    Annotations
    @throws( classOf[java.lang.Throwable] )
  44. def first(): R
    Definition Classes
    RDD
  45. def firstParent[U](implicit arg0: ClassTag[U]): RDD[U]
    Attributes
    protected[org.apache.spark]
    Definition Classes
    RDD
  46. def flatMap[U](f: (R) ⇒ TraversableOnce[U])(implicit arg0: ClassTag[U]): RDD[U]
    Definition Classes
    RDD
  47. def fold(zeroValue: R)(op: (R, R) ⇒ R): R
    Definition Classes
    RDD
  48. def foreach(f: (R) ⇒ Unit): Unit
    Definition Classes
    RDD
  49. def foreachPartition(f: (Iterator[R]) ⇒ Unit): Unit
    Definition Classes
    RDD
  50. def getCheckpointFile: Option[String]
    Definition Classes
    RDD
  51. final def getClass(): Class[_]
    Definition Classes
    AnyRef → Any
    Annotations
    @native()
  52. def getDependencies: Seq[Dependency[_]]
    Attributes
    protected
    Definition Classes
    RDD
  53. final def getNumPartitions: Int
    Definition Classes
    RDD
    Annotations
    @Since( "1.6.0" )
  54. def getOutputDeterministicLevel: org.apache.spark.rdd.DeterministicLevel.Value
    Attributes
    protected
    Definition Classes
    RDD
    Annotations
    @DeveloperApi()
  55. def getPreferredLocations(split: Partition): Seq[String]
    Attributes
    protected
    Definition Classes
    RDD
  56. def getStorageLevel: StorageLevel
    Definition Classes
    RDD
  57. def glom(): RDD[Array[R]]
    Definition Classes
    RDD
  58. def groupBy[K](f: (R) ⇒ K, p: Partitioner)(implicit kt: ClassTag[K], ord: Ordering[K]): RDD[(K, Iterable[R])]
    Definition Classes
    RDD
  59. def groupBy[K](f: (R) ⇒ K, numPartitions: Int)(implicit kt: ClassTag[K]): RDD[(K, Iterable[R])]
    Definition Classes
    RDD
  60. def groupBy[K](f: (R) ⇒ K)(implicit kt: ClassTag[K]): RDD[(K, Iterable[R])]
    Definition Classes
    RDD
  61. def hashCode(): Int
    Definition Classes
    AnyRef → Any
    Annotations
    @native()
  62. val id: Int
    Definition Classes
    RDD
  63. def initializeLogIfNecessary(isInterpreter: Boolean, silent: Boolean): Boolean
    Attributes
    protected
    Definition Classes
    Logging
  64. def initializeLogIfNecessary(isInterpreter: Boolean): Unit
    Attributes
    protected
    Definition Classes
    Logging
  65. def intersection(other: RDD[R], numPartitions: Int): RDD[R]
    Definition Classes
    RDD
  66. def intersection(other: RDD[R], partitioner: Partitioner)(implicit ord: Ordering[R]): RDD[R]
    Definition Classes
    RDD
  67. def intersection(other: RDD[R]): RDD[R]
    Definition Classes
    RDD
  68. lazy val isBarrier_: Boolean
    Attributes
    protected
    Definition Classes
    RDD
    Annotations
    @transient()
  69. def isCheckpointed: Boolean
    Definition Classes
    RDD
  70. def isEmpty(): Boolean
    Definition Classes
    RDD
  71. final def isInstanceOf[T0]: Boolean
    Definition Classes
    Any
  72. def isTraceEnabled(): Boolean
    Attributes
    protected
    Definition Classes
    Logging
  73. final def iterator(split: Partition, context: TaskContext): Iterator[R]
    Definition Classes
    RDD
  74. def keyBy[K](f: (R) ⇒ K): RDD[(K, R)]
    Definition Classes
    RDD
  75. def limit(rowLimit: Long): Self

    Adds the limit clause to CQL select statement.

    Adds the limit clause to CQL select statement. The limit will be applied for each created Spark partition. In other words, unless the data are fetched from a single Cassandra partition the number of results is unpredictable.

    The main purpose of passing limit clause is to fetch top n rows from a single Cassandra partition when the table is designed so that it uses clustering keys and a partition key predicate is passed to the where clause.

  76. def localCheckpoint(): CassandraRDD.this.type
    Definition Classes
    RDD
  77. def log: Logger
    Attributes
    protected
    Definition Classes
    Logging
  78. def logDebug(msg: ⇒ String, throwable: Throwable): Unit
    Attributes
    protected
    Definition Classes
    Logging
  79. def logDebug(msg: ⇒ String): Unit
    Attributes
    protected
    Definition Classes
    Logging
  80. def logError(msg: ⇒ String, throwable: Throwable): Unit
    Attributes
    protected
    Definition Classes
    Logging
  81. def logError(msg: ⇒ String): Unit
    Attributes
    protected
    Definition Classes
    Logging
  82. def logInfo(msg: ⇒ String, throwable: Throwable): Unit
    Attributes
    protected
    Definition Classes
    Logging
  83. def logInfo(msg: ⇒ String): Unit
    Attributes
    protected
    Definition Classes
    Logging
  84. def logName: String
    Attributes
    protected
    Definition Classes
    Logging
  85. def logTrace(msg: ⇒ String, throwable: Throwable): Unit
    Attributes
    protected
    Definition Classes
    Logging
  86. def logTrace(msg: ⇒ String): Unit
    Attributes
    protected
    Definition Classes
    Logging
  87. def logWarning(msg: ⇒ String, throwable: Throwable): Unit
    Attributes
    protected
    Definition Classes
    Logging
  88. def logWarning(msg: ⇒ String): Unit
    Attributes
    protected
    Definition Classes
    Logging
  89. def map[U](f: (R) ⇒ U)(implicit arg0: ClassTag[U]): RDD[U]
    Definition Classes
    RDD
  90. def mapPartitions[U](f: (Iterator[R]) ⇒ Iterator[U], preservesPartitioning: Boolean)(implicit arg0: ClassTag[U]): RDD[U]
    Definition Classes
    RDD
  91. def mapPartitionsWithIndex[U](f: (Int, Iterator[R]) ⇒ Iterator[U], preservesPartitioning: Boolean)(implicit arg0: ClassTag[U]): RDD[U]
    Definition Classes
    RDD
  92. def max()(implicit ord: Ordering[R]): R
    Definition Classes
    RDD
  93. def min()(implicit ord: Ordering[R]): R
    Definition Classes
    RDD
  94. var name: String
    Definition Classes
    RDD
  95. final def ne(arg0: AnyRef): Boolean
    Definition Classes
    AnyRef
  96. final def notify(): Unit
    Definition Classes
    AnyRef
    Annotations
    @native()
  97. final def notifyAll(): Unit
    Definition Classes
    AnyRef
    Annotations
    @native()
  98. def parent[U](j: Int)(implicit arg0: ClassTag[U]): RDD[U]
    Attributes
    protected[org.apache.spark]
    Definition Classes
    RDD
  99. val partitioner: Option[Partitioner]
    Definition Classes
    RDD
  100. final def partitions: Array[Partition]
    Definition Classes
    RDD
  101. def perPartitionLimit(rowLimit: Long): Self

    Adds the PER PARTITION LIMIT clause to CQL select statement.

    Adds the PER PARTITION LIMIT clause to CQL select statement. The limit will be applied for every Cassandra Partition. Only Valid For Cassandra 3.6+

  102. def persist(): CassandraRDD.this.type
    Definition Classes
    RDD
  103. def persist(newLevel: StorageLevel): CassandraRDD.this.type
    Definition Classes
    RDD
  104. def pipe(command: Seq[String], env: Map[String, String], printPipeContext: ((String) ⇒ Unit) ⇒ Unit, printRDDElement: (R, (String) ⇒ Unit) ⇒ Unit, separateWorkingDir: Boolean, bufferSize: Int, encoding: String): RDD[String]
    Definition Classes
    RDD
  105. def pipe(command: String, env: Map[String, String]): RDD[String]
    Definition Classes
    RDD
  106. def pipe(command: String): RDD[String]
    Definition Classes
    RDD
  107. final def preferredLocations(split: Partition): Seq[String]
    Definition Classes
    RDD
  108. def randomSplit(weights: Array[Double], seed: Long): Array[RDD[R]]
    Definition Classes
    RDD
  109. def reduce(f: (R, R) ⇒ R): R
    Definition Classes
    RDD
  110. def repartition(numPartitions: Int)(implicit ord: Ordering[R]): RDD[R]
    Definition Classes
    RDD
  111. def sample(withReplacement: Boolean, fraction: Double, seed: Long): RDD[R]
    Definition Classes
    RDD
  112. def saveAsObjectFile(path: String): Unit
    Definition Classes
    RDD
  113. def saveAsTextFile(path: String, codec: Class[_ <: CompressionCodec]): Unit
    Definition Classes
    RDD
  114. def saveAsTextFile(path: String): Unit
    Definition Classes
    RDD
  115. def select(columns: ColumnRef*): Self

    Narrows down the selected set of columns.

    Narrows down the selected set of columns. Use this for better performance, when you don't need all the columns in the result RDD. When called multiple times, it selects the subset of the already selected columns, so after a column was removed by the previous select call, it is not possible to add it back.

    The selected columns are ColumnRef instances. This type allows to specify columns for straightforward retrieval and to read TTL or write time of regular columns as well. Implicit conversions included in com.datastax.spark.connector package make it possible to provide just column names (which is also backward compatible) and optional add .ttl or .writeTime suffix in order to create an appropriate ColumnRef instance.

  116. def selectedColumnNames: Seq[String]
  117. def setName(_name: String): CassandraRDD.this.type
    Definition Classes
    RDD
  118. def sortBy[K](f: (R) ⇒ K, ascending: Boolean, numPartitions: Int)(implicit ord: Ordering[K], ctag: ClassTag[K]): RDD[R]
    Definition Classes
    RDD
  119. def sparkContext: SparkContext
    Definition Classes
    RDD
  120. def subtract(other: RDD[R], p: Partitioner)(implicit ord: Ordering[R]): RDD[R]
    Definition Classes
    RDD
  121. def subtract(other: RDD[R], numPartitions: Int): RDD[R]
    Definition Classes
    RDD
  122. def subtract(other: RDD[R]): RDD[R]
    Definition Classes
    RDD
  123. final def synchronized[T0](arg0: ⇒ T0): T0
    Definition Classes
    AnyRef
  124. def take(num: Int): Array[R]
    Definition Classes
    CassandraRDD → RDD
  125. def takeOrdered(num: Int)(implicit ord: Ordering[R]): Array[R]
    Definition Classes
    RDD
  126. def takeSample(withReplacement: Boolean, num: Int, seed: Long): Array[R]
    Definition Classes
    RDD
  127. def toDebugString: String
    Definition Classes
    RDD
  128. def toJavaRDD(): JavaRDD[R]
    Definition Classes
    RDD
  129. def toLocalIterator: Iterator[R]
    Definition Classes
    RDD
  130. def toString(): String
    Definition Classes
    RDD → AnyRef → Any
  131. def top(num: Int)(implicit ord: Ordering[R]): Array[R]
    Definition Classes
    RDD
  132. def treeAggregate[U](zeroValue: U)(seqOp: (U, R) ⇒ U, combOp: (U, U) ⇒ U, depth: Int)(implicit arg0: ClassTag[U]): U
    Definition Classes
    RDD
  133. def treeReduce(f: (R, R) ⇒ R, depth: Int): R
    Definition Classes
    RDD
  134. def union(other: RDD[R]): RDD[R]
    Definition Classes
    RDD
  135. def unpersist(blocking: Boolean): CassandraRDD.this.type
    Definition Classes
    RDD
  136. final def wait(): Unit
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  137. final def wait(arg0: Long, arg1: Int): Unit
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  138. final def wait(arg0: Long): Unit
    Definition Classes
    AnyRef
    Annotations
    @throws( ... ) @native()
  139. def where(cql: String, values: Any*): Self

    Adds a CQL WHERE predicate(s) to the query.

    Adds a CQL WHERE predicate(s) to the query. Useful for leveraging secondary indexes in Cassandra. Implicitly adds an ALLOW FILTERING clause to the WHERE clause, however beware that some predicates might be rejected by Cassandra, particularly in cases when they filter on an unindexed, non-clustering column.

  140. def withAscOrder: Self
  141. def withConnector(connector: CassandraConnector): Self

    Returns a copy of this Cassandra RDD with specified connector

  142. def withDescOrder: Self
  143. def withReadConf(readConf: ReadConf): Self

    Allows to set custom read configuration, e.g.

    Allows to set custom read configuration, e.g. consistency level or fetch size.

  144. def zip[U](other: RDD[U])(implicit arg0: ClassTag[U]): RDD[(R, U)]
    Definition Classes
    RDD
  145. def zipPartitions[B, C, D, V](rdd2: RDD[B], rdd3: RDD[C], rdd4: RDD[D])(f: (Iterator[R], Iterator[B], Iterator[C], Iterator[D]) ⇒ Iterator[V])(implicit arg0: ClassTag[B], arg1: ClassTag[C], arg2: ClassTag[D], arg3: ClassTag[V]): RDD[V]
    Definition Classes
    RDD
  146. def zipPartitions[B, C, D, V](rdd2: RDD[B], rdd3: RDD[C], rdd4: RDD[D], preservesPartitioning: Boolean)(f: (Iterator[R], Iterator[B], Iterator[C], Iterator[D]) ⇒ Iterator[V])(implicit arg0: ClassTag[B], arg1: ClassTag[C], arg2: ClassTag[D], arg3: ClassTag[V]): RDD[V]
    Definition Classes
    RDD
  147. def zipPartitions[B, C, V](rdd2: RDD[B], rdd3: RDD[C])(f: (Iterator[R], Iterator[B], Iterator[C]) ⇒ Iterator[V])(implicit arg0: ClassTag[B], arg1: ClassTag[C], arg2: ClassTag[V]): RDD[V]
    Definition Classes
    RDD
  148. def zipPartitions[B, C, V](rdd2: RDD[B], rdd3: RDD[C], preservesPartitioning: Boolean)(f: (Iterator[R], Iterator[B], Iterator[C]) ⇒ Iterator[V])(implicit arg0: ClassTag[B], arg1: ClassTag[C], arg2: ClassTag[V]): RDD[V]
    Definition Classes
    RDD
  149. def zipPartitions[B, V](rdd2: RDD[B])(f: (Iterator[R], Iterator[B]) ⇒ Iterator[V])(implicit arg0: ClassTag[B], arg1: ClassTag[V]): RDD[V]
    Definition Classes
    RDD
  150. def zipPartitions[B, V](rdd2: RDD[B], preservesPartitioning: Boolean)(f: (Iterator[R], Iterator[B]) ⇒ Iterator[V])(implicit arg0: ClassTag[B], arg1: ClassTag[V]): RDD[V]
    Definition Classes
    RDD
  151. def zipWithIndex(): RDD[(R, Long)]
    Definition Classes
    RDD
  152. def zipWithUniqueId(): RDD[(R, Long)]
    Definition Classes
    RDD

Inherited from RDD[R]

Inherited from Logging

Inherited from Serializable

Inherited from Serializable

Inherited from AnyRef

Inherited from Any

Ungrouped