| Method and Description | 
|---|
| org.apache.spark.streaming.StreamingContext.awaitTermination(long) As of 1.3.0, replaced by  awaitTerminationOrTimeout(Long). | 
| org.apache.spark.streaming.api.java.JavaStreamingContext.awaitTermination(long) As of 1.3.0, replaced by  awaitTerminationOrTimeout(Long). | 
| org.apache.spark.sql.functions.callUDF(Function0>, DataType) As of 1.5.0, since it's redundant with udf()
              This will be removed in Spark 2.0. | 
| org.apache.spark.sql.functions.callUDF(Function1, ?>, DataType, Column) As of 1.5.0, since it's redundant with udf()
              This will be removed in Spark 2.0. | 
| org.apache.spark.sql.functions.callUDF(Function10, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?>, DataType, Column, Column, Column, Column, Column, Column, Column, Column, Column, Column) As of 1.5.0, since it's redundant with udf().
              This will be removed in Spark 2.0. | 
| org.apache.spark.sql.functions.callUDF(Function2, ?, ?>, DataType, Column, Column) As of 1.5.0, since it's redundant with udf()
              This will be removed in Spark 2.0. | 
| org.apache.spark.sql.functions.callUDF(Function3, ?, ?, ?>, DataType, Column, Column, Column) As of 1.5.0, since it's redundant with udf()
              This will be removed in Spark 2.0. | 
| org.apache.spark.sql.functions.callUDF(Function4, ?, ?, ?, ?>, DataType, Column, Column, Column, Column) As of 1.5.0, since it's redundant with udf()
              This will be removed in Spark 2.0. | 
| org.apache.spark.sql.functions.callUDF(Function5, ?, ?, ?, ?, ?>, DataType, Column, Column, Column, Column, Column) As of 1.5.0, since it's redundant with udf()
              This will be removed in Spark 2.0. | 
| org.apache.spark.sql.functions.callUDF(Function6, ?, ?, ?, ?, ?, ?>, DataType, Column, Column, Column, Column, Column, Column) As of 1.5.0, since it's redundant with udf()
              This will be removed in Spark 2.0. | 
| org.apache.spark.sql.functions.callUDF(Function7, ?, ?, ?, ?, ?, ?, ?>, DataType, Column, Column, Column, Column, Column, Column, Column) As of 1.5.0, since it's redundant with udf()
              This will be removed in Spark 2.0. | 
| org.apache.spark.sql.functions.callUDF(Function8, ?, ?, ?, ?, ?, ?, ?, ?>, DataType, Column, Column, Column, Column, Column, Column, Column, Column) As of 1.5.0, since it's redundant with udf()
              This will be removed in Spark 2.0. | 
| org.apache.spark.sql.functions.callUDF(Function9, ?, ?, ?, ?, ?, ?, ?, ?, ?>, DataType, Column, Column, Column, Column, Column, Column, Column, Column, Column) As of 1.5.0, since it's redundant with udf().
              This will be removed in Spark 2.0. | 
| org.apache.spark.sql.functions.callUdf(String, Seq As of 1.5.0, since it was not coherent to have two functions callUdf and callUDF.
             This will be removed in Spark 2.0. | 
| org.apache.spark.api.java.StorageLevels.create(boolean, boolean, boolean, int) | 
| org.apache.spark.sql.DataFrame.createJDBCTable(String, String, boolean) As of 1.340, replaced by  write().jdbc(). This will be removed in Spark 2.0. | 
| org.apache.spark.sql.functions.cumeDist() As of 1.6.0, replaced by  cume_dist. This will be removed in Spark 2.0. | 
| org.apache.spark.api.java.JavaSparkContext.defaultMinSplits() As of Spark 1.0.0, defaultMinSplits is deprecated, use
             JavaSparkContext.defaultMinPartitions()instead | 
| org.apache.spark.sql.functions.denseRank() As of 1.6.0, replaced by  dense_rank. This will be removed in Spark 2.0. | 
| org.apache.spark.streaming.api.java.JavaDStreamLike.foreach(Function As of release 0.9.0, replaced by foreachRDD | 
| org.apache.spark.streaming.dstream.DStream.foreach(Function1 As of 0.9.0, replaced by  foreachRDD. | 
| org.apache.spark.streaming.dstream.DStream.foreach(Function2 As of 0.9.0, replaced by  foreachRDD. | 
| org.apache.spark.streaming.api.java.JavaDStreamLike.foreach(Function2 As of release 0.9.0, replaced by foreachRDD | 
| org.apache.spark.streaming.api.java.JavaDStreamLike.foreachRDD(Function As of release 1.6.0, replaced by foreachRDD(JVoidFunction) | 
| org.apache.spark.streaming.api.java.JavaDStreamLike.foreachRDD(Function2 As of release 1.6.0, replaced by foreachRDD(JVoidFunction2) | 
| org.apache.spark.sql.types.DataType.fromCaseClassString(String) As of 1.2.0, replaced by  DataType.fromJson() | 
| org.apache.spark.streaming.api.java.JavaStreamingContext.getOrCreate(String, Configuration, JavaStreamingContextFactory) As of 1.4.0, replaced by  getOrCreatewithout JavaStreamingContextFactory. | 
| org.apache.spark.streaming.api.java.JavaStreamingContext.getOrCreate(String, Configuration, JavaStreamingContextFactory, boolean) As of 1.4.0, replaced by  getOrCreatewithout JavaStreamingContextFactory. | 
| org.apache.spark.streaming.api.java.JavaStreamingContext.getOrCreate(String, JavaStreamingContextFactory) As of 1.4.0, replaced by  getOrCreatewithout JavaStreamingContextFactory. | 
| org.apache.spark.sql.Column.in(Object...) As of 1.5.0. Use isin. This will be removed in Spark 2.0. | 
| org.apache.spark.sql.Column.in(Seq As of 1.5.0. Use isin. This will be removed in Spark 2.0. | 
| org.apache.spark.sql.functions.inputFileName() As of 1.6.0, replaced by  input_file_name. This will be removed in Spark 2.0. | 
| org.apache.spark.sql.DataFrame.insertInto(String) As of 1.4.0, replaced by
             write().mode(SaveMode.Append).saveAsTable(tableName).
             This will be removed in Spark 2.0. | 
| org.apache.spark.sql.DataFrame.insertInto(String, boolean) As of 1.4.0, replaced by
             write().mode(SaveMode.Append|SaveMode.Overwrite).saveAsTable(tableName).
             This will be removed in Spark 2.0. | 
| org.apache.spark.sql.DataFrame.insertIntoJDBC(String, String, boolean) As of 1.4.0, replaced by  write().jdbc(). This will be removed in Spark 2.0. | 
| org.apache.spark.sql.functions.isNaN(Column) As of 1.6.0, replaced by  isnan. This will be removed in Spark 2.0. | 
| org.apache.spark.sql.SQLContext.jdbc(String, String) As of 1.4.0, replaced by  read().jdbc(). This will be removed in Spark 2.0. | 
| org.apache.spark.sql.SQLContext.jdbc(String, String, String[]) As of 1.4.0, replaced by  read().jdbc(). This will be removed in Spark 2.0. | 
| org.apache.spark.sql.SQLContext.jdbc(String, String, String, long, long, int) As of 1.4.0, replaced by  read().jdbc(). This will be removed in Spark 2.0. | 
| org.apache.spark.sql.SQLContext.jsonFile(String) As of 1.4.0, replaced by  read().json(). This will be removed in Spark 2.0. | 
| org.apache.spark.sql.SQLContext.jsonFile(String, double) As of 1.4.0, replaced by  read().json(). This will be removed in Spark 2.0. | 
| org.apache.spark.sql.SQLContext.jsonFile(String, StructType) As of 1.4.0, replaced by  read().json(). This will be removed in Spark 2.0. | 
| org.apache.spark.sql.SQLContext.jsonRDD(JavaRDD As of 1.4.0, replaced by  read().json(). This will be removed in Spark 2.0. | 
| org.apache.spark.sql.SQLContext.jsonRDD(JavaRDD As of 1.4.0, replaced by  read().json(). This will be removed in Spark 2.0. | 
| org.apache.spark.sql.SQLContext.jsonRDD(JavaRDD As of 1.4.0, replaced by  read().json(). This will be removed in Spark 2.0. | 
| org.apache.spark.sql.SQLContext.jsonRDD(RDD As of 1.4.0, replaced by  read().json(). This will be removed in Spark 2.0. | 
| org.apache.spark.sql.SQLContext.jsonRDD(RDD As of 1.4.0, replaced by  read().json(). This will be removed in Spark 2.0. | 
| org.apache.spark.sql.SQLContext.jsonRDD(RDD As of 1.4.0, replaced by  read().json(). This will be removed in Spark 2.0. | 
| org.apache.spark.sql.SQLContext.load(String) As of 1.4.0, replaced by  read().load(path). This will be removed in Spark 2.0. | 
| org.apache.spark.sql.SQLContext.load(String, Map As of 1.4.0, replaced by  read().format(source).options(options).load().
             This will be removed in Spark 2.0. | 
| org.apache.spark.sql.SQLContext.load(String, Map As of 1.4.0, replaced by  read().format(source).options(options).load(). | 
| org.apache.spark.sql.SQLContext.load(String, String) As of 1.4.0, replaced by  read().format(source).load(path).
             This will be removed in Spark 2.0. | 
| org.apache.spark.sql.SQLContext.load(String, StructType, Map As of 1.4.0, replaced by
             read().format(source).schema(schema).options(options).load(). | 
| org.apache.spark.sql.SQLContext.load(String, StructType, Map As of 1.4.0, replaced by
             read().format(source).schema(schema).options(options).load(). | 
| org.apache.spark.mllib.util.MLUtils.loadLabeledData(SparkContext, String) Should use  RDD.saveAsTextFile(java.lang.String)for saving andMLUtils.loadLabeledPoints(org.apache.spark.SparkContext, java.lang.String, int)for loading. | 
| org.apache.spark.streaming.StreamingContext.networkStream(Receiver As of 1.0.0", replaced by  receiverStream. | 
| org.apache.spark.sql.SQLContext.parquetFile(String...) As of 1.4.0, replaced by  read().parquet(). This will be removed in Spark 2.0. | 
| org.apache.spark.sql.functions.percentRank() As of 1.6.0, replaced by  percent_rank. This will be removed in Spark 2.0. | 
| org.apache.spark.streaming.api.java.JavaDStreamLike.reduceByWindow(Function2 As this API is not Java compatible. | 
| org.apache.spark.sql.functions.rowNumber() As of 1.6.0, replaced by  row_number. This will be removed in Spark 2.0. | 
| org.apache.spark.sql.DataFrame.save(String) As of 1.4.0, replaced by  write().save(path). This will be removed in Spark 2.0. | 
| org.apache.spark.sql.DataFrame.save(String, SaveMode) As of 1.4.0, replaced by  write().mode(mode).save(path).
             This will be removed in Spark 2.0. | 
| org.apache.spark.sql.DataFrame.save(String, SaveMode, Map As of 1.4.0, replaced by
             write().format(source).mode(mode).options(options).save(path).
             This will be removed in Spark 2.0. | 
| org.apache.spark.sql.DataFrame.save(String, SaveMode, Map As of 1.4.0, replaced by
             write().format(source).mode(mode).options(options).save(path).
             This will be removed in Spark 2.0. | 
| org.apache.spark.sql.DataFrame.save(String, String) As of 1.4.0, replaced by  write().format(source).save(path).
             This will be removed in Spark 2.0. | 
| org.apache.spark.sql.DataFrame.save(String, String, SaveMode) As of 1.4.0, replaced by  write().format(source).mode(mode).save(path).
             This will be removed in Spark 2.0. | 
| org.apache.spark.sql.DataFrame.saveAsParquetFile(String) As of 1.4.0, replaced by  write().parquet(). This will be removed in Spark 2.0. | 
| org.apache.spark.sql.DataFrame.saveAsTable(String) As of 1.4.0, replaced by  write().saveAsTable(tableName).
             This will be removed in Spark 2.0. | 
| org.apache.spark.sql.DataFrame.saveAsTable(String, SaveMode) As of 1.4.0, replaced by  write().mode(mode).saveAsTable(tableName).
              This will be removed in Spark 2.0. | 
| org.apache.spark.sql.DataFrame.saveAsTable(String, String) As of 1.4.0, replaced by  write().format(source).saveAsTable(tableName).
             This will be removed in Spark 2.0. | 
| org.apache.spark.sql.DataFrame.saveAsTable(String, String, SaveMode) As of 1.4.0, replaced by  write().mode(mode).saveAsTable(tableName).
             This will be removed in Spark 2.0. | 
| org.apache.spark.sql.DataFrame.saveAsTable(String, String, SaveMode, Map As of 1.4.0, replaced by
             write().format(source).mode(mode).options(options).saveAsTable(tableName).
             This will be removed in Spark 2.0. | 
| org.apache.spark.sql.DataFrame.saveAsTable(String, String, SaveMode, Map As of 1.4.0, replaced by
             write().format(source).mode(mode).options(options).saveAsTable(tableName).
             This will be removed in Spark 2.0. | 
| org.apache.spark.mllib.util.MLUtils.saveLabeledData(RDD Should use  RDD.saveAsTextFile(java.lang.String)for saving andMLUtils.loadLabeledPoints(org.apache.spark.SparkContext, java.lang.String, int)for loading. | 
| org.apache.spark.streaming.api.java.JavaStreamingContext.sc() As of 0.9.0, replaced by  sparkContext | 
| org.apache.spark.mllib.optimization.LBFGS.setMaxNumIterations(int) use  LBFGS.setNumIterations(int)instead | 
| org.apache.spark.ml.evaluation.BinaryClassificationEvaluator.setScoreCol(String) use  setRawPredictionCol()instead | 
| org.apache.spark.sql.functions.sparkPartitionId() As of 1.6.0, replaced by  spark_partition_id. This will be removed in Spark 2.0. | 
| org.apache.spark.api.java.JavaRDDLike.toArray() As of Spark 1.0.0, toArray() is deprecated, use  JavaRDDLike.collect()instead | 
| org.apache.spark.streaming.StreamingContext.toPairDStreamFunctions(DStream As of 1.3.0, replaced by implicit functions in the DStream companion object.
             This is kept here only for backward compatibility. | 
| org.apache.spark.sql.DataFrame.toSchemaRDD() As of 1.3.0, replaced by  toDF(). This will be removed in Spark 2.0. | 
| org.apache.spark.mllib.rdd.RDDFunctions.treeAggregate(U, Function2, Function2, int, ClassTag) | 
| org.apache.spark.mllib.rdd.RDDFunctions.treeReduce(Function2 Use  RDD.treeReduce(scala.Function2<T, T, T>, int)instead. |