(app-name)(app-name spark)Params:
Result: String
Source: https://spark.apache.org/docs/3.0.1/api/scala/org/apache/spark/api/java/JavaSparkContext.html
Timestamp: 2020-10-19T01:56:49.487Z
Params: Result: String Source: https://spark.apache.org/docs/3.0.1/api/scala/org/apache/spark/api/java/JavaSparkContext.html Timestamp: 2020-10-19T01:56:49.487Z
Params: (path: String, minPartitions: Int)
Result: JavaPairRDD[String, PortableDataStream]
Read a directory of binary files from HDFS, a local file system (available on all nodes), or any Hadoop-supported file system URI as a byte array. Each file is read as a single record and returned in a key-value pair, where the key is the path of each file, the value is the content of each file.
For example, if you have the following files:
Do
then rdd contains
A suggestion value of the minimal splitting number for input data.
Small files are preferred; very large files but may cause bad performance.
Source: https://spark.apache.org/docs/3.0.1/api/scala/org/apache/spark/api/java/JavaSparkContext.html
Timestamp: 2020-10-19T01:56:49.492Z
Params: (path: String, minPartitions: Int) Result: JavaPairRDD[String, PortableDataStream] Read a directory of binary files from HDFS, a local file system (available on all nodes), or any Hadoop-supported file system URI as a byte array. Each file is read as a single record and returned in a key-value pair, where the key is the path of each file, the value is the content of each file. For example, if you have the following files: Do then rdd contains A suggestion value of the minimal splitting number for input data. Small files are preferred; very large files but may cause bad performance. Source: https://spark.apache.org/docs/3.0.1/api/scala/org/apache/spark/api/java/JavaSparkContext.html Timestamp: 2020-10-19T01:56:49.492Z
(broadcast value)(broadcast spark value)Params: (value: T)
Result: Broadcast[T]
Broadcast a read-only variable to the cluster, returning a org.apache.spark.broadcast.Broadcast object for reading it in distributed functions. The variable will be sent to each cluster only once.
Source: https://spark.apache.org/docs/3.0.1/api/scala/org/apache/spark/api/java/JavaSparkContext.html
Timestamp: 2020-10-19T01:56:49.495Z
Params: (value: T) Result: Broadcast[T] Broadcast a read-only variable to the cluster, returning a org.apache.spark.broadcast.Broadcast object for reading it in distributed functions. The variable will be sent to each cluster only once. Source: https://spark.apache.org/docs/3.0.1/api/scala/org/apache/spark/api/java/JavaSparkContext.html Timestamp: 2020-10-19T01:56:49.495Z
(checkpoint-dir)(checkpoint-dir spark)Params:
Result: Optional[String]
Source: https://spark.apache.org/docs/3.0.1/api/scala/org/apache/spark/api/java/JavaSparkContext.html
Timestamp: 2020-10-19T01:56:49.509Z
Params: Result: Optional[String] Source: https://spark.apache.org/docs/3.0.1/api/scala/org/apache/spark/api/java/JavaSparkContext.html Timestamp: 2020-10-19T01:56:49.509Z
(conf)(conf spark)Params:
Result: SparkConf
Return a copy of this JavaSparkContext's configuration. The configuration cannot be changed at runtime.
Source: https://spark.apache.org/docs/3.0.1/api/scala/org/apache/spark/api/java/JavaSparkContext.html
Timestamp: 2020-10-19T01:56:49.511Z
Params: Result: SparkConf Return a copy of this JavaSparkContext's configuration. The configuration cannot be changed at runtime. Source: https://spark.apache.org/docs/3.0.1/api/scala/org/apache/spark/api/java/JavaSparkContext.html Timestamp: 2020-10-19T01:56:49.511Z
(default-min-partitions)(default-min-partitions spark)Params:
Result: Integer
Source: https://spark.apache.org/docs/3.0.1/api/scala/org/apache/spark/api/java/JavaSparkContext.html
Timestamp: 2020-10-19T01:56:49.503Z
Params: Result: Integer Source: https://spark.apache.org/docs/3.0.1/api/scala/org/apache/spark/api/java/JavaSparkContext.html Timestamp: 2020-10-19T01:56:49.503Z
(default-parallelism)(default-parallelism spark)Params:
Result: Integer
Default level of parallelism to use when not given by user (e.g. parallelize and makeRDD).
Source: https://spark.apache.org/docs/3.0.1/api/scala/org/apache/spark/api/java/JavaSparkContext.html
Timestamp: 2020-10-19T01:56:49.504Z
Params: Result: Integer Default level of parallelism to use when not given by user (e.g. parallelize and makeRDD). Source: https://spark.apache.org/docs/3.0.1/api/scala/org/apache/spark/api/java/JavaSparkContext.html Timestamp: 2020-10-19T01:56:49.504Z
(empty-rdd)(empty-rdd spark)Params:
Result: JavaRDD[T]
Source: https://spark.apache.org/docs/3.0.1/api/scala/org/apache/spark/api/java/JavaSparkContext.html
Timestamp: 2020-10-19T01:56:49.505Z
Params: Result: JavaRDD[T] Source: https://spark.apache.org/docs/3.0.1/api/scala/org/apache/spark/api/java/JavaSparkContext.html Timestamp: 2020-10-19T01:56:49.505Z
(get-checkpoint-dir)(get-checkpoint-dir spark)Params:
Result: Optional[String]
Source: https://spark.apache.org/docs/3.0.1/api/scala/org/apache/spark/api/java/JavaSparkContext.html
Timestamp: 2020-10-19T01:56:49.509Z
Params: Result: Optional[String] Source: https://spark.apache.org/docs/3.0.1/api/scala/org/apache/spark/api/java/JavaSparkContext.html Timestamp: 2020-10-19T01:56:49.509Z
(get-conf)(get-conf spark)Params:
Result: SparkConf
Return a copy of this JavaSparkContext's configuration. The configuration cannot be changed at runtime.
Source: https://spark.apache.org/docs/3.0.1/api/scala/org/apache/spark/api/java/JavaSparkContext.html
Timestamp: 2020-10-19T01:56:49.511Z
Params: Result: SparkConf Return a copy of this JavaSparkContext's configuration. The configuration cannot be changed at runtime. Source: https://spark.apache.org/docs/3.0.1/api/scala/org/apache/spark/api/java/JavaSparkContext.html Timestamp: 2020-10-19T01:56:49.511Z
(get-local-property k)(get-local-property spark k)Params: (key: String)
Result: String
Get a local property set in this thread, or null if it is missing. See org.apache.spark.api.java.JavaSparkContext.setLocalProperty.
Source: https://spark.apache.org/docs/3.0.1/api/scala/org/apache/spark/api/java/JavaSparkContext.html
Timestamp: 2020-10-19T01:56:49.512Z
Params: (key: String) Result: String Get a local property set in this thread, or null if it is missing. See org.apache.spark.api.java.JavaSparkContext.setLocalProperty. Source: https://spark.apache.org/docs/3.0.1/api/scala/org/apache/spark/api/java/JavaSparkContext.html Timestamp: 2020-10-19T01:56:49.512Z
(get-persistent-rd-ds)(get-persistent-rd-ds spark)Params:
Result: Map[Integer, JavaRDD[_]]
Returns a Java map of JavaRDDs that have marked themselves as persistent via cache() call.
This does not necessarily mean the caching or computation was successful.
Source: https://spark.apache.org/docs/3.0.1/api/scala/org/apache/spark/api/java/JavaSparkContext.html
Timestamp: 2020-10-19T01:56:49.513Z
Params: Result: Map[Integer, JavaRDD[_]] Returns a Java map of JavaRDDs that have marked themselves as persistent via cache() call. This does not necessarily mean the caching or computation was successful. Source: https://spark.apache.org/docs/3.0.1/api/scala/org/apache/spark/api/java/JavaSparkContext.html Timestamp: 2020-10-19T01:56:49.513Z
(get-persistent-rdds)(get-persistent-rdds spark)Params:
Result: Map[Integer, JavaRDD[_]]
Returns a Java map of JavaRDDs that have marked themselves as persistent via cache() call.
This does not necessarily mean the caching or computation was successful.
Source: https://spark.apache.org/docs/3.0.1/api/scala/org/apache/spark/api/java/JavaSparkContext.html
Timestamp: 2020-10-19T01:56:49.513Z
Params: Result: Map[Integer, JavaRDD[_]] Returns a Java map of JavaRDDs that have marked themselves as persistent via cache() call. This does not necessarily mean the caching or computation was successful. Source: https://spark.apache.org/docs/3.0.1/api/scala/org/apache/spark/api/java/JavaSparkContext.html Timestamp: 2020-10-19T01:56:49.513Z
(get-spark-home)(get-spark-home spark)Params: ()
Result: Optional[String]
Get Spark's home location from either a value set through the constructor, or the spark.home Java property, or the SPARK_HOME environment variable (in that order of preference). If neither of these is set, return None.
Source: https://spark.apache.org/docs/3.0.1/api/scala/org/apache/spark/api/java/JavaSparkContext.html
Timestamp: 2020-10-19T01:56:49.518Z
Params: () Result: Optional[String] Get Spark's home location from either a value set through the constructor, or the spark.home Java property, or the SPARK_HOME environment variable (in that order of preference). If neither of these is set, return None. Source: https://spark.apache.org/docs/3.0.1/api/scala/org/apache/spark/api/java/JavaSparkContext.html Timestamp: 2020-10-19T01:56:49.518Z
(is-local)(is-local spark)Params:
Result: Boolean
Source: https://spark.apache.org/docs/3.0.1/api/scala/org/apache/spark/api/java/JavaSparkContext.html
Timestamp: 2020-10-19T01:56:49.531Z
Params: Result: Boolean Source: https://spark.apache.org/docs/3.0.1/api/scala/org/apache/spark/api/java/JavaSparkContext.html Timestamp: 2020-10-19T01:56:49.531Z
(jars)(jars spark)Params:
Result: List[String]
Source: https://spark.apache.org/docs/3.0.1/api/scala/org/apache/spark/api/java/JavaSparkContext.html
Timestamp: 2020-10-19T01:56:49.532Z
Params: Result: List[String] Source: https://spark.apache.org/docs/3.0.1/api/scala/org/apache/spark/api/java/JavaSparkContext.html Timestamp: 2020-10-19T01:56:49.532Z
(java-spark-context spark)Converts a SparkSession to a JavaSparkContext.
Converts a SparkSession to a JavaSparkContext.
(local-property k)(local-property spark k)Params: (key: String)
Result: String
Get a local property set in this thread, or null if it is missing. See org.apache.spark.api.java.JavaSparkContext.setLocalProperty.
Source: https://spark.apache.org/docs/3.0.1/api/scala/org/apache/spark/api/java/JavaSparkContext.html
Timestamp: 2020-10-19T01:56:49.512Z
Params: (key: String) Result: String Get a local property set in this thread, or null if it is missing. See org.apache.spark.api.java.JavaSparkContext.setLocalProperty. Source: https://spark.apache.org/docs/3.0.1/api/scala/org/apache/spark/api/java/JavaSparkContext.html Timestamp: 2020-10-19T01:56:49.512Z
(local?)(local? spark)Params:
Result: Boolean
Source: https://spark.apache.org/docs/3.0.1/api/scala/org/apache/spark/api/java/JavaSparkContext.html
Timestamp: 2020-10-19T01:56:49.531Z
Params: Result: Boolean Source: https://spark.apache.org/docs/3.0.1/api/scala/org/apache/spark/api/java/JavaSparkContext.html Timestamp: 2020-10-19T01:56:49.531Z
(master)(master spark)Params:
Result: String
Source: https://spark.apache.org/docs/3.0.1/api/scala/org/apache/spark/api/java/JavaSparkContext.html
Timestamp: 2020-10-19T01:56:49.532Z
Params: Result: String Source: https://spark.apache.org/docs/3.0.1/api/scala/org/apache/spark/api/java/JavaSparkContext.html Timestamp: 2020-10-19T01:56:49.532Z
(parallelise data)(parallelise spark data)Params: (list: List[T], numSlices: Int)
Result: JavaRDD[T]
Source: https://spark.apache.org/docs/3.0.1/api/scala/org/apache/spark/api/java/JavaSparkContext.html
Timestamp: 2020-10-19T01:56:49.544Z
Params: (list: List[T], numSlices: Int) Result: JavaRDD[T] Source: https://spark.apache.org/docs/3.0.1/api/scala/org/apache/spark/api/java/JavaSparkContext.html Timestamp: 2020-10-19T01:56:49.544Z
(parallelise-doubles data)(parallelise-doubles spark data)Params: (list: List[Double], numSlices: Int)
Result: JavaDoubleRDD
Source: https://spark.apache.org/docs/3.0.1/api/scala/org/apache/spark/api/java/JavaSparkContext.html
Timestamp: 2020-10-19T01:56:49.546Z
Params: (list: List[Double], numSlices: Int) Result: JavaDoubleRDD Source: https://spark.apache.org/docs/3.0.1/api/scala/org/apache/spark/api/java/JavaSparkContext.html Timestamp: 2020-10-19T01:56:49.546Z
(parallelise-pairs data)(parallelise-pairs spark data)Params: (list: List[(K, V)], numSlices: Int)
Result: JavaPairRDD[K, V]
Source: https://spark.apache.org/docs/3.0.1/api/scala/org/apache/spark/api/java/JavaSparkContext.html
Timestamp: 2020-10-19T01:56:49.549Z
Params: (list: List[(K, V)], numSlices: Int) Result: JavaPairRDD[K, V] Source: https://spark.apache.org/docs/3.0.1/api/scala/org/apache/spark/api/java/JavaSparkContext.html Timestamp: 2020-10-19T01:56:49.549Z
(parallelize data)(parallelize spark data)Params: (list: List[T], numSlices: Int)
Result: JavaRDD[T]
Source: https://spark.apache.org/docs/3.0.1/api/scala/org/apache/spark/api/java/JavaSparkContext.html
Timestamp: 2020-10-19T01:56:49.544Z
Params: (list: List[T], numSlices: Int) Result: JavaRDD[T] Source: https://spark.apache.org/docs/3.0.1/api/scala/org/apache/spark/api/java/JavaSparkContext.html Timestamp: 2020-10-19T01:56:49.544Z
(parallelize-doubles data)(parallelize-doubles spark data)Params: (list: List[Double], numSlices: Int)
Result: JavaDoubleRDD
Source: https://spark.apache.org/docs/3.0.1/api/scala/org/apache/spark/api/java/JavaSparkContext.html
Timestamp: 2020-10-19T01:56:49.546Z
Params: (list: List[Double], numSlices: Int) Result: JavaDoubleRDD Source: https://spark.apache.org/docs/3.0.1/api/scala/org/apache/spark/api/java/JavaSparkContext.html Timestamp: 2020-10-19T01:56:49.546Z
(parallelize-pairs data)(parallelize-pairs spark data)Params: (list: List[(K, V)], numSlices: Int)
Result: JavaPairRDD[K, V]
Source: https://spark.apache.org/docs/3.0.1/api/scala/org/apache/spark/api/java/JavaSparkContext.html
Timestamp: 2020-10-19T01:56:49.549Z
Params: (list: List[(K, V)], numSlices: Int) Result: JavaPairRDD[K, V] Source: https://spark.apache.org/docs/3.0.1/api/scala/org/apache/spark/api/java/JavaSparkContext.html Timestamp: 2020-10-19T01:56:49.549Z
(persistent-rdds)(persistent-rdds spark)Params:
Result: Map[Integer, JavaRDD[_]]
Returns a Java map of JavaRDDs that have marked themselves as persistent via cache() call.
This does not necessarily mean the caching or computation was successful.
Source: https://spark.apache.org/docs/3.0.1/api/scala/org/apache/spark/api/java/JavaSparkContext.html
Timestamp: 2020-10-19T01:56:49.513Z
Params: Result: Map[Integer, JavaRDD[_]] Returns a Java map of JavaRDDs that have marked themselves as persistent via cache() call. This does not necessarily mean the caching or computation was successful. Source: https://spark.apache.org/docs/3.0.1/api/scala/org/apache/spark/api/java/JavaSparkContext.html Timestamp: 2020-10-19T01:56:49.513Z
(resources)(resources spark)Params:
Result: Map[String, ResourceInformation]
Source: https://spark.apache.org/docs/3.0.1/api/scala/org/apache/spark/api/java/JavaSparkContext.html
Timestamp: 2020-10-19T01:56:49.550Z
Params: Result: Map[String, ResourceInformation] Source: https://spark.apache.org/docs/3.0.1/api/scala/org/apache/spark/api/java/JavaSparkContext.html Timestamp: 2020-10-19T01:56:49.550Z
(sc)(sc spark)Params:
Result: SparkContext
Source: https://spark.apache.org/docs/3.0.1/api/scala/org/apache/spark/api/java/JavaSparkContext.html
Timestamp: 2020-10-19T01:56:49.550Z
Params: Result: SparkContext Source: https://spark.apache.org/docs/3.0.1/api/scala/org/apache/spark/api/java/JavaSparkContext.html Timestamp: 2020-10-19T01:56:49.550Z
(spark-context)(spark-context spark)Params:
Result: SparkContext
Source: https://spark.apache.org/docs/3.0.1/api/scala/org/apache/spark/api/java/JavaSparkContext.html
Timestamp: 2020-10-19T01:56:49.550Z
Params: Result: SparkContext Source: https://spark.apache.org/docs/3.0.1/api/scala/org/apache/spark/api/java/JavaSparkContext.html Timestamp: 2020-10-19T01:56:49.550Z
(spark-home)(spark-home spark)Params: ()
Result: Optional[String]
Get Spark's home location from either a value set through the constructor, or the spark.home Java property, or the SPARK_HOME environment variable (in that order of preference). If neither of these is set, return None.
Source: https://spark.apache.org/docs/3.0.1/api/scala/org/apache/spark/api/java/JavaSparkContext.html
Timestamp: 2020-10-19T01:56:49.518Z
Params: () Result: Optional[String] Get Spark's home location from either a value set through the constructor, or the spark.home Java property, or the SPARK_HOME environment variable (in that order of preference). If neither of these is set, return None. Source: https://spark.apache.org/docs/3.0.1/api/scala/org/apache/spark/api/java/JavaSparkContext.html Timestamp: 2020-10-19T01:56:49.518Z
Params: (path: String)
Result: JavaRDD[String]
Read a text file from HDFS, a local file system (available on all nodes), or any Hadoop-supported file system URI, and return it as an RDD of Strings. The text files must be encoded as UTF-8.
Source: https://spark.apache.org/docs/3.0.1/api/scala/org/apache/spark/api/java/JavaSparkContext.html
Timestamp: 2020-10-19T01:56:49.570Z
Params: (path: String) Result: JavaRDD[String] Read a text file from HDFS, a local file system (available on all nodes), or any Hadoop-supported file system URI, and return it as an RDD of Strings. The text files must be encoded as UTF-8. Source: https://spark.apache.org/docs/3.0.1/api/scala/org/apache/spark/api/java/JavaSparkContext.html Timestamp: 2020-10-19T01:56:49.570Z
(version)(version spark)Params:
Result: String
Source: https://spark.apache.org/docs/3.0.1/api/scala/org/apache/spark/api/java/JavaSparkContext.html
Timestamp: 2020-10-19T01:56:49.576Z
Params: Result: String Source: https://spark.apache.org/docs/3.0.1/api/scala/org/apache/spark/api/java/JavaSparkContext.html Timestamp: 2020-10-19T01:56:49.576Z
Params: (path: String, minPartitions: Int)
Result: JavaPairRDD[String, String]
Read a directory of text files from HDFS, a local file system (available on all nodes), or any Hadoop-supported file system URI. Each file is read as a single record and returned in a key-value pair, where the key is the path of each file, the value is the content of each file. The text files must be encoded as UTF-8.
For example, if you have the following files:
Do
then rdd contains
A suggestion value of the minimal splitting number for input data.
Small files are preferred, large file is also allowable, but may cause bad performance.
Source: https://spark.apache.org/docs/3.0.1/api/scala/org/apache/spark/api/java/JavaSparkContext.html
Timestamp: 2020-10-19T01:56:49.582Z
Params: (path: String, minPartitions: Int) Result: JavaPairRDD[String, String] Read a directory of text files from HDFS, a local file system (available on all nodes), or any Hadoop-supported file system URI. Each file is read as a single record and returned in a key-value pair, where the key is the path of each file, the value is the content of each file. The text files must be encoded as UTF-8. For example, if you have the following files: Do then rdd contains A suggestion value of the minimal splitting number for input data. Small files are preferred, large file is also allowable, but may cause bad performance. Source: https://spark.apache.org/docs/3.0.1/api/scala/org/apache/spark/api/java/JavaSparkContext.html Timestamp: 2020-10-19T01:56:49.582Z
cljdoc builds & hosts documentation for Clojure/Script libraries
| Ctrl+k | Jump to recent docs |
| ← | Move to previous article |
| → | Move to next article |
| Ctrl+/ | Jump to the search field |