pyspark.sql.SparkSession.sparkContext¶
-
property
SparkSession.
sparkContext
¶ Returns the underlying
SparkContext
.New in version 2.0.0.
- Returns
SparkContext
Examples
>>> spark.sparkContext <SparkContext master=... appName=...>
Create an RDD from the Spark context
>>> rdd = spark.sparkContext.parallelize([1, 2, 3]) >>> rdd.collect() [1, 2, 3]