pyspark.SparkContext.defaultMinPartitions#

property SparkContext.defaultMinPartitions#

Default min number of partitions for Hadoop RDDs when not given by user

New in version 1.1.0.

Examples

>>> sc.defaultMinPartitions > 0
True