pyspark.RDD.mapValues#
- RDD.mapValues(f)[source]#
Pass each value in the key-value pair RDD through a map function without changing the keys; this also retains the original RDD’s partitioning.
New in version 0.7.0.
- Parameters
- ffunction
a function to turn a V into a U
- Returns
See also
Examples
>>> rdd = sc.parallelize([("a", ["apple", "banana", "lemon"]), ("b", ["grapes"])]) >>> def f(x): return len(x) ... >>> rdd.mapValues(f).collect() [('a', 3), ('b', 1)]