pyspark.SparkContext.range¶
-
SparkContext.
range
(start: int, end: Optional[int] = None, step: int = 1, numSlices: Optional[int] = None) → pyspark.rdd.RDD[int]¶ Create a new RDD of int containing elements from start to end (exclusive), increased by step every element. Can be called the same way as python’s built-in range() function. If called with a single argument, the argument is interpreted as end, and start is set to 0.
- Parameters
- startint
the start value
- endint, optional
the end value (exclusive)
- stepint, optional
the incremental step (default: 1)
- numSlicesint, optional
the number of partitions of the new RDD
- Returns
pyspark.RDD
An RDD of int
Examples
>>> sc.range(5).collect() [0, 1, 2, 3, 4] >>> sc.range(2, 4).collect() [2, 3] >>> sc.range(1, 7, 2).collect() [1, 3, 5]