Distributed Dataset (RDD), the basic abstraction in Spark. Represents an immutable, partitioned collection of elements that can be operated on in parallel
That was my understanding
It could be understood as a collection semantically, but not as a collection in scala stdlib sense
Обсуждают сегодня