Packages

c

com.here.platform.data.processing.spark.partitioner

AdaptiveLevelingPartitioner

case class AdaptiveLevelingPartitioner(pattern: AdaptivePattern, fallbackPartitioner: Option[PartitionNamePartitioner] = None) extends PartitionNamePartitioner with Product with Serializable

A Partitioner for com.here.platform.data.processing.catalog.Partition.Keys that uses a precalculated com.here.platform.data.processing.leveling.AdaptivePattern.

Keys are distributed to Spark partitions strictly following the leveling points that the pattern specifies. Keys left not aggregated by the pattern are distributed among a disjoint set of Spark partitions using a fallback partitioner, if specified. Otherwise they are uniformly distributed over the existing partitions.

The number of partitions used for aggregated keys is fixed and matches the number of leveling points of the pattern.

pattern

The adaptive leveling pattern that controls the partitioning.

fallbackPartitioner

The optional partitioner used for non-aggregated keys. If undefined, non-aggregated keys are uniformly distributed over the existing partitions.

Linear Supertypes
Product, Equals, PartitionNamePartitioner, Partitioner[KeyOrName], org.apache.spark.Partitioner, Serializable, Serializable, AnyRef, Any
Ordering
  1. Alphabetic
  2. By Inheritance
Inherited
  1. AdaptiveLevelingPartitioner
  2. Product
  3. Equals
  4. PartitionNamePartitioner
  5. Partitioner
  6. Partitioner
  7. Serializable
  8. Serializable
  9. AnyRef
  10. Any
  1. Hide All
  2. Show All
Visibility
  1. Public
  2. All

Instance Constructors

  1. new AdaptiveLevelingPartitioner(pattern: AdaptivePattern, fallbackPartitioner: Option[PartitionNamePartitioner] = None)

    pattern

    The adaptive leveling pattern that controls the partitioning.

    fallbackPartitioner

    The optional partitioner used for non-aggregated keys. If undefined, non-aggregated keys are uniformly distributed over the existing partitions.

Value Members

  1. final def !=(arg0: Any): Boolean
    Definition Classes
    AnyRef → Any
  2. final def ##(): Int
    Definition Classes
    AnyRef → Any
  3. final def ==(arg0: Any): Boolean
    Definition Classes
    AnyRef → Any
  4. final def asInstanceOf[T0]: T0
    Definition Classes
    Any
  5. def clone(): AnyRef
    Attributes
    protected[lang]
    Definition Classes
    AnyRef
    Annotations
    @throws( ... ) @native()
  6. final def eq(arg0: AnyRef): Boolean
    Definition Classes
    AnyRef
  7. val fallbackPartitioner: Option[PartitionNamePartitioner]
  8. def finalize(): Unit
    Attributes
    protected[lang]
    Definition Classes
    AnyRef
    Annotations
    @throws( classOf[java.lang.Throwable] )
  9. final def getClass(): Class[_]
    Definition Classes
    AnyRef → Any
    Annotations
    @native()
  10. def getPartition(key: Any): Int

    Implements the Spark org.apache.spark.Partitioner interface by forwarding the calls to getPartitionForKey.

    Implements the Spark org.apache.spark.Partitioner interface by forwarding the calls to getPartitionForKey.

    If the object passed is not of type K or can't be converted to it (e.g. java.lang.Integer to Int), a IllegalArgumentException is thrown. This should be considered a bug that should not happen because the processing library uses Partitioner of type K only for RDDs for which it is aware and sure to have keys of type K.

    Basically, this function is a no-op call that forwards to getPartitionForKey, but the important point here is to have a type-safe Partitioner in the processing library.

    key

    the key for which the partition must be calculated

    returns

    the partition, identified by one scala.Int, in which the key should be located

    Definition Classes
    Partitioner → Partitioner
    Note

    This is called by Spark and should not be called by developer's code, as it may be unsafe.

  11. final def getPartitionForKey(key: KeyOrName): Int

    Gets the partition for a given key of type K.

    Gets the partition for a given key of type K. This is the function that must be implemented by children partitioners.

    key

    the key for which the partition must be calculated

    returns

    the partition, identified by one scala.Int, in which the key should be located

    Definition Classes
    PartitionNamePartitionerPartitioner
  12. def getPartitionForName(name: Name): Int
  13. final def isInstanceOf[T0]: Boolean
    Definition Classes
    Any
  14. final def ne(arg0: AnyRef): Boolean
    Definition Classes
    AnyRef
  15. final def notify(): Unit
    Definition Classes
    AnyRef
    Annotations
    @native()
  16. final def notifyAll(): Unit
    Definition Classes
    AnyRef
    Annotations
    @native()
  17. def numPartitions: Int
    Definition Classes
    AdaptiveLevelingPartitioner → Partitioner
  18. val pattern: AdaptivePattern
  19. final def synchronized[T0](arg0: ⇒ T0): T0
    Definition Classes
    AnyRef
  20. final def wait(): Unit
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  21. final def wait(arg0: Long, arg1: Int): Unit
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  22. final def wait(arg0: Long): Unit
    Definition Classes
    AnyRef
    Annotations
    @throws( ... ) @native()

Inherited from Product

Inherited from Equals

Inherited from PartitionNamePartitioner

Inherited from Partitioner[KeyOrName]

Inherited from org.apache.spark.Partitioner

Inherited from Serializable

Inherited from Serializable

Inherited from AnyRef

Inherited from Any

Ungrouped