Runs this query returning the result as an RDD.
A custom implementation modeled after the take function on RDDs but which never runs any job locally.
Specifies how data is partitioned across different nodes in the cluster.
Specifies how data is partitioned across different nodes in the cluster.
Specifies any partition requirements on the input data for this operator.
Specifies any partition requirements on the input data for this operator.
:: DeveloperApi :: Take the first limit elements. Note that the implementation is different depending on whether this is a terminal operator or not. If it is terminal and is invoked using executeCollect, this operator uses something similar to Spark's take method on the Spark driver. If it is not terminal or is invoked using execute, we first take the limit on each partition, and then repartition all the data to a single partition to compute the global limit.