Skip navigation links
A B C D E F G H I J K L M N O P Q R S T U V W X Y Z _ 

A

abort(Throwable) - Method in interface org.apache.spark.shuffle.api.ShuffleMapOutputWriter
Abort all of the writes done by any writers returned by ShuffleMapOutputWriter.getPartitionWriter(int).
abort(WriterCommitMessage[]) - Method in interface org.apache.spark.sql.connector.write.BatchWrite
Aborts this writing job because some data writers are failed and keep failing when retry, or the Spark job fails with some unknown reasons, or BatchWrite.onDataWriterCommit(WriterCommitMessage) fails, or BatchWrite.commit(WriterCommitMessage[]) fails.
abort() - Method in interface org.apache.spark.sql.connector.write.DataWriter
Aborts this writer if it is failed.
abort(long, WriterCommitMessage[]) - Method in interface org.apache.spark.sql.connector.write.streaming.StreamingWrite
Aborts this writing job because some data writers are failed and keep failing when retried, or the Spark job fails with some unknown reasons, or StreamingWrite.commit(long, WriterCommitMessage[]) fails.
abortStagedChanges() - Method in interface org.apache.spark.sql.connector.catalog.StagedTable
Abort the changes that were staged, both in metadata and from temporary outputs of this table's writers.
abs(Column) - Static method in class org.apache.spark.sql.functions
Computes the absolute value of a numeric value.
abs(T) - Static method in class org.apache.spark.sql.types.ByteExactNumeric
 
abs() - Method in class org.apache.spark.sql.types.Decimal
 
abs(T) - Static method in class org.apache.spark.sql.types.DecimalExactNumeric
 
abs(double) - Static method in class org.apache.spark.sql.types.DoubleExactNumeric
 
abs(double) - Method in interface org.apache.spark.sql.types.DoubleType.DoubleIsConflicted
 
abs(float) - Static method in class org.apache.spark.sql.types.FloatExactNumeric
 
abs(float) - Method in interface org.apache.spark.sql.types.FloatType.FloatIsConflicted
 
abs(T) - Static method in class org.apache.spark.sql.types.IntegerExactNumeric
 
abs(T) - Static method in class org.apache.spark.sql.types.LongExactNumeric
 
abs(T) - Static method in class org.apache.spark.sql.types.ShortExactNumeric
 
absent() - Static method in class org.apache.spark.api.java.Optional
 
AbsoluteError - Class in org.apache.spark.mllib.tree.loss
Class for absolute error loss calculation (for regression).
AbsoluteError() - Constructor for class org.apache.spark.mllib.tree.loss.AbsoluteError
 
AbstractLauncher<T extends AbstractLauncher<T>> - Class in org.apache.spark.launcher
Base class for launcher implementations.
accept(Parsers) - Static method in class org.apache.spark.ml.feature.RFormulaParser
 
accept(ES, Function1<ES, List<Object>>) - Static method in class org.apache.spark.ml.feature.RFormulaParser
 
accept(String, PartialFunction<Object, U>) - Static method in class org.apache.spark.ml.feature.RFormulaParser
 
accept(Path) - Method in class org.apache.spark.ml.image.SamplePathFilter
 
acceptIf(Function1<Object, Object>, Function1<Object, String>) - Static method in class org.apache.spark.ml.feature.RFormulaParser
 
acceptMatch(String, PartialFunction<Object, U>) - Static method in class org.apache.spark.ml.feature.RFormulaParser
 
acceptSeq(ES, Function1<ES, Iterable<Object>>) - Static method in class org.apache.spark.ml.feature.RFormulaParser
 
AcceptsLatestSeenOffset - Interface in org.apache.spark.sql.connector.read.streaming
Indicates that the source accepts the latest seen offset, which requires streaming execution to provide the latest seen offset when restarting the streaming query from checkpoint.
acceptsType(DataType) - Method in class org.apache.spark.sql.types.ObjectType
 
accessNonExistentAccumulatorError(long) - Static method in class org.apache.spark.errors.SparkCoreErrors
 
accId() - Method in class org.apache.spark.CleanAccum
 
accumCleaned(long) - Method in interface org.apache.spark.CleanerListener
 
AccumulableInfo - Class in org.apache.spark.scheduler
:: DeveloperApi :: Information about an AccumulatorV2 modified during a task or stage.
AccumulableInfo - Class in org.apache.spark.status.api.v1
 
accumulableInfoFromJson(JsonNode) - Static method in class org.apache.spark.util.JsonProtocol
 
AccumulableInfoSerializer - Class in org.apache.spark.status.protobuf
 
AccumulableInfoSerializer() - Constructor for class org.apache.spark.status.protobuf.AccumulableInfoSerializer
 
accumulableInfoToJson(AccumulableInfo, JsonGenerator) - Static method in class org.apache.spark.util.JsonProtocol
 
accumulables() - Method in class org.apache.spark.scheduler.StageInfo
Terminal values of accumulables updated during this stage, including all the user-defined accumulators.
accumulables() - Method in class org.apache.spark.scheduler.TaskInfo
Intermediate updates to accumulables during this task.
accumulablesToJson(Iterable<AccumulableInfo>, JsonGenerator) - Static method in class org.apache.spark.util.JsonProtocol
 
ACCUMULATOR_ID_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetric
 
ACCUMULATOR_UPDATES_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.StageData
 
ACCUMULATOR_UPDATES_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.TaskData
 
ACCUMULATOR_UPDATES_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
 
AccumulatorContext - Class in org.apache.spark.util
An internal class used to track accumulators by Spark itself.
AccumulatorContext() - Constructor for class org.apache.spark.util.AccumulatorContext
 
ACCUMULATORS() - Static method in class org.apache.spark.status.TaskIndexNames
 
accumulatorUpdates() - Method in class org.apache.spark.status.api.v1.StageData
 
accumulatorUpdates() - Method in class org.apache.spark.status.api.v1.TaskData
 
AccumulatorV2<IN,OUT> - Class in org.apache.spark.util
The base class for accumulators, that can accumulate inputs of type IN, and produce output of type OUT.
AccumulatorV2() - Constructor for class org.apache.spark.util.AccumulatorV2
 
accumUpdates() - Method in class org.apache.spark.ExceptionFailure
 
accumUpdates() - Method in class org.apache.spark.scheduler.SparkListenerExecutorMetricsUpdate
 
accumUpdates() - Method in class org.apache.spark.TaskKilled
 
accuracy() - Method in interface org.apache.spark.ml.classification.ClassificationSummary
Returns accuracy.
accuracy() - Method in class org.apache.spark.mllib.evaluation.MulticlassMetrics
 
accuracy() - Method in class org.apache.spark.mllib.evaluation.MultilabelMetrics
Returns accuracy
acos(Column) - Static method in class org.apache.spark.sql.functions
 
acos(String) - Static method in class org.apache.spark.sql.functions
 
acosh(Column) - Static method in class org.apache.spark.sql.functions
 
acosh(String) - Static method in class org.apache.spark.sql.functions
 
acquire(Seq<String>) - Method in interface org.apache.spark.resource.ResourceAllocator
Acquire a sequence of resource addresses (to a launched task), these addresses must be available.
actionNotAllowedOnTableSincePartitionMetadataNotStoredError(String, String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
actionNotAllowedOnTableWithFilesourcePartitionManagementDisabledError(String, String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
ActivationFunction - Interface in org.apache.spark.ml.ann
Trait for functions and their derivatives for functional layers
active() - Static method in class org.apache.spark.sql.SparkSession
Returns the currently active SparkSession, otherwise the default one.
active() - Method in class org.apache.spark.sql.streaming.StreamingQueryManager
Returns a list of active queries associated with this SQLContext
active() - Method in class org.apache.spark.streaming.scheduler.ReceiverInfo
 
ACTIVE() - Static method in class org.apache.spark.streaming.scheduler.ReceiverState
 
ACTIVE_TASKS_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
 
activeIterator() - Method in interface org.apache.spark.ml.linalg.Vector
Returns an iterator over all the active elements of this vector.
activeIterator() - Method in interface org.apache.spark.mllib.linalg.Vector
Returns an iterator over all the active elements of this vector.
activeStages() - Method in class org.apache.spark.status.LiveJob
 
activeTasks() - Method in class org.apache.spark.status.api.v1.ExecutorSummary
 
activeTasks() - Method in class org.apache.spark.status.LiveJob
 
activeTasks() - Method in class org.apache.spark.status.LiveStage
 
activeTasksPerExecutor() - Method in class org.apache.spark.status.LiveStage
 
add(Tuple2<Vector, Object>) - Method in class org.apache.spark.ml.clustering.ExpectationAggregator
Add a new training instance to this ExpectationAggregator, update the weights, means and covariances for each distributions, and update the log likelihood.
add(org.apache.spark.ml.feature.InstanceBlock) - Method in class org.apache.spark.ml.clustering.KMeansAggregator
 
add(Term) - Static method in class org.apache.spark.ml.feature.Dot
 
add(Term) - Static method in class org.apache.spark.ml.feature.EmptyTerm
 
add(Term) - Method in interface org.apache.spark.ml.feature.Term
Creates a summation term by concatenation of terms.
add(Datum) - Method in interface org.apache.spark.ml.optim.aggregator.DifferentiableLossAggregator
Add a single data point to this aggregator.
add(double[], MultivariateGaussian[], ExpectationSum, Vector<Object>) - Static method in class org.apache.spark.mllib.clustering.ExpectationSum
 
add(Vector) - Method in class org.apache.spark.mllib.feature.IDF.DocumentFrequencyAggregator
Adds a new document.
add(BlockMatrix) - Method in class org.apache.spark.mllib.linalg.distributed.BlockMatrix
Adds the given block matrix other to this block matrix: this + other.
add(Vector) - Method in class org.apache.spark.mllib.stat.MultivariateOnlineSummarizer
Add a new sample to this summarizer, and update the statistical summary.
add(StructField) - Method in class org.apache.spark.sql.types.StructType
Creates a new StructType by adding a new field.
add(String, DataType) - Method in class org.apache.spark.sql.types.StructType
Creates a new StructType by adding a new nullable field with no metadata.
add(String, DataType, boolean) - Method in class org.apache.spark.sql.types.StructType
Creates a new StructType by adding a new field with no metadata.
add(String, DataType, boolean, Metadata) - Method in class org.apache.spark.sql.types.StructType
Creates a new StructType by adding a new field and specifying metadata.
add(String, DataType, boolean, String) - Method in class org.apache.spark.sql.types.StructType
Creates a new StructType by adding a new field and specifying metadata.
add(String, String) - Method in class org.apache.spark.sql.types.StructType
Creates a new StructType by adding a new nullable field with no metadata where the dataType is specified as a String.
add(String, String, boolean) - Method in class org.apache.spark.sql.types.StructType
Creates a new StructType by adding a new field with no metadata where the dataType is specified as a String.
add(String, String, boolean, Metadata) - Method in class org.apache.spark.sql.types.StructType
Creates a new StructType by adding a new field and specifying metadata where the dataType is specified as a String.
add(String, String, boolean, String) - Method in class org.apache.spark.sql.types.StructType
Creates a new StructType by adding a new field and specifying metadata where the dataType is specified as a String.
add(Long) - Method in class org.apache.spark.sql.util.MapperRowCounter
 
add(double) - Method in class org.apache.spark.sql.util.NumericHistogram
Adds a new data point to the histogram approximation.
add(T) - Method in class org.apache.spark.sql.util.SQLOpenHashSet
 
add(long, long) - Static method in class org.apache.spark.streaming.util.RawTextHelper
 
add(IN) - Method in class org.apache.spark.util.AccumulatorV2
Takes the inputs and accumulates.
add(T) - Method in class org.apache.spark.util.CollectionAccumulator
 
add(Double) - Method in class org.apache.spark.util.DoubleAccumulator
Adds v to the accumulator, i.e.
add(double) - Method in class org.apache.spark.util.DoubleAccumulator
Adds v to the accumulator, i.e.
add(Long) - Method in class org.apache.spark.util.LongAccumulator
Adds v to the accumulator, i.e.
add(long) - Method in class org.apache.spark.util.LongAccumulator
Adds v to the accumulator, i.e.
add(Object) - Method in class org.apache.spark.util.sketch.CountMinSketch
Increments item's count by one.
add(Object, long) - Method in class org.apache.spark.util.sketch.CountMinSketch
Increments item's count by count.
add_months(Column, int) - Static method in class org.apache.spark.sql.functions
Returns the date that is numMonths after startDate.
add_months(Column, Column) - Static method in class org.apache.spark.sql.functions
Returns the date that is numMonths after startDate.
ADD_TIME_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
 
ADD_TIME_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary
 
addAccumulatorUpdates(StoreTypes.AccumulableInfo) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 44;
addAccumulatorUpdates(int, StoreTypes.AccumulableInfo) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 44;
addAccumulatorUpdates(StoreTypes.AccumulableInfo.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 44;
addAccumulatorUpdates(int, StoreTypes.AccumulableInfo.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 44;
addAccumulatorUpdates(StoreTypes.AccumulableInfo) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
addAccumulatorUpdates(int, StoreTypes.AccumulableInfo) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
addAccumulatorUpdates(StoreTypes.AccumulableInfo.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
addAccumulatorUpdates(int, StoreTypes.AccumulableInfo.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
addAccumulatorUpdates(StoreTypes.AccumulableInfo) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
addAccumulatorUpdates(int, StoreTypes.AccumulableInfo) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
addAccumulatorUpdates(StoreTypes.AccumulableInfo.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
addAccumulatorUpdates(int, StoreTypes.AccumulableInfo.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
addAccumulatorUpdatesBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 44;
addAccumulatorUpdatesBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 44;
addAccumulatorUpdatesBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
addAccumulatorUpdatesBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
addAccumulatorUpdatesBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
addAccumulatorUpdatesBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
addAddresses(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceInformation.Builder
repeated string addresses = 2;
addAddressesBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceInformation.Builder
repeated string addresses = 2;
addAllAccumulatorUpdates(Iterable<? extends StoreTypes.AccumulableInfo>) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 44;
addAllAccumulatorUpdates(Iterable<? extends StoreTypes.AccumulableInfo>) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
addAllAccumulatorUpdates(Iterable<? extends StoreTypes.AccumulableInfo>) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
addAllAddresses(Iterable<String>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceInformation.Builder
repeated string addresses = 2;
addAllAttempts(Iterable<? extends StoreTypes.ApplicationAttemptInfo>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
repeated .org.apache.spark.status.protobuf.ApplicationAttemptInfo attempts = 7;
addAllBlacklistedInStages(Iterable<? extends Long>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
repeated int64 blacklisted_in_stages = 25;
addAllBytesRead(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributions.Builder
repeated double bytes_read = 1;
addAllBytesWritten(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributions.Builder
repeated double bytes_written = 1;
addAllChildClusters(Iterable<? extends StoreTypes.RDDOperationClusterWrapper>) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationClusterWrapper child_clusters = 4;
addAllChildNodes(Iterable<? extends StoreTypes.RDDOperationNode>) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationNode child_nodes = 3;
addAllClasspathEntries(Iterable<? extends StoreTypes.PairStrings>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings classpath_entries = 6;
addAllCorruptMergedBlockChunks(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
repeated double corrupt_merged_block_chunks = 1;
addAllDataDistribution(Iterable<? extends StoreTypes.RDDDataDistribution>) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
repeated .org.apache.spark.status.protobuf.RDDDataDistribution data_distribution = 8;
addAllDiskBytesSpilled(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double disk_bytes_spilled = 15;
addAllDiskBytesSpilled(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double disk_bytes_spilled = 14;
addAllDuration(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double duration = 2;
addAllEdges(Iterable<? extends StoreTypes.RDDOperationEdge>) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationEdge edges = 2;
addAllEdges(Iterable<? extends StoreTypes.SparkPlanGraphEdge>) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.SparkPlanGraphEdge edges = 3;
addAllExcludedInStages(Iterable<? extends Long>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
repeated int64 excluded_in_stages = 31;
addAllExecutorCpuTime(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double executor_cpu_time = 6;
addAllExecutorDeserializeCpuTime(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double executor_deserialize_cpu_time = 4;
addAllExecutorDeserializeTime(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double executor_deserialize_time = 3;
addAllExecutorMetrics(Iterable<? extends StoreTypes.ExecutorMetrics>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions.Builder
repeated .org.apache.spark.status.protobuf.ExecutorMetrics executor_metrics = 2;
addAllExecutorRunTime(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double executor_run_time = 5;
addAllExecutors(Iterable<String>) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo.Builder
repeated string executors = 5;
addAllFailedTasks(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double failed_tasks = 3;
addAllFetchWaitTime(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
repeated double fetch_wait_time = 5;
addAllGettingResultTime(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double getting_result_time = 10;
addAllHadoopProperties(Iterable<? extends StoreTypes.PairStrings>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings hadoop_properties = 3;
addAllIncomingEdges(Iterable<? extends StoreTypes.RDDOperationEdge>) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationEdge incoming_edges = 4;
addAllInputBytes(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double input_bytes = 6;
addAllInputRecords(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double input_records = 7;
addAllJobIds(Iterable<? extends Long>) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper.Builder
repeated int64 job_ids = 2;
addAllJvmGcTime(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double jvm_gc_time = 8;
addAllKilledTasks(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double killed_tasks = 5;
addAllLocalBlocksFetched(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
repeated double local_blocks_fetched = 4;
addAllLocalMergedBlocksFetched(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
repeated double local_merged_blocks_fetched = 4;
addAllLocalMergedBytesRead(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
repeated double local_merged_bytes_read = 8;
addAllLocalMergedChunksFetched(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
repeated double local_merged_chunks_fetched = 6;
addAllMemoryBytesSpilled(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double memory_bytes_spilled = 14;
addAllMemoryBytesSpilled(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double memory_bytes_spilled = 13;
addAllMergedFetchFallbackCount(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
repeated double merged_fetch_fallback_count = 2;
addAllMetrics(Iterable<? extends StoreTypes.SQLPlanMetric>) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 5;
addAllMetrics(Iterable<? extends StoreTypes.SQLPlanMetric>) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 4;
addAllMetrics(Iterable<? extends StoreTypes.SQLPlanMetric>) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 7;
addAllMetricsProperties(Iterable<? extends StoreTypes.PairStrings>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings metrics_properties = 5;
addAllNodes(Iterable<? extends StoreTypes.SparkPlanGraphNodeWrapper>) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 4;
addAllNodes(Iterable<? extends StoreTypes.SparkPlanGraphNodeWrapper>) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 2;
addAllOutgoingEdges(Iterable<? extends StoreTypes.RDDOperationEdge>) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationEdge outgoing_edges = 3;
addAllOutputBytes(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double output_bytes = 8;
addAllOutputRecords(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double output_records = 9;
addAllPartitions(Iterable<? extends StoreTypes.RDDPartitionInfo>) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
repeated .org.apache.spark.status.protobuf.RDDPartitionInfo partitions = 9;
addAllPeakExecutionMemory(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double peak_execution_memory = 12;
addAllQuantiles(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double quantiles = 1;
addAllQuantiles(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions.Builder
repeated double quantiles = 1;
addAllQuantiles(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double quantiles = 1;
addAllRddIds(Iterable<? extends Long>) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
repeated int64 rdd_ids = 43;
addAllReadBytes(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
repeated double read_bytes = 1;
addAllReadRecords(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
repeated double read_records = 2;
addAllRecordsRead(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributions.Builder
repeated double records_read = 2;
addAllRecordsWritten(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributions.Builder
repeated double records_written = 2;
addAllRemoteBlocksFetched(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
repeated double remote_blocks_fetched = 3;
addAllRemoteBytesRead(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
repeated double remote_bytes_read = 6;
addAllRemoteBytesReadToDisk(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
repeated double remote_bytes_read_to_disk = 7;
addAllRemoteMergedBlocksFetched(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
repeated double remote_merged_blocks_fetched = 3;
addAllRemoteMergedBytesRead(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
repeated double remote_merged_bytes_read = 7;
addAllRemoteMergedChunksFetched(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
repeated double remote_merged_chunks_fetched = 5;
addAllRemoteMergedReqsDuration(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
repeated double remote_merged_reqs_duration = 9;
addAllRemoteReqsDuration(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
repeated double remote_reqs_duration = 9;
addAllResourceProfiles(Iterable<? extends StoreTypes.ResourceProfileInfo>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.ResourceProfileInfo resource_profiles = 7;
addAllResultSerializationTime(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double result_serialization_time = 9;
addAllResultSize(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double result_size = 7;
addAllSchedulerDelay(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double scheduler_delay = 11;
addAllShuffleRead(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double shuffle_read = 10;
addAllShuffleReadRecords(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double shuffle_read_records = 11;
addAllShuffleWrite(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double shuffle_write = 12;
addAllShuffleWriteRecords(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double shuffle_write_records = 13;
addAllSkippedStages(Iterable<? extends Integer>) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper.Builder
repeated int32 skipped_stages = 2;
addAllSources(Iterable<? extends StoreTypes.SourceProgress>) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
repeated .org.apache.spark.status.protobuf.SourceProgress sources = 10;
addAllSparkProperties(Iterable<? extends StoreTypes.PairStrings>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings spark_properties = 2;
addAllStageIds(Iterable<? extends Long>) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
repeated int64 stage_ids = 6;
addAllStageIds(Iterable<? extends Long>) - Method in class org.apache.spark.status.protobuf.StoreTypes.PoolData.Builder
repeated int64 stage_ids = 2;
addAllStages(Iterable<? extends Long>) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
repeated int64 stages = 12;
addAllStateOperators(Iterable<? extends StoreTypes.StateOperatorProgress>) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
repeated .org.apache.spark.status.protobuf.StateOperatorProgress state_operators = 9;
addAllSucceededTasks(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double succeeded_tasks = 4;
addAllSystemProperties(Iterable<? extends StoreTypes.PairStrings>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings system_properties = 4;
addAllTaskTime(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double task_time = 2;
addAllTotalBlocksFetched(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
repeated double total_blocks_fetched = 8;
addAllWriteBytes(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions.Builder
repeated double write_bytes = 1;
addAllWriteRecords(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions.Builder
repeated double write_records = 2;
addAllWriteTime(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions.Builder
repeated double write_time = 3;
addAppArgs(String...) - Method in class org.apache.spark.launcher.AbstractLauncher
Adds command line arguments for the application.
addAppArgs(String...) - Method in class org.apache.spark.launcher.SparkLauncher
 
addArchive(String) - Method in class org.apache.spark.SparkContext
:: Experimental :: Add an archive to be downloaded and unpacked with this Spark job on every node.
addAttempts(StoreTypes.ApplicationAttemptInfo) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
repeated .org.apache.spark.status.protobuf.ApplicationAttemptInfo attempts = 7;
addAttempts(int, StoreTypes.ApplicationAttemptInfo) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
repeated .org.apache.spark.status.protobuf.ApplicationAttemptInfo attempts = 7;
addAttempts(StoreTypes.ApplicationAttemptInfo.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
repeated .org.apache.spark.status.protobuf.ApplicationAttemptInfo attempts = 7;
addAttempts(int, StoreTypes.ApplicationAttemptInfo.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
repeated .org.apache.spark.status.protobuf.ApplicationAttemptInfo attempts = 7;
addAttemptsBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
repeated .org.apache.spark.status.protobuf.ApplicationAttemptInfo attempts = 7;
addAttemptsBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
repeated .org.apache.spark.status.protobuf.ApplicationAttemptInfo attempts = 7;
addBin(double, double, int) - Method in class org.apache.spark.sql.util.NumericHistogram
Set a particular histogram bin with index.
addBinary(byte[]) - Method in class org.apache.spark.util.sketch.CountMinSketch
Increments item's count by one.
addBinary(byte[], long) - Method in class org.apache.spark.util.sketch.CountMinSketch
Increments item's count by count.
addBlacklistedInStages(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
repeated int64 blacklisted_in_stages = 25;
addBytesRead(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributions.Builder
repeated double bytes_read = 1;
addBytesWritten(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributions.Builder
repeated double bytes_written = 1;
addCatalogInCacheTableAsSelectNotAllowedError(String, SqlBaseParser.CacheTableContext) - Static method in class org.apache.spark.sql.errors.QueryParsingErrors
 
addChildClusters(StoreTypes.RDDOperationClusterWrapper) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationClusterWrapper child_clusters = 4;
addChildClusters(int, StoreTypes.RDDOperationClusterWrapper) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationClusterWrapper child_clusters = 4;
addChildClusters(StoreTypes.RDDOperationClusterWrapper.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationClusterWrapper child_clusters = 4;
addChildClusters(int, StoreTypes.RDDOperationClusterWrapper.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationClusterWrapper child_clusters = 4;
addChildClustersBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationClusterWrapper child_clusters = 4;
addChildClustersBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationClusterWrapper child_clusters = 4;
addChildNodes(StoreTypes.RDDOperationNode) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationNode child_nodes = 3;
addChildNodes(int, StoreTypes.RDDOperationNode) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationNode child_nodes = 3;
addChildNodes(StoreTypes.RDDOperationNode.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationNode child_nodes = 3;
addChildNodes(int, StoreTypes.RDDOperationNode.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationNode child_nodes = 3;
addChildNodesBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationNode child_nodes = 3;
addChildNodesBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationNode child_nodes = 3;
addChunk(ShuffleBlockChunkId, RoaringBitmap) - Method in class org.apache.spark.storage.PushBasedFetchHelper
This is executed by the task thread when the iterator.next() is invoked and the iterator processes a response of type ShuffleBlockFetcherIterator.PushMergedLocalMetaFetchResult.
addClasspathEntries(StoreTypes.PairStrings) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings classpath_entries = 6;
addClasspathEntries(int, StoreTypes.PairStrings) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings classpath_entries = 6;
addClasspathEntries(StoreTypes.PairStrings.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings classpath_entries = 6;
addClasspathEntries(int, StoreTypes.PairStrings.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings classpath_entries = 6;
addClasspathEntriesBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings classpath_entries = 6;
addClasspathEntriesBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings classpath_entries = 6;
addColumn(String[], DataType) - Static method in interface org.apache.spark.sql.connector.catalog.TableChange
Create a TableChange for adding an optional column.
addColumn(String[], DataType, boolean) - Static method in interface org.apache.spark.sql.connector.catalog.TableChange
Create a TableChange for adding a column.
addColumn(String[], DataType, boolean, String) - Static method in interface org.apache.spark.sql.connector.catalog.TableChange
Create a TableChange for adding a column.
addColumn(String[], DataType, boolean, String, TableChange.ColumnPosition, ColumnDefaultValue) - Static method in interface org.apache.spark.sql.connector.catalog.TableChange
Create a TableChange for adding a column.
addColumnWithV1TableCannotSpecifyNotNullError() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
addCorruptMergedBlockChunks(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
repeated double corrupt_merged_block_chunks = 1;
addDataDistribution(StoreTypes.RDDDataDistribution) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
repeated .org.apache.spark.status.protobuf.RDDDataDistribution data_distribution = 8;
addDataDistribution(int, StoreTypes.RDDDataDistribution) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
repeated .org.apache.spark.status.protobuf.RDDDataDistribution data_distribution = 8;
addDataDistribution(StoreTypes.RDDDataDistribution.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
repeated .org.apache.spark.status.protobuf.RDDDataDistribution data_distribution = 8;
addDataDistribution(int, StoreTypes.RDDDataDistribution.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
repeated .org.apache.spark.status.protobuf.RDDDataDistribution data_distribution = 8;
addDataDistributionBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
repeated .org.apache.spark.status.protobuf.RDDDataDistribution data_distribution = 8;
addDataDistributionBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
repeated .org.apache.spark.status.protobuf.RDDDataDistribution data_distribution = 8;
addDiskBytesSpilled(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double disk_bytes_spilled = 15;
addDiskBytesSpilled(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double disk_bytes_spilled = 14;
addDuration(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double duration = 2;
addEdges(StoreTypes.RDDOperationEdge) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationEdge edges = 2;
addEdges(int, StoreTypes.RDDOperationEdge) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationEdge edges = 2;
addEdges(StoreTypes.RDDOperationEdge.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationEdge edges = 2;
addEdges(int, StoreTypes.RDDOperationEdge.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationEdge edges = 2;
addEdges(StoreTypes.SparkPlanGraphEdge) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.SparkPlanGraphEdge edges = 3;
addEdges(int, StoreTypes.SparkPlanGraphEdge) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.SparkPlanGraphEdge edges = 3;
addEdges(StoreTypes.SparkPlanGraphEdge.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.SparkPlanGraphEdge edges = 3;
addEdges(int, StoreTypes.SparkPlanGraphEdge.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.SparkPlanGraphEdge edges = 3;
addEdgesBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationEdge edges = 2;
addEdgesBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationEdge edges = 2;
addEdgesBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.SparkPlanGraphEdge edges = 3;
addEdgesBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.SparkPlanGraphEdge edges = 3;
addExcludedInStages(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
repeated int64 excluded_in_stages = 31;
addExecutorCpuTime(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double executor_cpu_time = 6;
addExecutorDeserializeCpuTime(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double executor_deserialize_cpu_time = 4;
addExecutorDeserializeTime(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double executor_deserialize_time = 3;
addExecutorMetrics(StoreTypes.ExecutorMetrics) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions.Builder
repeated .org.apache.spark.status.protobuf.ExecutorMetrics executor_metrics = 2;
addExecutorMetrics(int, StoreTypes.ExecutorMetrics) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions.Builder
repeated .org.apache.spark.status.protobuf.ExecutorMetrics executor_metrics = 2;
addExecutorMetrics(StoreTypes.ExecutorMetrics.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions.Builder
repeated .org.apache.spark.status.protobuf.ExecutorMetrics executor_metrics = 2;
addExecutorMetrics(int, StoreTypes.ExecutorMetrics.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions.Builder
repeated .org.apache.spark.status.protobuf.ExecutorMetrics executor_metrics = 2;
addExecutorMetricsBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions.Builder
repeated .org.apache.spark.status.protobuf.ExecutorMetrics executor_metrics = 2;
addExecutorMetricsBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions.Builder
repeated .org.apache.spark.status.protobuf.ExecutorMetrics executor_metrics = 2;
addExecutorRunTime(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double executor_run_time = 5;
addExecutors(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo.Builder
repeated string executors = 5;
addExecutorsBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo.Builder
repeated string executors = 5;
addFailedTasks(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double failed_tasks = 3;
addFetchWaitTime(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
repeated double fetch_wait_time = 5;
addFile(String) - Method in class org.apache.spark.api.java.JavaSparkContext
Add a file to be downloaded with this Spark job on every node.
addFile(String, boolean) - Method in class org.apache.spark.api.java.JavaSparkContext
Add a file to be downloaded with this Spark job on every node.
addFile(String) - Method in class org.apache.spark.launcher.AbstractLauncher
Adds a file to be submitted with the application.
addFile(String) - Method in class org.apache.spark.launcher.SparkLauncher
 
addFile(String) - Method in class org.apache.spark.SparkContext
Add a file to be downloaded with this Spark job on every node.
addFile(String, boolean) - Method in class org.apache.spark.SparkContext
Add a file to be downloaded with this Spark job on every node.
addFilesWithAbsolutePathUnsupportedError(String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
addFilter(ServletContextHandler, String, Map<String, String>) - Static method in class org.apache.spark.ui.JettyUtils
 
addGettingResultTime(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double getting_result_time = 10;
addGrid(Param<T>, Iterable<T>) - Method in class org.apache.spark.ml.tuning.ParamGridBuilder
Adds a param with multiple values (overwrites if the input param exists).
addGrid(DoubleParam, double[]) - Method in class org.apache.spark.ml.tuning.ParamGridBuilder
Adds a double param with multiple values.
addGrid(IntParam, int[]) - Method in class org.apache.spark.ml.tuning.ParamGridBuilder
Adds an int param with multiple values.
addGrid(FloatParam, float[]) - Method in class org.apache.spark.ml.tuning.ParamGridBuilder
Adds a float param with multiple values.
addGrid(LongParam, long[]) - Method in class org.apache.spark.ml.tuning.ParamGridBuilder
Adds a long param with multiple values.
addGrid(BooleanParam) - Method in class org.apache.spark.ml.tuning.ParamGridBuilder
Adds a boolean param with true and false.
addHadoopProperties(StoreTypes.PairStrings) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings hadoop_properties = 3;
addHadoopProperties(int, StoreTypes.PairStrings) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings hadoop_properties = 3;
addHadoopProperties(StoreTypes.PairStrings.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings hadoop_properties = 3;
addHadoopProperties(int, StoreTypes.PairStrings.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings hadoop_properties = 3;
addHadoopPropertiesBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings hadoop_properties = 3;
addHadoopPropertiesBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings hadoop_properties = 3;
addIncomingEdges(StoreTypes.RDDOperationEdge) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationEdge incoming_edges = 4;
addIncomingEdges(int, StoreTypes.RDDOperationEdge) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationEdge incoming_edges = 4;
addIncomingEdges(StoreTypes.RDDOperationEdge.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationEdge incoming_edges = 4;
addIncomingEdges(int, StoreTypes.RDDOperationEdge.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationEdge incoming_edges = 4;
addIncomingEdgesBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationEdge incoming_edges = 4;
addIncomingEdgesBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationEdge incoming_edges = 4;
addInputBytes(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double input_bytes = 6;
addInputRecords(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double input_records = 7;
addJar(String) - Method in class org.apache.spark.api.java.JavaSparkContext
Adds a JAR dependency for all tasks to be executed on this SparkContext in the future.
addJar(String) - Method in class org.apache.spark.launcher.AbstractLauncher
Adds a jar file to be submitted with the application.
addJar(String) - Method in class org.apache.spark.launcher.SparkLauncher
 
addJar(String) - Method in class org.apache.spark.SparkContext
Adds a JAR dependency for all tasks to be executed on this SparkContext in the future.
addJarsToClassPath(String, MutableURLClassLoader) - Static method in class org.apache.spark.util.DependencyUtils
 
addJarToClasspath(String, MutableURLClassLoader) - Static method in class org.apache.spark.util.DependencyUtils
 
addJobIds(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper.Builder
repeated int64 job_ids = 2;
addJvmGcTime(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double jvm_gc_time = 8;
addKilledTasks(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double killed_tasks = 5;
addListener(SparkAppHandle.Listener) - Method in interface org.apache.spark.launcher.SparkAppHandle
Adds a listener to be notified of changes to the handle's information.
addListener(StreamingQueryListener) - Method in class org.apache.spark.sql.streaming.StreamingQueryManager
Register a StreamingQueryListener to receive up-calls for life cycle events of StreamingQuery.
addListener(L) - Method in interface org.apache.spark.util.ListenerBus
Add a listener to listen events.
addLocalBlocksFetched(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
repeated double local_blocks_fetched = 4;
addLocalConfiguration(String, int, int, int, JobConf) - Static method in class org.apache.spark.rdd.HadoopRDD
Add Hadoop configuration specific to a single partition and attempt.
addLocalMergedBlocksFetched(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
repeated double local_merged_blocks_fetched = 4;
addLocalMergedBytesRead(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
repeated double local_merged_bytes_read = 8;
addLocalMergedChunksFetched(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
repeated double local_merged_chunks_fetched = 6;
addLong(long) - Method in class org.apache.spark.util.sketch.CountMinSketch
Increments item's count by one.
addLong(long, long) - Method in class org.apache.spark.util.sketch.CountMinSketch
Increments item's count by count.
addMapOutput(int, MapStatus) - Method in class org.apache.spark.ShuffleStatus
Register a map output.
addMemoryBytesSpilled(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double memory_bytes_spilled = 14;
addMemoryBytesSpilled(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double memory_bytes_spilled = 13;
addMergedFetchFallbackCount(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
repeated double merged_fetch_fallback_count = 2;
addMergeResult(int, org.apache.spark.scheduler.MergeStatus) - Method in class org.apache.spark.ShuffleStatus
Register a merge result.
addMetrics(TaskMetrics, TaskMetrics) - Static method in class org.apache.spark.status.LiveEntityHelpers
Add m2 values to m1.
addMetrics(StoreTypes.SQLPlanMetric) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 5;
addMetrics(int, StoreTypes.SQLPlanMetric) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 5;
addMetrics(StoreTypes.SQLPlanMetric.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 5;
addMetrics(int, StoreTypes.SQLPlanMetric.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 5;
addMetrics(StoreTypes.SQLPlanMetric) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 4;
addMetrics(int, StoreTypes.SQLPlanMetric) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 4;
addMetrics(StoreTypes.SQLPlanMetric.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 4;
addMetrics(int, StoreTypes.SQLPlanMetric.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 4;
addMetrics(StoreTypes.SQLPlanMetric) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 7;
addMetrics(int, StoreTypes.SQLPlanMetric) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 7;
addMetrics(StoreTypes.SQLPlanMetric.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 7;
addMetrics(int, StoreTypes.SQLPlanMetric.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 7;
addMetricsBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 5;
addMetricsBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 5;
addMetricsBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 4;
addMetricsBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 4;
addMetricsBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 7;
addMetricsBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 7;
addMetricsProperties(StoreTypes.PairStrings) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings metrics_properties = 5;
addMetricsProperties(int, StoreTypes.PairStrings) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings metrics_properties = 5;
addMetricsProperties(StoreTypes.PairStrings.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings metrics_properties = 5;
addMetricsProperties(int, StoreTypes.PairStrings.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings metrics_properties = 5;
addMetricsPropertiesBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings metrics_properties = 5;
addMetricsPropertiesBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings metrics_properties = 5;
addNaN() - Method in class org.apache.spark.sql.util.SQLOpenHashSet
 
addNewDefaultColumnToExistingTableNotAllowed(String, String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
addNewFunctionMismatchedWithFunctionError(String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
addNodes(StoreTypes.SparkPlanGraphNodeWrapper) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 4;
addNodes(int, StoreTypes.SparkPlanGraphNodeWrapper) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 4;
addNodes(StoreTypes.SparkPlanGraphNodeWrapper.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 4;
addNodes(int, StoreTypes.SparkPlanGraphNodeWrapper.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 4;
addNodes(StoreTypes.SparkPlanGraphNodeWrapper) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 2;
addNodes(int, StoreTypes.SparkPlanGraphNodeWrapper) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 2;
addNodes(StoreTypes.SparkPlanGraphNodeWrapper.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 2;
addNodes(int, StoreTypes.SparkPlanGraphNodeWrapper.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 2;
addNodesBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 4;
addNodesBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 4;
addNodesBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 2;
addNodesBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 2;
addNull() - Method in class org.apache.spark.sql.util.SQLOpenHashSet
 
addOutgoingEdges(StoreTypes.RDDOperationEdge) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationEdge outgoing_edges = 3;
addOutgoingEdges(int, StoreTypes.RDDOperationEdge) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationEdge outgoing_edges = 3;
addOutgoingEdges(StoreTypes.RDDOperationEdge.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationEdge outgoing_edges = 3;
addOutgoingEdges(int, StoreTypes.RDDOperationEdge.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationEdge outgoing_edges = 3;
addOutgoingEdgesBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationEdge outgoing_edges = 3;
addOutgoingEdgesBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationEdge outgoing_edges = 3;
addOutputBytes(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double output_bytes = 8;
addOutputRecords(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double output_records = 9;
addPartition(LiveRDDPartition) - Method in class org.apache.spark.status.RDDPartitionSeq
 
addPartitions(StoreTypes.RDDPartitionInfo) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
repeated .org.apache.spark.status.protobuf.RDDPartitionInfo partitions = 9;
addPartitions(int, StoreTypes.RDDPartitionInfo) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
repeated .org.apache.spark.status.protobuf.RDDPartitionInfo partitions = 9;
addPartitions(StoreTypes.RDDPartitionInfo.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
repeated .org.apache.spark.status.protobuf.RDDPartitionInfo partitions = 9;
addPartitions(int, StoreTypes.RDDPartitionInfo.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
repeated .org.apache.spark.status.protobuf.RDDPartitionInfo partitions = 9;
addPartitionsBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
repeated .org.apache.spark.status.protobuf.RDDPartitionInfo partitions = 9;
addPartitionsBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
repeated .org.apache.spark.status.protobuf.RDDPartitionInfo partitions = 9;
addPartToPGroup(Partition, PartitionGroup) - Method in class org.apache.spark.rdd.DefaultPartitionCoalescer
 
addPeakExecutionMemory(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double peak_execution_memory = 12;
addPyFile(String) - Method in class org.apache.spark.launcher.AbstractLauncher
Adds a python file / zip / egg to be submitted with the application.
addPyFile(String) - Method in class org.apache.spark.launcher.SparkLauncher
 
addQuantiles(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double quantiles = 1;
addQuantiles(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions.Builder
repeated double quantiles = 1;
addQuantiles(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double quantiles = 1;
addRddIds(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
repeated int64 rdd_ids = 43;
addReadBytes(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
repeated double read_bytes = 1;
addReadRecords(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
repeated double read_records = 2;
addRecordsRead(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributions.Builder
repeated double records_read = 2;
addRecordsWritten(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributions.Builder
repeated double records_written = 2;
addRemoteBlocksFetched(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
repeated double remote_blocks_fetched = 3;
addRemoteBytesRead(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
repeated double remote_bytes_read = 6;
addRemoteBytesReadToDisk(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
repeated double remote_bytes_read_to_disk = 7;
addRemoteMergedBlocksFetched(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
repeated double remote_merged_blocks_fetched = 3;
addRemoteMergedBytesRead(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
repeated double remote_merged_bytes_read = 7;
addRemoteMergedChunksFetched(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
repeated double remote_merged_chunks_fetched = 5;
addRemoteMergedReqsDuration(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
repeated double remote_merged_reqs_duration = 9;
addRemoteReqsDuration(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
repeated double remote_reqs_duration = 9;
addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo.Builder
 
addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo.Builder
 
addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
 
addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoWrapper.Builder
 
addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
 
addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoWrapper.Builder
 
addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.AppSummary.Builder
 
addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
 
addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetrics.Builder
 
addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
 
addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions.Builder
 
addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest.Builder
 
addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
 
addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper.Builder
 
addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
 
addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryWrapper.Builder
 
addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributions.Builder
 
addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetrics.Builder
 
addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
 
addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper.Builder
 
addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.MemoryMetrics.Builder
 
addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributions.Builder
 
addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetrics.Builder
 
addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.PairStrings.Builder
 
addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.PoolData.Builder
 
addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder
 
addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryWrapper.Builder
 
addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution.Builder
 
addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
 
addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationEdge.Builder
 
addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
 
addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode.Builder
 
addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo.Builder
 
addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
 
addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoWrapper.Builder
 
addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceInformation.Builder
 
addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo.Builder
 
addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileWrapper.Builder
 
addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.RuntimeInfo.Builder
 
addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
 
addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics.Builder
 
addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
 
addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics.Builder
 
addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions.Builder
 
addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetrics.Builder
 
addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress.Builder
 
addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
 
addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
 
addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphEdge.Builder
 
addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
 
addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper.Builder
 
addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
 
addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummary.Builder
 
addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryWrapper.Builder
 
addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
 
addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetric.Builder
 
addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
 
addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper.Builder
 
addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
 
addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
 
addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData.Builder
 
addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
 
addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressWrapper.Builder
 
addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
 
addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
 
addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
 
addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
 
addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskResourceRequest.Builder
 
addRequest(TaskResourceRequest) - Method in class org.apache.spark.resource.TaskResourceRequests
Add a certain TaskResourceRequest to the request set.
addResourceProfiles(StoreTypes.ResourceProfileInfo) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.ResourceProfileInfo resource_profiles = 7;
addResourceProfiles(int, StoreTypes.ResourceProfileInfo) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.ResourceProfileInfo resource_profiles = 7;
addResourceProfiles(StoreTypes.ResourceProfileInfo.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.ResourceProfileInfo resource_profiles = 7;
addResourceProfiles(int, StoreTypes.ResourceProfileInfo.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.ResourceProfileInfo resource_profiles = 7;
addResourceProfilesBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.ResourceProfileInfo resource_profiles = 7;
addResourceProfilesBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.ResourceProfileInfo resource_profiles = 7;
address() - Method in class org.apache.spark.BarrierTaskInfo
 
address() - Method in class org.apache.spark.status.api.v1.RDDDataDistribution
 
ADDRESS_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution
 
addresses() - Method in class org.apache.spark.resource.ResourceInformation
 
addresses() - Method in class org.apache.spark.resource.ResourceInformationJson
 
ADDRESSES_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.ResourceInformation
 
addResultSerializationTime(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double result_serialization_time = 9;
addResultSize(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double result_size = 7;
addSchedulable(Schedulable) - Method in interface org.apache.spark.scheduler.Schedulable
 
addSchedulerDelay(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double scheduler_delay = 11;
addShuffleRead(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double shuffle_read = 10;
addShuffleReadRecords(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double shuffle_read_records = 11;
addShuffleWrite(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double shuffle_write = 12;
addShuffleWriteRecords(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double shuffle_write_records = 13;
addShutdownHook(Function0<BoxedUnit>) - Static method in class org.apache.spark.util.ShutdownHookManager
Adds a shutdown hook with default priority.
addShutdownHook(int, Function0<BoxedUnit>) - Static method in class org.apache.spark.util.ShutdownHookManager
Adds a shutdown hook with the given priority.
addSkippedStages(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper.Builder
repeated int32 skipped_stages = 2;
addSources(StoreTypes.SourceProgress) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
repeated .org.apache.spark.status.protobuf.SourceProgress sources = 10;
addSources(int, StoreTypes.SourceProgress) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
repeated .org.apache.spark.status.protobuf.SourceProgress sources = 10;
addSources(StoreTypes.SourceProgress.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
repeated .org.apache.spark.status.protobuf.SourceProgress sources = 10;
addSources(int, StoreTypes.SourceProgress.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
repeated .org.apache.spark.status.protobuf.SourceProgress sources = 10;
addSourcesBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
repeated .org.apache.spark.status.protobuf.SourceProgress sources = 10;
addSourcesBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
repeated .org.apache.spark.status.protobuf.SourceProgress sources = 10;
addSparkArg(String) - Method in class org.apache.spark.launcher.AbstractLauncher
Adds a no-value argument to the Spark invocation.
addSparkArg(String, String) - Method in class org.apache.spark.launcher.AbstractLauncher
Adds an argument with a value to the Spark invocation.
addSparkArg(String) - Method in class org.apache.spark.launcher.SparkLauncher
 
addSparkArg(String, String) - Method in class org.apache.spark.launcher.SparkLauncher
 
addSparkListener(SparkListenerInterface) - Method in class org.apache.spark.SparkContext
:: DeveloperApi :: Register a listener to receive up-calls from events that happen during execution.
addSparkProperties(StoreTypes.PairStrings) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings spark_properties = 2;
addSparkProperties(int, StoreTypes.PairStrings) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings spark_properties = 2;
addSparkProperties(StoreTypes.PairStrings.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings spark_properties = 2;
addSparkProperties(int, StoreTypes.PairStrings.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings spark_properties = 2;
addSparkPropertiesBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings spark_properties = 2;
addSparkPropertiesBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings spark_properties = 2;
addStageIds(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
repeated int64 stage_ids = 6;
addStageIds(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.PoolData.Builder
repeated int64 stage_ids = 2;
addStages(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
repeated int64 stages = 12;
addStateOperators(StoreTypes.StateOperatorProgress) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
repeated .org.apache.spark.status.protobuf.StateOperatorProgress state_operators = 9;
addStateOperators(int, StoreTypes.StateOperatorProgress) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
repeated .org.apache.spark.status.protobuf.StateOperatorProgress state_operators = 9;
addStateOperators(StoreTypes.StateOperatorProgress.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
repeated .org.apache.spark.status.protobuf.StateOperatorProgress state_operators = 9;
addStateOperators(int, StoreTypes.StateOperatorProgress.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
repeated .org.apache.spark.status.protobuf.StateOperatorProgress state_operators = 9;
addStateOperatorsBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
repeated .org.apache.spark.status.protobuf.StateOperatorProgress state_operators = 9;
addStateOperatorsBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
repeated .org.apache.spark.status.protobuf.StateOperatorProgress state_operators = 9;
addStreamingListener(StreamingListener) - Method in class org.apache.spark.streaming.api.java.JavaStreamingContext
Deprecated.
Add a StreamingListener object for receiving system events related to streaming.
addStreamingListener(StreamingListener) - Method in class org.apache.spark.streaming.StreamingContext
Deprecated.
Add a StreamingListener object for receiving system events related to streaming.
addString(String) - Method in class org.apache.spark.util.sketch.CountMinSketch
Increments item's count by one.
addString(String, long) - Method in class org.apache.spark.util.sketch.CountMinSketch
Increments item's count by count.
addSucceededTasks(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double succeeded_tasks = 4;
addSystemProperties(StoreTypes.PairStrings) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings system_properties = 4;
addSystemProperties(int, StoreTypes.PairStrings) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings system_properties = 4;
addSystemProperties(StoreTypes.PairStrings.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings system_properties = 4;
addSystemProperties(int, StoreTypes.PairStrings.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings system_properties = 4;
addSystemPropertiesBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings system_properties = 4;
addSystemPropertiesBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings system_properties = 4;
addTaskCompletionListener(TaskCompletionListener) - Method in class org.apache.spark.BarrierTaskContext
 
addTaskCompletionListener(TaskCompletionListener) - Method in class org.apache.spark.TaskContext
Adds a (Java friendly) listener to be executed on task completion.
addTaskCompletionListener(Function1<TaskContext, U>) - Method in class org.apache.spark.TaskContext
Adds a listener in the form of a Scala closure to be executed on task completion.
addTaskFailureListener(TaskFailureListener) - Method in class org.apache.spark.BarrierTaskContext
 
addTaskFailureListener(TaskFailureListener) - Method in class org.apache.spark.TaskContext
Adds a listener to be executed on task failure (which includes completion listener failure, if the task body did not already fail).
addTaskFailureListener(Function2<TaskContext, Throwable, BoxedUnit>) - Method in class org.apache.spark.TaskContext
Adds a listener to be executed on task failure (which includes completion listener failure, if the task body did not already fail).
addTaskResourceRequests(SparkConf, TaskResourceRequests) - Static method in class org.apache.spark.resource.ResourceUtils
 
addTaskSetManager(Schedulable, Properties) - Method in interface org.apache.spark.scheduler.SchedulableBuilder
 
addTaskTime(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double task_time = 2;
addTime() - Method in class org.apache.spark.status.api.v1.ExecutorSummary
 
addTime() - Method in class org.apache.spark.status.api.v1.ProcessSummary
 
addTotalBlocksFetched(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
repeated double total_blocks_fetched = 8;
addURL(URL) - Method in class org.apache.spark.util.MutableURLClassLoader
 
AddWebUIFilter(String, Map<String, String>, String) - Constructor for class org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.AddWebUIFilter
 
AddWebUIFilter$() - Constructor for class org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.AddWebUIFilter$
 
addWriteBytes(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions.Builder
repeated double write_bytes = 1;
addWriteRecords(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions.Builder
repeated double write_records = 2;
addWriteTime(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions.Builder
repeated double write_time = 3;
aesCryptoError(String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
aesModeUnsupportedError(String, String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
after(String) - Static method in interface org.apache.spark.sql.connector.catalog.TableChange.ColumnPosition
 
AFTSurvivalRegression - Class in org.apache.spark.ml.regression
Fit a parametric survival regression model named accelerated failure time (AFT) model (see Accelerated failure time model (Wikipedia)) based on the Weibull distribution of the survival time.
AFTSurvivalRegression(String) - Constructor for class org.apache.spark.ml.regression.AFTSurvivalRegression
 
AFTSurvivalRegression() - Constructor for class org.apache.spark.ml.regression.AFTSurvivalRegression
 
AFTSurvivalRegressionModel - Class in org.apache.spark.ml.regression
Model produced by AFTSurvivalRegression.
AFTSurvivalRegressionParams - Interface in org.apache.spark.ml.regression
Params for accelerated failure time (AFT) regression.
agg(Column, Column...) - Method in class org.apache.spark.sql.Dataset
Aggregates on the entire Dataset without groups.
agg(Tuple2<String, String>, Seq<Tuple2<String, String>>) - Method in class org.apache.spark.sql.Dataset
(Scala-specific) Aggregates on the entire Dataset without groups.
agg(Map<String, String>) - Method in class org.apache.spark.sql.Dataset
(Scala-specific) Aggregates on the entire Dataset without groups.
agg(Map<String, String>) - Method in class org.apache.spark.sql.Dataset
(Java-specific) Aggregates on the entire Dataset without groups.
agg(Column, Seq<Column>) - Method in class org.apache.spark.sql.Dataset
Aggregates on the entire Dataset without groups.
agg(TypedColumn<V, U1>) - Method in class org.apache.spark.sql.KeyValueGroupedDataset
Computes the given aggregation, returning a Dataset of tuples for each unique key and the result of computing this aggregation over all elements in the group.
agg(TypedColumn<V, U1>, TypedColumn<V, U2>) - Method in class org.apache.spark.sql.KeyValueGroupedDataset
Computes the given aggregations, returning a Dataset of tuples for each unique key and the result of computing these aggregations over all elements in the group.
agg(TypedColumn<V, U1>, TypedColumn<V, U2>, TypedColumn<V, U3>) - Method in class org.apache.spark.sql.KeyValueGroupedDataset
Computes the given aggregations, returning a Dataset of tuples for each unique key and the result of computing these aggregations over all elements in the group.
agg(TypedColumn<V, U1>, TypedColumn<V, U2>, TypedColumn<V, U3>, TypedColumn<V, U4>) - Method in class org.apache.spark.sql.KeyValueGroupedDataset
Computes the given aggregations, returning a Dataset of tuples for each unique key and the result of computing these aggregations over all elements in the group.
agg(TypedColumn<V, U1>, TypedColumn<V, U2>, TypedColumn<V, U3>, TypedColumn<V, U4>, TypedColumn<V, U5>) - Method in class org.apache.spark.sql.KeyValueGroupedDataset
Computes the given aggregations, returning a Dataset of tuples for each unique key and the result of computing these aggregations over all elements in the group.
agg(TypedColumn<V, U1>, TypedColumn<V, U2>, TypedColumn<V, U3>, TypedColumn<V, U4>, TypedColumn<V, U5>, TypedColumn<V, U6>) - Method in class org.apache.spark.sql.KeyValueGroupedDataset
Computes the given aggregations, returning a Dataset of tuples for each unique key and the result of computing these aggregations over all elements in the group.
agg(TypedColumn<V, U1>, TypedColumn<V, U2>, TypedColumn<V, U3>, TypedColumn<V, U4>, TypedColumn<V, U5>, TypedColumn<V, U6>, TypedColumn<V, U7>) - Method in class org.apache.spark.sql.KeyValueGroupedDataset
Computes the given aggregations, returning a Dataset of tuples for each unique key and the result of computing these aggregations over all elements in the group.
agg(TypedColumn<V, U1>, TypedColumn<V, U2>, TypedColumn<V, U3>, TypedColumn<V, U4>, TypedColumn<V, U5>, TypedColumn<V, U6>, TypedColumn<V, U7>, TypedColumn<V, U8>) - Method in class org.apache.spark.sql.KeyValueGroupedDataset
Computes the given aggregations, returning a Dataset of tuples for each unique key and the result of computing these aggregations over all elements in the group.
agg(Column, Column...) - Method in class org.apache.spark.sql.RelationalGroupedDataset
Compute aggregates by specifying a series of aggregate columns.
agg(Tuple2<String, String>, Seq<Tuple2<String, String>>) - Method in class org.apache.spark.sql.RelationalGroupedDataset
(Scala-specific) Compute aggregates by specifying the column names and aggregate methods.
agg(Map<String, String>) - Method in class org.apache.spark.sql.RelationalGroupedDataset
(Scala-specific) Compute aggregates by specifying a map from column name to aggregate methods.
agg(Map<String, String>) - Method in class org.apache.spark.sql.RelationalGroupedDataset
(Java-specific) Compute aggregates by specifying a map from column name to aggregate methods.
agg(Column, Seq<Column>) - Method in class org.apache.spark.sql.RelationalGroupedDataset
Compute aggregates by specifying a series of aggregate columns.
aggregate(U, Function2<U, T, U>, Function2<U, U, U>) - Method in interface org.apache.spark.api.java.JavaRDDLike
Aggregate the elements of each partition, and then the results for all the partitions, using given combine functions and a neutral "zero value".
aggregate(U, Function2<U, T, U>, Function2<U, U, U>, ClassTag<U>) - Method in class org.apache.spark.rdd.RDD
Aggregate the elements of each partition, and then the results for all the partitions, using given combine functions and a neutral "zero value".
aggregate(Column, Column, Function2<Column, Column, Column>, Function1<Column, Column>) - Static method in class org.apache.spark.sql.functions
Applies a binary operator to an initial state and all elements in the array, and reduces this to a single state.
aggregate(Column, Column, Function2<Column, Column, Column>) - Static method in class org.apache.spark.sql.functions
Applies a binary operator to an initial state and all elements in the array, and reduces this to a single state.
aggregateByKey(U, Partitioner, Function2<U, V, U>, Function2<U, U, U>) - Method in class org.apache.spark.api.java.JavaPairRDD
Aggregate the values of each key, using given combine functions and a neutral "zero value".
aggregateByKey(U, int, Function2<U, V, U>, Function2<U, U, U>) - Method in class org.apache.spark.api.java.JavaPairRDD
Aggregate the values of each key, using given combine functions and a neutral "zero value".
aggregateByKey(U, Function2<U, V, U>, Function2<U, U, U>) - Method in class org.apache.spark.api.java.JavaPairRDD
Aggregate the values of each key, using given combine functions and a neutral "zero value".
aggregateByKey(U, Partitioner, Function2<U, V, U>, Function2<U, U, U>, ClassTag<U>) - Method in class org.apache.spark.rdd.PairRDDFunctions
Aggregate the values of each key, using given combine functions and a neutral "zero value".
aggregateByKey(U, int, Function2<U, V, U>, Function2<U, U, U>, ClassTag<U>) - Method in class org.apache.spark.rdd.PairRDDFunctions
Aggregate the values of each key, using given combine functions and a neutral "zero value".
aggregateByKey(U, Function2<U, V, U>, Function2<U, U, U>, ClassTag<U>) - Method in class org.apache.spark.rdd.PairRDDFunctions
Aggregate the values of each key, using given combine functions and a neutral "zero value".
AggregatedDialect - Class in org.apache.spark.sql.jdbc
AggregatedDialect can unify multiple dialects into one virtual Dialect.
AggregatedDialect(List<JdbcDialect>) - Constructor for class org.apache.spark.sql.jdbc.AggregatedDialect
 
aggregateExpressionRequiredForPivotError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
aggregateExpressions() - Method in class org.apache.spark.sql.connector.expressions.aggregate.Aggregation
 
AggregateFunc - Interface in org.apache.spark.sql.connector.expressions.aggregate
Base class of the Aggregate Functions.
AggregateFunction<S extends java.io.Serializable,R> - Interface in org.apache.spark.sql.connector.catalog.functions
Interface for a function that produces a result value by aggregating over multiple input rows.
aggregateInAggregateFilterError() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
aggregateMessages(Function1<EdgeContext<VD, ED, A>, BoxedUnit>, Function2<A, A, A>, TripletFields, ClassTag<A>) - Method in class org.apache.spark.graphx.Graph
Aggregates values from the neighboring edges and vertices of each vertex.
aggregateMessagesWithActiveSet(Function1<EdgeContext<VD, ED, A>, BoxedUnit>, Function2<A, A, A>, TripletFields, Option<Tuple2<VertexRDD<?>, EdgeDirection>>, ClassTag<A>) - Method in class org.apache.spark.graphx.impl.GraphImpl
 
aggregateTaskMetrics(long[]) - Method in class org.apache.spark.sql.connector.metric.CustomAvgMetric
 
aggregateTaskMetrics(long[]) - Method in interface org.apache.spark.sql.connector.metric.CustomMetric
Given an array of task metric values, returns aggregated final metric value.
aggregateTaskMetrics(long[]) - Method in class org.apache.spark.sql.connector.metric.CustomSumMetric
 
aggregateUsingIndex(RDD<Tuple2<Object, VD2>>, Function2<VD2, VD2, VD2>, ClassTag<VD2>) - Method in class org.apache.spark.graphx.impl.VertexRDDImpl
 
aggregateUsingIndex(RDD<Tuple2<Object, VD2>>, Function2<VD2, VD2, VD2>, ClassTag<VD2>) - Method in class org.apache.spark.graphx.VertexRDD
Aggregates vertices in messages that have the same ids using reduceFunc, returning a VertexRDD co-indexed with this.
AggregatingEdgeContext<VD,ED,A> - Class in org.apache.spark.graphx.impl
 
AggregatingEdgeContext(Function2<A, A, A>, Object, BitSet) - Constructor for class org.apache.spark.graphx.impl.AggregatingEdgeContext
 
Aggregation - Class in org.apache.spark.sql.connector.expressions.aggregate
Aggregation in SQL statement.
Aggregation(AggregateFunc[], Expression[]) - Constructor for class org.apache.spark.sql.connector.expressions.aggregate.Aggregation
 
aggregationDepth() - Method in class org.apache.spark.ml.classification.LinearSVC
 
aggregationDepth() - Method in class org.apache.spark.ml.classification.LinearSVCModel
 
aggregationDepth() - Method in class org.apache.spark.ml.classification.LogisticRegression
 
aggregationDepth() - Method in class org.apache.spark.ml.classification.LogisticRegressionModel
 
aggregationDepth() - Method in class org.apache.spark.ml.clustering.GaussianMixture
 
aggregationDepth() - Method in class org.apache.spark.ml.clustering.GaussianMixtureModel
 
aggregationDepth() - Method in interface org.apache.spark.ml.param.shared.HasAggregationDepth
Param for suggested depth for treeAggregate (&gt;= 2).
aggregationDepth() - Method in class org.apache.spark.ml.regression.AFTSurvivalRegression
 
aggregationDepth() - Method in class org.apache.spark.ml.regression.AFTSurvivalRegressionModel
 
aggregationDepth() - Method in class org.apache.spark.ml.regression.GeneralizedLinearRegression
 
aggregationDepth() - Method in class org.apache.spark.ml.regression.GeneralizedLinearRegressionModel
 
aggregationDepth() - Method in class org.apache.spark.ml.regression.LinearRegression
 
aggregationDepth() - Method in class org.apache.spark.ml.regression.LinearRegressionModel
 
aggregationFunctionAppliedOnNonNumericColumnError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
aggregationFunctionAppliedOnNonNumericColumnError(String, int) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
Aggregator<K,V,C> - Class in org.apache.spark
:: DeveloperApi :: A set of functions used to aggregate data.
Aggregator(Function1<V, C>, Function2<C, V, C>, Function2<C, C, C>) - Constructor for class org.apache.spark.Aggregator
 
aggregator() - Method in class org.apache.spark.ShuffleDependency
 
Aggregator<IN,BUF,OUT> - Class in org.apache.spark.sql.expressions
A base class for user-defined aggregations, which can be used in Dataset operations to take all of the elements of a group and reduce them to a single value.
Aggregator() - Constructor for class org.apache.spark.sql.expressions.Aggregator
 
aic(RDD<Tuple3<Object, Object, Object>>, double, double, double) - Method in class org.apache.spark.ml.regression.GeneralizedLinearRegression.Binomial$
 
aic(RDD<Tuple3<Object, Object, Object>>, double, double, double) - Method in class org.apache.spark.ml.regression.GeneralizedLinearRegression.Gamma$
 
aic(RDD<Tuple3<Object, Object, Object>>, double, double, double) - Method in class org.apache.spark.ml.regression.GeneralizedLinearRegression.Gaussian$
 
aic(RDD<Tuple3<Object, Object, Object>>, double, double, double) - Method in class org.apache.spark.ml.regression.GeneralizedLinearRegression.Poisson$
 
aic() - Method in class org.apache.spark.ml.regression.GeneralizedLinearRegressionSummary
 
Algo - Class in org.apache.spark.mllib.tree.configuration
Enum to select the algorithm for the decision tree
Algo() - Constructor for class org.apache.spark.mllib.tree.configuration.Algo
 
algo() - Method in class org.apache.spark.mllib.tree.configuration.Strategy
 
algo() - Method in class org.apache.spark.mllib.tree.model.DecisionTreeModel
 
algo() - Method in class org.apache.spark.mllib.tree.model.GradientBoostedTreesModel
 
algo() - Method in class org.apache.spark.mllib.tree.model.RandomForestModel
 
algorithm() - Method in class org.apache.spark.mllib.regression.StreamingLinearRegressionWithSGD
 
alias(String) - Method in class org.apache.spark.sql.Column
Gives the column an alias.
alias(String) - Method in class org.apache.spark.sql.Dataset
Returns a new Dataset with an alias set.
alias(Symbol) - Method in class org.apache.spark.sql.Dataset
(Scala-specific) Returns a new Dataset with an alias set.
aliasesNumberNotMatchUDTFOutputError(int, String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
aliasNumberNotMatchColumnNumberError(int, int, TreeNode<?>) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
All - Static variable in class org.apache.spark.graphx.TripletFields
Expose all the fields (source, edge, and destination).
ALL_GATHER() - Static method in class org.apache.spark.RequestMethod
 
ALL_REMOVALS_TIME_MS_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress
 
ALL_UPDATES_TIME_MS_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress
 
allAvailable() - Static method in interface org.apache.spark.sql.connector.read.streaming.ReadLimit
 
allGather(String) - Method in class org.apache.spark.BarrierTaskContext
:: Experimental :: Blocks until all tasks in the same stage have reached this routine.
AllJobsCancelled - Class in org.apache.spark.scheduler
 
AllJobsCancelled() - Constructor for class org.apache.spark.scheduler.AllJobsCancelled
 
allocate(int) - Method in class org.apache.spark.sql.util.NumericHistogram
Sets the number of histogram bins to use for approximating data.
allocator() - Method in class org.apache.spark.storage.memory.SerializedValuesHolder
 
AllReceiverIds - Class in org.apache.spark.streaming.scheduler
A message used by ReceiverTracker to ask all receiver's ids still stored in ReceiverTrackerEndpoint.
AllReceiverIds() - Constructor for class org.apache.spark.streaming.scheduler.AllReceiverIds
 
allRemovalsTimeMs() - Method in class org.apache.spark.sql.streaming.StateOperatorProgress
 
allSources() - Static method in class org.apache.spark.metrics.source.StaticSources
The set of all static sources.
allSupportedExecutorResources() - Static method in class org.apache.spark.resource.ResourceProfile
Return all supported Spark built-in executor resources, custom resources like GPUs/FPGAs are excluded.
allUpdatesTimeMs() - Method in class org.apache.spark.sql.streaming.StateOperatorProgress
 
alpha() - Method in class org.apache.spark.ml.recommendation.ALS
 
alpha() - Method in interface org.apache.spark.ml.recommendation.ALSParams
Param for the alpha parameter in the implicit preference formulation (nonnegative).
alpha() - Method in class org.apache.spark.mllib.random.WeibullGenerator
 
ALS - Class in org.apache.spark.ml.recommendation
Alternating Least Squares (ALS) matrix factorization.
ALS(String) - Constructor for class org.apache.spark.ml.recommendation.ALS
 
ALS() - Constructor for class org.apache.spark.ml.recommendation.ALS
 
ALS - Class in org.apache.spark.mllib.recommendation
Alternating Least Squares matrix factorization.
ALS() - Constructor for class org.apache.spark.mllib.recommendation.ALS
Constructs an ALS instance with default parameters: {numBlocks: -1, rank: 10, iterations: 10, lambda: 0.01, implicitPrefs: false, alpha: 1.0}.
ALS.InBlock$ - Class in org.apache.spark.ml.recommendation
 
ALS.LeastSquaresNESolver - Interface in org.apache.spark.ml.recommendation
Trait for least squares solvers applied to the normal equation.
ALS.Rating<ID> - Class in org.apache.spark.ml.recommendation
Rating class for better code readability.
ALS.Rating$ - Class in org.apache.spark.ml.recommendation
 
ALS.RatingBlock$ - Class in org.apache.spark.ml.recommendation
 
ALSModel - Class in org.apache.spark.ml.recommendation
Model fitted by ALS.
ALSModelParams - Interface in org.apache.spark.ml.recommendation
Common params for ALS and ALSModel.
ALSParams - Interface in org.apache.spark.ml.recommendation
Common params for ALS.
alterAddColNotSupportDatasourceTableError(Object, TableIdentifier) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
alterAddColNotSupportViewError(TableIdentifier) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
alterColumnCannotFindColumnInV1TableError(String, org.apache.spark.sql.connector.catalog.V1Table) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
alterColumnWithV1TableCannotSpecifyNotNullError() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
alterDatabaseLocationUnsupportedError() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
alterNamespace(String[], NamespaceChange...) - Method in class org.apache.spark.sql.connector.catalog.DelegatingCatalogExtension
 
alterNamespace(String[], NamespaceChange...) - Method in interface org.apache.spark.sql.connector.catalog.SupportsNamespaces
Apply a set of metadata changes to a namespace in the catalog.
alterTable(Identifier, TableChange...) - Method in class org.apache.spark.sql.connector.catalog.DelegatingCatalogExtension
 
alterTable(Identifier, TableChange...) - Method in interface org.apache.spark.sql.connector.catalog.TableCatalog
Apply a set of changes to a table in the catalog.
alterTable(String, Seq<TableChange>, int) - Static method in class org.apache.spark.sql.jdbc.DB2Dialect
 
alterTable(String, Seq<TableChange>, int) - Static method in class org.apache.spark.sql.jdbc.DerbyDialect
 
alterTable(String, Seq<TableChange>, int) - Static method in class org.apache.spark.sql.jdbc.H2Dialect
 
alterTable(String, Seq<TableChange>, int) - Method in class org.apache.spark.sql.jdbc.JdbcDialect
Alter an existing table.
alterTable(String, Seq<TableChange>, int) - Static method in class org.apache.spark.sql.jdbc.MsSqlServerDialect
 
alterTable(String, Seq<TableChange>, int) - Static method in class org.apache.spark.sql.jdbc.MySQLDialect
 
alterTable(String, Seq<TableChange>, int) - Static method in class org.apache.spark.sql.jdbc.NoopDialect
 
alterTable(String, Seq<TableChange>, int) - Static method in class org.apache.spark.sql.jdbc.OracleDialect
 
alterTable(String, Seq<TableChange>, int) - Static method in class org.apache.spark.sql.jdbc.PostgresDialect
 
alterTable(String, Seq<TableChange>, int) - Static method in class org.apache.spark.sql.jdbc.TeradataDialect
 
alterTableChangeColumnNotSupportedForColumnTypeError(StructField, StructField) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
alterTableRecoverPartitionsNotSupportedForV2TablesError() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
alterTableSerDePropertiesNotSupportedForV2TablesError() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
alterTableSetSerdeForSpecificPartitionNotSupportedError() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
alterTableSetSerdeNotSupportedError() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
alterTableWithDropPartitionAndPurgeUnsupportedError() - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
alterV2TableSetLocationWithPartitionNotSupportedError() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
alterView(Identifier, ViewChange...) - Method in interface org.apache.spark.sql.connector.catalog.ViewCatalog
Apply changes to a view in the catalog.
AlwaysFalse - Class in org.apache.spark.sql.connector.expressions.filter
A predicate that always evaluates to false.
AlwaysFalse() - Constructor for class org.apache.spark.sql.connector.expressions.filter.AlwaysFalse
 
AlwaysFalse - Class in org.apache.spark.sql.sources
A filter that always evaluates to false.
AlwaysFalse() - Constructor for class org.apache.spark.sql.sources.AlwaysFalse
 
AlwaysTrue - Class in org.apache.spark.sql.connector.expressions.filter
A predicate that always evaluates to true.
AlwaysTrue() - Constructor for class org.apache.spark.sql.connector.expressions.filter.AlwaysTrue
 
AlwaysTrue - Class in org.apache.spark.sql.sources
A filter that always evaluates to true.
AlwaysTrue() - Constructor for class org.apache.spark.sql.sources.AlwaysTrue
 
am() - Method in class org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.RegisterClusterManager
 
ambiguousAttributesInSelfJoinError(Seq<AttributeReference>) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
ambiguousColumnOrFieldError(Seq<String>, int, Origin) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
ambiguousColumnOrFieldError(Seq<String>, int) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
ambiguousLateralColumnAliasError(String, int) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
ambiguousLateralColumnAliasError(Seq<String>, int) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
ambiguousReferenceError(String, Seq<Attribute>) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
ambiguousReferenceToFieldsError(String, int) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
ambiguousRelationAliasNameInNestedCTEError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
amount() - Method in class org.apache.spark.resource.ExecutorResourceRequest
 
amount() - Method in class org.apache.spark.resource.ResourceRequest
 
AMOUNT() - Static method in class org.apache.spark.resource.ResourceUtils
 
amount() - Method in class org.apache.spark.resource.TaskResourceRequest
 
AMOUNT_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest
 
AMOUNT_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.TaskResourceRequest
 
AnalysisException - Exception in org.apache.spark.sql
Thrown when a query fails to analyze, usually because the query itself is invalid.
AnalysisException(String, Map<String, String>, Option<Throwable>) - Constructor for exception org.apache.spark.sql.AnalysisException
 
AnalysisException(String, Map<String, String>, QueryContext[], String) - Constructor for exception org.apache.spark.sql.AnalysisException
 
AnalysisException(String, Map<String, String>) - Constructor for exception org.apache.spark.sql.AnalysisException
 
AnalysisException(String, Map<String, String>, Origin) - Constructor for exception org.apache.spark.sql.AnalysisException
 
AnalysisException(String, Map<String, String>, Origin, Option<Throwable>) - Constructor for exception org.apache.spark.sql.AnalysisException
 
analyzeTableNotSupportedForV2TablesError() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
analyzeTableNotSupportedOnViewsError() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
analyzingColumnStatisticsNotSupportedForColumnTypeError(String, DataType) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
and(Column) - Method in class org.apache.spark.sql.Column
Boolean AND.
And - Class in org.apache.spark.sql.connector.expressions.filter
A predicate that evaluates to true iff both left and right evaluate to true.
And(Predicate, Predicate) - Constructor for class org.apache.spark.sql.connector.expressions.filter.And
 
And - Class in org.apache.spark.sql.sources
A filter that evaluates to true iff both left or right evaluate to true.
And(Filter, Filter) - Constructor for class org.apache.spark.sql.sources.And
 
ANOVATest - Class in org.apache.spark.ml.stat
ANOVA Test for continuous data.
ANOVATest() - Constructor for class org.apache.spark.ml.stat.ANOVATest
 
ansiDateTimeError(Exception) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
ansiDateTimeParseError(Exception) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
ansiIllegalArgumentError(String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
ansiIllegalArgumentError(IllegalArgumentException) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
antecedent() - Method in class org.apache.spark.mllib.fpm.AssociationRules.Rule
 
ANY() - Static method in class org.apache.spark.scheduler.TaskLocality
 
AnyDataType - Class in org.apache.spark.sql.types
An AbstractDataType that matches any concrete data types.
AnyDataType() - Constructor for class org.apache.spark.sql.types.AnyDataType
 
anyNull() - Method in interface org.apache.spark.sql.Row
Returns true if there are any NULL values in this row.
anyNull() - Method in class org.apache.spark.sql.vectorized.ColumnarBatchRow
 
anyNull() - Method in class org.apache.spark.sql.vectorized.ColumnarRow
 
AnyTimestampType - Class in org.apache.spark.sql.types
 
AnyTimestampType() - Constructor for class org.apache.spark.sql.types.AnyTimestampType
 
ApiHelper - Class in org.apache.spark.ui.jobs
 
ApiHelper() - Constructor for class org.apache.spark.ui.jobs.ApiHelper
 
ApiRequestContext - Interface in org.apache.spark.status.api.v1
 
APP_SPARK_VERSION_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo
 
appAttemptId() - Method in class org.apache.spark.scheduler.SparkListenerApplicationStart
 
append() - Method in class org.apache.spark.sql.DataFrameWriterV2
Append the contents of the data frame to the output table.
Append() - Static method in class org.apache.spark.sql.streaming.OutputMode
OutputMode in which only the new rows in the streaming DataFrame/Dataset will be written to the sink.
appendBias(Vector) - Static method in class org.apache.spark.mllib.util.MLUtils
Returns a new vector with 1.0 (bias) appended to the input vector.
appendColumn(StructType, String, DataType, boolean) - Static method in class org.apache.spark.ml.util.SchemaUtils
Appends a new column to the input schema.
appendColumn(StructType, StructField) - Static method in class org.apache.spark.ml.util.SchemaUtils
Appends a new column to the input schema.
AppHistoryServerPlugin - Interface in org.apache.spark.status
An interface for creating history listeners(to replay event logs) defined in other modules like SQL, and setup the UI of the plugin to rebuild the history UI.
appId() - Method in class org.apache.spark.scheduler.SparkListenerApplicationStart
 
appId() - Method in interface org.apache.spark.status.api.v1.BaseAppResource
 
appId() - Method in class org.apache.spark.storage.ShuffleMergedDataBlockId
 
appId() - Method in class org.apache.spark.storage.ShuffleMergedIndexBlockId
 
appId() - Method in class org.apache.spark.storage.ShuffleMergedMetaBlockId
 
APPLICATION_EXECUTOR_LIMIT() - Static method in class org.apache.spark.ui.ToolTips
 
APPLICATION_MASTER() - Static method in class org.apache.spark.metrics.MetricsSystemInstances
 
applicationAttemptId() - Method in interface org.apache.spark.scheduler.SchedulerBackend
Get the attempt ID for this run, if the cluster manager supports multiple attempts.
applicationAttemptId() - Method in interface org.apache.spark.scheduler.TaskScheduler
Get an application's attempt ID associated with the job.
applicationAttemptId() - Method in class org.apache.spark.SparkContext
 
ApplicationAttemptInfo - Class in org.apache.spark.status.api.v1
 
applicationEndFromJson(JsonNode) - Static method in class org.apache.spark.util.JsonProtocol
 
applicationEndToJson(SparkListenerApplicationEnd, JsonGenerator) - Static method in class org.apache.spark.util.JsonProtocol
 
ApplicationEnvironmentInfo - Class in org.apache.spark.status.api.v1
 
applicationId() - Method in interface org.apache.spark.scheduler.SchedulerBackend
Get an application ID associated with the job.
applicationId() - Method in interface org.apache.spark.scheduler.TaskScheduler
Get an application ID associated with the job.
applicationId() - Method in class org.apache.spark.SparkContext
A unique identifier for the Spark application.
ApplicationInfo - Class in org.apache.spark.status.api.v1
 
APPLICATIONS() - Static method in class org.apache.spark.metrics.MetricsSystemInstances
 
applicationStartFromJson(JsonNode) - Static method in class org.apache.spark.util.JsonProtocol
 
applicationStartToJson(SparkListenerApplicationStart, JsonGenerator) - Static method in class org.apache.spark.util.JsonProtocol
 
ApplicationStatus - Enum in org.apache.spark.status.api.v1
 
apply(T1) - Static method in class org.apache.spark.CleanAccum
 
apply(T1) - Static method in class org.apache.spark.CleanBroadcast
 
apply(T1) - Static method in class org.apache.spark.CleanCheckpoint
 
apply(T1) - Static method in class org.apache.spark.CleanRDD
 
apply(T1) - Static method in class org.apache.spark.CleanShuffle
 
apply(T1) - Static method in class org.apache.spark.CleanSparkListener
 
apply(T1, T2) - Static method in class org.apache.spark.ContextBarrierId
 
apply(T1, T2, T3) - Static method in class org.apache.spark.ErrorInfo
 
apply(int) - Static method in class org.apache.spark.ErrorMessageFormat
 
apply(T1) - Static method in class org.apache.spark.ErrorSubInfo
 
apply(T1, T2, T3, T4, T5, T6, T7, T8) - Static method in class org.apache.spark.ExceptionFailure
 
apply(T1, T2, T3) - Static method in class org.apache.spark.ExecutorLostFailure
 
apply(T1) - Static method in class org.apache.spark.ExecutorRegistered
 
apply(T1) - Static method in class org.apache.spark.ExecutorRemoved
 
apply(T1, T2, T3, T4, T5, T6) - Static method in class org.apache.spark.FetchFailed
 
apply(RDD<Tuple2<Object, VD>>, RDD<Edge<ED>>, VD, StorageLevel, StorageLevel, ClassTag<VD>, ClassTag<ED>) - Static method in class org.apache.spark.graphx.Graph
Construct a graph from a collection of vertices and edges with attributes.
apply(RDD<Edge<ED>>, VD, StorageLevel, StorageLevel, ClassTag<VD>, ClassTag<ED>) - Static method in class org.apache.spark.graphx.impl.GraphImpl
Create a graph from edges, setting referenced vertices to defaultVertexAttr.
apply(RDD<Tuple2<Object, VD>>, RDD<Edge<ED>>, VD, StorageLevel, StorageLevel, ClassTag<VD>, ClassTag<ED>) - Static method in class org.apache.spark.graphx.impl.GraphImpl
Create a graph from vertices and edges, setting missing vertices to defaultVertexAttr.
apply(VertexRDD<VD>, EdgeRDD<ED>, ClassTag<VD>, ClassTag<ED>) - Static method in class org.apache.spark.graphx.impl.GraphImpl
Create a graph from a VertexRDD and an EdgeRDD with arbitrary replicated vertices.
apply(Graph<VD, ED>, A, int, EdgeDirection, Function3<Object, VD, A, VD>, Function1<EdgeTriplet<VD, ED>, Iterator<Tuple2<Object, A>>>, Function2<A, A, A>, ClassTag<VD>, ClassTag<ED>, ClassTag<A>) - Static method in class org.apache.spark.graphx.Pregel
Execute a Pregel-like iterative vertex-parallel abstraction.
apply(RDD<Tuple2<Object, VD>>, ClassTag<VD>) - Static method in class org.apache.spark.graphx.VertexRDD
Constructs a standalone VertexRDD (one that is not set up for efficient joins with an EdgeRDD) from an RDD of vertex-attribute pairs.
apply(RDD<Tuple2<Object, VD>>, EdgeRDD<?>, VD, ClassTag<VD>) - Static method in class org.apache.spark.graphx.VertexRDD
Constructs a VertexRDD from an RDD of vertex-attribute pairs.
apply(RDD<Tuple2<Object, VD>>, EdgeRDD<?>, VD, Function2<VD, VD, VD>, ClassTag<VD>) - Static method in class org.apache.spark.graphx.VertexRDD
Constructs a VertexRDD from an RDD of vertex-attribute pairs.
apply(DenseMatrix<Object>, DenseMatrix<Object>, Function1<Object, Object>) - Static method in class org.apache.spark.ml.ann.ApplyInPlace
 
apply(DenseMatrix<Object>, DenseMatrix<Object>, DenseMatrix<Object>, Function2<Object, Object, Object>) - Static method in class org.apache.spark.ml.ann.ApplyInPlace
 
apply(String) - Method in class org.apache.spark.ml.attribute.AttributeGroup
Gets an attribute by its name.
apply(int) - Method in class org.apache.spark.ml.attribute.AttributeGroup
Gets an attribute by its index.
apply(T1, T2) - Static method in class org.apache.spark.ml.clustering.ClusterData
 
apply(T1, T2) - Static method in class org.apache.spark.ml.feature.LabeledPoint
 
apply(int, int) - Method in class org.apache.spark.ml.linalg.DenseMatrix
 
apply(int) - Method in class org.apache.spark.ml.linalg.DenseVector
 
apply(int, int) - Method in interface org.apache.spark.ml.linalg.Matrix
Gets the (i, j)-th element.
apply(int, int) - Method in class org.apache.spark.ml.linalg.SparseMatrix
 
apply(int) - Method in class org.apache.spark.ml.linalg.SparseVector
 
apply(int) - Method in interface org.apache.spark.ml.linalg.Vector
Gets the value of the ith element.
apply(Param<T>) - Method in class org.apache.spark.ml.param.ParamMap
Gets the value of the input param or its default value if it does not exist.
apply(GeneralizedLinearRegressionBase) - Method in class org.apache.spark.ml.regression.GeneralizedLinearRegression.FamilyAndLink$
Constructs the FamilyAndLink object from a parameter map
apply(T1) - Static method in class org.apache.spark.ml.SaveInstanceEnd
 
apply(T1) - Static method in class org.apache.spark.ml.SaveInstanceStart
 
apply() - Static method in class org.apache.spark.ml.TransformEnd
 
apply() - Static method in class org.apache.spark.ml.TransformStart
 
apply(Split) - Method in class org.apache.spark.ml.tree.DecisionTreeModelReadWrite.SplitData$
 
apply(T1, T2, T3) - Static method in class org.apache.spark.mllib.classification.impl.GLMClassificationModel.SaveLoadV1_0$.Data
 
apply(T1, T2, T3) - Static method in class org.apache.spark.mllib.classification.NaiveBayesModel.SaveLoadV1_0$.Data
 
apply(T1, T2, T3, T4) - Static method in class org.apache.spark.mllib.classification.NaiveBayesModel.SaveLoadV2_0$.Data
 
apply(Row) - Method in class org.apache.spark.mllib.clustering.KMeansModel.Cluster$
 
apply(BinaryConfusionMatrix) - Method in interface org.apache.spark.mllib.evaluation.binary.BinaryClassificationMetricComputer
 
apply(BinaryConfusionMatrix) - Static method in class org.apache.spark.mllib.evaluation.binary.FalsePositiveRate
 
apply(BinaryConfusionMatrix) - Static method in class org.apache.spark.mllib.evaluation.binary.Precision
 
apply(BinaryConfusionMatrix) - Static method in class org.apache.spark.mllib.evaluation.binary.Recall
 
apply(T1) - Static method in class org.apache.spark.mllib.feature.ChiSqSelectorModel.SaveLoadV1_0$.Data
 
apply(T1, T2, T3, T4, T5) - Static method in class org.apache.spark.mllib.feature.VocabWord
 
apply(int, int) - Method in class org.apache.spark.mllib.linalg.DenseMatrix
 
apply(int) - Method in class org.apache.spark.mllib.linalg.DenseVector
 
apply(T1, T2) - Static method in class org.apache.spark.mllib.linalg.distributed.IndexedRow
 
apply(T1, T2, T3) - Static method in class org.apache.spark.mllib.linalg.distributed.MatrixEntry
 
apply(int, int) - Method in interface org.apache.spark.mllib.linalg.Matrix
Gets the (i, j)-th element.
apply(int, int) - Method in class org.apache.spark.mllib.linalg.SparseMatrix
 
apply(int) - Method in class org.apache.spark.mllib.linalg.SparseVector
 
apply(int) - Method in interface org.apache.spark.mllib.linalg.Vector
Gets the value of the ith element.
apply(T1, T2, T3) - Static method in class org.apache.spark.mllib.recommendation.Rating
 
apply(T1, T2) - Static method in class org.apache.spark.mllib.regression.impl.GLMRegressionModel.SaveLoadV1_0$.Data
 
apply(T1, T2) - Static method in class org.apache.spark.mllib.stat.test.BinarySample
 
apply(int) - Static method in class org.apache.spark.mllib.tree.configuration.Algo
 
apply(int) - Static method in class org.apache.spark.mllib.tree.configuration.EnsembleCombiningStrategy
 
apply(int) - Static method in class org.apache.spark.mllib.tree.configuration.FeatureType
 
apply(int) - Static method in class org.apache.spark.mllib.tree.configuration.QuantileStrategy
 
apply(int, Node) - Method in class org.apache.spark.mllib.tree.model.DecisionTreeModel.SaveLoadV1_0$.NodeData$
 
apply(Row) - Method in class org.apache.spark.mllib.tree.model.DecisionTreeModel.SaveLoadV1_0$.NodeData$
 
apply(int, Node) - Static method in class org.apache.spark.mllib.tree.model.DecisionTreeModel.SaveLoadV1_0$.NodeData
 
apply(Row) - Static method in class org.apache.spark.mllib.tree.model.DecisionTreeModel.SaveLoadV1_0$.NodeData
 
apply(Predict) - Method in class org.apache.spark.mllib.tree.model.DecisionTreeModel.SaveLoadV1_0$.PredictData$
 
apply(Row) - Method in class org.apache.spark.mllib.tree.model.DecisionTreeModel.SaveLoadV1_0$.PredictData$
 
apply(Predict) - Static method in class org.apache.spark.mllib.tree.model.DecisionTreeModel.SaveLoadV1_0$.PredictData
 
apply(Row) - Static method in class org.apache.spark.mllib.tree.model.DecisionTreeModel.SaveLoadV1_0$.PredictData
 
apply(Split) - Method in class org.apache.spark.mllib.tree.model.DecisionTreeModel.SaveLoadV1_0$.SplitData$
 
apply(Row) - Method in class org.apache.spark.mllib.tree.model.DecisionTreeModel.SaveLoadV1_0$.SplitData$
 
apply(Split) - Static method in class org.apache.spark.mllib.tree.model.DecisionTreeModel.SaveLoadV1_0$.SplitData
 
apply(Row) - Static method in class org.apache.spark.mllib.tree.model.DecisionTreeModel.SaveLoadV1_0$.SplitData
 
apply(int, Predict, double, boolean) - Static method in class org.apache.spark.mllib.tree.model.Node
Construct a node with nodeIndex, predict, impurity and isLeaf parameters.
apply(T1, T2, T3, T4) - Static method in class org.apache.spark.mllib.tree.model.Split
 
apply(int) - Static method in class org.apache.spark.rdd.CheckpointState
 
apply(int) - Static method in class org.apache.spark.rdd.DeterministicLevel
 
apply(int) - Static method in class org.apache.spark.RequestMethod
 
apply(T1, T2) - Static method in class org.apache.spark.resource.ResourceInformationJson
 
apply(T1, T2, T3, T4, T5, T6, T7) - Static method in class org.apache.spark.scheduler.AccumulableInfo
 
apply(T1, T2, T3, T4) - Static method in class org.apache.spark.scheduler.AskPermissionToCommitOutput
 
apply(String, long, Enumeration.Value, ByteBuffer, int, Map<String, ResourceInformation>) - Method in class org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.StatusUpdate$
Alternate factory method that takes a ByteBuffer directly for the data field
apply(T1, T2) - Static method in class org.apache.spark.scheduler.ExcludedExecutor
 
apply(T1, T2, T3) - Static method in class org.apache.spark.scheduler.local.KillTask
 
apply() - Static method in class org.apache.spark.scheduler.local.ReviveOffers
 
apply(T1, T2, T3) - Static method in class org.apache.spark.scheduler.local.StatusUpdate
 
apply() - Static method in class org.apache.spark.scheduler.local.StopExecutor
 
apply(long, TaskMetrics) - Static method in class org.apache.spark.scheduler.RuntimePercentage
 
apply(int) - Static method in class org.apache.spark.scheduler.SchedulingMode
 
apply(T1) - Static method in class org.apache.spark.scheduler.SparkListenerApplicationEnd
 
apply(T1, T2, T3, T4, T5, T6, T7) - Static method in class org.apache.spark.scheduler.SparkListenerApplicationStart
 
apply(T1, T2, T3, T4, T5) - Static method in class org.apache.spark.scheduler.SparkListenerBlockManagerAdded
 
apply(T1, T2) - Static method in class org.apache.spark.scheduler.SparkListenerBlockManagerRemoved
 
apply(T1) - Static method in class org.apache.spark.scheduler.SparkListenerBlockUpdated
 
apply(T1) - Static method in class org.apache.spark.scheduler.SparkListenerEnvironmentUpdate
 
apply(T1, T2, T3) - Static method in class org.apache.spark.scheduler.SparkListenerExecutorAdded
 
apply(T1, T2, T3) - Static method in class org.apache.spark.scheduler.SparkListenerExecutorBlacklisted
Deprecated.
 
apply(T1, T2, T3, T4, T5) - Static method in class org.apache.spark.scheduler.SparkListenerExecutorBlacklistedForStage
Deprecated.
 
apply(T1, T2, T3) - Static method in class org.apache.spark.scheduler.SparkListenerExecutorExcluded
 
apply(T1, T2, T3, T4, T5) - Static method in class org.apache.spark.scheduler.SparkListenerExecutorExcludedForStage
 
apply(T1, T2, T3) - Static method in class org.apache.spark.scheduler.SparkListenerExecutorMetricsUpdate
 
apply(T1, T2, T3) - Static method in class org.apache.spark.scheduler.SparkListenerExecutorRemoved
 
apply(T1, T2) - Static method in class org.apache.spark.scheduler.SparkListenerExecutorUnblacklisted
Deprecated.
 
apply(T1, T2) - Static method in class org.apache.spark.scheduler.SparkListenerExecutorUnexcluded
 
apply(T1, T2, T3) - Static method in class org.apache.spark.scheduler.SparkListenerJobEnd
 
apply(T1, T2, T3, T4) - Static method in class org.apache.spark.scheduler.SparkListenerJobStart
 
apply(T1) - Static method in class org.apache.spark.scheduler.SparkListenerLogStart
 
apply(T1, T2, T3) - Static method in class org.apache.spark.scheduler.SparkListenerMiscellaneousProcessAdded
 
apply(T1, T2, T3) - Static method in class org.apache.spark.scheduler.SparkListenerNodeBlacklisted
Deprecated.
 
apply(T1, T2, T3, T4, T5) - Static method in class org.apache.spark.scheduler.SparkListenerNodeBlacklistedForStage
Deprecated.
 
apply(T1, T2, T3) - Static method in class org.apache.spark.scheduler.SparkListenerNodeExcluded
 
apply(T1, T2, T3, T4, T5) - Static method in class org.apache.spark.scheduler.SparkListenerNodeExcludedForStage
 
apply(T1, T2) - Static method in class org.apache.spark.scheduler.SparkListenerNodeUnblacklisted
Deprecated.
 
apply(T1, T2) - Static method in class org.apache.spark.scheduler.SparkListenerNodeUnexcluded
 
apply(T1) - Static method in class org.apache.spark.scheduler.SparkListenerResourceProfileAdded
 
apply(T1, T2) - Static method in class org.apache.spark.scheduler.SparkListenerSpeculativeTaskSubmitted
 
apply(T1) - Static method in class org.apache.spark.scheduler.SparkListenerStageCompleted
 
apply(T1, T2, T3, T4) - Static method in class org.apache.spark.scheduler.SparkListenerStageExecutorMetrics
 
apply(T1, T2) - Static method in class org.apache.spark.scheduler.SparkListenerStageSubmitted
 
apply(T1, T2, T3, T4, T5, T6, T7) - Static method in class org.apache.spark.scheduler.SparkListenerTaskEnd
 
apply(T1) - Static method in class org.apache.spark.scheduler.SparkListenerTaskGettingResult
 
apply(T1, T2, T3) - Static method in class org.apache.spark.scheduler.SparkListenerTaskStart
 
apply(T1) - Static method in class org.apache.spark.scheduler.SparkListenerUnpersistRDD
 
apply(T1, T2) - Static method in class org.apache.spark.scheduler.SparkListenerUnschedulableTaskSetAdded
 
apply(T1, T2) - Static method in class org.apache.spark.scheduler.SparkListenerUnschedulableTaskSetRemoved
 
apply(int) - Static method in class org.apache.spark.scheduler.TaskLocality
 
apply(Object) - Method in class org.apache.spark.sql.Column
Extracts a value or values from a complex type.
apply(String, Expression...) - Static method in class org.apache.spark.sql.connector.expressions.Expressions
Create a logical transform for applying a named transform.
apply(String, Seq<Expression>) - Static method in class org.apache.spark.sql.connector.expressions.LogicalExpressions
 
apply(String) - Method in class org.apache.spark.sql.Dataset
Selects column based on the column name and returns it as a Column.
apply(Column...) - Method in class org.apache.spark.sql.expressions.UserDefinedAggregateFunction
Deprecated.
Creates a Column for this UDAF using given Columns as input arguments.
apply(Seq<Column>) - Method in class org.apache.spark.sql.expressions.UserDefinedAggregateFunction
Deprecated.
Creates a Column for this UDAF using given Columns as input arguments.
apply(Column...) - Method in class org.apache.spark.sql.expressions.UserDefinedFunction
Returns an expression that invokes the UDF, using the given arguments.
apply(Seq<Column>) - Method in class org.apache.spark.sql.expressions.UserDefinedFunction
Returns an expression that invokes the UDF, using the given arguments.
apply(T1, T2) - Static method in class org.apache.spark.sql.jdbc.JdbcType
 
apply() - Static method in class org.apache.spark.sql.Observation
Observation constructor for creating an anonymous observation.
apply(String) - Static method in class org.apache.spark.sql.Observation
Observation constructor for creating a named observation.
apply(Dataset<Row>, Seq<Expression>, RelationalGroupedDataset.GroupType) - Static method in class org.apache.spark.sql.RelationalGroupedDataset
 
apply(int) - Method in interface org.apache.spark.sql.Row
Returns the value at position i.
apply(T1, T2) - Static method in class org.apache.spark.sql.sources.And
 
apply(T1, T2) - Static method in class org.apache.spark.sql.sources.EqualNullSafe
 
apply(T1, T2) - Static method in class org.apache.spark.sql.sources.EqualTo
 
apply(T1, T2) - Static method in class org.apache.spark.sql.sources.GreaterThan
 
apply(T1, T2) - Static method in class org.apache.spark.sql.sources.GreaterThanOrEqual
 
apply(T1, T2) - Static method in class org.apache.spark.sql.sources.In
 
apply(T1) - Static method in class org.apache.spark.sql.sources.IsNotNull
 
apply(T1) - Static method in class org.apache.spark.sql.sources.IsNull
 
apply(T1, T2) - Static method in class org.apache.spark.sql.sources.LessThan
 
apply(T1, T2) - Static method in class org.apache.spark.sql.sources.LessThanOrEqual
 
apply(T1) - Static method in class org.apache.spark.sql.sources.Not
 
apply(T1, T2) - Static method in class org.apache.spark.sql.sources.Or
 
apply(T1, T2) - Static method in class org.apache.spark.sql.sources.StringContains
 
apply(T1, T2) - Static method in class org.apache.spark.sql.sources.StringEndsWith
 
apply(T1, T2) - Static method in class org.apache.spark.sql.sources.StringStartsWith
 
apply(String, Option<Object>, Map<String, String>) - Static method in class org.apache.spark.sql.streaming.SinkProgress
 
apply(DataType) - Static method in class org.apache.spark.sql.types.ArrayType
Construct a ArrayType object with the given element type.
apply(T1) - Static method in class org.apache.spark.sql.types.CharType
 
apply() - Static method in class org.apache.spark.sql.types.DayTimeIntervalType
 
apply(byte) - Static method in class org.apache.spark.sql.types.DayTimeIntervalType
 
apply(double) - Static method in class org.apache.spark.sql.types.Decimal
 
apply(long) - Static method in class org.apache.spark.sql.types.Decimal
 
apply(int) - Static method in class org.apache.spark.sql.types.Decimal
 
apply(BigDecimal) - Static method in class org.apache.spark.sql.types.Decimal
 
apply(BigDecimal) - Static method in class org.apache.spark.sql.types.Decimal
 
apply(BigInteger) - Static method in class org.apache.spark.sql.types.Decimal
 
apply(BigInt) - Static method in class org.apache.spark.sql.types.Decimal
 
apply(BigDecimal, int, int) - Static method in class org.apache.spark.sql.types.Decimal
 
apply(BigDecimal, int, int) - Static method in class org.apache.spark.sql.types.Decimal
 
apply(long, int, int) - Static method in class org.apache.spark.sql.types.Decimal
 
apply(String) - Static method in class org.apache.spark.sql.types.Decimal
 
apply(DataType, DataType) - Static method in class org.apache.spark.sql.types.MapType
Construct a MapType object with the given key type and value type.
apply(T1, T2, T3, T4) - Static method in class org.apache.spark.sql.types.StructField
 
apply(String) - Method in class org.apache.spark.sql.types.StructType
Extracts the StructField with the given name.
apply(Set<String>) - Method in class org.apache.spark.sql.types.StructType
Returns a StructType containing StructFields of the given names, preserving the original order of fields.
apply(int) - Method in class org.apache.spark.sql.types.StructType
 
apply(T1) - Static method in class org.apache.spark.sql.types.VarcharType
 
apply() - Static method in class org.apache.spark.sql.types.YearMonthIntervalType
 
apply(byte) - Static method in class org.apache.spark.sql.types.YearMonthIntervalType
 
apply(T1, T2, T3, T4, T5, T6, T7, T8) - Static method in class org.apache.spark.status.api.v1.ApplicationAttemptInfo
 
apply(T1, T2, T3, T4, T5, T6, T7) - Static method in class org.apache.spark.status.api.v1.ApplicationInfo
 
apply(T1, T2) - Static method in class org.apache.spark.status.api.v1.sql.Metric
 
apply(T1, T2, T3, T4) - Static method in class org.apache.spark.status.api.v1.sql.Node
 
apply(T1) - Static method in class org.apache.spark.status.api.v1.StackTrace
 
apply(T1, T2, T3, T4, T5, T6, T7) - Static method in class org.apache.spark.status.api.v1.ThreadStackTrace
 
apply(int) - Method in class org.apache.spark.status.RDDPartitionSeq
 
apply(String) - Static method in class org.apache.spark.storage.BlockId
 
apply(String, String, int, Option<String>) - Static method in class org.apache.spark.storage.BlockManagerId
Returns a BlockManagerId for the given configuration.
apply(ObjectInput) - Static method in class org.apache.spark.storage.BlockManagerId
 
apply(T1, T2) - Static method in class org.apache.spark.storage.BroadcastBlockId
 
apply(T1, T2) - Static method in class org.apache.spark.storage.RDDBlockId
 
apply(T1, T2, T3, T4) - Static method in class org.apache.spark.storage.ShuffleBlockBatchId
 
apply(T1, T2, T3, T4) - Static method in class org.apache.spark.storage.ShuffleBlockChunkId
 
apply(T1, T2, T3) - Static method in class org.apache.spark.storage.ShuffleBlockId
 
apply(T1, T2, T3) - Static method in class org.apache.spark.storage.ShuffleChecksumBlockId
 
apply(T1, T2, T3) - Static method in class org.apache.spark.storage.ShuffleDataBlockId
 
apply(T1, T2, T3) - Static method in class org.apache.spark.storage.ShuffleIndexBlockId
 
apply(T1, T2, T3) - Static method in class org.apache.spark.storage.ShuffleMergedBlockId
 
apply(T1, T2, T3, T4) - Static method in class org.apache.spark.storage.ShuffleMergedDataBlockId
 
apply(T1, T2, T3, T4) - Static method in class org.apache.spark.storage.ShuffleMergedIndexBlockId
 
apply(T1, T2, T3, T4) - Static method in class org.apache.spark.storage.ShuffleMergedMetaBlockId
 
apply(T1, T2, T3, T4) - Static method in class org.apache.spark.storage.ShufflePushBlockId
 
apply(boolean, boolean, boolean, boolean, int) - Static method in class org.apache.spark.storage.StorageLevel
:: DeveloperApi :: Create a new StorageLevel object.
apply(boolean, boolean, boolean, int) - Static method in class org.apache.spark.storage.StorageLevel
:: DeveloperApi :: Create a new StorageLevel object without setting useOffHeap.
apply(int, int) - Static method in class org.apache.spark.storage.StorageLevel
:: DeveloperApi :: Create a new StorageLevel object from its integer representation.
apply(ObjectInput) - Static method in class org.apache.spark.storage.StorageLevel
:: DeveloperApi :: Read StorageLevel object from ObjectInput stream.
apply(T1, T2) - Static method in class org.apache.spark.storage.StreamBlockId
 
apply(T1) - Static method in class org.apache.spark.storage.TaskResultBlockId
 
apply(T1) - Static method in class org.apache.spark.streaming.Duration
 
apply(long) - Static method in class org.apache.spark.streaming.Milliseconds
 
apply(long) - Static method in class org.apache.spark.streaming.Minutes
 
apply(T1, T2, T3, T4, T5, T6) - Static method in class org.apache.spark.streaming.scheduler.BatchInfo
 
apply(T1, T2, T3, T4, T5, T6, T7) - Static method in class org.apache.spark.streaming.scheduler.OutputOperationInfo
 
apply(T1, T2, T3, T4, T5, T6, T7, T8) - Static method in class org.apache.spark.streaming.scheduler.ReceiverInfo
 
apply(int) - Static method in class org.apache.spark.streaming.scheduler.ReceiverState
 
apply(T1) - Static method in class org.apache.spark.streaming.scheduler.StreamingListenerBatchCompleted
 
apply(T1) - Static method in class org.apache.spark.streaming.scheduler.StreamingListenerBatchStarted
 
apply(T1) - Static method in class org.apache.spark.streaming.scheduler.StreamingListenerBatchSubmitted
 
apply(T1) - Static method in class org.apache.spark.streaming.scheduler.StreamingListenerOutputOperationCompleted
 
apply(T1) - Static method in class org.apache.spark.streaming.scheduler.StreamingListenerOutputOperationStarted
 
apply(T1) - Static method in class org.apache.spark.streaming.scheduler.StreamingListenerReceiverError
 
apply(T1) - Static method in class org.apache.spark.streaming.scheduler.StreamingListenerReceiverStarted
 
apply(T1) - Static method in class org.apache.spark.streaming.scheduler.StreamingListenerReceiverStopped
 
apply(T1) - Static method in class org.apache.spark.streaming.scheduler.StreamingListenerStreamingStarted
 
apply(long) - Static method in class org.apache.spark.streaming.Seconds
 
apply(T1, T2, T3) - Static method in class org.apache.spark.TaskCommitDenied
 
apply(T1, T2, T3, T4) - Static method in class org.apache.spark.TaskKilled
 
apply(int) - Static method in class org.apache.spark.TaskState
 
apply(TraversableOnce<Object>) - Static method in class org.apache.spark.util.StatCounter
Build a StatCounter from a list of values.
apply(Seq<Object>) - Static method in class org.apache.spark.util.StatCounter
Build a StatCounter from a list of values passed as variable-length arguments.
ApplyInPlace - Class in org.apache.spark.ml.ann
Implements in-place application of functions in the arrays
ApplyInPlace() - Constructor for class org.apache.spark.ml.ann.ApplyInPlace
 
applyNamespaceChanges(Map<String, String>, Seq<NamespaceChange>) - Static method in class org.apache.spark.sql.connector.catalog.CatalogV2Util
Apply properties changes to a map and return the result.
applyNamespaceChanges(Map<String, String>, Seq<NamespaceChange>) - Static method in class org.apache.spark.sql.connector.catalog.CatalogV2Util
Apply properties changes to a Java map and return the result.
applyPropertiesChanges(Map<String, String>, Seq<TableChange>) - Static method in class org.apache.spark.sql.connector.catalog.CatalogV2Util
Apply properties changes to a map and return the result.
applyPropertiesChanges(Map<String, String>, Seq<TableChange>) - Static method in class org.apache.spark.sql.connector.catalog.CatalogV2Util
Apply properties changes to a Java map and return the result.
applySchema(RDD<Row>, StructType) - Method in class org.apache.spark.sql.SQLContext
Deprecated.
Use createDataFrame instead. Since 1.3.0.
applySchema(JavaRDD<Row>, StructType) - Method in class org.apache.spark.sql.SQLContext
Deprecated.
Use createDataFrame instead. Since 1.3.0.
applySchema(RDD<?>, Class<?>) - Method in class org.apache.spark.sql.SQLContext
Deprecated.
Use createDataFrame instead. Since 1.3.0.
applySchema(JavaRDD<?>, Class<?>) - Method in class org.apache.spark.sql.SQLContext
Deprecated.
Use createDataFrame instead. Since 1.3.0.
applySchemaChanges(StructType, Seq<TableChange>, Option<String>, String) - Static method in class org.apache.spark.sql.connector.catalog.CatalogV2Util
Apply schema changes to a schema and return the result.
appName() - Method in class org.apache.spark.api.java.JavaSparkContext
 
appName() - Method in class org.apache.spark.scheduler.SparkListenerApplicationStart
 
appName() - Method in class org.apache.spark.SparkContext
 
appName(String) - Method in class org.apache.spark.sql.SparkSession.Builder
Sets a name for the application, which will be shown in the Spark web UI.
approx_count_distinct(Column) - Static method in class org.apache.spark.sql.functions
Aggregate function: returns the approximate number of distinct items in a group.
approx_count_distinct(String) - Static method in class org.apache.spark.sql.functions
Aggregate function: returns the approximate number of distinct items in a group.
approx_count_distinct(Column, double) - Static method in class org.apache.spark.sql.functions
Aggregate function: returns the approximate number of distinct items in a group.
approx_count_distinct(String, double) - Static method in class org.apache.spark.sql.functions
Aggregate function: returns the approximate number of distinct items in a group.
approxCountDistinct(Column) - Static method in class org.apache.spark.sql.functions
Deprecated.
Use approx_count_distinct. Since 2.1.0.
approxCountDistinct(String) - Static method in class org.apache.spark.sql.functions
Deprecated.
Use approx_count_distinct. Since 2.1.0.
approxCountDistinct(Column, double) - Static method in class org.apache.spark.sql.functions
Deprecated.
Use approx_count_distinct. Since 2.1.0.
approxCountDistinct(String, double) - Static method in class org.apache.spark.sql.functions
Deprecated.
Use approx_count_distinct. Since 2.1.0.
ApproxHist() - Static method in class org.apache.spark.mllib.tree.configuration.QuantileStrategy
 
ApproximateEvaluator<U,R> - Interface in org.apache.spark.partial
An object that computes a function incrementally by merging in results of type U from multiple tasks.
approxQuantile(String, double[], double) - Method in class org.apache.spark.sql.DataFrameStatFunctions
Calculates the approximate quantiles of a numerical column of a DataFrame.
approxQuantile(String[], double[], double) - Method in class org.apache.spark.sql.DataFrameStatFunctions
Calculates the approximate quantiles of numerical columns of a DataFrame.
appSparkVersion() - Method in class org.apache.spark.status.api.v1.ApplicationAttemptInfo
 
AppStatusUtils - Class in org.apache.spark.status
 
AppStatusUtils() - Constructor for class org.apache.spark.status.AppStatusUtils
 
archives() - Method in class org.apache.spark.SparkContext
 
AreaUnderCurve - Class in org.apache.spark.mllib.evaluation
Computes the area under the curve (AUC) using the trapezoidal rule.
AreaUnderCurve() - Constructor for class org.apache.spark.mllib.evaluation.AreaUnderCurve
 
areaUnderPR() - Method in class org.apache.spark.mllib.evaluation.BinaryClassificationMetrics
Computes the area under the precision-recall curve.
areaUnderROC() - Method in interface org.apache.spark.ml.classification.BinaryClassificationSummary
Computes the area under the receiver operating characteristic (ROC) curve.
areaUnderROC() - Method in class org.apache.spark.ml.classification.BinaryLogisticRegressionSummaryImpl
 
areaUnderROC() - Method in class org.apache.spark.ml.classification.BinaryRandomForestClassificationSummaryImpl
 
areaUnderROC() - Method in class org.apache.spark.ml.classification.FMClassificationSummaryImpl
 
areaUnderROC() - Method in class org.apache.spark.ml.classification.LinearSVCSummaryImpl
 
areaUnderROC() - Method in class org.apache.spark.mllib.evaluation.BinaryClassificationMetrics
Computes the area under the receiver operating characteristic (ROC) curve.
argmax() - Method in class org.apache.spark.ml.linalg.DenseVector
 
argmax() - Method in class org.apache.spark.ml.linalg.SparseVector
 
argmax() - Method in interface org.apache.spark.ml.linalg.Vector
Find the index of a maximal element.
argmax() - Method in class org.apache.spark.mllib.linalg.DenseVector
 
argmax() - Method in class org.apache.spark.mllib.linalg.SparseVector
 
argmax() - Method in interface org.apache.spark.mllib.linalg.Vector
Find the index of a maximal element.
arguments() - Method in interface org.apache.spark.sql.connector.expressions.Transform
Returns the arguments passed to the transform function.
arithmeticOverflowError(ArithmeticException) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
arithmeticOverflowError(String, String, SQLQueryContext) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
ARPACK - Class in org.apache.spark.mllib.linalg
ARPACK routines for MLlib's vectors and matrices.
ARPACK() - Constructor for class org.apache.spark.mllib.linalg.ARPACK
 
array(DataType) - Method in class org.apache.spark.sql.ColumnName
Creates a new StructField of type array.
array(Column...) - Static method in class org.apache.spark.sql.functions
Creates a new array column.
array(String, String...) - Static method in class org.apache.spark.sql.functions
Creates a new array column.
array(Seq<Column>) - Static method in class org.apache.spark.sql.functions
Creates a new array column.
array(String, Seq<String>) - Static method in class org.apache.spark.sql.functions
Creates a new array column.
array() - Method in class org.apache.spark.sql.vectorized.ColumnarArray
 
array_append(Column, Object) - Static method in class org.apache.spark.sql.functions
Returns an ARRAY containing all elements from the source ARRAY as well as the new element.
array_compact(Column) - Static method in class org.apache.spark.sql.functions
Remove all null elements from the given array.
array_contains(Column, Object) - Static method in class org.apache.spark.sql.functions
Returns null if the array is null, true if the array contains value, and false otherwise.
array_distinct(Column) - Static method in class org.apache.spark.sql.functions
Removes duplicate values from the array.
array_except(Column, Column) - Static method in class org.apache.spark.sql.functions
Returns an array of the elements in the first array but not in the second array, without duplicates.
array_insert(Column, Column, Column) - Static method in class org.apache.spark.sql.functions
Adds an item into a given array at a specified position
array_intersect(Column, Column) - Static method in class org.apache.spark.sql.functions
Returns an array of the elements in the intersection of the given two arrays, without duplicates.
array_join(Column, String, String) - Static method in class org.apache.spark.sql.functions
Concatenates the elements of column using the delimiter.
array_join(Column, String) - Static method in class org.apache.spark.sql.functions
Concatenates the elements of column using the delimiter.
array_max(Column) - Static method in class org.apache.spark.sql.functions
Returns the maximum value in the array.
array_min(Column) - Static method in class org.apache.spark.sql.functions
Returns the minimum value in the array.
array_position(Column, Object) - Static method in class org.apache.spark.sql.functions
Locates the position of the first occurrence of the value in the given array as long.
array_remove(Column, Object) - Static method in class org.apache.spark.sql.functions
Remove all elements that equal to element from the given array.
array_repeat(Column, Column) - Static method in class org.apache.spark.sql.functions
Creates an array containing the left argument repeated the number of times given by the right argument.
array_repeat(Column, int) - Static method in class org.apache.spark.sql.functions
Creates an array containing the left argument repeated the number of times given by the right argument.
array_sort(Column) - Static method in class org.apache.spark.sql.functions
Sorts the input array in ascending order.
array_sort(Column, Function2<Column, Column, Column>) - Static method in class org.apache.spark.sql.functions
Sorts the input array based on the given comparator function.
array_to_vector(Column) - Static method in class org.apache.spark.ml.functions
Converts a column of array of numeric type into a column of dense vectors in MLlib.
array_union(Column, Column) - Static method in class org.apache.spark.sql.functions
Returns an array of the elements in the union of the given two arrays, without duplicates.
arrayComponentTypeUnsupportedError(Class<?>) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
arrayLengthGt(double) - Static method in class org.apache.spark.ml.param.ParamValidators
Check that the array length is greater than lowerBound.
arrays_overlap(Column, Column) - Static method in class org.apache.spark.sql.functions
Returns true if a1 and a2 have at least one non-null element in common.
arrays_zip(Column...) - Static method in class org.apache.spark.sql.functions
Returns a merged array of structs in which the N-th struct contains all N-th values of input arrays.
arrays_zip(Seq<Column>) - Static method in class org.apache.spark.sql.functions
Returns a merged array of structs in which the N-th struct contains all N-th values of input arrays.
ArrayType - Class in org.apache.spark.sql.types
 
ArrayType(DataType, boolean) - Constructor for class org.apache.spark.sql.types.ArrayType
 
arrayValues() - Method in class org.apache.spark.storage.memory.DeserializedValuesHolder
 
ArrowColumnVector - Class in org.apache.spark.sql.vectorized
A column vector backed by Apache Arrow.
ArrowColumnVector(ValueVector) - Constructor for class org.apache.spark.sql.vectorized.ArrowColumnVector
 
ArrowUtils - Class in org.apache.spark.sql.util
 
ArrowUtils() - Constructor for class org.apache.spark.sql.util.ArrowUtils
 
as(Encoder<U>) - Method in class org.apache.spark.sql.Column
Provides a type hint about the expected return value of this column.
as(String) - Method in class org.apache.spark.sql.Column
Gives the column an alias.
as(Seq<String>) - Method in class org.apache.spark.sql.Column
(Scala-specific) Assigns the given aliases to the results of a table generating function.
as(String[]) - Method in class org.apache.spark.sql.Column
Assigns the given aliases to the results of a table generating function.
as(Symbol) - Method in class org.apache.spark.sql.Column
Gives the column an alias.
as(String, Metadata) - Method in class org.apache.spark.sql.Column
Gives the column an alias with metadata.
as(Encoder<U>) - Method in class org.apache.spark.sql.Dataset
Returns a new Dataset where each record has been mapped on to the specified type.
as(String) - Method in class org.apache.spark.sql.Dataset
Returns a new Dataset with an alias set.
as(Symbol) - Method in class org.apache.spark.sql.Dataset
(Scala-specific) Returns a new Dataset with an alias set.
as(Encoder<K>, Encoder<T>) - Method in class org.apache.spark.sql.RelationalGroupedDataset
Returns a KeyValueGroupedDataset where the data is grouped by the grouping expressions of current RelationalGroupedDataset.
asBinary() - Method in interface org.apache.spark.ml.classification.LogisticRegressionSummary
Convenient method for casting to binary logistic regression summary.
asBinary() - Method in interface org.apache.spark.ml.classification.RandomForestClassificationSummary
Convenient method for casting to BinaryRandomForestClassificationSummary.
asBreeze() - Method in interface org.apache.spark.ml.linalg.Matrix
Converts to a breeze matrix.
asBreeze() - Method in interface org.apache.spark.ml.linalg.Vector
Converts the instance to a breeze vector.
asBreeze() - Method in interface org.apache.spark.mllib.linalg.Matrix
Converts to a breeze matrix.
asBreeze() - Method in interface org.apache.spark.mllib.linalg.Vector
Converts the instance to a breeze vector.
asc() - Method in class org.apache.spark.sql.Column
Returns a sort expression based on ascending order of the column.
asc(String) - Static method in class org.apache.spark.sql.functions
Returns a sort expression based on ascending order of the column.
asc_nulls_first() - Method in class org.apache.spark.sql.Column
Returns a sort expression based on ascending order of the column, and null values return before non-null values.
asc_nulls_first(String) - Static method in class org.apache.spark.sql.functions
Returns a sort expression based on ascending order of the column, and null values return before non-null values.
asc_nulls_last() - Method in class org.apache.spark.sql.Column
Returns a sort expression based on ascending order of the column, and null values appear after non-null values.
asc_nulls_last(String) - Static method in class org.apache.spark.sql.functions
Returns a sort expression based on ascending order of the column, and null values appear after non-null values.
asCaseSensitiveMap() - Method in class org.apache.spark.sql.util.CaseInsensitiveStringMap
Returns the original case-sensitive map.
ascii(Column) - Static method in class org.apache.spark.sql.functions
Computes the numeric value of the first character of the string column, and returns the result as an int column.
asFunctionCatalog() - Method in class org.apache.spark.sql.connector.catalog.CatalogV2Implicits.CatalogHelper
 
asFunctionIdentifier() - Method in class org.apache.spark.sql.connector.catalog.CatalogV2Implicits.IdentifierHelper
 
asFunctionIdentifier() - Method in class org.apache.spark.sql.connector.catalog.CatalogV2Implicits.MultipartIdentifierHelper
 
asIdentifier() - Method in class org.apache.spark.sql.connector.catalog.CatalogV2Implicits.MultipartIdentifierHelper
 
asin(Column) - Static method in class org.apache.spark.sql.functions
 
asin(String) - Static method in class org.apache.spark.sql.functions
 
asinh(Column) - Static method in class org.apache.spark.sql.functions
 
asinh(String) - Static method in class org.apache.spark.sql.functions
 
asInteraction() - Static method in class org.apache.spark.ml.feature.Dot
 
asInteraction() - Method in interface org.apache.spark.ml.feature.InteractableTerm
Convert to ColumnInteraction to wrap all interactions.
asIterator() - Method in class org.apache.spark.serializer.DeserializationStream
Read the elements of this stream through an iterator.
asJavaPairRDD() - Method in class org.apache.spark.api.r.PairwiseRRDD
 
asJavaRDD() - Method in class org.apache.spark.api.r.RRDD
 
asJavaRDD() - Method in class org.apache.spark.api.r.StringRRDD
 
ask(Object) - Method in interface org.apache.spark.api.plugin.PluginContext
Send an RPC to the plugin's driver-side component.
asKeyValueIterator() - Method in class org.apache.spark.serializer.DeserializationStream
Read the elements of this stream through an iterator over key-value pairs.
AskPermissionToCommitOutput - Class in org.apache.spark.scheduler
 
AskPermissionToCommitOutput(int, int, int, int) - Constructor for class org.apache.spark.scheduler.AskPermissionToCommitOutput
 
askRpcTimeout(SparkConf) - Static method in class org.apache.spark.util.RpcUtils
Returns the default Spark timeout to use for RPC ask operations.
askStandaloneSchedulerToShutDownExecutorsError(Exception) - Static method in class org.apache.spark.errors.SparkCoreErrors
 
askStorageEndpoints() - Method in class org.apache.spark.storage.BlockManagerMessages.GetBlockStatus
 
askStorageEndpoints() - Method in class org.apache.spark.storage.BlockManagerMessages.GetMatchingBlockIds
 
asML() - Method in class org.apache.spark.mllib.linalg.DenseMatrix
 
asML() - Method in class org.apache.spark.mllib.linalg.DenseVector
 
asML() - Method in interface org.apache.spark.mllib.linalg.Matrix
Convert this matrix to the new mllib-local representation.
asML() - Method in class org.apache.spark.mllib.linalg.SparseMatrix
 
asML() - Method in class org.apache.spark.mllib.linalg.SparseVector
 
asML() - Method in interface org.apache.spark.mllib.linalg.Vector
Convert this vector to the new mllib-local representation.
asMultipart() - Method in class org.apache.spark.sql.connector.catalog.CatalogV2Implicits.FunctionIdentifierHelper
 
asMultipartIdentifier() - Method in class org.apache.spark.sql.connector.catalog.CatalogV2Implicits.IdentifierHelper
 
asNamespaceCatalog() - Method in class org.apache.spark.sql.connector.catalog.CatalogV2Implicits.CatalogHelper
 
asNondeterministic() - Method in class org.apache.spark.sql.expressions.UserDefinedFunction
Updates UserDefinedFunction to nondeterministic.
asNonNullable() - Method in class org.apache.spark.sql.expressions.UserDefinedFunction
Updates UserDefinedFunction to non-nullable.
asNullable() - Method in class org.apache.spark.sql.types.ObjectType
 
asRDDId() - Method in class org.apache.spark.storage.BlockId
 
asSchema() - Method in class org.apache.spark.sql.connector.catalog.CatalogV2Implicits.ColumnsHelper
 
assert_true(Column) - Static method in class org.apache.spark.sql.functions
Returns null if the condition is true, and throws an exception otherwise.
assert_true(Column, Column) - Static method in class org.apache.spark.sql.functions
Returns null if the condition is true; throws an exception with the error message otherwise.
assertExceptionMsg(Throwable, String, boolean, ClassTag<E>) - Static method in class org.apache.spark.TestUtils
Asserts that exception message contains the message.
assertNotSpilled(SparkContext, String, Function0<BoxedUnit>) - Static method in class org.apache.spark.TestUtils
Run some code involving jobs submitted to the given context and assert that the jobs did not spill.
assertSpilled(SparkContext, String, Function0<BoxedUnit>) - Static method in class org.apache.spark.TestUtils
Run some code involving jobs submitted to the given context and assert that the jobs spilled.
assignClusters(Dataset<?>) - Method in class org.apache.spark.ml.clustering.PowerIterationClustering
Run the PIC algorithm and returns a cluster assignment for each input vertex.
assignedAddrs() - Method in interface org.apache.spark.resource.ResourceAllocator
Sequence of currently assigned resource addresses.
Assignment(long, int) - Constructor for class org.apache.spark.mllib.clustering.PowerIterationClustering.Assignment
 
Assignment$() - Constructor for class org.apache.spark.mllib.clustering.PowerIterationClustering.Assignment$
 
assignments() - Method in class org.apache.spark.mllib.clustering.PowerIterationClusteringModel
 
AssociationRules - Class in org.apache.spark.ml.fpm
 
AssociationRules() - Constructor for class org.apache.spark.ml.fpm.AssociationRules
 
associationRules() - Method in class org.apache.spark.ml.fpm.FPGrowthModel
Get association rules fitted using the minConfidence.
AssociationRules - Class in org.apache.spark.mllib.fpm
Generates association rules from a RDD[FreqItemset[Item}.
AssociationRules() - Constructor for class org.apache.spark.mllib.fpm.AssociationRules
Constructs a default instance with default parameters {minConfidence = 0.8}.
AssociationRules.Rule<Item> - Class in org.apache.spark.mllib.fpm
An association rule between sets of items.
asTableCatalog() - Method in class org.apache.spark.sql.connector.catalog.CatalogV2Implicits.CatalogHelper
 
asTableIdentifier() - Method in class org.apache.spark.sql.connector.catalog.CatalogV2Implicits.IdentifierHelper
 
asTableIdentifier() - Method in class org.apache.spark.sql.connector.catalog.CatalogV2Implicits.MultipartIdentifierHelper
 
AsTableIdentifier() - Method in interface org.apache.spark.sql.connector.catalog.LookupCatalog
 
AsTableIdentifier() - Constructor for class org.apache.spark.sql.connector.catalog.LookupCatalog.AsTableIdentifier
 
AsTableIdentifier$() - Constructor for class org.apache.spark.sql.connector.catalog.LookupCatalog.AsTableIdentifier$
 
asTerms() - Static method in class org.apache.spark.ml.feature.Dot
 
asTerms() - Static method in class org.apache.spark.ml.feature.EmptyTerm
 
asTerms() - Method in interface org.apache.spark.ml.feature.Term
Default representation of a single Term as a part of summed terms.
asTransform() - Method in class org.apache.spark.sql.connector.catalog.CatalogV2Implicits.BucketSpecHelper
 
asTransforms() - Method in class org.apache.spark.sql.connector.catalog.CatalogV2Implicits.PartitionTypeHelper
 
AsyncEventQueue - Class in org.apache.spark.scheduler
An asynchronous queue for events.
AsyncEventQueue(String, SparkConf, LiveListenerBusMetrics, LiveListenerBus) - Constructor for class org.apache.spark.scheduler.AsyncEventQueue
 
AsyncRDDActions<T> - Class in org.apache.spark.rdd
A set of asynchronous RDD actions available through an implicit conversion.
AsyncRDDActions(RDD<T>, ClassTag<T>) - Constructor for class org.apache.spark.rdd.AsyncRDDActions
 
atan(Column) - Static method in class org.apache.spark.sql.functions
 
atan(String) - Static method in class org.apache.spark.sql.functions
 
atan2(Column, Column) - Static method in class org.apache.spark.sql.functions
 
atan2(Column, String) - Static method in class org.apache.spark.sql.functions
 
atan2(String, Column) - Static method in class org.apache.spark.sql.functions
 
atan2(String, String) - Static method in class org.apache.spark.sql.functions
 
atan2(Column, double) - Static method in class org.apache.spark.sql.functions
 
atan2(String, double) - Static method in class org.apache.spark.sql.functions
 
atan2(double, Column) - Static method in class org.apache.spark.sql.functions
 
atan2(double, String) - Static method in class org.apache.spark.sql.functions
 
atanh(Column) - Static method in class org.apache.spark.sql.functions
 
atanh(String) - Static method in class org.apache.spark.sql.functions
 
attempt() - Method in class org.apache.spark.status.api.v1.TaskData
 
ATTEMPT() - Static method in class org.apache.spark.status.TaskIndexNames
 
ATTEMPT_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.TaskData
 
ATTEMPT_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
 
ATTEMPT_ID_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo
 
ATTEMPT_ID_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.StageData
 
attemptId() - Method in class org.apache.spark.status.api.v1.ApplicationAttemptInfo
 
attemptId() - Method in interface org.apache.spark.status.api.v1.BaseAppResource
 
attemptId() - Method in class org.apache.spark.status.api.v1.StageData
 
attemptNumber() - Method in class org.apache.spark.BarrierTaskContext
 
attemptNumber() - Method in class org.apache.spark.scheduler.AskPermissionToCommitOutput
 
attemptNumber() - Method in class org.apache.spark.scheduler.StageInfo
 
attemptNumber() - Method in class org.apache.spark.scheduler.TaskInfo
 
attemptNumber() - Method in class org.apache.spark.TaskCommitDenied
 
attemptNumber() - Method in class org.apache.spark.TaskContext
How many times this task has been attempted.
attempts() - Method in class org.apache.spark.status.api.v1.ApplicationInfo
 
ATTEMPTS_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo
 
AtTimestamp(Date) - Constructor for class org.apache.spark.streaming.kinesis.KinesisInitialPositions.AtTimestamp
 
attr() - Method in class org.apache.spark.graphx.Edge
 
attr() - Method in class org.apache.spark.graphx.EdgeContext
The attribute associated with the edge.
attr() - Method in class org.apache.spark.graphx.impl.AggregatingEdgeContext
 
Attribute - Class in org.apache.spark.ml.attribute
Abstract class for ML attributes.
Attribute() - Constructor for class org.apache.spark.ml.attribute.Attribute
 
attribute() - Method in class org.apache.spark.sql.sources.EqualNullSafe
 
attribute() - Method in class org.apache.spark.sql.sources.EqualTo
 
attribute() - Method in class org.apache.spark.sql.sources.GreaterThan
 
attribute() - Method in class org.apache.spark.sql.sources.GreaterThanOrEqual
 
attribute() - Method in class org.apache.spark.sql.sources.In
 
attribute() - Method in class org.apache.spark.sql.sources.IsNotNull
 
attribute() - Method in class org.apache.spark.sql.sources.IsNull
 
attribute() - Method in class org.apache.spark.sql.sources.LessThan
 
attribute() - Method in class org.apache.spark.sql.sources.LessThanOrEqual
 
attribute() - Method in class org.apache.spark.sql.sources.StringContains
 
attribute() - Method in class org.apache.spark.sql.sources.StringEndsWith
 
attribute() - Method in class org.apache.spark.sql.sources.StringStartsWith
 
AttributeFactory - Interface in org.apache.spark.ml.attribute
Trait for ML attribute factories.
AttributeGroup - Class in org.apache.spark.ml.attribute
Attributes that describe a vector ML column.
AttributeGroup(String) - Constructor for class org.apache.spark.ml.attribute.AttributeGroup
Creates an attribute group without attribute info.
AttributeGroup(String, int) - Constructor for class org.apache.spark.ml.attribute.AttributeGroup
Creates an attribute group knowing only the number of attributes.
AttributeGroup(String, Attribute[]) - Constructor for class org.apache.spark.ml.attribute.AttributeGroup
Creates an attribute group with attributes.
AttributeKeys - Class in org.apache.spark.ml.attribute
Keys used to store attributes.
AttributeKeys() - Constructor for class org.apache.spark.ml.attribute.AttributeKeys
 
attributeNameSyntaxError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
attributes() - Method in class org.apache.spark.ml.attribute.AttributeGroup
Optional array of attributes.
ATTRIBUTES() - Static method in class org.apache.spark.ml.attribute.AttributeKeys
 
attributes() - Method in class org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.RegisterExecutor
 
attributes() - Method in class org.apache.spark.scheduler.cluster.ExecutorInfo
 
attributes() - Method in class org.apache.spark.status.api.v1.ExecutorSummary
 
ATTRIBUTES_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
 
attributesForTypeUnsupportedError(ScalaReflection.Schema) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
AttributeType - Class in org.apache.spark.ml.attribute
An enum-like type for attribute types: AttributeType$.Numeric, AttributeType$.Nominal, and AttributeType$.Binary.
AttributeType(String) - Constructor for class org.apache.spark.ml.attribute.AttributeType
 
attrType() - Method in class org.apache.spark.ml.attribute.Attribute
Attribute type.
attrType() - Method in class org.apache.spark.ml.attribute.BinaryAttribute
 
attrType() - Method in class org.apache.spark.ml.attribute.NominalAttribute
 
attrType() - Method in class org.apache.spark.ml.attribute.NumericAttribute
 
attrType() - Static method in class org.apache.spark.ml.attribute.UnresolvedAttribute
 
available() - Method in class org.apache.spark.io.NioBufferedFileInputStream
 
available() - Method in class org.apache.spark.io.ReadAheadInputStream
 
available() - Method in class org.apache.spark.storage.BufferReleasingInputStream
 
availableAddrs() - Method in interface org.apache.spark.resource.ResourceAllocator
Sequence of currently available resource addresses.
AvailableNow() - Static method in class org.apache.spark.sql.streaming.Trigger
A trigger that processes all available data at the start of the query in one or multiple batches, then terminates the query.
Average() - Static method in class org.apache.spark.mllib.tree.configuration.EnsembleCombiningStrategy
 
Avg - Class in org.apache.spark.sql.connector.expressions.aggregate
An aggregate function that returns the mean of all the values in a group.
Avg(Expression, boolean) - Constructor for class org.apache.spark.sql.connector.expressions.aggregate.Avg
 
avg(MapFunction<T, Double>) - Static method in class org.apache.spark.sql.expressions.javalang.typed
Deprecated.
Average aggregate function.
avg(Function1<IN, Object>) - Static method in class org.apache.spark.sql.expressions.scalalang.typed
Deprecated.
Average aggregate function.
avg(Column) - Static method in class org.apache.spark.sql.functions
Aggregate function: returns the average of the values in a group.
avg(String) - Static method in class org.apache.spark.sql.functions
Aggregate function: returns the average of the values in a group.
avg(String...) - Method in class org.apache.spark.sql.RelationalGroupedDataset
Compute the mean value for each numeric columns for each group.
avg(Seq<String>) - Method in class org.apache.spark.sql.RelationalGroupedDataset
Compute the mean value for each numeric columns for each group.
avg() - Method in class org.apache.spark.util.DoubleAccumulator
Returns the average of elements added to the accumulator.
avg() - Method in class org.apache.spark.util.LongAccumulator
Returns the average of elements added to the accumulator.
avgEventRate() - Method in class org.apache.spark.status.api.v1.streaming.ReceiverInfo
 
avgInputRate() - Method in class org.apache.spark.status.api.v1.streaming.StreamingStatistics
 
avgLen() - Method in interface org.apache.spark.sql.connector.read.colstats.ColumnStatistics
 
avgMetrics() - Method in class org.apache.spark.ml.tuning.CrossValidatorModel
 
avgProcessingTime() - Method in class org.apache.spark.status.api.v1.streaming.StreamingStatistics
 
avgSchedulingDelay() - Method in class org.apache.spark.status.api.v1.streaming.StreamingStatistics
 
avgTotalDelay() - Method in class org.apache.spark.status.api.v1.streaming.StreamingStatistics
 
AvroMatchedField$() - Constructor for class org.apache.spark.sql.avro.AvroUtils.AvroMatchedField$
 
AvroSchemaHelper(Schema, StructType, Seq<String>, Seq<String>, boolean) - Constructor for class org.apache.spark.sql.avro.AvroUtils.AvroSchemaHelper
 
AvroUtils - Class in org.apache.spark.sql.avro
 
AvroUtils() - Constructor for class org.apache.spark.sql.avro.AvroUtils
 
AvroUtils.AvroMatchedField$ - Class in org.apache.spark.sql.avro
 
AvroUtils.AvroSchemaHelper - Class in org.apache.spark.sql.avro
Helper class to perform field lookup/matching on Avro schemas.
AvroUtils.RowReader - Interface in org.apache.spark.sql.avro
 
awaitAnyTermination() - Method in class org.apache.spark.sql.streaming.StreamingQueryManager
Wait until any of the queries on the associated SQLContext has terminated since the creation of the context, or since resetTerminated() was called.
awaitAnyTermination(long) - Method in class org.apache.spark.sql.streaming.StreamingQueryManager
Wait until any of the queries on the associated SQLContext has terminated since the creation of the context, or since resetTerminated() was called.
awaitReady(Awaitable<T>, Duration) - Static method in class org.apache.spark.util.ThreadUtils
Preferred alternative to Await.ready().
awaitResult(Awaitable<T>, Duration) - Static method in class org.apache.spark.util.ThreadUtils
Preferred alternative to Await.result().
awaitResult(Future<T>, Duration) - Static method in class org.apache.spark.util.ThreadUtils
 
awaitTermination() - Method in interface org.apache.spark.sql.streaming.StreamingQuery
Waits for the termination of this query, either by query.stop() or by an exception.
awaitTermination(long) - Method in interface org.apache.spark.sql.streaming.StreamingQuery
Waits for the termination of this query, either by query.stop() or by an exception.
awaitTermination() - Method in class org.apache.spark.streaming.api.java.JavaStreamingContext
Deprecated.
Wait for the execution to stop.
awaitTermination() - Method in class org.apache.spark.streaming.StreamingContext
Deprecated.
Wait for the execution to stop.
awaitTerminationOrTimeout(long) - Method in class org.apache.spark.streaming.api.java.JavaStreamingContext
Deprecated.
Wait for the execution to stop.
awaitTerminationOrTimeout(long) - Method in class org.apache.spark.streaming.StreamingContext
Deprecated.
Wait for the execution to stop.
axpy(double, Vector, Vector) - Static method in class org.apache.spark.ml.linalg.BLAS
y += a * x
axpy(double, Vector, Vector) - Static method in class org.apache.spark.mllib.linalg.BLAS
y += a * x

B

BACKUP_STANDALONE_MASTER_PREFIX() - Static method in class org.apache.spark.util.Utils
An identifier that backup masters use in their responses.
balanceSlack() - Method in class org.apache.spark.rdd.DefaultPartitionCoalescer
 
barrier() - Method in class org.apache.spark.BarrierTaskContext
:: Experimental :: Sets a global barrier and waits until all tasks in this stage hit this barrier.
barrier() - Method in class org.apache.spark.rdd.RDD
:: Experimental :: Marks the current stage as a barrier stage, where Spark must launch all tasks together.
BARRIER() - Static method in class org.apache.spark.RequestMethod
 
BARRIER_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode
 
BarrierCoordinatorMessage - Interface in org.apache.spark
 
barrierStageWithDynamicAllocationError() - Static method in class org.apache.spark.errors.SparkCoreErrors
 
barrierStageWithRDDChainPatternError() - Static method in class org.apache.spark.errors.SparkCoreErrors
 
BarrierTaskContext - Class in org.apache.spark
:: Experimental :: A TaskContext with extra contextual info and tooling for tasks in a barrier stage.
BarrierTaskInfo - Class in org.apache.spark
:: Experimental :: Carries all task infos of a barrier task.
base64(Column) - Static method in class org.apache.spark.sql.functions
Computes the BASE64 encoding of a binary column and returns it as a string column.
BaseAppResource - Interface in org.apache.spark.status.api.v1
Base class for resource handlers that use app-specific data.
baseOn(ParamPair<?>...) - Method in class org.apache.spark.ml.tuning.ParamGridBuilder
Sets the given parameters in this grid to fixed values.
baseOn(ParamMap) - Method in class org.apache.spark.ml.tuning.ParamGridBuilder
Sets the given parameters in this grid to fixed values.
baseOn(Seq<ParamPair<?>>) - Method in class org.apache.spark.ml.tuning.ParamGridBuilder
Sets the given parameters in this grid to fixed values.
BaseReadWrite - Interface in org.apache.spark.ml.util
Trait for MLWriter and MLReader.
BaseRelation - Class in org.apache.spark.sql.sources
Represents a collection of tuples with a known schema.
BaseRelation() - Constructor for class org.apache.spark.sql.sources.BaseRelation
 
baseRelationToDataFrame(BaseRelation) - Method in class org.apache.spark.sql.SparkSession
Convert a BaseRelation created for external data sources into a DataFrame.
baseRelationToDataFrame(BaseRelation) - Method in class org.apache.spark.sql.SQLContext
 
BaseRRDD<T,U> - Class in org.apache.spark.api.r
 
BaseRRDD(RDD<T>, int, byte[], String, String, byte[], Broadcast<Object>[], ClassTag<T>, ClassTag<U>) - Constructor for class org.apache.spark.api.r.BaseRRDD
 
BaseStreamingAppResource - Interface in org.apache.spark.status.api.v1.streaming
Base class for streaming API handlers, provides easy access to the streaming listener that holds the app's information.
BasicBlockReplicationPolicy - Class in org.apache.spark.storage
 
BasicBlockReplicationPolicy() - Constructor for class org.apache.spark.storage.BasicBlockReplicationPolicy
 
basicCredentials(String, String) - Method in class org.apache.spark.streaming.kinesis.SparkAWSCredentials.Builder
Use a basic AWS keypair for long-lived authorization.
basicSparkPage(HttpServletRequest, Function0<Seq<Node>>, String, boolean) - Static method in class org.apache.spark.ui.UIUtils
Returns a page with the spark css/js and a simple format.
Batch - Interface in org.apache.spark.sql.connector.read
A physical representation of a data source scan for batch queries.
BATCH_DURATION_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
 
BATCH_ID_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
 
batchDuration() - Method in class org.apache.spark.sql.streaming.StreamingQueryProgress
 
batchDuration() - Method in class org.apache.spark.status.api.v1.streaming.BatchInfo
 
batchDuration() - Method in class org.apache.spark.status.api.v1.streaming.StreamingStatistics
 
BATCHES() - Static method in class org.apache.spark.mllib.clustering.StreamingKMeans
 
batchId() - Method in class org.apache.spark.sql.streaming.StreamingQueryProgress
 
batchId() - Method in class org.apache.spark.status.api.v1.streaming.BatchInfo
 
BatchInfo - Class in org.apache.spark.status.api.v1.streaming
 
BatchInfo - Class in org.apache.spark.streaming.scheduler
:: DeveloperApi :: Class having information on completed batches.
BatchInfo(Time, Map<Object, StreamInputInfo>, long, Option<Object>, Option<Object>, Map<Object, OutputOperationInfo>) - Constructor for class org.apache.spark.streaming.scheduler.BatchInfo
 
batchInfo() - Method in class org.apache.spark.streaming.scheduler.StreamingListenerBatchCompleted
 
batchInfo() - Method in class org.apache.spark.streaming.scheduler.StreamingListenerBatchStarted
 
batchInfo() - Method in class org.apache.spark.streaming.scheduler.StreamingListenerBatchSubmitted
 
batchInfos() - Method in class org.apache.spark.streaming.scheduler.StatsReportListener
 
batchMetadataFileNotFoundError(Path) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
BatchStatus - Enum in org.apache.spark.status.api.v1.streaming
 
batchTime() - Method in class org.apache.spark.status.api.v1.streaming.BatchInfo
 
batchTime() - Method in class org.apache.spark.streaming.scheduler.BatchInfo
 
batchTime() - Method in class org.apache.spark.streaming.scheduler.OutputOperationInfo
 
BatchWrite - Interface in org.apache.spark.sql.connector.write
An interface that defines how to write the data to data source for batch processing.
batchWriteCapabilityError(Table, String, String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
bbos() - Method in class org.apache.spark.storage.memory.SerializedValuesHolder
 
bean(Class<T>) - Static method in class org.apache.spark.sql.Encoders
Creates an encoder for Java Bean of type T.
beforeFetch(Connection, Map<String, String>) - Static method in class org.apache.spark.sql.jdbc.DB2Dialect
 
beforeFetch(Connection, Map<String, String>) - Static method in class org.apache.spark.sql.jdbc.DerbyDialect
 
beforeFetch(Connection, Map<String, String>) - Static method in class org.apache.spark.sql.jdbc.H2Dialect
 
beforeFetch(Connection, Map<String, String>) - Method in class org.apache.spark.sql.jdbc.JdbcDialect
Override connection specific properties to run before a select is made.
beforeFetch(Connection, Map<String, String>) - Static method in class org.apache.spark.sql.jdbc.MsSqlServerDialect
 
beforeFetch(Connection, Map<String, String>) - Static method in class org.apache.spark.sql.jdbc.MySQLDialect
 
beforeFetch(Connection, Map<String, String>) - Static method in class org.apache.spark.sql.jdbc.NoopDialect
 
beforeFetch(Connection, Map<String, String>) - Static method in class org.apache.spark.sql.jdbc.OracleDialect
 
beforeFetch(Connection, Map<String, String>) - Static method in class org.apache.spark.sql.jdbc.PostgresDialect
 
beforeFetch(Connection, Map<String, String>) - Static method in class org.apache.spark.sql.jdbc.TeradataDialect
 
BernoulliCellSampler<T> - Class in org.apache.spark.util.random
:: DeveloperApi :: A sampler based on Bernoulli trials for partitioning a data sequence.
BernoulliCellSampler(double, double, boolean) - Constructor for class org.apache.spark.util.random.BernoulliCellSampler
 
BernoulliSampler<T> - Class in org.apache.spark.util.random
:: DeveloperApi :: A sampler based on Bernoulli trials.
BernoulliSampler(double, ClassTag<T>) - Constructor for class org.apache.spark.util.random.BernoulliSampler
 
bestModel() - Method in class org.apache.spark.ml.tuning.CrossValidatorModel
 
bestModel() - Method in class org.apache.spark.ml.tuning.TrainValidationSplitModel
 
beta() - Method in class org.apache.spark.ml.evaluation.MulticlassClassificationEvaluator
The beta value, which controls precision vs recall weighting, used in "weightedFMeasure", "fMeasureByLabel".
beta() - Method in class org.apache.spark.mllib.random.WeibullGenerator
 
between(Object, Object) - Method in class org.apache.spark.sql.Column
True if the current column is between the lower bound and upper bound, inclusive.
bin(Column) - Static method in class org.apache.spark.sql.functions
An expression that returns the string representation of the binary value of the given long column.
bin(String) - Static method in class org.apache.spark.sql.functions
An expression that returns the string representation of the binary value of the given long column.
Binarizer - Class in org.apache.spark.ml.feature
Binarize a column of continuous features given a threshold.
Binarizer(String) - Constructor for class org.apache.spark.ml.feature.Binarizer
 
Binarizer() - Constructor for class org.apache.spark.ml.feature.Binarizer
 
Binary() - Static method in class org.apache.spark.ml.attribute.AttributeType
Binary type.
binary() - Method in class org.apache.spark.ml.feature.CountVectorizer
 
binary() - Method in class org.apache.spark.ml.feature.CountVectorizerModel
 
binary() - Method in interface org.apache.spark.ml.feature.CountVectorizerParams
Binary toggle to control the output vector values.
binary() - Method in class org.apache.spark.ml.feature.HashingTF
Binary toggle to control term frequency counts.
binary() - Method in class org.apache.spark.sql.ColumnName
Creates a new StructField of type binary.
BINARY() - Static method in class org.apache.spark.sql.Encoders
An encoder for arrays of bytes.
binaryArithmeticCauseOverflowError(short, String, short) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
BinaryAttribute - Class in org.apache.spark.ml.attribute
A binary attribute.
BinaryClassificationEvaluator - Class in org.apache.spark.ml.evaluation
Evaluator for binary classification, which expects input columns rawPrediction, label and an optional weight column.
BinaryClassificationEvaluator(String) - Constructor for class org.apache.spark.ml.evaluation.BinaryClassificationEvaluator
 
BinaryClassificationEvaluator() - Constructor for class org.apache.spark.ml.evaluation.BinaryClassificationEvaluator
 
BinaryClassificationMetricComputer - Interface in org.apache.spark.mllib.evaluation.binary
Trait for a binary classification evaluation metric computer.
BinaryClassificationMetrics - Class in org.apache.spark.mllib.evaluation
Evaluator for binary classification.
BinaryClassificationMetrics(RDD<? extends Product>, int) - Constructor for class org.apache.spark.mllib.evaluation.BinaryClassificationMetrics
 
BinaryClassificationMetrics(RDD<Tuple2<Object, Object>>) - Constructor for class org.apache.spark.mllib.evaluation.BinaryClassificationMetrics
Defaults numBins to 0.
BinaryClassificationSummary - Interface in org.apache.spark.ml.classification
Abstraction for binary classification results for a given model.
binaryColumn(String[]) - Static method in class org.apache.parquet.filter2.predicate.SparkFilterApi
 
BinaryConfusionMatrix - Interface in org.apache.spark.mllib.evaluation.binary
Trait for a binary confusion matrix.
binaryFiles(String, int) - Method in class org.apache.spark.api.java.JavaSparkContext
Read a directory of binary files from HDFS, a local file system (available on all nodes), or any Hadoop-supported file system URI as a byte array.
binaryFiles(String) - Method in class org.apache.spark.api.java.JavaSparkContext
Read a directory of binary files from HDFS, a local file system (available on all nodes), or any Hadoop-supported file system URI as a byte array.
binaryFiles(String, int) - Method in class org.apache.spark.SparkContext
Get an RDD for a Hadoop-readable dataset as PortableDataStream for each file (useful for binary data)
binaryLabelValidator() - Static method in class org.apache.spark.mllib.util.DataValidators
Function to check if labels used for classification are either zero or one.
BinaryLogisticRegressionSummary - Interface in org.apache.spark.ml.classification
Abstraction for binary logistic regression results for a given model.
BinaryLogisticRegressionSummaryImpl - Class in org.apache.spark.ml.classification
Binary logistic regression results for a given model.
BinaryLogisticRegressionSummaryImpl(Dataset<Row>, String, String, String, String, String) - Constructor for class org.apache.spark.ml.classification.BinaryLogisticRegressionSummaryImpl
 
BinaryLogisticRegressionTrainingSummary - Interface in org.apache.spark.ml.classification
Abstraction for binary logistic regression training results.
BinaryLogisticRegressionTrainingSummaryImpl - Class in org.apache.spark.ml.classification
Binary logistic regression training results.
BinaryLogisticRegressionTrainingSummaryImpl(Dataset<Row>, String, String, String, String, String, double[]) - Constructor for class org.apache.spark.ml.classification.BinaryLogisticRegressionTrainingSummaryImpl
 
BinaryRandomForestClassificationSummary - Interface in org.apache.spark.ml.classification
Abstraction for BinaryRandomForestClassification results for a given model.
BinaryRandomForestClassificationSummaryImpl - Class in org.apache.spark.ml.classification
Binary RandomForestClassification for a given model.
BinaryRandomForestClassificationSummaryImpl(Dataset<Row>, String, String, String, String) - Constructor for class org.apache.spark.ml.classification.BinaryRandomForestClassificationSummaryImpl
 
BinaryRandomForestClassificationTrainingSummary - Interface in org.apache.spark.ml.classification
Abstraction for BinaryRandomForestClassification training results.
BinaryRandomForestClassificationTrainingSummaryImpl - Class in org.apache.spark.ml.classification
Binary RandomForestClassification training results.
BinaryRandomForestClassificationTrainingSummaryImpl(Dataset<Row>, String, String, String, String, double[]) - Constructor for class org.apache.spark.ml.classification.BinaryRandomForestClassificationTrainingSummaryImpl
 
binaryRecords(String, int) - Method in class org.apache.spark.api.java.JavaSparkContext
Load data from a flat binary file, assuming the length of each record is constant.
binaryRecords(String, int, Configuration) - Method in class org.apache.spark.SparkContext
Load data from a flat binary file, assuming the length of each record is constant.
binaryRecordsStream(String, int) - Method in class org.apache.spark.streaming.api.java.JavaStreamingContext
Deprecated.
Create an input stream that monitors a Hadoop-compatible filesystem for new files and reads them as flat binary files with fixed record lengths, yielding byte arrays
binaryRecordsStream(String, int) - Method in class org.apache.spark.streaming.StreamingContext
Deprecated.
Create an input stream that monitors a Hadoop-compatible filesystem for new files and reads them as flat binary files, assuming a fixed length per record, generating one byte array per record.
BinarySample - Class in org.apache.spark.mllib.stat.test
Class that represents the group and value of a sample.
BinarySample(boolean, double) - Constructor for class org.apache.spark.mllib.stat.test.BinarySample
 
binarySummary() - Method in class org.apache.spark.ml.classification.LogisticRegressionModel
Gets summary of model on training set.
binarySummary() - Method in class org.apache.spark.ml.classification.RandomForestClassificationModel
Gets summary of model on training set.
BinaryType - Class in org.apache.spark.sql.types
The data type representing Array[Byte] values.
BinaryType() - Constructor for class org.apache.spark.sql.types.BinaryType
 
BinaryType - Static variable in class org.apache.spark.sql.types.DataTypes
Gets the BinaryType object.
bind(StructType) - Method in interface org.apache.spark.sql.connector.catalog.functions.UnboundFunction
Bind this function to an input type.
Binomial$() - Constructor for class org.apache.spark.ml.regression.GeneralizedLinearRegression.Binomial$
 
BinomialBounds - Class in org.apache.spark.util.random
Utility functions that help us determine bounds on adjusted sampling rate to guarantee exact sample size with high confidence when sampling without replacement.
BinomialBounds() - Constructor for class org.apache.spark.util.random.BinomialBounds
 
bins() - Method in interface org.apache.spark.sql.connector.read.colstats.Histogram
 
BisectingKMeans - Class in org.apache.spark.ml.clustering
A bisecting k-means algorithm based on the paper "A comparison of document clustering techniques" by Steinbach, Karypis, and Kumar, with modification to fit Spark.
BisectingKMeans(String) - Constructor for class org.apache.spark.ml.clustering.BisectingKMeans
 
BisectingKMeans() - Constructor for class org.apache.spark.ml.clustering.BisectingKMeans
 
BisectingKMeans - Class in org.apache.spark.mllib.clustering
A bisecting k-means algorithm based on the paper "A comparison of document clustering techniques" by Steinbach, Karypis, and Kumar, with modification to fit Spark.
BisectingKMeans() - Constructor for class org.apache.spark.mllib.clustering.BisectingKMeans
Constructs with the default configuration
BisectingKMeansModel - Class in org.apache.spark.ml.clustering
Model fitted by BisectingKMeans.
BisectingKMeansModel - Class in org.apache.spark.mllib.clustering
Clustering model produced by BisectingKMeans.
BisectingKMeansModel(ClusteringTreeNode) - Constructor for class org.apache.spark.mllib.clustering.BisectingKMeansModel
 
BisectingKMeansModel.SaveLoadV1_0$ - Class in org.apache.spark.mllib.clustering
 
BisectingKMeansModel.SaveLoadV2_0$ - Class in org.apache.spark.mllib.clustering
 
BisectingKMeansModel.SaveLoadV3_0$ - Class in org.apache.spark.mllib.clustering
 
BisectingKMeansParams - Interface in org.apache.spark.ml.clustering
Common params for BisectingKMeans and BisectingKMeansModel
BisectingKMeansSummary - Class in org.apache.spark.ml.clustering
Summary of BisectingKMeans.
bit_length(Column) - Static method in class org.apache.spark.sql.functions
Calculates the bit length for the specified string column.
bitSize() - Method in class org.apache.spark.util.sketch.BloomFilter
Returns the number of bits in the underlying bit array.
bitwise_not(Column) - Static method in class org.apache.spark.sql.functions
Computes bitwise NOT (~) of a number.
bitwiseAND(Object) - Method in class org.apache.spark.sql.Column
Compute bitwise AND of this expression with another expression.
bitwiseNOT(Column) - Static method in class org.apache.spark.sql.functions
Deprecated.
Use bitwise_not. Since 3.2.0.
bitwiseOR(Object) - Method in class org.apache.spark.sql.Column
Compute bitwise OR of this expression with another expression.
bitwiseXOR(Object) - Method in class org.apache.spark.sql.Column
Compute bitwise XOR of this expression with another expression.
BLACKLISTED_IN_STAGES_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
 
blacklistedInStages() - Method in class org.apache.spark.status.api.v1.ExecutorSummary
Deprecated.
use excludedInStages instead. Since 3.1.0.
BLAS - Class in org.apache.spark.ml.linalg
BLAS routines for MLlib's vectors and matrices.
BLAS() - Constructor for class org.apache.spark.ml.linalg.BLAS
 
BLAS - Class in org.apache.spark.mllib.linalg
BLAS routines for MLlib's vectors and matrices.
BLAS() - Constructor for class org.apache.spark.mllib.linalg.BLAS
 
BLOCK_NAME_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo
 
BlockData - Interface in org.apache.spark.storage
Abstracts away how blocks are stored and provides different ways to read the underlying block data.
blockDoesNotExistError(BlockId) - Static method in class org.apache.spark.errors.SparkCoreErrors
 
blockedByLock() - Method in class org.apache.spark.status.api.v1.ThreadStackTrace
 
blockedByThreadId() - Method in class org.apache.spark.status.api.v1.ThreadStackTrace
 
BlockEvictionHandler - Interface in org.apache.spark.storage.memory
 
BlockGeneratorListener - Interface in org.apache.spark.streaming.receiver
Listener object for BlockGenerator events
blockHaveBeenRemovedError(String) - Static method in class org.apache.spark.errors.SparkCoreErrors
 
BlockId - Class in org.apache.spark.storage
:: DeveloperApi :: Identifies a particular Block of data, usually associated with a single file.
BlockId() - Constructor for class org.apache.spark.storage.BlockId
 
blockId() - Method in class org.apache.spark.storage.BlockManagerMessages.GetBlockStatus
 
blockId() - Method in class org.apache.spark.storage.BlockManagerMessages.GetLocations
 
blockId() - Method in class org.apache.spark.storage.BlockManagerMessages.GetLocationsAndStatus
 
blockId() - Method in class org.apache.spark.storage.BlockManagerMessages.RemoveBlock
 
blockId() - Method in class org.apache.spark.storage.BlockManagerMessages.ReplicateBlock
 
blockId() - Method in class org.apache.spark.storage.BlockManagerMessages.UpdateBlockInfo
 
blockId() - Method in class org.apache.spark.storage.BlockUpdatedInfo
 
blockId() - Method in interface org.apache.spark.streaming.receiver.ReceivedBlockStoreResult
 
blockIds() - Method in class org.apache.spark.storage.BlockManagerMessages.GetLocationsMultipleBlockIds
 
BlockInfoWrapper - Class in org.apache.spark.storage
 
BlockInfoWrapper(BlockInfo, Lock, Condition) - Constructor for class org.apache.spark.storage.BlockInfoWrapper
 
BlockInfoWrapper(BlockInfo, Lock) - Constructor for class org.apache.spark.storage.BlockInfoWrapper
 
BlockLocationsAndStatus(Seq<BlockManagerId>, BlockStatus, Option<String[]>) - Constructor for class org.apache.spark.storage.BlockManagerMessages.BlockLocationsAndStatus
 
BlockLocationsAndStatus$() - Constructor for class org.apache.spark.storage.BlockManagerMessages.BlockLocationsAndStatus$
 
blockManager() - Method in class org.apache.spark.SparkEnv
 
blockManagerAddedFromJson(JsonNode) - Static method in class org.apache.spark.util.JsonProtocol
 
blockManagerAddedToJson(SparkListenerBlockManagerAdded, JsonGenerator) - Static method in class org.apache.spark.util.JsonProtocol
 
BlockManagerHeartbeat(BlockManagerId) - Constructor for class org.apache.spark.storage.BlockManagerMessages.BlockManagerHeartbeat
 
BlockManagerHeartbeat$() - Constructor for class org.apache.spark.storage.BlockManagerMessages.BlockManagerHeartbeat$
 
blockManagerId() - Method in class org.apache.spark.scheduler.SparkListenerBlockManagerAdded
 
blockManagerId() - Method in class org.apache.spark.scheduler.SparkListenerBlockManagerRemoved
 
BlockManagerId - Class in org.apache.spark.storage
:: DeveloperApi :: This class represent a unique identifier for a BlockManager.
BlockManagerId() - Constructor for class org.apache.spark.storage.BlockManagerId
 
blockManagerId() - Method in class org.apache.spark.storage.BlockManagerMessages.BlockManagerHeartbeat
 
blockManagerId() - Method in class org.apache.spark.storage.BlockManagerMessages.GetPeers
 
blockManagerId() - Method in class org.apache.spark.storage.BlockManagerMessages.GetReplicateInfoForRDDBlocks
 
blockManagerId() - Method in class org.apache.spark.storage.BlockManagerMessages.RegisterBlockManager
 
blockManagerId() - Method in class org.apache.spark.storage.BlockManagerMessages.UpdateBlockInfo
 
blockManagerId() - Method in class org.apache.spark.storage.BlockUpdatedInfo
 
blockManagerIdCache() - Static method in class org.apache.spark.storage.BlockManagerId
The max cache size is hardcoded to 10000, since the size of a BlockManagerId object is about 48B, the total memory cost should be below 1MB which is feasible.
blockManagerIdFromJson(JsonNode) - Static method in class org.apache.spark.util.JsonProtocol
 
blockManagerIdToJson(BlockManagerId, JsonGenerator) - Static method in class org.apache.spark.util.JsonProtocol
 
BlockManagerMessages - Class in org.apache.spark.storage
 
BlockManagerMessages() - Constructor for class org.apache.spark.storage.BlockManagerMessages
 
BlockManagerMessages.BlockLocationsAndStatus - Class in org.apache.spark.storage
The response message of GetLocationsAndStatus request.
BlockManagerMessages.BlockLocationsAndStatus$ - Class in org.apache.spark.storage
 
BlockManagerMessages.BlockManagerHeartbeat - Class in org.apache.spark.storage
 
BlockManagerMessages.BlockManagerHeartbeat$ - Class in org.apache.spark.storage
 
BlockManagerMessages.DecommissionBlockManager$ - Class in org.apache.spark.storage
 
BlockManagerMessages.DecommissionBlockManagers - Class in org.apache.spark.storage
 
BlockManagerMessages.DecommissionBlockManagers$ - Class in org.apache.spark.storage
 
BlockManagerMessages.GetBlockStatus - Class in org.apache.spark.storage
 
BlockManagerMessages.GetBlockStatus$ - Class in org.apache.spark.storage
 
BlockManagerMessages.GetExecutorEndpointRef - Class in org.apache.spark.storage
 
BlockManagerMessages.GetExecutorEndpointRef$ - Class in org.apache.spark.storage
 
BlockManagerMessages.GetLocations - Class in org.apache.spark.storage
 
BlockManagerMessages.GetLocations$ - Class in org.apache.spark.storage
 
BlockManagerMessages.GetLocationsAndStatus - Class in org.apache.spark.storage
 
BlockManagerMessages.GetLocationsAndStatus$ - Class in org.apache.spark.storage
 
BlockManagerMessages.GetLocationsMultipleBlockIds - Class in org.apache.spark.storage
 
BlockManagerMessages.GetLocationsMultipleBlockIds$ - Class in org.apache.spark.storage
 
BlockManagerMessages.GetMatchingBlockIds - Class in org.apache.spark.storage
 
BlockManagerMessages.GetMatchingBlockIds$ - Class in org.apache.spark.storage
 
BlockManagerMessages.GetMemoryStatus$ - Class in org.apache.spark.storage
 
BlockManagerMessages.GetPeers - Class in org.apache.spark.storage
 
BlockManagerMessages.GetPeers$ - Class in org.apache.spark.storage
 
BlockManagerMessages.GetReplicateInfoForRDDBlocks - Class in org.apache.spark.storage
 
BlockManagerMessages.GetReplicateInfoForRDDBlocks$ - Class in org.apache.spark.storage
 
BlockManagerMessages.GetShufflePushMergerLocations - Class in org.apache.spark.storage
 
BlockManagerMessages.GetShufflePushMergerLocations$ - Class in org.apache.spark.storage
 
BlockManagerMessages.GetStorageStatus$ - Class in org.apache.spark.storage
 
BlockManagerMessages.IsExecutorAlive - Class in org.apache.spark.storage
 
BlockManagerMessages.IsExecutorAlive$ - Class in org.apache.spark.storage
 
BlockManagerMessages.RegisterBlockManager - Class in org.apache.spark.storage
 
BlockManagerMessages.RegisterBlockManager$ - Class in org.apache.spark.storage
 
BlockManagerMessages.RemoveBlock - Class in org.apache.spark.storage
 
BlockManagerMessages.RemoveBlock$ - Class in org.apache.spark.storage
 
BlockManagerMessages.RemoveBroadcast - Class in org.apache.spark.storage
 
BlockManagerMessages.RemoveBroadcast$ - Class in org.apache.spark.storage
 
BlockManagerMessages.RemoveExecutor - Class in org.apache.spark.storage
 
BlockManagerMessages.RemoveExecutor$ - Class in org.apache.spark.storage
 
BlockManagerMessages.RemoveRdd - Class in org.apache.spark.storage
 
BlockManagerMessages.RemoveRdd$ - Class in org.apache.spark.storage
 
BlockManagerMessages.RemoveShuffle - Class in org.apache.spark.storage
 
BlockManagerMessages.RemoveShuffle$ - Class in org.apache.spark.storage
 
BlockManagerMessages.RemoveShufflePushMergerLocation - Class in org.apache.spark.storage
 
BlockManagerMessages.RemoveShufflePushMergerLocation$ - Class in org.apache.spark.storage
 
BlockManagerMessages.ReplicateBlock - Class in org.apache.spark.storage
 
BlockManagerMessages.ReplicateBlock$ - Class in org.apache.spark.storage
 
BlockManagerMessages.StopBlockManagerMaster$ - Class in org.apache.spark.storage
 
BlockManagerMessages.ToBlockManagerMaster - Interface in org.apache.spark.storage
 
BlockManagerMessages.ToBlockManagerMasterStorageEndpoint - Interface in org.apache.spark.storage
 
BlockManagerMessages.TriggerThreadDump$ - Class in org.apache.spark.storage
Driver to Executor message to trigger a thread dump.
BlockManagerMessages.UpdateBlockInfo - Class in org.apache.spark.storage
 
BlockManagerMessages.UpdateBlockInfo$ - Class in org.apache.spark.storage
 
blockManagerRemovedFromJson(JsonNode) - Static method in class org.apache.spark.util.JsonProtocol
 
blockManagerRemovedToJson(SparkListenerBlockManagerRemoved, JsonGenerator) - Static method in class org.apache.spark.util.JsonProtocol
 
BlockMatrix - Class in org.apache.spark.mllib.linalg.distributed
Represents a distributed matrix in blocks of local matrices.
BlockMatrix(RDD<Tuple2<Tuple2<Object, Object>, Matrix>>, int, int, long, long) - Constructor for class org.apache.spark.mllib.linalg.distributed.BlockMatrix
 
BlockMatrix(RDD<Tuple2<Tuple2<Object, Object>, Matrix>>, int, int) - Constructor for class org.apache.spark.mllib.linalg.distributed.BlockMatrix
Alternate constructor for BlockMatrix without the input of the number of rows and columns.
blockName() - Method in class org.apache.spark.status.api.v1.RDDPartitionInfo
 
blockName() - Method in class org.apache.spark.status.LiveRDDPartition
 
blockNotFoundError(BlockId) - Static method in class org.apache.spark.errors.SparkCoreErrors
 
BlockNotFoundException - Exception in org.apache.spark.storage
 
BlockNotFoundException(String) - Constructor for exception org.apache.spark.storage.BlockNotFoundException
 
BlockReplicationPolicy - Interface in org.apache.spark.storage
::DeveloperApi:: BlockReplicationPrioritization provides logic for prioritizing a sequence of peers for replicating blocks.
BlockReplicationUtils - Class in org.apache.spark.storage
 
BlockReplicationUtils() - Constructor for class org.apache.spark.storage.BlockReplicationUtils
 
blocks() - Method in class org.apache.spark.mllib.linalg.distributed.BlockMatrix
 
blockSize() - Method in class org.apache.spark.ml.classification.MultilayerPerceptronClassificationModel
 
blockSize() - Method in class org.apache.spark.ml.classification.MultilayerPerceptronClassifier
 
blockSize() - Method in interface org.apache.spark.ml.param.shared.HasBlockSize
Param for block size for stacking input data in matrices.
blockSize() - Method in class org.apache.spark.ml.recommendation.ALS
 
blockSize() - Method in class org.apache.spark.ml.recommendation.ALSModel
 
BlockStatus - Class in org.apache.spark.storage
 
BlockStatus(StorageLevel, long, long) - Constructor for class org.apache.spark.storage.BlockStatus
 
blockStatusFromJson(JsonNode) - Static method in class org.apache.spark.util.JsonProtocol
 
blockStatusQueryReturnedNullError(BlockId) - Static method in class org.apache.spark.errors.SparkCoreErrors
 
blockStatusToJson(BlockStatus, JsonGenerator) - Static method in class org.apache.spark.util.JsonProtocol
 
blockUpdatedInfo() - Method in class org.apache.spark.scheduler.SparkListenerBlockUpdated
 
BlockUpdatedInfo - Class in org.apache.spark.storage
:: DeveloperApi :: Stores information about a block status in a block manager.
BlockUpdatedInfo(BlockManagerId, BlockId, StorageLevel, long, long) - Constructor for class org.apache.spark.storage.BlockUpdatedInfo
 
blockUpdatedInfoFromJson(JsonNode) - Static method in class org.apache.spark.util.JsonProtocol
 
blockUpdatedInfoToJson(BlockUpdatedInfo, JsonGenerator) - Static method in class org.apache.spark.util.JsonProtocol
 
blockUpdateFromJson(JsonNode) - Static method in class org.apache.spark.util.JsonProtocol
 
blockUpdateToJson(SparkListenerBlockUpdated, JsonGenerator) - Static method in class org.apache.spark.util.JsonProtocol
 
bloomFilter(String, long, double) - Method in class org.apache.spark.sql.DataFrameStatFunctions
Builds a Bloom filter over a specified column.
bloomFilter(Column, long, double) - Method in class org.apache.spark.sql.DataFrameStatFunctions
Builds a Bloom filter over a specified column.
bloomFilter(String, long, long) - Method in class org.apache.spark.sql.DataFrameStatFunctions
Builds a Bloom filter over a specified column.
bloomFilter(Column, long, long) - Method in class org.apache.spark.sql.DataFrameStatFunctions
Builds a Bloom filter over a specified column.
BloomFilter - Class in org.apache.spark.util.sketch
A Bloom filter is a space-efficient probabilistic data structure that offers an approximate containment test with one-sided error: if it claims that an item is contained in it, this might be in error, but if it claims that an item is not contained in it, then this is definitely true.
BloomFilter() - Constructor for class org.apache.spark.util.sketch.BloomFilter
 
BloomFilter.Version - Enum in org.apache.spark.util.sketch
 
bmAddress() - Method in class org.apache.spark.FetchFailed
 
BOOLEAN() - Static method in class org.apache.spark.sql.Encoders
An encoder for nullable boolean type.
booleanColumn(String[]) - Static method in class org.apache.parquet.filter2.predicate.SparkFilterApi
 
BooleanParam - Class in org.apache.spark.ml.param
Specialized version of Param[Boolean] for Java.
BooleanParam(String, String, String) - Constructor for class org.apache.spark.ml.param.BooleanParam
 
BooleanParam(Identifiable, String, String) - Constructor for class org.apache.spark.ml.param.BooleanParam
 
BooleanType - Class in org.apache.spark.sql.types
The data type representing Boolean values.
BooleanType() - Constructor for class org.apache.spark.sql.types.BooleanType
 
BooleanType - Static variable in class org.apache.spark.sql.types.DataTypes
Gets the BooleanType object.
boost(RDD<org.apache.spark.ml.feature.Instance>, RDD<org.apache.spark.ml.feature.Instance>, BoostingStrategy, boolean, long, String, Option<org.apache.spark.ml.util.Instrumentation>) - Static method in class org.apache.spark.ml.tree.impl.GradientBoostedTrees
Internal method for performing regression using trees as base learners.
BoostingStrategy - Class in org.apache.spark.mllib.tree.configuration
Configuration options for GradientBoostedTrees.
BoostingStrategy(Strategy, Loss, int, double, double) - Constructor for class org.apache.spark.mllib.tree.configuration.BoostingStrategy
 
bootstrap() - Method in class org.apache.spark.ml.classification.RandomForestClassificationModel
 
bootstrap() - Method in class org.apache.spark.ml.classification.RandomForestClassifier
 
bootstrap() - Method in class org.apache.spark.ml.regression.RandomForestRegressionModel
 
bootstrap() - Method in class org.apache.spark.ml.regression.RandomForestRegressor
 
bootstrap() - Method in interface org.apache.spark.ml.tree.RandomForestParams
Whether bootstrap samples are used when building trees.
Both() - Static method in class org.apache.spark.graphx.EdgeDirection
Edges originating from *and* arriving at a vertex of interest.
boundaries() - Method in class org.apache.spark.ml.regression.IsotonicRegressionModel
Boundaries in increasing order for which predictions are known.
boundaries() - Method in class org.apache.spark.mllib.regression.IsotonicRegressionModel
 
BoundedDouble - Class in org.apache.spark.partial
A Double value with error bars and associated confidence.
BoundedDouble(double, double, double, double) - Constructor for class org.apache.spark.partial.BoundedDouble
 
BoundFunction - Interface in org.apache.spark.sql.connector.catalog.functions
Represents a function that is bound to an input type.
BreezeUtil - Class in org.apache.spark.ml.ann
In-place DGEMM and DGEMV for Breeze
BreezeUtil() - Constructor for class org.apache.spark.ml.ann.BreezeUtil
 
broadcast(T) - Method in class org.apache.spark.api.java.JavaSparkContext
Broadcast a read-only variable to the cluster, returning a Broadcast object for reading it in distributed functions.
Broadcast<T> - Class in org.apache.spark.broadcast
A broadcast variable.
Broadcast(long, ClassTag<T>) - Constructor for class org.apache.spark.broadcast.Broadcast
 
broadcast(T, ClassTag<T>) - Method in class org.apache.spark.SparkContext
Broadcast a read-only variable to the cluster, returning a Broadcast object for reading it in distributed functions.
broadcast(Dataset<T>) - Static method in class org.apache.spark.sql.functions
Marks a DataFrame as small enough for use in broadcast joins.
BROADCAST() - Static method in class org.apache.spark.storage.BlockId
 
BroadcastBlockId - Class in org.apache.spark.storage
 
BroadcastBlockId(long, String) - Constructor for class org.apache.spark.storage.BroadcastBlockId
 
broadcastCleaned(long) - Method in interface org.apache.spark.CleanerListener
 
BroadcastFactory - Interface in org.apache.spark.broadcast
An interface for all the broadcast implementations in Spark (to allow multiple broadcast implementations).
broadcastId() - Method in class org.apache.spark.CleanBroadcast
 
broadcastId() - Method in class org.apache.spark.storage.BlockManagerMessages.RemoveBroadcast
 
broadcastId() - Method in class org.apache.spark.storage.BroadcastBlockId
 
broadcastManager() - Method in class org.apache.spark.SparkEnv
 
bround(Column) - Static method in class org.apache.spark.sql.functions
Returns the value of the column e rounded to 0 decimal places with HALF_EVEN round mode.
bround(Column, int) - Static method in class org.apache.spark.sql.functions
Round the value of e to scale decimal places with HALF_EVEN round mode if scale is greater than or equal to 0 or at integral part when scale is less than 0.
bucket(int, String...) - Static method in class org.apache.spark.sql.connector.expressions.Expressions
Create a bucket transform for one or more columns.
bucket(int, NamedReference[]) - Static method in class org.apache.spark.sql.connector.expressions.LogicalExpressions
 
bucket(int, NamedReference[], NamedReference[]) - Static method in class org.apache.spark.sql.connector.expressions.LogicalExpressions
 
bucket(Column, Column) - Static method in class org.apache.spark.sql.functions
A transform for any type that partitions by a hash of the input column.
bucket(int, Column) - Static method in class org.apache.spark.sql.functions
A transform for any type that partitions by a hash of the input column.
bucketBy(int, String, String...) - Method in class org.apache.spark.sql.DataFrameWriter
Buckets the output by the given columns.
bucketBy(int, String, Seq<String>) - Method in class org.apache.spark.sql.DataFrameWriter
Buckets the output by the given columns.
bucketByAndSortByUnsupportedByOperationError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
bucketByUnsupportedByOperationError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
BucketedRandomProjectionLSH - Class in org.apache.spark.ml.feature
This BucketedRandomProjectionLSH implements Locality Sensitive Hashing functions for Euclidean distance metrics.
BucketedRandomProjectionLSH(String) - Constructor for class org.apache.spark.ml.feature.BucketedRandomProjectionLSH
 
BucketedRandomProjectionLSH() - Constructor for class org.apache.spark.ml.feature.BucketedRandomProjectionLSH
 
BucketedRandomProjectionLSHModel - Class in org.apache.spark.ml.feature
Model produced by BucketedRandomProjectionLSH, where multiple random vectors are stored.
BucketedRandomProjectionLSHParams - Interface in org.apache.spark.ml.feature
bucketingColumnCannotBePartOfPartitionColumnsError(String, Seq<String>) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
Bucketizer - Class in org.apache.spark.ml.feature
Bucketizer maps a column of continuous features to a column of feature buckets.
Bucketizer(String) - Constructor for class org.apache.spark.ml.feature.Bucketizer
 
Bucketizer() - Constructor for class org.apache.spark.ml.feature.Bucketizer
 
bucketLength() - Method in class org.apache.spark.ml.feature.BucketedRandomProjectionLSH
 
bucketLength() - Method in class org.apache.spark.ml.feature.BucketedRandomProjectionLSHModel
 
bucketLength() - Method in interface org.apache.spark.ml.feature.BucketedRandomProjectionLSHParams
The length of each hash bucket, a larger bucket lowers the false negative rate.
bucketSortingColumnCannotBePartOfPartitionColumnsError(String, Seq<String>) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
BucketSpecHelper(BucketSpec) - Constructor for class org.apache.spark.sql.connector.catalog.CatalogV2Implicits.BucketSpecHelper
 
buffer() - Method in class org.apache.spark.storage.memory.SerializedMemoryEntry
 
bufferEncoder() - Method in class org.apache.spark.ml.feature.StringIndexerAggregator
 
bufferEncoder() - Method in class org.apache.spark.sql.expressions.Aggregator
Specifies the Encoder for the intermediate value type.
BufferReleasingInputStream - Class in org.apache.spark.storage
Helper class that ensures a ManagedBuffer is released upon InputStream.close() and also detects stream corruption if streamCompressedOrEncrypted is true
BufferReleasingInputStream(InputStream, ShuffleBlockFetcherIterator, BlockId, int, BlockManagerId, boolean, boolean, Option<CheckedInputStream>) - Constructor for class org.apache.spark.storage.BufferReleasingInputStream
 
bufferSchema() - Method in class org.apache.spark.sql.expressions.UserDefinedAggregateFunction
Deprecated.
A StructType represents data types of values in the aggregation buffer.
build(Node, int) - Method in class org.apache.spark.ml.tree.DecisionTreeModelReadWrite.NodeData$
Create DecisionTreeModelReadWrite.NodeData instances for this node and all children.
build(DecisionTreeModel, int) - Method in class org.apache.spark.ml.tree.EnsembleModelReadWrite.EnsembleNodeData$
Create EnsembleModelReadWrite.EnsembleNodeData instances for the given tree.
build() - Method in class org.apache.spark.ml.tuning.ParamGridBuilder
Builds and returns all combinations of parameters specified by the param grid.
build() - Method in class org.apache.spark.resource.ResourceProfileBuilder
 
build() - Method in interface org.apache.spark.sql.connector.read.ScanBuilder
 
build(Expression) - Method in class org.apache.spark.sql.connector.util.V2ExpressionSQLBuilder
 
build() - Method in interface org.apache.spark.sql.connector.write.DeltaWriteBuilder
 
build() - Method in interface org.apache.spark.sql.connector.write.RowLevelOperationBuilder
Returns a RowLevelOperation that controls how Spark rewrites data for DELETE, UPDATE, MERGE commands.
build() - Method in interface org.apache.spark.sql.connector.write.WriteBuilder
Returns a logical Write shared between batch and streaming.
build() - Method in class org.apache.spark.sql.types.MetadataBuilder
Builds the Metadata instance.
build() - Method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo.Builder
 
build() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo.Builder
 
build() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
 
build() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoWrapper.Builder
 
build() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
 
build() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoWrapper.Builder
 
build() - Method in class org.apache.spark.status.protobuf.StoreTypes.AppSummary.Builder
 
build() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
 
build() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetrics.Builder
 
build() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
 
build() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions.Builder
 
build() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest.Builder
 
build() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
 
build() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper.Builder
 
build() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
 
build() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryWrapper.Builder
 
build() - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributions.Builder
 
build() - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetrics.Builder
 
build() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
 
build() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper.Builder
 
build() - Method in class org.apache.spark.status.protobuf.StoreTypes.MemoryMetrics.Builder
 
build() - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributions.Builder
 
build() - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetrics.Builder
 
build() - Method in class org.apache.spark.status.protobuf.StoreTypes.PairStrings.Builder
 
build() - Method in class org.apache.spark.status.protobuf.StoreTypes.PoolData.Builder
 
build() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder
 
build() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryWrapper.Builder
 
build() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution.Builder
 
build() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
 
build() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationEdge.Builder
 
build() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
 
build() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode.Builder
 
build() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo.Builder
 
build() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
 
build() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoWrapper.Builder
 
build() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceInformation.Builder
 
build() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo.Builder
 
build() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileWrapper.Builder
 
build() - Method in class org.apache.spark.status.protobuf.StoreTypes.RuntimeInfo.Builder
 
build() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
 
build() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics.Builder
 
build() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
 
build() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics.Builder
 
build() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions.Builder
 
build() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetrics.Builder
 
build() - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress.Builder
 
build() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
 
build() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
 
build() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphEdge.Builder
 
build() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
 
build() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper.Builder
 
build() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
 
build() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummary.Builder
 
build() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryWrapper.Builder
 
build() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
 
build() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetric.Builder
 
build() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
 
build() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper.Builder
 
build() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
 
build() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
 
build() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData.Builder
 
build() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
 
build() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressWrapper.Builder
 
build() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
 
build() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
 
build() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
 
build() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
 
build() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskResourceRequest.Builder
 
build() - Method in interface org.apache.spark.storage.memory.MemoryEntryBuilder
 
build() - Method in class org.apache.spark.streaming.kinesis.SparkAWSCredentials.Builder
Returns the appropriate instance of SparkAWSCredentials given the configured parameters.
builder() - Static method in class org.apache.spark.sql.SparkSession
Creates a SparkSession.Builder for constructing a SparkSession.
Builder() - Constructor for class org.apache.spark.sql.SparkSession.Builder
 
Builder() - Constructor for class org.apache.spark.streaming.kinesis.SparkAWSCredentials.Builder
 
buildErrorResponse(Response.Status, String) - Static method in class org.apache.spark.ui.UIUtils
 
buildFilter(Seq<Expression>, Seq<Attribute>) - Method in interface org.apache.spark.sql.columnar.CachedBatchSerializer
Builds a function that can be used to filter batches prior to being decompressed.
buildFilter(Seq<Expression>, Seq<Attribute>) - Method in class org.apache.spark.sql.columnar.SimpleMetricsCachedBatchSerializer
 
buildForBatch() - Method in interface org.apache.spark.sql.connector.write.WriteBuilder
Deprecated.
buildForStreaming() - Method in interface org.apache.spark.sql.connector.write.WriteBuilder
Deprecated.
buildLocationMetadata(Seq<Path>, int) - Static method in class org.apache.spark.util.Utils
Convert a sequence of Paths to a metadata string.
buildPartial() - Method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo.Builder
 
buildPartial() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo.Builder
 
buildPartial() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
 
buildPartial() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoWrapper.Builder
 
buildPartial() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
 
buildPartial() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoWrapper.Builder
 
buildPartial() - Method in class org.apache.spark.status.protobuf.StoreTypes.AppSummary.Builder
 
buildPartial() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
 
buildPartial() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetrics.Builder
 
buildPartial() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
 
buildPartial() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions.Builder
 
buildPartial() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest.Builder
 
buildPartial() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
 
buildPartial() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper.Builder
 
buildPartial() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
 
buildPartial() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryWrapper.Builder
 
buildPartial() - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributions.Builder
 
buildPartial() - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetrics.Builder
 
buildPartial() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
 
buildPartial() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper.Builder
 
buildPartial() - Method in class org.apache.spark.status.protobuf.StoreTypes.MemoryMetrics.Builder
 
buildPartial() - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributions.Builder
 
buildPartial() - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetrics.Builder
 
buildPartial() - Method in class org.apache.spark.status.protobuf.StoreTypes.PairStrings.Builder
 
buildPartial() - Method in class org.apache.spark.status.protobuf.StoreTypes.PoolData.Builder
 
buildPartial() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder
 
buildPartial() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryWrapper.Builder
 
buildPartial() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution.Builder
 
buildPartial() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
 
buildPartial() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationEdge.Builder
 
buildPartial() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
 
buildPartial() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode.Builder
 
buildPartial() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo.Builder
 
buildPartial() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
 
buildPartial() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoWrapper.Builder
 
buildPartial() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceInformation.Builder
 
buildPartial() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo.Builder
 
buildPartial() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileWrapper.Builder
 
buildPartial() - Method in class org.apache.spark.status.protobuf.StoreTypes.RuntimeInfo.Builder
 
buildPartial() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
 
buildPartial() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics.Builder
 
buildPartial() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
 
buildPartial() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics.Builder
 
buildPartial() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions.Builder
 
buildPartial() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetrics.Builder
 
buildPartial() - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress.Builder
 
buildPartial() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
 
buildPartial() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
 
buildPartial() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphEdge.Builder
 
buildPartial() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
 
buildPartial() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper.Builder
 
buildPartial() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
 
buildPartial() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummary.Builder
 
buildPartial() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryWrapper.Builder
 
buildPartial() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
 
buildPartial() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetric.Builder
 
buildPartial() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
 
buildPartial() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper.Builder
 
buildPartial() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
 
buildPartial() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
 
buildPartial() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData.Builder
 
buildPartial() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
 
buildPartial() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressWrapper.Builder
 
buildPartial() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
 
buildPartial() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
 
buildPartial() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
 
buildPartial() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
 
buildPartial() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskResourceRequest.Builder
 
buildPlanNormalizationRules(SparkSession) - Method in class org.apache.spark.sql.SparkSessionExtensions
 
buildPools() - Method in interface org.apache.spark.scheduler.SchedulableBuilder
 
buildReaderUnsupportedForFileFormatError(String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
buildScan(Seq<Attribute>, Seq<Expression>) - Method in interface org.apache.spark.sql.sources.CatalystScan
 
buildScan(String[], Filter[]) - Method in interface org.apache.spark.sql.sources.PrunedFilteredScan
 
buildScan(String[]) - Method in interface org.apache.spark.sql.sources.PrunedScan
 
buildScan() - Method in interface org.apache.spark.sql.sources.TableScan
 
buildTreeFromNodes(DecisionTreeModelReadWrite.NodeData[], String) - Static method in class org.apache.spark.ml.tree.DecisionTreeModelReadWrite
Given all data for all nodes in a tree, rebuild the tree.
BYTE() - Static method in class org.apache.spark.api.r.SerializationFormats
 
BYTE() - Static method in class org.apache.spark.sql.Encoders
An encoder for nullable byte type.
BytecodeUtils - Class in org.apache.spark.graphx.util
Includes an utility function to test whether a function accesses a specific attribute of an object.
BytecodeUtils() - Constructor for class org.apache.spark.graphx.util.BytecodeUtils
 
ByteExactNumeric - Class in org.apache.spark.sql.types
 
ByteExactNumeric() - Constructor for class org.apache.spark.sql.types.ByteExactNumeric
 
BYTES_READ() - Method in class org.apache.spark.InternalAccumulator.input$
 
BYTES_READ_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
 
BYTES_READ_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributions
 
BYTES_READ_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.InputMetrics
 
BYTES_WRITTEN() - Method in class org.apache.spark.InternalAccumulator.output$
 
BYTES_WRITTEN() - Method in class org.apache.spark.InternalAccumulator.shuffleWrite$
 
BYTES_WRITTEN_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
 
BYTES_WRITTEN_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributions
 
BYTES_WRITTEN_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.OutputMetrics
 
BYTES_WRITTEN_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetrics
 
bytesRead() - Method in class org.apache.spark.status.api.v1.InputMetricDistributions
 
bytesRead() - Method in class org.apache.spark.status.api.v1.InputMetrics
 
bytesToString(long) - Static method in class org.apache.spark.util.Utils
Convert a quantity in bytes to a human-readable string such as "4.0 MiB".
bytesToString(BigInt) - Static method in class org.apache.spark.util.Utils
 
byteStringAsBytes(String) - Static method in class org.apache.spark.util.Utils
Convert a passed byte string (e.g.
byteStringAsGb(String) - Static method in class org.apache.spark.util.Utils
Convert a passed byte string (e.g.
byteStringAsKb(String) - Static method in class org.apache.spark.util.Utils
Convert a passed byte string (e.g.
byteStringAsMb(String) - Static method in class org.apache.spark.util.Utils
Convert a passed byte string (e.g.
bytesWritten() - Method in class org.apache.spark.status.api.v1.OutputMetricDistributions
 
bytesWritten() - Method in class org.apache.spark.status.api.v1.OutputMetrics
 
bytesWritten() - Method in class org.apache.spark.status.api.v1.ShuffleWriteMetrics
 
bytesWritten(long) - Method in interface org.apache.spark.util.logging.RollingPolicy
Notify that bytes have been written
ByteType - Class in org.apache.spark.sql.types
The data type representing Byte values.
ByteType() - Constructor for class org.apache.spark.sql.types.ByteType
 
ByteType - Static variable in class org.apache.spark.sql.types.DataTypes
Gets the ByteType object.

C

cache() - Method in class org.apache.spark.api.java.JavaDoubleRDD
Persist this RDD with the default storage level (MEMORY_ONLY).
cache() - Method in class org.apache.spark.api.java.JavaPairRDD
Persist this RDD with the default storage level (MEMORY_ONLY).
cache() - Method in class org.apache.spark.api.java.JavaRDD
Persist this RDD with the default storage level (MEMORY_ONLY).
cache() - Method in class org.apache.spark.graphx.Graph
Caches the vertices and edges associated with this graph at the previously-specified target storage levels, which default to MEMORY_ONLY.
cache() - Method in class org.apache.spark.graphx.impl.EdgeRDDImpl
Persists the edge partitions using targetStorageLevel, which defaults to MEMORY_ONLY.
cache() - Method in class org.apache.spark.graphx.impl.GraphImpl
 
cache() - Method in class org.apache.spark.graphx.impl.VertexRDDImpl
Persists the vertex partitions at targetStorageLevel, which defaults to MEMORY_ONLY.
cache() - Method in class org.apache.spark.mllib.linalg.distributed.BlockMatrix
Caches the underlying RDD.
cache() - Method in class org.apache.spark.rdd.RDD
Persist this RDD with the default storage level (MEMORY_ONLY).
cache() - Method in class org.apache.spark.sql.Dataset
Persist this Dataset with the default storage level (MEMORY_AND_DISK).
cache() - Method in class org.apache.spark.streaming.api.java.JavaDStream
Persist RDDs of this DStream with the default storage level (MEMORY_ONLY_SER)
cache() - Method in class org.apache.spark.streaming.api.java.JavaPairDStream
Persist RDDs of this DStream with the default storage level (MEMORY_ONLY_SER)
cache() - Method in class org.apache.spark.streaming.dstream.DStream
Persist RDDs of this DStream with the default storage level (MEMORY_ONLY_SER)
CACHED_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode
 
CACHED_PARTITIONS() - Static method in class org.apache.spark.ui.storage.ToolTips
 
CachedBatch - Interface in org.apache.spark.sql.columnar
Basic interface that all cached batches of data must support.
CachedBatchSerializer - Interface in org.apache.spark.sql.columnar
Provides APIs that handle transformations of SQL data associated with the cache/persist APIs.
cacheNodeIds() - Method in class org.apache.spark.ml.classification.DecisionTreeClassificationModel
 
cacheNodeIds() - Method in class org.apache.spark.ml.classification.DecisionTreeClassifier
 
cacheNodeIds() - Method in class org.apache.spark.ml.classification.GBTClassificationModel
 
cacheNodeIds() - Method in class org.apache.spark.ml.classification.GBTClassifier
 
cacheNodeIds() - Method in class org.apache.spark.ml.classification.RandomForestClassificationModel
 
cacheNodeIds() - Method in class org.apache.spark.ml.classification.RandomForestClassifier
 
cacheNodeIds() - Method in class org.apache.spark.ml.regression.DecisionTreeRegressionModel
 
cacheNodeIds() - Method in class org.apache.spark.ml.regression.DecisionTreeRegressor
 
cacheNodeIds() - Method in class org.apache.spark.ml.regression.GBTRegressionModel
 
cacheNodeIds() - Method in class org.apache.spark.ml.regression.GBTRegressor
 
cacheNodeIds() - Method in class org.apache.spark.ml.regression.RandomForestRegressionModel
 
cacheNodeIds() - Method in class org.apache.spark.ml.regression.RandomForestRegressor
 
cacheNodeIds() - Method in interface org.apache.spark.ml.tree.DecisionTreeParams
If false, the algorithm will pass trees to executors to match instances with nodes.
cacheSize() - Method in interface org.apache.spark.SparkExecutorInfo
 
cacheSize() - Method in class org.apache.spark.SparkExecutorInfoImpl
 
cacheTable(String) - Method in class org.apache.spark.sql.catalog.Catalog
Caches the specified table in-memory.
cacheTable(String, StorageLevel) - Method in class org.apache.spark.sql.catalog.Catalog
Caches the specified table with the given storage level.
cacheTable(String) - Method in class org.apache.spark.sql.SQLContext
Caches the specified table in-memory.
calculate(double[], double) - Static method in class org.apache.spark.mllib.tree.impurity.Entropy
information calculation for multiclass classification
calculate(double, double, double) - Static method in class org.apache.spark.mllib.tree.impurity.Entropy
variance calculation
calculate(double[], double) - Static method in class org.apache.spark.mllib.tree.impurity.Gini
information calculation for multiclass classification
calculate(double, double, double) - Static method in class org.apache.spark.mllib.tree.impurity.Gini
variance calculation
calculate(double[], double) - Method in interface org.apache.spark.mllib.tree.impurity.Impurity
information calculation for multiclass classification
calculate(double, double, double) - Method in interface org.apache.spark.mllib.tree.impurity.Impurity
information calculation for regression
calculate(double[], double) - Static method in class org.apache.spark.mllib.tree.impurity.Variance
information calculation for multiclass classification
calculate(double, double, double) - Static method in class org.apache.spark.mllib.tree.impurity.Variance
variance calculation
calculateAmountAndPartsForFraction(double) - Static method in class org.apache.spark.resource.ResourceUtils
 
calculateNumberOfPartitions(long, int, int) - Method in class org.apache.spark.ml.feature.Word2VecModel.Word2VecModelWriter$
Calculate the number of partitions to use in saving the model.
CalendarInterval - Class in org.apache.spark.unsafe.types
The class representing calendar intervals.
CalendarInterval(int, int, long) - Constructor for class org.apache.spark.unsafe.types.CalendarInterval
 
CalendarIntervalType - Class in org.apache.spark.sql.types
The data type representing calendar intervals.
CalendarIntervalType() - Constructor for class org.apache.spark.sql.types.CalendarIntervalType
 
CalendarIntervalType - Static variable in class org.apache.spark.sql.types.DataTypes
Gets the CalendarIntervalType object.
call(K, Iterator<V1>, Iterator<V2>) - Method in interface org.apache.spark.api.java.function.CoGroupFunction
 
call(T) - Method in interface org.apache.spark.api.java.function.DoubleFlatMapFunction
 
call(T) - Method in interface org.apache.spark.api.java.function.DoubleFunction
 
call(T) - Method in interface org.apache.spark.api.java.function.FilterFunction
 
call(T) - Method in interface org.apache.spark.api.java.function.FlatMapFunction
 
call(T1, T2) - Method in interface org.apache.spark.api.java.function.FlatMapFunction2
 
call(K, Iterator<V>) - Method in interface org.apache.spark.api.java.function.FlatMapGroupsFunction
 
call(K, Iterator<V>, GroupState<S>) - Method in interface org.apache.spark.api.java.function.FlatMapGroupsWithStateFunction
 
call(T) - Method in interface org.apache.spark.api.java.function.ForeachFunction
 
call(Iterator<T>) - Method in interface org.apache.spark.api.java.function.ForeachPartitionFunction
 
call(T1) - Method in interface org.apache.spark.api.java.function.Function
 
call() - Method in interface org.apache.spark.api.java.function.Function0
 
call(T1, T2) - Method in interface org.apache.spark.api.java.function.Function2
 
call(T1, T2, T3) - Method in interface org.apache.spark.api.java.function.Function3
 
call(T1, T2, T3, T4) - Method in interface org.apache.spark.api.java.function.Function4
 
call(T) - Method in interface org.apache.spark.api.java.function.MapFunction
 
call(K, Iterator<V>) - Method in interface org.apache.spark.api.java.function.MapGroupsFunction
 
call(K, Iterator<V>, GroupState<S>) - Method in interface org.apache.spark.api.java.function.MapGroupsWithStateFunction
 
call(Iterator<T>) - Method in interface org.apache.spark.api.java.function.MapPartitionsFunction
 
call(T) - Method in interface org.apache.spark.api.java.function.PairFlatMapFunction
 
call(T) - Method in interface org.apache.spark.api.java.function.PairFunction
 
call(T, T) - Method in interface org.apache.spark.api.java.function.ReduceFunction
 
call(T) - Method in interface org.apache.spark.api.java.function.VoidFunction
 
call(T1, T2) - Method in interface org.apache.spark.api.java.function.VoidFunction2
 
call() - Method in interface org.apache.spark.sql.api.java.UDF0
 
call(T1) - Method in interface org.apache.spark.sql.api.java.UDF1
 
call(T1, T2, T3, T4, T5, T6, T7, T8, T9, T10) - Method in interface org.apache.spark.sql.api.java.UDF10
 
call(T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11) - Method in interface org.apache.spark.sql.api.java.UDF11
 
call(T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12) - Method in interface org.apache.spark.sql.api.java.UDF12
 
call(T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13) - Method in interface org.apache.spark.sql.api.java.UDF13
 
call(T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14) - Method in interface org.apache.spark.sql.api.java.UDF14
 
call(T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14, T15) - Method in interface org.apache.spark.sql.api.java.UDF15
 
call(T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14, T15, T16) - Method in interface org.apache.spark.sql.api.java.UDF16
 
call(T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14, T15, T16, T17) - Method in interface org.apache.spark.sql.api.java.UDF17
 
call(T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14, T15, T16, T17, T18) - Method in interface org.apache.spark.sql.api.java.UDF18
 
call(T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14, T15, T16, T17, T18, T19) - Method in interface org.apache.spark.sql.api.java.UDF19
 
call(T1, T2) - Method in interface org.apache.spark.sql.api.java.UDF2
 
call(T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14, T15, T16, T17, T18, T19, T20) - Method in interface org.apache.spark.sql.api.java.UDF20
 
call(T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14, T15, T16, T17, T18, T19, T20, T21) - Method in interface org.apache.spark.sql.api.java.UDF21
 
call(T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14, T15, T16, T17, T18, T19, T20, T21, T22) - Method in interface org.apache.spark.sql.api.java.UDF22
 
call(T1, T2, T3) - Method in interface org.apache.spark.sql.api.java.UDF3
 
call(T1, T2, T3, T4) - Method in interface org.apache.spark.sql.api.java.UDF4
 
call(T1, T2, T3, T4, T5) - Method in interface org.apache.spark.sql.api.java.UDF5
 
call(T1, T2, T3, T4, T5, T6) - Method in interface org.apache.spark.sql.api.java.UDF6
 
call(T1, T2, T3, T4, T5, T6, T7) - Method in interface org.apache.spark.sql.api.java.UDF7
 
call(T1, T2, T3, T4, T5, T6, T7, T8) - Method in interface org.apache.spark.sql.api.java.UDF8
 
call(T1, T2, T3, T4, T5, T6, T7, T8, T9) - Method in interface org.apache.spark.sql.api.java.UDF9
 
call_udf(String, Column...) - Static method in class org.apache.spark.sql.functions
Call an user-defined function.
call_udf(String, Seq<Column>) - Static method in class org.apache.spark.sql.functions
Call an user-defined function.
callSite() - Method in class org.apache.spark.storage.RDDInfo
 
CALLSITE_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode
 
callUDF(String, Column...) - Static method in class org.apache.spark.sql.functions
Call an user-defined function.
callUDF(String, Seq<Column>) - Static method in class org.apache.spark.sql.functions
Deprecated.
Use call_udf. Since .
cancel() - Method in class org.apache.spark.ComplexFutureAction
 
cancel() - Method in interface org.apache.spark.FutureAction
Cancels the execution of this action.
cancel() - Method in class org.apache.spark.SimpleFutureAction
 
cancelAllJobs() - Method in class org.apache.spark.api.java.JavaSparkContext
Cancel all jobs that have been scheduled or are running.
cancelAllJobs() - Method in class org.apache.spark.SparkContext
Cancel all jobs that have been scheduled or are running.
cancelJob(int, String) - Method in class org.apache.spark.SparkContext
Cancel a given job if it's scheduled or running.
cancelJob(int) - Method in class org.apache.spark.SparkContext
Cancel a given job if it's scheduled or running.
cancelJobGroup(String) - Method in class org.apache.spark.api.java.JavaSparkContext
Cancel active jobs for the specified group.
cancelJobGroup(String) - Method in class org.apache.spark.SparkContext
Cancel active jobs for the specified group.
cancelStage(int, String) - Method in class org.apache.spark.SparkContext
Cancel a given stage and all jobs associated with it.
cancelStage(int) - Method in class org.apache.spark.SparkContext
Cancel a given stage and all jobs associated with it.
cancelTasks(int, boolean) - Method in interface org.apache.spark.scheduler.TaskScheduler
 
canCreate(String) - Method in interface org.apache.spark.scheduler.ExternalClusterManager
Check if this cluster manager instance can create scheduler components for a certain master URL.
canDeleteWhere(Filter[]) - Method in interface org.apache.spark.sql.connector.catalog.SupportsDelete
Checks whether it is possible to delete data from a data source table that matches filter expressions.
canDeleteWhere(Predicate[]) - Method in interface org.apache.spark.sql.connector.catalog.SupportsDelete
 
canDeleteWhere(Predicate[]) - Method in interface org.apache.spark.sql.connector.catalog.SupportsDeleteV2
Checks whether it is possible to delete data from a data source table that matches filter expressions.
canEqual(Object) - Static method in class org.apache.spark.ExpireDeadHosts
 
canEqual(Object) - Static method in class org.apache.spark.metrics.DirectPoolMemory
 
canEqual(Object) - Static method in class org.apache.spark.metrics.GarbageCollectionMetrics
 
canEqual(Object) - Static method in class org.apache.spark.metrics.JVMHeapMemory
 
canEqual(Object) - Static method in class org.apache.spark.metrics.JVMOffHeapMemory
 
canEqual(Object) - Static method in class org.apache.spark.metrics.MappedPoolMemory
 
canEqual(Object) - Static method in class org.apache.spark.metrics.OffHeapExecutionMemory
 
canEqual(Object) - Static method in class org.apache.spark.metrics.OffHeapStorageMemory
 
canEqual(Object) - Static method in class org.apache.spark.metrics.OffHeapUnifiedMemory
 
canEqual(Object) - Static method in class org.apache.spark.metrics.OnHeapExecutionMemory
 
canEqual(Object) - Static method in class org.apache.spark.metrics.OnHeapStorageMemory
 
canEqual(Object) - Static method in class org.apache.spark.metrics.OnHeapUnifiedMemory
 
canEqual(Object) - Static method in class org.apache.spark.metrics.ProcessTreeMetrics
 
canEqual(Object) - Static method in class org.apache.spark.ml.feature.Dot
 
canEqual(Object) - Static method in class org.apache.spark.ml.feature.EmptyTerm
 
canEqual(Object) - Static method in class org.apache.spark.Resubmitted
 
canEqual(Object) - Static method in class org.apache.spark.scheduler.AllJobsCancelled
 
canEqual(Object) - Method in class org.apache.spark.scheduler.cluster.ExecutorInfo
 
canEqual(Object) - Static method in class org.apache.spark.scheduler.JobSucceeded
 
canEqual(Object) - Static method in class org.apache.spark.scheduler.ResubmitFailedStages
 
canEqual(Object) - Static method in class org.apache.spark.scheduler.StopCoordinator
 
canEqual(Object) - Static method in class org.apache.spark.sql.jdbc.MySQLDialect
 
canEqual(Object) - Static method in class org.apache.spark.sql.jdbc.OracleDialect
 
canEqual(Object) - Static method in class org.apache.spark.sql.jdbc.TeradataDialect
 
canEqual(Object) - Static method in class org.apache.spark.sql.sources.AlwaysFalse
 
canEqual(Object) - Static method in class org.apache.spark.sql.sources.AlwaysTrue
 
canEqual(Object) - Static method in class org.apache.spark.sql.types.BinaryType
 
canEqual(Object) - Static method in class org.apache.spark.sql.types.BooleanType
 
canEqual(Object) - Static method in class org.apache.spark.sql.types.ByteType
 
canEqual(Object) - Static method in class org.apache.spark.sql.types.CalendarIntervalType
 
canEqual(Object) - Static method in class org.apache.spark.sql.types.DateType
 
canEqual(Object) - Static method in class org.apache.spark.sql.types.DayTimeIntervalType
 
canEqual(Object) - Static method in class org.apache.spark.sql.types.DoubleType
 
canEqual(Object) - Static method in class org.apache.spark.sql.types.FloatType
 
canEqual(Object) - Static method in class org.apache.spark.sql.types.IntegerType
 
canEqual(Object) - Static method in class org.apache.spark.sql.types.LongType
 
canEqual(Object) - Static method in class org.apache.spark.sql.types.NullType
 
canEqual(Object) - Static method in class org.apache.spark.sql.types.ShortType
 
canEqual(Object) - Static method in class org.apache.spark.sql.types.StringType
 
canEqual(Object) - Static method in class org.apache.spark.sql.types.TimestampNTZType
 
canEqual(Object) - Static method in class org.apache.spark.sql.types.TimestampType
 
canEqual(Object) - Static method in class org.apache.spark.sql.types.YearMonthIntervalType
 
canEqual(Object) - Static method in class org.apache.spark.StopMapOutputTracker
 
canEqual(Object) - Static method in class org.apache.spark.streaming.kinesis.DefaultCredentials
 
canEqual(Object) - Static method in class org.apache.spark.streaming.scheduler.AllReceiverIds
 
canEqual(Object) - Static method in class org.apache.spark.streaming.scheduler.GetAllReceiverInfo
 
canEqual(Object) - Static method in class org.apache.spark.streaming.scheduler.StopAllReceivers
 
canEqual(Object) - Static method in class org.apache.spark.Success
 
canEqual(Object) - Static method in class org.apache.spark.TaskResultLost
 
canEqual(Object) - Static method in class org.apache.spark.TaskSchedulerIsSet
 
canEqual(Object) - Static method in class org.apache.spark.UnknownReason
 
canEqual(Object) - Method in class org.apache.spark.util.MutablePair
 
canHandle(String) - Method in class org.apache.spark.sql.jdbc.AggregatedDialect
 
canHandle(String) - Static method in class org.apache.spark.sql.jdbc.DB2Dialect
 
canHandle(String) - Static method in class org.apache.spark.sql.jdbc.DerbyDialect
 
canHandle(String) - Static method in class org.apache.spark.sql.jdbc.H2Dialect
 
canHandle(Driver, Map<String, String>) - Method in class org.apache.spark.sql.jdbc.JdbcConnectionProvider
Checks if this connection provider instance can handle the connection initiated by the driver.
canHandle(String) - Method in class org.apache.spark.sql.jdbc.JdbcDialect
Check if this dialect instance can handle a certain jdbc url.
canHandle(String) - Static method in class org.apache.spark.sql.jdbc.MsSqlServerDialect
 
canHandle(String) - Static method in class org.apache.spark.sql.jdbc.MySQLDialect
 
canHandle(String) - Static method in class org.apache.spark.sql.jdbc.NoopDialect
 
canHandle(String) - Static method in class org.apache.spark.sql.jdbc.OracleDialect
 
canHandle(String) - Static method in class org.apache.spark.sql.jdbc.PostgresDialect
 
canHandle(String) - Static method in class org.apache.spark.sql.jdbc.TeradataDialect
 
cannotAcquireMemoryToBuildLongHashedRelationError(long, long) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
cannotAcquireMemoryToBuildUnsafeHashedRelationError() - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
cannotAddMultiPartitionsOnNonatomicPartitionTableError(String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
cannotAllocateMemoryToGrowBytesToBytesMapError() - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
cannotAlterTableWithAlterViewError() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
cannotAlterViewWithAlterTableError() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
cannotBroadcastTableOverMaxTableBytesError(long, long) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
cannotBroadcastTableOverMaxTableRowsError(long, long) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
cannotBuildHashedRelationLargerThan8GError() - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
cannotBuildHashedRelationWithUniqueKeysExceededError() - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
cannotCastError(DataType, DataType) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
cannotCastFromNullTypeError(DataType) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
cannotChangeDecimalPrecisionError(Decimal, int, int, SQLQueryContext) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
cannotChangeStorageLevelError() - Static method in class org.apache.spark.errors.SparkCoreErrors
 
cannotCleanReservedNamespacePropertyError(String, ParserRuleContext, String) - Static method in class org.apache.spark.sql.errors.QueryParsingErrors
 
cannotCleanReservedTablePropertyError(String, ParserRuleContext, String) - Static method in class org.apache.spark.sql.errors.QueryParsingErrors
 
cannotClearOutputDirectoryError(Path) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
cannotClearPartitionDirectoryError(Path) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
cannotCloneOrCopyReadOnlySQLConfError() - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
cannotCompareCostWithTargetCostError(String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
cannotConvertCatalystTypeToProtobufEnumTypeError(Seq<String>, String, String, String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
cannotConvertCatalystTypeToProtobufTypeError(Seq<String>, String, DataType, String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
cannotConvertDataTypeToParquetTypeError(StructField) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
cannotConvertOrcTimestampNTZToTimestampLTZError() - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
cannotConvertOrcTimestampToTimestampNTZError() - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
cannotConvertProtobufTypeToCatalystTypeError(String, DataType, Throwable) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
cannotConvertProtobufTypeToSqlTypeError(String, Seq<String>, String, DataType) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
cannotConvertSqlTypeToProtobufError(String, DataType, Throwable) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
cannotCreateArrayWithElementsExceedLimitError(long, String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
cannotCreateColumnarReaderError() - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
cannotCreateDatabaseWithSameNameAsPreservedDatabaseError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
cannotCreateJDBCNamespaceUsingProviderError() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
cannotCreateJDBCNamespaceWithPropertyError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
cannotCreateJDBCTableUsingLocationError() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
cannotCreateJDBCTableUsingProviderError() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
cannotCreateJDBCTableWithPartitionsError() - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
cannotCreateParquetConverterForDataTypeError(DataType, String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
cannotCreateParquetConverterForDecimalTypeError(DecimalType, String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
cannotCreateParquetConverterForTypeError(DecimalType, String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
cannotCreateStagingDirError(String, IOException) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
cannotCreateTableWithBothProviderAndSerdeError(Option<String>, Option<SerdeInfo>) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
cannotCreateTempViewUsingHiveDataSourceError() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
cannotDeleteTableWhereFiltersError(Table, Predicate[]) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
cannotDropBuiltinFuncError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
cannotDropDefaultDatabaseError() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
cannotDropMultiPartitionsOnNonatomicPartitionTableError(String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
cannotDropNonemptyDatabaseError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
cannotDropNonemptyNamespaceError(Seq<String>) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
cannotEvaluateExpressionError(Expression) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
cannotExecuteStreamingRelationExecError() - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
cannotFetchTablesOfDatabaseError(String, Exception) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
cannotFindCatalogToHandleIdentifierError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
cannotFindCatalystTypeInProtobufSchemaError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
cannotFindColumnError(String, String[]) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
cannotFindColumnInRelationOutputError(String, LogicalPlan) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
cannotFindConstructorForTypeError(String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
cannotFindDescriptorFileError(String, Throwable) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
cannotFindEncoderForTypeError(String, String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
cannotFindPartitionColumnInPartitionSchemaError(StructField, StructType) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
cannotFindProtobufFieldInCatalystError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
cannotGenerateCodeForExpressionError(Expression) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
cannotGenerateCodeForIncomparableTypeError(String, DataType) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
cannotGetEventTimeWatermarkError() - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
cannotGetJdbcTypeError(DataType) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
cannotGetOuterPointerForInnerClassError(Class<?>) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
cannotGetSQLConfInSchedulerEventLoopThreadError() - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
cannotHaveCircularReferencesInBeanClassError(Class<?>) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
cannotHaveCircularReferencesInClassError(String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
cannotInstantiateAbstractCatalogPluginClassError(String, String, Exception) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
cannotInterpolateClassIntoCodeBlockError(Object) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
cannotLoadClassNotOnClassPathError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
cannotLoadClassWhenRegisteringFunctionError(String, FunctionIdentifier) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
cannotLoadUserDefinedTypeError(String, String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
cannotMergeClassWithOtherClassError(String, String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
cannotMergeDecimalTypesWithIncompatibleScaleError(int, int) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
cannotMergeIncompatibleDataTypesError(DataType, DataType) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
cannotModifyValueOfSparkConfigError(String, String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
cannotModifyValueOfStaticConfigError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
cannotMutateReadOnlySQLConfError() - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
cannotOperateOnHiveDataSourceFilesError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
cannotOverwritePathBeingReadFromError() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
cannotOverwriteTableThatIsBeingReadFromError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
cannotParseDecimalError() - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
cannotParseIntervalError(String, Throwable) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
cannotParseJsonArraysAsStructsError() - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
cannotParseJSONFieldError(JsonParser, JsonToken, DataType) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
cannotParseJSONFieldError(String, String, JsonToken, DataType) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
cannotParseStatisticAsPercentileError(String, NumberFormatException) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
cannotParseStringAsDataTypeError(JsonParser, JsonToken, DataType) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
cannotParseStringAsDataTypeError(String, String, DataType) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
cannotParseValueTypeError(String, String, SqlBaseParser.TypeConstructorContext) - Static method in class org.apache.spark.sql.errors.QueryParsingErrors
 
cannotPassTypedColumnInUntypedSelectError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
cannotPurgeAsBreakInternalStateError() - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
cannotReadCorruptedTablePropertyError(String, String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
cannotReadFilesError(Throwable, String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
cannotReadFooterForFileError(Path, Exception) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
cannotRecognizeHiveTypeError(ParseException, String, String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
cannotRefreshBuiltInFuncError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
cannotRefreshTempFuncError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
cannotRemovePartitionDirError(Path) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
cannotRemoveReservedPropertyError(String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
cannotRenameTableWithAlterViewError() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
cannotRenameTempViewToExistingTableError(TableIdentifier) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
cannotRenameTempViewWithDatabaseSpecifiedError(TableIdentifier, TableIdentifier) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
cannotReplaceMissingTableError(Identifier) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
cannotReplaceMissingTableError(Identifier, Option<Throwable>) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
cannotResolveAttributeError(String, String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
cannotResolveColumnGivenInputColumnsError(String, String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
cannotResolveColumnNameAmongAttributesError(String, String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
cannotResolveStarExpandGivenInputColumnsError(String, String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
cannotResolveWindowReferenceError(String, SqlBaseParser.WindowClauseContext) - Static method in class org.apache.spark.sql.errors.QueryParsingErrors
 
cannotRestorePermissionsForPathError(FsPermission, Path) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
cannotRetrieveTableOrViewNotInSameDatabaseError(Seq<QualifiedTableName>) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
cannotRunSubmitMapStageOnZeroPartitionRDDError() - Static method in class org.apache.spark.errors.SparkCoreErrors
 
cannotSafelyMergeSerdePropertiesError(Map<String, String>, Map<String, String>, Set<String>) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
cannotSaveBlockOnDecommissionedExecutorError(BlockId) - Static method in class org.apache.spark.errors.SparkCoreErrors
 
cannotSaveIntervalIntoExternalStorageError() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
cannotSetJDBCNamespaceWithPropertyError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
cannotSetTimeoutDurationError() - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
cannotSetTimeoutTimestampError() - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
cannotSpecifyBothJdbcTableNameAndQueryError(String, String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
cannotSpecifyDatabaseForTempViewError(TableIdentifier) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
cannotSpecifyWindowFrameError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
cannotTerminateGeneratorError(UnresolvedGenerator) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
cannotTranslateExpressionToSourceFilterError(Expression) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
cannotTranslateNonNullValueForFieldError(int) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
cannotUnsetJDBCNamespaceWithPropertyError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
cannotUseAllColumnsForPartitionColumnsError() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
cannotUseDataTypeForPartitionColumnError(StructField) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
cannotUseIntervalTypeInTableSchemaError() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
cannotUseInvalidJavaIdentifierAsFieldNameError(String, WalkedTypePath) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
cannotUseMapSideCombiningWithArrayKeyError() - Static method in class org.apache.spark.errors.SparkCoreErrors
 
cannotUsePreservedDatabaseAsCurrentDatabaseError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
cannotWriteDataToRelationsWithMultiplePathsError() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
cannotWriteIncompatibleDataToTableError(String, Seq<String>) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
cannotWriteNotEnoughColumnsToTableError(String, Seq<Attribute>, LogicalPlan) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
cannotWriteTooManyColumnsToTableError(String, Seq<Attribute>, LogicalPlan) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
canonicalName() - Method in interface org.apache.spark.sql.connector.catalog.functions.BoundFunction
Returns the canonical name of this function, used to determine if functions are equivalent.
canonicalName() - Method in class org.apache.spark.sql.connector.expressions.aggregate.UserDefinedAggregateFunc
 
canonicalName() - Method in class org.apache.spark.sql.connector.expressions.UserDefinedScalarFunc
 
CanonicalRandomVertexCut$() - Constructor for class org.apache.spark.graphx.PartitionStrategy.CanonicalRandomVertexCut$
 
canOnlyZipRDDsWithSamePartitionSizeError() - Static method in class org.apache.spark.errors.SparkCoreErrors
 
canOverwrite(Filter[]) - Method in interface org.apache.spark.sql.connector.write.SupportsOverwrite
Checks whether it is possible to overwrite data from a data source table that matches filter expressions.
canOverwrite(Predicate[]) - Method in interface org.apache.spark.sql.connector.write.SupportsOverwrite
 
canOverwrite(Predicate[]) - Method in interface org.apache.spark.sql.connector.write.SupportsOverwriteV2
Checks whether it is possible to overwrite data from a data source table that matches filter expressions.
canWrite(DataType, DataType, boolean, Function2<String, String, Object>, String, Enumeration.Value, Function1<String, BoxedUnit>) - Static method in class org.apache.spark.sql.types.DataType
Returns true if the write data type can be read using the read data type.
capabilities() - Method in class org.apache.spark.sql.connector.catalog.DelegatingCatalogExtension
 
capabilities() - Method in interface org.apache.spark.sql.connector.catalog.Table
Returns the set of capabilities for this table.
capabilities() - Method in interface org.apache.spark.sql.connector.catalog.TableCatalog
 
cardinality() - Method in class org.apache.spark.util.sketch.BloomFilter
 
cartesian(JavaRDDLike<U, ?>) - Method in interface org.apache.spark.api.java.JavaRDDLike
Return the Cartesian product of this RDD and another one, that is, the RDD of all pairs of elements (a, b) where a is in this and b is in other.
cartesian(RDD<U>, ClassTag<U>) - Method in class org.apache.spark.rdd.RDD
Return the Cartesian product of this RDD and another one, that is, the RDD of all pairs of elements (a, b) where a is in this and b is in other.
CaseInsensitiveStringMap - Class in org.apache.spark.sql.util
Case-insensitive map of string keys to string values.
CaseInsensitiveStringMap(Map<String, String>) - Constructor for class org.apache.spark.sql.util.CaseInsensitiveStringMap
 
caseSensitive() - Method in class org.apache.spark.ml.feature.StopWordsRemover
Whether to do a case sensitive comparison over the stop words.
cast(DataType) - Method in class org.apache.spark.sql.Column
Casts the column to a different data type.
cast(String) - Method in class org.apache.spark.sql.Column
Casts the column to a different data type, using the canonical string representation of the type.
Cast - Class in org.apache.spark.sql.connector.expressions
Represents a cast expression in the public logical expression API.
Cast(Expression, DataType) - Constructor for class org.apache.spark.sql.connector.expressions.Cast
 
castingCauseOverflowError(Object, DataType, DataType) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
castingCauseOverflowErrorInTableInsert(DataType, DataType, String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
castPartitionSpec(String, DataType, SQLConf) - Static method in class org.apache.spark.sql.util.PartitioningUtils
 
Catalog - Class in org.apache.spark.sql.catalog
Catalog interface for Spark.
Catalog() - Constructor for class org.apache.spark.sql.catalog.Catalog
 
catalog() - Method in class org.apache.spark.sql.catalog.Database
 
catalog() - Method in class org.apache.spark.sql.catalog.Function
 
catalog() - Method in class org.apache.spark.sql.catalog.Table
 
catalog() - Method in class org.apache.spark.sql.SparkSession
 
CatalogAndIdentifier() - Method in interface org.apache.spark.sql.connector.catalog.LookupCatalog
 
CatalogAndIdentifier() - Constructor for class org.apache.spark.sql.connector.catalog.LookupCatalog.CatalogAndIdentifier
 
CatalogAndIdentifier$() - Constructor for class org.apache.spark.sql.connector.catalog.LookupCatalog.CatalogAndIdentifier$
 
CatalogAndMultipartIdentifier() - Method in interface org.apache.spark.sql.connector.catalog.LookupCatalog
 
CatalogAndNamespace() - Method in interface org.apache.spark.sql.connector.catalog.LookupCatalog
 
CatalogAndNamespace() - Constructor for class org.apache.spark.sql.connector.catalog.LookupCatalog.CatalogAndNamespace
 
CatalogAndNamespace$() - Constructor for class org.apache.spark.sql.connector.catalog.LookupCatalog.CatalogAndNamespace$
 
CatalogExtension - Interface in org.apache.spark.sql.connector.catalog
An API to extend the Spark built-in session catalog.
catalogFailToCallPublicNoArgConstructorError(String, String, Exception) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
catalogFailToFindPublicNoArgConstructorError(String, String, Exception) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
CatalogHelper(CatalogPlugin) - Constructor for class org.apache.spark.sql.connector.catalog.CatalogV2Implicits.CatalogHelper
 
catalogManager() - Method in interface org.apache.spark.sql.connector.catalog.LookupCatalog
 
CatalogMetadata - Class in org.apache.spark.sql.catalog
A catalog in Spark, as returned by the listCatalogs method defined in Catalog.
CatalogMetadata(String, String) - Constructor for class org.apache.spark.sql.catalog.CatalogMetadata
 
CatalogNotFoundException - Exception in org.apache.spark.sql.connector.catalog
 
CatalogNotFoundException(String, Throwable) - Constructor for exception org.apache.spark.sql.connector.catalog.CatalogNotFoundException
 
CatalogNotFoundException(String) - Constructor for exception org.apache.spark.sql.connector.catalog.CatalogNotFoundException
 
catalogOperationNotSupported(CatalogPlugin, String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
CatalogPlugin - Interface in org.apache.spark.sql.connector.catalog
A marker interface to provide a catalog implementation for Spark.
catalogPluginClassNotFoundError(String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
catalogPluginClassNotFoundForCatalogError(String, String, Exception) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
catalogPluginClassNotImplementedError(String, String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
Catalogs - Class in org.apache.spark.sql.connector.catalog
 
Catalogs() - Constructor for class org.apache.spark.sql.connector.catalog.Catalogs
 
catalogString() - Method in class org.apache.spark.sql.types.ArrayType
 
catalogString() - Static method in class org.apache.spark.sql.types.BinaryType
 
catalogString() - Static method in class org.apache.spark.sql.types.BooleanType
 
catalogString() - Static method in class org.apache.spark.sql.types.ByteType
 
catalogString() - Static method in class org.apache.spark.sql.types.CalendarIntervalType
 
catalogString() - Method in class org.apache.spark.sql.types.DataType
String representation for the type saved in external catalogs.
catalogString() - Static method in class org.apache.spark.sql.types.DateType
 
catalogString() - Static method in class org.apache.spark.sql.types.DoubleType
 
catalogString() - Static method in class org.apache.spark.sql.types.FloatType
 
catalogString() - Static method in class org.apache.spark.sql.types.IntegerType
 
catalogString() - Static method in class org.apache.spark.sql.types.LongType
 
catalogString() - Method in class org.apache.spark.sql.types.MapType
 
catalogString() - Static method in class org.apache.spark.sql.types.NullType
 
catalogString() - Static method in class org.apache.spark.sql.types.ShortType
 
catalogString() - Static method in class org.apache.spark.sql.types.StringType
 
catalogString() - Method in class org.apache.spark.sql.types.StructType
 
catalogString() - Static method in class org.apache.spark.sql.types.TimestampNTZType
 
catalogString() - Static method in class org.apache.spark.sql.types.TimestampType
 
catalogString() - Method in class org.apache.spark.sql.types.UserDefinedType
 
CatalogV2Implicits - Class in org.apache.spark.sql.connector.catalog
Conversion helpers for working with v2 CatalogPlugin.
CatalogV2Implicits() - Constructor for class org.apache.spark.sql.connector.catalog.CatalogV2Implicits
 
CatalogV2Implicits.BucketSpecHelper - Class in org.apache.spark.sql.connector.catalog
 
CatalogV2Implicits.CatalogHelper - Class in org.apache.spark.sql.connector.catalog
 
CatalogV2Implicits.ColumnsHelper - Class in org.apache.spark.sql.connector.catalog
 
CatalogV2Implicits.FunctionIdentifierHelper - Class in org.apache.spark.sql.connector.catalog
 
CatalogV2Implicits.IdentifierHelper - Class in org.apache.spark.sql.connector.catalog
 
CatalogV2Implicits.MultipartIdentifierHelper - Class in org.apache.spark.sql.connector.catalog
 
CatalogV2Implicits.NamespaceHelper - Class in org.apache.spark.sql.connector.catalog
 
CatalogV2Implicits.PartitionTypeHelper - Class in org.apache.spark.sql.connector.catalog
 
CatalogV2Implicits.TableIdentifierHelper - Class in org.apache.spark.sql.connector.catalog
 
CatalogV2Implicits.TransformHelper - Class in org.apache.spark.sql.connector.catalog
 
CatalogV2Util - Class in org.apache.spark.sql.connector.catalog
 
CatalogV2Util() - Constructor for class org.apache.spark.sql.connector.catalog.CatalogV2Util
 
CatalystScan - Interface in org.apache.spark.sql.sources
::Experimental:: An interface for experimenting with a more direct connection to the query planner.
Categorical() - Static method in class org.apache.spark.mllib.tree.configuration.FeatureType
 
categoricalCols() - Method in class org.apache.spark.ml.feature.FeatureHasher
Numeric columns to treat as categorical features.
categoricalFeaturesInfo() - Method in class org.apache.spark.mllib.tree.configuration.Strategy
 
CategoricalSplit - Class in org.apache.spark.ml.tree
Split which tests a categorical feature.
categories() - Method in class org.apache.spark.mllib.tree.model.DecisionTreeModel.SaveLoadV1_0$.SplitData
 
categories() - Method in class org.apache.spark.mllib.tree.model.Split
 
categoryMaps() - Method in class org.apache.spark.ml.feature.VectorIndexerModel
 
categorySizes() - Method in class org.apache.spark.ml.feature.OneHotEncoderModel
 
cause() - Method in exception org.apache.spark.sql.AnalysisException
 
cause() - Method in exception org.apache.spark.sql.streaming.StreamingQueryException
 
CausedBy - Class in org.apache.spark.util
Extractor Object for pulling out the root cause of an error.
CausedBy() - Constructor for class org.apache.spark.util.CausedBy
 
cbrt(Column) - Static method in class org.apache.spark.sql.functions
Computes the cube-root of the given value.
cbrt(String) - Static method in class org.apache.spark.sql.functions
Computes the cube-root of the given column.
ceil(Column, Column) - Static method in class org.apache.spark.sql.functions
Computes the ceiling of the given value of e to scale decimal places.
ceil(Column) - Static method in class org.apache.spark.sql.functions
Computes the ceiling of the given value of e to 0 decimal places.
ceil(String) - Static method in class org.apache.spark.sql.functions
Computes the ceiling of the given value of e to 0 decimal places.
ceil() - Method in class org.apache.spark.sql.types.Decimal
 
censorCol() - Method in class org.apache.spark.ml.regression.AFTSurvivalRegression
 
censorCol() - Method in class org.apache.spark.ml.regression.AFTSurvivalRegressionModel
 
censorCol() - Method in interface org.apache.spark.ml.regression.AFTSurvivalRegressionParams
Param for censor column name.
centerMatrix() - Method in class org.apache.spark.ml.clustering.KMeansAggregator
 
chainl1(Function0<Parsers.Parser<T>>, Function0<Parsers.Parser<Function2<T, T, T>>>) - Static method in class org.apache.spark.ml.feature.RFormulaParser
 
chainl1(Function0<Parsers.Parser<T>>, Function0<Parsers.Parser<U>>, Function0<Parsers.Parser<Function2<T, U, T>>>) - Static method in class org.apache.spark.ml.feature.RFormulaParser
 
chainr1(Function0<Parsers.Parser<T>>, Function0<Parsers.Parser<Function2<T, U, U>>>, Function2<T, U, U>, U) - Static method in class org.apache.spark.ml.feature.RFormulaParser
 
changePrecision(int, int) - Method in class org.apache.spark.sql.types.Decimal
Update precision and scale while keeping our value the same, and return true if successful.
channel() - Method in interface org.apache.spark.shuffle.api.WritableByteChannelWrapper
The underlying channel to write bytes into.
channelRead0(ChannelHandlerContext, byte[]) - Method in class org.apache.spark.api.r.RBackendAuthHandler
 
charOrVarcharTypeAsStringUnsupportedError() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
CharType - Class in org.apache.spark.sql.types
 
CharType(int) - Constructor for class org.apache.spark.sql.types.CharType
 
charTypeMissingLengthError(String, SqlBaseParser.PrimitiveDataTypeContext) - Static method in class org.apache.spark.sql.errors.QueryParsingErrors
 
checkAndGetK8sMasterUrl(String) - Static method in class org.apache.spark.util.Utils
Check the validity of the given Kubernetes master URL and return the resolved URL.
checkColumnNameDuplication(Seq<String>, Function2<String, String, Object>) - Static method in class org.apache.spark.sql.util.SchemaUtils
Checks if input column names have duplicate identifiers.
checkColumnNameDuplication(Seq<String>, boolean) - Static method in class org.apache.spark.sql.util.SchemaUtils
Checks if input column names have duplicate identifiers.
checkColumnType(StructType, String, DataType, String) - Static method in class org.apache.spark.ml.util.SchemaUtils
Check whether the given schema contains a column of the required data type.
checkColumnTypes(StructType, String, Seq<DataType>, String) - Static method in class org.apache.spark.ml.util.SchemaUtils
Check whether the given schema contains a column of one of the require data types.
checkDataColumns(RFormula, Dataset<?>) - Static method in class org.apache.spark.ml.r.RWrapperUtils
DataFrame column check.
checkFileExists(String, Configuration) - Static method in class org.apache.spark.streaming.util.HdfsUtils
Check if the file exists at the given path.
checkHost(String) - Static method in class org.apache.spark.util.Utils
Checks if the host contains only valid hostname/ip without port NOTE: Incase of IPV6 ip it should be enclosed inside []
checkHostPort(String) - Static method in class org.apache.spark.util.Utils
 
checkIntegers(Dataset<?>, String) - Method in interface org.apache.spark.ml.recommendation.ALSModelParams
Attempts to safely cast a user/item id to an Int.
checkNumericType(StructType, String, String) - Static method in class org.apache.spark.ml.util.SchemaUtils
Check whether the given schema contains a column of the numeric data type.
checkOffHeapEnabled(SparkConf, long) - Static method in class org.apache.spark.util.Utils
return 0 if MEMORY_OFFHEAP_ENABLED is false.
checkpoint() - Method in interface org.apache.spark.api.java.JavaRDDLike
Mark this RDD for checkpointing.
checkpoint() - Method in class org.apache.spark.graphx.Graph
Mark this Graph for checkpointing.
checkpoint() - Method in class org.apache.spark.graphx.impl.EdgeRDDImpl
 
checkpoint() - Method in class org.apache.spark.graphx.impl.GraphImpl
 
checkpoint() - Method in class org.apache.spark.graphx.impl.VertexRDDImpl
 
checkpoint() - Method in class org.apache.spark.rdd.HadoopRDD
 
checkpoint() - Method in class org.apache.spark.rdd.RDD
Mark this RDD for checkpointing.
checkpoint() - Method in class org.apache.spark.sql.Dataset
Eagerly checkpoint a Dataset and return the new Dataset.
checkpoint(boolean) - Method in class org.apache.spark.sql.Dataset
Returns a checkpointed version of this Dataset.
checkpoint(Duration) - Method in interface org.apache.spark.streaming.api.java.JavaDStreamLike
Enable periodic checkpointing of RDDs of this DStream.
checkpoint(String) - Method in class org.apache.spark.streaming.api.java.JavaStreamingContext
Deprecated.
Sets the context to periodically checkpoint the DStream operations for master fault-tolerance.
checkpoint(Duration) - Method in class org.apache.spark.streaming.dstream.DStream
Enable periodic checkpointing of RDDs of this DStream
checkpoint(String) - Method in class org.apache.spark.streaming.StreamingContext
Deprecated.
Set the context to periodically checkpoint the DStream operations for driver fault-tolerance.
checkpointCleaned(long) - Method in interface org.apache.spark.CleanerListener
 
checkpointDirectoryHasNotBeenSetInSparkContextError() - Static method in class org.apache.spark.errors.SparkCoreErrors
 
Checkpointed() - Static method in class org.apache.spark.rdd.CheckpointState
 
checkpointFailedToSaveError(int, Path) - Static method in class org.apache.spark.errors.SparkCoreErrors
 
CheckpointingInProgress() - Static method in class org.apache.spark.rdd.CheckpointState
 
checkpointInterval() - Method in class org.apache.spark.ml.classification.DecisionTreeClassificationModel
 
checkpointInterval() - Method in class org.apache.spark.ml.classification.DecisionTreeClassifier
 
checkpointInterval() - Method in class org.apache.spark.ml.classification.GBTClassificationModel
 
checkpointInterval() - Method in class org.apache.spark.ml.classification.GBTClassifier
 
checkpointInterval() - Method in class org.apache.spark.ml.classification.RandomForestClassificationModel
 
checkpointInterval() - Method in class org.apache.spark.ml.classification.RandomForestClassifier
 
checkpointInterval() - Method in class org.apache.spark.ml.clustering.LDA
 
checkpointInterval() - Method in class org.apache.spark.ml.clustering.LDAModel
 
checkpointInterval() - Method in interface org.apache.spark.ml.param.shared.HasCheckpointInterval
Param for set checkpoint interval (&gt;= 1) or disable checkpoint (-1).
checkpointInterval() - Method in class org.apache.spark.ml.recommendation.ALS
 
checkpointInterval() - Method in class org.apache.spark.ml.regression.DecisionTreeRegressionModel
 
checkpointInterval() - Method in class org.apache.spark.ml.regression.DecisionTreeRegressor
 
checkpointInterval() - Method in class org.apache.spark.ml.regression.GBTRegressionModel
 
checkpointInterval() - Method in class org.apache.spark.ml.regression.GBTRegressor
 
checkpointInterval() - Method in class org.apache.spark.ml.regression.RandomForestRegressionModel
 
checkpointInterval() - Method in class org.apache.spark.ml.regression.RandomForestRegressor
 
checkpointInterval() - Method in class org.apache.spark.mllib.tree.configuration.Strategy
 
checkpointLocationNotSpecifiedError() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
checkpointRDDBlockIdNotFoundError(RDDBlockId) - Static method in class org.apache.spark.errors.SparkCoreErrors
 
checkpointRDDHasDifferentNumberOfPartitionsFromOriginalRDDError(int, int, int, int) - Static method in class org.apache.spark.errors.SparkCoreErrors
 
CheckpointReader - Class in org.apache.spark.streaming
 
CheckpointReader() - Constructor for class org.apache.spark.streaming.CheckpointReader
 
CheckpointState - Class in org.apache.spark.rdd
Enumeration to manage state transitions of an RDD through checkpointing
CheckpointState() - Constructor for class org.apache.spark.rdd.CheckpointState
 
checkSchemaColumnNameDuplication(DataType, boolean) - Static method in class org.apache.spark.sql.util.SchemaUtils
Checks if an input schema has duplicate column names.
checkSchemaColumnNameDuplication(StructType, Function2<String, String, Object>) - Static method in class org.apache.spark.sql.util.SchemaUtils
Checks if an input schema has duplicate column names.
checkSingleVsMultiColumnParams(Params, Seq<Param<?>>, Seq<Param<?>>) - Static method in class org.apache.spark.ml.param.ParamValidators
Utility for Param validity checks for Transformers which have both single- and multi-column support.
checkSpeculatableTasks(long) - Method in interface org.apache.spark.scheduler.Schedulable
 
checkState(boolean, Function0<String>) - Static method in class org.apache.spark.streaming.util.HdfsUtils
 
checkThresholdConsistency() - Method in interface org.apache.spark.ml.classification.LogisticRegressionParams
If threshold and thresholds are both set, ensures they are consistent.
checkTransformDuplication(Seq<Transform>, String, boolean) - Static method in class org.apache.spark.sql.util.SchemaUtils
Checks if the partitioning transforms are being duplicated or not.
checkUIViewPermissions() - Method in interface org.apache.spark.status.api.v1.BaseAppResource
 
checkUIViewPermissions(String, Option<String>, String) - Method in interface org.apache.spark.status.api.v1.UIRoot
 
child() - Method in class org.apache.spark.sql.connector.expressions.filter.Not
 
child() - Method in class org.apache.spark.sql.sources.Not
 
CHILD_CLUSTERS_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper
 
CHILD_CONNECTION_TIMEOUT - Static variable in class org.apache.spark.launcher.SparkLauncher
Maximum time (in ms) to wait for a child process to connect back to the launcher server when using @link{#start()}.
CHILD_NODES_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper
 
CHILD_PROCESS_LOGGER_NAME - Static variable in class org.apache.spark.launcher.SparkLauncher
Logger name to use when launching a child process.
ChildFirstURLClassLoader - Class in org.apache.spark.util
A mutable class loader that gives preference to its own URLs over the parent class loader when loading classes and resources.
ChildFirstURLClassLoader(URL[], ClassLoader) - Constructor for class org.apache.spark.util.ChildFirstURLClassLoader
 
children() - Method in class org.apache.spark.sql.connector.expressions.aggregate.Avg
 
children() - Method in class org.apache.spark.sql.connector.expressions.aggregate.Count
 
children() - Method in class org.apache.spark.sql.connector.expressions.aggregate.CountStar
 
children() - Method in class org.apache.spark.sql.connector.expressions.aggregate.GeneralAggregateFunc
 
children() - Method in class org.apache.spark.sql.connector.expressions.aggregate.Max
 
children() - Method in class org.apache.spark.sql.connector.expressions.aggregate.Min
 
children() - Method in class org.apache.spark.sql.connector.expressions.aggregate.Sum
 
children() - Method in class org.apache.spark.sql.connector.expressions.aggregate.UserDefinedAggregateFunc
 
children() - Method in class org.apache.spark.sql.connector.expressions.Cast
 
children() - Method in interface org.apache.spark.sql.connector.expressions.Expression
Returns an array of the children of this node.
children() - Method in class org.apache.spark.sql.connector.expressions.Extract
 
children() - Method in class org.apache.spark.sql.connector.expressions.GeneralScalarExpression
 
children() - Method in interface org.apache.spark.sql.connector.expressions.Literal
 
children() - Method in interface org.apache.spark.sql.connector.expressions.NamedReference
 
children() - Method in interface org.apache.spark.sql.connector.expressions.SortOrder
 
children() - Method in interface org.apache.spark.sql.connector.expressions.Transform
 
children() - Method in class org.apache.spark.sql.connector.expressions.UserDefinedScalarFunc
 
chiSqFunc() - Method in class org.apache.spark.mllib.stat.test.ChiSqTest.Method
 
ChiSqSelector - Class in org.apache.spark.ml.feature
Deprecated.
use UnivariateFeatureSelector instead. Since 3.1.1.
ChiSqSelector(String) - Constructor for class org.apache.spark.ml.feature.ChiSqSelector
Deprecated.
 
ChiSqSelector() - Constructor for class org.apache.spark.ml.feature.ChiSqSelector
Deprecated.
 
ChiSqSelector - Class in org.apache.spark.mllib.feature
Creates a ChiSquared feature selector.
ChiSqSelector() - Constructor for class org.apache.spark.mllib.feature.ChiSqSelector
 
ChiSqSelector(int) - Constructor for class org.apache.spark.mllib.feature.ChiSqSelector
The is the same to call this() and setNumTopFeatures(numTopFeatures)
ChiSqSelectorModel - Class in org.apache.spark.ml.feature
Model fitted by ChiSqSelector.
ChiSqSelectorModel - Class in org.apache.spark.mllib.feature
Chi Squared selector model.
ChiSqSelectorModel(int[]) - Constructor for class org.apache.spark.mllib.feature.ChiSqSelectorModel
 
ChiSqSelectorModel.ChiSqSelectorModelWriter - Class in org.apache.spark.ml.feature
 
ChiSqSelectorModel.SaveLoadV1_0$ - Class in org.apache.spark.mllib.feature
 
ChiSqSelectorModel.SaveLoadV1_0$.Data - Class in org.apache.spark.mllib.feature
Model data for import/export
ChiSqSelectorModel.SaveLoadV1_0$.Data$ - Class in org.apache.spark.mllib.feature
 
ChiSqSelectorModelWriter(ChiSqSelectorModel) - Constructor for class org.apache.spark.ml.feature.ChiSqSelectorModel.ChiSqSelectorModelWriter
 
chiSqTest(Vector, Vector) - Static method in class org.apache.spark.mllib.stat.Statistics
Conduct Pearson's chi-squared goodness of fit test of the observed data against the expected distribution.
chiSqTest(Vector) - Static method in class org.apache.spark.mllib.stat.Statistics
Conduct Pearson's chi-squared goodness of fit test of the observed data against the uniform distribution, with each category having an expected frequency of 1 / observed.size.
chiSqTest(Matrix) - Static method in class org.apache.spark.mllib.stat.Statistics
Conduct Pearson's independence test on the input contingency matrix, which cannot contain negative entries or columns or rows that sum up to 0.
chiSqTest(RDD<LabeledPoint>) - Static method in class org.apache.spark.mllib.stat.Statistics
Conduct Pearson's independence test for every feature against the label across the input RDD.
chiSqTest(JavaRDD<LabeledPoint>) - Static method in class org.apache.spark.mllib.stat.Statistics
Java-friendly version of chiSqTest()
ChiSqTest - Class in org.apache.spark.mllib.stat.test
Conduct the chi-squared test for the input RDDs using the specified method.
ChiSqTest() - Constructor for class org.apache.spark.mllib.stat.test.ChiSqTest
 
ChiSqTest.Method - Class in org.apache.spark.mllib.stat.test
param: name String name for the method.
ChiSqTest.Method$ - Class in org.apache.spark.mllib.stat.test
 
ChiSqTest.NullHypothesis$ - Class in org.apache.spark.mllib.stat.test
 
ChiSqTestResult - Class in org.apache.spark.mllib.stat.test
Object containing the test results for the chi-squared hypothesis test.
chiSquared(Vector, Vector, String) - Static method in class org.apache.spark.mllib.stat.test.ChiSqTest
 
chiSquaredFeatures(RDD<LabeledPoint>, String) - Static method in class org.apache.spark.mllib.stat.test.ChiSqTest
Conduct Pearson's independence test for each feature against the label across the input RDD.
chiSquaredMatrix(Matrix, String) - Static method in class org.apache.spark.mllib.stat.test.ChiSqTest
 
ChiSquareTest - Class in org.apache.spark.ml.stat
Chi-square hypothesis testing for categorical data.
ChiSquareTest() - Constructor for class org.apache.spark.ml.stat.ChiSquareTest
 
chmod700(File) - Static method in class org.apache.spark.util.Utils
JDK equivalent of chmod 700 file.
CholeskyDecomposition - Class in org.apache.spark.mllib.linalg
Compute Cholesky decomposition.
CholeskyDecomposition() - Constructor for class org.apache.spark.mllib.linalg.CholeskyDecomposition
 
chunkId() - Method in class org.apache.spark.storage.ShuffleBlockChunkId
 
cipherStream() - Method in interface org.apache.spark.security.CryptoStreamUtils.BaseErrorHandler
The encrypted stream that may get into an unhealthy state.
classDoesNotImplementUserDefinedAggregateFunctionError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
classForName(String, boolean, boolean) - Static method in class org.apache.spark.util.Utils
Preferred alternative to Class.forName(className), as well as Class.forName(className, initialize, loader) with current thread's ContextClassLoader.
classHasUnexpectedSerializerError(String, Expression) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
Classification() - Static method in class org.apache.spark.mllib.tree.configuration.Algo
 
ClassificationLoss - Interface in org.apache.spark.mllib.tree.loss
 
ClassificationModel<FeaturesType,M extends ClassificationModel<FeaturesType,M>> - Class in org.apache.spark.ml.classification
Model produced by a Classifier.
ClassificationModel() - Constructor for class org.apache.spark.ml.classification.ClassificationModel
 
ClassificationModel - Interface in org.apache.spark.mllib.classification
Represents a classification model that predicts to which of a set of categories an example belongs.
ClassificationSummary - Interface in org.apache.spark.ml.classification
Abstraction for multiclass classification results for a given model.
Classifier<FeaturesType,E extends Classifier<FeaturesType,E,M>,M extends ClassificationModel<FeaturesType,M>> - Class in org.apache.spark.ml.classification
Single-label binary or multiclass classification.
Classifier() - Constructor for class org.apache.spark.ml.classification.Classifier
 
classifier() - Method in class org.apache.spark.ml.classification.OneVsRest
 
classifier() - Method in class org.apache.spark.ml.classification.OneVsRestModel
 
classifier() - Method in interface org.apache.spark.ml.classification.OneVsRestParams
param for the base binary classifier that we reduce multiclass classification into.
ClassifierParams - Interface in org.apache.spark.ml.classification
(private[spark]) Params for classification.
ClassifierTypeTrait - Interface in org.apache.spark.ml.classification
 
classifyException(String, Throwable) - Static method in class org.apache.spark.sql.jdbc.DB2Dialect
 
classifyException(String, Throwable) - Static method in class org.apache.spark.sql.jdbc.DerbyDialect
 
classifyException(String, Throwable) - Static method in class org.apache.spark.sql.jdbc.H2Dialect
 
classifyException(String, Throwable) - Method in class org.apache.spark.sql.jdbc.JdbcDialect
Gets a dialect exception, classifies it and wraps it by AnalysisException.
classifyException(String, Throwable) - Static method in class org.apache.spark.sql.jdbc.MsSqlServerDialect
 
classifyException(String, Throwable) - Static method in class org.apache.spark.sql.jdbc.MySQLDialect
 
classifyException(String, Throwable) - Static method in class org.apache.spark.sql.jdbc.NoopDialect
 
classifyException(String, Throwable) - Static method in class org.apache.spark.sql.jdbc.OracleDialect
 
classifyException(String, Throwable) - Static method in class org.apache.spark.sql.jdbc.PostgresDialect
 
classifyException(String, Throwable) - Static method in class org.apache.spark.sql.jdbc.TeradataDialect
 
classIsLoadable(String) - Static method in class org.apache.spark.util.Utils
Determines whether the provided class is loadable in the current thread.
className() - Method in class org.apache.spark.ExceptionFailure
 
className() - Static method in class org.apache.spark.ml.linalg.JsonMatrixConverter
Unique class name for identifying JSON object encoded by this class.
className() - Method in class org.apache.spark.sql.catalog.Function
 
CLASSPATH_ENTRIES_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
 
classpathEntries() - Method in class org.apache.spark.status.api.v1.ApplicationEnvironmentInfo
 
classTag() - Method in class org.apache.spark.api.java.JavaDoubleRDD
 
classTag() - Method in class org.apache.spark.api.java.JavaPairRDD
 
classTag() - Method in class org.apache.spark.api.java.JavaRDD
 
classTag() - Method in interface org.apache.spark.api.java.JavaRDDLike
 
classTag() - Method in class org.apache.spark.sql.Dataset
 
classTag() - Method in class org.apache.spark.storage.memory.DeserializedMemoryEntry
 
classTag() - Method in interface org.apache.spark.storage.memory.MemoryEntry
 
classTag() - Method in class org.apache.spark.storage.memory.SerializedMemoryEntry
 
classTag() - Method in class org.apache.spark.streaming.api.java.JavaDStream
 
classTag() - Method in interface org.apache.spark.streaming.api.java.JavaDStreamLike
 
classTag() - Method in class org.apache.spark.streaming.api.java.JavaInputDStream
 
classTag() - Method in class org.apache.spark.streaming.api.java.JavaPairDStream
 
classTag() - Method in class org.apache.spark.streaming.api.java.JavaReceiverInputDStream
 
classUnsupportedByMapObjectsError(Class<?>) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
classWithoutPublicNonArgumentConstructorError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
clean(long, boolean) - Method in class org.apache.spark.streaming.util.WriteAheadLog
Clean all the records that are older than the threshold time.
clean(Object, boolean, boolean) - Static method in class org.apache.spark.util.ClosureCleaner
Clean the given closure in place.
CleanAccum - Class in org.apache.spark
 
CleanAccum(long) - Constructor for class org.apache.spark.CleanAccum
 
CleanBroadcast - Class in org.apache.spark
 
CleanBroadcast(long) - Constructor for class org.apache.spark.CleanBroadcast
 
CleanCheckpoint - Class in org.apache.spark
 
CleanCheckpoint(int) - Constructor for class org.apache.spark.CleanCheckpoint
 
CleanerListener - Interface in org.apache.spark
Listener class used when any item has been cleaned by the Cleaner class.
cleaning() - Method in class org.apache.spark.status.LiveStage
 
CleanRDD - Class in org.apache.spark
 
CleanRDD(int) - Constructor for class org.apache.spark.CleanRDD
 
CleanShuffle - Class in org.apache.spark
 
CleanShuffle(int) - Constructor for class org.apache.spark.CleanShuffle
 
cleanShuffleDependencies(boolean) - Method in class org.apache.spark.rdd.RDD
Removes an RDD's shuffles and it's non-persisted ancestors.
CleanSparkListener - Class in org.apache.spark
 
CleanSparkListener(SparkListener) - Constructor for class org.apache.spark.CleanSparkListener
 
cleanupApplication() - Method in interface org.apache.spark.shuffle.api.ShuffleDriverComponents
Called once at the end of the Spark application to clean up any existing shuffle state.
cleanupOldBlocks(long) - Method in interface org.apache.spark.streaming.receiver.ReceivedBlockHandler
Cleanup old blocks older than the given threshold time
cleanUpSourceFilesUnsupportedError() - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
CleanupTask - Interface in org.apache.spark
Classes that represent cleaning tasks.
CleanupTaskWeakReference - Class in org.apache.spark
A WeakReference associated with a CleanupTask.
CleanupTaskWeakReference(CleanupTask, Object, ReferenceQueue<Object>) - Constructor for class org.apache.spark.CleanupTaskWeakReference
 
clear(Param<?>) - Method in interface org.apache.spark.ml.param.Params
Clears the user-supplied value for the input param.
clear() - Method in class org.apache.spark.sql.util.CaseInsensitiveStringMap
 
clear() - Method in class org.apache.spark.sql.util.ExecutionListenerManager
Removes all the registered QueryExecutionListener.
clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo.Builder
 
clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo.Builder
 
clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
 
clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoWrapper.Builder
 
clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
 
clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoWrapper.Builder
 
clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.AppSummary.Builder
 
clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
 
clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetrics.Builder
 
clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
 
clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions.Builder
 
clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest.Builder
 
clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
 
clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper.Builder
 
clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
 
clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryWrapper.Builder
 
clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributions.Builder
 
clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetrics.Builder
 
clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
 
clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper.Builder
 
clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.MemoryMetrics.Builder
 
clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributions.Builder
 
clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetrics.Builder
 
clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.PairStrings.Builder
 
clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.PoolData.Builder
 
clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder
 
clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryWrapper.Builder
 
clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution.Builder
 
clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
 
clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationEdge.Builder
 
clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
 
clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode.Builder
 
clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo.Builder
 
clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
 
clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoWrapper.Builder
 
clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceInformation.Builder
 
clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo.Builder
 
clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileWrapper.Builder
 
clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.RuntimeInfo.Builder
 
clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
 
clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics.Builder
 
clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
 
clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics.Builder
 
clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions.Builder
 
clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetrics.Builder
 
clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress.Builder
 
clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
 
clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
 
clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphEdge.Builder
 
clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
 
clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper.Builder
 
clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
 
clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummary.Builder
 
clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryWrapper.Builder
 
clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
 
clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetric.Builder
 
clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
 
clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper.Builder
 
clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
 
clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
 
clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData.Builder
 
clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
 
clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressWrapper.Builder
 
clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
 
clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
 
clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
 
clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
 
clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskResourceRequest.Builder
 
clear() - Static method in class org.apache.spark.util.AccumulatorContext
Clears all registered AccumulatorV2s.
clearAccumulatorId() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetric.Builder
int64 accumulator_id = 2;
clearAccumulatorUpdates() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 44;
clearAccumulatorUpdates() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
clearAccumulatorUpdates() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
clearActive() - Static method in class org.apache.spark.sql.SQLContext
Deprecated.
Use SparkSession.clearActiveSession instead. Since 2.0.0.
clearActiveSession() - Static method in class org.apache.spark.sql.SparkSession
Clears the active SparkSession for current thread.
clearActiveTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
int32 active_tasks = 9;
clearAddress() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution.Builder
optional string address = 1;
clearAddresses() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceInformation.Builder
repeated string addresses = 2;
clearAddTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
int64 add_time = 20;
clearAddTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder
int64 add_time = 5;
clearAllRemovalsTimeMs() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
int64 all_removals_time_ms = 6;
clearAllUpdatesTimeMs() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
int64 all_updates_time_ms = 4;
clearAmount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest.Builder
int64 amount = 2;
clearAmount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskResourceRequest.Builder
double amount = 2;
clearAppSparkVersion() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo.Builder
optional string app_spark_version = 8;
clearAttempt() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
int32 attempt = 3;
clearAttempt() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int32 attempt = 3;
clearAttemptId() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo.Builder
optional string attempt_id = 1;
clearAttemptId() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int32 attempt_id = 3;
clearAttempts() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
repeated .org.apache.spark.status.protobuf.ApplicationAttemptInfo attempts = 7;
clearAttributes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
 
clearBarrier() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode.Builder
bool barrier = 4;
clearBatchDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
int64 batch_duration = 6;
clearBatchId() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
int64 batch_id = 5;
clearBlacklistedInStages() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
repeated int64 blacklisted_in_stages = 25;
clearBlockName() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo.Builder
optional string block_name = 1;
clearBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double bytes_read = 18;
clearBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributions.Builder
repeated double bytes_read = 1;
clearBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetrics.Builder
int64 bytes_read = 1;
clearBytesWritten() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double bytes_written = 20;
clearBytesWritten() - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributions.Builder
repeated double bytes_written = 1;
clearBytesWritten() - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetrics.Builder
int64 bytes_written = 1;
clearBytesWritten() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetrics.Builder
int64 bytes_written = 1;
clearCache() - Method in class org.apache.spark.sql.catalog.Catalog
Removes all cached tables from the in-memory cache.
clearCache() - Method in class org.apache.spark.sql.SQLContext
Removes all cached tables from the in-memory cache.
clearCached() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode.Builder
bool cached = 3;
clearCallSite() - Method in class org.apache.spark.api.java.JavaSparkContext
Pass-through to SparkContext.setCallSite.
clearCallSite() - Method in class org.apache.spark.SparkContext
Clear the thread-local property for overriding the call sites of actions and RDDs.
clearCallsite() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode.Builder
optional string callsite = 5;
clearChildClusters() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationClusterWrapper child_clusters = 4;
clearChildNodes() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationNode child_nodes = 3;
clearClasspathEntries() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings classpath_entries = 6;
clearCluster() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper.Builder
.org.apache.spark.status.protobuf.SparkPlanGraphClusterWrapper cluster = 2;
clearCommitTimeMs() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
int64 commit_time_ms = 7;
clearCompleted() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo.Builder
bool completed = 7;
clearCompletedTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
int32 completed_tasks = 11;
clearCompletionTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
optional int64 completion_time = 5;
clearCompletionTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
optional int64 completion_time = 9;
clearCompletionTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
optional int64 completion_time = 12;
clearCoresGranted() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
optional int32 cores_granted = 3;
clearCoresPerExecutor() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
optional int32 cores_per_executor = 5;
clearCorruptMergedBlockChunks() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
repeated double corrupt_merged_block_chunks = 1;
clearCorruptMergedBlockChunks() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics.Builder
int64 corrupt_merged_block_chunks = 1;
clearCurrentDefaultValue() - Method in class org.apache.spark.sql.types.StructField
Clears the StructField of its current default value, if any.
clearCustomMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
 
clearDataDistribution() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
repeated .org.apache.spark.status.protobuf.RDDDataDistribution data_distribution = 8;
clearDefaultSession() - Static method in class org.apache.spark.sql.SparkSession
Clears the default SparkSession that is returned by the builder.
clearDependencies() - Method in class org.apache.spark.rdd.CoGroupedRDD
 
clearDependencies() - Method in class org.apache.spark.rdd.ShuffledRDD
 
clearDependencies() - Method in class org.apache.spark.rdd.UnionRDD
 
clearDesc() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
optional string desc = 3;
clearDesc() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
optional string desc = 3;
clearDescription() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
optional string description = 3;
clearDescription() - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress.Builder
optional string description = 1;
clearDescription() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
optional string description = 1;
clearDescription() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
optional string description = 3;
clearDescription() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
optional string description = 40;
clearDeserialized() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
bool deserialized = 7;
clearDetails() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
optional string details = 4;
clearDetails() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
optional string details = 41;
clearDiscoveryScript() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest.Builder
optional string discoveryScript = 3;
clearDiskBytesSpilled() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double disk_bytes_spilled = 17;
clearDiskBytesSpilled() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double disk_bytes_spilled = 15;
clearDiskBytesSpilled() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
int64 disk_bytes_spilled = 14;
clearDiskBytesSpilled() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int64 disk_bytes_spilled = 22;
clearDiskBytesSpilled() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 disk_bytes_spilled = 24;
clearDiskBytesSpilled() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double disk_bytes_spilled = 14;
clearDiskBytesSpilled() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
int64 disk_bytes_spilled = 9;
clearDiskSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
int64 disk_size = 9;
clearDiskUsed() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
int64 disk_used = 6;
clearDiskUsed() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution.Builder
int64 disk_used = 4;
clearDiskUsed() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo.Builder
int64 disk_used = 4;
clearDiskUsed() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
int64 disk_used = 7;
clearDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo.Builder
int64 duration = 5;
clearDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double duration = 5;
clearDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
optional int64 duration = 7;
clearDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 duration = 7;
clearDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double duration = 2;
clearDurationMs() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
 
clearEdges() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationEdge edges = 2;
clearEdges() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.SparkPlanGraphEdge edges = 3;
clearEndOffset() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
optional string end_offset = 3;
clearEndTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo.Builder
int64 end_time = 3;
clearEndTimestamp() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData.Builder
optional int64 end_timestamp = 7;
clearErrorMessage() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
optional string error_message = 10;
clearErrorMessage() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
optional string error_message = 14;
clearErrorMessage() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
optional string error_message = 14;
clearEventTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
 
clearException() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData.Builder
optional string exception = 5;
clearExcludedInStages() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
repeated int64 excluded_in_stages = 31;
clearExecutionId() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
int64 execution_id = 1;
clearExecutionId() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
int64 execution_id = 1;
clearExecutorCpuTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double executor_cpu_time = 9;
clearExecutorCpuTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int64 executor_cpu_time = 17;
clearExecutorCpuTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 executor_cpu_time = 19;
clearExecutorCpuTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double executor_cpu_time = 6;
clearExecutorCpuTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
int64 executor_cpu_time = 4;
clearExecutorDeserializeCpuTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double executor_deserialize_cpu_time = 7;
clearExecutorDeserializeCpuTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int64 executor_deserialize_cpu_time = 15;
clearExecutorDeserializeCpuTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 executor_deserialize_cpu_time = 17;
clearExecutorDeserializeCpuTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double executor_deserialize_cpu_time = 4;
clearExecutorDeserializeCpuTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
int64 executor_deserialize_cpu_time = 2;
clearExecutorDeserializeTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double executor_deserialize_time = 6;
clearExecutorDeserializeTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int64 executor_deserialize_time = 14;
clearExecutorDeserializeTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 executor_deserialize_time = 16;
clearExecutorDeserializeTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double executor_deserialize_time = 3;
clearExecutorDeserializeTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
int64 executor_deserialize_time = 1;
clearExecutorId() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper.Builder
optional string executor_id = 3;
clearExecutorId() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
optional string executor_id = 2;
clearExecutorId() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
optional string executor_id = 8;
clearExecutorId() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
optional string executor_id = 8;
clearExecutorLogs() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
 
clearExecutorLogs() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
 
clearExecutorMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions.Builder
repeated .org.apache.spark.status.protobuf.ExecutorMetrics executor_metrics = 2;
clearExecutorMetricsDistributions() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
optional .org.apache.spark.status.protobuf.ExecutorMetricsDistributions executor_metrics_distributions = 52;
clearExecutorResourceRequests() - Method in class org.apache.spark.resource.ResourceProfileBuilder
 
clearExecutorResources() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo.Builder
 
clearExecutorRunTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double executor_run_time = 8;
clearExecutorRunTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int64 executor_run_time = 16;
clearExecutorRunTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 executor_run_time = 18;
clearExecutorRunTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double executor_run_time = 5;
clearExecutorRunTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
int64 executor_run_time = 3;
clearExecutors() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo.Builder
repeated string executors = 5;
clearExecutorSummary() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
 
clearFailedTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double failed_tasks = 3;
clearFailedTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
int32 failed_tasks = 2;
clearFailedTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
int32 failed_tasks = 10;
clearFailureReason() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
optional string failure_reason = 13;
clearFetchWaitTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
repeated double fetch_wait_time = 5;
clearFetchWaitTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics.Builder
int64 fetch_wait_time = 3;
clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo.Builder
 
clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo.Builder
 
clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
 
clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoWrapper.Builder
 
clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
 
clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoWrapper.Builder
 
clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.AppSummary.Builder
 
clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
 
clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetrics.Builder
 
clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
 
clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions.Builder
 
clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest.Builder
 
clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
 
clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper.Builder
 
clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
 
clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryWrapper.Builder
 
clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributions.Builder
 
clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetrics.Builder
 
clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
 
clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper.Builder
 
clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.MemoryMetrics.Builder
 
clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributions.Builder
 
clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetrics.Builder
 
clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.PairStrings.Builder
 
clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.PoolData.Builder
 
clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder
 
clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryWrapper.Builder
 
clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution.Builder
 
clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
 
clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationEdge.Builder
 
clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
 
clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode.Builder
 
clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo.Builder
 
clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
 
clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoWrapper.Builder
 
clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceInformation.Builder
 
clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo.Builder
 
clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileWrapper.Builder
 
clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.RuntimeInfo.Builder
 
clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
 
clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics.Builder
 
clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
 
clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics.Builder
 
clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions.Builder
 
clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetrics.Builder
 
clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress.Builder
 
clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
 
clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
 
clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphEdge.Builder
 
clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
 
clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper.Builder
 
clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
 
clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummary.Builder
 
clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryWrapper.Builder
 
clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
 
clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetric.Builder
 
clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
 
clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper.Builder
 
clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
 
clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
 
clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData.Builder
 
clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
 
clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressWrapper.Builder
 
clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
 
clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
 
clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
 
clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
 
clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskResourceRequest.Builder
 
clearFirstTaskLaunchedTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
optional int64 first_task_launched_time = 11;
clearFromId() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationEdge.Builder
int32 from_id = 1;
clearFromId() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphEdge.Builder
int64 from_id = 1;
clearFunctions() - Static method in class org.apache.spark.sql.jdbc.H2Dialect
 
clearGettingResultTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double getting_result_time = 13;
clearGettingResultTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
int64 getting_result_time = 18;
clearGettingResultTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double getting_result_time = 10;
clearHadoopProperties() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings hadoop_properties = 3;
clearHasMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
bool has_metrics = 15;
clearHost() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
optional string host = 9;
clearHost() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
optional string host = 9;
clearHostPort() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
optional string host_port = 2;
clearHostPort() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder
optional string host_port = 2;
clearHostPort() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
optional string host_port = 3;
clearId() - Method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo.Builder
int64 id = 1;
clearId() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
optional string id = 1;
clearId() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
optional string id = 1;
clearId() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder
optional string id = 1;
clearId() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
optional string id = 1;
clearId() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode.Builder
int32 id = 1;
clearId() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
int32 id = 1;
clearId() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo.Builder
int32 id = 1;
clearId() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
int64 id = 1;
clearId() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
int64 id = 1;
clearId() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData.Builder
optional string id = 2;
clearId() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
optional string id = 1;
clearIncomingEdges() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationEdge incoming_edges = 4;
clearIndex() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
int32 index = 2;
clearIndex() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int32 index = 2;
clearInfo() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoWrapper.Builder
.org.apache.spark.status.protobuf.ApplicationEnvironmentInfo info = 1;
clearInfo() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoWrapper.Builder
.org.apache.spark.status.protobuf.ApplicationInfo info = 1;
clearInfo() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper.Builder
.org.apache.spark.status.protobuf.ExecutorStageSummary info = 4;
clearInfo() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryWrapper.Builder
.org.apache.spark.status.protobuf.ExecutorSummary info = 1;
clearInfo() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper.Builder
.org.apache.spark.status.protobuf.JobData info = 1;
clearInfo() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryWrapper.Builder
.org.apache.spark.status.protobuf.ProcessSummary info = 1;
clearInfo() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoWrapper.Builder
.org.apache.spark.status.protobuf.RDDStorageInfo info = 1;
clearInfo() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryWrapper.Builder
.org.apache.spark.status.protobuf.SpeculationStageSummary info = 3;
clearInfo() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper.Builder
.org.apache.spark.status.protobuf.StageData info = 1;
clearInputBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double input_bytes = 6;
clearInputBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
int64 input_bytes = 5;
clearInputBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int64 input_bytes = 24;
clearInputBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 input_bytes_read = 26;
clearInputMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
.org.apache.spark.status.protobuf.InputMetricDistributions input_metrics = 15;
clearInputMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
.org.apache.spark.status.protobuf.InputMetrics input_metrics = 11;
clearInputRecords() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double input_records = 7;
clearInputRecords() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
int64 input_records = 6;
clearInputRecords() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int64 input_records = 25;
clearInputRecordsRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 input_records_read = 27;
clearInputRowsPerSecond() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
double input_rows_per_second = 6;
clearIsActive() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
bool is_active = 3;
clearIsActive() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder
bool is_active = 3;
clearIsActive() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData.Builder
bool is_active = 4;
clearIsBlacklisted() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
bool is_blacklisted = 18;
clearIsBlacklistedForStage() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
bool is_blacklisted_for_stage = 15;
clearIsExcluded() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
bool is_excluded = 30;
clearIsExcludedForStage() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
bool is_excluded_for_stage = 17;
clearIsShufflePushEnabled() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
bool is_shuffle_push_enabled = 63;
clearJavaHome() - Method in class org.apache.spark.status.protobuf.StoreTypes.RuntimeInfo.Builder
optional string java_home = 2;
clearJavaVersion() - Method in class org.apache.spark.status.protobuf.StoreTypes.RuntimeInfo.Builder
optional string java_version = 1;
clearJobGroup() - Method in class org.apache.spark.api.java.JavaSparkContext
Clear the current thread's job group ID and its description.
clearJobGroup() - Method in class org.apache.spark.SparkContext
Clear the current thread's job group ID and its description.
clearJobGroup() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
optional string job_group = 7;
clearJobId() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
All IDs are int64 for extendability, even when they are currently int32 in Spark.
clearJobIds() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper.Builder
repeated int64 job_ids = 2;
clearJobs() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
 
clearJvmGcTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double jvm_gc_time = 11;
clearJvmGcTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int64 jvm_gc_time = 19;
clearJvmGcTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 jvm_gc_time = 21;
clearJvmGcTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double jvm_gc_time = 8;
clearJvmGcTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
int64 jvm_gc_time = 6;
clearKilledTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double killed_tasks = 5;
clearKilledTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
int32 killed_tasks = 4;
clearKilledTasksSummary() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
 
clearKillTasksSummary() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
 
clearLastUpdated() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo.Builder
int64 last_updated = 4;
clearLatestOffset() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
optional string latest_offset = 4;
clearLaunchTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
int64 launch_time = 5;
clearLaunchTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 launch_time = 5;
clearLocalBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
repeated double local_blocks_fetched = 4;
clearLocalBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics.Builder
int64 local_blocks_fetched = 2;
clearLocalBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics.Builder
int64 local_bytes_read = 6;
clearLocality() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper.Builder
 
clearLocalMergedBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
repeated double local_merged_blocks_fetched = 4;
clearLocalMergedBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics.Builder
int64 local_merged_blocks_fetched = 4;
clearLocalMergedBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
repeated double local_merged_bytes_read = 8;
clearLocalMergedBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics.Builder
int64 local_merged_bytes_read = 8;
clearLocalMergedChunksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
repeated double local_merged_chunks_fetched = 6;
clearLocalMergedChunksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics.Builder
int64 local_merged_chunks_fetched = 6;
clearMaxCores() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
optional int32 max_cores = 4;
clearMaxMemory() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
int64 max_memory = 19;
clearMaxTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
int32 max_tasks = 8;
clearMemoryBytesSpilled() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double memory_bytes_spilled = 16;
clearMemoryBytesSpilled() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double memory_bytes_spilled = 14;
clearMemoryBytesSpilled() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
int64 memory_bytes_spilled = 13;
clearMemoryBytesSpilled() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int64 memory_bytes_spilled = 21;
clearMemoryBytesSpilled() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 memory_bytes_spilled = 23;
clearMemoryBytesSpilled() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double memory_bytes_spilled = 13;
clearMemoryBytesSpilled() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
int64 memory_bytes_spilled = 8;
clearMemoryMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
optional .org.apache.spark.status.protobuf.MemoryMetrics memory_metrics = 24;
clearMemoryPerExecutorMb() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
optional int32 memory_per_executor_mb = 6;
clearMemoryRemaining() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution.Builder
int64 memory_remaining = 3;
clearMemoryUsed() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
int64 memory_used = 5;
clearMemoryUsed() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution.Builder
int64 memory_used = 2;
clearMemoryUsed() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo.Builder
int64 memory_used = 3;
clearMemoryUsed() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
int64 memory_used = 6;
clearMemoryUsedBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
int64 memory_used_bytes = 8;
clearMemSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
int64 mem_size = 8;
clearMergedFetchFallbackCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
repeated double merged_fetch_fallback_count = 2;
clearMergedFetchFallbackCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics.Builder
int64 merged_fetch_fallback_count = 2;
clearMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetrics.Builder
 
clearMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress.Builder
 
clearMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
 
clearMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 5;
clearMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 4;
clearMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 7;
clearMetricsProperties() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings metrics_properties = 5;
clearMetricType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetric.Builder
optional string metric_type = 3;
clearMetricValues() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
 
clearMetricValuesIsNull() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
bool metric_values_is_null = 13;
clearModifiedConfigs() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
 
clearName() - Method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo.Builder
optional string name = 2;
clearName() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
optional string name = 2;
clearName() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
optional string name = 2;
clearName() - Method in class org.apache.spark.status.protobuf.StoreTypes.PoolData.Builder
optional string name = 1;
clearName() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
optional string name = 2;
clearName() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode.Builder
optional string name = 2;
clearName() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
optional string name = 2;
clearName() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceInformation.Builder
optional string name = 1;
clearName() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
optional string name = 2;
clearName() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
optional string name = 2;
clearName() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetric.Builder
optional string name = 1;
clearName() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
optional string name = 39;
clearName() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
optional string name = 1;
clearName() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData.Builder
optional string name = 1;
clearName() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
optional string name = 3;
clearNode() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper.Builder
.org.apache.spark.status.protobuf.SparkPlanGraphNode node = 1;
clearNodes() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 4;
clearNodes() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 2;
clearNumActiveStages() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
int32 num_active_stages = 16;
clearNumActiveTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
int32 num_active_tasks = 10;
clearNumActiveTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummary.Builder
int32 num_active_tasks = 2;
clearNumActiveTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int32 num_active_tasks = 5;
clearNumCachedPartitions() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
int32 num_cached_partitions = 4;
clearNumCompletedIndices() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
int32 num_completed_indices = 15;
clearNumCompletedIndices() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int32 num_completed_indices = 9;
clearNumCompletedJobs() - Method in class org.apache.spark.status.protobuf.StoreTypes.AppSummary.Builder
int32 num_completed_jobs = 1;
clearNumCompletedStages() - Method in class org.apache.spark.status.protobuf.StoreTypes.AppSummary.Builder
int32 num_completed_stages = 2;
clearNumCompletedStages() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
int32 num_completed_stages = 17;
clearNumCompletedTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
int32 num_completed_tasks = 11;
clearNumCompletedTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummary.Builder
int32 num_completed_tasks = 3;
clearNumCompleteTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int32 num_complete_tasks = 6;
clearNumFailedStages() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
int32 num_failed_stages = 19;
clearNumFailedTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
int32 num_failed_tasks = 13;
clearNumFailedTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummary.Builder
int32 num_failed_tasks = 4;
clearNumFailedTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int32 num_failed_tasks = 7;
clearNumInputRows() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
int64 num_input_rows = 5;
clearNumKilledTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
int32 num_killed_tasks = 14;
clearNumKilledTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummary.Builder
int32 num_killed_tasks = 5;
clearNumKilledTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int32 num_killed_tasks = 8;
clearNumOutputRows() - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress.Builder
int64 num_output_rows = 2;
clearNumPartitions() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
int32 num_partitions = 3;
clearNumRowsDroppedByWatermark() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
int64 num_rows_dropped_by_watermark = 9;
clearNumRowsRemoved() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
int64 num_rows_removed = 5;
clearNumRowsTotal() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
int64 num_rows_total = 2;
clearNumRowsUpdated() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
int64 num_rows_updated = 3;
clearNumShufflePartitions() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
int64 num_shuffle_partitions = 10;
clearNumSkippedStages() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
int32 num_skipped_stages = 18;
clearNumSkippedTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
int32 num_skipped_tasks = 12;
clearNumStateStoreInstances() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
int64 num_state_store_instances = 11;
clearNumTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
int32 num_tasks = 9;
clearNumTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummary.Builder
int32 num_tasks = 1;
clearNumTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int32 num_tasks = 4;
clearObservedMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
 
clearOffHeapMemoryRemaining() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution.Builder
optional int64 off_heap_memory_remaining = 8;
clearOffHeapMemoryUsed() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution.Builder
optional int64 off_heap_memory_used = 6;
clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo.Builder
 
clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo.Builder
 
clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
 
clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoWrapper.Builder
 
clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
 
clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoWrapper.Builder
 
clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.AppSummary.Builder
 
clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
 
clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetrics.Builder
 
clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
 
clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions.Builder
 
clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest.Builder
 
clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
 
clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper.Builder
 
clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
 
clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryWrapper.Builder
 
clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributions.Builder
 
clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetrics.Builder
 
clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
 
clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper.Builder
 
clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.MemoryMetrics.Builder
 
clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributions.Builder
 
clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetrics.Builder
 
clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.PairStrings.Builder
 
clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.PoolData.Builder
 
clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder
 
clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryWrapper.Builder
 
clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution.Builder
 
clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
 
clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationEdge.Builder
 
clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
 
clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode.Builder
 
clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo.Builder
 
clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
 
clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoWrapper.Builder
 
clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceInformation.Builder
 
clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo.Builder
 
clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileWrapper.Builder
 
clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.RuntimeInfo.Builder
 
clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
 
clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics.Builder
 
clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
 
clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics.Builder
 
clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions.Builder
 
clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetrics.Builder
 
clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress.Builder
 
clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
 
clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
 
clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphEdge.Builder
 
clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
 
clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper.Builder
 
clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
 
clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummary.Builder
 
clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryWrapper.Builder
 
clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
 
clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetric.Builder
 
clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
 
clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper.Builder
 
clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
 
clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
 
clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData.Builder
 
clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
 
clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressWrapper.Builder
 
clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
 
clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
 
clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
 
clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
 
clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskResourceRequest.Builder
 
clearOnHeapMemoryRemaining() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution.Builder
optional int64 on_heap_memory_remaining = 7;
clearOnHeapMemoryUsed() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution.Builder
optional int64 on_heap_memory_used = 5;
clearOperatorName() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
optional string operator_name = 1;
clearOutgoingEdges() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationEdge outgoing_edges = 3;
clearOutputBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double output_bytes = 8;
clearOutputBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
int64 output_bytes = 7;
clearOutputBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int64 output_bytes = 26;
clearOutputBytesWritten() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 output_bytes_written = 28;
clearOutputDeterministicLevel() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode.Builder
.org.apache.spark.status.protobuf.DeterministicLevel output_deterministic_level = 6;
clearOutputMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
.org.apache.spark.status.protobuf.OutputMetricDistributions output_metrics = 16;
clearOutputMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
.org.apache.spark.status.protobuf.OutputMetrics output_metrics = 12;
clearOutputRecords() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double output_records = 9;
clearOutputRecords() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
int64 output_records = 8;
clearOutputRecords() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int64 output_records = 27;
clearOutputRecordsWritten() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 output_records_written = 29;
clearPartitionId() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
int32 partition_id = 4;
clearPartitionId() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int32 partition_id = 4;
clearPartitions() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
repeated .org.apache.spark.status.protobuf.RDDPartitionInfo partitions = 9;
clearPeakExecutionMemory() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double peak_execution_memory = 15;
clearPeakExecutionMemory() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int64 peak_execution_memory = 23;
clearPeakExecutionMemory() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 peak_execution_memory = 25;
clearPeakExecutionMemory() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double peak_execution_memory = 12;
clearPeakExecutionMemory() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
int64 peak_execution_memory = 10;
clearPeakExecutorMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
optional .org.apache.spark.status.protobuf.ExecutorMetrics peak_executor_metrics = 50;
clearPeakMemoryMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
.org.apache.spark.status.protobuf.ExecutorPeakMetricsDistributions peak_memory_metrics = 16;
clearPeakMemoryMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
optional .org.apache.spark.status.protobuf.ExecutorMetrics peak_memory_metrics = 16;
clearPeakMemoryMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
optional .org.apache.spark.status.protobuf.ExecutorMetrics peak_memory_metrics = 26;
clearPhysicalPlanDescription() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
optional string physical_plan_description = 5;
clearProcessedRowsPerSecond() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
double processed_rows_per_second = 7;
clearProcessLogs() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder
 
clearProgress() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressWrapper.Builder
.org.apache.spark.status.protobuf.StreamingQueryProgress progress = 1;
clearQuantile() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
optional string quantile = 3;
clearQuantiles() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double quantiles = 1;
clearQuantiles() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions.Builder
repeated double quantiles = 1;
clearQuantiles() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double quantiles = 1;
clearRddBlocks() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
int32 rdd_blocks = 4;
clearRddIds() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
repeated int64 rdd_ids = 43;
clearReadBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
repeated double read_bytes = 1;
clearReadRecords() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
repeated double read_records = 2;
clearRecordsRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double records_read = 19;
clearRecordsRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributions.Builder
repeated double records_read = 2;
clearRecordsRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetrics.Builder
int64 records_read = 2;
clearRecordsRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics.Builder
int64 records_read = 7;
clearRecordsWritten() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double records_written = 21;
clearRecordsWritten() - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributions.Builder
repeated double records_written = 2;
clearRecordsWritten() - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetrics.Builder
int64 records_written = 2;
clearRecordsWritten() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetrics.Builder
int64 records_written = 3;
clearRemoteBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
repeated double remote_blocks_fetched = 3;
clearRemoteBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics.Builder
int64 remote_blocks_fetched = 1;
clearRemoteBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
repeated double remote_bytes_read = 6;
clearRemoteBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics.Builder
int64 remote_bytes_read = 4;
clearRemoteBytesReadToDisk() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
repeated double remote_bytes_read_to_disk = 7;
clearRemoteBytesReadToDisk() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics.Builder
int64 remote_bytes_read_to_disk = 5;
clearRemoteMergedBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
repeated double remote_merged_blocks_fetched = 3;
clearRemoteMergedBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics.Builder
int64 remote_merged_blocks_fetched = 3;
clearRemoteMergedBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
repeated double remote_merged_bytes_read = 7;
clearRemoteMergedBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics.Builder
int64 remote_merged_bytes_read = 7;
clearRemoteMergedChunksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
repeated double remote_merged_chunks_fetched = 5;
clearRemoteMergedChunksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics.Builder
int64 remote_merged_chunks_fetched = 5;
clearRemoteMergedReqsDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
repeated double remote_merged_reqs_duration = 9;
clearRemoteMergedReqsDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics.Builder
int64 remote_merged_reqs_duration = 9;
clearRemoteReqsDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
repeated double remote_reqs_duration = 9;
clearRemoteReqsDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics.Builder
int64 remote_reqs_duration = 8;
clearRemoveReason() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
optional string remove_reason = 22;
clearRemoveTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
optional int64 remove_time = 21;
clearRemoveTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder
optional int64 remove_time = 6;
clearResourceName() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest.Builder
optional string resource_name = 1;
clearResourceName() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskResourceRequest.Builder
optional string resource_name = 1;
clearResourceProfileId() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
int32 resource_profile_id = 29;
clearResourceProfileId() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int32 resource_profile_id = 49;
clearResourceProfiles() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.ResourceProfileInfo resource_profiles = 7;
clearResources() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
 
clearResultFetchStart() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
optional int64 result_fetch_start = 6;
clearResultFetchStart() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 result_fetch_start = 6;
clearResultSerializationTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double result_serialization_time = 12;
clearResultSerializationTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int64 result_serialization_time = 20;
clearResultSerializationTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 result_serialization_time = 22;
clearResultSerializationTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double result_serialization_time = 9;
clearResultSerializationTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
int64 result_serialization_time = 7;
clearResultSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double result_size = 10;
clearResultSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int64 result_size = 18;
clearResultSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 result_size = 20;
clearResultSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double result_size = 7;
clearResultSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
int64 result_size = 5;
clearRootCluster() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
.org.apache.spark.status.protobuf.RDDOperationClusterWrapper root_cluster = 5;
clearRootExecutionId() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
int64 root_execution_id = 2;
clearRpInfo() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileWrapper.Builder
.org.apache.spark.status.protobuf.ResourceProfileInfo rpInfo = 1;
clearRunId() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData.Builder
optional string run_id = 3;
clearRunId() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
optional string run_id = 2;
clearRuntime() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
.org.apache.spark.status.protobuf.RuntimeInfo runtime = 1;
clearScalaVersion() - Method in class org.apache.spark.status.protobuf.StoreTypes.RuntimeInfo.Builder
optional string scala_version = 3;
clearSchedulerDelay() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double scheduler_delay = 14;
clearSchedulerDelay() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
int64 scheduler_delay = 17;
clearSchedulerDelay() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double scheduler_delay = 11;
clearSchedulingPool() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
optional string scheduling_pool = 42;
clearShuffleBytesWritten() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 shuffle_bytes_written = 37;
clearShuffleCorruptMergedBlockChunks() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double shuffle_corrupt_merged_block_chunks = 33;
clearShuffleCorruptMergedBlockChunks() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int64 shuffle_corrupt_merged_block_chunks = 53;
clearShuffleCorruptMergedBlockChunks() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 shuffle_corrupt_merged_block_chunks = 42;
clearShuffleFetchWaitTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double shuffle_fetch_wait_time = 26;
clearShuffleFetchWaitTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int64 shuffle_fetch_wait_time = 30;
clearShuffleFetchWaitTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 shuffle_fetch_wait_time = 32;
clearShuffleLocalBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double shuffle_local_blocks_fetched = 25;
clearShuffleLocalBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int64 shuffle_local_blocks_fetched = 29;
clearShuffleLocalBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 shuffle_local_blocks_fetched = 31;
clearShuffleLocalBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int64 shuffle_local_bytes_read = 33;
clearShuffleLocalBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 shuffle_local_bytes_read = 35;
clearShuffleMergedFetchFallbackCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double shuffle_merged_fetch_fallback_count = 34;
clearShuffleMergedFetchFallbackCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int64 shuffle_merged_fetch_fallback_count = 54;
clearShuffleMergedFetchFallbackCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 shuffle_merged_fetch_fallback_count = 43;
clearShuffleMergedLocalBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double shuffle_merged_local_blocks_fetched = 36;
clearShuffleMergedLocalBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int64 shuffle_merged_local_blocks_fetched = 56;
clearShuffleMergedLocalBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 shuffle_merged_local_blocks_fetched = 45;
clearShuffleMergedLocalBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double shuffle_merged_local_bytes_read = 40;
clearShuffleMergedLocalBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int64 shuffle_merged_local_bytes_read = 60;
clearShuffleMergedLocalBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 shuffle_merged_local_bytes_read = 49;
clearShuffleMergedLocalChunksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double shuffle_merged_local_chunks_fetched = 38;
clearShuffleMergedLocalChunksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int64 shuffle_merged_local_chunks_fetched = 58;
clearShuffleMergedLocalChunksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 shuffle_merged_local_chunks_fetched = 47;
clearShuffleMergedRemoteBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double shuffle_merged_remote_blocks_fetched = 35;
clearShuffleMergedRemoteBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int64 shuffle_merged_remote_blocks_fetched = 55;
clearShuffleMergedRemoteBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 shuffle_merged_remote_blocks_fetched = 44;
clearShuffleMergedRemoteBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double shuffle_merged_remote_bytes_read = 39;
clearShuffleMergedRemoteBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int64 shuffle_merged_remote_bytes_read = 59;
clearShuffleMergedRemoteBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 shuffle_merged_remote_bytes_read = 48;
clearShuffleMergedRemoteChunksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double shuffle_merged_remote_chunks_fetched = 37;
clearShuffleMergedRemoteChunksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int64 shuffle_merged_remote_chunks_fetched = 57;
clearShuffleMergedRemoteChunksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 shuffle_merged_remote_chunks_fetched = 46;
clearShuffleMergedRemoteReqDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 shuffle_merged_remote_req_duration = 51;
clearShuffleMergedRemoteReqsDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double shuffle_merged_remote_reqs_duration = 42;
clearShuffleMergedRemoteReqsDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int64 shuffle_merged_remote_reqs_duration = 62;
clearShuffleMergersCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int32 shuffle_mergers_count = 64;
clearShufflePushReadMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics.Builder
.org.apache.spark.status.protobuf.ShufflePushReadMetrics shuffle_push_read_metrics = 9;
clearShufflePushReadMetricsDist() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
.org.apache.spark.status.protobuf.ShufflePushReadMetricDistributions shuffle_push_read_metrics_dist = 10;
clearShuffleRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double shuffle_read = 10;
clearShuffleRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
int64 shuffle_read = 9;
clearShuffleReadBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double shuffle_read_bytes = 22;
clearShuffleReadBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int64 shuffle_read_bytes = 34;
clearShuffleReadMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
.org.apache.spark.status.protobuf.ShuffleReadMetricDistributions shuffle_read_metrics = 17;
clearShuffleReadMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
.org.apache.spark.status.protobuf.ShuffleReadMetrics shuffle_read_metrics = 13;
clearShuffleReadRecords() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double shuffle_read_records = 11;
clearShuffleReadRecords() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
int64 shuffle_read_records = 10;
clearShuffleReadRecords() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int64 shuffle_read_records = 35;
clearShuffleRecordsRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double shuffle_records_read = 23;
clearShuffleRecordsRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 shuffle_records_read = 36;
clearShuffleRecordsWritten() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 shuffle_records_written = 39;
clearShuffleRemoteBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double shuffle_remote_blocks_fetched = 24;
clearShuffleRemoteBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int64 shuffle_remote_blocks_fetched = 28;
clearShuffleRemoteBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 shuffle_remote_blocks_fetched = 30;
clearShuffleRemoteBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double shuffle_remote_bytes_read = 27;
clearShuffleRemoteBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int64 shuffle_remote_bytes_read = 31;
clearShuffleRemoteBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 shuffle_remote_bytes_read = 33;
clearShuffleRemoteBytesReadToDisk() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double shuffle_remote_bytes_read_to_disk = 28;
clearShuffleRemoteBytesReadToDisk() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int64 shuffle_remote_bytes_read_to_disk = 32;
clearShuffleRemoteBytesReadToDisk() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 shuffle_remote_bytes_read_to_disk = 34;
clearShuffleRemoteReqsDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double shuffle_remote_reqs_duration = 41;
clearShuffleRemoteReqsDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int64 shuffle_remote_reqs_duration = 61;
clearShuffleRemoteReqsDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 shuffle_remote_reqs_duration = 50;
clearShuffleTotalBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double shuffle_total_blocks_fetched = 29;
clearShuffleWrite() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double shuffle_write = 12;
clearShuffleWrite() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
int64 shuffle_write = 11;
clearShuffleWriteBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double shuffle_write_bytes = 30;
clearShuffleWriteBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int64 shuffle_write_bytes = 36;
clearShuffleWriteMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
.org.apache.spark.status.protobuf.ShuffleWriteMetricDistributions shuffle_write_metrics = 18;
clearShuffleWriteMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
.org.apache.spark.status.protobuf.ShuffleWriteMetrics shuffle_write_metrics = 14;
clearShuffleWriteRecords() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double shuffle_write_records = 31;
clearShuffleWriteRecords() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double shuffle_write_records = 13;
clearShuffleWriteRecords() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
int64 shuffle_write_records = 12;
clearShuffleWriteRecords() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int64 shuffle_write_records = 38;
clearShuffleWriteTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double shuffle_write_time = 32;
clearShuffleWriteTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int64 shuffle_write_time = 37;
clearShuffleWriteTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 shuffle_write_time = 38;
clearSink() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
.org.apache.spark.status.protobuf.SinkProgress sink = 11;
clearSkippedStages() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper.Builder
repeated int32 skipped_stages = 2;
clearSources() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
repeated .org.apache.spark.status.protobuf.SourceProgress sources = 10;
clearSparkProperties() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings spark_properties = 2;
clearSparkUser() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo.Builder
optional string spark_user = 6;
clearSpeculationSummary() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
optional .org.apache.spark.status.protobuf.SpeculationStageSummary speculation_summary = 47;
clearSpeculative() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
bool speculative = 12;
clearSpeculative() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
bool speculative = 12;
clearSqlExecutionId() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper.Builder
optional int64 sql_execution_id = 3;
clearStageAttemptId() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
int32 stage_attempt_id = 2;
clearStageAttemptId() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper.Builder
int32 stage_attempt_id = 2;
clearStageAttemptId() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryWrapper.Builder
int32 stage_attempt_id = 2;
clearStageAttemptId() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int32 stage_attempt_id = 41;
clearStageId() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
int64 stage_id = 1;
clearStageId() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper.Builder
int64 stage_id = 1;
clearStageId() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
int64 stage_id = 1;
clearStageId() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryWrapper.Builder
int64 stage_id = 1;
clearStageId() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int64 stage_id = 2;
clearStageId() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 stage_id = 40;
clearStageIds() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
repeated int64 stage_ids = 6;
clearStageIds() - Method in class org.apache.spark.status.protobuf.StoreTypes.PoolData.Builder
repeated int64 stage_ids = 2;
clearStages() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
repeated int64 stages = 12;
clearStartOffset() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
optional string start_offset = 2;
clearStartTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo.Builder
int64 start_time = 2;
clearStartTimestamp() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData.Builder
int64 start_timestamp = 6;
clearStateOperators() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
repeated .org.apache.spark.status.protobuf.StateOperatorProgress state_operators = 9;
clearStatus() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
.org.apache.spark.status.protobuf.JobExecutionStatus status = 8;
clearStatus() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
.org.apache.spark.status.protobuf.StageStatus status = 1;
clearStatus() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
optional string status = 10;
clearStatus() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
optional string status = 10;
clearStorageLevel() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo.Builder
optional string storage_level = 2;
clearStorageLevel() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
optional string storage_level = 5;
clearStorageLevel() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
optional string storage_level = 4;
clearSubmissionTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
optional int64 submission_time = 4;
clearSubmissionTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
int64 submission_time = 8;
clearSubmissionTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
optional int64 submission_time = 10;
clearSucceededTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double succeeded_tasks = 4;
clearSucceededTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
int32 succeeded_tasks = 3;
clearSystemProperties() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings system_properties = 4;
clearTaskCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
int64 task_count = 4;
clearTaskId() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
int64 task_id = 1;
clearTaskId() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 task_id = 1;
clearTaskLocality() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
optional string task_locality = 11;
clearTaskLocality() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
optional string task_locality = 11;
clearTaskMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
optional .org.apache.spark.status.protobuf.TaskMetrics task_metrics = 15;
clearTaskMetricsDistributions() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
optional .org.apache.spark.status.protobuf.TaskMetricDistributions task_metrics_distributions = 51;
clearTaskResourceRequests() - Method in class org.apache.spark.resource.ResourceProfileBuilder
 
clearTaskResources() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo.Builder
 
clearTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
 
clearTaskTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double task_time = 2;
clearTaskTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
int64 task_time = 1;
clearThreshold() - Method in class org.apache.spark.mllib.classification.LogisticRegressionModel
Clears the threshold so that predict will output raw prediction scores.
clearThreshold() - Method in class org.apache.spark.mllib.classification.SVMModel
Clears the threshold so that predict will output raw prediction scores.
clearTimestamp() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
optional string timestamp = 4;
clearToId() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationEdge.Builder
int32 to_id = 2;
clearToId() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphEdge.Builder
int64 to_id = 2;
clearTotalBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
repeated double total_blocks_fetched = 8;
clearTotalCores() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
int32 total_cores = 7;
clearTotalCores() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder
int32 total_cores = 4;
clearTotalDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
int64 total_duration = 13;
clearTotalGcTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
int64 total_gc_time = 14;
clearTotalInputBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
int64 total_input_bytes = 15;
clearTotalOffHeapStorageMemory() - Method in class org.apache.spark.status.protobuf.StoreTypes.MemoryMetrics.Builder
int64 total_off_heap_storage_memory = 4;
clearTotalOnHeapStorageMemory() - Method in class org.apache.spark.status.protobuf.StoreTypes.MemoryMetrics.Builder
int64 total_on_heap_storage_memory = 3;
clearTotalShuffleRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
int64 total_shuffle_read = 16;
clearTotalShuffleWrite() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
int64 total_shuffle_write = 17;
clearTotalTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
int32 total_tasks = 12;
clearUpdate() - Method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo.Builder
optional string update = 3;
clearUseDisk() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
bool use_disk = 6;
clearUsedOffHeapStorageMemory() - Method in class org.apache.spark.status.protobuf.StoreTypes.MemoryMetrics.Builder
int64 used_off_heap_storage_memory = 2;
clearUsedOnHeapStorageMemory() - Method in class org.apache.spark.status.protobuf.StoreTypes.MemoryMetrics.Builder
int64 used_on_heap_storage_memory = 1;
clearUseMemory() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
bool use_memory = 5;
clearValue() - Method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo.Builder
optional string value = 4;
clearValue1() - Method in class org.apache.spark.status.protobuf.StoreTypes.PairStrings.Builder
optional string value1 = 1;
clearValue2() - Method in class org.apache.spark.status.protobuf.StoreTypes.PairStrings.Builder
optional string value2 = 2;
clearVendor() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest.Builder
optional string vendor = 4;
clearWrapper() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper.Builder
 
clearWriteBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions.Builder
repeated double write_bytes = 1;
clearWriteRecords() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions.Builder
repeated double write_records = 2;
clearWriteTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions.Builder
repeated double write_time = 3;
clearWriteTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetrics.Builder
int64 write_time = 2;
Clock - Interface in org.apache.spark.util
An interface to represent clocks, so that they can be mocked out in unit tests.
CLogLog$() - Constructor for class org.apache.spark.ml.regression.GeneralizedLinearRegression.CLogLog$
 
clone() - Method in class org.apache.spark.SparkConf
Copy this object
clone() - Method in class org.apache.spark.sql.ExperimentalMethods
 
clone() - Method in class org.apache.spark.sql.types.Decimal
 
clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo.Builder
 
clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo.Builder
 
clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
 
clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoWrapper.Builder
 
clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
 
clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoWrapper.Builder
 
clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.AppSummary.Builder
 
clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
 
clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetrics.Builder
 
clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
 
clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions.Builder
 
clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest.Builder
 
clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
 
clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper.Builder
 
clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
 
clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryWrapper.Builder
 
clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributions.Builder
 
clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetrics.Builder
 
clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
 
clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper.Builder
 
clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.MemoryMetrics.Builder
 
clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributions.Builder
 
clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetrics.Builder
 
clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.PairStrings.Builder
 
clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.PoolData.Builder
 
clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder
 
clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryWrapper.Builder
 
clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution.Builder
 
clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
 
clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationEdge.Builder
 
clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
 
clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode.Builder
 
clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo.Builder
 
clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
 
clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoWrapper.Builder
 
clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceInformation.Builder
 
clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo.Builder
 
clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileWrapper.Builder
 
clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.RuntimeInfo.Builder
 
clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
 
clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics.Builder
 
clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
 
clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics.Builder
 
clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions.Builder
 
clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetrics.Builder
 
clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress.Builder
 
clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
 
clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
 
clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphEdge.Builder
 
clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
 
clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper.Builder
 
clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
 
clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummary.Builder
 
clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryWrapper.Builder
 
clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
 
clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetric.Builder
 
clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
 
clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper.Builder
 
clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
 
clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
 
clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData.Builder
 
clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
 
clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressWrapper.Builder
 
clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
 
clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
 
clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
 
clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
 
clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskResourceRequest.Builder
 
clone() - Method in class org.apache.spark.storage.StorageLevel
 
clone() - Method in class org.apache.spark.util.random.BernoulliCellSampler
 
clone() - Method in class org.apache.spark.util.random.BernoulliSampler
 
clone() - Method in class org.apache.spark.util.random.PoissonSampler
 
clone() - Method in interface org.apache.spark.util.random.RandomSampler
return a copy of the RandomSampler object
clone(T, SerializerInstance, ClassTag<T>) - Static method in class org.apache.spark.util.Utils
Clone an object using a Spark serializer.
cloneComplement() - Method in class org.apache.spark.util.random.BernoulliCellSampler
Return a sampler that is the complement of the range specified of the current sampler.
cloneProperties(Properties) - Static method in class org.apache.spark.util.Utils
Create a new properties object with the same values as `props`
close() - Method in class org.apache.spark.api.java.JavaSparkContext
 
close() - Method in class org.apache.spark.io.NioBufferedFileInputStream
 
close() - Method in class org.apache.spark.io.ReadAheadInputStream
 
close() - Method in interface org.apache.spark.security.CryptoStreamUtils.BaseErrorHandler
 
close() - Method in class org.apache.spark.serializer.DeserializationStream
 
close() - Method in class org.apache.spark.serializer.SerializationStream
 
close(Throwable) - Method in class org.apache.spark.sql.ForeachWriter
Called when stopping to process one partition of new data in the executor side.
close() - Method in class org.apache.spark.sql.SparkSession
Synonym for stop().
close() - Method in class org.apache.spark.sql.vectorized.ArrowColumnVector
 
close() - Method in class org.apache.spark.sql.vectorized.ColumnarBatch
Called to close all the columns in this batch.
close() - Method in class org.apache.spark.sql.vectorized.ColumnVector
Cleans up memory for this column vector.
close() - Method in class org.apache.spark.storage.BufferReleasingInputStream
 
close() - Method in class org.apache.spark.storage.CountingWritableChannel
 
close() - Method in class org.apache.spark.storage.TimeTrackingOutputStream
 
close() - Method in class org.apache.spark.streaming.api.java.JavaStreamingContext
Deprecated.
 
close() - Method in class org.apache.spark.streaming.util.WriteAheadLog
Close this log and release any resources.
ClosureCleaner - Class in org.apache.spark.util
A cleaner that renders closures serializable if they can be done so safely.
ClosureCleaner() - Constructor for class org.apache.spark.util.ClosureCleaner
 
closureSerializer() - Method in class org.apache.spark.SparkEnv
 
cls() - Method in class org.apache.spark.sql.types.ObjectType
 
cls() - Method in class org.apache.spark.util.MethodIdentifier
 
clsTag() - Method in interface org.apache.spark.sql.Encoder
A ClassTag that can be used to construct an Array to contain a collection of T.
cluster() - Method in class org.apache.spark.ml.clustering.ClusteringSummary
 
cluster() - Method in class org.apache.spark.mllib.clustering.PowerIterationClustering.Assignment
 
Cluster$() - Constructor for class org.apache.spark.mllib.clustering.KMeansModel.Cluster$
 
CLUSTER_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper
 
clusterCenter() - Method in class org.apache.spark.ml.clustering.ClusterData
 
clusterCenters() - Method in class org.apache.spark.ml.clustering.BisectingKMeansModel
 
clusterCenters() - Method in class org.apache.spark.ml.clustering.KMeansModel
 
clusterCenters() - Method in class org.apache.spark.mllib.clustering.BisectingKMeansModel
Leaf cluster centers.
clusterCenters() - Method in class org.apache.spark.mllib.clustering.KMeansModel
 
clusterCenters() - Method in class org.apache.spark.mllib.clustering.StreamingKMeansModel
 
ClusterData - Class in org.apache.spark.ml.clustering
Helper class for storing model data
ClusterData(int, Vector) - Constructor for class org.apache.spark.ml.clustering.ClusterData
 
clustered(Expression[]) - Static method in class org.apache.spark.sql.connector.distributions.Distributions
Creates a distribution where tuples that share the same values for clustering expressions are co-located in the same partition.
clustered(Expression[]) - Static method in class org.apache.spark.sql.connector.distributions.LogicalDistributions
 
ClusteredDistribution - Interface in org.apache.spark.sql.connector.distributions
A distribution where tuples that share the same values for clustering expressions are co-located in the same partition.
clusterIdx() - Method in class org.apache.spark.ml.clustering.ClusterData
 
clustering() - Method in interface org.apache.spark.sql.connector.distributions.ClusteredDistribution
Returns clustering expressions.
ClusteringEvaluator - Class in org.apache.spark.ml.evaluation
Evaluator for clustering results.
ClusteringEvaluator(String) - Constructor for class org.apache.spark.ml.evaluation.ClusteringEvaluator
 
ClusteringEvaluator() - Constructor for class org.apache.spark.ml.evaluation.ClusteringEvaluator
 
ClusteringMetrics - Class in org.apache.spark.ml.evaluation
Metrics for clustering, which expects two input columns: prediction and label.
ClusteringSummary - Class in org.apache.spark.ml.clustering
Summary of clustering algorithms.
CLUSTERS_CONFIG_PREFIX() - Static method in class org.apache.spark.kafka010.KafkaTokenSparkConf
 
clusterSchedulerError(String) - Static method in class org.apache.spark.errors.SparkCoreErrors
 
clusterSizes() - Method in class org.apache.spark.ml.clustering.ClusteringSummary
 
ClusterStats(Vector, double, double) - Constructor for class org.apache.spark.ml.evaluation.SquaredEuclideanSilhouette.ClusterStats
 
ClusterStats$() - Constructor for class org.apache.spark.ml.evaluation.SquaredEuclideanSilhouette.ClusterStats$
 
clusterWeights() - Method in class org.apache.spark.mllib.clustering.StreamingKMeansModel
 
cmdOnlyWorksOnPartitionedTablesError(String, String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
cmdOnlyWorksOnTableWithLocationError(String, String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
cn() - Method in class org.apache.spark.mllib.feature.VocabWord
 
coalesce(int) - Method in class org.apache.spark.api.java.JavaDoubleRDD
Return a new RDD that is reduced into numPartitions partitions.
coalesce(int, boolean) - Method in class org.apache.spark.api.java.JavaDoubleRDD
Return a new RDD that is reduced into numPartitions partitions.
coalesce(int) - Method in class org.apache.spark.api.java.JavaPairRDD
Return a new RDD that is reduced into numPartitions partitions.
coalesce(int, boolean) - Method in class org.apache.spark.api.java.JavaPairRDD
Return a new RDD that is reduced into numPartitions partitions.
coalesce(int) - Method in class org.apache.spark.api.java.JavaRDD
Return a new RDD that is reduced into numPartitions partitions.
coalesce(int, boolean) - Method in class org.apache.spark.api.java.JavaRDD
Return a new RDD that is reduced into numPartitions partitions.
coalesce(int, RDD<?>) - Method in class org.apache.spark.rdd.DefaultPartitionCoalescer
Runs the packing algorithm and returns an array of PartitionGroups that if possible are load balanced and grouped by locality
coalesce(int, RDD<?>) - Method in interface org.apache.spark.rdd.PartitionCoalescer
Coalesce the partitions of the given RDD.
coalesce(int, boolean, Option<PartitionCoalescer>, Ordering<T>) - Method in class org.apache.spark.rdd.RDD
Return a new RDD that is reduced into numPartitions partitions.
coalesce(int) - Method in class org.apache.spark.sql.Dataset
Returns a new Dataset that has exactly numPartitions partitions, when the fewer partitions are requested.
coalesce(Column...) - Static method in class org.apache.spark.sql.functions
Returns the first column that is not null, or null if all inputs are null.
coalesce(Seq<Column>) - Static method in class org.apache.spark.sql.functions
Returns the first column that is not null, or null if all inputs are null.
CoarseGrainedClusterMessage - Interface in org.apache.spark.scheduler.cluster
 
CoarseGrainedClusterMessages - Class in org.apache.spark.scheduler.cluster
 
CoarseGrainedClusterMessages() - Constructor for class org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages
 
CoarseGrainedClusterMessages.AddWebUIFilter - Class in org.apache.spark.scheduler.cluster
 
CoarseGrainedClusterMessages.AddWebUIFilter$ - Class in org.apache.spark.scheduler.cluster
 
CoarseGrainedClusterMessages.DecommissionExecutor$ - Class in org.apache.spark.scheduler.cluster
 
CoarseGrainedClusterMessages.DecommissionExecutorsOnHost - Class in org.apache.spark.scheduler.cluster
 
CoarseGrainedClusterMessages.DecommissionExecutorsOnHost$ - Class in org.apache.spark.scheduler.cluster
 
CoarseGrainedClusterMessages.ExecutorDecommissioning - Class in org.apache.spark.scheduler.cluster
 
CoarseGrainedClusterMessages.ExecutorDecommissioning$ - Class in org.apache.spark.scheduler.cluster
 
CoarseGrainedClusterMessages.ExecutorDecommissionSigReceived$ - Class in org.apache.spark.scheduler.cluster
 
CoarseGrainedClusterMessages.GetExecutorLossReason - Class in org.apache.spark.scheduler.cluster
 
CoarseGrainedClusterMessages.GetExecutorLossReason$ - Class in org.apache.spark.scheduler.cluster
 
CoarseGrainedClusterMessages.IsExecutorAlive - Class in org.apache.spark.scheduler.cluster
 
CoarseGrainedClusterMessages.IsExecutorAlive$ - Class in org.apache.spark.scheduler.cluster
 
CoarseGrainedClusterMessages.KillExecutors - Class in org.apache.spark.scheduler.cluster
 
CoarseGrainedClusterMessages.KillExecutors$ - Class in org.apache.spark.scheduler.cluster
 
CoarseGrainedClusterMessages.KillExecutorsOnHost - Class in org.apache.spark.scheduler.cluster
 
CoarseGrainedClusterMessages.KillExecutorsOnHost$ - Class in org.apache.spark.scheduler.cluster
 
CoarseGrainedClusterMessages.KillTask - Class in org.apache.spark.scheduler.cluster
 
CoarseGrainedClusterMessages.KillTask$ - Class in org.apache.spark.scheduler.cluster
 
CoarseGrainedClusterMessages.LaunchedExecutor - Class in org.apache.spark.scheduler.cluster
 
CoarseGrainedClusterMessages.LaunchedExecutor$ - Class in org.apache.spark.scheduler.cluster
 
CoarseGrainedClusterMessages.LaunchTask - Class in org.apache.spark.scheduler.cluster
 
CoarseGrainedClusterMessages.LaunchTask$ - Class in org.apache.spark.scheduler.cluster
 
CoarseGrainedClusterMessages.MiscellaneousProcessAdded - Class in org.apache.spark.scheduler.cluster
 
CoarseGrainedClusterMessages.MiscellaneousProcessAdded$ - Class in org.apache.spark.scheduler.cluster
 
CoarseGrainedClusterMessages.RegisterClusterManager - Class in org.apache.spark.scheduler.cluster
 
CoarseGrainedClusterMessages.RegisterClusterManager$ - Class in org.apache.spark.scheduler.cluster
 
CoarseGrainedClusterMessages.RegisterExecutor - Class in org.apache.spark.scheduler.cluster
 
CoarseGrainedClusterMessages.RegisterExecutor$ - Class in org.apache.spark.scheduler.cluster
 
CoarseGrainedClusterMessages.RemoveExecutor - Class in org.apache.spark.scheduler.cluster
 
CoarseGrainedClusterMessages.RemoveExecutor$ - Class in org.apache.spark.scheduler.cluster
 
CoarseGrainedClusterMessages.RemoveWorker - Class in org.apache.spark.scheduler.cluster
 
CoarseGrainedClusterMessages.RemoveWorker$ - Class in org.apache.spark.scheduler.cluster
 
CoarseGrainedClusterMessages.RequestExecutors - Class in org.apache.spark.scheduler.cluster
 
CoarseGrainedClusterMessages.RequestExecutors$ - Class in org.apache.spark.scheduler.cluster
 
CoarseGrainedClusterMessages.RetrieveDelegationTokens$ - Class in org.apache.spark.scheduler.cluster
 
CoarseGrainedClusterMessages.RetrieveLastAllocatedExecutorId$ - Class in org.apache.spark.scheduler.cluster
 
CoarseGrainedClusterMessages.RetrieveSparkAppConfig - Class in org.apache.spark.scheduler.cluster
 
CoarseGrainedClusterMessages.RetrieveSparkAppConfig$ - Class in org.apache.spark.scheduler.cluster
 
CoarseGrainedClusterMessages.ReviveOffers$ - Class in org.apache.spark.scheduler.cluster
 
CoarseGrainedClusterMessages.SetupDriver - Class in org.apache.spark.scheduler.cluster
 
CoarseGrainedClusterMessages.SetupDriver$ - Class in org.apache.spark.scheduler.cluster
 
CoarseGrainedClusterMessages.ShufflePushCompletion - Class in org.apache.spark.scheduler.cluster
 
CoarseGrainedClusterMessages.ShufflePushCompletion$ - Class in org.apache.spark.scheduler.cluster
 
CoarseGrainedClusterMessages.Shutdown - Class in org.apache.spark.scheduler.cluster
 
CoarseGrainedClusterMessages.Shutdown$ - Class in org.apache.spark.scheduler.cluster
 
CoarseGrainedClusterMessages.SparkAppConfig - Class in org.apache.spark.scheduler.cluster
 
CoarseGrainedClusterMessages.SparkAppConfig$ - Class in org.apache.spark.scheduler.cluster
 
CoarseGrainedClusterMessages.StatusUpdate - Class in org.apache.spark.scheduler.cluster
 
CoarseGrainedClusterMessages.StatusUpdate$ - Class in org.apache.spark.scheduler.cluster
 
CoarseGrainedClusterMessages.StopDriver$ - Class in org.apache.spark.scheduler.cluster
 
CoarseGrainedClusterMessages.StopExecutor$ - Class in org.apache.spark.scheduler.cluster
 
CoarseGrainedClusterMessages.StopExecutors$ - Class in org.apache.spark.scheduler.cluster
 
CoarseGrainedClusterMessages.UpdateDelegationTokens - Class in org.apache.spark.scheduler.cluster
 
CoarseGrainedClusterMessages.UpdateDelegationTokens$ - Class in org.apache.spark.scheduler.cluster
 
code() - Method in class org.apache.spark.mllib.feature.VocabWord
 
CodegenMetrics - Class in org.apache.spark.metrics.source
Metrics for code generation.
CodegenMetrics() - Constructor for class org.apache.spark.metrics.source.CodegenMetrics
 
codeLen() - Method in class org.apache.spark.mllib.feature.VocabWord
 
coefficientMatrix() - Method in class org.apache.spark.ml.classification.LogisticRegressionModel
 
coefficients() - Method in class org.apache.spark.ml.classification.LinearSVCModel
 
coefficients() - Method in class org.apache.spark.ml.classification.LogisticRegressionModel
A vector of model coefficients for "binomial" logistic regression.
coefficients() - Method in class org.apache.spark.ml.regression.AFTSurvivalRegressionModel
 
coefficients() - Method in class org.apache.spark.ml.regression.GeneralizedLinearRegressionModel
 
coefficients() - Method in class org.apache.spark.ml.regression.LinearRegressionModel
 
coefficientStandardErrors() - Method in class org.apache.spark.ml.regression.GeneralizedLinearRegressionTrainingSummary
 
coefficientStandardErrors() - Method in class org.apache.spark.ml.regression.LinearRegressionSummary
 
cogroup(JavaPairRDD<K, W>, Partitioner) - Method in class org.apache.spark.api.java.JavaPairRDD
For each key k in this or other, return a resulting RDD that contains a tuple with the list of values for that key in this as well as other.
cogroup(JavaPairRDD<K, W1>, JavaPairRDD<K, W2>, Partitioner) - Method in class org.apache.spark.api.java.JavaPairRDD
For each key k in this or other1 or other2, return a resulting RDD that contains a tuple with the list of values for that key in this, other1 and other2.
cogroup(JavaPairRDD<K, W1>, JavaPairRDD<K, W2>, JavaPairRDD<K, W3>, Partitioner) - Method in class org.apache.spark.api.java.JavaPairRDD
For each key k in this or other1 or other2 or other3, return a resulting RDD that contains a tuple with the list of values for that key in this, other1, other2 and other3.
cogroup(JavaPairRDD<K, W>) - Method in class org.apache.spark.api.java.JavaPairRDD
For each key k in this or other, return a resulting RDD that contains a tuple with the list of values for that key in this as well as other.
cogroup(JavaPairRDD<K, W1>, JavaPairRDD<K, W2>) - Method in class org.apache.spark.api.java.JavaPairRDD
For each key k in this or other1 or other2, return a resulting RDD that contains a tuple with the list of values for that key in this, other1 and other2.
cogroup(JavaPairRDD<K, W1>, JavaPairRDD<K, W2>, JavaPairRDD<K, W3>) - Method in class org.apache.spark.api.java.JavaPairRDD
For each key k in this or other1 or other2 or other3, return a resulting RDD that contains a tuple with the list of values for that key in this, other1, other2 and other3.
cogroup(JavaPairRDD<K, W>, int) - Method in class org.apache.spark.api.java.JavaPairRDD
For each key k in this or other, return a resulting RDD that contains a tuple with the list of values for that key in this as well as other.
cogroup(JavaPairRDD<K, W1>, JavaPairRDD<K, W2>, int) - Method in class org.apache.spark.api.java.JavaPairRDD
For each key k in this or other1 or other2, return a resulting RDD that contains a tuple with the list of values for that key in this, other1 and other2.
cogroup(JavaPairRDD<K, W1>, JavaPairRDD<K, W2>, JavaPairRDD<K, W3>, int) - Method in class org.apache.spark.api.java.JavaPairRDD
For each key k in this or other1 or other2 or other3, return a resulting RDD that contains a tuple with the list of values for that key in this, other1, other2 and other3.
cogroup(RDD<Tuple2<K, W1>>, RDD<Tuple2<K, W2>>, RDD<Tuple2<K, W3>>, Partitioner) - Method in class org.apache.spark.rdd.PairRDDFunctions
For each key k in this or other1 or other2 or other3, return a resulting RDD that contains a tuple with the list of values for that key in this, other1, other2 and other3.
cogroup(RDD<Tuple2<K, W>>, Partitioner) - Method in class org.apache.spark.rdd.PairRDDFunctions
For each key k in this or other, return a resulting RDD that contains a tuple with the list of values for that key in this as well as other.
cogroup(RDD<Tuple2<K, W1>>, RDD<Tuple2<K, W2>>, Partitioner) - Method in class org.apache.spark.rdd.PairRDDFunctions
For each key k in this or other1 or other2, return a resulting RDD that contains a tuple with the list of values for that key in this, other1 and other2.
cogroup(RDD<Tuple2<K, W1>>, RDD<Tuple2<K, W2>>, RDD<Tuple2<K, W3>>) - Method in class org.apache.spark.rdd.PairRDDFunctions
For each key k in this or other1 or other2 or other3, return a resulting RDD that contains a tuple with the list of values for that key in this, other1, other2 and other3.
cogroup(RDD<Tuple2<K, W>>) - Method in class org.apache.spark.rdd.PairRDDFunctions
For each key k in this or other, return a resulting RDD that contains a tuple with the list of values for that key in this as well as other.
cogroup(RDD<Tuple2<K, W1>>, RDD<Tuple2<K, W2>>) - Method in class org.apache.spark.rdd.PairRDDFunctions
For each key k in this or other1 or other2, return a resulting RDD that contains a tuple with the list of values for that key in this, other1 and other2.
cogroup(RDD<Tuple2<K, W>>, int) - Method in class org.apache.spark.rdd.PairRDDFunctions
For each key k in this or other, return a resulting RDD that contains a tuple with the list of values for that key in this as well as other.
cogroup(RDD<Tuple2<K, W1>>, RDD<Tuple2<K, W2>>, int) - Method in class org.apache.spark.rdd.PairRDDFunctions
For each key k in this or other1 or other2, return a resulting RDD that contains a tuple with the list of values for that key in this, other1 and other2.
cogroup(RDD<Tuple2<K, W1>>, RDD<Tuple2<K, W2>>, RDD<Tuple2<K, W3>>, int) - Method in class org.apache.spark.rdd.PairRDDFunctions
For each key k in this or other1 or other2 or other3, return a resulting RDD that contains a tuple with the list of values for that key in this, other1, other2 and other3.
cogroup(KeyValueGroupedDataset<K, U>, Function3<K, Iterator<V>, Iterator<U>, TraversableOnce<R>>, Encoder<R>) - Method in class org.apache.spark.sql.KeyValueGroupedDataset
(Scala-specific) Applies the given function to each cogrouped data.
cogroup(KeyValueGroupedDataset<K, U>, CoGroupFunction<K, V, U, R>, Encoder<R>) - Method in class org.apache.spark.sql.KeyValueGroupedDataset
(Java-specific) Applies the given function to each cogrouped data.
cogroup(JavaPairDStream<K, W>) - Method in class org.apache.spark.streaming.api.java.JavaPairDStream
Return a new DStream by applying 'cogroup' between RDDs of this DStream and other DStream.
cogroup(JavaPairDStream<K, W>, int) - Method in class org.apache.spark.streaming.api.java.JavaPairDStream
Return a new DStream by applying 'cogroup' between RDDs of this DStream and other DStream.
cogroup(JavaPairDStream<K, W>, Partitioner) - Method in class org.apache.spark.streaming.api.java.JavaPairDStream
Return a new DStream by applying 'cogroup' between RDDs of this DStream and other DStream.
cogroup(DStream<Tuple2<K, W>>, ClassTag<W>) - Method in class org.apache.spark.streaming.dstream.PairDStreamFunctions
Return a new DStream by applying 'cogroup' between RDDs of this DStream and other DStream.
cogroup(DStream<Tuple2<K, W>>, int, ClassTag<W>) - Method in class org.apache.spark.streaming.dstream.PairDStreamFunctions
Return a new DStream by applying 'cogroup' between RDDs of this DStream and other DStream.
cogroup(DStream<Tuple2<K, W>>, Partitioner, ClassTag<W>) - Method in class org.apache.spark.streaming.dstream.PairDStreamFunctions
Return a new DStream by applying 'cogroup' between RDDs of this DStream and other DStream.
CoGroupedRDD<K> - Class in org.apache.spark.rdd
:: DeveloperApi :: An RDD that cogroups its parents.
CoGroupedRDD(Seq<RDD<? extends Product2<K, ?>>>, Partitioner, ClassTag<K>) - Constructor for class org.apache.spark.rdd.CoGroupedRDD
 
CoGroupFunction<K,V1,V2,R> - Interface in org.apache.spark.api.java.function
A function that returns zero or more output records from each grouping key and its values from 2 Datasets.
cogroupSorted(KeyValueGroupedDataset<K, U>, Seq<Column>, Seq<Column>, Function3<K, Iterator<V>, Iterator<U>, TraversableOnce<R>>, Encoder<R>) - Method in class org.apache.spark.sql.KeyValueGroupedDataset
(Scala-specific) Applies the given function to each sorted cogrouped data.
cogroupSorted(KeyValueGroupedDataset<K, U>, Column[], Column[], CoGroupFunction<K, V, U, R>, Encoder<R>) - Method in class org.apache.spark.sql.KeyValueGroupedDataset
(Java-specific) Applies the given function to each sorted cogrouped data.
col(String) - Method in class org.apache.spark.sql.Dataset
Selects column based on the column name and returns it as a Column.
col(String) - Static method in class org.apache.spark.sql.functions
Returns a Column based on the given column name.
COL_POS_KEY() - Static method in class org.apache.spark.sql.Dataset
 
coldStartStrategy() - Method in class org.apache.spark.ml.recommendation.ALS
 
coldStartStrategy() - Method in class org.apache.spark.ml.recommendation.ALSModel
 
coldStartStrategy() - Method in interface org.apache.spark.ml.recommendation.ALSModelParams
Param for strategy for dealing with unknown or new users/items at prediction time.
colIter() - Method in class org.apache.spark.ml.linalg.DenseMatrix
 
colIter() - Method in interface org.apache.spark.ml.linalg.Matrix
Returns an iterator of column vectors.
colIter() - Method in class org.apache.spark.ml.linalg.SparseMatrix
 
colIter() - Method in class org.apache.spark.mllib.linalg.DenseMatrix
 
colIter() - Method in interface org.apache.spark.mllib.linalg.Matrix
Returns an iterator of column vectors.
colIter() - Method in class org.apache.spark.mllib.linalg.SparseMatrix
 
collect() - Method in interface org.apache.spark.api.java.JavaRDDLike
Return an array that contains all of the elements in this RDD.
collect() - Method in class org.apache.spark.graphx.impl.EdgeRDDImpl
 
collect() - Method in class org.apache.spark.rdd.RDD
Return an array that contains all of the elements in this RDD.
collect(PartialFunction<T, U>, ClassTag<U>) - Method in class org.apache.spark.rdd.RDD
Return an RDD that contains all matching values by applying f.
collect() - Method in class org.apache.spark.sql.Dataset
Returns an array that contains all rows in this Dataset.
collect_list(Column) - Static method in class org.apache.spark.sql.functions
Aggregate function: returns a list of objects with duplicates.
collect_list(String) - Static method in class org.apache.spark.sql.functions
Aggregate function: returns a list of objects with duplicates.
collect_set(Column) - Static method in class org.apache.spark.sql.functions
Aggregate function: returns a set of objects with duplicate elements eliminated.
collect_set(String) - Static method in class org.apache.spark.sql.functions
Aggregate function: returns a set of objects with duplicate elements eliminated.
collectAsList() - Method in class org.apache.spark.sql.Dataset
Returns a Java list that contains all rows in this Dataset.
collectAsMap() - Method in class org.apache.spark.api.java.JavaPairRDD
Return the key-value pairs in this RDD to the master as a Map.
collectAsMap() - Method in class org.apache.spark.rdd.PairRDDFunctions
Return the key-value pairs in this RDD to the master as a Map.
collectAsync() - Method in interface org.apache.spark.api.java.JavaRDDLike
The asynchronous version of collect, which returns a future for retrieving an array containing all of the elements in this RDD.
collectAsync() - Method in class org.apache.spark.rdd.AsyncRDDActions
Returns a future for retrieving all elements of this RDD.
collectEdges(EdgeDirection) - Method in class org.apache.spark.graphx.GraphOps
Returns an RDD that contains for each vertex v its local edges, i.e., the edges that are incident on v, in the user-specified direction.
collectionAccumulator() - Method in class org.apache.spark.SparkContext
Create and register a CollectionAccumulator, which starts with empty list and accumulates inputs by adding them into the list.
collectionAccumulator(String) - Method in class org.apache.spark.SparkContext
Create and register a CollectionAccumulator, which starts with empty list and accumulates inputs by adding them into the list.
CollectionAccumulator<T> - Class in org.apache.spark.util
An accumulator for collecting a list of elements.
CollectionAccumulator() - Constructor for class org.apache.spark.util.CollectionAccumulator
 
CollectionsUtils - Class in org.apache.spark.util
 
CollectionsUtils() - Constructor for class org.apache.spark.util.CollectionsUtils
 
collectNeighborIds(EdgeDirection) - Method in class org.apache.spark.graphx.GraphOps
Collect the neighbor vertex ids for each vertex.
collectNeighbors(EdgeDirection) - Method in class org.apache.spark.graphx.GraphOps
Collect the neighbor vertex attributes for each vertex.
collectPartitions(int[]) - Method in interface org.apache.spark.api.java.JavaRDDLike
Return an array that contains all of the elements in a specific partition of this RDD.
collectSubModels() - Method in interface org.apache.spark.ml.param.shared.HasCollectSubModels
Param for whether to collect a list of sub-models trained during tuning.
collectSubModels() - Method in class org.apache.spark.ml.tuning.CrossValidator
 
collectSubModels() - Method in class org.apache.spark.ml.tuning.TrainValidationSplit
 
colPtrs() - Method in class org.apache.spark.ml.linalg.SparseMatrix
 
colPtrs() - Method in class org.apache.spark.mllib.linalg.SparseMatrix
 
colRegex(String) - Method in class org.apache.spark.sql.Dataset
Selects column based on the column name specified as a regex and returns it as Column.
colsPerBlock() - Method in class org.apache.spark.mllib.linalg.distributed.BlockMatrix
 
colStats(RDD<Vector>) - Static method in class org.apache.spark.mllib.stat.Statistics
Computes column-wise summary statistics for the input RDD[Vector].
Column - Class in org.apache.spark.sql.catalog
A column in Spark, as returned by listColumns method in Catalog.
Column(String, String, String, boolean, boolean, boolean) - Constructor for class org.apache.spark.sql.catalog.Column
 
Column - Class in org.apache.spark.sql
A column that will be computed based on the data in a DataFrame.
Column(Expression) - Constructor for class org.apache.spark.sql.Column
 
Column(String) - Constructor for class org.apache.spark.sql.Column
 
Column - Interface in org.apache.spark.sql.connector.catalog
An interface representing a column of a Table.
column() - Method in class org.apache.spark.sql.connector.catalog.TableChange.After
 
column() - Method in class org.apache.spark.sql.connector.expressions.aggregate.Avg
 
column() - Method in class org.apache.spark.sql.connector.expressions.aggregate.Count
 
column() - Method in class org.apache.spark.sql.connector.expressions.aggregate.Max
 
column() - Method in class org.apache.spark.sql.connector.expressions.aggregate.Min
 
column() - Method in class org.apache.spark.sql.connector.expressions.aggregate.Sum
 
column(String) - Static method in class org.apache.spark.sql.connector.expressions.Expressions
Create a named reference expression for a (nested) column.
column(String) - Static method in class org.apache.spark.sql.functions
Returns a Column based on the given column name.
column(int) - Method in class org.apache.spark.sql.vectorized.ColumnarBatch
Returns the column at `ordinal`.
columnAliases() - Method in interface org.apache.spark.sql.connector.catalog.View
The view column aliases.
columnAliasInOperationNotAllowedError(String, SqlBaseParser.TableAliasContext) - Static method in class org.apache.spark.sql.errors.QueryParsingErrors
 
columnAlreadyExistsError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
ColumnarArray - Class in org.apache.spark.sql.vectorized
Array abstraction in ColumnVector.
ColumnarArray(ColumnVector, int, int) - Constructor for class org.apache.spark.sql.vectorized.ColumnarArray
 
ColumnarBatch - Class in org.apache.spark.sql.vectorized
This class wraps multiple ColumnVectors as a row-wise table.
ColumnarBatch(ColumnVector[]) - Constructor for class org.apache.spark.sql.vectorized.ColumnarBatch
 
ColumnarBatch(ColumnVector[], int) - Constructor for class org.apache.spark.sql.vectorized.ColumnarBatch
Create a new batch from existing column vectors.
ColumnarBatchRow - Class in org.apache.spark.sql.vectorized
This class wraps an array of ColumnVector and provides a row view.
ColumnarBatchRow(ColumnVector[]) - Constructor for class org.apache.spark.sql.vectorized.ColumnarBatchRow
 
ColumnarMap - Class in org.apache.spark.sql.vectorized
Map abstraction in ColumnVector.
ColumnarMap(ColumnVector, ColumnVector, int, int) - Constructor for class org.apache.spark.sql.vectorized.ColumnarMap
 
ColumnarRow - Class in org.apache.spark.sql.vectorized
Row abstraction in ColumnVector.
ColumnarRow(ColumnVector, int) - Constructor for class org.apache.spark.sql.vectorized.ColumnarRow
 
columnComments() - Method in interface org.apache.spark.sql.connector.catalog.View
The view column comments.
ColumnDefaultValue - Class in org.apache.spark.sql.connector.catalog
A class representing the default value of a column.
ColumnDefaultValue(String, Literal<?>) - Constructor for class org.apache.spark.sql.connector.catalog.ColumnDefaultValue
 
ColumnName - Class in org.apache.spark.sql
A convenient class used for constructing schema.
ColumnName(String) - Constructor for class org.apache.spark.sql.ColumnName
 
columnNotDefinedInTableError(String, String, String, Seq<String>) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
columnNotFoundError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
columnNotFoundInExistingColumnsError(String, String, Seq<String>) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
columnNotFoundInSchemaError(StructField, Option<StructType>) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
columnNotInGroupByClauseError(Expression) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
columnProperties() - Method in class org.apache.spark.sql.connector.catalog.index.TableIndex
 
ColumnPruner - Class in org.apache.spark.ml.feature
Utility transformer for removing temporary columns from a DataFrame.
ColumnPruner(String, Set<String>) - Constructor for class org.apache.spark.ml.feature.ColumnPruner
 
ColumnPruner(Set<String>) - Constructor for class org.apache.spark.ml.feature.ColumnPruner
 
columns() - Method in class org.apache.spark.sql.connector.catalog.index.TableIndex
 
columns() - Method in interface org.apache.spark.sql.connector.catalog.Table
Returns the columns of this table.
columns() - Method in class org.apache.spark.sql.Dataset
Returns all column names as an array.
columnSchema() - Static method in class org.apache.spark.ml.image.ImageSchema
Schema for the image column: Row(String, Int, Int, Int, Int, Array[Byte])
ColumnsHelper(Column[]) - Constructor for class org.apache.spark.sql.connector.catalog.CatalogV2Implicits.ColumnsHelper
 
columnSimilarities() - Method in class org.apache.spark.mllib.linalg.distributed.IndexedRowMatrix
Compute all cosine similarities between columns of this matrix using the brute-force approach of computing normalized dot products.
columnSimilarities() - Method in class org.apache.spark.mllib.linalg.distributed.RowMatrix
Compute all cosine similarities between columns of this matrix using the brute-force approach of computing normalized dot products.
columnSimilarities(double) - Method in class org.apache.spark.mllib.linalg.distributed.RowMatrix
Compute similarities between columns of this matrix using a sampling approach.
ColumnStatistics - Interface in org.apache.spark.sql.connector.read.colstats
An interface to represent column statistics, which is part of Statistics.
columnStatisticsDeserializationNotSupportedError(String, DataType) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
columnStatisticsSerializationNotSupportedError(String, DataType) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
columnStats() - Method in interface org.apache.spark.sql.connector.read.Statistics
 
columnsToPrune() - Method in class org.apache.spark.ml.feature.ColumnPruner
 
columnToOldVector(Dataset<?>, String) - Static method in class org.apache.spark.ml.util.DatasetUtils
 
columnToVector(Dataset<?>, String) - Static method in class org.apache.spark.ml.util.DatasetUtils
Cast a column in a Dataset to Vector type.
columnTypeNotSupportStatisticsCollectionError(String, TableIdentifier, DataType) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
ColumnVector - Class in org.apache.spark.sql.vectorized
An interface representing in-memory columnar data in Spark.
combinationQueryResultClausesUnsupportedError(SqlBaseParser.QueryOrganizationContext) - Static method in class org.apache.spark.sql.errors.QueryParsingErrors
 
combineByKey(Function<V, C>, Function2<C, V, C>, Function2<C, C, C>, Partitioner, boolean, Serializer) - Method in class org.apache.spark.api.java.JavaPairRDD
Generic function to combine the elements for each key using a custom set of aggregation functions.
combineByKey(Function<V, C>, Function2<C, V, C>, Function2<C, C, C>, Partitioner) - Method in class org.apache.spark.api.java.JavaPairRDD
Generic function to combine the elements for each key using a custom set of aggregation functions.
combineByKey(Function<V, C>, Function2<C, V, C>, Function2<C, C, C>, int) - Method in class org.apache.spark.api.java.JavaPairRDD
Simplified version of combineByKey that hash-partitions the output RDD and uses map-side aggregation.
combineByKey(Function<V, C>, Function2<C, V, C>, Function2<C, C, C>) - Method in class org.apache.spark.api.java.JavaPairRDD
Simplified version of combineByKey that hash-partitions the resulting RDD using the existing partitioner/parallelism level and using map-side aggregation.
combineByKey(Function1<V, C>, Function2<C, V, C>, Function2<C, C, C>, Partitioner, boolean, Serializer) - Method in class org.apache.spark.rdd.PairRDDFunctions
Generic function to combine the elements for each key using a custom set of aggregation functions.
combineByKey(Function1<V, C>, Function2<C, V, C>, Function2<C, C, C>, int) - Method in class org.apache.spark.rdd.PairRDDFunctions
Simplified version of combineByKeyWithClassTag that hash-partitions the output RDD.
combineByKey(Function1<V, C>, Function2<C, V, C>, Function2<C, C, C>) - Method in class org.apache.spark.rdd.PairRDDFunctions
Simplified version of combineByKeyWithClassTag that hash-partitions the resulting RDD using the existing partitioner/parallelism level.
combineByKey(Function<V, C>, Function2<C, V, C>, Function2<C, C, C>, Partitioner) - Method in class org.apache.spark.streaming.api.java.JavaPairDStream
Combine elements of each key in DStream's RDDs using custom function.
combineByKey(Function<V, C>, Function2<C, V, C>, Function2<C, C, C>, Partitioner, boolean) - Method in class org.apache.spark.streaming.api.java.JavaPairDStream
Combine elements of each key in DStream's RDDs using custom function.
combineByKey(Function1<V, C>, Function2<C, V, C>, Function2<C, C, C>, Partitioner, boolean, ClassTag<C>) - Method in class org.apache.spark.streaming.dstream.PairDStreamFunctions
Combine elements of each key in DStream's RDDs using custom functions.
combineByKeyWithClassTag(Function1<V, C>, Function2<C, V, C>, Function2<C, C, C>, Partitioner, boolean, Serializer, ClassTag<C>) - Method in class org.apache.spark.rdd.PairRDDFunctions
Generic function to combine the elements for each key using a custom set of aggregation functions.
combineByKeyWithClassTag(Function1<V, C>, Function2<C, V, C>, Function2<C, C, C>, int, ClassTag<C>) - Method in class org.apache.spark.rdd.PairRDDFunctions
Simplified version of combineByKeyWithClassTag that hash-partitions the output RDD.
combineByKeyWithClassTag(Function1<V, C>, Function2<C, V, C>, Function2<C, C, C>, ClassTag<C>) - Method in class org.apache.spark.rdd.PairRDDFunctions
Simplified version of combineByKeyWithClassTag that hash-partitions the resulting RDD using the existing partitioner/parallelism level.
combineCombinersByKey(Iterator<? extends Product2<K, C>>, TaskContext) - Method in class org.apache.spark.Aggregator
 
combineValuesByKey(Iterator<? extends Product2<K, V>>, TaskContext) - Method in class org.apache.spark.Aggregator
 
command() - Method in interface org.apache.spark.sql.connector.write.RowLevelOperation
Returns the SQL command that is being performed.
command() - Method in interface org.apache.spark.sql.connector.write.RowLevelOperationInfo
Returns the row-level SQL command (e.g.
commandExecutionInRunnerUnsupportedError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
CommandLineLoggingUtils - Interface in org.apache.spark.util
 
CommandLineUtils - Interface in org.apache.spark.util
Contains basic command line parsing functionality and methods to parse some common Spark CLI options.
commandNotSupportNestedColumnError(String, String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
commandUnsupportedInV2TableError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
comment() - Method in interface org.apache.spark.sql.connector.catalog.Column
Returns the comment of this table column.
comment() - Method in interface org.apache.spark.sql.connector.catalog.MetadataColumn
Documentation for this metadata column, or null.
comment() - Method in class org.apache.spark.sql.connector.catalog.TableChange.AddColumn
 
commentOnTableUnsupportedError() - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
commit(Function0<Parsers.Parser<T>>) - Static method in class org.apache.spark.ml.feature.RFormulaParser
 
commit(Offset) - Method in interface org.apache.spark.sql.connector.read.streaming.SparkDataStream
Informs the source that Spark has completed processing all data for offsets less than or equal to `end` and will only request offsets greater than `end` in the future.
commit(WriterCommitMessage[]) - Method in interface org.apache.spark.sql.connector.write.BatchWrite
Commits this writing job with a list of commit messages.
commit() - Method in interface org.apache.spark.sql.connector.write.DataWriter
Commits this writer after all records are written successfully, returns a commit message which will be sent back to driver side and passed to BatchWrite.commit(WriterCommitMessage[]).
commit(long, WriterCommitMessage[]) - Method in interface org.apache.spark.sql.connector.write.streaming.StreamingWrite
Commits this writing job for the specified epoch with a list of commit messages.
COMMIT_TIME_MS_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress
 
commitAllPartitions(long[]) - Method in interface org.apache.spark.shuffle.api.ShuffleMapOutputWriter
Commits the writes done by all partition writers returned by all calls to this object's ShuffleMapOutputWriter.getPartitionWriter(int), and returns the number of bytes written for each partition.
commitDeniedError(int, long, int, int, int) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
commitStagedChanges() - Method in interface org.apache.spark.sql.connector.catalog.StagedTable
Finalize the creation or replacement of this table.
commitTask(OutputCommitter, TaskAttemptContext, int, int) - Static method in class org.apache.spark.mapred.SparkHadoopMapRedUtil
Commits a task output.
commitTimeMs() - Method in class org.apache.spark.sql.streaming.StateOperatorProgress
 
commonHeaderNodes(HttpServletRequest) - Static method in class org.apache.spark.ui.UIUtils
 
comparator(Schedulable, Schedulable) - Method in interface org.apache.spark.scheduler.SchedulingAlgorithm
 
comparatorReturnsNull(String, String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
compare(PartitionGroup, PartitionGroup) - Method in class org.apache.spark.rdd.DefaultPartitionCoalescer.partitionGroupOrdering$
 
compare(byte, byte) - Static method in class org.apache.spark.sql.types.ByteExactNumeric
 
compare(Decimal) - Method in class org.apache.spark.sql.types.Decimal
 
compare(Decimal, Decimal) - Method in interface org.apache.spark.sql.types.Decimal.DecimalIsConflicted
 
compare(Decimal, Decimal) - Static method in class org.apache.spark.sql.types.DecimalExactNumeric
 
compare(double, double) - Static method in class org.apache.spark.sql.types.DoubleExactNumeric
 
compare(double, double) - Method in class org.apache.spark.sql.types.DoubleType.DoubleAsIfIntegral$
 
compare(float, float) - Static method in class org.apache.spark.sql.types.FloatExactNumeric
 
compare(float, float) - Method in class org.apache.spark.sql.types.FloatType.FloatAsIfIntegral$
 
compare(int, int) - Static method in class org.apache.spark.sql.types.IntegerExactNumeric
 
compare(long, long) - Static method in class org.apache.spark.sql.types.LongExactNumeric
 
compare(short, short) - Static method in class org.apache.spark.sql.types.ShortExactNumeric
 
compare(RDDInfo) - Method in class org.apache.spark.storage.RDDInfo
 
compareTo(NumericHistogram.Coord) - Method in class org.apache.spark.sql.util.NumericHistogram.Coord
 
compareTo(SparkShutdownHook) - Method in class org.apache.spark.util.SparkShutdownHook
 
compileAggregate(AggregateFunc) - Static method in class org.apache.spark.sql.jdbc.DB2Dialect
 
compileAggregate(AggregateFunc) - Static method in class org.apache.spark.sql.jdbc.DerbyDialect
 
compileAggregate(AggregateFunc) - Static method in class org.apache.spark.sql.jdbc.H2Dialect
 
compileAggregate(AggregateFunc) - Method in class org.apache.spark.sql.jdbc.JdbcDialect
Deprecated.
use org.apache.spark.sql.jdbc.JdbcDialect.compileExpression instead. Since 3.4.0.
compileAggregate(AggregateFunc) - Static method in class org.apache.spark.sql.jdbc.MsSqlServerDialect
 
compileAggregate(AggregateFunc) - Static method in class org.apache.spark.sql.jdbc.MySQLDialect
 
compileAggregate(AggregateFunc) - Static method in class org.apache.spark.sql.jdbc.NoopDialect
 
compileAggregate(AggregateFunc) - Static method in class org.apache.spark.sql.jdbc.OracleDialect
 
compileAggregate(AggregateFunc) - Static method in class org.apache.spark.sql.jdbc.PostgresDialect
 
compileAggregate(AggregateFunc) - Static method in class org.apache.spark.sql.jdbc.TeradataDialect
 
compileExpression(Expression) - Static method in class org.apache.spark.sql.jdbc.DB2Dialect
 
compileExpression(Expression) - Static method in class org.apache.spark.sql.jdbc.DerbyDialect
 
compileExpression(Expression) - Static method in class org.apache.spark.sql.jdbc.H2Dialect
 
compileExpression(Expression) - Method in class org.apache.spark.sql.jdbc.JdbcDialect
Converts V2 expression to String representing a SQL expression.
compileExpression(Expression) - Static method in class org.apache.spark.sql.jdbc.MsSqlServerDialect
 
compileExpression(Expression) - Static method in class org.apache.spark.sql.jdbc.MySQLDialect
 
compileExpression(Expression) - Static method in class org.apache.spark.sql.jdbc.NoopDialect
 
compileExpression(Expression) - Static method in class org.apache.spark.sql.jdbc.OracleDialect
 
compileExpression(Expression) - Static method in class org.apache.spark.sql.jdbc.PostgresDialect
 
compileExpression(Expression) - Static method in class org.apache.spark.sql.jdbc.TeradataDialect
 
compilerError(CompileException) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
compileValue(Object) - Static method in class org.apache.spark.sql.jdbc.DB2Dialect
 
compileValue(Object) - Static method in class org.apache.spark.sql.jdbc.DerbyDialect
 
compileValue(Object) - Static method in class org.apache.spark.sql.jdbc.H2Dialect
 
compileValue(Object) - Method in class org.apache.spark.sql.jdbc.JdbcDialect
Converts value to SQL expression.
compileValue(Object) - Static method in class org.apache.spark.sql.jdbc.MsSqlServerDialect
 
compileValue(Object) - Static method in class org.apache.spark.sql.jdbc.MySQLDialect
 
compileValue(Object) - Static method in class org.apache.spark.sql.jdbc.NoopDialect
 
compileValue(Object) - Static method in class org.apache.spark.sql.jdbc.OracleDialect
 
compileValue(Object) - Static method in class org.apache.spark.sql.jdbc.PostgresDialect
 
compileValue(Object) - Static method in class org.apache.spark.sql.jdbc.TeradataDialect
 
Complete() - Static method in class org.apache.spark.sql.streaming.OutputMode
OutputMode in which all the rows in the streaming DataFrame/Dataset will be written to the sink every time there are some updates.
completed() - Method in class org.apache.spark.status.api.v1.ApplicationAttemptInfo
 
COMPLETED_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo
 
COMPLETED_TASKS_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
 
completedIndices() - Method in class org.apache.spark.status.LiveJob
 
completedIndices() - Method in class org.apache.spark.status.LiveStage
 
completedStages() - Method in class org.apache.spark.status.LiveJob
 
completedTasks() - Method in class org.apache.spark.status.api.v1.ExecutorSummary
 
completedTasks() - Method in class org.apache.spark.status.LiveJob
 
completedTasks() - Method in class org.apache.spark.status.LiveStage
 
COMPLETION_TIME() - Static method in class org.apache.spark.status.TaskIndexNames
 
COMPLETION_TIME_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.JobData
 
COMPLETION_TIME_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
 
COMPLETION_TIME_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.StageData
 
completionTime() - Method in class org.apache.spark.scheduler.StageInfo
Time when the stage completed or when the stage was cancelled.
completionTime() - Method in class org.apache.spark.status.api.v1.JobData
 
completionTime() - Method in class org.apache.spark.status.api.v1.StageData
 
completionTime() - Method in class org.apache.spark.status.LiveJob
 
ComplexFutureAction<T> - Class in org.apache.spark
A FutureAction for actions that could trigger multiple Spark jobs.
ComplexFutureAction(Function1<JobSubmitter, Future<T>>) - Constructor for class org.apache.spark.ComplexFutureAction
 
componentName() - Method in class org.apache.spark.resource.ResourceID
 
compositeLimit(ReadLimit[]) - Static method in interface org.apache.spark.sql.connector.read.streaming.ReadLimit
 
CompositeReadLimit - Class in org.apache.spark.sql.connector.read.streaming
/** Represents a ReadLimit where the MicroBatchStream should scan approximately given maximum number of rows with at least the given minimum number of rows.
compressed() - Method in interface org.apache.spark.ml.linalg.Matrix
Returns a matrix in dense column major, dense row major, sparse row major, or sparse column major format, whichever uses less storage.
compressed() - Method in interface org.apache.spark.ml.linalg.Vector
Returns a vector in either dense or sparse format, whichever uses less storage.
compressed() - Method in interface org.apache.spark.mllib.linalg.Vector
Returns a vector in either dense or sparse format, whichever uses less storage.
compressedColMajor() - Method in interface org.apache.spark.ml.linalg.Matrix
Returns a matrix in dense or sparse column major format, whichever uses less storage.
compressedContinuousInputStream(InputStream) - Method in interface org.apache.spark.io.CompressionCodec
 
compressedContinuousInputStream(InputStream) - Method in class org.apache.spark.io.ZStdCompressionCodec
 
compressedContinuousOutputStream(OutputStream) - Method in interface org.apache.spark.io.CompressionCodec
 
compressedInputStream(InputStream) - Method in interface org.apache.spark.io.CompressionCodec
 
compressedInputStream(InputStream) - Method in class org.apache.spark.io.LZ4CompressionCodec
 
compressedInputStream(InputStream) - Method in class org.apache.spark.io.LZFCompressionCodec
 
compressedInputStream(InputStream) - Method in class org.apache.spark.io.SnappyCompressionCodec
 
compressedInputStream(InputStream) - Method in class org.apache.spark.io.ZStdCompressionCodec
 
compressedOutputStream(OutputStream) - Method in interface org.apache.spark.io.CompressionCodec
 
compressedOutputStream(OutputStream) - Method in class org.apache.spark.io.LZ4CompressionCodec
 
compressedOutputStream(OutputStream) - Method in class org.apache.spark.io.LZFCompressionCodec
 
compressedOutputStream(OutputStream) - Method in class org.apache.spark.io.SnappyCompressionCodec
 
compressedOutputStream(OutputStream) - Method in class org.apache.spark.io.ZStdCompressionCodec
 
compressedRowMajor() - Method in interface org.apache.spark.ml.linalg.Matrix
Returns a matrix in dense or sparse row major format, whichever uses less storage.
CompressionCodec - Interface in org.apache.spark.io
:: DeveloperApi :: CompressionCodec allows the customization of choosing different compression implementations to be used in block storage.
compute(Partition, TaskContext) - Method in class org.apache.spark.api.r.BaseRRDD
 
compute(Partition, TaskContext) - Method in class org.apache.spark.graphx.EdgeRDD
 
compute(Partition, TaskContext) - Method in class org.apache.spark.graphx.VertexRDD
Provides the RDD[(VertexId, VD)] equivalent output.
compute(Vector, double, Vector) - Method in class org.apache.spark.mllib.optimization.Gradient
Compute the gradient and loss given the features of a single data point.
compute(Vector, double, Vector, Vector) - Method in class org.apache.spark.mllib.optimization.Gradient
Compute the gradient and loss given the features of a single data point, add the gradient to a provided vector to avoid creating new objects, and return loss.
compute(Vector, double, Vector) - Method in class org.apache.spark.mllib.optimization.HingeGradient
 
compute(Vector, double, Vector, Vector) - Method in class org.apache.spark.mllib.optimization.HingeGradient
 
compute(Vector, Vector, double, int, double) - Method in class org.apache.spark.mllib.optimization.L1Updater
 
compute(Vector, double, Vector) - Method in class org.apache.spark.mllib.optimization.LeastSquaresGradient
 
compute(Vector, double, Vector, Vector) - Method in class org.apache.spark.mllib.optimization.LeastSquaresGradient
 
compute(Vector, double, Vector, Vector) - Method in class org.apache.spark.mllib.optimization.LogisticGradient
 
compute(Vector, Vector, double, int, double) - Method in class org.apache.spark.mllib.optimization.SimpleUpdater
 
compute(Vector, Vector, double, int, double) - Method in class org.apache.spark.mllib.optimization.SquaredL2Updater
 
compute(Vector, Vector, double, int, double) - Method in class org.apache.spark.mllib.optimization.Updater
Compute an updated value for weights given the gradient, stepSize, iteration number and regularization parameter.
compute(Partition, TaskContext) - Method in class org.apache.spark.rdd.CoGroupedRDD
 
compute(Partition, TaskContext) - Method in class org.apache.spark.rdd.HadoopRDD
 
compute(Partition, TaskContext) - Method in class org.apache.spark.rdd.JdbcRDD
 
compute(Partition, TaskContext) - Method in class org.apache.spark.rdd.NewHadoopRDD
 
compute(Partition, TaskContext) - Method in class org.apache.spark.rdd.PartitionPruningRDD
 
compute(Partition, TaskContext) - Method in class org.apache.spark.rdd.RDD
:: DeveloperApi :: Implemented by subclasses to compute a given partition.
compute(Partition, TaskContext) - Method in class org.apache.spark.rdd.ShuffledRDD
 
compute(Partition, TaskContext) - Method in class org.apache.spark.rdd.UnionRDD
 
compute(Time) - Method in class org.apache.spark.streaming.api.java.JavaDStream
Generate an RDD for the given duration
compute(Time) - Method in class org.apache.spark.streaming.api.java.JavaPairDStream
Method that generates an RDD for the given Duration
compute(Time) - Method in class org.apache.spark.streaming.dstream.ConstantInputDStream
 
compute(Time) - Method in class org.apache.spark.streaming.dstream.DStream
Method that generates an RDD for the given time
compute(Time) - Method in class org.apache.spark.streaming.dstream.ReceiverInputDStream
 
compute(long, long, long, long) - Method in interface org.apache.spark.streaming.scheduler.rate.RateEstimator
Computes the number of records the stream attached to this RateEstimator should ingest per second, given an update on the size and completion times of the latest batch.
computeClusterStats(Dataset<Row>, String, String, String) - Static method in class org.apache.spark.ml.evaluation.CosineSilhouette
The method takes the input dataset and computes the aggregated values about a cluster which are needed by the algorithm.
computeClusterStats(Dataset<Row>, String, String, String) - Static method in class org.apache.spark.ml.evaluation.SquaredEuclideanSilhouette
The method takes the input dataset and computes the aggregated values about a cluster which are needed by the algorithm.
computeColumnSummaryStatistics() - Method in class org.apache.spark.mllib.linalg.distributed.RowMatrix
Computes column-wise summary statistics.
computeCorrelation(RDD<Object>, RDD<Object>) - Method in interface org.apache.spark.mllib.stat.correlation.Correlation
Compute correlation for two datasets.
computeCorrelation(RDD<Object>, RDD<Object>) - Static method in class org.apache.spark.mllib.stat.correlation.PearsonCorrelation
Compute the Pearson correlation for two datasets.
computeCorrelation(RDD<Object>, RDD<Object>) - Static method in class org.apache.spark.mllib.stat.correlation.SpearmanCorrelation
Compute Spearman's correlation for two datasets.
computeCorrelationMatrix(RDD<Vector>) - Method in interface org.apache.spark.mllib.stat.correlation.Correlation
Compute the correlation matrix S, for the input matrix, where S(i, j) is the correlation between column i and j.
computeCorrelationMatrix(RDD<Vector>) - Static method in class org.apache.spark.mllib.stat.correlation.PearsonCorrelation
Compute the Pearson correlation matrix S, for the input matrix, where S(i, j) is the correlation between column i and j.
computeCorrelationMatrix(RDD<Vector>) - Static method in class org.apache.spark.mllib.stat.correlation.SpearmanCorrelation
Compute Spearman's correlation matrix S, for the input matrix, where S(i, j) is the correlation between column i and j.
computeCorrelationMatrixFromCovariance(Matrix) - Static method in class org.apache.spark.mllib.stat.correlation.PearsonCorrelation
Compute the Pearson correlation matrix from the covariance matrix.
computeCorrelationWithMatrixImpl(RDD<Object>, RDD<Object>) - Method in interface org.apache.spark.mllib.stat.correlation.Correlation
Combine the two input RDD[Double]s into an RDD[Vector] and compute the correlation using the correlation implementation for RDD[Vector].
computeCorrelationWithMatrixImpl(RDD<Object>, RDD<Object>) - Static method in class org.apache.spark.mllib.stat.correlation.PearsonCorrelation
 
computeCorrelationWithMatrixImpl(RDD<Object>, RDD<Object>) - Static method in class org.apache.spark.mllib.stat.correlation.SpearmanCorrelation
 
computeCost(Dataset<?>) - Method in class org.apache.spark.ml.clustering.BisectingKMeansModel
Deprecated.
This method is deprecated and will be removed in future versions. Use ClusteringEvaluator instead. You can also get the cost on the training dataset in the summary.
computeCost(Vector) - Method in class org.apache.spark.mllib.clustering.BisectingKMeansModel
Computes the squared distance between the input point and the cluster center it belongs to.
computeCost(RDD<Vector>) - Method in class org.apache.spark.mllib.clustering.BisectingKMeansModel
Computes the sum of squared distances between the input points and their corresponding cluster centers.
computeCost(JavaRDD<Vector>) - Method in class org.apache.spark.mllib.clustering.BisectingKMeansModel
Java-friendly version of computeCost().
computeCost(RDD<Vector>) - Method in class org.apache.spark.mllib.clustering.KMeansModel
Return the K-means cost (sum of squared distances of points to their nearest center) for this model on the given data.
computeCovariance() - Method in class org.apache.spark.mllib.linalg.distributed.RowMatrix
Computes the covariance matrix, treating each row as an observation.
computeError(org.apache.spark.mllib.tree.model.TreeEnsembleModel, RDD<LabeledPoint>) - Method in interface org.apache.spark.mllib.tree.loss.Loss
Method to calculate error of the base learner for the gradient boosting calculation.
computeError(double, double) - Method in interface org.apache.spark.mllib.tree.loss.Loss
Method to calculate loss when the predictions are already known.
computeFractionForSampleSize(int, long, boolean) - Static method in class org.apache.spark.util.random.SamplingUtils
Returns a sampling rate that guarantees a sample of size greater than or equal to sampleSizeLowerBound 99.99% of the time.
computeGradient(DenseMatrix<Object>, DenseMatrix<Object>, Vector, int) - Method in interface org.apache.spark.ml.ann.TopologyModel
Computes gradient for the network
computeGramianMatrix() - Method in class org.apache.spark.mllib.linalg.distributed.IndexedRowMatrix
Computes the Gramian matrix A^T A.
computeGramianMatrix() - Method in class org.apache.spark.mllib.linalg.distributed.RowMatrix
Computes the Gramian matrix A^T A.
computeInitialPredictionAndError(RDD<TreePoint>, double, DecisionTreeRegressionModel, Loss, Broadcast<Split[][]>) - Static method in class org.apache.spark.ml.tree.impl.GradientBoostedTrees
Compute the initial predictions and errors for a dataset for the first iteration of gradient boosting.
computeInitialPredictionAndError(RDD<LabeledPoint>, double, DecisionTreeModel, Loss) - Static method in class org.apache.spark.mllib.tree.model.GradientBoostedTreesModel
Compute the initial predictions and errors for a dataset for the first iteration of gradient boosting.
computePreferredLocations(Seq<InputFormatInfo>) - Static method in class org.apache.spark.scheduler.InputFormatInfo
Computes the preferred locations based on input(s) and returned a location to block map.
computePrevDelta(DenseMatrix<Object>, DenseMatrix<Object>, DenseMatrix<Object>) - Method in interface org.apache.spark.ml.ann.LayerModel
Computes the delta for back propagation.
computePrincipalComponents(int) - Method in class org.apache.spark.mllib.linalg.distributed.RowMatrix
Computes the top k principal components only.
computePrincipalComponentsAndExplainedVariance(int) - Method in class org.apache.spark.mllib.linalg.distributed.RowMatrix
Computes the top k principal components and a vector of proportions of variance explained by each principal component.
computeProbability(double) - Method in interface org.apache.spark.mllib.tree.loss.ClassificationLoss
Computes the class probability given the margin.
computeSilhouetteCoefficient(Broadcast<Map<Object, Tuple2<Vector, Object>>>, Vector, double, double) - Static method in class org.apache.spark.ml.evaluation.CosineSilhouette
It computes the Silhouette coefficient for a point.
computeSilhouetteCoefficient(Broadcast<Map<Object, SquaredEuclideanSilhouette.ClusterStats>>, Vector, double, double, double) - Static method in class org.apache.spark.ml.evaluation.SquaredEuclideanSilhouette
It computes the Silhouette coefficient for a point.
computeSilhouetteScore(Dataset<?>, String, String, String) - Static method in class org.apache.spark.ml.evaluation.CosineSilhouette
Compute the Silhouette score of the dataset using the cosine distance measure.
computeSilhouetteScore(Dataset<?>, String, String, String) - Static method in class org.apache.spark.ml.evaluation.SquaredEuclideanSilhouette
Compute the Silhouette score of the dataset using squared Euclidean distance measure.
computeStatisticsNotExpectedError(SqlBaseParser.IdentifierContext) - Static method in class org.apache.spark.sql.errors.QueryParsingErrors
 
computeSVD(int, boolean, double) - Method in class org.apache.spark.mllib.linalg.distributed.IndexedRowMatrix
Computes the singular value decomposition of this IndexedRowMatrix.
computeSVD(int, boolean, double) - Method in class org.apache.spark.mllib.linalg.distributed.RowMatrix
Computes singular value decomposition of this matrix.
computeThresholdByKey(Map<K, AcceptanceResult>, Map<K, Object>) - Static method in class org.apache.spark.util.random.StratifiedSamplingUtils
Given the result returned by getCounts, determine the threshold for accepting items to generate exact sample size.
computeWeightedError(RDD<org.apache.spark.ml.feature.Instance>, DecisionTreeRegressionModel[], double[], Loss) - Static method in class org.apache.spark.ml.tree.impl.GradientBoostedTrees
Method to calculate error of the base learner for the gradient boosting calculation.
computeWeightedError(RDD<TreePoint>, RDD<Tuple2<Object, Object>>) - Static method in class org.apache.spark.ml.tree.impl.GradientBoostedTrees
Method to calculate error of the base learner for the gradient boosting calculation.
concat(Column...) - Static method in class org.apache.spark.sql.functions
Concatenates multiple input columns together into a single column.
concat(Seq<Column>) - Static method in class org.apache.spark.sql.functions
Concatenates multiple input columns together into a single column.
concat_ws(String, Column...) - Static method in class org.apache.spark.sql.functions
Concatenates multiple input string columns together into a single string column, using the given separator.
concat_ws(String, Seq<Column>) - Static method in class org.apache.spark.sql.functions
Concatenates multiple input string columns together into a single string column, using the given separator.
concatArraysWithElementsExceedLimitError(long) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
concurrentModificationOnExternalAppendOnlyUnsafeRowArrayError(String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
concurrentQueryInstanceError() - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
conf() - Method in interface org.apache.spark.api.plugin.PluginContext
Configuration of the Spark application.
Conf(int, int, double, double, double, double, double, double) - Constructor for class org.apache.spark.graphx.lib.SVDPlusPlus.Conf
 
conf() - Method in class org.apache.spark.SparkEnv
 
conf() - Static method in class org.apache.spark.sql.jdbc.MySQLDialect
 
conf() - Static method in class org.apache.spark.sql.jdbc.PostgresDialect
 
conf() - Method in class org.apache.spark.sql.SparkSession
 
confidence() - Method in class org.apache.spark.mllib.fpm.AssociationRules.Rule
Returns the confidence of the rule.
confidence() - Method in class org.apache.spark.partial.BoundedDouble
 
confidence() - Method in class org.apache.spark.util.sketch.CountMinSketch
Returns the confidence (or delta) of this CountMinSketch.
config(String, String) - Method in class org.apache.spark.sql.SparkSession.Builder
Sets a config option.
config(String, long) - Method in class org.apache.spark.sql.SparkSession.Builder
Sets a config option.
config(String, double) - Method in class org.apache.spark.sql.SparkSession.Builder
Sets a config option.
config(String, boolean) - Method in class org.apache.spark.sql.SparkSession.Builder
Sets a config option.
config(Map<String, Object>) - Method in class org.apache.spark.sql.SparkSession.Builder
Sets a config option.
config(Map<String, Object>) - Method in class org.apache.spark.sql.SparkSession.Builder
Sets a config option.
config(SparkConf) - Method in class org.apache.spark.sql.SparkSession.Builder
Sets a list of config options based on the given SparkConf.
configRemovedInVersionError(String, String, String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
configTestLog4j2(String) - Static method in class org.apache.spark.TestUtils
config a log4j2 properties used for testsuite
Configurable - Interface in org.apache.spark.input
A trait to implement Configurable interface.
configuration() - Method in class org.apache.spark.scheduler.InputFormatInfo
 
CONFIGURATION_INSTANTIATION_LOCK() - Static method in class org.apache.spark.rdd.HadoopRDD
Configuration's constructor is not threadsafe (see SPARK-1097 and HADOOP-10456).
CONFIGURATION_INSTANTIATION_LOCK() - Static method in class org.apache.spark.rdd.NewHadoopRDD
Configuration's constructor is not threadsafe (see SPARK-1097 and HADOOP-10456).
conflictingAttributesInJoinConditionError(AttributeSet, LogicalPlan, LogicalPlan) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
confusionMatrix() - Method in class org.apache.spark.mllib.evaluation.MulticlassMetrics
Returns confusion matrix: predicted classes are in columns, they are ordered by class label ascending, as in "labels"
connectedComponents() - Method in class org.apache.spark.graphx.GraphOps
Compute the connected component membership of each vertex and return a graph with the vertex value containing the lowest vertex id in the connected component containing that vertex.
connectedComponents(int) - Method in class org.apache.spark.graphx.GraphOps
Compute the connected component membership of each vertex and return a graph with the vertex value containing the lowest vertex id in the connected component containing that vertex.
ConnectedComponents - Class in org.apache.spark.graphx.lib
Connected components algorithm.
ConnectedComponents() - Constructor for class org.apache.spark.graphx.lib.ConnectedComponents
 
consequent() - Method in class org.apache.spark.mllib.fpm.AssociationRules.Rule
 
ConstantInputDStream<T> - Class in org.apache.spark.streaming.dstream
An input stream that always returns the same RDD on each time step.
ConstantInputDStream(StreamingContext, RDD<T>, ClassTag<T>) - Constructor for class org.apache.spark.streaming.dstream.ConstantInputDStream
 
constructorNotFoundError(String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
constructTree(org.apache.spark.mllib.tree.model.DecisionTreeModel.SaveLoadV1_0.NodeData[]) - Method in class org.apache.spark.mllib.tree.model.DecisionTreeModel.SaveLoadV1_0$
Given a list of nodes from a tree, construct the tree.
constructTrees(RDD<org.apache.spark.mllib.tree.model.DecisionTreeModel.SaveLoadV1_0.NodeData>) - Method in class org.apache.spark.mllib.tree.model.DecisionTreeModel.SaveLoadV1_0$
 
contains(Param<?>) - Method in class org.apache.spark.ml.param.ParamMap
Checks whether a parameter is explicitly specified.
contains(String) - Method in class org.apache.spark.SparkConf
Does the configuration contain a given parameter?
contains(Object) - Method in class org.apache.spark.sql.Column
Contains the other element.
contains(String) - Method in class org.apache.spark.sql.types.Metadata
Tests whether this Metadata contains a binding for a key.
contains(T) - Method in class org.apache.spark.sql.util.SQLOpenHashSet
 
containsAttributes(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
map<string, string> attributes = 27;
containsAttributes(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
map<string, string> attributes = 27;
containsAttributes(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
map<string, string> attributes = 27;
containsCustomMetrics(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
map<string, int64> custom_metrics = 12;
containsCustomMetrics(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress
map<string, int64> custom_metrics = 12;
containsCustomMetrics(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgressOrBuilder
map<string, int64> custom_metrics = 12;
containsDurationMs(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
map<string, int64> duration_ms = 7;
containsDurationMs(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
map<string, int64> duration_ms = 7;
containsDurationMs(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
map<string, int64> duration_ms = 7;
containsEventTime(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
map<string, string> event_time = 8;
containsEventTime(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
map<string, string> event_time = 8;
containsEventTime(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
map<string, string> event_time = 8;
containsExecutorLogs(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
map<string, string> executor_logs = 23;
containsExecutorLogs(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
map<string, string> executor_logs = 23;
containsExecutorLogs(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
map<string, string> executor_logs = 23;
containsExecutorLogs(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
map<string, string> executor_logs = 16;
containsExecutorLogs(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
map<string, string> executor_logs = 16;
containsExecutorLogs(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataOrBuilder
map<string, string> executor_logs = 16;
containsExecutorResources(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo.Builder
map<string, .org.apache.spark.status.protobuf.ExecutorResourceRequest> executor_resources = 2;
containsExecutorResources(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo
map<string, .org.apache.spark.status.protobuf.ExecutorResourceRequest> executor_resources = 2;
containsExecutorResources(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfoOrBuilder
map<string, .org.apache.spark.status.protobuf.ExecutorResourceRequest> executor_resources = 2;
containsExecutorSummary(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
map<string, .org.apache.spark.status.protobuf.ExecutorStageSummary> executor_summary = 46;
containsExecutorSummary(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
map<string, .org.apache.spark.status.protobuf.ExecutorStageSummary> executor_summary = 46;
containsExecutorSummary(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
map<string, .org.apache.spark.status.protobuf.ExecutorStageSummary> executor_summary = 46;
containsJobs(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
map<int64, .org.apache.spark.status.protobuf.JobExecutionStatus> jobs = 11;
containsJobs(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
map<int64, .org.apache.spark.status.protobuf.JobExecutionStatus> jobs = 11;
containsJobs(long) - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
map<int64, .org.apache.spark.status.protobuf.JobExecutionStatus> jobs = 11;
containsKey(Object) - Method in class org.apache.spark.api.java.JavaUtils.SerializableMapWrapper
 
containsKey(Object) - Method in class org.apache.spark.sql.util.CaseInsensitiveStringMap
 
containsKilledTasksSummary(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
map<string, int32> killed_tasks_summary = 48;
containsKilledTasksSummary(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
map<string, int32> killed_tasks_summary = 48;
containsKilledTasksSummary(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
map<string, int32> killed_tasks_summary = 48;
containsKillTasksSummary(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
map<string, int32> kill_tasks_summary = 20;
containsKillTasksSummary(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
map<string, int32> kill_tasks_summary = 20;
containsKillTasksSummary(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataOrBuilder
map<string, int32> kill_tasks_summary = 20;
containsLocality(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper.Builder
map<string, int64> locality = 3;
containsLocality(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper
map<string, int64> locality = 3;
containsLocality(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataWrapperOrBuilder
map<string, int64> locality = 3;
containsMetrics(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetrics.Builder
map<string, int64> metrics = 1;
containsMetrics(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetrics
map<string, int64> metrics = 1;
containsMetrics(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsOrBuilder
map<string, int64> metrics = 1;
containsMetrics(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress.Builder
map<string, string> metrics = 3;
containsMetrics(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress
map<string, string> metrics = 3;
containsMetrics(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.SinkProgressOrBuilder
map<string, string> metrics = 3;
containsMetrics(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
map<string, string> metrics = 8;
containsMetrics(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress
map<string, string> metrics = 8;
containsMetrics(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.SourceProgressOrBuilder
map<string, string> metrics = 8;
containsMetricValues(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
map<int64, string> metric_values = 14;
containsMetricValues(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
map<int64, string> metric_values = 14;
containsMetricValues(long) - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
map<int64, string> metric_values = 14;
containsModifiedConfigs(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
map<string, string> modified_configs = 6;
containsModifiedConfigs(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
map<string, string> modified_configs = 6;
containsModifiedConfigs(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
map<string, string> modified_configs = 6;
containsNaN() - Method in class org.apache.spark.sql.util.SQLOpenHashSet
 
containsNull() - Method in class org.apache.spark.sql.types.ArrayType
 
containsNull() - Method in class org.apache.spark.sql.util.SQLOpenHashSet
 
containsObservedMetrics(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
map<string, string> observed_metrics = 12;
containsObservedMetrics(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
map<string, string> observed_metrics = 12;
containsObservedMetrics(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
map<string, string> observed_metrics = 12;
containsProcessLogs(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder
map<string, string> process_logs = 7;
containsProcessLogs(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary
map<string, string> process_logs = 7;
containsProcessLogs(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryOrBuilder
map<string, string> process_logs = 7;
containsResources(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
map<string, .org.apache.spark.status.protobuf.ResourceInformation> resources = 28;
containsResources(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
map<string, .org.apache.spark.status.protobuf.ResourceInformation> resources = 28;
containsResources(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
map<string, .org.apache.spark.status.protobuf.ResourceInformation> resources = 28;
containsTaskResources(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo.Builder
map<string, .org.apache.spark.status.protobuf.TaskResourceRequest> task_resources = 3;
containsTaskResources(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo
map<string, .org.apache.spark.status.protobuf.TaskResourceRequest> task_resources = 3;
containsTaskResources(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfoOrBuilder
map<string, .org.apache.spark.status.protobuf.TaskResourceRequest> task_resources = 3;
containsTasks(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
map<int64, .org.apache.spark.status.protobuf.TaskData> tasks = 45;
containsTasks(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
map<int64, .org.apache.spark.status.protobuf.TaskData> tasks = 45;
containsTasks(long) - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
map<int64, .org.apache.spark.status.protobuf.TaskData> tasks = 45;
containsValue(Object) - Method in class org.apache.spark.sql.util.CaseInsensitiveStringMap
 
contentType() - Method in class org.apache.spark.ui.JettyUtils.ServletParams
 
context() - Method in interface org.apache.spark.api.java.JavaRDDLike
The SparkContext that this RDD was created on.
context() - Method in class org.apache.spark.ContextAwareIterator
 
context() - Method in class org.apache.spark.InterruptibleIterator
 
context() - Method in class org.apache.spark.rdd.RDD
The SparkContext that this RDD was created on.
context() - Method in exception org.apache.spark.sql.AnalysisException
 
context() - Method in interface org.apache.spark.streaming.api.java.JavaDStreamLike
Return the StreamingContext associated with this DStream
context() - Method in class org.apache.spark.streaming.dstream.DStream
Return the StreamingContext associated with this DStream
ContextAwareIterator<T> - Class in org.apache.spark
:: DeveloperApi :: A TaskContext aware iterator.
ContextAwareIterator(TaskContext, Iterator<T>) - Constructor for class org.apache.spark.ContextAwareIterator
 
ContextBarrierId - Class in org.apache.spark
For each barrier stage attempt, only at most one barrier() call can be active at any time, thus we can use (stageId, stageAttemptId) to identify the stage attempt where the barrier() call is from.
ContextBarrierId(int, int) - Constructor for class org.apache.spark.ContextBarrierId
 
Continuous() - Static method in class org.apache.spark.mllib.tree.configuration.FeatureType
 
Continuous(long) - Static method in class org.apache.spark.sql.streaming.Trigger
A trigger that continuously processes streaming data, asynchronously checkpointing at the specified interval.
Continuous(long, TimeUnit) - Static method in class org.apache.spark.sql.streaming.Trigger
A trigger that continuously processes streaming data, asynchronously checkpointing at the specified interval.
Continuous(Duration) - Static method in class org.apache.spark.sql.streaming.Trigger
(Scala-friendly) A trigger that continuously processes streaming data, asynchronously checkpointing at the specified interval.
Continuous(String) - Static method in class org.apache.spark.sql.streaming.Trigger
A trigger that continuously processes streaming data, asynchronously checkpointing at the specified interval.
ContinuousPartitionReader<T> - Interface in org.apache.spark.sql.connector.read.streaming
A variation on PartitionReader for use with continuous streaming processing.
ContinuousPartitionReaderFactory - Interface in org.apache.spark.sql.connector.read.streaming
A variation on PartitionReaderFactory that returns ContinuousPartitionReader instead of PartitionReader.
continuousProcessingUnsupportedByDataSourceError(String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
ContinuousSplit - Class in org.apache.spark.ml.tree
Split which tests a continuous feature.
ContinuousStream - Interface in org.apache.spark.sql.connector.read.streaming
A SparkDataStream for streaming queries with continuous mode.
conv(Column, int, int) - Static method in class org.apache.spark.sql.functions
Convert a number in a string column from one base to another.
convertCachedBatchToColumnarBatch(RDD<CachedBatch>, Seq<Attribute>, Seq<Attribute>, SQLConf) - Method in interface org.apache.spark.sql.columnar.CachedBatchSerializer
Convert the cached data into a ColumnarBatch.
convertCachedBatchToInternalRow(RDD<CachedBatch>, Seq<Attribute>, Seq<Attribute>, SQLConf) - Method in interface org.apache.spark.sql.columnar.CachedBatchSerializer
Convert the cached batch into InternalRows.
convertColumnarBatchToCachedBatch(RDD<ColumnarBatch>, Seq<Attribute>, StorageLevel, SQLConf) - Method in interface org.apache.spark.sql.columnar.CachedBatchSerializer
Convert an RDD[ColumnarBatch] into an RDD[CachedBatch] in preparation for caching the data.
convertHiveTableToCatalogTableError(SparkException, String, String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
convertInternalRowToCachedBatch(RDD<InternalRow>, Seq<Attribute>, StorageLevel, SQLConf) - Method in interface org.apache.spark.sql.columnar.CachedBatchSerializer
Convert an RDD[InternalRow] into an RDD[CachedBatch] in preparation for caching the data.
convertMatrixColumnsFromML(Dataset<?>, String...) - Static method in class org.apache.spark.mllib.util.MLUtils
Converts matrix columns in an input Dataset to the Matrix type from the new Matrix type under the spark.ml package.
convertMatrixColumnsFromML(Dataset<?>, Seq<String>) - Static method in class org.apache.spark.mllib.util.MLUtils
Converts matrix columns in an input Dataset to the Matrix type from the new Matrix type under the spark.ml package.
convertMatrixColumnsToML(Dataset<?>, String...) - Static method in class org.apache.spark.mllib.util.MLUtils
Converts Matrix columns in an input Dataset from the Matrix type to the new Matrix type under the spark.ml package.
convertMatrixColumnsToML(Dataset<?>, Seq<String>) - Static method in class org.apache.spark.mllib.util.MLUtils
Converts Matrix columns in an input Dataset from the Matrix type to the new Matrix type under the spark.ml package.
convertTableProperties(TableSpec) - Static method in class org.apache.spark.sql.connector.catalog.CatalogV2Util
 
convertToCanonicalEdges(Function2<ED, ED, ED>) - Method in class org.apache.spark.graphx.GraphOps
Convert bi-directional edges into uni-directional ones.
convertToOldLossType(String) - Method in interface org.apache.spark.ml.tree.GBTRegressorParams
 
convertToTimeUnit(long, TimeUnit) - Static method in class org.apache.spark.streaming.ui.UIUtils
Convert milliseconds to the specified unit.
convertTransforms() - Method in class org.apache.spark.sql.connector.catalog.CatalogV2Implicits.TransformHelper
 
convertVectorColumnsFromML(Dataset<?>, String...) - Static method in class org.apache.spark.mllib.util.MLUtils
Converts vector columns in an input Dataset to the Vector type from the new Vector type under the spark.ml package.
convertVectorColumnsFromML(Dataset<?>, Seq<String>) - Static method in class org.apache.spark.mllib.util.MLUtils
Converts vector columns in an input Dataset to the Vector type from the new Vector type under the spark.ml package.
convertVectorColumnsToML(Dataset<?>, String...) - Static method in class org.apache.spark.mllib.util.MLUtils
Converts vector columns in an input Dataset from the Vector type to the new Vector type under the spark.ml package.
convertVectorColumnsToML(Dataset<?>, Seq<String>) - Static method in class org.apache.spark.mllib.util.MLUtils
Converts vector columns in an input Dataset from the Vector type to the new Vector type under the spark.ml package.
Coord() - Constructor for class org.apache.spark.sql.util.NumericHistogram.Coord
 
CoordinateMatrix - Class in org.apache.spark.mllib.linalg.distributed
Represents a matrix in coordinate format.
CoordinateMatrix(RDD<MatrixEntry>, long, long) - Constructor for class org.apache.spark.mllib.linalg.distributed.CoordinateMatrix
 
CoordinateMatrix(RDD<MatrixEntry>) - Constructor for class org.apache.spark.mllib.linalg.distributed.CoordinateMatrix
Alternative constructor leaving matrix dimensions to be determined automatically.
copy(ParamMap) - Method in class org.apache.spark.ml.classification.DecisionTreeClassificationModel
 
copy(ParamMap) - Method in class org.apache.spark.ml.classification.DecisionTreeClassifier
 
copy(ParamMap) - Method in class org.apache.spark.ml.classification.FMClassificationModel
 
copy(ParamMap) - Method in class org.apache.spark.ml.classification.FMClassifier
 
copy(ParamMap) - Method in class org.apache.spark.ml.classification.GBTClassificationModel
 
copy(ParamMap) - Method in class org.apache.spark.ml.classification.GBTClassifier
 
copy(ParamMap) - Method in class org.apache.spark.ml.classification.LinearSVC
 
copy(ParamMap) - Method in class org.apache.spark.ml.classification.LinearSVCModel
 
copy(ParamMap) - Method in class org.apache.spark.ml.classification.LogisticRegression
 
copy(ParamMap) - Method in class org.apache.spark.ml.classification.LogisticRegressionModel
 
copy(ParamMap) - Method in class org.apache.spark.ml.classification.MultilayerPerceptronClassificationModel
 
copy(ParamMap) - Method in class org.apache.spark.ml.classification.MultilayerPerceptronClassifier
 
copy(ParamMap) - Method in class org.apache.spark.ml.classification.NaiveBayes
 
copy(ParamMap) - Method in class org.apache.spark.ml.classification.NaiveBayesModel
 
copy(ParamMap) - Method in class org.apache.spark.ml.classification.OneVsRest
 
copy(ParamMap) - Method in class org.apache.spark.ml.classification.OneVsRestModel
 
copy(ParamMap) - Method in class org.apache.spark.ml.classification.RandomForestClassificationModel
 
copy(ParamMap) - Method in class org.apache.spark.ml.classification.RandomForestClassifier
 
copy(ParamMap) - Method in class org.apache.spark.ml.clustering.BisectingKMeans
 
copy(ParamMap) - Method in class org.apache.spark.ml.clustering.BisectingKMeansModel
 
copy(ParamMap) - Method in class org.apache.spark.ml.clustering.DistributedLDAModel
 
copy(ParamMap) - Method in class org.apache.spark.ml.clustering.GaussianMixture
 
copy(ParamMap) - Method in class org.apache.spark.ml.clustering.GaussianMixtureModel
 
copy(ParamMap) - Method in class org.apache.spark.ml.clustering.KMeans
 
copy(ParamMap) - Method in class org.apache.spark.ml.clustering.KMeansModel
 
copy(ParamMap) - Method in class org.apache.spark.ml.clustering.LDA
 
copy(ParamMap) - Method in class org.apache.spark.ml.clustering.LocalLDAModel
 
copy(ParamMap) - Method in class org.apache.spark.ml.clustering.PowerIterationClustering
 
copy(ParamMap) - Method in class org.apache.spark.ml.Estimator
 
copy(ParamMap) - Method in class org.apache.spark.ml.evaluation.BinaryClassificationEvaluator
 
copy(ParamMap) - Method in class org.apache.spark.ml.evaluation.ClusteringEvaluator
 
copy(ParamMap) - Method in class org.apache.spark.ml.evaluation.Evaluator
 
copy(ParamMap) - Method in class org.apache.spark.ml.evaluation.MulticlassClassificationEvaluator
 
copy(ParamMap) - Method in class org.apache.spark.ml.evaluation.MultilabelClassificationEvaluator
 
copy(ParamMap) - Method in class org.apache.spark.ml.evaluation.RankingEvaluator
 
copy(ParamMap) - Method in class org.apache.spark.ml.evaluation.RegressionEvaluator
 
copy(ParamMap) - Method in class org.apache.spark.ml.feature.Binarizer
 
copy(ParamMap) - Method in class org.apache.spark.ml.feature.BucketedRandomProjectionLSH
 
copy(ParamMap) - Method in class org.apache.spark.ml.feature.BucketedRandomProjectionLSHModel
 
copy(ParamMap) - Method in class org.apache.spark.ml.feature.Bucketizer
 
copy(ParamMap) - Method in class org.apache.spark.ml.feature.ChiSqSelector
Deprecated.
 
copy(ParamMap) - Method in class org.apache.spark.ml.feature.ChiSqSelectorModel
 
copy(ParamMap) - Method in class org.apache.spark.ml.feature.ColumnPruner
 
copy(ParamMap) - Method in class org.apache.spark.ml.feature.CountVectorizer
 
copy(ParamMap) - Method in class org.apache.spark.ml.feature.CountVectorizerModel
 
copy(ParamMap) - Method in class org.apache.spark.ml.feature.FeatureHasher
 
copy(ParamMap) - Method in class org.apache.spark.ml.feature.HashingTF
 
copy(ParamMap) - Method in class org.apache.spark.ml.feature.IDF
 
copy(ParamMap) - Method in class org.apache.spark.ml.feature.IDFModel
 
copy(ParamMap) - Method in class org.apache.spark.ml.feature.Imputer
 
copy(ParamMap) - Method in class org.apache.spark.ml.feature.ImputerModel
 
copy(ParamMap) - Method in class org.apache.spark.ml.feature.IndexToString
 
copy(ParamMap) - Method in class org.apache.spark.ml.feature.Interaction
 
copy(ParamMap) - Method in class org.apache.spark.ml.feature.MaxAbsScaler
 
copy(ParamMap) - Method in class org.apache.spark.ml.feature.MaxAbsScalerModel
 
copy(ParamMap) - Method in class org.apache.spark.ml.feature.MinHashLSH
 
copy(ParamMap) - Method in class org.apache.spark.ml.feature.MinHashLSHModel
 
copy(ParamMap) - Method in class org.apache.spark.ml.feature.MinMaxScaler
 
copy(ParamMap) - Method in class org.apache.spark.ml.feature.MinMaxScalerModel
 
copy(ParamMap) - Method in class org.apache.spark.ml.feature.OneHotEncoder
 
copy(ParamMap) - Method in class org.apache.spark.ml.feature.OneHotEncoderModel
 
copy(ParamMap) - Method in class org.apache.spark.ml.feature.PCA
 
copy(ParamMap) - Method in class org.apache.spark.ml.feature.PCAModel
 
copy(ParamMap) - Method in class org.apache.spark.ml.feature.PolynomialExpansion
 
copy(ParamMap) - Method in class org.apache.spark.ml.feature.QuantileDiscretizer
 
copy(ParamMap) - Method in class org.apache.spark.ml.feature.RegexTokenizer
 
copy(ParamMap) - Method in class org.apache.spark.ml.feature.RFormula
 
copy(ParamMap) - Method in class org.apache.spark.ml.feature.RFormulaModel
 
copy(ParamMap) - Method in class org.apache.spark.ml.feature.RobustScaler
 
copy(ParamMap) - Method in class org.apache.spark.ml.feature.RobustScalerModel
 
copy(ParamMap) - Method in class org.apache.spark.ml.feature.SQLTransformer
 
copy(ParamMap) - Method in class org.apache.spark.ml.feature.StandardScaler
 
copy(ParamMap) - Method in class org.apache.spark.ml.feature.StandardScalerModel
 
copy(ParamMap) - Method in class org.apache.spark.ml.feature.StopWordsRemover
 
copy(ParamMap) - Method in class org.apache.spark.ml.feature.StringIndexer
 
copy(ParamMap) - Method in class org.apache.spark.ml.feature.StringIndexerModel
 
copy(ParamMap) - Method in class org.apache.spark.ml.feature.Tokenizer
 
copy(ParamMap) - Method in class org.apache.spark.ml.feature.UnivariateFeatureSelector
 
copy(ParamMap) - Method in class org.apache.spark.ml.feature.UnivariateFeatureSelectorModel
 
copy(ParamMap) - Method in class org.apache.spark.ml.feature.VarianceThresholdSelector
 
copy(ParamMap) - Method in class org.apache.spark.ml.feature.VarianceThresholdSelectorModel
 
copy(ParamMap) - Method in class org.apache.spark.ml.feature.VectorAssembler
 
copy(ParamMap) - Method in class org.apache.spark.ml.feature.VectorAttributeRewriter
 
copy(ParamMap) - Method in class org.apache.spark.ml.feature.VectorIndexer
 
copy(ParamMap) - Method in class org.apache.spark.ml.feature.VectorIndexerModel
 
copy(ParamMap) - Method in class org.apache.spark.ml.feature.VectorSizeHint
 
copy(ParamMap) - Method in class org.apache.spark.ml.feature.VectorSlicer
 
copy(ParamMap) - Method in class org.apache.spark.ml.feature.Word2Vec
 
copy(ParamMap) - Method in class org.apache.spark.ml.feature.Word2VecModel
 
copy(ParamMap) - Method in class org.apache.spark.ml.fpm.FPGrowth
 
copy(ParamMap) - Method in class org.apache.spark.ml.fpm.FPGrowthModel
 
copy(ParamMap) - Method in class org.apache.spark.ml.fpm.PrefixSpan
 
copy(Vector, Vector) - Static method in class org.apache.spark.ml.linalg.BLAS
y = x
copy() - Method in class org.apache.spark.ml.linalg.DenseMatrix
 
copy() - Method in class org.apache.spark.ml.linalg.DenseVector
 
copy() - Method in interface org.apache.spark.ml.linalg.Matrix
Get a deep copy of the matrix.
copy() - Method in class org.apache.spark.ml.linalg.SparseMatrix
 
copy() - Method in class org.apache.spark.ml.linalg.SparseVector
 
copy() - Method in interface org.apache.spark.ml.linalg.Vector
Makes a deep copy of this vector.
copy(ParamMap) - Method in class org.apache.spark.ml.Model
 
copy() - Method in class org.apache.spark.ml.param.ParamMap
Creates a copy of this param map.
copy(ParamMap) - Method in interface org.apache.spark.ml.param.Params
Creates a copy of this instance with the same UID and some extra params.
copy(ParamMap) - Method in class org.apache.spark.ml.Pipeline
 
copy(ParamMap) - Method in class org.apache.spark.ml.PipelineModel
 
copy(ParamMap) - Method in class org.apache.spark.ml.PipelineStage
 
copy(ParamMap) - Method in class org.apache.spark.ml.Predictor
 
copy(ParamMap) - Method in class org.apache.spark.ml.recommendation.ALS
 
copy(ParamMap) - Method in class org.apache.spark.ml.recommendation.ALSModel
 
copy(ParamMap) - Method in class org.apache.spark.ml.regression.AFTSurvivalRegression
 
copy(ParamMap) - Method in class org.apache.spark.ml.regression.AFTSurvivalRegressionModel
 
copy(ParamMap) - Method in class org.apache.spark.ml.regression.DecisionTreeRegressionModel
 
copy(ParamMap) - Method in class org.apache.spark.ml.regression.DecisionTreeRegressor
 
copy(ParamMap) - Method in class org.apache.spark.ml.regression.FMRegressionModel
 
copy(ParamMap) - Method in class org.apache.spark.ml.regression.FMRegressor
 
copy(ParamMap) - Method in class org.apache.spark.ml.regression.GBTRegressionModel
 
copy(ParamMap) - Method in class org.apache.spark.ml.regression.GBTRegressor
 
copy(ParamMap) - Method in class org.apache.spark.ml.regression.GeneralizedLinearRegression
 
copy(ParamMap) - Method in class org.apache.spark.ml.regression.GeneralizedLinearRegressionModel
 
copy(ParamMap) - Method in class org.apache.spark.ml.regression.IsotonicRegression
 
copy(ParamMap) - Method in class org.apache.spark.ml.regression.IsotonicRegressionModel
 
copy(ParamMap) - Method in class org.apache.spark.ml.regression.LinearRegression
 
copy(ParamMap) - Method in class org.apache.spark.ml.regression.LinearRegressionModel
 
copy(ParamMap) - Method in class org.apache.spark.ml.regression.RandomForestRegressionModel
 
copy(ParamMap) - Method in class org.apache.spark.ml.regression.RandomForestRegressor
 
copy(ParamMap) - Method in class org.apache.spark.ml.Transformer
 
copy(ParamMap) - Method in class org.apache.spark.ml.tuning.CrossValidator
 
copy(ParamMap) - Method in class org.apache.spark.ml.tuning.CrossValidatorModel
 
copy(ParamMap) - Method in class org.apache.spark.ml.tuning.TrainValidationSplit
 
copy(ParamMap) - Method in class org.apache.spark.ml.tuning.TrainValidationSplitModel
 
copy(ParamMap) - Method in class org.apache.spark.ml.UnaryTransformer
 
copy(Vector, Vector) - Static method in class org.apache.spark.mllib.linalg.BLAS
y = x
copy() - Method in class org.apache.spark.mllib.linalg.DenseMatrix
 
copy() - Method in class org.apache.spark.mllib.linalg.DenseVector
 
copy() - Method in interface org.apache.spark.mllib.linalg.Matrix
Get a deep copy of the matrix.
copy() - Method in class org.apache.spark.mllib.linalg.SparseMatrix
 
copy() - Method in class org.apache.spark.mllib.linalg.SparseVector
 
copy() - Method in interface org.apache.spark.mllib.linalg.Vector
Makes a deep copy of this vector.
copy() - Method in class org.apache.spark.mllib.random.ExponentialGenerator
 
copy() - Method in class org.apache.spark.mllib.random.GammaGenerator
 
copy() - Method in class org.apache.spark.mllib.random.LogNormalGenerator
 
copy() - Method in class org.apache.spark.mllib.random.PoissonGenerator
 
copy() - Method in interface org.apache.spark.mllib.random.RandomDataGenerator
Returns a copy of the RandomDataGenerator with a new instance of the rng object used in the class when applicable for non-locking concurrent usage.
copy() - Method in class org.apache.spark.mllib.random.StandardNormalGenerator
 
copy() - Method in class org.apache.spark.mllib.random.UniformGenerator
 
copy() - Method in class org.apache.spark.mllib.random.WeibullGenerator
 
copy() - Method in class org.apache.spark.mllib.tree.configuration.Strategy
Returns a shallow copy of this instance.
copy(String, Option<Object>, Option<Object>, Option<LogicalPlan>, Option<Throwable>, Option<String>, Map<String, String>, QueryContext[]) - Method in exception org.apache.spark.sql.AnalysisException
 
copy() - Method in interface org.apache.spark.sql.Row
Make a copy of the current Row object.
copy() - Static method in class org.apache.spark.sql.sources.AlwaysFalse
 
copy() - Static method in class org.apache.spark.sql.sources.AlwaysTrue
 
copy() - Method in class org.apache.spark.sql.util.MapperRowCounter
 
copy() - Method in class org.apache.spark.sql.vectorized.ColumnarArray
 
copy() - Method in class org.apache.spark.sql.vectorized.ColumnarBatchRow
 
copy() - Method in class org.apache.spark.sql.vectorized.ColumnarMap
 
copy() - Method in class org.apache.spark.sql.vectorized.ColumnarRow
Revisit this.
copy() - Method in class org.apache.spark.util.AccumulatorV2
Creates a new copy of this accumulator.
copy() - Method in class org.apache.spark.util.CollectionAccumulator
 
copy() - Method in class org.apache.spark.util.DoubleAccumulator
 
copy() - Method in class org.apache.spark.util.LongAccumulator
 
copy() - Method in class org.apache.spark.util.StatCounter
Clone this StatCounter
copyAndReset() - Method in class org.apache.spark.sql.util.MapperRowCounter
 
copyAndReset() - Method in class org.apache.spark.util.AccumulatorV2
Creates a new copy of this accumulator, which is zero value.
copyAndReset() - Method in class org.apache.spark.util.CollectionAccumulator
 
copyFileStreamNIO(FileChannel, WritableByteChannel, long, long) - Static method in class org.apache.spark.util.Utils
 
copyStream(InputStream, OutputStream, boolean, boolean) - Static method in class org.apache.spark.util.Utils
Copy all data from an InputStream to an OutputStream.
copyStreamUpTo(InputStream, long) - Static method in class org.apache.spark.util.Utils
Copy the first maxSize bytes of data from the InputStream to an in-memory buffer, primarily to check for corruption.
copyValues(T, ParamMap) - Method in interface org.apache.spark.ml.param.Params
Copies param values from this instance to another instance for params shared by them.
cores(int) - Method in class org.apache.spark.resource.ExecutorResourceRequests
Specify number of cores per Executor.
CORES() - Static method in class org.apache.spark.resource.ResourceProfile
built-in executor resource: cores
cores() - Method in class org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.RegisterExecutor
 
cores() - Method in class org.apache.spark.scheduler.MiscellaneousProcessDetails
 
CORES_GRANTED_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo
 
CORES_PER_EXECUTOR_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo
 
coresGranted() - Method in class org.apache.spark.status.api.v1.ApplicationInfo
 
coresPerExecutor() - Method in class org.apache.spark.status.api.v1.ApplicationInfo
 
corr(Dataset<?>, String, String) - Static method in class org.apache.spark.ml.stat.Correlation
Compute the correlation matrix for the input Dataset of Vectors using the specified method.
corr(Dataset<?>, String) - Static method in class org.apache.spark.ml.stat.Correlation
Compute the Pearson correlation matrix for the input Dataset of Vectors.
corr(RDD<Object>, RDD<Object>, String) - Static method in class org.apache.spark.mllib.stat.correlation.Correlations
 
corr(RDD<Vector>) - Static method in class org.apache.spark.mllib.stat.Statistics
Compute the Pearson correlation matrix for the input RDD of Vectors.
corr(RDD<Vector>, String) - Static method in class org.apache.spark.mllib.stat.Statistics
Compute the correlation matrix for the input RDD of Vectors using the specified method.
corr(RDD<Object>, RDD<Object>) - Static method in class org.apache.spark.mllib.stat.Statistics
Compute the Pearson correlation for the input RDDs.
corr(JavaRDD<Double>, JavaRDD<Double>) - Static method in class org.apache.spark.mllib.stat.Statistics
Java-friendly version of corr()
corr(RDD<Object>, RDD<Object>, String) - Static method in class org.apache.spark.mllib.stat.Statistics
Compute the correlation for the input RDDs using the specified method.
corr(JavaRDD<Double>, JavaRDD<Double>, String) - Static method in class org.apache.spark.mllib.stat.Statistics
Java-friendly version of corr()
corr(String, String, String) - Method in class org.apache.spark.sql.DataFrameStatFunctions
Calculates the correlation of two columns of a DataFrame.
corr(String, String) - Method in class org.apache.spark.sql.DataFrameStatFunctions
Calculates the Pearson Correlation Coefficient of two columns of a DataFrame.
corr(Column, Column) - Static method in class org.apache.spark.sql.functions
Aggregate function: returns the Pearson Correlation Coefficient for two columns.
corr(String, String) - Static method in class org.apache.spark.sql.functions
Aggregate function: returns the Pearson Correlation Coefficient for two columns.
Correlation - Class in org.apache.spark.ml.stat
API for correlation functions in MLlib, compatible with DataFrames and Datasets.
Correlation() - Constructor for class org.apache.spark.ml.stat.Correlation
 
Correlation - Interface in org.apache.spark.mllib.stat.correlation
Trait for correlation algorithms.
CorrelationNames - Class in org.apache.spark.mllib.stat.correlation
Maintains supported and default correlation names.
CorrelationNames() - Constructor for class org.apache.spark.mllib.stat.correlation.CorrelationNames
 
Correlations - Class in org.apache.spark.mllib.stat.correlation
Delegates computation to the specific correlation object based on the input method name.
Correlations() - Constructor for class org.apache.spark.mllib.stat.correlation.Correlations
 
corrMatrix(RDD<Vector>, String) - Static method in class org.apache.spark.mllib.stat.correlation.Correlations
 
CORRUPT_MERGED_BLOCK_CHUNKS() - Method in class org.apache.spark.InternalAccumulator.shuffleRead$
 
CORRUPT_MERGED_BLOCK_CHUNKS_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions
 
CORRUPT_MERGED_BLOCK_CHUNKS_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics
 
corruptedTableNameContextInCatalogError(int, int) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
corruptedViewQueryOutputColumnsInCatalogError(String, int) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
corruptedViewReferredTempFunctionsInCatalogError(Exception) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
corruptedViewReferredTempViewInCatalogError(Exception) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
corruptedViewSQLConfigsInCatalogError(Exception) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
corruptMergedBlockChunks() - Method in class org.apache.spark.status.api.v1.ShufflePushReadMetricDistributions
 
corruptMergedBlockChunks() - Method in class org.apache.spark.status.api.v1.ShufflePushReadMetrics
 
cos(Column) - Static method in class org.apache.spark.sql.functions
 
cos(String) - Static method in class org.apache.spark.sql.functions
 
cosh(Column) - Static method in class org.apache.spark.sql.functions
 
cosh(String) - Static method in class org.apache.spark.sql.functions
 
CosineSilhouette - Class in org.apache.spark.ml.evaluation
The algorithm which is implemented in this object, instead, is an efficient and parallel implementation of the Silhouette using the cosine distance measure.
CosineSilhouette() - Constructor for class org.apache.spark.ml.evaluation.CosineSilhouette
 
costSum() - Method in class org.apache.spark.ml.clustering.KMeansAggregator
 
cot(Column) - Static method in class org.apache.spark.sql.functions
 
count() - Method in interface org.apache.spark.api.java.JavaRDDLike
Return the number of elements in the RDD.
count() - Method in class org.apache.spark.graphx.impl.EdgeRDDImpl
The number of edges in the RDD.
count() - Method in class org.apache.spark.graphx.impl.VertexRDDImpl
The number of vertices in the RDD.
count() - Method in class org.apache.spark.ml.clustering.ExpectationAggregator
 
count() - Method in class org.apache.spark.ml.clustering.KMeansAggregator
 
count(Column, Column) - Static method in class org.apache.spark.ml.stat.Summarizer
 
count(Column) - Static method in class org.apache.spark.ml.stat.Summarizer
 
count() - Method in class org.apache.spark.mllib.stat.MultivariateOnlineSummarizer
Sample size.
count() - Method in interface org.apache.spark.mllib.stat.MultivariateStatisticalSummary
Sample size.
count() - Method in class org.apache.spark.rdd.RDD
Return the number of elements in the RDD.
Count - Class in org.apache.spark.sql.connector.expressions.aggregate
An aggregate function that returns the number of the specific row in a group.
Count(Expression, boolean) - Constructor for class org.apache.spark.sql.connector.expressions.aggregate.Count
 
count() - Method in class org.apache.spark.sql.Dataset
Returns the number of rows in the Dataset.
count(MapFunction<T, Object>) - Static method in class org.apache.spark.sql.expressions.javalang.typed
Deprecated.
Count aggregate function.
count(Function1<IN, Object>) - Static method in class org.apache.spark.sql.expressions.scalalang.typed
Deprecated.
Count aggregate function.
count(Column) - Static method in class org.apache.spark.sql.functions
Aggregate function: returns the number of items in a group.
count(String) - Static method in class org.apache.spark.sql.functions
Aggregate function: returns the number of items in a group.
count() - Method in class org.apache.spark.sql.KeyValueGroupedDataset
Returns a Dataset that contains a tuple with each key and the number of items present for that key.
count() - Method in class org.apache.spark.sql.RelationalGroupedDataset
Count the number of rows for each group.
count(KVStoreView<T>, Function1<T, Object>) - Static method in class org.apache.spark.status.KVUtils
Counts the number of elements in the KVStoreView which satisfy a predicate.
count() - Method in class org.apache.spark.status.RDDPartitionSeq
 
count() - Method in class org.apache.spark.storage.ReadableChannelFileRegion
 
count() - Method in interface org.apache.spark.streaming.api.java.JavaDStreamLike
Return a new DStream in which each RDD has a single element generated by counting each RDD of this DStream.
count() - Method in class org.apache.spark.streaming.dstream.DStream
Return a new DStream in which each RDD has a single element generated by counting each RDD of this DStream.
count() - Method in class org.apache.spark.util.DoubleAccumulator
Returns the number of elements added to the accumulator.
count() - Method in class org.apache.spark.util.LongAccumulator
Returns the number of elements added to the accumulator.
count() - Method in class org.apache.spark.util.StatCounter
 
count_distinct(Column, Column...) - Static method in class org.apache.spark.sql.functions
Aggregate function: returns the number of distinct items in a group.
count_distinct(Column, Seq<Column>) - Static method in class org.apache.spark.sql.functions
Aggregate function: returns the number of distinct items in a group.
countApprox(long, double) - Method in interface org.apache.spark.api.java.JavaRDDLike
Approximate version of count() that returns a potentially incomplete result within a timeout, even if not all tasks have finished.
countApprox(long) - Method in interface org.apache.spark.api.java.JavaRDDLike
Approximate version of count() that returns a potentially incomplete result within a timeout, even if not all tasks have finished.
countApprox(long, double) - Method in class org.apache.spark.rdd.RDD
Approximate version of count() that returns a potentially incomplete result within a timeout, even if not all tasks have finished.
countApproxDistinct(double) - Method in interface org.apache.spark.api.java.JavaRDDLike
Return approximate number of distinct elements in the RDD.
countApproxDistinct(int, int) - Method in class org.apache.spark.rdd.RDD
Return approximate number of distinct elements in the RDD.
countApproxDistinct(double) - Method in class org.apache.spark.rdd.RDD
Return approximate number of distinct elements in the RDD.
countApproxDistinctByKey(double, Partitioner) - Method in class org.apache.spark.api.java.JavaPairRDD
Return approximate number of distinct values for each key in this RDD.
countApproxDistinctByKey(double, int) - Method in class org.apache.spark.api.java.JavaPairRDD
Return approximate number of distinct values for each key in this RDD.
countApproxDistinctByKey(double) - Method in class org.apache.spark.api.java.JavaPairRDD
Return approximate number of distinct values for each key in this RDD.
countApproxDistinctByKey(int, int, Partitioner) - Method in class org.apache.spark.rdd.PairRDDFunctions
Return approximate number of distinct values for each key in this RDD.
countApproxDistinctByKey(double, Partitioner) - Method in class org.apache.spark.rdd.PairRDDFunctions
Return approximate number of distinct values for each key in this RDD.
countApproxDistinctByKey(double, int) - Method in class org.apache.spark.rdd.PairRDDFunctions
Return approximate number of distinct values for each key in this RDD.
countApproxDistinctByKey(double) - Method in class org.apache.spark.rdd.PairRDDFunctions
Return approximate number of distinct values for each key in this RDD.
countAsync() - Method in interface org.apache.spark.api.java.JavaRDDLike
The asynchronous version of count, which returns a future for counting the number of elements in this RDD.
countAsync() - Method in class org.apache.spark.rdd.AsyncRDDActions
Returns a future for counting the number of elements in the RDD.
countByKey() - Method in class org.apache.spark.api.java.JavaPairRDD
Count the number of elements for each key, and return the result to the master as a Map.
countByKey() - Method in class org.apache.spark.rdd.PairRDDFunctions
Count the number of elements for each key, collecting the results to a local Map.
countByKeyApprox(long) - Method in class org.apache.spark.api.java.JavaPairRDD
Approximate version of countByKey that can return a partial result if it does not finish within a timeout.
countByKeyApprox(long, double) - Method in class org.apache.spark.api.java.JavaPairRDD
Approximate version of countByKey that can return a partial result if it does not finish within a timeout.
countByKeyApprox(long, double) - Method in class org.apache.spark.rdd.PairRDDFunctions
Approximate version of countByKey that can return a partial result if it does not finish within a timeout.
countByValue() - Method in interface org.apache.spark.api.java.JavaRDDLike
Return the count of each unique value in this RDD as a map of (value, count) pairs.
countByValue(Ordering<T>) - Method in class org.apache.spark.rdd.RDD
Return the count of each unique value in this RDD as a local map of (value, count) pairs.
countByValue() - Method in interface org.apache.spark.streaming.api.java.JavaDStreamLike
Return a new DStream in which each RDD contains the counts of each distinct value in each RDD of this DStream.
countByValue(int) - Method in interface org.apache.spark.streaming.api.java.JavaDStreamLike
Return a new DStream in which each RDD contains the counts of each distinct value in each RDD of this DStream.
countByValue(int, Ordering<T>) - Method in class org.apache.spark.streaming.dstream.DStream
Return a new DStream in which each RDD contains the counts of each distinct value in each RDD of this DStream.
countByValueAndWindow(Duration, Duration) - Method in interface org.apache.spark.streaming.api.java.JavaDStreamLike
Return a new DStream in which each RDD contains the count of distinct elements in RDDs in a sliding window over this DStream.
countByValueAndWindow(Duration, Duration, int) - Method in interface org.apache.spark.streaming.api.java.JavaDStreamLike
Return a new DStream in which each RDD contains the count of distinct elements in RDDs in a sliding window over this DStream.
countByValueAndWindow(Duration, Duration, int, Ordering<T>) - Method in class org.apache.spark.streaming.dstream.DStream
Return a new DStream in which each RDD contains the count of distinct elements in RDDs in a sliding window over this DStream.
countByValueApprox(long, double) - Method in interface org.apache.spark.api.java.JavaRDDLike
Approximate version of countByValue().
countByValueApprox(long) - Method in interface org.apache.spark.api.java.JavaRDDLike
Approximate version of countByValue().
countByValueApprox(long, double, Ordering<T>) - Method in class org.apache.spark.rdd.RDD
Approximate version of countByValue().
countByValueApproxNotSupportArraysError() - Static method in class org.apache.spark.errors.SparkCoreErrors
 
countByWindow(Duration, Duration) - Method in interface org.apache.spark.streaming.api.java.JavaDStreamLike
Return a new DStream in which each RDD has a single element generated by counting the number of elements in a window over this DStream.
countByWindow(Duration, Duration) - Method in class org.apache.spark.streaming.dstream.DStream
Return a new DStream in which each RDD has a single element generated by counting the number of elements in a sliding window over this DStream.
countDistinct(Column, Column...) - Static method in class org.apache.spark.sql.functions
Aggregate function: returns the number of distinct items in a group.
countDistinct(String, String...) - Static method in class org.apache.spark.sql.functions
Aggregate function: returns the number of distinct items in a group.
countDistinct(Column, Seq<Column>) - Static method in class org.apache.spark.sql.functions
Aggregate function: returns the number of distinct items in a group.
countDistinct(String, Seq<String>) - Static method in class org.apache.spark.sql.functions
Aggregate function: returns the number of distinct items in a group.
COUNTER() - Static method in class org.apache.spark.metrics.sink.StatsdMetricType
 
CountingWritableChannel - Class in org.apache.spark.storage
 
CountingWritableChannel(WritableByteChannel) - Constructor for class org.apache.spark.storage.CountingWritableChannel
 
countMinSketch(String, int, int, int) - Method in class org.apache.spark.sql.DataFrameStatFunctions
Builds a Count-min Sketch over a specified column.
countMinSketch(String, double, double, int) - Method in class org.apache.spark.sql.DataFrameStatFunctions
Builds a Count-min Sketch over a specified column.
countMinSketch(Column, int, int, int) - Method in class org.apache.spark.sql.DataFrameStatFunctions
Builds a Count-min Sketch over a specified column.
countMinSketch(Column, double, double, int) - Method in class org.apache.spark.sql.DataFrameStatFunctions
Builds a Count-min Sketch over a specified column.
CountMinSketch - Class in org.apache.spark.util.sketch
A Count-min sketch is a probabilistic data structure used for cardinality estimation using sub-linear space.
CountMinSketch() - Constructor for class org.apache.spark.util.sketch.CountMinSketch
 
CountMinSketch.Version - Enum in org.apache.spark.util.sketch
 
CountStar - Class in org.apache.spark.sql.connector.expressions.aggregate
An aggregate function that returns the number of rows in a group.
CountStar() - Constructor for class org.apache.spark.sql.connector.expressions.aggregate.CountStar
 
countTowardsTaskFailures() - Method in class org.apache.spark.ExecutorLostFailure
 
countTowardsTaskFailures() - Method in class org.apache.spark.FetchFailed
Fetch failures lead to a different failure handling path: (1) we don't abort the stage after 4 task failures, instead we immediately go back to the stage which generated the map output, and regenerate the missing data.
countTowardsTaskFailures() - Static method in class org.apache.spark.Resubmitted
 
countTowardsTaskFailures() - Method in class org.apache.spark.TaskCommitDenied
If a task failed because its attempt to commit was denied, do not count this failure towards failing the stage.
countTowardsTaskFailures() - Method in interface org.apache.spark.TaskFailedReason
Whether this task failure should be counted towards the maximum number of times the task is allowed to fail before the stage is aborted.
countTowardsTaskFailures() - Method in class org.apache.spark.TaskKilled
 
countTowardsTaskFailures() - Static method in class org.apache.spark.TaskResultLost
 
countTowardsTaskFailures() - Static method in class org.apache.spark.UnknownReason
 
CountVectorizer - Class in org.apache.spark.ml.feature
Extracts a vocabulary from document collections and generates a CountVectorizerModel.
CountVectorizer(String) - Constructor for class org.apache.spark.ml.feature.CountVectorizer
 
CountVectorizer() - Constructor for class org.apache.spark.ml.feature.CountVectorizer
 
CountVectorizerModel - Class in org.apache.spark.ml.feature
Converts a text document to a sparse vector of token counts.
CountVectorizerModel(String, String[]) - Constructor for class org.apache.spark.ml.feature.CountVectorizerModel
 
CountVectorizerModel(String[]) - Constructor for class org.apache.spark.ml.feature.CountVectorizerModel
 
CountVectorizerParams - Interface in org.apache.spark.ml.feature
cov() - Method in class org.apache.spark.ml.stat.distribution.MultivariateGaussian
 
cov(String, String) - Method in class org.apache.spark.sql.DataFrameStatFunctions
Calculate the sample covariance of two numerical columns of a DataFrame.
covar_pop(Column, Column) - Static method in class org.apache.spark.sql.functions
Aggregate function: returns the population covariance for two columns.
covar_pop(String, String) - Static method in class org.apache.spark.sql.functions
Aggregate function: returns the population covariance for two columns.
covar_samp(Column, Column) - Static method in class org.apache.spark.sql.functions
Aggregate function: returns the sample covariance for two columns.
covar_samp(String, String) - Static method in class org.apache.spark.sql.functions
Aggregate function: returns the sample covariance for two columns.
covs() - Method in class org.apache.spark.ml.clustering.ExpectationAggregator
 
cpus() - Method in class org.apache.spark.BarrierTaskContext
 
CPUS() - Static method in class org.apache.spark.resource.ResourceProfile
built-in task resource: cpus
cpus(int) - Method in class org.apache.spark.resource.TaskResourceRequests
Specify number of cpus per Task.
cpus() - Method in class org.apache.spark.TaskContext
CPUs allocated to the task.
crc32(Column) - Static method in class org.apache.spark.sql.functions
Calculates the cyclic redundancy check value (CRC32) of a binary column and returns the value as a bigint.
CreatableRelationProvider - Interface in org.apache.spark.sql.sources
 
create(boolean, boolean, boolean, boolean, int) - Static method in class org.apache.spark.api.java.StorageLevels
Create a new StorageLevel object.
create(JavaSparkContext, JdbcRDD.ConnectionFactory, String, long, long, int, Function<ResultSet, T>) - Static method in class org.apache.spark.rdd.JdbcRDD
Create an RDD that executes a SQL query on a JDBC connection and reads results.
create(JavaSparkContext, JdbcRDD.ConnectionFactory, String, long, long, int) - Static method in class org.apache.spark.rdd.JdbcRDD
Create an RDD that executes a SQL query on a JDBC connection and reads results.
create(RDD<T>, Function1<Object, Object>) - Static method in class org.apache.spark.rdd.PartitionPruningRDD
Create a PartitionPruningRDD.
create(String, DataType) - Static method in interface org.apache.spark.sql.connector.catalog.Column
 
create(String, DataType, boolean) - Static method in interface org.apache.spark.sql.connector.catalog.Column
 
create(String, DataType, boolean, String, String) - Static method in interface org.apache.spark.sql.connector.catalog.Column
 
create(String, DataType, boolean, String, ColumnDefaultValue, String) - Static method in interface org.apache.spark.sql.connector.catalog.Column
 
create(String, DataType, boolean, String, String, String) - Static method in interface org.apache.spark.sql.connector.catalog.Column
 
create() - Method in interface org.apache.spark.sql.CreateTableWriter
Create a new table from the contents of the data frame.
create() - Method in class org.apache.spark.sql.DataFrameWriterV2
 
create(Object...) - Static method in class org.apache.spark.sql.RowFactory
Create a Row from the given arguments.
create(long) - Static method in class org.apache.spark.util.sketch.BloomFilter
Creates a BloomFilter with the expected number of insertions and a default expected false positive probability of 3%.
create(long, double) - Static method in class org.apache.spark.util.sketch.BloomFilter
Creates a BloomFilter with the expected number of insertions and expected false positive probability.
create(long, long) - Static method in class org.apache.spark.util.sketch.BloomFilter
Creates a BloomFilter with given expectedNumItems and numBits, it will pick an optimal numHashFunctions which can minimize fpp for the bloom filter.
create(int, int, int) - Static method in class org.apache.spark.util.sketch.CountMinSketch
Creates a CountMinSketch with given depth, width, and random seed.
create(double, double, int) - Static method in class org.apache.spark.util.sketch.CountMinSketch
Creates a CountMinSketch with given relative error (eps), confidence, and random seed.
createArrayType(Column) - Static method in class org.apache.spark.sql.api.r.SQLUtils
 
createArrayType(DataType) - Static method in class org.apache.spark.sql.types.DataTypes
Creates an ArrayType by specifying the data type of elements (elementType).
createArrayType(DataType, boolean) - Static method in class org.apache.spark.sql.types.DataTypes
Creates an ArrayType by specifying the data type of elements (elementType) and whether the array contains null values (containsNull).
createArrayWithElementsExceedLimitError(Object) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
createAttrGroupForAttrNames(String, int, boolean, boolean) - Static method in class org.apache.spark.ml.feature.OneHotEncoderCommon
Creates an `AttributeGroup` with the required number of `BinaryAttribute`.
createBatchWriterFactory(PhysicalWriteInfo) - Method in interface org.apache.spark.sql.connector.write.BatchWrite
Creates a writer factory which will be serialized and sent to executors.
createBatchWriterFactory(PhysicalWriteInfo) - Method in interface org.apache.spark.sql.connector.write.DeltaBatchWrite
 
createChunkBlockInfosFromMetaResponse(int, int, int, long, RoaringBitmap[]) - Method in class org.apache.spark.storage.PushBasedFetchHelper
This is executed by the task thread when the iterator.next() is invoked and the iterator processes a response of type ShuffleBlockFetcherIterator.PushMergedRemoteMetaFetchResult.
createColumnarReader(InputPartition) - Method in interface org.apache.spark.sql.connector.read.PartitionReaderFactory
Returns a columnar partition reader to read data from the given InputPartition.
createColumnarReader(InputPartition) - Method in interface org.apache.spark.sql.connector.read.streaming.ContinuousPartitionReaderFactory
 
createCombiner() - Method in class org.apache.spark.Aggregator
 
createCompiledClass(String, File, TestUtils.JavaSourceFromString, Seq<URL>) - Static method in class org.apache.spark.TestUtils
Creates a compiled class with the source file.
createCompiledClass(String, File, String, String, Seq<URL>, Seq<String>, String) - Static method in class org.apache.spark.TestUtils
Creates a compiled class with the given name.
createConnectionFactory(JDBCOptions) - Static method in class org.apache.spark.sql.jdbc.DB2Dialect
 
createConnectionFactory(JDBCOptions) - Static method in class org.apache.spark.sql.jdbc.DerbyDialect
 
createConnectionFactory(JDBCOptions) - Static method in class org.apache.spark.sql.jdbc.H2Dialect
 
createConnectionFactory(JDBCOptions) - Method in class org.apache.spark.sql.jdbc.JdbcDialect
Returns a factory for creating connections to the given JDBC URL.
createConnectionFactory(JDBCOptions) - Static method in class org.apache.spark.sql.jdbc.MsSqlServerDialect
 
createConnectionFactory(JDBCOptions) - Static method in class org.apache.spark.sql.jdbc.MySQLDialect
 
createConnectionFactory(JDBCOptions) - Static method in class org.apache.spark.sql.jdbc.NoopDialect
 
createConnectionFactory(JDBCOptions) - Static method in class org.apache.spark.sql.jdbc.OracleDialect
 
createConnectionFactory(JDBCOptions) - Static method in class org.apache.spark.sql.jdbc.PostgresDialect
 
createConnectionFactory(JDBCOptions) - Static method in class org.apache.spark.sql.jdbc.TeradataDialect
 
createContinuousReaderFactory() - Method in interface org.apache.spark.sql.connector.read.streaming.ContinuousStream
Returns a factory to create a ContinuousPartitionReader for each InputPartition.
createCryptoInputStream(InputStream, SparkConf, byte[]) - Static method in class org.apache.spark.security.CryptoStreamUtils
Helper method to wrap InputStream with CryptoInputStream for decryption.
createCryptoOutputStream(OutputStream, SparkConf, byte[]) - Static method in class org.apache.spark.security.CryptoStreamUtils
Helper method to wrap OutputStream with CryptoOutputStream for encryption.
createDataFrame(RDD<A>, TypeTags.TypeTag<A>) - Method in class org.apache.spark.sql.SparkSession
Creates a DataFrame from an RDD of Product (e.g.
createDataFrame(Seq<A>, TypeTags.TypeTag<A>) - Method in class org.apache.spark.sql.SparkSession
Creates a DataFrame from a local Seq of Product.
createDataFrame(RDD<Row>, StructType) - Method in class org.apache.spark.sql.SparkSession
:: DeveloperApi :: Creates a DataFrame from an RDD containing Rows using the given schema.
createDataFrame(JavaRDD<Row>, StructType) - Method in class org.apache.spark.sql.SparkSession
:: DeveloperApi :: Creates a DataFrame from a JavaRDD containing Rows using the given schema.
createDataFrame(List<Row>, StructType) - Method in class org.apache.spark.sql.SparkSession
:: DeveloperApi :: Creates a DataFrame from a java.util.List containing Rows using the given schema.
createDataFrame(RDD<?>, Class<?>) - Method in class org.apache.spark.sql.SparkSession
Applies a schema to an RDD of Java Beans.
createDataFrame(JavaRDD<?>, Class<?>) - Method in class org.apache.spark.sql.SparkSession
Applies a schema to an RDD of Java Beans.
createDataFrame(List<?>, Class<?>) - Method in class org.apache.spark.sql.SparkSession
Applies a schema to a List of Java Beans.
createDataFrame(RDD<A>, TypeTags.TypeTag<A>) - Method in class org.apache.spark.sql.SQLContext
 
createDataFrame(Seq<A>, TypeTags.TypeTag<A>) - Method in class org.apache.spark.sql.SQLContext
 
createDataFrame(RDD<Row>, StructType) - Method in class org.apache.spark.sql.SQLContext
 
createDataFrame(JavaRDD<Row>, StructType) - Method in class org.apache.spark.sql.SQLContext
 
createDataFrame(List<Row>, StructType) - Method in class org.apache.spark.sql.SQLContext
 
createDataFrame(RDD<?>, Class<?>) - Method in class org.apache.spark.sql.SQLContext
 
createDataFrame(JavaRDD<?>, Class<?>) - Method in class org.apache.spark.sql.SQLContext
 
createDataFrame(List<?>, Class<?>) - Method in class org.apache.spark.sql.SQLContext
 
createDataset(Seq<T>, Encoder<T>) - Method in class org.apache.spark.sql.SparkSession
Creates a Dataset from a local Seq of data of a given type.
createDataset(RDD<T>, Encoder<T>) - Method in class org.apache.spark.sql.SparkSession
Creates a Dataset from an RDD of a given type.
createDataset(List<T>, Encoder<T>) - Method in class org.apache.spark.sql.SparkSession
Creates a Dataset from a java.util.List of a given type.
createDataset(Seq<T>, Encoder<T>) - Method in class org.apache.spark.sql.SQLContext
 
createDataset(RDD<T>, Encoder<T>) - Method in class org.apache.spark.sql.SQLContext
 
createDataset(List<T>, Encoder<T>) - Method in class org.apache.spark.sql.SQLContext
 
createDayTimeIntervalType(byte, byte) - Static method in class org.apache.spark.sql.types.DataTypes
Creates a DayTimeIntervalType by specifying the start and end fields.
createDayTimeIntervalType() - Static method in class org.apache.spark.sql.types.DataTypes
Creates a DayTimeIntervalType with default start and end fields: interval day to second.
createDecimalType(int, int) - Static method in class org.apache.spark.sql.types.DataTypes
Creates a DecimalType by specifying the precision and scale.
createDecimalType() - Static method in class org.apache.spark.sql.types.DataTypes
Creates a DecimalType with default precision and scale, which are 10 and 0.
createDF(RDD<byte[]>, StructType, SparkSession) - Static method in class org.apache.spark.sql.api.r.SQLUtils
 
createDirectory(File) - Static method in class org.apache.spark.util.Utils
Create a directory given the abstract pathname
createDirectory(String, String) - Static method in class org.apache.spark.util.Utils
Create a directory inside the given parent directory.
createExternalTable(String, String) - Method in class org.apache.spark.sql.catalog.Catalog
Deprecated.
use createTable instead. Since 2.2.0.
createExternalTable(String, String, String) - Method in class org.apache.spark.sql.catalog.Catalog
Deprecated.
use createTable instead. Since 2.2.0.
createExternalTable(String, String, Map<String, String>) - Method in class org.apache.spark.sql.catalog.Catalog
Deprecated.
use createTable instead. Since 2.2.0.
createExternalTable(String, String, Map<String, String>) - Method in class org.apache.spark.sql.catalog.Catalog
Deprecated.
use createTable instead. Since 2.2.0.
createExternalTable(String, String, StructType, Map<String, String>) - Method in class org.apache.spark.sql.catalog.Catalog
Deprecated.
use createTable instead. Since 2.2.0.
createExternalTable(String, String, StructType, Map<String, String>) - Method in class org.apache.spark.sql.catalog.Catalog
Deprecated.
use createTable instead. Since 2.2.0.
createExternalTable(String, String) - Method in class org.apache.spark.sql.SQLContext
Deprecated.
use sparkSession.catalog.createTable instead. Since 2.2.0.
createExternalTable(String, String, String) - Method in class org.apache.spark.sql.SQLContext
Deprecated.
use sparkSession.catalog.createTable instead. Since 2.2.0.
createExternalTable(String, String, Map<String, String>) - Method in class org.apache.spark.sql.SQLContext
Deprecated.
use sparkSession.catalog.createTable instead. Since 2.2.0.
createExternalTable(String, String, Map<String, String>) - Method in class org.apache.spark.sql.SQLContext
Deprecated.
use sparkSession.catalog.createTable instead. Since 2.2.0.
createExternalTable(String, String, StructType, Map<String, String>) - Method in class org.apache.spark.sql.SQLContext
Deprecated.
use sparkSession.catalog.createTable instead. Since 2.2.0.
createExternalTable(String, String, StructType, Map<String, String>) - Method in class org.apache.spark.sql.SQLContext
Deprecated.
use sparkSession.catalog.createTable instead. Since 2.2.0.
createExternalTableWithoutLocationError() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
createFailedToGetTokenMessage(String, Throwable) - Static method in class org.apache.spark.util.Utils
Returns a string message about delegation token generation failure
createFuncWithBothIfNotExistsAndReplaceError(SqlBaseParser.CreateFunctionContext) - Static method in class org.apache.spark.sql.errors.QueryParsingErrors
 
createGlobalTempView(String) - Method in class org.apache.spark.sql.Dataset
Creates a global temporary view using the given name.
createIndex(String, NamedReference[], Map<NamedReference, Map<String, String>>, Map<String, String>) - Method in interface org.apache.spark.sql.connector.catalog.index.SupportsIndex
Creates an index.
createIndex(String, Identifier, NamedReference[], Map<NamedReference, Map<String, String>>, Map<String, String>) - Static method in class org.apache.spark.sql.jdbc.DB2Dialect
 
createIndex(String, Identifier, NamedReference[], Map<NamedReference, Map<String, String>>, Map<String, String>) - Static method in class org.apache.spark.sql.jdbc.DerbyDialect
 
createIndex(String, Identifier, NamedReference[], Map<NamedReference, Map<String, String>>, Map<String, String>) - Static method in class org.apache.spark.sql.jdbc.H2Dialect
 
createIndex(String, Identifier, NamedReference[], Map<NamedReference, Map<String, String>>, Map<String, String>) - Method in class org.apache.spark.sql.jdbc.JdbcDialect
Build a create index SQL statement.
createIndex(String, Identifier, NamedReference[], Map<NamedReference, Map<String, String>>, Map<String, String>) - Static method in class org.apache.spark.sql.jdbc.MsSqlServerDialect
 
createIndex(String, Identifier, NamedReference[], Map<NamedReference, Map<String, String>>, Map<String, String>) - Static method in class org.apache.spark.sql.jdbc.MySQLDialect
 
createIndex(String, Identifier, NamedReference[], Map<NamedReference, Map<String, String>>, Map<String, String>) - Static method in class org.apache.spark.sql.jdbc.NoopDialect
 
createIndex(String, Identifier, NamedReference[], Map<NamedReference, Map<String, String>>, Map<String, String>) - Static method in class org.apache.spark.sql.jdbc.OracleDialect
 
createIndex(String, Identifier, NamedReference[], Map<NamedReference, Map<String, String>>, Map<String, String>) - Static method in class org.apache.spark.sql.jdbc.PostgresDialect
 
createIndex(String, Identifier, NamedReference[], Map<NamedReference, Map<String, String>>, Map<String, String>) - Static method in class org.apache.spark.sql.jdbc.TeradataDialect
 
createJar(Seq<File>, File, Option<String>, Option<String>) - Static method in class org.apache.spark.TestUtils
Create a jar file that contains this set of files.
createJarWithClasses(Seq<String>, String, Seq<Tuple2<String, String>>, Seq<URL>) - Static method in class org.apache.spark.TestUtils
Create a jar that defines classes with the given names.
createJarWithFiles(Map<String, String>, File) - Static method in class org.apache.spark.TestUtils
Create a jar file containing multiple files.
createKey(SparkConf) - Static method in class org.apache.spark.security.CryptoStreamUtils
Creates a new encryption key.
createKVStore(Option<File>, boolean, SparkConf) - Static method in class org.apache.spark.status.KVUtils
 
createListeners(SparkConf, ElementTrackingStore) - Method in interface org.apache.spark.status.AppHistoryServerPlugin
Creates listeners to replay the event logs.
createLogForDriver(SparkConf, String, Configuration) - Static method in class org.apache.spark.streaming.util.WriteAheadLogUtils
Create a WriteAheadLog for the driver.
createLogForReceiver(SparkConf, String, Configuration) - Static method in class org.apache.spark.streaming.util.WriteAheadLogUtils
Create a WriteAheadLog for the receiver.
createMapOutputWriter(int, long, int) - Method in interface org.apache.spark.shuffle.api.ShuffleExecutorComponents
Called once per map task to create a writer that will be responsible for persisting all the partitioned bytes written by that map task.
createMapType(DataType, DataType) - Static method in class org.apache.spark.sql.types.DataTypes
Creates a MapType by specifying the data type of keys (keyType) and values (keyType).
createMapType(DataType, DataType, boolean) - Static method in class org.apache.spark.sql.types.DataTypes
Creates a MapType by specifying the data type of keys (keyType), the data type of values (keyType), and whether values contain any null value (valueContainsNull).
createMetrics(long, long, long, long, long, long, long, long, long, long, long, long, long, long, long, long, long, long, long, long, long, long, long, long, long, long, long, long, long, long, long, long, long, long) - Static method in class org.apache.spark.status.LiveEntityHelpers
 
createMetrics(long) - Static method in class org.apache.spark.status.LiveEntityHelpers
 
createModel(DenseVector<Object>) - Method in interface org.apache.spark.ml.ann.Layer
Returns the instance of the layer based on weights provided.
createNamespace(String[], Map<String, String>) - Method in class org.apache.spark.sql.connector.catalog.DelegatingCatalogExtension
 
createNamespace(String[], Map<String, String>) - Method in interface org.apache.spark.sql.connector.catalog.SupportsNamespaces
Create a namespace in the catalog.
createOrReplace() - Method in interface org.apache.spark.sql.CreateTableWriter
Create a new table or replace an existing table with the contents of the data frame.
createOrReplace() - Method in class org.apache.spark.sql.DataFrameWriterV2
 
createOrReplaceGlobalTempView(String) - Method in class org.apache.spark.sql.Dataset
Creates or replaces a global temporary view using the given name.
createOrReplaceTempView(String) - Method in class org.apache.spark.sql.Dataset
Creates a local temporary view using the given name.
createOutputOperationFailureForUI(String) - Static method in class org.apache.spark.streaming.ui.UIUtils
 
createPartition(InternalRow, Map<String, String>) - Method in interface org.apache.spark.sql.connector.catalog.SupportsAtomicPartitionManagement
 
createPartition(InternalRow, Map<String, String>) - Method in interface org.apache.spark.sql.connector.catalog.SupportsPartitionManagement
Create a partition in table.
createPartitions(InternalRow[], Map<String, String>[]) - Method in interface org.apache.spark.sql.connector.catalog.SupportsAtomicPartitionManagement
Create an array of partitions atomically in table.
createPersistedViewFromDatasetAPINotAllowedError() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
createPMMLModelExport(Object) - Static method in class org.apache.spark.mllib.pmml.export.PMMLModelExportFactory
Factory object to help creating the necessary PMMLModelExport implementation taking as input the machine learning model (for example KMeansModel).
createProxyHandler(Function1<String, Option<String>>) - Static method in class org.apache.spark.ui.JettyUtils
Create a handler for proxying request to Workers and Application Drivers
createProxyLocationHeader(String, HttpServletRequest, URI) - Static method in class org.apache.spark.ui.JettyUtils
 
createProxyURI(String, String, String, String) - Static method in class org.apache.spark.ui.JettyUtils
 
createRDDFromArray(JavaSparkContext, byte[][]) - Static method in class org.apache.spark.api.r.RRDD
Create an RRDD given a sequence of byte arrays.
createRDDFromFile(JavaSparkContext, String, int) - Static method in class org.apache.spark.api.r.RRDD
Create an RRDD given a temporary file name.
createReadableChannel(ReadableByteChannel, SparkConf, byte[]) - Static method in class org.apache.spark.security.CryptoStreamUtils
Wrap a ReadableByteChannel for decryption.
createReader(InputPartition) - Method in interface org.apache.spark.sql.connector.read.PartitionReaderFactory
Returns a row-based partition reader to read data from the given InputPartition.
createReader(InputPartition) - Method in interface org.apache.spark.sql.connector.read.streaming.ContinuousPartitionReaderFactory
 
createReaderFactory() - Method in interface org.apache.spark.sql.connector.read.Batch
Returns a factory to create a PartitionReader for each InputPartition.
createReaderFactory() - Method in interface org.apache.spark.sql.connector.read.streaming.MicroBatchStream
Returns a factory to create a PartitionReader for each InputPartition.
createRedirectHandler(String, String, Function1<HttpServletRequest, BoxedUnit>, String, Set<String>) - Static method in class org.apache.spark.ui.JettyUtils
Create a handler that always redirects the user to the given path
createRelation(SQLContext, SaveMode, Map<String, String>, Dataset<Row>) - Method in interface org.apache.spark.sql.sources.CreatableRelationProvider
Saves a DataFrame to a destination (using data source-specific parameters)
createRelation(SQLContext, Map<String, String>) - Method in interface org.apache.spark.sql.sources.RelationProvider
Returns a new base relation with the given parameters.
createRelation(SQLContext, Map<String, String>, StructType) - Method in interface org.apache.spark.sql.sources.SchemaRelationProvider
Returns a new base relation with the given parameters and user defined schema.
createSchedulerBackend(SparkContext, String, TaskScheduler) - Method in interface org.apache.spark.scheduler.ExternalClusterManager
Create a scheduler backend for the given SparkContext and scheduler.
createSchema(Statement, String, String) - Static method in class org.apache.spark.sql.jdbc.DB2Dialect
 
createSchema(Statement, String, String) - Static method in class org.apache.spark.sql.jdbc.DerbyDialect
 
createSchema(Statement, String, String) - Static method in class org.apache.spark.sql.jdbc.H2Dialect
 
createSchema(Statement, String, String) - Method in class org.apache.spark.sql.jdbc.JdbcDialect
Create schema with an optional comment.
createSchema(Statement, String, String) - Static method in class org.apache.spark.sql.jdbc.MsSqlServerDialect
 
createSchema(Statement, String, String) - Static method in class org.apache.spark.sql.jdbc.MySQLDialect
 
createSchema(Statement, String, String) - Static method in class org.apache.spark.sql.jdbc.NoopDialect
 
createSchema(Statement, String, String) - Static method in class org.apache.spark.sql.jdbc.OracleDialect
 
createSchema(Statement, String, String) - Static method in class org.apache.spark.sql.jdbc.PostgresDialect
 
createSchema(Statement, String, String) - Static method in class org.apache.spark.sql.jdbc.TeradataDialect
 
createSecret(SparkConf) - Static method in class org.apache.spark.util.Utils
 
createServletHandler(String, JettyUtils.ServletParams<T>, SparkConf, String) - Static method in class org.apache.spark.ui.JettyUtils
Create a context handler that responds to a request with the given path prefix
createServletHandler(String, HttpServlet, String) - Static method in class org.apache.spark.ui.JettyUtils
Create a context handler that responds to a request with the given path prefix
createSingleFileMapOutputWriter(int, long) - Method in interface org.apache.spark.shuffle.api.ShuffleExecutorComponents
An optional extension for creating a map output writer that can optimize the transfer of a single partition file, as the entire result of a map task, to the backing store.
createSink(SQLContext, Map<String, String>, Seq<String>, OutputMode) - Method in interface org.apache.spark.sql.sources.StreamSinkProvider
 
createSource(SQLContext, String, Option<StructType>, String, Map<String, String>) - Method in interface org.apache.spark.sql.sources.StreamSourceProvider
 
createSparkContext(String, String, String, String[], Map<Object, Object>, Map<Object, Object>) - Static method in class org.apache.spark.api.r.RRDD
 
createStaticHandler(String, String) - Static method in class org.apache.spark.ui.JettyUtils
Create a handler for serving files from a static directory
createStream(JavaStreamingContext, String, String, String, String, int, Duration, int, StorageLevel, String, String, String, String, String) - Method in class org.apache.spark.streaming.kinesis.KinesisUtilsPythonHelper
 
createStreamingSourceNotSpecifySchemaError() - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
createStreamingWriterFactory(PhysicalWriteInfo) - Method in interface org.apache.spark.sql.connector.write.streaming.StreamingWrite
Creates a writer factory which will be serialized and sent to executors.
createStructField(String, String, boolean) - Static method in class org.apache.spark.sql.api.r.SQLUtils
 
createStructField(String, DataType, boolean, Metadata) - Static method in class org.apache.spark.sql.types.DataTypes
Creates a StructField by specifying the name (name), data type (dataType) and whether values of this field can be null values (nullable).
createStructField(String, DataType, boolean) - Static method in class org.apache.spark.sql.types.DataTypes
Creates a StructField with empty metadata.
createStructType(Seq<StructField>) - Static method in class org.apache.spark.sql.api.r.SQLUtils
 
createStructType(List<StructField>) - Static method in class org.apache.spark.sql.types.DataTypes
Creates a StructType with the given list of StructFields (fields).
createStructType(StructField[]) - Static method in class org.apache.spark.sql.types.DataTypes
Creates a StructType with the given StructField array (fields).
createTable(String, String) - Method in class org.apache.spark.sql.catalog.Catalog
Creates a table from the given path and returns the corresponding DataFrame.
createTable(String, String, String) - Method in class org.apache.spark.sql.catalog.Catalog
Creates a table from the given path based on a data source and returns the corresponding DataFrame.
createTable(String, String, Map<String, String>) - Method in class org.apache.spark.sql.catalog.Catalog
Creates a table based on the dataset in a data source and a set of options.
createTable(String, String, Map<String, String>) - Method in class org.apache.spark.sql.catalog.Catalog
(Scala-specific) Creates a table based on the dataset in a data source and a set of options.
createTable(String, String, String, Map<String, String>) - Method in class org.apache.spark.sql.catalog.Catalog
Creates a table based on the dataset in a data source and a set of options.
createTable(String, String, String, Map<String, String>) - Method in class org.apache.spark.sql.catalog.Catalog
(Scala-specific) Creates a table based on the dataset in a data source and a set of options.
createTable(String, String, StructType, Map<String, String>) - Method in class org.apache.spark.sql.catalog.Catalog
Create a table based on the dataset in a data source, a schema and a set of options.
createTable(String, String, StructType, Map<String, String>) - Method in class org.apache.spark.sql.catalog.Catalog
(Scala-specific) Create a table based on the dataset in a data source, a schema and a set of options.
createTable(String, String, StructType, String, Map<String, String>) - Method in class org.apache.spark.sql.catalog.Catalog
Create a table based on the dataset in a data source, a schema and a set of options.
createTable(String, String, StructType, String, Map<String, String>) - Method in class org.apache.spark.sql.catalog.Catalog
(Scala-specific) Create a table based on the dataset in a data source, a schema and a set of options.
createTable(Identifier, StructType, Transform[], Map<String, String>) - Method in class org.apache.spark.sql.connector.catalog.DelegatingCatalogExtension
 
createTable(Identifier, Column[], Transform[], Map<String, String>) - Method in class org.apache.spark.sql.connector.catalog.DelegatingCatalogExtension
 
createTable(Identifier, StructType, Transform[], Map<String, String>) - Method in interface org.apache.spark.sql.connector.catalog.TableCatalog
Deprecated.
createTable(Identifier, Column[], Transform[], Map<String, String>) - Method in interface org.apache.spark.sql.connector.catalog.TableCatalog
Create a table in the catalog.
createTableAsSelectWithNonEmptyDirectoryError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
createTableColumnTypesOptionColumnNotFoundInSchemaError(String, StructType) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
CreateTableWriter<T> - Interface in org.apache.spark.sql
Trait to restrict calls to create and replace operations.
createTaskScheduler(SparkContext, String) - Method in interface org.apache.spark.scheduler.ExternalClusterManager
Create a task scheduler instance for the given SparkContext
createTempDir() - Static method in class org.apache.spark.util.Utils
Create a temporary directory inside the java.io.tmpdir prefixed with spark.
createTempDir(String, String) - Static method in class org.apache.spark.util.Utils
Create a temporary directory inside the given parent directory.
createTempJsonFile(File, String, JsonAST.JValue) - Static method in class org.apache.spark.TestUtils
Creates a temp JSON file that contains the input JSON record.
createTempScriptWithExpectedOutput(File, String, String) - Static method in class org.apache.spark.TestUtils
Creates a temp bash script that prints the given output.
createTempTableNotSpecifyProviderError(SqlBaseParser.CreateTableContext) - Static method in class org.apache.spark.sql.errors.QueryParsingErrors
 
createTempView(String) - Method in class org.apache.spark.sql.Dataset
Creates a local temporary view using the given name.
createUnsafe(long, int, int) - Static method in class org.apache.spark.sql.types.Decimal
Creates a decimal from unscaled, precision and scale without checking the bounds.
createView(Identifier, String, String, String[], StructType, String[], String[], String[], Map<String, String>) - Method in interface org.apache.spark.sql.connector.catalog.ViewCatalog
Create a view in the catalog.
createViewNumColumnsMismatchUserSpecifiedColumnLengthError(int, int) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
createViewWithBothIfNotExistsAndReplaceError(SqlBaseParser.CreateViewContext) - Static method in class org.apache.spark.sql.errors.QueryParsingErrors
 
createWorkspace(int) - Static method in class org.apache.spark.mllib.optimization.NNLS
 
createWritableChannel(WritableByteChannel, SparkConf, byte[]) - Static method in class org.apache.spark.security.CryptoStreamUtils
Wrap a WritableByteChannel for encryption.
createWriter(int, long) - Method in interface org.apache.spark.sql.connector.write.DataWriterFactory
Returns a data writer to do the actual writing work.
createWriter(int, long) - Method in interface org.apache.spark.sql.connector.write.DeltaWriterFactory
 
createWriter(int, long, long) - Method in interface org.apache.spark.sql.connector.write.streaming.StreamingDataWriterFactory
Returns a data writer to do the actual writing work.
createYearMonthIntervalType(byte, byte) - Static method in class org.apache.spark.sql.types.DataTypes
Creates a YearMonthIntervalType by specifying the start and end fields.
createYearMonthIntervalType() - Static method in class org.apache.spark.sql.types.DataTypes
Creates a YearMonthIntervalType with default start and end fields: interval year to month.
crossJoin(Dataset<?>) - Method in class org.apache.spark.sql.Dataset
Explicit cartesian join with another DataFrame.
crosstab(String, String) - Method in class org.apache.spark.sql.DataFrameStatFunctions
Computes a pair-wise frequency table of the given columns.
CrossValidator - Class in org.apache.spark.ml.tuning
K-fold cross validation performs model selection by splitting the dataset into a set of non-overlapping randomly partitioned folds which are used as separate training and test datasets e.g., with k=3 folds, K-fold cross validation will generate 3 (training, test) dataset pairs, each of which uses 2/3 of the data for training and 1/3 for testing.
CrossValidator(String) - Constructor for class org.apache.spark.ml.tuning.CrossValidator
 
CrossValidator() - Constructor for class org.apache.spark.ml.tuning.CrossValidator
 
CrossValidatorModel - Class in org.apache.spark.ml.tuning
CrossValidatorModel contains the model with the highest average cross-validation metric across folds and uses this model to transform input data.
CrossValidatorModel.CrossValidatorModelWriter - Class in org.apache.spark.ml.tuning
Writer for CrossValidatorModel.
CrossValidatorParams - Interface in org.apache.spark.ml.tuning
CryptoStreamUtils - Class in org.apache.spark.security
A util class for manipulating IO encryption and decryption streams.
CryptoStreamUtils() - Constructor for class org.apache.spark.security.CryptoStreamUtils
 
CryptoStreamUtils.BaseErrorHandler - Interface in org.apache.spark.security
SPARK-25535.
CryptoStreamUtils.ErrorHandlingReadableChannel - Class in org.apache.spark.security
 
csc(Column) - Static method in class org.apache.spark.sql.functions
 
csv(String...) - Method in class org.apache.spark.sql.DataFrameReader
Loads CSV files and returns the result as a DataFrame.
csv(String) - Method in class org.apache.spark.sql.DataFrameReader
Loads a CSV file and returns the result as a DataFrame.
csv(Dataset<String>) - Method in class org.apache.spark.sql.DataFrameReader
Loads an Dataset[String] storing CSV rows and returns the result as a DataFrame.
csv(Seq<String>) - Method in class org.apache.spark.sql.DataFrameReader
Loads CSV files and returns the result as a DataFrame.
csv(String) - Method in class org.apache.spark.sql.DataFrameWriter
Saves the content of the DataFrame in CSV format at the specified path.
csv(String) - Method in class org.apache.spark.sql.streaming.DataStreamReader
Loads a CSV file stream and returns the result as a DataFrame.
cube(Column...) - Method in class org.apache.spark.sql.Dataset
Create a multi-dimensional cube for the current Dataset using the specified columns, so we can run aggregation on them.
cube(String, String...) - Method in class org.apache.spark.sql.Dataset
Create a multi-dimensional cube for the current Dataset using the specified columns, so we can run aggregation on them.
cube(Seq<Column>) - Method in class org.apache.spark.sql.Dataset
Create a multi-dimensional cube for the current Dataset using the specified columns, so we can run aggregation on them.
cube(String, Seq<String>) - Method in class org.apache.spark.sql.Dataset
Create a multi-dimensional cube for the current Dataset using the specified columns, so we can run aggregation on them.
CubeType$() - Constructor for class org.apache.spark.sql.RelationalGroupedDataset.CubeType$
 
cume_dist() - Static method in class org.apache.spark.sql.functions
Window function: returns the cumulative distribution of values within a window partition, i.e.
curId() - Static method in class org.apache.spark.sql.Dataset
 
current_date() - Static method in class org.apache.spark.sql.functions
Returns the current date at the start of query evaluation as a date column.
current_timestamp() - Static method in class org.apache.spark.sql.functions
Returns the current timestamp at the start of query evaluation as a timestamp column.
currentAttemptId() - Method in interface org.apache.spark.SparkStageInfo
 
currentAttemptId() - Method in class org.apache.spark.SparkStageInfoImpl
 
currentCatalog() - Method in class org.apache.spark.sql.catalog.Catalog
Returns the current catalog in this session.
currentCatalog() - Method in interface org.apache.spark.sql.connector.catalog.LookupCatalog
Returns the current catalog set.
currentCatalog() - Method in interface org.apache.spark.sql.connector.catalog.View
The current catalog when the view is created.
currentDatabase() - Method in class org.apache.spark.sql.catalog.Catalog
Returns the current database (namespace) in this session.
currentMetricsValues() - Method in interface org.apache.spark.sql.connector.read.PartitionReader
Returns an array of custom task metrics.
currentMetricsValues() - Method in interface org.apache.spark.sql.connector.write.DataWriter
Returns an array of custom task metrics.
currentNamespace() - Method in interface org.apache.spark.sql.connector.catalog.View
The current namespace when the view is created.
currentResult() - Method in interface org.apache.spark.partial.ApproximateEvaluator
 
currentRow() - Static method in class org.apache.spark.sql.expressions.Window
Value representing the current row.
currPrefLocs(Partition, RDD<?>) - Method in class org.apache.spark.rdd.DefaultPartitionCoalescer
 
CUSTOM_METRICS_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress
 
CustomAvgMetric - Class in org.apache.spark.sql.connector.metric
Built-in `CustomMetric` that computes average of metric values.
CustomAvgMetric() - Constructor for class org.apache.spark.sql.connector.metric.CustomAvgMetric
 
customCollectionClsNotResolvedError() - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
CustomMetric - Interface in org.apache.spark.sql.connector.metric
A custom metric.
customMetrics() - Method in class org.apache.spark.sql.streaming.StateOperatorProgress
 
CustomSumMetric - Class in org.apache.spark.sql.connector.metric
Built-in `CustomMetric` that sums up metric values.
CustomSumMetric() - Constructor for class org.apache.spark.sql.connector.metric.CustomSumMetric
 
CustomTaskMetric - Interface in org.apache.spark.sql.connector.metric
A custom task metric.

D

DAGSchedulerEvent - Interface in org.apache.spark.scheduler
Types of events that can be handled by the DAGScheduler.
dapply(Dataset<Row>, byte[], byte[], Object[], StructType) - Static method in class org.apache.spark.sql.api.r.SQLUtils
The helper function for dapply() on R side.
Data(Vector, double, Option<Object>) - Constructor for class org.apache.spark.mllib.classification.impl.GLMClassificationModel.SaveLoadV1_0$.Data
 
Data(double[], double[], double[][]) - Constructor for class org.apache.spark.mllib.classification.NaiveBayesModel.SaveLoadV1_0$.Data
 
Data(double[], double[], double[][], String) - Constructor for class org.apache.spark.mllib.classification.NaiveBayesModel.SaveLoadV2_0$.Data
 
Data(int) - Constructor for class org.apache.spark.mllib.feature.ChiSqSelectorModel.SaveLoadV1_0$.Data
 
Data(Vector, double) - Constructor for class org.apache.spark.mllib.regression.impl.GLMRegressionModel.SaveLoadV1_0$.Data
 
data() - Method in class org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.LaunchTask
 
data() - Method in class org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.StatusUpdate
 
data() - Method in class org.apache.spark.storage.ShuffleFetchCompletionListener
 
Data$() - Constructor for class org.apache.spark.ml.feature.Word2VecModel.Data$
 
Data$() - Constructor for class org.apache.spark.mllib.classification.impl.GLMClassificationModel.SaveLoadV1_0$.Data$
 
Data$() - Constructor for class org.apache.spark.mllib.classification.NaiveBayesModel.SaveLoadV1_0$.Data$
 
Data$() - Constructor for class org.apache.spark.mllib.classification.NaiveBayesModel.SaveLoadV2_0$.Data$
 
Data$() - Constructor for class org.apache.spark.mllib.feature.ChiSqSelectorModel.SaveLoadV1_0$.Data$
 
Data$() - Constructor for class org.apache.spark.mllib.regression.impl.GLMRegressionModel.SaveLoadV1_0$.Data$
 
DATA_DISTRIBUTION_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo
 
Database - Class in org.apache.spark.sql.catalog
A database in Spark, as returned by the listDatabases method defined in Catalog.
Database(String, String, String, String) - Constructor for class org.apache.spark.sql.catalog.Database
 
Database(String, String, String) - Constructor for class org.apache.spark.sql.catalog.Database
 
database() - Method in class org.apache.spark.sql.catalog.Function
 
database() - Method in class org.apache.spark.sql.catalog.Table
 
databaseExists(String) - Method in class org.apache.spark.sql.catalog.Catalog
Check if the database (namespace) with the specified name exists (the name can be qualified with catalog).
databaseFromV1SessionCatalogNotSpecifiedError() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
databaseNameConflictWithSystemPreservedDatabaseError(String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
databaseTypeDefinition() - Method in class org.apache.spark.sql.jdbc.JdbcType
 
dataDistribution() - Method in class org.apache.spark.status.api.v1.RDDStorageInfo
 
DATAFRAME_DAPPLY() - Static method in class org.apache.spark.api.r.RRunnerModes
 
DATAFRAME_GAPPLY() - Static method in class org.apache.spark.api.r.RRunnerModes
 
DataFrameNaFunctions - Class in org.apache.spark.sql
Functionality for working with missing data in DataFrames.
DataFrameReader - Class in org.apache.spark.sql
Interface used to load a Dataset from external storage systems (e.g.
DataFrameStatFunctions - Class in org.apache.spark.sql
Statistic functions for DataFrames.
DataFrameWriter<T> - Class in org.apache.spark.sql
Interface used to write a Dataset to external storage systems (e.g.
DataFrameWriterV2<T> - Class in org.apache.spark.sql
Interface used to write a Dataset to external storage using the v2 API.
dataPathNotExistError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
dataPathNotSpecifiedError() - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
dataSchemaNotSpecifiedError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
dataSchemaNotSpecifiedError(String, String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
dataset() - Method in class org.apache.spark.ml.FitStart
 
Dataset<T> - Class in org.apache.spark.sql
A Dataset is a strongly typed collection of domain-specific objects that can be transformed in parallel using functional or relational operations.
Dataset(SparkSession, LogicalPlan, Encoder<T>) - Constructor for class org.apache.spark.sql.Dataset
 
Dataset(SQLContext, LogicalPlan, Encoder<T>) - Constructor for class org.apache.spark.sql.Dataset
 
DATASET_ID_KEY() - Static method in class org.apache.spark.sql.Dataset
 
DATASET_ID_TAG() - Static method in class org.apache.spark.sql.Dataset
 
DatasetHolder<T> - Class in org.apache.spark.sql
A container for a Dataset, used for implicit conversions in Scala.
DatasetUtils - Class in org.apache.spark.ml.util
 
DatasetUtils() - Constructor for class org.apache.spark.ml.util.DatasetUtils
 
dataSource() - Method in interface org.apache.spark.ui.PagedTable
 
dataSourceNotFoundError(String, Throwable) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
dataSourceOutputModeUnsupportedError(String, OutputMode) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
DataSourceRegister - Interface in org.apache.spark.sql.sources
Data sources should implement this trait so that they can register an alias to their data source.
DataStreamReader - Class in org.apache.spark.sql.streaming
Interface used to load a streaming Dataset from external storage systems (e.g.
DataStreamWriter<T> - Class in org.apache.spark.sql.streaming
Interface used to write a streaming Dataset to external storage systems (e.g.
dataTablesHeaderNodes(HttpServletRequest) - Static method in class org.apache.spark.ui.UIUtils
 
dataType() - Method in class org.apache.spark.sql.avro.SchemaConverters.SchemaType
 
dataType() - Method in class org.apache.spark.sql.catalog.Column
 
dataType() - Method in interface org.apache.spark.sql.connector.catalog.Column
Returns the data type of this table column.
dataType() - Method in interface org.apache.spark.sql.connector.catalog.MetadataColumn
The data type of values in this metadata column.
dataType() - Method in class org.apache.spark.sql.connector.catalog.TableChange.AddColumn
 
dataType() - Method in class org.apache.spark.sql.connector.expressions.Cast
 
dataType() - Method in class org.apache.spark.sql.connector.expressions.filter.AlwaysFalse
 
dataType() - Method in class org.apache.spark.sql.connector.expressions.filter.AlwaysTrue
 
dataType() - Method in interface org.apache.spark.sql.connector.expressions.Literal
Returns the SQL data type of the literal.
dataType() - Method in class org.apache.spark.sql.expressions.UserDefinedAggregateFunction
Deprecated.
The DataType of the returned value of this UserDefinedAggregateFunction.
DataType - Class in org.apache.spark.sql.types
The base type of all Spark SQL data types.
DataType() - Constructor for class org.apache.spark.sql.types.DataType
 
dataType() - Method in class org.apache.spark.sql.types.StructField
 
dataType() - Method in class org.apache.spark.sql.vectorized.ColumnVector
Returns the data type of this column vector.
dataTypeMismatchForDeserializerError(DataType, String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
dataTypeOperationUnsupportedError() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
DataTypes - Class in org.apache.spark.sql.types
To get/create specific data type, users should use singleton objects and factory methods provided by this class.
DataTypes() - Constructor for class org.apache.spark.sql.types.DataTypes
 
dataTypeUnexpectedError(DataType) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
dataTypeUnsupportedByClassError(DataType, String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
dataTypeUnsupportedByDataSourceError(String, StructField) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
dataTypeUnsupportedByExtractValueError(DataType, Expression, Expression) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
dataTypeUnsupportedError(String, String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
dataTypeUnsupportedError(String, SqlBaseParser.PrimitiveDataTypeContext) - Static method in class org.apache.spark.sql.errors.QueryParsingErrors
 
dataTypeUnsupportedYetError(DataType) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
DataValidators - Class in org.apache.spark.mllib.util
A collection of methods used to validate data before applying ML algorithms.
DataValidators() - Constructor for class org.apache.spark.mllib.util.DataValidators
 
DataWriter<T> - Interface in org.apache.spark.sql.connector.write
A data writer returned by DataWriterFactory.createWriter(int, long) and is responsible for writing data for an input RDD partition.
DataWriterFactory - Interface in org.apache.spark.sql.connector.write
A factory of DataWriter returned by BatchWrite.createBatchWriterFactory(PhysicalWriteInfo), which is responsible for creating and initializing the actual data writer at executor side.
date() - Method in class org.apache.spark.sql.ColumnName
Creates a new StructField of type date.
DATE() - Static method in class org.apache.spark.sql.Encoders
An encoder for nullable date type.
date_add(Column, int) - Static method in class org.apache.spark.sql.functions
Returns the date that is days days after start
date_add(Column, Column) - Static method in class org.apache.spark.sql.functions
Returns the date that is days days after start
date_format(Column, String) - Static method in class org.apache.spark.sql.functions
Converts a date/timestamp/string to a value of string in the format specified by the date format given by the second argument.
date_sub(Column, int) - Static method in class org.apache.spark.sql.functions
Returns the date that is days days before start
date_sub(Column, Column) - Static method in class org.apache.spark.sql.functions
Returns the date that is days days before start
date_trunc(String, Column) - Static method in class org.apache.spark.sql.functions
Returns timestamp truncated to the unit specified by the format.
datediff(Column, Column) - Static method in class org.apache.spark.sql.functions
Returns the number of days from start to end.
DateType - Static variable in class org.apache.spark.sql.types.DataTypes
Gets the DateType object.
DateType - Class in org.apache.spark.sql.types
The date type represents a valid date in the proleptic Gregorian calendar.
DateType() - Constructor for class org.apache.spark.sql.types.DateType
 
DAY() - Static method in class org.apache.spark.sql.types.DayTimeIntervalType
 
dayofmonth(Column) - Static method in class org.apache.spark.sql.functions
Extracts the day of the month as an integer from a given date/timestamp/string.
dayofweek(Column) - Static method in class org.apache.spark.sql.functions
Extracts the day of the week as an integer from a given date/timestamp/string.
dayofyear(Column) - Static method in class org.apache.spark.sql.functions
Extracts the day of the year as an integer from a given date/timestamp/string.
days(String) - Static method in class org.apache.spark.sql.connector.expressions.Expressions
Create a daily transform for a timestamp or date column.
days(NamedReference) - Static method in class org.apache.spark.sql.connector.expressions.LogicalExpressions
 
days(Column) - Static method in class org.apache.spark.sql.functions
A transform for timestamps and dates to partition data into days.
days - Variable in class org.apache.spark.unsafe.types.CalendarInterval
 
dayTimeFields() - Static method in class org.apache.spark.sql.types.DayTimeIntervalType
 
DayTimeIntervalType - Class in org.apache.spark.sql.types
The type represents day-time intervals of the SQL standard.
DayTimeIntervalType(byte, byte) - Constructor for class org.apache.spark.sql.types.DayTimeIntervalType
 
DB2Dialect - Class in org.apache.spark.sql.jdbc
 
DB2Dialect() - Constructor for class org.apache.spark.sql.jdbc.DB2Dialect
 
DB2Dialect.DB2SQLBuilder - Class in org.apache.spark.sql.jdbc
 
DB2SQLBuilder() - Constructor for class org.apache.spark.sql.jdbc.DB2Dialect.DB2SQLBuilder
 
DCT - Class in org.apache.spark.ml.feature
A feature transformer that takes the 1D discrete cosine transform of a real vector.
DCT(String) - Constructor for class org.apache.spark.ml.feature.DCT
 
DCT() - Constructor for class org.apache.spark.ml.feature.DCT
 
ddlUnsupportedTemporarilyError(String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
ddlWithoutHiveSupportEnabledError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
deallocate() - Method in class org.apache.spark.storage.ReadableChannelFileRegion
 
decayFactor() - Method in class org.apache.spark.mllib.clustering.StreamingKMeans
 
decimal() - Method in class org.apache.spark.sql.ColumnName
Creates a new StructField of type decimal.
decimal(int, int) - Method in class org.apache.spark.sql.ColumnName
Creates a new StructField of type decimal.
DECIMAL() - Static method in class org.apache.spark.sql.Encoders
An encoder for nullable decimal type.
Decimal - Class in org.apache.spark.sql.types
A mutable implementation of BigDecimal that can hold a Long if values are small enough.
Decimal() - Constructor for class org.apache.spark.sql.types.Decimal
 
Decimal.DecimalAsIfIntegral$ - Class in org.apache.spark.sql.types
A Integral evidence parameter for Decimals.
Decimal.DecimalIsConflicted - Interface in org.apache.spark.sql.types
Common methods for Decimal evidence parameters
Decimal.DecimalIsFractional$ - Class in org.apache.spark.sql.types
A Fractional evidence parameter for Decimals.
DecimalAsIfIntegral$() - Constructor for class org.apache.spark.sql.types.Decimal.DecimalAsIfIntegral$
 
decimalCannotGreaterThanPrecisionError(int, int) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
DecimalExactNumeric - Class in org.apache.spark.sql.types
 
DecimalExactNumeric() - Constructor for class org.apache.spark.sql.types.DecimalExactNumeric
 
DecimalIsFractional$() - Constructor for class org.apache.spark.sql.types.Decimal.DecimalIsFractional$
 
decimalPrecisionExceedsMaxPrecisionError(int, int) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
DecimalType - Class in org.apache.spark.sql.types
The data type representing java.math.BigDecimal values.
DecimalType(int, int) - Constructor for class org.apache.spark.sql.types.DecimalType
 
DecimalType(int) - Constructor for class org.apache.spark.sql.types.DecimalType
 
DecimalType() - Constructor for class org.apache.spark.sql.types.DecimalType
 
DecimalType.Expression$ - Class in org.apache.spark.sql.types
 
DecimalType.Fixed$ - Class in org.apache.spark.sql.types
 
DecisionTree - Class in org.apache.spark.mllib.tree
A class which implements a decision tree learning algorithm for classification and regression.
DecisionTree(Strategy) - Constructor for class org.apache.spark.mllib.tree.DecisionTree
 
DecisionTreeClassificationModel - Class in org.apache.spark.ml.classification
Decision tree model (http://en.wikipedia.org/wiki/Decision_tree_learning) for classification.
DecisionTreeClassifier - Class in org.apache.spark.ml.classification
Decision tree learning algorithm (http://en.wikipedia.org/wiki/Decision_tree_learning) for classification.
DecisionTreeClassifier(String) - Constructor for class org.apache.spark.ml.classification.DecisionTreeClassifier
 
DecisionTreeClassifier() - Constructor for class org.apache.spark.ml.classification.DecisionTreeClassifier
 
DecisionTreeClassifierParams - Interface in org.apache.spark.ml.tree
 
DecisionTreeModel - Interface in org.apache.spark.ml.tree
Abstraction for Decision Tree models.
DecisionTreeModel - Class in org.apache.spark.mllib.tree.model
Decision tree model for classification or regression.
DecisionTreeModel(Node, Enumeration.Value) - Constructor for class org.apache.spark.mllib.tree.model.DecisionTreeModel
 
DecisionTreeModel.SaveLoadV1_0$ - Class in org.apache.spark.mllib.tree.model
 
DecisionTreeModel.SaveLoadV1_0$.NodeData - Class in org.apache.spark.mllib.tree.model
Model data for model import/export
DecisionTreeModel.SaveLoadV1_0$.NodeData$ - Class in org.apache.spark.mllib.tree.model
 
DecisionTreeModel.SaveLoadV1_0$.PredictData - Class in org.apache.spark.mllib.tree.model
 
DecisionTreeModel.SaveLoadV1_0$.PredictData$ - Class in org.apache.spark.mllib.tree.model
 
DecisionTreeModel.SaveLoadV1_0$.SplitData - Class in org.apache.spark.mllib.tree.model
 
DecisionTreeModel.SaveLoadV1_0$.SplitData$ - Class in org.apache.spark.mllib.tree.model
 
DecisionTreeModelReadWrite - Class in org.apache.spark.ml.tree
Helper classes for tree model persistence
DecisionTreeModelReadWrite() - Constructor for class org.apache.spark.ml.tree.DecisionTreeModelReadWrite
 
DecisionTreeModelReadWrite.NodeData - Class in org.apache.spark.ml.tree
Info for a Node
DecisionTreeModelReadWrite.NodeData$ - Class in org.apache.spark.ml.tree
 
DecisionTreeModelReadWrite.SplitData - Class in org.apache.spark.ml.tree
Info for a Split
DecisionTreeModelReadWrite.SplitData$ - Class in org.apache.spark.ml.tree
 
DecisionTreeParams - Interface in org.apache.spark.ml.tree
Parameters for Decision Tree-based algorithms.
DecisionTreeRegressionModel - Class in org.apache.spark.ml.regression
Decision tree (Wikipedia) model for regression.
DecisionTreeRegressor - Class in org.apache.spark.ml.regression
Decision tree learning algorithm for regression.
DecisionTreeRegressor(String) - Constructor for class org.apache.spark.ml.regression.DecisionTreeRegressor
 
DecisionTreeRegressor() - Constructor for class org.apache.spark.ml.regression.DecisionTreeRegressor
 
DecisionTreeRegressorParams - Interface in org.apache.spark.ml.tree
 
decode(Column, String) - Static method in class org.apache.spark.sql.functions
Computes the first argument into a string from a binary using the provided character set (one of 'US-ASCII', 'ISO-8859-1', 'UTF-8', 'UTF-16BE', 'UTF-16LE', 'UTF-16').
decodeFileNameInURI(URI) - Static method in class org.apache.spark.util.Utils
Get the file name from uri's raw path and decode it.
decodeStructField(StructField, boolean) - Method in interface org.apache.spark.ml.attribute.AttributeFactory
Creates an Attribute from a StructField instance, optionally preserving name.
decodeURLParameter(String) - Static method in class org.apache.spark.ui.UIUtils
Decode URLParameter if URL is encoded by YARN-WebAppProxyServlet.
decodeURLParameter(MultivaluedMap<String, String>) - Static method in class org.apache.spark.ui.UIUtils
Decode URLParameter if URL is encoded by YARN-WebAppProxyServlet.
DecommissionBlockManager$() - Constructor for class org.apache.spark.storage.BlockManagerMessages.DecommissionBlockManager$
 
DecommissionBlockManagers(Seq<String>) - Constructor for class org.apache.spark.storage.BlockManagerMessages.DecommissionBlockManagers
 
DecommissionBlockManagers$() - Constructor for class org.apache.spark.storage.BlockManagerMessages.DecommissionBlockManagers$
 
DecommissionExecutor$() - Constructor for class org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.DecommissionExecutor$
 
DecommissionExecutorsOnHost(String) - Constructor for class org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.DecommissionExecutorsOnHost
 
DecommissionExecutorsOnHost$() - Constructor for class org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.DecommissionExecutorsOnHost$
 
decommissionFinished() - Static method in class org.apache.spark.scheduler.ExecutorLossMessage
 
decorrelateInnerQueryThroughPlanUnsupportedError(LogicalPlan) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
DEFAULT() - Static method in class org.apache.spark.sql.types.DayTimeIntervalType
 
DEFAULT() - Static method in class org.apache.spark.sql.types.YearMonthIntervalType
 
DEFAULT_DRIVER_MEM_MB() - Static method in class org.apache.spark.util.Utils
Define a default value for driver memory here since this value is referenced across the code base and nearly all files already use Utils.scala
DEFAULT_NUM_OUTPUT_ROWS() - Static method in class org.apache.spark.sql.streaming.SinkProgress
 
DEFAULT_NUMBER_EXECUTORS() - Static method in class org.apache.spark.scheduler.cluster.SchedulerBackendUtils
 
DEFAULT_RESOURCE_PROFILE_ID() - Static method in class org.apache.spark.resource.ResourceProfile
 
DEFAULT_SASL_KERBEROS_SERVICE_NAME() - Static method in class org.apache.spark.kafka010.KafkaTokenSparkConf
 
DEFAULT_SASL_TOKEN_MECHANISM() - Static method in class org.apache.spark.kafka010.KafkaTokenSparkConf
 
DEFAULT_SCALE() - Static method in class org.apache.spark.sql.types.DecimalType
 
DEFAULT_SECURITY_PROTOCOL_CONFIG() - Static method in class org.apache.spark.kafka010.KafkaTokenSparkConf
 
DEFAULT_SHUTDOWN_PRIORITY() - Static method in class org.apache.spark.util.ShutdownHookManager
 
DEFAULT_TARGET_SERVERS_REGEX() - Static method in class org.apache.spark.kafka010.KafkaTokenSparkConf
 
defaultAttr() - Static method in class org.apache.spark.ml.attribute.BinaryAttribute
The default binary attribute.
defaultAttr() - Static method in class org.apache.spark.ml.attribute.NominalAttribute
The default nominal attribute.
defaultAttr() - Static method in class org.apache.spark.ml.attribute.NumericAttribute
The default numeric attribute.
defaultColumnNotEnabledError(ParserRuleContext) - Static method in class org.apache.spark.sql.errors.QueryParsingErrors
 
defaultColumnNotImplementedYetError(ParserRuleContext) - Static method in class org.apache.spark.sql.errors.QueryParsingErrors
 
defaultColumnReferencesNotAllowedInPartitionSpec(ParserRuleContext) - Static method in class org.apache.spark.sql.errors.QueryParsingErrors
 
defaultCopy(ParamMap) - Method in interface org.apache.spark.ml.param.Params
Default implementation of copy with extra params.
defaultCorrName() - Static method in class org.apache.spark.mllib.stat.correlation.CorrelationNames
 
DefaultCredentials - Class in org.apache.spark.streaming.kinesis
Returns DefaultAWSCredentialsProviderChain for authentication.
DefaultCredentials() - Constructor for class org.apache.spark.streaming.kinesis.DefaultCredentials
 
defaultDatabaseNotExistsError(String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
defaultLink() - Method in class org.apache.spark.ml.regression.GeneralizedLinearRegression.Binomial$
 
defaultLink() - Method in class org.apache.spark.ml.regression.GeneralizedLinearRegression.Gamma$
 
defaultLink() - Method in class org.apache.spark.ml.regression.GeneralizedLinearRegression.Gaussian$
 
defaultLink() - Method in class org.apache.spark.ml.regression.GeneralizedLinearRegression.Poisson$
 
defaultMinPartitions() - Method in class org.apache.spark.api.java.JavaSparkContext
Default min number of partitions for Hadoop RDDs when not given by user
defaultMinPartitions() - Method in class org.apache.spark.SparkContext
Default min number of partitions for Hadoop RDDs when not given by user Notice that we use math.min so the "defaultMinPartitions" cannot be higher than 2.
defaultModuleOptions() - Static method in class org.apache.spark.launcher.JavaModuleOptions
Returns the default Java options related to `--add-opens' and `-XX:+IgnoreUnrecognizedVMOptions` used by Spark.
defaultNamespace() - Method in interface org.apache.spark.sql.connector.catalog.CatalogPlugin
Return a default namespace for the catalog.
defaultNamespace() - Method in class org.apache.spark.sql.connector.catalog.DelegatingCatalogExtension
 
defaultNullOrdering() - Method in enum org.apache.spark.sql.connector.expressions.SortDirection
Returns the default null ordering to use if no null ordering is specified explicitly.
defaultParallelism() - Method in class org.apache.spark.api.java.JavaSparkContext
Default level of parallelism to use when not given by user (e.g.
defaultParallelism() - Method in interface org.apache.spark.scheduler.SchedulerBackend
 
defaultParallelism() - Method in interface org.apache.spark.scheduler.TaskScheduler
 
defaultParallelism() - Method in class org.apache.spark.SparkContext
Default level of parallelism to use when not given by user (e.g.
defaultParamMap() - Method in interface org.apache.spark.ml.param.Params
Internal param map for default values.
defaultParams(String) - Static method in class org.apache.spark.mllib.tree.configuration.BoostingStrategy
Returns default configuration for the boosting algorithm
defaultParams(Enumeration.Value) - Static method in class org.apache.spark.mllib.tree.configuration.BoostingStrategy
Returns default configuration for the boosting algorithm
DefaultParamsReadable<T> - Interface in org.apache.spark.ml.util
Helper trait for making simple Params types readable.
DefaultParamsWritable - Interface in org.apache.spark.ml.util
Helper trait for making simple Params types writable.
DefaultPartitionCoalescer - Class in org.apache.spark.rdd
Coalesce the partitions of a parent RDD (prev) into fewer partitions, so that each partition of this RDD computes one or more of the parent ones.
DefaultPartitionCoalescer(double) - Constructor for class org.apache.spark.rdd.DefaultPartitionCoalescer
 
DefaultPartitionCoalescer.partitionGroupOrdering$ - Class in org.apache.spark.rdd
 
defaultPartitioner(RDD<?>, Seq<RDD<?>>) - Static method in class org.apache.spark.Partitioner
Choose a partitioner to use for a cogroup-like operation between a number of RDDs.
DefaultProfileExecutorResources$() - Constructor for class org.apache.spark.resource.ResourceProfile.DefaultProfileExecutorResources$
 
defaultReferencesNotAllowedInComplexExpressionsInInsertValuesList() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
defaultReferencesNotAllowedInComplexExpressionsInMergeInsertsOrUpdates() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
defaultReferencesNotAllowedInComplexExpressionsInUpdateSetClause() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
defaultReferencesNotAllowedInDataSource(String, String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
defaultReferencesNotAllowedInMergeCondition() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
defaultReferencesNotAllowedInUpdateWhereClause() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
defaultSize() - Method in class org.apache.spark.sql.types.ArrayType
The default size of a value of the ArrayType is the default size of the element type.
defaultSize() - Method in class org.apache.spark.sql.types.BinaryType
The default size of a value of the BinaryType is 100 bytes.
defaultSize() - Method in class org.apache.spark.sql.types.BooleanType
The default size of a value of the BooleanType is 1 byte.
defaultSize() - Method in class org.apache.spark.sql.types.ByteType
The default size of a value of the ByteType is 1 byte.
defaultSize() - Method in class org.apache.spark.sql.types.CalendarIntervalType
 
defaultSize() - Method in class org.apache.spark.sql.types.CharType
 
defaultSize() - Method in class org.apache.spark.sql.types.DataType
The default size of a value of this data type, used internally for size estimation.
defaultSize() - Method in class org.apache.spark.sql.types.DateType
The default size of a value of the DateType is 4 bytes.
defaultSize() - Method in class org.apache.spark.sql.types.DayTimeIntervalType
The day-time interval type has constant precision.
defaultSize() - Method in class org.apache.spark.sql.types.DecimalType
The default size of a value of the DecimalType is 8 bytes when precision is at most 18, and 16 bytes otherwise.
defaultSize() - Method in class org.apache.spark.sql.types.DoubleType
The default size of a value of the DoubleType is 8 bytes.
defaultSize() - Method in class org.apache.spark.sql.types.FloatType
The default size of a value of the FloatType is 4 bytes.
defaultSize() - Method in class org.apache.spark.sql.types.IntegerType
The default size of a value of the IntegerType is 4 bytes.
defaultSize() - Method in class org.apache.spark.sql.types.LongType
The default size of a value of the LongType is 8 bytes.
defaultSize() - Method in class org.apache.spark.sql.types.MapType
The default size of a value of the MapType is (the default size of the key type + the default size of the value type).
defaultSize() - Method in class org.apache.spark.sql.types.NullType
 
defaultSize() - Method in class org.apache.spark.sql.types.ObjectType
 
defaultSize() - Method in class org.apache.spark.sql.types.ShortType
The default size of a value of the ShortType is 2 bytes.
defaultSize() - Method in class org.apache.spark.sql.types.StringType
The default size of a value of the StringType is 20 bytes.
defaultSize() - Method in class org.apache.spark.sql.types.StructType
The default size of a value of the StructType is the total default sizes of all field types.
defaultSize() - Method in class org.apache.spark.sql.types.TimestampNTZType
The default size of a value of the TimestampNTZType is 8 bytes.
defaultSize() - Method in class org.apache.spark.sql.types.TimestampType
The default size of a value of the TimestampType is 8 bytes.
defaultSize() - Method in class org.apache.spark.sql.types.UserDefinedType
 
defaultSize() - Method in class org.apache.spark.sql.types.VarcharType
 
defaultSize() - Method in class org.apache.spark.sql.types.YearMonthIntervalType
Year-month interval values always occupy 4 bytes.
defaultStrategy(String) - Static method in class org.apache.spark.mllib.tree.configuration.Strategy
Construct a default set of parameters for DecisionTree
defaultStrategy(Enumeration.Value) - Static method in class org.apache.spark.mllib.tree.configuration.Strategy
Construct a default set of parameters for DecisionTree
DefaultTopologyMapper - Class in org.apache.spark.storage
A TopologyMapper that assumes all nodes are in the same rack
DefaultTopologyMapper(SparkConf) - Constructor for class org.apache.spark.storage.DefaultTopologyMapper
 
defaultValue() - Method in interface org.apache.spark.sql.connector.catalog.Column
Returns the default value of this table column.
defaultValue() - Method in class org.apache.spark.sql.connector.catalog.TableChange.AddColumn
 
defaultValuesMayNotContainSubQueryExpressions() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
defineTempFuncWithIfNotExistsError(SqlBaseParser.CreateFunctionContext) - Static method in class org.apache.spark.sql.errors.QueryParsingErrors
 
defineTempViewWithIfNotExistsError(SqlBaseParser.CreateViewContext) - Static method in class org.apache.spark.sql.errors.QueryParsingErrors
 
degree() - Method in class org.apache.spark.ml.feature.PolynomialExpansion
The polynomial degree to expand, which should be greater than equal to 1.
degrees() - Method in class org.apache.spark.graphx.GraphOps
 
degrees(Column) - Static method in class org.apache.spark.sql.functions
Converts an angle measured in radians to an approximately equivalent angle measured in degrees.
degrees(String) - Static method in class org.apache.spark.sql.functions
Converts an angle measured in radians to an approximately equivalent angle measured in degrees.
degreesOfFreedom() - Method in class org.apache.spark.ml.regression.GeneralizedLinearRegressionSummary
 
degreesOfFreedom() - Method in class org.apache.spark.ml.regression.LinearRegressionSummary
Degrees of freedom
degreesOfFreedom() - Method in class org.apache.spark.mllib.stat.test.ChiSqTestResult
 
degreesOfFreedom() - Method in class org.apache.spark.mllib.stat.test.KolmogorovSmirnovTestResult
 
degreesOfFreedom() - Method in interface org.apache.spark.mllib.stat.test.TestResult
Returns the degree(s) of freedom of the hypothesis test.
delegate() - Method in class org.apache.spark.ContextAwareIterator
 
delegate() - Method in class org.apache.spark.InterruptibleIterator
 
DelegatingCatalogExtension - Class in org.apache.spark.sql.connector.catalog
A simple implementation of CatalogExtension, which implements all the catalog functions by calling the built-in session catalog directly.
DelegatingCatalogExtension() - Constructor for class org.apache.spark.sql.connector.catalog.DelegatingCatalogExtension
 
delegationTokensRequired(SparkConf, Configuration) - Method in interface org.apache.spark.security.HadoopDelegationTokenProvider
Returns true if delegation tokens are required for this service.
delete(T, T) - Method in interface org.apache.spark.sql.connector.write.DeltaWriter
Deletes a row.
deleteCheckpointFiles() - Method in class org.apache.spark.ml.clustering.DistributedLDAModel
Remove any remaining checkpoint files from training.
deleteColumn(String[], Boolean) - Static method in interface org.apache.spark.sql.connector.catalog.TableChange
Create a TableChange for deleting a field.
deleteRecursively(File) - Static method in class org.apache.spark.util.Utils
Delete a file or directory and its contents recursively.
deleteWhere(Filter[]) - Method in interface org.apache.spark.sql.connector.catalog.SupportsDelete
Delete data from a data source table that matches filter expressions.
deleteWhere(Predicate[]) - Method in interface org.apache.spark.sql.connector.catalog.SupportsDelete
 
deleteWhere(Predicate[]) - Method in interface org.apache.spark.sql.connector.catalog.SupportsDeleteV2
Delete data from a data source table that matches filter expressions.
delta() - Method in class org.apache.spark.ml.regression.GeneralizedLinearRegression.Tweedie$
Constant used in initialization and deviance to avoid numerical issues.
DeltaBatchWrite - Interface in org.apache.spark.sql.connector.write
An interface that defines how to write a delta of rows during batch processing.
DeltaWrite - Interface in org.apache.spark.sql.connector.write
A logical representation of a data source write that handles a delta of rows.
DeltaWriteBuilder - Interface in org.apache.spark.sql.connector.write
An interface for building a DeltaWrite.
DeltaWriter<T> - Interface in org.apache.spark.sql.connector.write
A data writer returned by DeltaWriterFactory.createWriter(int, long) and is responsible for writing a delta of rows.
DeltaWriterFactory - Interface in org.apache.spark.sql.connector.write
A factory for creating DeltaWriters returned by DeltaBatchWrite.createBatchWriterFactory(PhysicalWriteInfo), which is responsible for creating and initializing writers at the executor side.
dense(int, int, double[]) - Static method in class org.apache.spark.ml.linalg.Matrices
Creates a column-major dense matrix.
dense(double, double...) - Static method in class org.apache.spark.ml.linalg.Vectors
Creates a dense vector from its values.
dense(double, Seq<Object>) - Static method in class org.apache.spark.ml.linalg.Vectors
Creates a dense vector from its values.
dense(double[]) - Static method in class org.apache.spark.ml.linalg.Vectors
Creates a dense vector from a double array.
dense(int, int, double[]) - Static method in class org.apache.spark.mllib.linalg.Matrices
Creates a column-major dense matrix.
dense(double, double...) - Static method in class org.apache.spark.mllib.linalg.Vectors
Creates a dense vector from its values.
dense(double, Seq<Object>) - Static method in class org.apache.spark.mllib.linalg.Vectors
Creates a dense vector from its values.
dense(double[]) - Static method in class org.apache.spark.mllib.linalg.Vectors
Creates a dense vector from a double array.
dense_rank() - Static method in class org.apache.spark.sql.functions
Window function: returns the rank of rows within a window partition, without any gaps.
DenseMatrix - Class in org.apache.spark.ml.linalg
Column-major dense matrix.
DenseMatrix(int, int, double[], boolean) - Constructor for class org.apache.spark.ml.linalg.DenseMatrix
 
DenseMatrix(int, int, double[]) - Constructor for class org.apache.spark.ml.linalg.DenseMatrix
Column-major dense matrix.
DenseMatrix - Class in org.apache.spark.mllib.linalg
Column-major dense matrix.
DenseMatrix(int, int, double[], boolean) - Constructor for class org.apache.spark.mllib.linalg.DenseMatrix
 
DenseMatrix(int, int, double[]) - Constructor for class org.apache.spark.mllib.linalg.DenseMatrix
Column-major dense matrix.
DenseVector - Class in org.apache.spark.ml.linalg
A dense vector represented by a value array.
DenseVector(double[]) - Constructor for class org.apache.spark.ml.linalg.DenseVector
 
DenseVector - Class in org.apache.spark.mllib.linalg
A dense vector represented by a value array.
DenseVector(double[]) - Constructor for class org.apache.spark.mllib.linalg.DenseVector
 
dependencies() - Method in class org.apache.spark.rdd.RDD
Get the list of dependencies of this RDD, taking into account whether the RDD is checkpointed or not.
dependencies() - Method in class org.apache.spark.streaming.dstream.DStream
List of parent DStreams on which this DStream depends on
dependencies() - Method in class org.apache.spark.streaming.dstream.InputDStream
 
Dependency<T> - Class in org.apache.spark
:: DeveloperApi :: Base class for dependencies.
Dependency() - Constructor for class org.apache.spark.Dependency
 
DependencyUtils - Class in org.apache.spark.util
 
DependencyUtils() - Constructor for class org.apache.spark.util.DependencyUtils
 
DEPLOY_MODE - Static variable in class org.apache.spark.launcher.SparkLauncher
The Spark deploy mode.
deployMode() - Method in class org.apache.spark.SparkContext
 
DEPRECATED_CHILD_CONNECTION_TIMEOUT - Static variable in class org.apache.spark.launcher.SparkLauncher
Deprecated.
use `CHILD_CONNECTION_TIMEOUT`
depth() - Method in class org.apache.spark.ml.classification.DecisionTreeClassificationModel
 
depth() - Method in class org.apache.spark.ml.regression.DecisionTreeRegressionModel
 
depth() - Method in interface org.apache.spark.ml.tree.DecisionTreeModel
Depth of the tree.
depth() - Method in class org.apache.spark.mllib.tree.model.DecisionTreeModel
Get depth of tree.
depth() - Method in class org.apache.spark.util.sketch.CountMinSketch
Depth of this CountMinSketch.
DerbyDialect - Class in org.apache.spark.sql.jdbc
 
DerbyDialect() - Constructor for class org.apache.spark.sql.jdbc.DerbyDialect
 
deriv(double) - Method in class org.apache.spark.ml.regression.GeneralizedLinearRegression.CLogLog$
 
deriv(double) - Method in class org.apache.spark.ml.regression.GeneralizedLinearRegression.Identity$
 
deriv(double) - Method in class org.apache.spark.ml.regression.GeneralizedLinearRegression.Inverse$
 
deriv(double) - Method in class org.apache.spark.ml.regression.GeneralizedLinearRegression.Log$
 
deriv(double) - Method in class org.apache.spark.ml.regression.GeneralizedLinearRegression.Logit$
 
deriv(double) - Method in class org.apache.spark.ml.regression.GeneralizedLinearRegression.Probit$
 
deriv(double) - Method in class org.apache.spark.ml.regression.GeneralizedLinearRegression.Sqrt$
 
derivative() - Method in interface org.apache.spark.ml.ann.ActivationFunction
Implements a derivative of a function (needed for the back propagation)
desc() - Method in class org.apache.spark.serializer.SerializationDebugger.ObjectStreamClassMethods
 
desc() - Method in class org.apache.spark.sql.Column
Returns a sort expression based on the descending order of the column.
desc(String) - Static method in class org.apache.spark.sql.functions
Returns a sort expression based on the descending order of the column.
desc() - Method in class org.apache.spark.util.MethodIdentifier
 
DESC_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper
 
DESC_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode
 
desc_nulls_first() - Method in class org.apache.spark.sql.Column
Returns a sort expression based on the descending order of the column, and null values appear before non-null values.
desc_nulls_first(String) - Static method in class org.apache.spark.sql.functions
Returns a sort expression based on the descending order of the column, and null values appear before non-null values.
desc_nulls_last() - Method in class org.apache.spark.sql.Column
Returns a sort expression based on the descending order of the column, and null values appear after non-null values.
desc_nulls_last(String) - Static method in class org.apache.spark.sql.functions
Returns a sort expression based on the descending order of the column, and null values appear after non-null values.
descColumnForPartitionUnsupportedError(SqlBaseParser.DescribeRelationContext) - Static method in class org.apache.spark.sql.errors.QueryParsingErrors
 
descPartitionNotAllowedOnTempView(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
descPartitionNotAllowedOnView(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
describe() - Method in interface org.apache.spark.sql.connector.expressions.Expression
Format the expression as a human readable SQL-like string.
describe(String...) - Method in class org.apache.spark.sql.Dataset
Computes basic statistics for numeric and string columns, including count, mean, stddev, min, and max.
describe(Seq<String>) - Method in class org.apache.spark.sql.Dataset
Computes basic statistics for numeric and string columns, including count, mean, stddev, min, and max.
describeDoesNotSupportPartitionForV2TablesError() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
describeTopics(int) - Method in class org.apache.spark.ml.clustering.LDAModel
Return the topics described by their top-weighted terms.
describeTopics() - Method in class org.apache.spark.ml.clustering.LDAModel
 
describeTopics(int) - Method in class org.apache.spark.mllib.clustering.DistributedLDAModel
 
describeTopics(int) - Method in class org.apache.spark.mllib.clustering.LDAModel
Return the topics described by weighted terms.
describeTopics() - Method in class org.apache.spark.mllib.clustering.LDAModel
Return the topics described by weighted terms.
describeTopics(int) - Method in class org.apache.spark.mllib.clustering.LocalLDAModel
 
description() - Method in class org.apache.spark.ExceptionFailure
 
description() - Method in class org.apache.spark.sql.catalog.CatalogMetadata
 
description() - Method in class org.apache.spark.sql.catalog.Column
 
description() - Method in class org.apache.spark.sql.catalog.Database
 
description() - Method in class org.apache.spark.sql.catalog.Function
 
description() - Method in class org.apache.spark.sql.catalog.Table
 
description() - Method in interface org.apache.spark.sql.connector.catalog.functions.UnboundFunction
Returns Function documentation.
description() - Method in interface org.apache.spark.sql.connector.metric.CustomMetric
Returns the description of custom metric.
description() - Method in interface org.apache.spark.sql.connector.read.Scan
A description string of this scan, which may includes information like: what filters are configured for this scan, what's the value of some important options like path, etc.
description() - Method in interface org.apache.spark.sql.connector.write.RowLevelOperation
Returns the description associated with this row-level operation.
description() - Method in interface org.apache.spark.sql.connector.write.Write
Returns the description associated with this write.
description() - Method in class org.apache.spark.sql.streaming.SinkProgress
 
description() - Method in class org.apache.spark.sql.streaming.SourceProgress
 
description() - Method in class org.apache.spark.status.api.v1.JobData
 
description() - Method in class org.apache.spark.status.api.v1.sql.ExecutionData
 
description() - Method in class org.apache.spark.status.api.v1.StageData
 
description() - Method in class org.apache.spark.status.api.v1.streaming.OutputOperationInfo
 
description() - Method in class org.apache.spark.status.LiveStage
 
description() - Method in class org.apache.spark.storage.StorageLevel
 
description() - Method in class org.apache.spark.streaming.scheduler.OutputOperationInfo
 
DESCRIPTION_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.JobData
 
DESCRIPTION_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress
 
DESCRIPTION_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress
 
DESCRIPTION_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
 
DESCRIPTION_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.StageData
 
descriptorParseError(String, Throwable) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
DESER_CPU_TIME() - Static method in class org.apache.spark.status.TaskIndexNames
 
DESER_TIME() - Static method in class org.apache.spark.status.TaskIndexNames
 
DeserializationStream - Class in org.apache.spark.serializer
:: DeveloperApi :: A stream for reading serialized objects.
DeserializationStream() - Constructor for class org.apache.spark.serializer.DeserializationStream
 
deserialize(Object) - Method in class org.apache.spark.mllib.linalg.VectorUDT
 
deserialize(ByteBuffer, ClassLoader, ClassTag<T>) - Method in class org.apache.spark.serializer.DummySerializerInstance
 
deserialize(ByteBuffer, ClassTag<T>) - Method in class org.apache.spark.serializer.DummySerializerInstance
 
deserialize(ByteBuffer, ClassTag<T>) - Method in class org.apache.spark.serializer.SerializerInstance
 
deserialize(ByteBuffer, ClassLoader, ClassTag<T>) - Method in class org.apache.spark.serializer.SerializerInstance
 
deserialize(Object) - Method in class org.apache.spark.sql.types.UserDefinedType
Convert a SQL datum to the user type
deserialize(List<StoreTypes.AccumulableInfo>) - Static method in class org.apache.spark.status.protobuf.AccumulableInfoSerializer
 
deserialize(StoreTypes.DeterministicLevel) - Static method in class org.apache.spark.status.protobuf.DeterministicLevelSerializer
 
deserialize(StoreTypes.ExecutorMetrics) - Static method in class org.apache.spark.status.protobuf.ExecutorMetricsSerializer
 
deserialize(StoreTypes.ExecutorStageSummary) - Static method in class org.apache.spark.status.protobuf.ExecutorStageSummarySerializer
 
deserialize(StoreTypes.JobExecutionStatus) - Static method in class org.apache.spark.status.protobuf.JobExecutionStatusSerializer
 
deserialize(byte[]) - Method in interface org.apache.spark.status.protobuf.ProtobufSerDe
Deserialize the input Array[Byte] to an object of the type T.
deserialize(StoreTypes.SinkProgress) - Static method in class org.apache.spark.status.protobuf.sql.SinkProgressSerializer
 
deserialize(StoreTypes.SQLPlanMetric) - Static method in class org.apache.spark.status.protobuf.sql.SQLPlanMetricSerializer
 
deserialize(StoreTypes.StreamingQueryProgress) - Static method in class org.apache.spark.status.protobuf.sql.StreamingQueryProgressSerializer
 
deserialize(StoreTypes.StageStatus) - Static method in class org.apache.spark.status.protobuf.StageStatusSerializer
 
deserialize(byte[]) - Static method in class org.apache.spark.util.Utils
Deserialize an object using Java serialization
deserialize(byte[], ClassLoader) - Static method in class org.apache.spark.util.Utils
Deserialize an object using Java serialization and the given ClassLoader
deserialized() - Method in class org.apache.spark.storage.StorageLevel
 
DESERIALIZED_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData
 
DeserializedMemoryEntry<T> - Class in org.apache.spark.storage.memory
 
DeserializedMemoryEntry(Object, long, MemoryMode, ClassTag<T>) - Constructor for class org.apache.spark.storage.memory.DeserializedMemoryEntry
 
DeserializedValuesHolder<T> - Class in org.apache.spark.storage.memory
A holder for storing the deserialized values.
DeserializedValuesHolder(ClassTag<T>, MemoryMode) - Constructor for class org.apache.spark.storage.memory.DeserializedValuesHolder
 
deserializeFromChunkedBuffer(SerializerInstance, ChunkedByteBuffer, ClassTag<T>) - Static method in class org.apache.spark.serializer.SerializerHelper
 
deserializeLongValue(byte[]) - Static method in class org.apache.spark.util.Utils
Deserialize a Long value (used for org.apache.spark.api.python.PythonPartitioner)
deserializeOffset(String) - Method in interface org.apache.spark.sql.connector.read.streaming.SparkDataStream
Deserialize a JSON string into an Offset of the implementation-defined offset type.
deserializer() - Method in interface org.apache.spark.sql.avro.AvroUtils.RowReader
 
deserializeStream(InputStream) - Method in class org.apache.spark.serializer.DummySerializerInstance
 
deserializeStream(InputStream) - Method in class org.apache.spark.serializer.SerializerInstance
 
deserializeToArray(List<StoreTypes.SourceProgress>) - Static method in class org.apache.spark.status.protobuf.sql.SourceProgressSerializer
 
deserializeToArray(List<StoreTypes.StateOperatorProgress>) - Static method in class org.apache.spark.status.protobuf.sql.StateOperatorProgressSerializer
 
deserializeViaNestedStream(InputStream, SerializerInstance, Function1<DeserializationStream, BoxedUnit>) - Static method in class org.apache.spark.util.Utils
Deserialize via nested stream using specific serializer
destroy() - Method in class org.apache.spark.broadcast.Broadcast
Destroy all data and metadata related to this broadcast variable.
destroy() - Method in class org.apache.spark.ui.HttpSecurityFilter
 
details() - Method in class org.apache.spark.scheduler.StageInfo
 
details() - Method in class org.apache.spark.status.api.v1.StageData
 
DETAILS_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
 
DETAILS_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.StageData
 
detailsUINode(boolean, String) - Static method in class org.apache.spark.ui.UIUtils
 
DETERMINATE() - Static method in class org.apache.spark.rdd.DeterministicLevel
 
determineBounds(ArrayBuffer<Tuple2<K, Object>>, int, Ordering<K>, ClassTag<K>) - Static method in class org.apache.spark.RangePartitioner
Determines the bounds for range partitioning from candidates with weights indicating how many items each represents.
deterministic() - Method in class org.apache.spark.sql.expressions.UserDefinedAggregateFunction
Deprecated.
Returns true iff this function is deterministic, i.e.
deterministic() - Method in class org.apache.spark.sql.expressions.UserDefinedFunction
Returns true iff the UDF is deterministic, i.e.
DETERMINISTIC_LEVEL_DETERMINATE_VALUE - Static variable in enum org.apache.spark.status.protobuf.StoreTypes.DeterministicLevel
DETERMINISTIC_LEVEL_DETERMINATE = 1;
DETERMINISTIC_LEVEL_INDETERMINATE_VALUE - Static variable in enum org.apache.spark.status.protobuf.StoreTypes.DeterministicLevel
DETERMINISTIC_LEVEL_INDETERMINATE = 3;
DETERMINISTIC_LEVEL_UNORDERED_VALUE - Static variable in enum org.apache.spark.status.protobuf.StoreTypes.DeterministicLevel
DETERMINISTIC_LEVEL_UNORDERED = 2;
DETERMINISTIC_LEVEL_UNSPECIFIED_VALUE - Static variable in enum org.apache.spark.status.protobuf.StoreTypes.DeterministicLevel
DETERMINISTIC_LEVEL_UNSPECIFIED = 0;
DeterministicLevel - Class in org.apache.spark.rdd
The deterministic level of RDD's output (i.e.
DeterministicLevel() - Constructor for class org.apache.spark.rdd.DeterministicLevel
 
DeterministicLevelSerializer - Class in org.apache.spark.status.protobuf
 
DeterministicLevelSerializer() - Constructor for class org.apache.spark.status.protobuf.DeterministicLevelSerializer
 
deviance(double, double, double) - Method in class org.apache.spark.ml.regression.GeneralizedLinearRegression.Binomial$
 
deviance(double, double, double) - Method in class org.apache.spark.ml.regression.GeneralizedLinearRegression.Gamma$
 
deviance(double, double, double) - Method in class org.apache.spark.ml.regression.GeneralizedLinearRegression.Gaussian$
 
deviance(double, double, double) - Method in class org.apache.spark.ml.regression.GeneralizedLinearRegression.Poisson$
 
deviance() - Method in class org.apache.spark.ml.regression.GeneralizedLinearRegressionSummary
 
devianceResiduals() - Method in class org.apache.spark.ml.regression.LinearRegressionSummary
 
dfToCols(Dataset<Row>) - Static method in class org.apache.spark.sql.api.r.SQLUtils
 
dfToRowRDD(Dataset<Row>) - Static method in class org.apache.spark.sql.api.r.SQLUtils
 
dgemm(double, DenseMatrix<Object>, DenseMatrix<Object>, double, DenseMatrix<Object>) - Static method in class org.apache.spark.ml.ann.BreezeUtil
DGEMM: C := alpha * A * B + beta * C
dgemv(double, DenseMatrix<Object>, DenseVector<Object>, double, DenseVector<Object>) - Static method in class org.apache.spark.ml.ann.BreezeUtil
DGEMV: y := alpha * A * x + beta * y
diag(Vector) - Static method in class org.apache.spark.ml.linalg.DenseMatrix
Generate a diagonal matrix in DenseMatrix format from the supplied values.
diag(Vector) - Static method in class org.apache.spark.ml.linalg.Matrices
Generate a diagonal matrix in Matrix format from the supplied values.
diag(Vector) - Static method in class org.apache.spark.mllib.linalg.DenseMatrix
Generate a diagonal matrix in DenseMatrix format from the supplied values.
diag(Vector) - Static method in class org.apache.spark.mllib.linalg.Matrices
Generate a diagonal matrix in Matrix format from the supplied values.
dialectFunctionName(String) - Method in class org.apache.spark.sql.jdbc.DB2Dialect.DB2SQLBuilder
 
dialectFunctionName(String) - Method in class org.apache.spark.sql.jdbc.MsSqlServerDialect.MsSqlServerSQLBuilder
 
diff(RDD<Tuple2<Object, VD>>) - Method in class org.apache.spark.graphx.impl.VertexRDDImpl
 
diff(VertexRDD<VD>) - Method in class org.apache.spark.graphx.impl.VertexRDDImpl
 
diff(RDD<Tuple2<Object, VD>>) - Method in class org.apache.spark.graphx.VertexRDD
For each vertex present in both this and other, diff returns only those vertices with differing values; for values that are different, keeps the values from other.
diff(VertexRDD<VD>) - Method in class org.apache.spark.graphx.VertexRDD
For each vertex present in both this and other, diff returns only those vertices with differing values; for values that are different, keeps the values from other.
DifferentiableLossAggregator<Datum,Agg extends DifferentiableLossAggregator<Datum,Agg>> - Interface in org.apache.spark.ml.optim.aggregator
A parent trait for aggregators used in fitting MLlib models.
DifferentiableRegularization<T> - Interface in org.apache.spark.ml.optim.loss
A Breeze diff function which represents a cost function for differentiable regularization of parameters.
dim() - Method in interface org.apache.spark.ml.optim.aggregator.DifferentiableLossAggregator
The dimension of the gradient array.
dir() - Method in class org.apache.spark.mllib.optimization.NNLS.Workspace
 
direction() - Method in interface org.apache.spark.sql.connector.expressions.SortOrder
Returns the sort direction.
directory(File) - Method in class org.apache.spark.launcher.SparkLauncher
Sets the working directory of spark-submit.
directoryPathAndOptionsPathBothSpecifiedError(SqlBaseParser.InsertOverwriteDirContext) - Static method in class org.apache.spark.sql.errors.QueryParsingErrors
 
DirectPoolMemory - Class in org.apache.spark.metrics
 
DirectPoolMemory() - Constructor for class org.apache.spark.metrics.DirectPoolMemory
 
disconnect() - Method in interface org.apache.spark.launcher.SparkAppHandle
Disconnects the handle from the application, without stopping it.
discoverResource(ResourceRequest, SparkConf) - Method in interface org.apache.spark.api.resource.ResourceDiscoveryPlugin
Discover the addresses of the requested resource.
discoverResource(ResourceRequest, SparkConf) - Method in class org.apache.spark.resource.ResourceDiscoveryScriptPlugin
 
DISCOVERY_SCRIPT() - Static method in class org.apache.spark.resource.ResourceUtils
 
discoveryScript() - Method in class org.apache.spark.resource.ExecutorResourceRequest
 
discoveryScript() - Method in class org.apache.spark.resource.ResourceRequest
 
DISCOVERYSCRIPT_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest
 
DISK_BYTES_SPILLED() - Static method in class org.apache.spark.InternalAccumulator
 
DISK_BYTES_SPILLED_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
 
DISK_BYTES_SPILLED_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
 
DISK_BYTES_SPILLED_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary
 
DISK_BYTES_SPILLED_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.StageData
 
DISK_BYTES_SPILLED_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
 
DISK_BYTES_SPILLED_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
 
DISK_BYTES_SPILLED_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics
 
DISK_ONLY - Static variable in class org.apache.spark.api.java.StorageLevels
 
DISK_ONLY() - Static method in class org.apache.spark.storage.StorageLevel
 
DISK_ONLY_2 - Static variable in class org.apache.spark.api.java.StorageLevels
 
DISK_ONLY_2() - Static method in class org.apache.spark.storage.StorageLevel
 
DISK_ONLY_3 - Static variable in class org.apache.spark.api.java.StorageLevels
 
DISK_ONLY_3() - Static method in class org.apache.spark.storage.StorageLevel
 
DISK_SIZE_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData
 
DISK_SPILL() - Static method in class org.apache.spark.status.TaskIndexNames
 
DISK_USED_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
 
DISK_USED_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution
 
DISK_USED_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo
 
DISK_USED_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo
 
DiskBlockData - Class in org.apache.spark.storage
 
DiskBlockData(long, long, File, long) - Constructor for class org.apache.spark.storage.DiskBlockData
 
diskBytesSpilled() - Method in class org.apache.spark.status.api.v1.ExecutorMetricsDistributions
 
diskBytesSpilled() - Method in class org.apache.spark.status.api.v1.ExecutorStageSummary
 
diskBytesSpilled() - Method in class org.apache.spark.status.api.v1.StageData
 
diskBytesSpilled() - Method in class org.apache.spark.status.api.v1.TaskMetricDistributions
 
diskBytesSpilled() - Method in class org.apache.spark.status.api.v1.TaskMetrics
 
diskSize() - Method in class org.apache.spark.storage.BlockManagerMessages.UpdateBlockInfo
 
diskSize() - Method in class org.apache.spark.storage.BlockStatus
 
diskSize() - Method in class org.apache.spark.storage.BlockUpdatedInfo
 
diskSize() - Method in class org.apache.spark.storage.RDDInfo
 
diskUsed() - Method in class org.apache.spark.status.api.v1.ExecutorSummary
 
diskUsed() - Method in class org.apache.spark.status.api.v1.RDDDataDistribution
 
diskUsed() - Method in class org.apache.spark.status.api.v1.RDDPartitionInfo
 
diskUsed() - Method in class org.apache.spark.status.api.v1.RDDStorageInfo
 
diskUsed() - Method in class org.apache.spark.status.LiveRDD
 
diskUsed() - Method in class org.apache.spark.status.LiveRDDDistribution
 
diskUsed() - Method in class org.apache.spark.status.LiveRDDPartition
 
dispersion() - Method in class org.apache.spark.ml.regression.GeneralizedLinearRegressionSummary
 
displayOrder() - Method in interface org.apache.spark.status.AppHistoryServerPlugin
The position of a plugin tab relative to the other plugin tabs in the history UI.
dispose() - Method in interface org.apache.spark.storage.BlockData
 
dispose() - Method in class org.apache.spark.storage.DiskBlockData
 
dispose(ByteBuffer) - Static method in class org.apache.spark.storage.StorageUtils
Attempt to clean up a ByteBuffer if it is direct or memory-mapped.
distanceMeasure() - Method in class org.apache.spark.ml.clustering.BisectingKMeans
 
distanceMeasure() - Method in class org.apache.spark.ml.clustering.BisectingKMeansModel
 
distanceMeasure() - Method in class org.apache.spark.ml.clustering.KMeans
 
distanceMeasure() - Method in class org.apache.spark.ml.clustering.KMeansAggregator
 
distanceMeasure() - Method in class org.apache.spark.ml.clustering.KMeansModel
 
distanceMeasure() - Method in class org.apache.spark.ml.evaluation.ClusteringEvaluator
param for distance measure to be used in evaluation (supports "squaredEuclidean" (default), "cosine")
distanceMeasure() - Method in interface org.apache.spark.ml.param.shared.HasDistanceMeasure
Param for The distance measure.
distanceMeasure() - Method in class org.apache.spark.mllib.clustering.BisectingKMeansModel
 
distanceMeasure() - Method in class org.apache.spark.mllib.clustering.KMeansModel
 
distinct() - Method in class org.apache.spark.api.java.JavaDoubleRDD
Return a new RDD containing the distinct elements in this RDD.
distinct(int) - Method in class org.apache.spark.api.java.JavaDoubleRDD
Return a new RDD containing the distinct elements in this RDD.
distinct() - Method in class org.apache.spark.api.java.JavaPairRDD
Return a new RDD containing the distinct elements in this RDD.
distinct(int) - Method in class org.apache.spark.api.java.JavaPairRDD
Return a new RDD containing the distinct elements in this RDD.
distinct() - Method in class org.apache.spark.api.java.JavaRDD
Return a new RDD containing the distinct elements in this RDD.
distinct(int) - Method in class org.apache.spark.api.java.JavaRDD
Return a new RDD containing the distinct elements in this RDD.
distinct(int, Ordering<T>) - Method in class org.apache.spark.rdd.RDD
Return a new RDD containing the distinct elements in this RDD.
distinct() - Method in class org.apache.spark.rdd.RDD
Return a new RDD containing the distinct elements in this RDD.
distinct() - Method in class org.apache.spark.sql.Dataset
Returns a new Dataset that contains only the unique rows from this Dataset.
distinct(Column...) - Method in class org.apache.spark.sql.expressions.UserDefinedAggregateFunction
Deprecated.
Creates a Column for this UDAF using the distinct values of the given Columns as input arguments.
distinct(Seq<Column>) - Method in class org.apache.spark.sql.expressions.UserDefinedAggregateFunction
Deprecated.
Creates a Column for this UDAF using the distinct values of the given Columns as input arguments.
distinctCount() - Method in interface org.apache.spark.sql.connector.read.colstats.ColumnStatistics
 
distributeByUnsupportedError(SqlBaseParser.QueryOrganizationContext) - Static method in class org.apache.spark.sql.errors.QueryParsingErrors
 
DistributedLDAModel - Class in org.apache.spark.ml.clustering
Distributed model fitted by LDA.
DistributedLDAModel - Class in org.apache.spark.mllib.clustering
Distributed LDA model.
DistributedMatrix - Interface in org.apache.spark.mllib.linalg.distributed
Represents a distributively stored matrix backed by one or more RDDs.
Distribution - Interface in org.apache.spark.sql.connector.distributions
An interface that defines how data is distributed across partitions.
distribution(LiveExecutor) - Method in class org.apache.spark.status.LiveRDD
 
distributionOpt(LiveExecutor) - Method in class org.apache.spark.status.LiveRDD
 
Distributions - Class in org.apache.spark.sql.connector.distributions
Helper methods to create distributions to pass into Spark.
distributionStrictlyRequired() - Method in interface org.apache.spark.sql.connector.write.RequiresDistributionAndOrdering
Returns if the distribution required by this write is strictly required or best effort only.
div(Decimal, Decimal) - Method in class org.apache.spark.sql.types.Decimal.DecimalIsFractional$
 
div(double, double) - Static method in class org.apache.spark.sql.types.DoubleExactNumeric
 
div(float, float) - Static method in class org.apache.spark.sql.types.FloatExactNumeric
 
div(Duration) - Method in class org.apache.spark.streaming.Duration
 
divide(Object) - Method in class org.apache.spark.sql.Column
Division this expression by another expression.
divideByZeroError(SQLQueryContext) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
doc() - Method in class org.apache.spark.ml.param.Param
 
docConcentration() - Method in class org.apache.spark.ml.clustering.LDA
 
docConcentration() - Method in class org.apache.spark.ml.clustering.LDAModel
 
docConcentration() - Method in interface org.apache.spark.ml.clustering.LDAParams
Concentration parameter (commonly named "alpha") for the prior placed on documents' distributions over topics ("theta").
docConcentration() - Method in class org.apache.spark.mllib.clustering.DistributedLDAModel
 
docConcentration() - Method in class org.apache.spark.mllib.clustering.LDAModel
Concentration parameter (commonly named "alpha") for the prior placed on documents' distributions over topics ("theta").
docConcentration() - Method in class org.apache.spark.mllib.clustering.LocalLDAModel
 
docFreq() - Method in class org.apache.spark.ml.feature.IDFModel
Returns the document frequency
docFreq() - Method in class org.apache.spark.mllib.feature.IDFModel
 
DocumentFrequencyAggregator(int) - Constructor for class org.apache.spark.mllib.feature.IDF.DocumentFrequencyAggregator
 
DocumentFrequencyAggregator() - Constructor for class org.apache.spark.mllib.feature.IDF.DocumentFrequencyAggregator
 
doesDirectoryContainAnyNewFiles(File, long) - Static method in class org.apache.spark.util.Utils
Determines if a directory contains any files newer than cutoff seconds.
doExecuteBroadcastNotImplementedError(String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
doFetchFile(String, File, String, SparkConf, Configuration) - Static method in class org.apache.spark.util.Utils
Download a file or directory to target directory.
doFilter(ServletRequest, ServletResponse, FilterChain) - Method in class org.apache.spark.ui.HttpSecurityFilter
 
doPostEvent(SparkListenerInterface, SparkListenerEvent) - Method in interface org.apache.spark.scheduler.SparkListenerBus
 
doPostEvent(L, E) - Method in interface org.apache.spark.util.ListenerBus
Post an event to the specified listener.
Dot - Class in org.apache.spark.ml.feature
 
Dot() - Constructor for class org.apache.spark.ml.feature.Dot
 
dot(Vector, Vector) - Static method in class org.apache.spark.ml.linalg.BLAS
dot(x, y)
dot(Vector) - Method in interface org.apache.spark.ml.linalg.Vector
Calculate the dot product of this vector with another.
dot(Vector, Vector) - Static method in class org.apache.spark.mllib.linalg.BLAS
dot(x, y)
dot(Vector) - Method in interface org.apache.spark.mllib.linalg.Vector
Calculate the dot product of this vector with another.
doTest(DStream<Tuple2<StatCounter, StatCounter>>) - Method in interface org.apache.spark.mllib.stat.test.StreamingTestMethod
Perform streaming 2-sample statistical significance testing.
doTest(DStream<Tuple2<StatCounter, StatCounter>>) - Static method in class org.apache.spark.mllib.stat.test.StudentTTest
 
doTest(DStream<Tuple2<StatCounter, StatCounter>>) - Static method in class org.apache.spark.mllib.stat.test.WelchTTest
 
DOUBLE() - Static method in class org.apache.spark.sql.Encoders
An encoder for nullable double type.
doubleAccumulator() - Method in class org.apache.spark.SparkContext
Create and register a double accumulator, which starts with 0 and accumulates inputs by add.
doubleAccumulator(String) - Method in class org.apache.spark.SparkContext
Create and register a double accumulator, which starts with 0 and accumulates inputs by add.
DoubleAccumulator - Class in org.apache.spark.util
An accumulator for computing sum, count, and averages for double precision floating numbers.
DoubleAccumulator() - Constructor for class org.apache.spark.util.DoubleAccumulator
 
DoubleAccumulatorSource - Class in org.apache.spark.metrics.source
 
DoubleAccumulatorSource() - Constructor for class org.apache.spark.metrics.source.DoubleAccumulatorSource
 
DoubleArrayArrayParam - Class in org.apache.spark.ml.param
Specialized version of Param[Array[Array[Double}] for Java.
DoubleArrayArrayParam(Params, String, String, Function1<double[][], Object>) - Constructor for class org.apache.spark.ml.param.DoubleArrayArrayParam
 
DoubleArrayArrayParam(Params, String, String) - Constructor for class org.apache.spark.ml.param.DoubleArrayArrayParam
 
DoubleArrayParam - Class in org.apache.spark.ml.param
Specialized version of Param[Array[Double} for Java.
DoubleArrayParam(Params, String, String, Function1<double[], Object>) - Constructor for class org.apache.spark.ml.param.DoubleArrayParam
 
DoubleArrayParam(Params, String, String) - Constructor for class org.apache.spark.ml.param.DoubleArrayParam
 
DoubleAsIfIntegral$() - Constructor for class org.apache.spark.sql.types.DoubleType.DoubleAsIfIntegral$
 
doubleColumn(String[]) - Static method in class org.apache.parquet.filter2.predicate.SparkFilterApi
 
DoubleExactNumeric - Class in org.apache.spark.sql.types
 
DoubleExactNumeric() - Constructor for class org.apache.spark.sql.types.DoubleExactNumeric
 
DoubleFlatMapFunction<T> - Interface in org.apache.spark.api.java.function
A function that returns zero or more records of type Double from each input record.
DoubleFunction<T> - Interface in org.apache.spark.api.java.function
A function that returns Doubles, and can be used to construct DoubleRDDs.
DoubleParam - Class in org.apache.spark.ml.param
Specialized version of Param[Double] for Java.
DoubleParam(String, String, String, Function1<Object, Object>) - Constructor for class org.apache.spark.ml.param.DoubleParam
 
DoubleParam(String, String, String) - Constructor for class org.apache.spark.ml.param.DoubleParam
 
DoubleParam(Identifiable, String, String, Function1<Object, Object>) - Constructor for class org.apache.spark.ml.param.DoubleParam
 
DoubleParam(Identifiable, String, String) - Constructor for class org.apache.spark.ml.param.DoubleParam
 
DoubleRDDFunctions - Class in org.apache.spark.rdd
Extra functions available on RDDs of Doubles through an implicit conversion.
DoubleRDDFunctions(RDD<Object>) - Constructor for class org.apache.spark.rdd.DoubleRDDFunctions
 
doubleRDDToDoubleRDDFunctions(RDD<Object>) - Static method in class org.apache.spark.rdd.RDD
 
DoubleType - Static variable in class org.apache.spark.sql.types.DataTypes
Gets the DoubleType object.
DoubleType - Class in org.apache.spark.sql.types
The data type representing Double values.
DoubleType() - Constructor for class org.apache.spark.sql.types.DoubleType
 
DoubleType.DoubleAsIfIntegral - Interface in org.apache.spark.sql.types
 
DoubleType.DoubleAsIfIntegral$ - Class in org.apache.spark.sql.types
 
DoubleType.DoubleIsConflicted - Interface in org.apache.spark.sql.types
 
downloadFile(String, File, SparkConf, Configuration) - Static method in class org.apache.spark.util.DependencyUtils
Download a file from the remote to a local temporary directory.
downloadFileList(String, File, SparkConf, Configuration) - Static method in class org.apache.spark.util.DependencyUtils
Download a list of remote files to temp local files.
DRIVER() - Static method in class org.apache.spark.metrics.MetricsSystemInstances
 
driver() - Method in class org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.SetupDriver
 
driver() - Method in interface org.apache.spark.shuffle.api.ShuffleDataIO
Called once on driver process to bootstrap the shuffle metadata modules that are maintained by the driver.
DRIVER_DEFAULT_JAVA_OPTIONS - Static variable in class org.apache.spark.launcher.SparkLauncher
Configuration key for the default driver VM options.
DRIVER_EXTRA_CLASSPATH - Static variable in class org.apache.spark.launcher.SparkLauncher
Configuration key for the driver class path.
DRIVER_EXTRA_JAVA_OPTIONS - Static variable in class org.apache.spark.launcher.SparkLauncher
Configuration key for the driver VM options.
DRIVER_EXTRA_LIBRARY_PATH - Static variable in class org.apache.spark.launcher.SparkLauncher
Configuration key for the driver native library path.
DRIVER_MEMORY - Static variable in class org.apache.spark.launcher.SparkLauncher
Configuration key for the driver memory.
driverAttributes() - Method in class org.apache.spark.scheduler.SparkListenerApplicationStart
 
driverLogs() - Method in class org.apache.spark.scheduler.SparkListenerApplicationStart
 
DriverPlugin - Interface in org.apache.spark.api.plugin
:: DeveloperApi :: Driver component of a SparkPlugin.
driverPlugin() - Method in interface org.apache.spark.api.plugin.SparkPlugin
Return the plugin's driver-side component.
drop() - Method in class org.apache.spark.sql.DataFrameNaFunctions
Returns a new DataFrame that drops rows containing any null or NaN values.
drop(String) - Method in class org.apache.spark.sql.DataFrameNaFunctions
Returns a new DataFrame that drops rows containing null or NaN values.
drop(String[]) - Method in class org.apache.spark.sql.DataFrameNaFunctions
Returns a new DataFrame that drops rows containing any null or NaN values in the specified columns.
drop(Seq<String>) - Method in class org.apache.spark.sql.DataFrameNaFunctions
(Scala-specific) Returns a new DataFrame that drops rows containing any null or NaN values in the specified columns.
drop(String, String[]) - Method in class org.apache.spark.sql.DataFrameNaFunctions
Returns a new DataFrame that drops rows containing null or NaN values in the specified columns.
drop(String, Seq<String>) - Method in class org.apache.spark.sql.DataFrameNaFunctions
(Scala-specific) Returns a new DataFrame that drops rows containing null or NaN values in the specified columns.
drop(int) - Method in class org.apache.spark.sql.DataFrameNaFunctions
Returns a new DataFrame that drops rows containing less than minNonNulls non-null and non-NaN values.
drop(int, String[]) - Method in class org.apache.spark.sql.DataFrameNaFunctions
Returns a new DataFrame that drops rows containing less than minNonNulls non-null and non-NaN values in the specified columns.
drop(int, Seq<String>) - Method in class org.apache.spark.sql.DataFrameNaFunctions
(Scala-specific) Returns a new DataFrame that drops rows containing less than minNonNulls non-null and non-NaN values in the specified columns.
drop(String...) - Method in class org.apache.spark.sql.Dataset
Returns a new Dataset with columns dropped.
drop(Column, Column...) - Method in class org.apache.spark.sql.Dataset
Returns a new Dataset with columns dropped.
drop(String) - Method in class org.apache.spark.sql.Dataset
Returns a new Dataset with a column dropped.
drop(Seq<String>) - Method in class org.apache.spark.sql.Dataset
Returns a new Dataset with columns dropped.
drop(Column) - Method in class org.apache.spark.sql.Dataset
Returns a new Dataset with column dropped.
drop(Column, Seq<Column>) - Method in class org.apache.spark.sql.Dataset
Returns a new Dataset with columns dropped.
dropDuplicates(String, String...) - Method in class org.apache.spark.sql.Dataset
Returns a new Dataset with duplicate rows removed, considering only the subset of columns.
dropDuplicates() - Method in class org.apache.spark.sql.Dataset
Returns a new Dataset that contains only the unique rows from this Dataset.
dropDuplicates(Seq<String>) - Method in class org.apache.spark.sql.Dataset
(Scala-specific) Returns a new Dataset with duplicate rows removed, considering only the subset of columns.
dropDuplicates(String[]) - Method in class org.apache.spark.sql.Dataset
Returns a new Dataset with duplicate rows removed, considering only the subset of columns.
dropDuplicates(String, Seq<String>) - Method in class org.apache.spark.sql.Dataset
Returns a new Dataset with duplicate rows removed, considering only the subset of columns.
dropFields(Seq<String>) - Method in class org.apache.spark.sql.Column
An expression that drops fields in StructType by name.
dropFromMemory(BlockId, Function0<Either<Object, ChunkedByteBuffer>>, ClassTag<T>) - Method in interface org.apache.spark.storage.memory.BlockEvictionHandler
Drop a block from memory, possibly putting it on disk if applicable.
dropGlobalTempView(String) - Method in class org.apache.spark.sql.catalog.Catalog
Drops the global temporary view with the given view name in the catalog.
dropIndex(String) - Method in interface org.apache.spark.sql.connector.catalog.index.SupportsIndex
Drops the index with the given name.
dropIndex(String, Identifier) - Static method in class org.apache.spark.sql.jdbc.DB2Dialect
 
dropIndex(String, Identifier) - Static method in class org.apache.spark.sql.jdbc.DerbyDialect
 
dropIndex(String, Identifier) - Static method in class org.apache.spark.sql.jdbc.H2Dialect
 
dropIndex(String, Identifier) - Method in class org.apache.spark.sql.jdbc.JdbcDialect
Build a drop index SQL statement.
dropIndex(String, Identifier) - Static method in class org.apache.spark.sql.jdbc.MsSqlServerDialect
 
dropIndex(String, Identifier) - Static method in class org.apache.spark.sql.jdbc.MySQLDialect
 
dropIndex(String, Identifier) - Static method in class org.apache.spark.sql.jdbc.NoopDialect
 
dropIndex(String, Identifier) - Static method in class org.apache.spark.sql.jdbc.OracleDialect
 
dropIndex(String, Identifier) - Static method in class org.apache.spark.sql.jdbc.PostgresDialect
 
dropIndex(String, Identifier) - Static method in class org.apache.spark.sql.jdbc.TeradataDialect
 
dropLast() - Method in class org.apache.spark.ml.feature.OneHotEncoder
 
dropLast() - Method in interface org.apache.spark.ml.feature.OneHotEncoderBase
Whether to drop the last category in the encoded vector (default: true)
dropLast() - Method in class org.apache.spark.ml.feature.OneHotEncoderModel
 
dropNamespace(String[], boolean) - Method in class org.apache.spark.sql.connector.catalog.DelegatingCatalogExtension
 
dropNamespace(String[], boolean) - Method in interface org.apache.spark.sql.connector.catalog.SupportsNamespaces
Drop a namespace from the catalog with cascade mode, recursively dropping all objects within the namespace if cascade is true.
dropNonExistentColumnsNotSupportedError(Seq<String>) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
dropPartition(InternalRow) - Method in interface org.apache.spark.sql.connector.catalog.SupportsAtomicPartitionManagement
 
dropPartition(InternalRow) - Method in interface org.apache.spark.sql.connector.catalog.SupportsPartitionManagement
Drop a partition from table.
dropPartitions(InternalRow[]) - Method in interface org.apache.spark.sql.connector.catalog.SupportsAtomicPartitionManagement
Drop an array of partitions atomically from table.
dropSchema(String, boolean) - Static method in class org.apache.spark.sql.jdbc.DB2Dialect
 
dropSchema(String, boolean) - Static method in class org.apache.spark.sql.jdbc.DerbyDialect
 
dropSchema(String, boolean) - Static method in class org.apache.spark.sql.jdbc.H2Dialect
 
dropSchema(String, boolean) - Method in class org.apache.spark.sql.jdbc.JdbcDialect
 
dropSchema(String, boolean) - Static method in class org.apache.spark.sql.jdbc.MsSqlServerDialect
 
dropSchema(String, boolean) - Static method in class org.apache.spark.sql.jdbc.MySQLDialect
 
dropSchema(String, boolean) - Static method in class org.apache.spark.sql.jdbc.NoopDialect
 
dropSchema(String, boolean) - Static method in class org.apache.spark.sql.jdbc.OracleDialect
 
dropSchema(String, boolean) - Static method in class org.apache.spark.sql.jdbc.PostgresDialect
 
dropSchema(String, boolean) - Static method in class org.apache.spark.sql.jdbc.TeradataDialect
 
dropTable(Identifier) - Method in class org.apache.spark.sql.connector.catalog.DelegatingCatalogExtension
 
dropTable(Identifier) - Method in interface org.apache.spark.sql.connector.catalog.TableCatalog
Drop a table in the catalog.
dropTableWithPurgeUnsupportedError() - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
dropTempTable(String) - Method in class org.apache.spark.sql.SQLContext
 
dropTempView(String) - Method in class org.apache.spark.sql.catalog.Catalog
Drops the local temporary view with the given view name in the catalog.
dropView(Identifier) - Method in interface org.apache.spark.sql.connector.catalog.ViewCatalog
Drop a view in the catalog.
dspmv(int, double, DenseVector, DenseVector, double, DenseVector) - Static method in class org.apache.spark.ml.linalg.BLAS
y := alpha*A*x + beta*y
Dst - Static variable in class org.apache.spark.graphx.TripletFields
Expose the destination and edge fields but not the source field.
dstAttr() - Method in class org.apache.spark.graphx.EdgeContext
The vertex attribute of the edge's destination vertex.
dstAttr() - Method in class org.apache.spark.graphx.EdgeTriplet
The destination vertex attribute
dstAttr() - Method in class org.apache.spark.graphx.impl.AggregatingEdgeContext
 
dstCol() - Method in class org.apache.spark.ml.clustering.PowerIterationClustering
 
dstCol() - Method in interface org.apache.spark.ml.clustering.PowerIterationClusteringParams
Name of the input column for destination vertex IDs.
dstId() - Method in class org.apache.spark.graphx.Edge
 
dstId() - Method in class org.apache.spark.graphx.EdgeContext
The vertex id of the edge's destination vertex.
dstId() - Method in class org.apache.spark.graphx.impl.AggregatingEdgeContext
 
dstream() - Method in class org.apache.spark.streaming.api.java.JavaDStream
 
dstream() - Method in interface org.apache.spark.streaming.api.java.JavaDStreamLike
 
dstream() - Method in class org.apache.spark.streaming.api.java.JavaPairDStream
 
DStream<T> - Class in org.apache.spark.streaming.dstream
A Discretized Stream (DStream), the basic abstraction in Spark Streaming, is a continuous sequence of RDDs (of the same type) representing a continuous stream of data (see org.apache.spark.rdd.RDD in the Spark core documentation for more details on RDDs).
DStream(StreamingContext, ClassTag<T>) - Constructor for class org.apache.spark.streaming.dstream.DStream
 
dtypes() - Method in class org.apache.spark.sql.Dataset
Returns all column names and their data types as an array.
DummySerializerInstance - Class in org.apache.spark.serializer
Unfortunately, we need a serializer instance in order to construct a DiskBlockObjectWriter.
duplicateClausesError(String, ParserRuleContext) - Static method in class org.apache.spark.sql.errors.QueryParsingErrors
 
duplicateCreateTableColumnOption(ParserRuleContext, String, String) - Static method in class org.apache.spark.sql.errors.QueryParsingErrors
 
duplicateCteDefinitionNamesError(String, SqlBaseParser.CtesContext) - Static method in class org.apache.spark.sql.errors.QueryParsingErrors
 
duplicatedTablePathsFoundError(String, String, ParserRuleContext) - Static method in class org.apache.spark.sql.errors.QueryParsingErrors
 
duplicateKeysError(String, ParserRuleContext) - Static method in class org.apache.spark.sql.errors.QueryParsingErrors
 
duplicateMapKeyFoundError(Object) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
duration() - Method in class org.apache.spark.scheduler.TaskInfo
 
DURATION() - Static method in class org.apache.spark.sql.Encoders
Creates an encoder that serializes instances of the java.time.Duration class to the internal representation of nullable Catalyst's DayTimeIntervalType.
duration() - Method in class org.apache.spark.status.api.v1.ApplicationAttemptInfo
 
duration() - Method in class org.apache.spark.status.api.v1.sql.ExecutionData
 
duration() - Method in class org.apache.spark.status.api.v1.streaming.OutputOperationInfo
 
duration() - Method in class org.apache.spark.status.api.v1.TaskData
 
duration() - Method in class org.apache.spark.status.api.v1.TaskMetricDistributions
 
DURATION() - Static method in class org.apache.spark.status.TaskIndexNames
 
Duration - Class in org.apache.spark.streaming
 
Duration(long) - Constructor for class org.apache.spark.streaming.Duration
 
duration() - Method in class org.apache.spark.streaming.scheduler.OutputOperationInfo
Return the duration of this output operation.
DURATION() - Static method in class org.apache.spark.ui.ToolTips
 
DURATION_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo
 
DURATION_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
 
DURATION_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.TaskData
 
DURATION_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
 
DURATION_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
 
DURATION_MS_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
 
durationCalledOnUnfinishedTaskError() - Static method in class org.apache.spark.errors.SparkCoreErrors
 
durationDataPadding(Tuple2<Object, Map<String, Long>>[]) - Static method in class org.apache.spark.ui.UIUtils
There may be different duration labels in each batch.
durationMs() - Method in class org.apache.spark.sql.streaming.StreamingQueryProgress
 
Durations - Class in org.apache.spark.streaming
 
Durations() - Constructor for class org.apache.spark.streaming.Durations
 
dynamicPartitionKeyNotAmongWrittenPartitionPathsError(String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
dynamicPartitionOverwriteUnsupportedByTableError(Table) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 

E

Edge<ED> - Class in org.apache.spark.graphx
A single directed edge consisting of a source id, target id, and the data associated with the edge.
Edge(long, long, ED) - Constructor for class org.apache.spark.graphx.Edge
 
EdgeActiveness - Enum in org.apache.spark.graphx.impl
Criteria for filtering edges based on activeness.
EdgeContext<VD,ED,A> - Class in org.apache.spark.graphx
Represents an edge along with its neighboring vertices and allows sending messages along the edge.
EdgeContext() - Constructor for class org.apache.spark.graphx.EdgeContext
 
EdgeDirection - Class in org.apache.spark.graphx
The direction of a directed edge relative to a vertex.
EdgeDirection() - Constructor for class org.apache.spark.graphx.EdgeDirection
 
edgeListFile(SparkContext, String, boolean, int, StorageLevel, StorageLevel) - Static method in class org.apache.spark.graphx.GraphLoader
Loads a graph from an edge list formatted file where each line contains two integers: a source id and a target id.
EdgeOnly - Static variable in class org.apache.spark.graphx.TripletFields
Expose only the edge field and not the source or destination field.
EdgePartition1D$() - Constructor for class org.apache.spark.graphx.PartitionStrategy.EdgePartition1D$
 
EdgePartition2D$() - Constructor for class org.apache.spark.graphx.PartitionStrategy.EdgePartition2D$
 
EdgeRDD<ED> - Class in org.apache.spark.graphx
EdgeRDD[ED, VD] extends RDD[Edge[ED} by storing the edges in columnar format on each partition for performance.
EdgeRDD(SparkContext, Seq<Dependency<?>>) - Constructor for class org.apache.spark.graphx.EdgeRDD
 
EdgeRDDImpl<ED,VD> - Class in org.apache.spark.graphx.impl
 
edges() - Method in class org.apache.spark.graphx.Graph
An RDD containing the edges and their associated attributes.
edges() - Method in class org.apache.spark.graphx.impl.GraphImpl
 
edges() - Method in class org.apache.spark.status.api.v1.sql.ExecutionData
 
EDGES_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper
 
EDGES_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper
 
EdgeTriplet<VD,ED> - Class in org.apache.spark.graphx
An edge triplet represents an edge along with the vertex attributes of its neighboring vertices.
EdgeTriplet() - Constructor for class org.apache.spark.graphx.EdgeTriplet
 
EigenValueDecomposition - Class in org.apache.spark.mllib.linalg
Compute eigen-decomposition.
EigenValueDecomposition() - Constructor for class org.apache.spark.mllib.linalg.EigenValueDecomposition
 
Either() - Static method in class org.apache.spark.graphx.EdgeDirection
Edges originating from *or* arriving at a vertex of interest.
elasticNetParam() - Method in class org.apache.spark.ml.classification.LogisticRegression
 
elasticNetParam() - Method in class org.apache.spark.ml.classification.LogisticRegressionModel
 
elasticNetParam() - Method in interface org.apache.spark.ml.param.shared.HasElasticNetParam
Param for the ElasticNet mixing parameter, in range [0, 1].
elasticNetParam() - Method in class org.apache.spark.ml.regression.LinearRegression
 
elasticNetParam() - Method in class org.apache.spark.ml.regression.LinearRegressionModel
 
elem(String, Function1<Object, Object>) - Static method in class org.apache.spark.ml.feature.RFormulaParser
 
elem(Parsers) - Static method in class org.apache.spark.ml.feature.RFormulaParser
 
element_at(Column, Object) - Static method in class org.apache.spark.sql.functions
Returns element of array at given index in value if column is array.
elementsOfTupleExceedLimitError() - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
elementType() - Method in class org.apache.spark.sql.types.ArrayType
 
ElementwiseProduct - Class in org.apache.spark.ml.feature
Outputs the Hadamard product (i.e., the element-wise product) of each input vector with a provided "weight" vector.
ElementwiseProduct(String) - Constructor for class org.apache.spark.ml.feature.ElementwiseProduct
 
ElementwiseProduct() - Constructor for class org.apache.spark.ml.feature.ElementwiseProduct
 
ElementwiseProduct - Class in org.apache.spark.mllib.feature
Outputs the Hadamard product (i.e., the element-wise product) of each input vector with a provided "weight" vector.
ElementwiseProduct(Vector) - Constructor for class org.apache.spark.mllib.feature.ElementwiseProduct
 
elems() - Method in class org.apache.spark.status.api.v1.StackTrace
 
EMLDAOptimizer - Class in org.apache.spark.mllib.clustering
Optimizer for EM algorithm which stores data + parameter graph, plus algorithm parameters.
EMLDAOptimizer() - Constructor for class org.apache.spark.mllib.clustering.EMLDAOptimizer
 
empty() - Static method in class org.apache.spark.api.java.Optional
 
empty() - Static method in class org.apache.spark.ml.param.ParamMap
Returns an empty param map.
empty() - Method in class org.apache.spark.mllib.fpm.PrefixSpan.Prefix$
An empty Prefix instance.
empty() - Static method in class org.apache.spark.sql.types.Metadata
Returns an empty Metadata.
empty() - Static method in class org.apache.spark.sql.util.CaseInsensitiveStringMap
 
empty() - Static method in class org.apache.spark.storage.BlockStatus
 
EMPTY_EXPRESSION - Static variable in interface org.apache.spark.sql.connector.expressions.Expression
 
EMPTY_NAMED_REFERENCE - Static variable in interface org.apache.spark.sql.connector.expressions.Expression
`EMPTY_EXPRESSION` is only used as an input when the default `references` method builds the result array to avoid repeatedly allocating an empty array.
EMPTY_USER_GROUPS() - Static method in class org.apache.spark.util.Utils
 
emptyCollectionError() - Static method in class org.apache.spark.errors.SparkCoreErrors
 
emptyDataFrame() - Method in class org.apache.spark.sql.SparkSession
 
emptyDataFrame() - Method in class org.apache.spark.sql.SQLContext
Returns a DataFrame with no rows or columns.
emptyDataset(Encoder<T>) - Method in class org.apache.spark.sql.SparkSession
Creates a new Dataset of type T containing zero elements.
emptyInputForTableSampleError(ParserRuleContext) - Static method in class org.apache.spark.sql.errors.QueryParsingErrors
 
emptyJsonFieldValueError(DataType) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
emptyMultipartIdentifierError() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
emptyNode(int) - Static method in class org.apache.spark.mllib.tree.model.Node
Return a node with the given node id (but nothing else set).
emptyOptionError(String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
emptyPartitionKeyError(String, SqlBaseParser.PartitionSpecContext) - Static method in class org.apache.spark.sql.errors.QueryParsingErrors
 
emptyRDD() - Method in class org.apache.spark.api.java.JavaSparkContext
Get an RDD that has no partitions or elements.
emptyRDD(ClassTag<T>) - Method in class org.apache.spark.SparkContext
Get an RDD that has no partitions or elements.
emptyRDDError() - Static method in class org.apache.spark.errors.SparkCoreErrors
 
emptySourceForMergeError(SqlBaseParser.MergeIntoTableContext) - Static method in class org.apache.spark.sql.errors.QueryParsingErrors
 
EmptyTerm - Class in org.apache.spark.ml.feature
Placeholder term for the result of undefined interactions, e.g.
EmptyTerm() - Constructor for class org.apache.spark.ml.feature.EmptyTerm
 
emptyWindowExpressionError(Window) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
enableHiveSupport() - Method in class org.apache.spark.sql.SparkSession.Builder
Enables Hive support, including connectivity to a persistent Hive metastore, support for Hive serdes, and Hive user-defined functions.
enableReceiverLog(SparkConf) - Static method in class org.apache.spark.streaming.util.WriteAheadLogUtils
 
encode(Column, String) - Static method in class org.apache.spark.sql.functions
Computes the first argument into a binary from a string using the provided character set (one of 'US-ASCII', 'ISO-8859-1', 'UTF-8', 'UTF-16BE', 'UTF-16LE', 'UTF-16').
encodeFileNameToURIRawPath(String) - Static method in class org.apache.spark.util.Utils
A file name may contain some invalid URI characters, such as " ".
encoder() - Method in class org.apache.spark.sql.Dataset
 
Encoder<T> - Interface in org.apache.spark.sql
Used to convert a JVM object of type T to and from the internal Spark SQL representation.
Encoders - Class in org.apache.spark.sql
Methods for creating an Encoder.
Encoders() - Constructor for class org.apache.spark.sql.Encoders
 
END_OFFSET_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress
 
END_TIME_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo
 
END_TIMESTAMP_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData
 
endField() - Method in class org.apache.spark.sql.types.DayTimeIntervalType
 
endField() - Method in class org.apache.spark.sql.types.YearMonthIntervalType
 
endOffset() - Method in class org.apache.spark.sql.streaming.SourceProgress
 
endOffset() - Method in exception org.apache.spark.sql.streaming.StreamingQueryException
 
endOfIteratorError() - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
endOfStreamError() - Static method in class org.apache.spark.errors.SparkCoreErrors
 
endOfStreamError() - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
endReduceId() - Method in class org.apache.spark.storage.ShuffleBlockBatchId
 
endsWith(Column) - Method in class org.apache.spark.sql.Column
String ends with.
endsWith(String) - Method in class org.apache.spark.sql.Column
String ends with another string literal.
endTime() - Method in class org.apache.spark.status.api.v1.ApplicationAttemptInfo
 
endTime() - Method in class org.apache.spark.status.api.v1.streaming.OutputOperationInfo
 
endTime() - Method in class org.apache.spark.streaming.scheduler.OutputOperationInfo
 
EnsembleCombiningStrategy - Class in org.apache.spark.mllib.tree.configuration
Enum to select ensemble combining strategy for base learners
EnsembleCombiningStrategy() - Constructor for class org.apache.spark.mllib.tree.configuration.EnsembleCombiningStrategy
 
EnsembleModelReadWrite - Class in org.apache.spark.ml.tree
 
EnsembleModelReadWrite() - Constructor for class org.apache.spark.ml.tree.EnsembleModelReadWrite
 
EnsembleModelReadWrite.EnsembleNodeData - Class in org.apache.spark.ml.tree
Info for one Node in a tree ensemble
EnsembleModelReadWrite.EnsembleNodeData$ - Class in org.apache.spark.ml.tree
 
EnsembleNodeData(int, DecisionTreeModelReadWrite.NodeData) - Constructor for class org.apache.spark.ml.tree.EnsembleModelReadWrite.EnsembleNodeData
 
EnsembleNodeData$() - Constructor for class org.apache.spark.ml.tree.EnsembleModelReadWrite.EnsembleNodeData$
 
entries() - Method in class org.apache.spark.mllib.linalg.distributed.CoordinateMatrix
 
Entropy - Class in org.apache.spark.mllib.tree.impurity
Class for calculating entropy during multiclass classification.
Entropy() - Constructor for class org.apache.spark.mllib.tree.impurity.Entropy
 
entrySet() - Method in class org.apache.spark.api.java.JavaUtils.SerializableMapWrapper
 
entrySet() - Method in class org.apache.spark.sql.util.CaseInsensitiveStringMap
 
EnumUtil - Class in org.apache.spark.util
 
EnumUtil() - Constructor for class org.apache.spark.util.EnumUtil
 
environmentDetails() - Method in class org.apache.spark.scheduler.SparkListenerEnvironmentUpdate
 
environmentUpdateFromJson(JsonNode) - Static method in class org.apache.spark.util.JsonProtocol
 
environmentUpdateToJson(SparkListenerEnvironmentUpdate, JsonGenerator) - Static method in class org.apache.spark.util.JsonProtocol
 
eofExceptionWhileReadPortNumberError(String, Option<Object>) - Static method in class org.apache.spark.errors.SparkCoreErrors
 
eps() - Method in class org.apache.spark.ml.evaluation.MulticlassClassificationEvaluator
param for eps.
EPSILON() - Static method in class org.apache.spark.ml.impl.Utils
 
epsilon() - Method in class org.apache.spark.ml.regression.LinearRegression
 
epsilon() - Method in class org.apache.spark.ml.regression.LinearRegressionModel
 
epsilon() - Method in interface org.apache.spark.ml.regression.LinearRegressionParams
The shape parameter to control the amount of robustness.
eqNullSafe(Object) - Method in class org.apache.spark.sql.Column
Equality test that is safe for null values.
EqualNullSafe - Class in org.apache.spark.sql.sources
Performs equality comparison, similar to EqualTo.
EqualNullSafe(String, Object) - Constructor for class org.apache.spark.sql.sources.EqualNullSafe
 
equals(Object) - Method in class org.apache.spark.api.java.Optional
 
equals(Object) - Static method in class org.apache.spark.ExpireDeadHosts
 
equals(Object) - Method in class org.apache.spark.graphx.EdgeDirection
 
equals(Object) - Method in class org.apache.spark.HashPartitioner
 
equals(Object) - Static method in class org.apache.spark.metrics.DirectPoolMemory
 
equals(Object) - Static method in class org.apache.spark.metrics.GarbageCollectionMetrics
 
equals(Object) - Static method in class org.apache.spark.metrics.JVMHeapMemory
 
equals(Object) - Static method in class org.apache.spark.metrics.JVMOffHeapMemory
 
equals(Object) - Static method in class org.apache.spark.metrics.MappedPoolMemory
 
equals(Object) - Static method in class org.apache.spark.metrics.OffHeapExecutionMemory
 
equals(Object) - Static method in class org.apache.spark.metrics.OffHeapStorageMemory
 
equals(Object) - Static method in class org.apache.spark.metrics.OffHeapUnifiedMemory
 
equals(Object) - Static method in class org.apache.spark.metrics.OnHeapExecutionMemory
 
equals(Object) - Static method in class org.apache.spark.metrics.OnHeapStorageMemory
 
equals(Object) - Static method in class org.apache.spark.metrics.OnHeapUnifiedMemory
 
equals(Object) - Static method in class org.apache.spark.metrics.ProcessTreeMetrics
 
equals(Object) - Method in class org.apache.spark.ml.attribute.AttributeGroup
 
equals(Object) - Method in class org.apache.spark.ml.attribute.BinaryAttribute
 
equals(Object) - Method in class org.apache.spark.ml.attribute.NominalAttribute
 
equals(Object) - Method in class org.apache.spark.ml.attribute.NumericAttribute
 
equals(Object) - Static method in class org.apache.spark.ml.feature.Dot
 
equals(Object) - Static method in class org.apache.spark.ml.feature.EmptyTerm
 
equals(Object) - Method in class org.apache.spark.ml.linalg.DenseMatrix
 
equals(Object) - Method in class org.apache.spark.ml.linalg.DenseVector
 
equals(Object) - Method in class org.apache.spark.ml.linalg.SparseMatrix
 
equals(Object) - Method in class org.apache.spark.ml.linalg.SparseVector
 
equals(Object) - Method in interface org.apache.spark.ml.linalg.Vector
 
equals(Object) - Method in class org.apache.spark.ml.param.Param
 
equals(Object) - Method in class org.apache.spark.ml.tree.CategoricalSplit
 
equals(Object) - Method in class org.apache.spark.ml.tree.ContinuousSplit
 
equals(Object) - Method in class org.apache.spark.mllib.linalg.DenseMatrix
 
equals(Object) - Method in class org.apache.spark.mllib.linalg.DenseVector
 
equals(Object) - Method in class org.apache.spark.mllib.linalg.SparseMatrix
 
equals(Object) - Method in class org.apache.spark.mllib.linalg.SparseVector
 
equals(Object) - Method in interface org.apache.spark.mllib.linalg.Vector
 
equals(Object) - Method in class org.apache.spark.mllib.linalg.VectorUDT
 
equals(Object) - Method in class org.apache.spark.mllib.tree.model.InformationGainStats
 
equals(Object) - Method in class org.apache.spark.mllib.tree.model.Predict
 
equals(Object) - Method in class org.apache.spark.partial.BoundedDouble
 
equals(Object) - Method in interface org.apache.spark.Partition
 
equals(Object) - Method in class org.apache.spark.RangePartitioner
 
equals(Object) - Method in class org.apache.spark.resource.ExecutorResourceRequest
 
equals(Object) - Method in class org.apache.spark.resource.ResourceID
 
equals(Object) - Method in class org.apache.spark.resource.ResourceInformation
 
equals(Object) - Method in class org.apache.spark.resource.ResourceProfile
 
equals(Object) - Method in class org.apache.spark.resource.ResourceRequest
 
equals(Object) - Method in class org.apache.spark.resource.TaskResourceRequest
 
equals(Object) - Static method in class org.apache.spark.Resubmitted
 
equals(Object) - Static method in class org.apache.spark.scheduler.AllJobsCancelled
 
equals(Object) - Method in class org.apache.spark.scheduler.cluster.ExecutorInfo
 
equals(Object) - Method in class org.apache.spark.scheduler.InputFormatInfo
 
equals(Object) - Static method in class org.apache.spark.scheduler.JobSucceeded
 
equals(Object) - Static method in class org.apache.spark.scheduler.ResubmitFailedStages
 
equals(Object) - Method in class org.apache.spark.scheduler.SplitInfo
 
equals(Object) - Static method in class org.apache.spark.scheduler.StopCoordinator
 
equals(Object) - Method in class org.apache.spark.sql.Column
 
equals(Object) - Method in class org.apache.spark.sql.connector.catalog.ColumnDefaultValue
 
equals(Object) - Method in class org.apache.spark.sql.connector.catalog.TableChange.AddColumn
 
equals(Object) - Method in class org.apache.spark.sql.connector.catalog.TableChange.After
 
equals(Object) - Method in class org.apache.spark.sql.connector.catalog.TableChange.DeleteColumn
 
equals(Object) - Method in class org.apache.spark.sql.connector.catalog.TableChange.RemoveProperty
 
equals(Object) - Method in class org.apache.spark.sql.connector.catalog.TableChange.RenameColumn
 
equals(Object) - Method in class org.apache.spark.sql.connector.catalog.TableChange.SetProperty
 
equals(Object) - Method in class org.apache.spark.sql.connector.catalog.TableChange.UpdateColumnComment
 
equals(Object) - Method in class org.apache.spark.sql.connector.catalog.TableChange.UpdateColumnDefaultValue
 
equals(Object) - Method in class org.apache.spark.sql.connector.catalog.TableChange.UpdateColumnNullability
 
equals(Object) - Method in class org.apache.spark.sql.connector.catalog.TableChange.UpdateColumnPosition
 
equals(Object) - Method in class org.apache.spark.sql.connector.catalog.TableChange.UpdateColumnType
 
equals(Object) - Method in class org.apache.spark.sql.connector.expressions.aggregate.GeneralAggregateFunc
 
equals(Object) - Method in class org.apache.spark.sql.connector.expressions.aggregate.UserDefinedAggregateFunc
 
equals(Object) - Method in class org.apache.spark.sql.connector.expressions.GeneralScalarExpression
 
equals(Object) - Method in class org.apache.spark.sql.connector.expressions.UserDefinedScalarFunc
 
equals(Object) - Method in class org.apache.spark.sql.connector.read.streaming.CompositeReadLimit
 
equals(Object) - Method in class org.apache.spark.sql.connector.read.streaming.Offset
Equality based on JSON string representation.
equals(Object) - Method in class org.apache.spark.sql.connector.read.streaming.ReadMaxFiles
 
equals(Object) - Method in class org.apache.spark.sql.connector.read.streaming.ReadMaxRows
 
equals(Object) - Method in class org.apache.spark.sql.connector.read.streaming.ReadMinRows
 
equals(Object) - Static method in class org.apache.spark.sql.jdbc.MySQLDialect
 
equals(Object) - Static method in class org.apache.spark.sql.jdbc.OracleDialect
 
equals(Object) - Static method in class org.apache.spark.sql.jdbc.TeradataDialect
 
equals(Object) - Method in interface org.apache.spark.sql.Row
 
equals(Object) - Static method in class org.apache.spark.sql.sources.AlwaysFalse
 
equals(Object) - Static method in class org.apache.spark.sql.sources.AlwaysTrue
 
equals(Object) - Method in class org.apache.spark.sql.sources.In
 
equals(Object) - Static method in class org.apache.spark.sql.types.BinaryType
 
equals(Object) - Static method in class org.apache.spark.sql.types.BooleanType
 
equals(Object) - Static method in class org.apache.spark.sql.types.ByteExactNumeric
 
equals(Object) - Static method in class org.apache.spark.sql.types.ByteType
 
equals(Object) - Static method in class org.apache.spark.sql.types.CalendarIntervalType
 
equals(Object) - Static method in class org.apache.spark.sql.types.DateType
 
equals(Object) - Static method in class org.apache.spark.sql.types.DayTimeIntervalType
 
equals(Object) - Method in class org.apache.spark.sql.types.Decimal
 
equals(Object) - Static method in class org.apache.spark.sql.types.DecimalExactNumeric
 
equals(Object) - Static method in class org.apache.spark.sql.types.DoubleExactNumeric
 
equals(Object) - Static method in class org.apache.spark.sql.types.DoubleType
 
equals(Object) - Static method in class org.apache.spark.sql.types.FloatExactNumeric
 
equals(Object) - Static method in class org.apache.spark.sql.types.FloatType
 
equals(Object) - Static method in class org.apache.spark.sql.types.IntegerExactNumeric
 
equals(Object) - Static method in class org.apache.spark.sql.types.IntegerType
 
equals(Object) - Static method in class org.apache.spark.sql.types.LongExactNumeric
 
equals(Object) - Static method in class org.apache.spark.sql.types.LongType
 
equals(Object) - Method in class org.apache.spark.sql.types.Metadata
 
equals(Object) - Static method in class org.apache.spark.sql.types.NullType
 
equals(Object) - Static method in class org.apache.spark.sql.types.ShortExactNumeric
 
equals(Object) - Static method in class org.apache.spark.sql.types.ShortType
 
equals(Object) - Static method in class org.apache.spark.sql.types.StringType
 
equals(Object) - Method in class org.apache.spark.sql.types.StructType
 
equals(Object) - Static method in class org.apache.spark.sql.types.TimestampNTZType
 
equals(Object) - Static method in class org.apache.spark.sql.types.TimestampType
 
equals(Object) - Method in class org.apache.spark.sql.types.UserDefinedType
 
equals(Object) - Static method in class org.apache.spark.sql.types.YearMonthIntervalType
 
equals(Object) - Method in class org.apache.spark.sql.util.CaseInsensitiveStringMap
 
equals(Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo
 
equals(Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo
 
equals(Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
 
equals(Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoWrapper
 
equals(Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo
 
equals(Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoWrapper
 
equals(Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.AppSummary
 
equals(Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
 
equals(Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetrics
 
equals(Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
 
equals(Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions
 
equals(Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest
 
equals(Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary
 
equals(Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper
 
equals(Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
 
equals(Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryWrapper
 
equals(Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributions
 
equals(Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetrics
 
equals(Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
 
equals(Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper
 
equals(Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.MemoryMetrics
 
equals(Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributions
 
equals(Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetrics
 
equals(Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.PairStrings
 
equals(Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.PoolData
 
equals(Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary
 
equals(Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryWrapper
 
equals(Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution
 
equals(Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper
 
equals(Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationEdge
 
equals(Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper
 
equals(Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode
 
equals(Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo
 
equals(Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo
 
equals(Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoWrapper
 
equals(Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceInformation
 
equals(Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo
 
equals(Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileWrapper
 
equals(Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.RuntimeInfo
 
equals(Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions
 
equals(Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics
 
equals(Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions
 
equals(Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics
 
equals(Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions
 
equals(Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetrics
 
equals(Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress
 
equals(Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress
 
equals(Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper
 
equals(Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphEdge
 
equals(Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode
 
equals(Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper
 
equals(Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper
 
equals(Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummary
 
equals(Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryWrapper
 
equals(Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
 
equals(Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetric
 
equals(Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
 
equals(Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper
 
equals(Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress
 
equals(Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData
 
equals(Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData
 
equals(Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
 
equals(Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressWrapper
 
equals(Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
 
equals(Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
 
equals(Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
 
equals(Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics
 
equals(Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskResourceRequest
 
equals(Object) - Static method in class org.apache.spark.StopMapOutputTracker
 
equals(Object) - Method in class org.apache.spark.storage.BlockManagerId
 
equals(Object) - Method in class org.apache.spark.storage.StorageLevel
 
equals(Object) - Static method in class org.apache.spark.streaming.kinesis.DefaultCredentials
 
equals(Object) - Static method in class org.apache.spark.streaming.scheduler.AllReceiverIds
 
equals(Object) - Static method in class org.apache.spark.streaming.scheduler.GetAllReceiverInfo
 
equals(Object) - Static method in class org.apache.spark.streaming.scheduler.StopAllReceivers
 
equals(Object) - Static method in class org.apache.spark.Success
 
equals(Object) - Static method in class org.apache.spark.TaskResultLost
 
equals(Object) - Static method in class org.apache.spark.TaskSchedulerIsSet
 
equals(Object) - Static method in class org.apache.spark.UnknownReason
 
equals(Object) - Method in class org.apache.spark.unsafe.types.CalendarInterval
 
equalsStructurally(DataType, DataType, boolean) - Static method in class org.apache.spark.sql.types.DataType
Returns true if the two data types share the same "shape", i.e.
equalsStructurallyByName(DataType, DataType, Function2<String, String, Object>) - Static method in class org.apache.spark.sql.types.DataType
Returns true if the two data types have the same field names in order recursively.
equalTo(Object) - Method in class org.apache.spark.sql.Column
Equality test.
EqualTo - Class in org.apache.spark.sql.sources
A filter that evaluates to true iff the column evaluates to a value equal to value.
EqualTo(String, Object) - Constructor for class org.apache.spark.sql.sources.EqualTo
 
equiv(T, T) - Static method in class org.apache.spark.sql.types.ByteExactNumeric
 
equiv(T, T) - Static method in class org.apache.spark.sql.types.DecimalExactNumeric
 
equiv(T, T) - Static method in class org.apache.spark.sql.types.DoubleExactNumeric
 
equiv(T, T) - Static method in class org.apache.spark.sql.types.FloatExactNumeric
 
equiv(T, T) - Static method in class org.apache.spark.sql.types.IntegerExactNumeric
 
equiv(T, T) - Static method in class org.apache.spark.sql.types.LongExactNumeric
 
equiv(T, T) - Static method in class org.apache.spark.sql.types.ShortExactNumeric
 
err(String) - Static method in class org.apache.spark.ml.feature.RFormulaParser
 
Error() - Static method in class org.apache.spark.ml.feature.RFormulaParser
 
ERROR() - Static method in class org.apache.spark.status.TaskIndexNames
 
ERROR_COMMAND_NOT_FOUND() - Static method in class org.apache.spark.util.SparkExitCode
Exception indicate command not found.
ERROR_MESSAGE_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
 
ERROR_MESSAGE_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.TaskData
 
ERROR_MESSAGE_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
 
ERROR_MISUSE_SHELL_BUILTIN() - Static method in class org.apache.spark.util.SparkExitCode
Exception indicate invalid usage of some shell built-in command.
ERROR_PATH_NOT_FOUND() - Static method in class org.apache.spark.util.SparkExitCode
Exception appears when the computer cannot find the specified path.
errorClass() - Method in exception org.apache.spark.sql.AnalysisException
 
ErrorClassesJsonReader - Class in org.apache.spark
A reader to load error information from one or more JSON files.
ErrorClassesJsonReader(Seq<URL>) - Constructor for class org.apache.spark.ErrorClassesJsonReader
 
ErrorHandlingReadableChannel(ReadableByteChannel, ReadableByteChannel) - Constructor for class org.apache.spark.security.CryptoStreamUtils.ErrorHandlingReadableChannel
 
ErrorInfo - Class in org.apache.spark
Information associated with an error class.
ErrorInfo(Seq<String>, Option<Map<String, ErrorSubInfo>>, Option<String>) - Constructor for class org.apache.spark.ErrorInfo
 
errorMessage() - Method in class org.apache.spark.status.api.v1.TaskData
 
errorMessage() - Method in class org.apache.spark.status.LiveTask
 
ErrorMessageFormat - Class in org.apache.spark
 
ErrorMessageFormat() - Constructor for class org.apache.spark.ErrorMessageFormat
 
errorReader() - Static method in class org.apache.spark.SparkThrowableHelper
 
ErrorSubInfo - Class in org.apache.spark
Information associated with an error subclass.
ErrorSubInfo(Seq<String>) - Constructor for class org.apache.spark.ErrorSubInfo
 
escapeCharacterAtTheEndError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
escapeCharacterInTheMiddleError(String, String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
escapeMetaCharacters(String) - Static method in class org.apache.spark.sql.util.SchemaUtils
 
estimate(double[]) - Method in class org.apache.spark.mllib.stat.KernelDensity
Estimates probability density function at the given array of points.
estimate(Object) - Static method in class org.apache.spark.util.SizeEstimator
Estimate the number of bytes that the given object takes up on the JVM heap.
estimateCount(Object) - Method in class org.apache.spark.util.sketch.CountMinSketch
Returns the estimated frequency of item.
estimatedDocConcentration() - Method in class org.apache.spark.ml.clustering.LDAModel
Value for docConcentration estimated from data.
estimatedSize() - Method in class org.apache.spark.storage.memory.DeserializedValuesHolder
 
estimatedSize() - Method in class org.apache.spark.storage.memory.SerializedValuesHolder
 
estimatedSize() - Method in interface org.apache.spark.storage.memory.ValuesHolder
 
estimatedSize() - Method in interface org.apache.spark.util.KnownSizeEstimation
 
estimateStatistics() - Method in interface org.apache.spark.sql.connector.read.SupportsReportStatistics
Returns the estimated statistics of this data source scan.
Estimator<M extends Model<M>> - Class in org.apache.spark.ml
Abstract class for estimators that fit models to data.
Estimator() - Constructor for class org.apache.spark.ml.Estimator
 
estimator() - Method in class org.apache.spark.ml.FitEnd
 
estimator() - Method in class org.apache.spark.ml.FitStart
 
estimator() - Method in class org.apache.spark.ml.tuning.CrossValidator
 
estimator() - Method in class org.apache.spark.ml.tuning.CrossValidatorModel
 
estimator() - Method in class org.apache.spark.ml.tuning.TrainValidationSplit
 
estimator() - Method in class org.apache.spark.ml.tuning.TrainValidationSplitModel
 
estimator() - Method in interface org.apache.spark.ml.tuning.ValidatorParams
param for the estimator to be validated
estimatorParamMaps() - Method in class org.apache.spark.ml.tuning.CrossValidator
 
estimatorParamMaps() - Method in class org.apache.spark.ml.tuning.CrossValidatorModel
 
estimatorParamMaps() - Method in class org.apache.spark.ml.tuning.TrainValidationSplit
 
estimatorParamMaps() - Method in class org.apache.spark.ml.tuning.TrainValidationSplitModel
 
estimatorParamMaps() - Method in interface org.apache.spark.ml.tuning.ValidatorParams
param for estimator param maps
eval() - Method in interface org.apache.spark.ml.ann.ActivationFunction
Implements a function
eval(DenseMatrix<Object>, DenseMatrix<Object>) - Method in interface org.apache.spark.ml.ann.LayerModel
Evaluates the data (process the data through the layer).
evaluate(Dataset<?>) - Method in class org.apache.spark.ml.classification.FMClassificationModel
Evaluates the model on a test dataset.
evaluate(Dataset<?>) - Method in class org.apache.spark.ml.classification.LinearSVCModel
Evaluates the model on a test dataset.
evaluate(Dataset<?>) - Method in class org.apache.spark.ml.classification.LogisticRegressionModel
Evaluates the model on a test dataset.
evaluate(Dataset<?>) - Method in class org.apache.spark.ml.classification.MultilayerPerceptronClassificationModel
Evaluates the model on a test dataset.
evaluate(Dataset<?>) - Method in class org.apache.spark.ml.classification.RandomForestClassificationModel
Evaluates the model on a test dataset.
evaluate(Dataset<?>) - Method in class org.apache.spark.ml.evaluation.BinaryClassificationEvaluator
 
evaluate(Dataset<?>) - Method in class org.apache.spark.ml.evaluation.ClusteringEvaluator
 
evaluate(Dataset<?>, ParamMap) - Method in class org.apache.spark.ml.evaluation.Evaluator
Evaluates model output and returns a scalar metric.
evaluate(Dataset<?>) - Method in class org.apache.spark.ml.evaluation.Evaluator
Evaluates model output and returns a scalar metric.
evaluate(Dataset<?>) - Method in class org.apache.spark.ml.evaluation.MulticlassClassificationEvaluator
 
evaluate(Dataset<?>) - Method in class org.apache.spark.ml.evaluation.MultilabelClassificationEvaluator
 
evaluate(Dataset<?>) - Method in class org.apache.spark.ml.evaluation.RankingEvaluator
 
evaluate(Dataset<?>) - Method in class org.apache.spark.ml.evaluation.RegressionEvaluator
 
evaluate(Dataset<?>) - Method in class org.apache.spark.ml.regression.GeneralizedLinearRegressionModel
Evaluate the model on the given dataset, returning a summary of the results.
evaluate(Dataset<?>) - Method in class org.apache.spark.ml.regression.LinearRegressionModel
Evaluates the model on a test dataset.
evaluate(Row) - Method in class org.apache.spark.sql.expressions.UserDefinedAggregateFunction
Deprecated.
Calculates the final result of this UserDefinedAggregateFunction based on the given aggregation buffer.
evaluateEachIteration(Dataset<?>) - Method in class org.apache.spark.ml.classification.GBTClassificationModel
Method to compute error or loss for every iteration of gradient boosting.
evaluateEachIteration(Dataset<?>, String) - Method in class org.apache.spark.ml.regression.GBTRegressionModel
Method to compute error or loss for every iteration of gradient boosting.
evaluateEachIteration(RDD<org.apache.spark.ml.feature.Instance>, DecisionTreeRegressionModel[], double[], Loss, Enumeration.Value) - Static method in class org.apache.spark.ml.tree.impl.GradientBoostedTrees
Method to compute error or loss for every iteration of gradient boosting.
evaluateEachIteration(RDD<LabeledPoint>, Loss) - Method in class org.apache.spark.mllib.tree.model.GradientBoostedTreesModel
Method to compute error or loss for every iteration of gradient boosting.
Evaluator - Class in org.apache.spark.ml.evaluation
Abstract class for evaluators that compute metrics from predictions.
Evaluator() - Constructor for class org.apache.spark.ml.evaluation.Evaluator
 
evaluator() - Method in class org.apache.spark.ml.tuning.CrossValidator
 
evaluator() - Method in class org.apache.spark.ml.tuning.CrossValidatorModel
 
evaluator() - Method in class org.apache.spark.ml.tuning.TrainValidationSplit
 
evaluator() - Method in class org.apache.spark.ml.tuning.TrainValidationSplitModel
 
evaluator() - Method in interface org.apache.spark.ml.tuning.ValidatorParams
param for the evaluator used to select hyper-parameters that maximize the validated metric
EVENT_TIME_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
 
eventRates() - Method in class org.apache.spark.status.api.v1.streaming.ReceiverInfo
 
eventTime() - Method in class org.apache.spark.sql.streaming.StreamingQueryProgress
 
EventTimeTimeout() - Static method in class org.apache.spark.sql.streaming.GroupStateTimeout
Timeout based on event-time.
exceedMapSizeLimitError(int) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
except(Dataset<T>) - Method in class org.apache.spark.sql.Dataset
Returns a new Dataset containing rows in this Dataset but not in another Dataset.
exceptAll(Dataset<T>) - Method in class org.apache.spark.sql.Dataset
Returns a new Dataset containing rows in this Dataset but not in another Dataset while preserving the duplicates.
exception() - Method in class org.apache.spark.ExceptionFailure
 
exception() - Method in interface org.apache.spark.sql.streaming.StreamingQuery
Returns the StreamingQueryException if the query was terminated by an exception.
exception() - Method in class org.apache.spark.sql.streaming.StreamingQueryListener.QueryTerminatedEvent
 
EXCEPTION_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData
 
ExceptionFailure - Class in org.apache.spark
:: DeveloperApi :: Task failed due to a runtime exception.
ExceptionFailure(String, String, StackTraceElement[], String, Option<ThrowableSerializationWrapper>, Seq<AccumulableInfo>, Seq<AccumulatorV2<?, ?>>, Seq<Object>) - Constructor for class org.apache.spark.ExceptionFailure
 
exceptionFromJson(JsonNode) - Static method in class org.apache.spark.util.JsonProtocol
 
exceptionString(Throwable) - Static method in class org.apache.spark.util.Utils
Return a nice string representation of the exception.
exceptionToJson(Exception, JsonGenerator) - Static method in class org.apache.spark.util.JsonProtocol
 
EXCLUDED_IN_STAGES_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
 
ExcludedExecutor - Class in org.apache.spark.scheduler
 
ExcludedExecutor(String, long) - Constructor for class org.apache.spark.scheduler.ExcludedExecutor
 
excludedExecutors() - Method in class org.apache.spark.status.LiveStage
 
excludedInStages() - Method in class org.apache.spark.status.api.v1.ExecutorSummary
 
excludedNodes() - Method in class org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.RequestExecutors
 
EXEC_CPU_TIME() - Static method in class org.apache.spark.status.TaskIndexNames
 
EXEC_RUN_TIME() - Static method in class org.apache.spark.status.TaskIndexNames
 
execId() - Method in class org.apache.spark.ExecutorLostFailure
 
execId() - Method in class org.apache.spark.scheduler.SparkListenerExecutorMetricsUpdate
 
execId() - Method in class org.apache.spark.scheduler.SparkListenerStageExecutorMetrics
 
execId() - Method in class org.apache.spark.storage.BlockManagerMessages.RemoveExecutor
 
executeAndGetOutput(Seq<String>, File, Map<String, String>, boolean) - Static method in class org.apache.spark.util.Utils
Execute a command and get its output, throwing an exception if it yields a code other than 0.
executeBroadcastTimeoutError(long, Option<TimeoutException>) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
executeCodePathUnsupportedError(String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
executeCommand(String, CaseInsensitiveStringMap) - Method in interface org.apache.spark.sql.connector.ExternalCommandRunner
Execute the given command.
executeCommand(String, String, Map<String, String>) - Method in class org.apache.spark.sql.SparkSession
Execute an arbitrary string command inside an external execution engine rather than Spark.
executeCommand(Seq<String>, File, Map<String, String>, boolean) - Static method in class org.apache.spark.util.Utils
Execute a command and return the process running the command.
EXECUTION_ID_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper
 
EXECUTION_ID_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
 
ExecutionData - Class in org.apache.spark.status.api.v1.sql
 
ExecutionListenerManager - Class in org.apache.spark.sql.util
EXECUTOR() - Static method in class org.apache.spark.metrics.MetricsSystemInstances
 
executor() - Method in class org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.KillTask
 
executor() - Method in interface org.apache.spark.shuffle.api.ShuffleDataIO
Called once on executor processes to bootstrap the shuffle data storage modules that are only invoked on the executors.
EXECUTOR() - Static method in class org.apache.spark.status.TaskIndexNames
 
EXECUTOR_CORES - Static variable in class org.apache.spark.launcher.SparkLauncher
Configuration key for the number of executor CPU cores.
EXECUTOR_CPU_TIME() - Static method in class org.apache.spark.InternalAccumulator
 
EXECUTOR_CPU_TIME_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
 
EXECUTOR_CPU_TIME_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.StageData
 
EXECUTOR_CPU_TIME_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
 
EXECUTOR_CPU_TIME_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
 
EXECUTOR_CPU_TIME_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics
 
EXECUTOR_DEFAULT_JAVA_OPTIONS - Static variable in class org.apache.spark.launcher.SparkLauncher
Configuration key for the default executor VM options.
EXECUTOR_DESERIALIZE_CPU_TIME() - Static method in class org.apache.spark.InternalAccumulator
 
EXECUTOR_DESERIALIZE_CPU_TIME_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
 
EXECUTOR_DESERIALIZE_CPU_TIME_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.StageData
 
EXECUTOR_DESERIALIZE_CPU_TIME_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
 
EXECUTOR_DESERIALIZE_CPU_TIME_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
 
EXECUTOR_DESERIALIZE_CPU_TIME_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics
 
EXECUTOR_DESERIALIZE_TIME() - Static method in class org.apache.spark.InternalAccumulator
 
EXECUTOR_DESERIALIZE_TIME_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
 
EXECUTOR_DESERIALIZE_TIME_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.StageData
 
EXECUTOR_DESERIALIZE_TIME_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
 
EXECUTOR_DESERIALIZE_TIME_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
 
EXECUTOR_DESERIALIZE_TIME_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics
 
EXECUTOR_EXTRA_CLASSPATH - Static variable in class org.apache.spark.launcher.SparkLauncher
Configuration key for the executor class path.
EXECUTOR_EXTRA_JAVA_OPTIONS - Static variable in class org.apache.spark.launcher.SparkLauncher
Configuration key for the executor VM options.
EXECUTOR_EXTRA_LIBRARY_PATH - Static variable in class org.apache.spark.launcher.SparkLauncher
Configuration key for the executor native library path.
EXECUTOR_ID_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper
 
EXECUTOR_ID_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData
 
EXECUTOR_ID_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.TaskData
 
EXECUTOR_ID_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
 
EXECUTOR_LOGS_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
 
EXECUTOR_LOGS_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.TaskData
 
EXECUTOR_MEMORY - Static variable in class org.apache.spark.launcher.SparkLauncher
Configuration key for the executor memory.
EXECUTOR_METRICS_DISTRIBUTIONS_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.StageData
 
EXECUTOR_METRICS_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions
 
EXECUTOR_RESOURCES_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo
 
EXECUTOR_RUN_TIME() - Static method in class org.apache.spark.InternalAccumulator
 
EXECUTOR_RUN_TIME_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
 
EXECUTOR_RUN_TIME_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.StageData
 
EXECUTOR_RUN_TIME_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
 
EXECUTOR_RUN_TIME_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
 
EXECUTOR_RUN_TIME_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics
 
EXECUTOR_SUMMARY_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.StageData
 
executorAddedFromJson(JsonNode) - Static method in class org.apache.spark.util.JsonProtocol
 
executorAddedToJson(SparkListenerExecutorAdded, JsonGenerator) - Static method in class org.apache.spark.util.JsonProtocol
 
executorCpuTime() - Method in class org.apache.spark.status.api.v1.StageData
 
executorCpuTime() - Method in class org.apache.spark.status.api.v1.TaskMetricDistributions
 
executorCpuTime() - Method in class org.apache.spark.status.api.v1.TaskMetrics
 
executorDecommission(String) - Method in interface org.apache.spark.scheduler.Schedulable
 
executorDecommission(String, ExecutorDecommissionInfo) - Method in interface org.apache.spark.scheduler.TaskScheduler
Process a decommissioning executor.
ExecutorDecommissioning(String) - Constructor for class org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.ExecutorDecommissioning
 
ExecutorDecommissioning$() - Constructor for class org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.ExecutorDecommissioning$
 
ExecutorDecommissionSigReceived$() - Constructor for class org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.ExecutorDecommissionSigReceived$
 
executorDeserializeCpuTime() - Method in class org.apache.spark.status.api.v1.StageData
 
executorDeserializeCpuTime() - Method in class org.apache.spark.status.api.v1.TaskMetricDistributions
 
executorDeserializeCpuTime() - Method in class org.apache.spark.status.api.v1.TaskMetrics
 
executorDeserializeTime() - Method in class org.apache.spark.status.api.v1.StageData
 
executorDeserializeTime() - Method in class org.apache.spark.status.api.v1.TaskMetricDistributions
 
executorDeserializeTime() - Method in class org.apache.spark.status.api.v1.TaskMetrics
 
executorFailures() - Method in class org.apache.spark.scheduler.SparkListenerNodeBlacklisted
Deprecated.
 
executorFailures() - Method in class org.apache.spark.scheduler.SparkListenerNodeBlacklistedForStage
Deprecated.
 
executorFailures() - Method in class org.apache.spark.scheduler.SparkListenerNodeExcluded
 
executorFailures() - Method in class org.apache.spark.scheduler.SparkListenerNodeExcludedForStage
 
executorHeartbeatReceived(String, Tuple2<Object, Seq<AccumulatorV2<?, ?>>>[], BlockManagerId, Map<Tuple2<Object, Object>, ExecutorMetrics>) - Method in interface org.apache.spark.scheduler.TaskScheduler
Update metrics for in-progress tasks and executor metrics, and let the master know that the BlockManager is still alive.
executorHost() - Method in class org.apache.spark.scheduler.cluster.ExecutorInfo
 
executorHost() - Method in class org.apache.spark.status.api.v1.streaming.ReceiverInfo
 
executorID() - Method in interface org.apache.spark.api.plugin.PluginContext
Executor ID of the process.
executorId() - Method in class org.apache.spark.ExecutorRegistered
 
executorId() - Method in class org.apache.spark.ExecutorRemoved
 
executorId() - Method in class org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.ExecutorDecommissioning
 
executorId() - Method in class org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.GetExecutorLossReason
 
executorId() - Method in class org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.IsExecutorAlive
 
executorId() - Method in class org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.LaunchedExecutor
 
executorId() - Method in class org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.RegisterExecutor
 
executorId() - Method in class org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.RemoveExecutor
 
executorId() - Method in class org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.StatusUpdate
 
executorId() - Method in class org.apache.spark.scheduler.SparkListenerExecutorAdded
 
executorId() - Method in class org.apache.spark.scheduler.SparkListenerExecutorBlacklisted
Deprecated.
 
executorId() - Method in class org.apache.spark.scheduler.SparkListenerExecutorBlacklistedForStage
Deprecated.
 
executorId() - Method in class org.apache.spark.scheduler.SparkListenerExecutorExcluded
 
executorId() - Method in class org.apache.spark.scheduler.SparkListenerExecutorExcludedForStage
 
executorId() - Method in class org.apache.spark.scheduler.SparkListenerExecutorRemoved
 
executorId() - Method in class org.apache.spark.scheduler.SparkListenerExecutorUnblacklisted
Deprecated.
 
executorId() - Method in class org.apache.spark.scheduler.SparkListenerExecutorUnexcluded
 
executorId() - Method in class org.apache.spark.scheduler.TaskInfo
 
executorId() - Method in class org.apache.spark.SparkEnv
 
executorId() - Method in class org.apache.spark.status.api.v1.streaming.ReceiverInfo
 
executorId() - Method in class org.apache.spark.status.api.v1.TaskData
 
executorId() - Method in class org.apache.spark.status.LiveRDDDistribution
 
executorId() - Method in class org.apache.spark.storage.BlockManagerId
 
executorId() - Method in class org.apache.spark.storage.BlockManagerMessages.GetExecutorEndpointRef
 
executorId() - Method in class org.apache.spark.storage.BlockManagerMessages.IsExecutorAlive
 
executorId() - Method in class org.apache.spark.streaming.scheduler.ReceiverInfo
 
executorId() - Method in class org.apache.spark.ui.storage.ExecutorStreamSummary
 
executorIds() - Method in class org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.KillExecutors
 
executorIds() - Method in class org.apache.spark.storage.BlockManagerMessages.DecommissionBlockManagers
 
ExecutorInfo - Class in org.apache.spark.scheduler.cluster
:: DeveloperApi :: Stores information about an executor to pass from the scheduler to SparkListeners.
ExecutorInfo(String, int, Map<String, String>, Map<String, String>, Map<String, ResourceInformation>, int, Option<Object>, Option<Object>) - Constructor for class org.apache.spark.scheduler.cluster.ExecutorInfo
 
ExecutorInfo(String, int, Map<String, String>, Map<String, String>, Map<String, ResourceInformation>, int) - Constructor for class org.apache.spark.scheduler.cluster.ExecutorInfo
 
ExecutorInfo(String, int, Map<String, String>) - Constructor for class org.apache.spark.scheduler.cluster.ExecutorInfo
 
ExecutorInfo(String, int, Map<String, String>, Map<String, String>) - Constructor for class org.apache.spark.scheduler.cluster.ExecutorInfo
 
ExecutorInfo(String, int, Map<String, String>, Map<String, String>, Map<String, ResourceInformation>) - Constructor for class org.apache.spark.scheduler.cluster.ExecutorInfo
 
executorInfo() - Method in class org.apache.spark.scheduler.SparkListenerExecutorAdded
 
executorInfoFromJson(JsonNode) - Static method in class org.apache.spark.util.JsonProtocol
 
executorInfoToJson(ExecutorInfo, JsonGenerator) - Static method in class org.apache.spark.util.JsonProtocol
 
ExecutorKilled - Class in org.apache.spark.scheduler
 
ExecutorKilled() - Constructor for class org.apache.spark.scheduler.ExecutorKilled
 
executorLogs() - Method in class org.apache.spark.status.api.v1.ExecutorSummary
 
executorLogs() - Method in class org.apache.spark.status.api.v1.TaskData
 
ExecutorLossMessage - Class in org.apache.spark.scheduler
 
ExecutorLossMessage() - Constructor for class org.apache.spark.scheduler.ExecutorLossMessage
 
executorLost(String, String, ExecutorLossReason) - Method in interface org.apache.spark.scheduler.Schedulable
 
executorLost(String, ExecutorLossReason) - Method in interface org.apache.spark.scheduler.TaskScheduler
Process a lost executor
ExecutorLostFailure - Class in org.apache.spark
:: DeveloperApi :: The task failed because the executor that it was running on was lost.
ExecutorLostFailure(String, boolean, Option<String>) - Constructor for class org.apache.spark.ExecutorLostFailure
 
executorMetrics() - Method in class org.apache.spark.scheduler.SparkListenerStageExecutorMetrics
 
executorMetrics() - Method in class org.apache.spark.status.api.v1.ExecutorPeakMetricsDistributions
 
ExecutorMetricsDistributions - Class in org.apache.spark.status.api.v1
 
executorMetricsDistributions() - Method in class org.apache.spark.status.api.v1.StageData
 
executorMetricsFromJson(JsonNode) - Static method in class org.apache.spark.util.JsonProtocol
Extract the executor metrics from JSON.
ExecutorMetricsSerializer - Class in org.apache.spark.status.protobuf
 
ExecutorMetricsSerializer() - Constructor for class org.apache.spark.status.protobuf.ExecutorMetricsSerializer
 
executorMetricsToJson(ExecutorMetrics, JsonGenerator) - Static method in class org.apache.spark.util.JsonProtocol
Convert executor metrics to JSON.
executorMetricsUpdateFromJson(JsonNode) - Static method in class org.apache.spark.util.JsonProtocol
 
executorMetricsUpdateToJson(SparkListenerExecutorMetricsUpdate, JsonGenerator) - Static method in class org.apache.spark.util.JsonProtocol
 
ExecutorMetricType - Interface in org.apache.spark.metrics
Executor metric types for executor-level metrics stored in ExecutorMetrics.
executorOffHeapMemorySizeAsMb(SparkConf) - Static method in class org.apache.spark.util.Utils
Convert MEMORY_OFFHEAP_SIZE to MB Unit, return 0 if MEMORY_OFFHEAP_ENABLED is false.
executorPct() - Method in class org.apache.spark.scheduler.RuntimePercentage
 
ExecutorPeakMetricsDistributions - Class in org.apache.spark.status.api.v1
 
ExecutorPlugin - Interface in org.apache.spark.api.plugin
:: DeveloperApi :: Executor component of a SparkPlugin.
executorPlugin() - Method in interface org.apache.spark.api.plugin.SparkPlugin
Return the plugin's executor-side component.
executorRef() - Method in class org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.RegisterExecutor
 
ExecutorRegistered - Class in org.apache.spark
 
ExecutorRegistered(String) - Constructor for class org.apache.spark.ExecutorRegistered
 
ExecutorRemoved - Class in org.apache.spark
 
ExecutorRemoved(String) - Constructor for class org.apache.spark.ExecutorRemoved
 
executorRemovedFromJson(JsonNode) - Static method in class org.apache.spark.util.JsonProtocol
 
executorRemovedToJson(SparkListenerExecutorRemoved, JsonGenerator) - Static method in class org.apache.spark.util.JsonProtocol
 
ExecutorResourceRequest - Class in org.apache.spark.resource
An Executor resource request.
ExecutorResourceRequest(String, long, String, String) - Constructor for class org.apache.spark.resource.ExecutorResourceRequest
 
executorResourceRequestFromJson(JsonNode) - Static method in class org.apache.spark.util.JsonProtocol
 
executorResourceRequestMapFromJson(JsonNode) - Static method in class org.apache.spark.util.JsonProtocol
 
executorResourceRequestMapToJson(Map<String, ExecutorResourceRequest>, JsonGenerator) - Static method in class org.apache.spark.util.JsonProtocol
 
ExecutorResourceRequests - Class in org.apache.spark.resource
A set of Executor resource requests.
ExecutorResourceRequests() - Constructor for class org.apache.spark.resource.ExecutorResourceRequests
 
executorResourceRequestToJson(ExecutorResourceRequest, JsonGenerator) - Static method in class org.apache.spark.util.JsonProtocol
 
executorResourceRequestToRequirement(Seq<ExecutorResourceRequest>) - Static method in class org.apache.spark.resource.ResourceUtils
 
executorResources() - Method in class org.apache.spark.resource.ResourceProfile
 
executorResources() - Method in class org.apache.spark.resource.ResourceProfileBuilder
 
executorResources() - Method in class org.apache.spark.status.api.v1.ResourceProfileInfo
 
executorResources() - Method in class org.apache.spark.status.LiveResourceProfile
 
executorResourcesJMap() - Method in class org.apache.spark.resource.ResourceProfile
(Java-specific) gets a Java Map of resources to ExecutorResourceRequest
executorResourcesJMap() - Method in class org.apache.spark.resource.ResourceProfileBuilder
(Java-specific) gets a Java Map of resources to ExecutorResourceRequest
ExecutorResourcesOrDefaults$() - Constructor for class org.apache.spark.resource.ResourceProfile.ExecutorResourcesOrDefaults$
 
executorRunTime() - Method in class org.apache.spark.status.api.v1.StageData
 
executorRunTime() - Method in class org.apache.spark.status.api.v1.TaskMetricDistributions
 
executorRunTime() - Method in class org.apache.spark.status.api.v1.TaskMetrics
 
executors() - Method in class org.apache.spark.status.api.v1.RDDPartitionInfo
 
executors() - Method in class org.apache.spark.status.LiveRDDPartition
 
EXECUTORS_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo
 
ExecutorStageSummary - Class in org.apache.spark.status.api.v1
 
ExecutorStageSummarySerializer - Class in org.apache.spark.status.protobuf
 
ExecutorStageSummarySerializer() - Constructor for class org.apache.spark.status.protobuf.ExecutorStageSummarySerializer
 
ExecutorStreamSummary - Class in org.apache.spark.ui.storage
 
ExecutorStreamSummary(Seq<org.apache.spark.status.StreamBlockData>) - Constructor for class org.apache.spark.ui.storage.ExecutorStreamSummary
 
executorSummaries() - Method in class org.apache.spark.status.LiveStage
 
ExecutorSummary - Class in org.apache.spark.status.api.v1
 
executorSummary() - Method in class org.apache.spark.status.api.v1.StageData
 
executorSummary(String) - Method in class org.apache.spark.status.LiveStage
 
executorUpdates() - Method in class org.apache.spark.scheduler.SparkListenerExecutorMetricsUpdate
 
exists(Column, Function1<Column, Column>) - Static method in class org.apache.spark.sql.functions
Returns whether a predicate holds for one or more elements in the array.
exists() - Method in interface org.apache.spark.sql.streaming.GroupState
Whether state exists or not.
exists(String) - Static method in class org.apache.spark.sql.types.UDTRegistration
Queries if a given user class is already registered or not.
exists() - Method in class org.apache.spark.streaming.State
Whether the state already exists
EXIT_FAILURE() - Static method in class org.apache.spark.util.SparkExitCode
Failed termination.
EXIT_SUCCESS() - Static method in class org.apache.spark.util.SparkExitCode
Successful termination.
exitCausedByApp() - Method in class org.apache.spark.ExecutorLostFailure
 
exitCode() - Method in class org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.Shutdown
 
exitFn() - Method in interface org.apache.spark.util.CommandLineLoggingUtils
 
exp(Column) - Static method in class org.apache.spark.sql.functions
Computes the exponential of the given value.
exp(String) - Static method in class org.apache.spark.sql.functions
Computes the exponential of the given column.
ExpectationAggregator - Class in org.apache.spark.ml.clustering
ExpectationAggregator computes the partial expectation results.
ExpectationAggregator(int, Broadcast<double[]>, Broadcast<Tuple2<DenseVector, DenseVector>[]>) - Constructor for class org.apache.spark.ml.clustering.ExpectationAggregator
 
ExpectationSum - Class in org.apache.spark.mllib.clustering
 
ExpectationSum(double, double[], DenseVector<Object>[], DenseMatrix<Object>[]) - Constructor for class org.apache.spark.mllib.clustering.ExpectationSum
 
expectedFpp() - Method in class org.apache.spark.util.sketch.BloomFilter
Returns the probability that BloomFilter.mightContain(Object) erroneously return true for an object that has not actually been put in the BloomFilter.
expectPersistentFuncError(String, String, Option<String>, TreeNode<?>) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
expectTableNotViewError(Seq<String>, boolean, String, Option<String>, TreeNode<?>) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
expectTableOrPermanentViewNotTempViewError(Seq<String>, String, TreeNode<?>) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
expectViewNotTableError(ResolvedTable, String, Option<String>, TreeNode<?>) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
expectViewNotTempViewError(Seq<String>, String, TreeNode<?>) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
experimental() - Method in class org.apache.spark.sql.SparkSession
:: Experimental :: A collection of methods that are considered experimental, but can be used to hook into the query planner for advanced functionality.
experimental() - Method in class org.apache.spark.sql.SQLContext
:: Experimental :: A collection of methods that are considered experimental, but can be used to hook into the query planner for advanced functionality.
ExperimentalMethods - Class in org.apache.spark.sql
:: Experimental :: Holder for experimental methods for the bravest.
ExpireDeadHosts - Class in org.apache.spark
 
ExpireDeadHosts() - Constructor for class org.apache.spark.ExpireDeadHosts
 
expiryTime() - Method in class org.apache.spark.scheduler.ExcludedExecutor
 
explain(boolean) - Method in class org.apache.spark.sql.Column
Prints the expression to the console for debugging purposes.
explain(String) - Method in class org.apache.spark.sql.Dataset
Prints the plans (logical and physical) with a format specified by a given explain mode.
explain(boolean) - Method in class org.apache.spark.sql.Dataset
Prints the plans (logical and physical) to the console for debugging purposes.
explain() - Method in class org.apache.spark.sql.Dataset
Prints the physical plan to the console for debugging purposes.
explain() - Method in interface org.apache.spark.sql.streaming.StreamingQuery
Prints the physical plan to the console for debugging purposes.
explain(boolean) - Method in interface org.apache.spark.sql.streaming.StreamingQuery
Prints the physical plan to the console for debugging purposes.
explainedVariance() - Method in class org.apache.spark.ml.feature.PCAModel
 
explainedVariance() - Method in class org.apache.spark.ml.regression.LinearRegressionSummary
Returns the explained variance regression score.
explainedVariance() - Method in class org.apache.spark.mllib.evaluation.RegressionMetrics
Returns the variance explained by regression.
explainedVariance() - Method in class org.apache.spark.mllib.feature.PCAModel
 
explainParam(Param<?>) - Method in interface org.apache.spark.ml.param.Params
Explains a param.
explainParams() - Method in interface org.apache.spark.ml.param.Params
Explains all params of this instance.
explode(Seq<Column>, Function1<Row, TraversableOnce<A>>, TypeTags.TypeTag<A>) - Method in class org.apache.spark.sql.Dataset
Deprecated.
use flatMap() or select() with functions.explode() instead. Since 2.0.0.
explode(String, String, Function1<A, TraversableOnce<B>>, TypeTags.TypeTag<B>) - Method in class org.apache.spark.sql.Dataset
Deprecated.
use flatMap() or select() with functions.explode() instead. Since 2.0.0.
explode(Column) - Static method in class org.apache.spark.sql.functions
Creates a new row for each element in the given array or map column.
explode_outer(Column) - Static method in class org.apache.spark.sql.functions
Creates a new row for each element in the given array or map column.
explodeNestedFieldNames(StructType) - Static method in class org.apache.spark.sql.util.SchemaUtils
Returns all column names in this schema as a flat list.
expm1(Column) - Static method in class org.apache.spark.sql.functions
Computes the exponential of the given value minus one.
expm1(String) - Static method in class org.apache.spark.sql.functions
Computes the exponential of the given column minus one.
ExponentialGenerator - Class in org.apache.spark.mllib.random
Generates i.i.d.
ExponentialGenerator(double) - Constructor for class org.apache.spark.mllib.random.ExponentialGenerator
 
exponentialJavaRDD(JavaSparkContext, double, long, int, long) - Static method in class org.apache.spark.mllib.random.RandomRDDs
Java-friendly version of RandomRDDs.exponentialRDD.
exponentialJavaRDD(JavaSparkContext, double, long, int) - Static method in class org.apache.spark.mllib.random.RandomRDDs
RandomRDDs.exponentialJavaRDD with the default seed.
exponentialJavaRDD(JavaSparkContext, double, long) - Static method in class org.apache.spark.mllib.random.RandomRDDs
RandomRDDs.exponentialJavaRDD with the default number of partitions and the default seed.
exponentialJavaVectorRDD(JavaSparkContext, double, long, int, int, long) - Static method in class org.apache.spark.mllib.random.RandomRDDs
Java-friendly version of RandomRDDs.exponentialVectorRDD.
exponentialJavaVectorRDD(JavaSparkContext, double, long, int, int) - Static method in class org.apache.spark.mllib.random.RandomRDDs
RandomRDDs.exponentialJavaVectorRDD with the default seed.
exponentialJavaVectorRDD(JavaSparkContext, double, long, int) - Static method in class org.apache.spark.mllib.random.RandomRDDs
RandomRDDs.exponentialJavaVectorRDD with the default number of partitions and the default seed.
exponentialRDD(SparkContext, double, long, int, long) - Static method in class org.apache.spark.mllib.random.RandomRDDs
Generates an RDD comprised of i.i.d. samples from the exponential distribution with the input mean.
exponentialVectorRDD(SparkContext, double, long, int, int, long) - Static method in class org.apache.spark.mllib.random.RandomRDDs
Generates an RDD[Vector] with vectors containing i.i.d. samples drawn from the exponential distribution with the input mean.
expr() - Method in class org.apache.spark.sql.Column
 
expr(String) - Static method in class org.apache.spark.sql.functions
Parses the expression string into the column that it represents, similar to Dataset.selectExpr(java.lang.String...).
expression() - Method in class org.apache.spark.sql.connector.expressions.Cast
 
Expression - Interface in org.apache.spark.sql.connector.expressions
Base class of the public logical expression API.
expression() - Method in interface org.apache.spark.sql.connector.expressions.SortOrder
Returns the sort expression.
Expression$() - Constructor for class org.apache.spark.sql.types.DecimalType.Expression$
 
expressionDecodingError(Exception, Seq<Expression>) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
expressionEncodingError(Exception, Seq<Expression>) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
Expressions - Class in org.apache.spark.sql.connector.expressions
Helper methods to create logical transforms to pass into Spark.
expressionWithMultiWindowExpressionsError(NamedExpression, Seq<WindowSpecDefinition>) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
expressionWithoutWindowExpressionError(NamedExpression) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
ExternalClusterManager - Interface in org.apache.spark.scheduler
A cluster manager interface to plugin external scheduler.
ExternalCommandRunner - Interface in org.apache.spark.sql.connector
An interface to execute an arbitrary string command inside an external execution engine rather than Spark.
externalShuffleServicePort(SparkConf) - Static method in class org.apache.spark.storage.StorageUtils
Get the port used by the external shuffle service.
Extract - Class in org.apache.spark.sql.connector.expressions
Represent an extract function, which extracts and returns the value of a specified datetime field from a datetime or interval value expression.
Extract(String, Expression) - Constructor for class org.apache.spark.sql.connector.expressions.Extract
 
ExtractableLiteral - Class in org.apache.spark.sql.columnar
 
ExtractableLiteral() - Constructor for class org.apache.spark.sql.columnar.ExtractableLiteral
 
extractAsDuration() - Method in class org.apache.spark.unsafe.types.CalendarInterval
Extracts the time part of the interval.
extractAsPeriod() - Method in class org.apache.spark.unsafe.types.CalendarInterval
Extracts the date part of the interval.
extractCatalog(CaseInsensitiveStringMap) - Method in interface org.apache.spark.sql.connector.catalog.SupportsCatalogOptions
Return the name of a catalog that can be used to check the existence of, load, and create a table for this DataSource given the identifier that will be extracted by extractIdentifier.
extractDistribution(Function1<BatchInfo, Option<Object>>) - Method in class org.apache.spark.streaming.scheduler.StatsReportListener
 
extractDoubleDistribution(Seq<Tuple2<TaskInfo, TaskMetrics>>, Function2<TaskInfo, TaskMetrics, Object>) - Static method in class org.apache.spark.scheduler.StatsReportListener
 
extractFn() - Method in class org.apache.spark.ui.JettyUtils.ServletParams
 
extractHostPortFromSparkUrl(String) - Static method in class org.apache.spark.util.Utils
Return a pair of host and port extracted from the sparkUrl.
extractIdentifier(CaseInsensitiveStringMap) - Method in interface org.apache.spark.sql.connector.catalog.SupportsCatalogOptions
Return a Identifier instance that can identify a table for a DataSource given DataFrame[Reader|Writer] options.
extractLongDistribution(Seq<Tuple2<TaskInfo, TaskMetrics>>, Function2<TaskInfo, TaskMetrics, Object>) - Static method in class org.apache.spark.scheduler.StatsReportListener
 
extractParamMap(ParamMap) - Method in interface org.apache.spark.ml.param.Params
Extracts the embedded default param values and user-supplied values, and then merges them with extra values from input into a flat param map, where the latter value is used if there exist conflicts, i.e., with ordering: default param values less than user-supplied values less than extra.
extractParamMap() - Method in interface org.apache.spark.ml.param.Params
extractParamMap with no extra values.
extractTimeTravelTimestamp(CaseInsensitiveStringMap) - Method in interface org.apache.spark.sql.connector.catalog.SupportsCatalogOptions
Extracts the timestamp string for time travel from the given options.
extractTimeTravelVersion(CaseInsensitiveStringMap) - Method in interface org.apache.spark.sql.connector.catalog.SupportsCatalogOptions
Extracts the version string for time travel from the given options.
extractWeightedLabeledPoints(Dataset<?>) - Method in interface org.apache.spark.ml.regression.IsotonicRegressionBase
Extracts (label, feature, weight) from input dataset.
extraOptimizations() - Method in class org.apache.spark.sql.ExperimentalMethods
 
extraStrategies() - Method in class org.apache.spark.sql.ExperimentalMethods
Allows extra strategies to be injected into the query planner at runtime.
eye(int) - Static method in class org.apache.spark.ml.linalg.DenseMatrix
Generate an Identity Matrix in DenseMatrix format.
eye(int) - Static method in class org.apache.spark.ml.linalg.Matrices
Generate a dense Identity Matrix in Matrix format.
eye(int) - Static method in class org.apache.spark.mllib.linalg.DenseMatrix
Generate an Identity Matrix in DenseMatrix format.
eye(int) - Static method in class org.apache.spark.mllib.linalg.Matrices
Generate a dense Identity Matrix in Matrix format.

F

f1Measure() - Method in class org.apache.spark.mllib.evaluation.MultilabelMetrics
Returns document-based f1-measure averaged by the number of documents
f1Measure(double) - Method in class org.apache.spark.mllib.evaluation.MultilabelMetrics
Returns f1-measure for a given label (category)
factorial(Column) - Static method in class org.apache.spark.sql.functions
Computes the factorial of the given value.
FactorizationMachines - Interface in org.apache.spark.ml.regression
 
FactorizationMachinesParams - Interface in org.apache.spark.ml.regression
Params for Factorization Machines
factors() - Method in class org.apache.spark.ml.classification.FMClassificationModel
 
factors() - Method in class org.apache.spark.ml.regression.FMRegressionModel
 
factorSize() - Method in class org.apache.spark.ml.classification.FMClassificationModel
 
factorSize() - Method in class org.apache.spark.ml.classification.FMClassifier
 
factorSize() - Method in interface org.apache.spark.ml.regression.FactorizationMachinesParams
Param for dimensionality of the factors (&gt;= 0)
factorSize() - Method in class org.apache.spark.ml.regression.FMRegressionModel
 
factorSize() - Method in class org.apache.spark.ml.regression.FMRegressor
 
failed() - Method in class org.apache.spark.scheduler.TaskInfo
 
FAILED() - Static method in class org.apache.spark.TaskState
 
FAILED_TASKS_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
 
FAILED_TASKS_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary
 
FAILED_TASKS_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
 
failedExecuteUserDefinedFunctionError(String, String, String, Throwable) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
failedJobIds() - Method in class org.apache.spark.status.api.v1.sql.ExecutionData
 
failedMergingSchemaError(StructType, SparkException) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
failedParsingDescriptorError(String, Throwable) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
failedParsingStructTypeError(String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
failedRenameTempFileError(Path, Path) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
failedStages() - Method in class org.apache.spark.status.LiveJob
 
failedTasks() - Method in class org.apache.spark.status.api.v1.ExecutorMetricsDistributions
 
failedTasks() - Method in class org.apache.spark.status.api.v1.ExecutorStageSummary
 
failedTasks() - Method in class org.apache.spark.status.api.v1.ExecutorSummary
 
failedTasks() - Method in class org.apache.spark.status.LiveExecutorStageSummary
 
failedTasks() - Method in class org.apache.spark.status.LiveJob
 
failedTasks() - Method in class org.apache.spark.status.LiveStage
 
failedToCastValueToDataTypeForPartitionColumnError(String, DataType, String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
failedToCompileMsg(Exception) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
failedToExecuteQueryError(Throwable) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
failedToFindAvroDataSourceError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
failedToFindKafkaDataSourceError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
failedToGenerateEpochMarkerError(Throwable) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
failedToInstantiateConstructorForCatalogError(String, String, Exception) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
failedToMergeIncompatibleSchemasError(StructType, StructType, Throwable) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
failedToParseExistenceDefaultAsLiteral(String, String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
failedToPushRowIntoRowQueueError(String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
failedToReadDataError(Throwable) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
failedToReadDeltaFileError(Path, String, int) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
failedToReadSnapshotFileError(Path, String, String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
failedToRebuildExpressionError(Filter) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
failToConvertValueToJsonError(Object, Class<?>, DataType) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
failToCreateCheckpointPathError(Path) - Static method in class org.apache.spark.errors.SparkCoreErrors
 
failToCreateDirectoryError(String, int) - Static method in class org.apache.spark.errors.SparkCoreErrors
 
failToGetBlockWithLockError(BlockId) - Static method in class org.apache.spark.errors.SparkCoreErrors
 
failToGetNonShuffleBlockError(BlockId, Throwable) - Static method in class org.apache.spark.errors.SparkCoreErrors
 
failToParseDateTimeInNewParserError(String, Throwable) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
failToRecognizePatternAfterUpgradeError(String, Throwable, String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
failToRecognizePatternError(String, Throwable, String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
failToResolveDataSourceForTableError(CatalogTable, String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
failToSerializeTaskError(Throwable) - Static method in class org.apache.spark.errors.SparkCoreErrors
 
failToSetOriginalACLBackError(String, Path, Throwable) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
failToStoreBlockOnBlockManagerError(BlockManagerId, BlockId) - Static method in class org.apache.spark.errors.SparkCoreErrors
 
failToTruncateTableWhenRemovingDataError(String, Path, Throwable) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
Failure() - Static method in class org.apache.spark.ml.feature.RFormulaParser
 
failure(String) - Static method in class org.apache.spark.ml.feature.RFormulaParser
 
FAILURE_REASON_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.StageData
 
failureReason() - Method in class org.apache.spark.scheduler.StageInfo
If the stage failed, the reason why.
failureReason() - Method in class org.apache.spark.status.api.v1.StageData
 
failureReason() - Method in class org.apache.spark.status.api.v1.streaming.OutputOperationInfo
 
failureReason() - Method in class org.apache.spark.streaming.scheduler.OutputOperationInfo
 
failureReasonCell(String, int, boolean) - Static method in class org.apache.spark.streaming.ui.UIUtils
 
FAIR() - Static method in class org.apache.spark.scheduler.SchedulingMode
 
fallbackV1RelationReportsInconsistentSchemaError(StructType, StructType) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
FalsePositiveRate - Class in org.apache.spark.mllib.evaluation.binary
False positive rate.
FalsePositiveRate() - Constructor for class org.apache.spark.mllib.evaluation.binary.FalsePositiveRate
 
falsePositiveRate(double) - Method in class org.apache.spark.mllib.evaluation.MulticlassMetrics
Returns false positive rate for a given label (category)
falsePositiveRateByLabel() - Method in interface org.apache.spark.ml.classification.ClassificationSummary
Returns false positive rate for each label (category).
family() - Method in class org.apache.spark.ml.classification.LogisticRegression
 
family() - Method in class org.apache.spark.ml.classification.LogisticRegressionModel
 
family() - Method in interface org.apache.spark.ml.classification.LogisticRegressionParams
Param for the name of family which is a description of the label distribution to be used in the model.
family() - Method in class org.apache.spark.ml.regression.GeneralizedLinearRegression
 
family() - Method in interface org.apache.spark.ml.regression.GeneralizedLinearRegressionBase
Param for the name of family which is a description of the error distribution to be used in the model.
family() - Method in class org.apache.spark.ml.regression.GeneralizedLinearRegressionModel
 
Family$() - Constructor for class org.apache.spark.ml.regression.GeneralizedLinearRegression.Family$
 
FamilyAndLink$() - Constructor for class org.apache.spark.ml.regression.GeneralizedLinearRegression.FamilyAndLink$
 
fdr() - Method in interface org.apache.spark.ml.feature.SelectorParams
The upper bound of the expected false discovery rate.
fdr() - Method in class org.apache.spark.mllib.feature.ChiSqSelector
 
feature() - Method in class org.apache.spark.mllib.feature.ChiSqSelectorModel.SaveLoadV1_0$.Data
 
feature() - Method in class org.apache.spark.mllib.tree.model.DecisionTreeModel.SaveLoadV1_0$.SplitData
 
feature() - Method in class org.apache.spark.mllib.tree.model.Split
 
FeatureHasher - Class in org.apache.spark.ml.feature
Feature hashing projects a set of categorical or numerical features into a feature vector of specified dimension (typically substantially smaller than that of the original feature space).
FeatureHasher(String) - Constructor for class org.apache.spark.ml.feature.FeatureHasher
 
FeatureHasher() - Constructor for class org.apache.spark.ml.feature.FeatureHasher
 
featureImportances() - Method in class org.apache.spark.ml.classification.DecisionTreeClassificationModel
 
featureImportances() - Method in class org.apache.spark.ml.classification.GBTClassificationModel
 
featureImportances() - Method in class org.apache.spark.ml.classification.RandomForestClassificationModel
 
featureImportances() - Method in class org.apache.spark.ml.regression.DecisionTreeRegressionModel
 
featureImportances() - Method in class org.apache.spark.ml.regression.GBTRegressionModel
 
featureImportances() - Method in class org.apache.spark.ml.regression.RandomForestRegressionModel
 
featureIndex() - Method in class org.apache.spark.ml.regression.IsotonicRegression
 
featureIndex() - Method in interface org.apache.spark.ml.regression.IsotonicRegressionBase
Param for the index of the feature if featuresCol is a vector column (default: 0), no effect otherwise.
featureIndex() - Method in class org.apache.spark.ml.regression.IsotonicRegressionModel
 
featureIndex() - Method in class org.apache.spark.ml.tree.CategoricalSplit
 
featureIndex() - Method in class org.apache.spark.ml.tree.ContinuousSplit
 
featureIndex() - Method in class org.apache.spark.ml.tree.DecisionTreeModelReadWrite.SplitData
 
featureIndex() - Method in interface org.apache.spark.ml.tree.Split
Index of feature which this split tests
features() - Method in class org.apache.spark.ml.feature.LabeledPoint
 
features() - Method in class org.apache.spark.mllib.regression.LabeledPoint
 
featuresCol() - Method in interface org.apache.spark.ml.classification.LogisticRegressionSummary
Field in "predictions" which gives the features of each instance as a vector.
featuresCol() - Method in class org.apache.spark.ml.classification.LogisticRegressionSummaryImpl
 
featuresCol() - Method in class org.apache.spark.ml.classification.OneVsRest
 
featuresCol() - Method in class org.apache.spark.ml.classification.OneVsRestModel
 
featuresCol() - Method in class org.apache.spark.ml.clustering.BisectingKMeans
 
featuresCol() - Method in class org.apache.spark.ml.clustering.BisectingKMeansModel
 
featuresCol() - Method in class org.apache.spark.ml.clustering.ClusteringSummary
 
featuresCol() - Method in class org.apache.spark.ml.clustering.GaussianMixture
 
featuresCol() - Method in class org.apache.spark.ml.clustering.GaussianMixtureModel
 
featuresCol() - Method in class org.apache.spark.ml.clustering.KMeans
 
featuresCol() - Method in class org.apache.spark.ml.clustering.KMeansModel
 
featuresCol() - Method in class org.apache.spark.ml.clustering.LDA
 
featuresCol() - Method in class org.apache.spark.ml.clustering.LDAModel
 
featuresCol() - Method in class org.apache.spark.ml.evaluation.ClusteringEvaluator
 
featuresCol() - Method in class org.apache.spark.ml.feature.RFormula
 
featuresCol() - Method in class org.apache.spark.ml.feature.RFormulaModel
 
featuresCol() - Method in class org.apache.spark.ml.feature.UnivariateFeatureSelector
 
featuresCol() - Method in class org.apache.spark.ml.feature.UnivariateFeatureSelectorModel
 
featuresCol() - Method in class org.apache.spark.ml.feature.VarianceThresholdSelector
 
featuresCol() - Method in class org.apache.spark.ml.feature.VarianceThresholdSelectorModel
 
featuresCol() - Method in interface org.apache.spark.ml.param.shared.HasFeaturesCol
Param for features column name.
featuresCol() - Method in class org.apache.spark.ml.PredictionModel
 
featuresCol() - Method in class org.apache.spark.ml.Predictor
 
featuresCol() - Method in class org.apache.spark.ml.regression.IsotonicRegression
 
featuresCol() - Method in class org.apache.spark.ml.regression.IsotonicRegressionModel
 
featuresCol() - Method in class org.apache.spark.ml.regression.LinearRegressionSummary
 
featureSubsetStrategy() - Method in class org.apache.spark.ml.classification.GBTClassificationModel
 
featureSubsetStrategy() - Method in class org.apache.spark.ml.classification.GBTClassifier
 
featureSubsetStrategy() - Method in class org.apache.spark.ml.classification.RandomForestClassificationModel
 
featureSubsetStrategy() - Method in class org.apache.spark.ml.classification.RandomForestClassifier
 
featureSubsetStrategy() - Method in class org.apache.spark.ml.regression.GBTRegressionModel
 
featureSubsetStrategy() - Method in class org.apache.spark.ml.regression.GBTRegressor
 
featureSubsetStrategy() - Method in class org.apache.spark.ml.regression.RandomForestRegressionModel
 
featureSubsetStrategy() - Method in class org.apache.spark.ml.regression.RandomForestRegressor
 
featureSubsetStrategy() - Method in interface org.apache.spark.ml.tree.TreeEnsembleParams
The number of features to consider for splits at each tree node.
featureSum() - Method in class org.apache.spark.ml.evaluation.SquaredEuclideanSilhouette.ClusterStats
 
featureType() - Method in class org.apache.spark.ml.feature.UnivariateFeatureSelector
 
featureType() - Method in class org.apache.spark.ml.feature.UnivariateFeatureSelectorModel
 
featureType() - Method in interface org.apache.spark.ml.feature.UnivariateFeatureSelectorParams
The feature type.
FeatureType - Class in org.apache.spark.mllib.tree.configuration
Enum to describe whether a feature is "continuous" or "categorical"
FeatureType() - Constructor for class org.apache.spark.mllib.tree.configuration.FeatureType
 
featureType() - Method in class org.apache.spark.mllib.tree.model.DecisionTreeModel.SaveLoadV1_0$.SplitData
 
featureType() - Method in class org.apache.spark.mllib.tree.model.Split
 
FETCH_WAIT_TIME() - Method in class org.apache.spark.InternalAccumulator.shuffleRead$
 
FETCH_WAIT_TIME_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions
 
FETCH_WAIT_TIME_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics
 
fetchAllPushMergedLocalBlocks(LinkedHashSet<BlockId>) - Method in class org.apache.spark.storage.PushBasedFetchHelper
This is executed by the task thread when the iterator is initialized.
FetchFailed - Class in org.apache.spark
:: DeveloperApi :: Task failed to fetch shuffle data from a remote node.
FetchFailed(BlockManagerId, int, long, int, int, String) - Constructor for class org.apache.spark.FetchFailed
 
fetchFailedError(BlockManagerId, int, long, int, int, String, Throwable) - Static method in class org.apache.spark.errors.SparkCoreErrors
 
fetchFile(String, File, SparkConf, Configuration, long, boolean, boolean) - Static method in class org.apache.spark.util.Utils
Download a file or directory to target directory.
fetchPct() - Method in class org.apache.spark.scheduler.RuntimePercentage
 
fetchWaitTime() - Method in class org.apache.spark.status.api.v1.ShuffleReadMetricDistributions
 
fetchWaitTime() - Method in class org.apache.spark.status.api.v1.ShuffleReadMetrics
 
field() - Method in class org.apache.spark.sql.connector.expressions.Extract
 
field() - Method in class org.apache.spark.storage.BroadcastBlockId
 
fieldCannotBeNullError(int, String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
fieldCannotBeNullMsg(int, String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
fieldDiffersFromDerivedLocalDateError(ChronoField, int, int, LocalDate) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
fieldIndex(String) - Method in interface org.apache.spark.sql.Row
Returns the index of a given field name.
fieldIndex(String) - Method in class org.apache.spark.sql.types.StructType
Returns the index of a given field.
fieldIndexOnRowWithoutSchemaError() - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
fieldNames() - Method in class org.apache.spark.sql.connector.catalog.TableChange.AddColumn
 
fieldNames() - Method in interface org.apache.spark.sql.connector.catalog.TableChange.ColumnChange
 
fieldNames() - Method in class org.apache.spark.sql.connector.catalog.TableChange.DeleteColumn
 
fieldNames() - Method in class org.apache.spark.sql.connector.catalog.TableChange.RenameColumn
 
fieldNames() - Method in class org.apache.spark.sql.connector.catalog.TableChange.UpdateColumnComment
 
fieldNames() - Method in class org.apache.spark.sql.connector.catalog.TableChange.UpdateColumnDefaultValue
 
fieldNames() - Method in class org.apache.spark.sql.connector.catalog.TableChange.UpdateColumnNullability
 
fieldNames() - Method in class org.apache.spark.sql.connector.catalog.TableChange.UpdateColumnPosition
 
fieldNames() - Method in class org.apache.spark.sql.connector.catalog.TableChange.UpdateColumnType
 
fieldNames() - Method in interface org.apache.spark.sql.connector.expressions.NamedReference
Returns the referenced field name as an array of String parts.
fieldNames() - Method in class org.apache.spark.sql.types.StructType
Returns all field names in an array.
fieldNumberMismatchForDeserializerError(StructType, int) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
fields() - Method in class org.apache.spark.sql.types.StructType
 
fieldToString(byte) - Static method in class org.apache.spark.sql.types.DayTimeIntervalType
 
fieldToString(byte) - Static method in class org.apache.spark.sql.types.YearMonthIntervalType
 
FIFO() - Static method in class org.apache.spark.scheduler.SchedulingMode
 
FileBasedTopologyMapper - Class in org.apache.spark.storage
A simple file based topology mapper.
FileBasedTopologyMapper(SparkConf) - Constructor for class org.apache.spark.storage.FileBasedTopologyMapper
 
fileLengthExceedsMaxLengthError(FileStatus, int) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
fileNotFoundError(FileNotFoundException) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
fileReader() - Method in interface org.apache.spark.sql.avro.AvroUtils.RowReader
 
files() - Method in class org.apache.spark.SparkContext
 
fileStream(String, Class<K>, Class<V>, Class<F>) - Method in class org.apache.spark.streaming.api.java.JavaStreamingContext
Deprecated.
Create an input stream that monitors a Hadoop-compatible filesystem for new files and reads them using the given key-value types and input format.
fileStream(String, Class<K>, Class<V>, Class<F>, Function<Path, Boolean>, boolean) - Method in class org.apache.spark.streaming.api.java.JavaStreamingContext
Deprecated.
Create an input stream that monitors a Hadoop-compatible filesystem for new files and reads them using the given key-value types and input format.
fileStream(String, Class<K>, Class<V>, Class<F>, Function<Path, Boolean>, boolean, Configuration) - Method in class org.apache.spark.streaming.api.java.JavaStreamingContext
Deprecated.
Create an input stream that monitors a Hadoop-compatible filesystem for new files and reads them using the given key-value types and input format.
fileStream(String, ClassTag<K>, ClassTag<V>, ClassTag<F>) - Method in class org.apache.spark.streaming.StreamingContext
Deprecated.
Create an input stream that monitors a Hadoop-compatible filesystem for new files and reads them using the given key-value types and input format.
fileStream(String, Function1<Path, Object>, boolean, ClassTag<K>, ClassTag<V>, ClassTag<F>) - Method in class org.apache.spark.streaming.StreamingContext
Deprecated.
Create an input stream that monitors a Hadoop-compatible filesystem for new files and reads them using the given key-value types and input format.
fileStream(String, Function1<Path, Object>, boolean, Configuration, ClassTag<K>, ClassTag<V>, ClassTag<F>) - Method in class org.apache.spark.streaming.StreamingContext
Deprecated.
Create an input stream that monitors a Hadoop-compatible filesystem for new files and reads them using the given key-value types and input format.
fill(long) - Method in class org.apache.spark.sql.DataFrameNaFunctions
Returns a new DataFrame that replaces null or NaN values in numeric columns with value.
fill(double) - Method in class org.apache.spark.sql.DataFrameNaFunctions
Returns a new DataFrame that replaces null or NaN values in numeric columns with value.
fill(String) - Method in class org.apache.spark.sql.DataFrameNaFunctions
Returns a new DataFrame that replaces null values in string columns with value.
fill(long, String[]) - Method in class org.apache.spark.sql.DataFrameNaFunctions
Returns a new DataFrame that replaces null or NaN values in specified numeric columns.
fill(double, String[]) - Method in class org.apache.spark.sql.DataFrameNaFunctions
Returns a new DataFrame that replaces null or NaN values in specified numeric columns.
fill(long, Seq<String>) - Method in class org.apache.spark.sql.DataFrameNaFunctions
(Scala-specific) Returns a new DataFrame that replaces null or NaN values in specified numeric columns.
fill(double, Seq<String>) - Method in class org.apache.spark.sql.DataFrameNaFunctions
(Scala-specific) Returns a new DataFrame that replaces null or NaN values in specified numeric columns.
fill(String, String[]) - Method in class org.apache.spark.sql.DataFrameNaFunctions
Returns a new DataFrame that replaces null values in specified string columns.
fill(String, Seq<String>) - Method in class org.apache.spark.sql.DataFrameNaFunctions
(Scala-specific) Returns a new DataFrame that replaces null values in specified string columns.
fill(boolean) - Method in class org.apache.spark.sql.DataFrameNaFunctions
Returns a new DataFrame that replaces null values in boolean columns with value.
fill(boolean, Seq<String>) - Method in class org.apache.spark.sql.DataFrameNaFunctions
(Scala-specific) Returns a new DataFrame that replaces null values in specified boolean columns.
fill(boolean, String[]) - Method in class org.apache.spark.sql.DataFrameNaFunctions
Returns a new DataFrame that replaces null values in specified boolean columns.
fill(Map<String, Object>) - Method in class org.apache.spark.sql.DataFrameNaFunctions
Returns a new DataFrame that replaces null values.
fill(Map<String, Object>) - Method in class org.apache.spark.sql.DataFrameNaFunctions
(Scala-specific) Returns a new DataFrame that replaces null values.
filter(Function<Double, Boolean>) - Method in class org.apache.spark.api.java.JavaDoubleRDD
Return a new RDD containing only the elements that satisfy a predicate.
filter(Function<Tuple2<K, V>, Boolean>) - Method in class org.apache.spark.api.java.JavaPairRDD
Return a new RDD containing only the elements that satisfy a predicate.
filter(Function<T, Boolean>) - Method in class org.apache.spark.api.java.JavaRDD
Return a new RDD containing only the elements that satisfy a predicate.
filter(Function1<Graph<VD, ED>, Graph<VD2, ED2>>, Function1<EdgeTriplet<VD2, ED2>, Object>, Function2<Object, VD2, Object>, ClassTag<VD2>, ClassTag<ED2>) - Method in class org.apache.spark.graphx.GraphOps
Filter the graph by computing some values to filter on, and applying the predicates.
filter(Function1<EdgeTriplet<VD, ED>, Object>, Function2<Object, VD, Object>) - Method in class org.apache.spark.graphx.impl.EdgeRDDImpl
 
filter(Function1<Tuple2<Object, VD>, Object>) - Method in class org.apache.spark.graphx.VertexRDD
Restricts the vertex set to the set of vertices satisfying the given predicate.
filter(Params) - Method in class org.apache.spark.ml.param.ParamMap
Filters this param map for the given parent.
filter(Function1<T, Object>) - Method in class org.apache.spark.rdd.RDD
Return a new RDD containing only the elements that satisfy a predicate.
filter(Filter[]) - Method in interface org.apache.spark.sql.connector.read.SupportsRuntimeFiltering
Filters this scan using runtime filters.
filter(Predicate[]) - Method in interface org.apache.spark.sql.connector.read.SupportsRuntimeFiltering
 
filter(Predicate[]) - Method in interface org.apache.spark.sql.connector.read.SupportsRuntimeV2Filtering
Filters this scan using runtime predicates.
filter(Column) - Method in class org.apache.spark.sql.Dataset
Filters rows using the given condition.
filter(String) - Method in class org.apache.spark.sql.Dataset
Filters rows using the given SQL expression.
filter(Function1<T, Object>) - Method in class org.apache.spark.sql.Dataset
(Scala-specific) Returns a new Dataset that only contains elements where func returns true.
filter(FilterFunction<T>) - Method in class org.apache.spark.sql.Dataset
(Java-specific) Returns a new Dataset that only contains elements where func returns true.
filter(Column, Function1<Column, Column>) - Static method in class org.apache.spark.sql.functions
Returns an array of elements for which a predicate holds in a given array.
filter(Column, Function2<Column, Column, Column>) - Static method in class org.apache.spark.sql.functions
Returns an array of elements for which a predicate holds in a given array.
Filter - Class in org.apache.spark.sql.sources
A filter predicate for data sources.
Filter() - Constructor for class org.apache.spark.sql.sources.Filter
 
filter() - Method in class org.apache.spark.storage.BlockManagerMessages.GetMatchingBlockIds
 
filter(Function<T, Boolean>) - Method in class org.apache.spark.streaming.api.java.JavaDStream
Return a new DStream containing only the elements that satisfy a predicate.
filter(Function<Tuple2<K, V>, Boolean>) - Method in class org.apache.spark.streaming.api.java.JavaPairDStream
Return a new DStream containing only the elements that satisfy a predicate.
filter(Function1<T, Object>) - Method in class org.apache.spark.streaming.dstream.DStream
Return a new DStream containing only the elements that satisfy a predicate.
filterAttributes() - Method in interface org.apache.spark.sql.connector.read.SupportsRuntimeFiltering
Returns attributes this scan can be filtered by at runtime.
filterAttributes() - Method in interface org.apache.spark.sql.connector.read.SupportsRuntimeV2Filtering
Returns attributes this scan can be filtered by at runtime.
filterByRange(K, K) - Method in class org.apache.spark.api.java.JavaPairRDD
Return a RDD containing only the elements in the inclusive range lower to upper.
filterByRange(Comparator<K>, K, K) - Method in class org.apache.spark.api.java.JavaPairRDD
Return a RDD containing only the elements in the inclusive range lower to upper.
filterByRange(K, K) - Method in class org.apache.spark.rdd.OrderedRDDFunctions
Returns an RDD containing only the elements in the inclusive range lower to upper.
FilterFunction<T> - Interface in org.apache.spark.api.java.function
Base interface for a function used in Dataset's filter function.
filterName() - Method in class org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.AddWebUIFilter
 
filterParams() - Method in class org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.AddWebUIFilter
 
finalStorageLevel() - Method in class org.apache.spark.ml.recommendation.ALS
 
finalStorageLevel() - Method in interface org.apache.spark.ml.recommendation.ALSParams
Param for StorageLevel for ALS model factors.
findAccessedFields(SerializedLambda, ClassLoader, Map<Class<?>, Set<String>>, boolean) - Static method in class org.apache.spark.util.IndylambdaScalaClosures
Scans an indylambda Scala closure, along with its lexically nested closures, and populate the accessed fields info on which fields on the outer object are accessed.
findClass(String) - Method in class org.apache.spark.util.ParentClassLoader
 
findColumnPosition(Seq<String>, StructType, Function2<String, String, Object>) - Static method in class org.apache.spark.sql.util.SchemaUtils
Returns the given column's ordinal within the given schema.
findFrequentSequentialPatterns(Dataset<?>) - Method in class org.apache.spark.ml.fpm.PrefixSpan
Finds the complete set of frequent sequential patterns in the input sequences of itemsets.
findListenersByClass(ClassTag<T>) - Method in interface org.apache.spark.util.ListenerBus
 
findMatchingTokenClusterConfig(SparkConf, String) - Static method in class org.apache.spark.kafka010.KafkaTokenUtil
 
findMissingFields(StructType, StructType, Function2<String, String, Object>) - Static method in class org.apache.spark.sql.types.StructType
Returns a StructType that contains missing fields recursively from source to target.
findMissingPartitions() - Method in class org.apache.spark.ShuffleStatus
Returns the sequence of partition ids that are missing (i.e.
findMultipleDataSourceError(String, Seq<String>) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
findSynonyms(String, int) - Method in class org.apache.spark.ml.feature.Word2VecModel
Find "num" number of words closest in similarity to the given word, not including the word itself.
findSynonyms(Vector, int) - Method in class org.apache.spark.ml.feature.Word2VecModel
Find "num" number of words whose vector representation is most similar to the supplied vector.
findSynonyms(String, int) - Method in class org.apache.spark.mllib.feature.Word2VecModel
Find synonyms of a word; do not include the word itself in results.
findSynonyms(Vector, int) - Method in class org.apache.spark.mllib.feature.Word2VecModel
Find synonyms of the vector representation of a word, possibly including any words in the model vocabulary whose vector representation is the supplied vector.
findSynonymsArray(Vector, int) - Method in class org.apache.spark.ml.feature.Word2VecModel
Find "num" number of words whose vector representation is most similar to the supplied vector.
findSynonymsArray(String, int) - Method in class org.apache.spark.ml.feature.Word2VecModel
Find "num" number of words closest in similarity to the given word, not including the word itself.
finish(OpenHashMap<String, Object>[]) - Method in class org.apache.spark.ml.feature.StringIndexerAggregator
 
finish(BUF) - Method in class org.apache.spark.sql.expressions.Aggregator
Transform the output of the reduction.
finished() - Method in class org.apache.spark.scheduler.TaskInfo
 
FINISHED() - Static method in class org.apache.spark.TaskState
 
finishTime() - Method in class org.apache.spark.scheduler.TaskInfo
The time when the task has completed successfully (including the time to remotely fetch results, if necessary).
first() - Method in class org.apache.spark.api.java.JavaDoubleRDD
 
first() - Method in class org.apache.spark.api.java.JavaPairRDD
 
first() - Method in interface org.apache.spark.api.java.JavaRDDLike
Return the first element in this RDD.
first() - Method in class org.apache.spark.rdd.RDD
Return the first element in this RDD.
first() - Static method in interface org.apache.spark.sql.connector.catalog.TableChange.ColumnPosition
 
first() - Method in class org.apache.spark.sql.Dataset
Returns the first row.
first(Column, boolean) - Static method in class org.apache.spark.sql.functions
Aggregate function: returns the first value in a group.
first(String, boolean) - Static method in class org.apache.spark.sql.functions
Aggregate function: returns the first value of a column in a group.
first(Column) - Static method in class org.apache.spark.sql.functions
Aggregate function: returns the first value in a group.
first(String) - Static method in class org.apache.spark.sql.functions
Aggregate function: returns the first value of a column in a group.
FIRST_TASK_LAUNCHED_TIME_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.StageData
 
firstFailureReason() - Method in class org.apache.spark.status.api.v1.streaming.BatchInfo
 
firstLaunchTime() - Method in class org.apache.spark.status.LiveStage
 
firstTaskLaunchedTime() - Method in class org.apache.spark.status.api.v1.StageData
 
fit(Dataset<?>) - Method in class org.apache.spark.ml.classification.OneVsRest
 
fit(Dataset<?>) - Method in class org.apache.spark.ml.clustering.BisectingKMeans
 
fit(Dataset<?>) - Method in class org.apache.spark.ml.clustering.GaussianMixture
 
fit(Dataset<?>) - Method in class org.apache.spark.ml.clustering.KMeans
 
fit(Dataset<?>) - Method in class org.apache.spark.ml.clustering.LDA
 
fit(Dataset<?>, ParamPair<?>, ParamPair<?>...) - Method in class org.apache.spark.ml.Estimator
Fits a single model to the input data with optional parameters.
fit(Dataset<?>, ParamPair<?>, Seq<ParamPair<?>>) - Method in class org.apache.spark.ml.Estimator
Fits a single model to the input data with optional parameters.
fit(Dataset<?>, ParamMap) - Method in class org.apache.spark.ml.Estimator
Fits a single model to the input data with provided parameter map.
fit(Dataset<?>) - Method in class org.apache.spark.ml.Estimator
Fits a model to the input data.
fit(Dataset<?>, Seq<ParamMap>) - Method in class org.apache.spark.ml.Estimator
Fits multiple models to the input data with multiple sets of parameters.
fit(Dataset<?>) - Method in class org.apache.spark.ml.feature.ChiSqSelector
Deprecated.
 
fit(Dataset<?>) - Method in class org.apache.spark.ml.feature.CountVectorizer
 
fit(Dataset<?>) - Method in class org.apache.spark.ml.feature.IDF
 
fit(Dataset<?>) - Method in class org.apache.spark.ml.feature.Imputer
 
fit(Dataset<?>) - Method in class org.apache.spark.ml.feature.MaxAbsScaler
 
fit(Dataset<?>) - Method in class org.apache.spark.ml.feature.MinMaxScaler
 
fit(Dataset<?>) - Method in class org.apache.spark.ml.feature.OneHotEncoder
 
fit(Dataset<?>) - Method in class org.apache.spark.ml.feature.PCA
Computes a PCAModel that contains the principal components of the input vectors.
fit(Dataset<?>) - Method in class org.apache.spark.ml.feature.QuantileDiscretizer
 
fit(Dataset<?>) - Method in class org.apache.spark.ml.feature.RFormula
 
fit(Dataset<?>) - Method in class org.apache.spark.ml.feature.RobustScaler
 
fit(Dataset<?>) - Method in class org.apache.spark.ml.feature.StandardScaler
 
fit(Dataset<?>) - Method in class org.apache.spark.ml.feature.StringIndexer
 
fit(Dataset<?>) - Method in class org.apache.spark.ml.feature.UnivariateFeatureSelector
 
fit(Dataset<?>) - Method in class org.apache.spark.ml.feature.VarianceThresholdSelector
 
fit(Dataset<?>) - Method in class org.apache.spark.ml.feature.VectorIndexer
 
fit(Dataset<?>) - Method in class org.apache.spark.ml.feature.Word2Vec
 
fit(Dataset<?>) - Method in class org.apache.spark.ml.fpm.FPGrowth
 
fit(Dataset<?>) - Method in class org.apache.spark.ml.Pipeline
Fits the pipeline to the input dataset with additional parameters.
fit(Dataset<?>) - Method in class org.apache.spark.ml.Predictor
 
fit(Dataset<?>) - Method in class org.apache.spark.ml.recommendation.ALS
 
fit(Dataset<?>) - Method in class org.apache.spark.ml.regression.IsotonicRegression
 
fit(Dataset<?>) - Method in class org.apache.spark.ml.tuning.CrossValidator
 
fit(Dataset<?>) - Method in class org.apache.spark.ml.tuning.TrainValidationSplit
 
fit(RDD<LabeledPoint>) - Method in class org.apache.spark.mllib.feature.ChiSqSelector
Returns a ChiSquared feature selector.
fit(RDD<Vector>) - Method in class org.apache.spark.mllib.feature.IDF
Computes the inverse document frequency.
fit(JavaRDD<Vector>) - Method in class org.apache.spark.mllib.feature.IDF
Computes the inverse document frequency.
fit(RDD<Vector>) - Method in class org.apache.spark.mllib.feature.PCA
Computes a PCAModel that contains the principal components of the input vectors.
fit(JavaRDD<Vector>) - Method in class org.apache.spark.mllib.feature.PCA
Java-friendly version of fit().
fit(RDD<Vector>) - Method in class org.apache.spark.mllib.feature.StandardScaler
Computes the mean and variance and stores as a model to be used for later scaling.
fit(RDD<S>) - Method in class org.apache.spark.mllib.feature.Word2Vec
Computes the vector representation of each word in vocabulary.
fit(JavaRDD<S>) - Method in class org.apache.spark.mllib.feature.Word2Vec
Computes the vector representation of each word in vocabulary (Java version).
FitEnd<M extends Model<M>> - Class in org.apache.spark.ml
Event fired after Estimator.fit.
FitEnd() - Constructor for class org.apache.spark.ml.FitEnd
 
fitIntercept() - Method in class org.apache.spark.ml.classification.FMClassificationModel
 
fitIntercept() - Method in class org.apache.spark.ml.classification.FMClassifier
 
fitIntercept() - Method in class org.apache.spark.ml.classification.LinearSVC
 
fitIntercept() - Method in class org.apache.spark.ml.classification.LinearSVCModel
 
fitIntercept() - Method in class org.apache.spark.ml.classification.LogisticRegression
 
fitIntercept() - Method in class org.apache.spark.ml.classification.LogisticRegressionModel
 
fitIntercept() - Method in interface org.apache.spark.ml.param.shared.HasFitIntercept
Param for whether to fit an intercept term.
fitIntercept() - Method in class org.apache.spark.ml.regression.AFTSurvivalRegression
 
fitIntercept() - Method in class org.apache.spark.ml.regression.AFTSurvivalRegressionModel
 
fitIntercept() - Method in class org.apache.spark.ml.regression.FMRegressionModel
 
fitIntercept() - Method in class org.apache.spark.ml.regression.FMRegressor
 
fitIntercept() - Method in class org.apache.spark.ml.regression.GeneralizedLinearRegression
 
fitIntercept() - Method in class org.apache.spark.ml.regression.GeneralizedLinearRegressionModel
 
fitIntercept() - Method in class org.apache.spark.ml.regression.LinearRegression
 
fitIntercept() - Method in class org.apache.spark.ml.regression.LinearRegressionModel
 
fitLinear() - Method in class org.apache.spark.ml.classification.FMClassificationModel
 
fitLinear() - Method in class org.apache.spark.ml.classification.FMClassifier
 
fitLinear() - Method in interface org.apache.spark.ml.regression.FactorizationMachinesParams
Param for whether to fit linear term (aka 1-way term)
fitLinear() - Method in class org.apache.spark.ml.regression.FMRegressionModel
 
fitLinear() - Method in class org.apache.spark.ml.regression.FMRegressor
 
FitStart<M extends Model<M>> - Class in org.apache.spark.ml
Event fired before Estimator.fit.
FitStart() - Constructor for class org.apache.spark.ml.FitStart
 
Fixed$() - Constructor for class org.apache.spark.sql.types.DecimalType.Fixed$
 
flatMap(FlatMapFunction<T, U>) - Method in interface org.apache.spark.api.java.JavaRDDLike
Return a new RDD by first applying a function to all elements of this RDD, and then flattening the results.
flatMap(Function1<T, TraversableOnce<U>>, ClassTag<U>) - Method in class org.apache.spark.rdd.RDD
Return a new RDD by first applying a function to all elements of this RDD, and then flattening the results.
flatMap(Function1<T, TraversableOnce<U>>, Encoder<U>) - Method in class org.apache.spark.sql.Dataset
(Scala-specific) Returns a new Dataset by first applying a function to all elements of this Dataset, and then flattening the results.
flatMap(FlatMapFunction<T, U>, Encoder<U>) - Method in class org.apache.spark.sql.Dataset
(Java-specific) Returns a new Dataset by first applying a function to all elements of this Dataset, and then flattening the results.
flatMap(FlatMapFunction<T, U>) - Method in interface org.apache.spark.streaming.api.java.JavaDStreamLike
Return a new DStream by applying a function to all elements of this DStream, and then flattening the results
flatMap(Function1<T, TraversableOnce<U>>, ClassTag<U>) - Method in class org.apache.spark.streaming.dstream.DStream
Return a new DStream by applying a function to all elements of this DStream, and then flattening the results
FlatMapFunction<T,R> - Interface in org.apache.spark.api.java.function
A function that returns zero or more output records from each input record.
FlatMapFunction2<T1,T2,R> - Interface in org.apache.spark.api.java.function
A function that takes two inputs and returns zero or more output records.
flatMapGroups(Function2<K, Iterator<V>, TraversableOnce<U>>, Encoder<U>) - Method in class org.apache.spark.sql.KeyValueGroupedDataset
(Scala-specific) Applies the given function to each group of data.
flatMapGroups(FlatMapGroupsFunction<K, V, U>, Encoder<U>) - Method in class org.apache.spark.sql.KeyValueGroupedDataset
(Java-specific) Applies the given function to each group of data.
FlatMapGroupsFunction<K,V,R> - Interface in org.apache.spark.api.java.function
A function that returns zero or more output records from each grouping key and its values.
flatMapGroupsWithState(OutputMode, GroupStateTimeout, Function3<K, Iterator<V>, GroupState<S>, Iterator<U>>, Encoder<S>, Encoder<U>) - Method in class org.apache.spark.sql.KeyValueGroupedDataset
(Scala-specific) Applies the given function to each group of data, while maintaining a user-defined per-group state.
flatMapGroupsWithState(OutputMode, GroupStateTimeout, KeyValueGroupedDataset<K, S>, Function3<K, Iterator<V>, GroupState<S>, Iterator<U>>, Encoder<S>, Encoder<U>) - Method in class org.apache.spark.sql.KeyValueGroupedDataset
(Scala-specific) Applies the given function to each group of data, while maintaining a user-defined per-group state.
flatMapGroupsWithState(FlatMapGroupsWithStateFunction<K, V, S, U>, OutputMode, Encoder<S>, Encoder<U>, GroupStateTimeout) - Method in class org.apache.spark.sql.KeyValueGroupedDataset
(Java-specific) Applies the given function to each group of data, while maintaining a user-defined per-group state.
flatMapGroupsWithState(FlatMapGroupsWithStateFunction<K, V, S, U>, OutputMode, Encoder<S>, Encoder<U>, GroupStateTimeout, KeyValueGroupedDataset<K, S>) - Method in class org.apache.spark.sql.KeyValueGroupedDataset
(Java-specific) Applies the given function to each group of data, while maintaining a user-defined per-group state.
FlatMapGroupsWithStateFunction<K,V,S,R> - Interface in org.apache.spark.api.java.function
::Experimental:: Base interface for a map function used in org.apache.spark.sql.KeyValueGroupedDataset.flatMapGroupsWithState( FlatMapGroupsWithStateFunction, org.apache.spark.sql.streaming.OutputMode, org.apache.spark.sql.Encoder, org.apache.spark.sql.Encoder)
flatMapSortedGroups(Seq<Column>, Function2<K, Iterator<V>, TraversableOnce<U>>, Encoder<U>) - Method in class org.apache.spark.sql.KeyValueGroupedDataset
(Scala-specific) Applies the given function to each group of data.
flatMapSortedGroups(Column[], FlatMapGroupsFunction<K, V, U>, Encoder<U>) - Method in class org.apache.spark.sql.KeyValueGroupedDataset
(Java-specific) Applies the given function to each group of data.
flatMapToDouble(DoubleFlatMapFunction<T>) - Method in interface org.apache.spark.api.java.JavaRDDLike
Return a new RDD by first applying a function to all elements of this RDD, and then flattening the results.
flatMapToPair(PairFlatMapFunction<T, K2, V2>) - Method in interface org.apache.spark.api.java.JavaRDDLike
Return a new RDD by first applying a function to all elements of this RDD, and then flattening the results.
flatMapToPair(PairFlatMapFunction<T, K2, V2>) - Method in interface org.apache.spark.streaming.api.java.JavaDStreamLike
Return a new DStream by applying a function to all elements of this DStream, and then flattening the results
flatMapValues(FlatMapFunction<V, U>) - Method in class org.apache.spark.api.java.JavaPairRDD
Pass each value in the key-value pair RDD through a flatMap function without changing the keys; this also retains the original RDD's partitioning.
flatMapValues(Function1<V, TraversableOnce<U>>) - Method in class org.apache.spark.rdd.PairRDDFunctions
Pass each value in the key-value pair RDD through a flatMap function without changing the keys; this also retains the original RDD's partitioning.
flatMapValues(FlatMapFunction<V, U>) - Method in class org.apache.spark.streaming.api.java.JavaPairDStream
Return a new DStream by applying a flatmap function to the value of each key-value pairs in 'this' DStream without changing the key.
flatMapValues(Function1<V, TraversableOnce<U>>, ClassTag<U>) - Method in class org.apache.spark.streaming.dstream.PairDStreamFunctions
Return a new DStream by applying a flatmap function to the value of each key-value pairs in 'this' DStream without changing the key.
flatten(Column) - Static method in class org.apache.spark.sql.functions
Creates a single array from an array of arrays.
flattenArraysWithElementsExceedLimitError(long) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
FLOAT() - Static method in class org.apache.spark.sql.Encoders
An encoder for nullable float type.
FloatAsIfIntegral$() - Constructor for class org.apache.spark.sql.types.FloatType.FloatAsIfIntegral$
 
floatColumn(String[]) - Static method in class org.apache.parquet.filter2.predicate.SparkFilterApi
 
FloatExactNumeric - Class in org.apache.spark.sql.types
 
FloatExactNumeric() - Constructor for class org.apache.spark.sql.types.FloatExactNumeric
 
FloatParam - Class in org.apache.spark.ml.param
Specialized version of Param[Float] for Java.
FloatParam(String, String, String, Function1<Object, Object>) - Constructor for class org.apache.spark.ml.param.FloatParam
 
FloatParam(String, String, String) - Constructor for class org.apache.spark.ml.param.FloatParam
 
FloatParam(Identifiable, String, String, Function1<Object, Object>) - Constructor for class org.apache.spark.ml.param.FloatParam
 
FloatParam(Identifiable, String, String) - Constructor for class org.apache.spark.ml.param.FloatParam
 
FloatType - Static variable in class org.apache.spark.sql.types.DataTypes
Gets the FloatType object.
FloatType - Class in org.apache.spark.sql.types
The data type representing Float values.
FloatType() - Constructor for class org.apache.spark.sql.types.FloatType
 
FloatType.FloatAsIfIntegral - Interface in org.apache.spark.sql.types
 
FloatType.FloatAsIfIntegral$ - Class in org.apache.spark.sql.types
 
FloatType.FloatIsConflicted - Interface in org.apache.spark.sql.types
 
floor(Column, Column) - Static method in class org.apache.spark.sql.functions
Computes the floor of the given value of e to scale decimal places.
floor(Column) - Static method in class org.apache.spark.sql.functions
Computes the floor of the given value of e to 0 decimal places.
floor(String) - Static method in class org.apache.spark.sql.functions
Computes the floor of the given column value to 0 decimal places.
floor() - Method in class org.apache.spark.sql.types.Decimal
 
floor(Duration) - Method in class org.apache.spark.streaming.Time
 
floor(Duration, Time) - Method in class org.apache.spark.streaming.Time
 
flush() - Method in class org.apache.spark.serializer.SerializationStream
 
flush() - Method in class org.apache.spark.storage.TimeTrackingOutputStream
 
FMClassificationModel - Class in org.apache.spark.ml.classification
Model produced by FMClassifier
FMClassificationSummary - Interface in org.apache.spark.ml.classification
Abstraction for FMClassifier results for a given model.
FMClassificationSummaryImpl - Class in org.apache.spark.ml.classification
FMClassifier results for a given model.
FMClassificationSummaryImpl(Dataset<Row>, String, String, String, String) - Constructor for class org.apache.spark.ml.classification.FMClassificationSummaryImpl
 
FMClassificationTrainingSummary - Interface in org.apache.spark.ml.classification
Abstraction for FMClassifier training results.
FMClassificationTrainingSummaryImpl - Class in org.apache.spark.ml.classification
FMClassifier training results.
FMClassificationTrainingSummaryImpl(Dataset<Row>, String, String, String, String, double[]) - Constructor for class org.apache.spark.ml.classification.FMClassificationTrainingSummaryImpl
 
FMClassifier - Class in org.apache.spark.ml.classification
Factorization Machines learning algorithm for classification.
FMClassifier(String) - Constructor for class org.apache.spark.ml.classification.FMClassifier
 
FMClassifier() - Constructor for class org.apache.spark.ml.classification.FMClassifier
 
FMClassifierParams - Interface in org.apache.spark.ml.classification
Params for FMClassifier.
fMeasure(double, double) - Method in class org.apache.spark.mllib.evaluation.MulticlassMetrics
Returns f-measure for a given label (category)
fMeasure(double) - Method in class org.apache.spark.mllib.evaluation.MulticlassMetrics
Returns f1-measure for a given label (category)
fMeasureByLabel(double) - Method in interface org.apache.spark.ml.classification.ClassificationSummary
Returns f-measure for each label (category).
fMeasureByLabel() - Method in interface org.apache.spark.ml.classification.ClassificationSummary
Returns f1-measure for each label (category).
fMeasureByThreshold() - Method in interface org.apache.spark.ml.classification.BinaryClassificationSummary
Returns a dataframe with two fields (threshold, F-Measure) curve with beta = 1.0.
fMeasureByThreshold() - Method in class org.apache.spark.ml.classification.BinaryLogisticRegressionSummaryImpl
 
fMeasureByThreshold() - Method in class org.apache.spark.ml.classification.BinaryRandomForestClassificationSummaryImpl
 
fMeasureByThreshold() - Method in class org.apache.spark.ml.classification.FMClassificationSummaryImpl
 
fMeasureByThreshold() - Method in class org.apache.spark.ml.classification.LinearSVCSummaryImpl
 
fMeasureByThreshold(double) - Method in class org.apache.spark.mllib.evaluation.BinaryClassificationMetrics
Returns the (threshold, F-Measure) curve.
fMeasureByThreshold() - Method in class org.apache.spark.mllib.evaluation.BinaryClassificationMetrics
Returns the (threshold, F-Measure) curve with beta = 1.0.
FMRegressionModel - Class in org.apache.spark.ml.regression
Model produced by FMRegressor.
FMRegressor - Class in org.apache.spark.ml.regression
Factorization Machines learning algorithm for regression.
FMRegressor(String) - Constructor for class org.apache.spark.ml.regression.FMRegressor
 
FMRegressor() - Constructor for class org.apache.spark.ml.regression.FMRegressor
 
FMRegressorParams - Interface in org.apache.spark.ml.regression
Params for FMRegressor
fold(T, Function2<T, T, T>) - Method in interface org.apache.spark.api.java.JavaRDDLike
Aggregate the elements of each partition, and then the results for all the partitions, using a given associative function and a neutral "zero value".
fold(T, Function2<T, T, T>) - Method in class org.apache.spark.rdd.RDD
Aggregate the elements of each partition, and then the results for all the partitions, using a given associative function and a neutral "zero value".
foldByKey(V, Partitioner, Function2<V, V, V>) - Method in class org.apache.spark.api.java.JavaPairRDD
Merge the values for each key using an associative function and a neutral "zero value" which may be added to the result an arbitrary number of times, and must not change the result (e.g ., Nil for list concatenation, 0 for addition, or 1 for multiplication.).
foldByKey(V, int, Function2<V, V, V>) - Method in class org.apache.spark.api.java.JavaPairRDD
Merge the values for each key using an associative function and a neutral "zero value" which may be added to the result an arbitrary number of times, and must not change the result (e.g ., Nil for list concatenation, 0 for addition, or 1 for multiplication.).
foldByKey(V, Function2<V, V, V>) - Method in class org.apache.spark.api.java.JavaPairRDD
Merge the values for each key using an associative function and a neutral "zero value" which may be added to the result an arbitrary number of times, and must not change the result (e.g., Nil for list concatenation, 0 for addition, or 1 for multiplication.).
foldByKey(V, Partitioner, Function2<V, V, V>) - Method in class org.apache.spark.rdd.PairRDDFunctions
Merge the values for each key using an associative function and a neutral "zero value" which may be added to the result an arbitrary number of times, and must not change the result (e.g., Nil for list concatenation, 0 for addition, or 1 for multiplication.).
foldByKey(V, int, Function2<V, V, V>) - Method in class org.apache.spark.rdd.PairRDDFunctions
Merge the values for each key using an associative function and a neutral "zero value" which may be added to the result an arbitrary number of times, and must not change the result (e.g., Nil for list concatenation, 0 for addition, or 1 for multiplication.).
foldByKey(V, Function2<V, V, V>) - Method in class org.apache.spark.rdd.PairRDDFunctions
Merge the values for each key using an associative function and a neutral "zero value" which may be added to the result an arbitrary number of times, and must not change the result (e.g., Nil for list concatenation, 0 for addition, or 1 for multiplication.).
foldCol() - Method in class org.apache.spark.ml.tuning.CrossValidator
 
foldCol() - Method in class org.apache.spark.ml.tuning.CrossValidatorModel
 
foldCol() - Method in interface org.apache.spark.ml.tuning.CrossValidatorParams
Param for the column name of user specified fold number.
forall(Column, Function1<Column, Column>) - Static method in class org.apache.spark.sql.functions
Returns whether a predicate holds for every element in the array.
forceIndexLabel() - Method in class org.apache.spark.ml.feature.RFormula
 
forceIndexLabel() - Method in interface org.apache.spark.ml.feature.RFormulaBase
Force to index label whether it is numeric or string type.
forceIndexLabel() - Method in class org.apache.spark.ml.feature.RFormulaModel
 
foreach(VoidFunction<T>) - Method in interface org.apache.spark.api.java.JavaRDDLike
Applies a function f to all elements of this RDD.
foreach(Function2<Object, Object, BoxedUnit>) - Method in interface org.apache.spark.ml.linalg.Vector
Applies a function f to all the elements of dense and sparse vector.
foreach(Function2<Object, Object, BoxedUnit>) - Method in interface org.apache.spark.mllib.linalg.Vector
Applies a function f to all the elements of dense and sparse vector.
foreach(Function1<T, BoxedUnit>) - Method in class org.apache.spark.rdd.RDD
Applies a function f to all elements of this RDD.
foreach(Function1<T, BoxedUnit>) - Method in class org.apache.spark.sql.Dataset
Applies a function f to all rows.
foreach(ForeachFunction<T>) - Method in class org.apache.spark.sql.Dataset
(Java-specific) Runs func on each element of this Dataset.
foreach(ForeachWriter<T>) - Method in class org.apache.spark.sql.streaming.DataStreamWriter
Sets the output of the streaming query to be processed using the provided writer object.
foreach(KVStoreView<T>, Function1<T, BoxedUnit>) - Static method in class org.apache.spark.status.KVUtils
Applies a function f to all values produced by KVStoreView.
foreachActive(Function3<Object, Object, Object, BoxedUnit>) - Method in class org.apache.spark.ml.linalg.DenseMatrix
 
foreachActive(Function3<Object, Object, Object, BoxedUnit>) - Method in interface org.apache.spark.ml.linalg.Matrix
Applies a function f to all the active elements of dense and sparse matrix.
foreachActive(Function3<Object, Object, Object, BoxedUnit>) - Method in class org.apache.spark.ml.linalg.SparseMatrix
 
foreachActive(Function2<Object, Object, BoxedUnit>) - Method in interface org.apache.spark.ml.linalg.Vector
Applies a function f to all the active elements of dense and sparse vector.
foreachActive(Function3<Object, Object, Object, BoxedUnit>) - Method in interface org.apache.spark.mllib.linalg.Matrix
Applies a function f to all the active elements of dense and sparse matrix.
foreachActive(Function2<Object, Object, BoxedUnit>) - Method in interface org.apache.spark.mllib.linalg.Vector
Applies a function f to all the active elements of dense and sparse vector.
foreachAsync(VoidFunction<T>) - Method in interface org.apache.spark.api.java.JavaRDDLike
The asynchronous version of the foreach action, which applies a function f to all the elements of this RDD.
foreachAsync(Function1<T, BoxedUnit>) - Method in class org.apache.spark.rdd.AsyncRDDActions
Applies a function f to all elements of this RDD.
foreachBatch(Function2<Dataset<T>, Object, BoxedUnit>) - Method in class org.apache.spark.sql.streaming.DataStreamWriter
:: Experimental ::
foreachBatch(VoidFunction2<Dataset<T>, Long>) - Method in class org.apache.spark.sql.streaming.DataStreamWriter
:: Experimental ::
ForeachFunction<T> - Interface in org.apache.spark.api.java.function
Base interface for a function used in Dataset's foreach function.
foreachNonZero(Function2<Object, Object, BoxedUnit>) - Method in interface org.apache.spark.ml.linalg.Vector
Applies a function f to all the non-zero elements of dense and sparse vector.
foreachNonZero(Function2<Object, Object, BoxedUnit>) - Method in interface org.apache.spark.mllib.linalg.Vector
Applies a function f to all the non-zero elements of dense and sparse vector.
foreachPartition(VoidFunction<Iterator<T>>) - Method in interface org.apache.spark.api.java.JavaRDDLike
Applies a function f to each partition of this RDD.
foreachPartition(Function1<Iterator<T>, BoxedUnit>) - Method in class org.apache.spark.rdd.RDD
Applies a function f to each partition of this RDD.
foreachPartition(Function1<Iterator<T>, BoxedUnit>) - Method in class org.apache.spark.sql.Dataset
Applies a function f to each partition of this Dataset.
foreachPartition(ForeachPartitionFunction<T>) - Method in class org.apache.spark.sql.Dataset
(Java-specific) Runs func on each partition of this Dataset.
foreachPartitionAsync(VoidFunction<Iterator<T>>) - Method in interface org.apache.spark.api.java.JavaRDDLike
The asynchronous version of the foreachPartition action, which applies a function f to each partition of this RDD.
foreachPartitionAsync(Function1<Iterator<T>, BoxedUnit>) - Method in class org.apache.spark.rdd.AsyncRDDActions
Applies a function f to each partition of this RDD.
ForeachPartitionFunction<T> - Interface in org.apache.spark.api.java.function
Base interface for a function used in Dataset's foreachPartition function.
foreachRDD(VoidFunction<R>) - Method in interface org.apache.spark.streaming.api.java.JavaDStreamLike
Apply a function to each RDD in this DStream.
foreachRDD(VoidFunction2<R, Time>) - Method in interface org.apache.spark.streaming.api.java.JavaDStreamLike
Apply a function to each RDD in this DStream.
foreachRDD(Function1<RDD<T>, BoxedUnit>) - Method in class org.apache.spark.streaming.dstream.DStream
Apply a function to each RDD in this DStream.
foreachRDD(Function2<RDD<T>, Time, BoxedUnit>) - Method in class org.apache.spark.streaming.dstream.DStream
Apply a function to each RDD in this DStream.
ForeachWriter<T> - Class in org.apache.spark.sql
The abstract class for writing custom logic to process data generated by a query.
ForeachWriter() - Constructor for class org.apache.spark.sql.ForeachWriter
 
foreachWriterAbortedDueToTaskFailureError() - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
format() - Method in class org.apache.spark.ml.clustering.InternalKMeansModelWriter
 
format() - Method in class org.apache.spark.ml.clustering.PMMLKMeansModelWriter
 
format() - Method in class org.apache.spark.ml.regression.InternalLinearRegressionModelWriter
 
format() - Method in class org.apache.spark.ml.regression.PMMLLinearRegressionModelWriter
 
format(String) - Method in class org.apache.spark.ml.util.GeneralMLWriter
Specifies the format of ML export (e.g.
format() - Method in interface org.apache.spark.ml.util.MLFormatRegister
The string that represents the format that this format provider uses.
format(String) - Method in class org.apache.spark.sql.DataFrameReader
Specifies the input data source format.
format(String) - Method in class org.apache.spark.sql.DataFrameWriter
Specifies the underlying output data source.
format(String) - Method in class org.apache.spark.sql.streaming.DataStreamReader
Specifies the input data source format.
format(String) - Method in class org.apache.spark.sql.streaming.DataStreamWriter
Specifies the underlying output data source.
format_number(Column, int) - Static method in class org.apache.spark.sql.functions
Formats numeric column x to a format like '#,###,###.##', rounded to d decimal places with HALF_EVEN round mode, and returns the result as a string column.
format_string(String, Column...) - Static method in class org.apache.spark.sql.functions
Formats the arguments in printf-style and returns the result as a string column.
format_string(String, Seq<Column>) - Static method in class org.apache.spark.sql.functions
Formats the arguments in printf-style and returns the result as a string column.
formatBatchTime(long, long, boolean, TimeZone) - Static method in class org.apache.spark.ui.UIUtils
If batchInterval is less than 1 second, format batchTime with milliseconds.
formatDate(Date) - Static method in class org.apache.spark.ui.UIUtils
 
formatDate(long) - Static method in class org.apache.spark.ui.UIUtils
 
formatDuration(long) - Static method in class org.apache.spark.ui.UIUtils
 
formatDurationVerbose(long) - Static method in class org.apache.spark.ui.UIUtils
Generate a verbose human-readable string representing a duration such as "5 second 35 ms"
formatNumber(double) - Static method in class org.apache.spark.ui.UIUtils
Generate a human-readable string representing a number (e.g.
formula() - Method in class org.apache.spark.ml.feature.RFormula
 
formula() - Method in interface org.apache.spark.ml.feature.RFormulaBase
R formula parameter.
formula() - Method in class org.apache.spark.ml.feature.RFormulaModel
 
forNumber(int) - Static method in enum org.apache.spark.status.protobuf.StoreTypes.DeterministicLevel
 
forNumber(int) - Static method in enum org.apache.spark.status.protobuf.StoreTypes.JobExecutionStatus
 
forNumber(int) - Static method in enum org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper.WrapperCase
 
forNumber(int) - Static method in enum org.apache.spark.status.protobuf.StoreTypes.StageStatus
 
forward(DenseMatrix<Object>, boolean) - Method in interface org.apache.spark.ml.ann.TopologyModel
Forward propagation
foundDifferentWindowFunctionTypeError(Seq<NamedExpression>) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
foundDuplicateFieldInCaseInsensitiveModeError(String, String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
foundDuplicateFieldInFieldIdLookupModeError(int, String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
foundNullValueForNotNullableFieldError(String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
foundRecursionInProtobufSchema(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
FPGA() - Static method in class org.apache.spark.resource.ResourceUtils
 
FPGrowth - Class in org.apache.spark.ml.fpm
A parallel FP-growth algorithm to mine frequent itemsets.
FPGrowth(String) - Constructor for class org.apache.spark.ml.fpm.FPGrowth
 
FPGrowth() - Constructor for class org.apache.spark.ml.fpm.FPGrowth
 
FPGrowth - Class in org.apache.spark.mllib.fpm
A parallel FP-growth algorithm to mine frequent itemsets.
FPGrowth() - Constructor for class org.apache.spark.mllib.fpm.FPGrowth
Constructs a default instance with default parameters {minSupport: 0.3, numPartitions: same as the input data}.
FPGrowth.FreqItemset<Item> - Class in org.apache.spark.mllib.fpm
Frequent itemset.
FPGrowthModel - Class in org.apache.spark.ml.fpm
Model fitted by FPGrowth.
FPGrowthModel<Item> - Class in org.apache.spark.mllib.fpm
Model trained by FPGrowth, which holds frequent itemsets.
FPGrowthModel(RDD<FPGrowth.FreqItemset<Item>>, Map<Item, Object>, ClassTag<Item>) - Constructor for class org.apache.spark.mllib.fpm.FPGrowthModel
 
FPGrowthModel(RDD<FPGrowth.FreqItemset<Item>>, ClassTag<Item>) - Constructor for class org.apache.spark.mllib.fpm.FPGrowthModel
 
FPGrowthModel.SaveLoadV1_0$ - Class in org.apache.spark.mllib.fpm
 
FPGrowthParams - Interface in org.apache.spark.ml.fpm
Common params for FPGrowth and FPGrowthModel
fpr() - Method in interface org.apache.spark.ml.feature.SelectorParams
The highest p-value for features to be kept.
fpr() - Method in class org.apache.spark.mllib.feature.ChiSqSelector
 
FRACTION_CACHED() - Static method in class org.apache.spark.ui.storage.ToolTips
 
fragment() - Method in interface org.apache.spark.QueryContext
 
freq() - Method in class org.apache.spark.mllib.fpm.FPGrowth.FreqItemset
 
freq() - Method in class org.apache.spark.mllib.fpm.PrefixSpan.FreqSequence
 
freqItems(String[], double) - Method in class org.apache.spark.sql.DataFrameStatFunctions
Finding frequent items for columns, possibly with false positives.
freqItems(String[]) - Method in class org.apache.spark.sql.DataFrameStatFunctions
Finding frequent items for columns, possibly with false positives.
freqItems(Seq<String>, double) - Method in class org.apache.spark.sql.DataFrameStatFunctions
(Scala-specific) Finding frequent items for columns, possibly with false positives.
freqItems(Seq<String>) - Method in class org.apache.spark.sql.DataFrameStatFunctions
(Scala-specific) Finding frequent items for columns, possibly with false positives.
FreqItemset(Object, long) - Constructor for class org.apache.spark.mllib.fpm.FPGrowth.FreqItemset
 
freqItemsets() - Method in class org.apache.spark.ml.fpm.FPGrowthModel
 
freqItemsets() - Method in class org.apache.spark.mllib.fpm.FPGrowthModel
 
FreqSequence(Object[], long) - Constructor for class org.apache.spark.mllib.fpm.PrefixSpan.FreqSequence
 
freqSequences() - Method in class org.apache.spark.mllib.fpm.PrefixSpanModel
 
from_avro(Column, String) - Static method in class org.apache.spark.sql.avro.functions
Converts a binary column of avro format into its corresponding catalyst value.
from_avro(Column, String, Map<String, String>) - Static method in class org.apache.spark.sql.avro.functions
Converts a binary column of Avro format into its corresponding catalyst value.
from_csv(Column, StructType, Map<String, String>) - Static method in class org.apache.spark.sql.functions
Parses a column containing a CSV string into a StructType with the specified schema.
from_csv(Column, Column, Map<String, String>) - Static method in class org.apache.spark.sql.functions
(Java-specific) Parses a column containing a CSV string into a StructType with the specified schema.
FROM_ID_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationEdge
 
FROM_ID_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphEdge
 
from_json(Column, StructType, Map<String, String>) - Static method in class org.apache.spark.sql.functions
(Scala-specific) Parses a column containing a JSON string into a StructType with the specified schema.
from_json(Column, DataType, Map<String, String>) - Static method in class org.apache.spark.sql.functions
(Scala-specific) Parses a column containing a JSON string into a MapType with StringType as keys type, StructType or ArrayType with the specified schema.
from_json(Column, StructType, Map<String, String>) - Static method in class org.apache.spark.sql.functions
(Java-specific) Parses a column containing a JSON string into a StructType with the specified schema.
from_json(Column, DataType, Map<String, String>) - Static method in class org.apache.spark.sql.functions
(Java-specific) Parses a column containing a JSON string into a MapType with StringType as keys type, StructType or ArrayType with the specified schema.
from_json(Column, StructType) - Static method in class org.apache.spark.sql.functions
Parses a column containing a JSON string into a StructType with the specified schema.
from_json(Column, DataType) - Static method in class org.apache.spark.sql.functions
Parses a column containing a JSON string into a MapType with StringType as keys type, StructType or ArrayType with the specified schema.
from_json(Column, String, Map<String, String>) - Static method in class org.apache.spark.sql.functions
(Java-specific) Parses a column containing a JSON string into a MapType with StringType as keys type, StructType or ArrayType with the specified schema.
from_json(Column, String, Map<String, String>) - Static method in class org.apache.spark.sql.functions
(Scala-specific) Parses a column containing a JSON string into a MapType with StringType as keys type, StructType or ArrayType with the specified schema.
from_json(Column, Column) - Static method in class org.apache.spark.sql.functions
(Scala-specific) Parses a column containing a JSON string into a MapType with StringType as keys type, StructType or ArrayType of StructTypes with the specified schema.
from_json(Column, Column, Map<String, String>) - Static method in class org.apache.spark.sql.functions
(Java-specific) Parses a column containing a JSON string into a MapType with StringType as keys type, StructType or ArrayType of StructTypes with the specified schema.
from_unixtime(Column) - Static method in class org.apache.spark.sql.functions
Converts the number of seconds from unix epoch (1970-01-01 00:00:00 UTC) to a string representing the timestamp of that moment in the current system time zone in the yyyy-MM-dd HH:mm:ss format.
from_unixtime(Column, String) - Static method in class org.apache.spark.sql.functions
Converts the number of seconds from unix epoch (1970-01-01 00:00:00 UTC) to a string representing the timestamp of that moment in the current system time zone in the given format.
from_utc_timestamp(Column, String) - Static method in class org.apache.spark.sql.functions
Given a timestamp like '2017-07-14 02:40:00.0', interprets it as a time in UTC, and renders that time as a timestamp in the given time zone.
from_utc_timestamp(Column, Column) - Static method in class org.apache.spark.sql.functions
Given a timestamp like '2017-07-14 02:40:00.0', interprets it as a time in UTC, and renders that time as a timestamp in the given time zone.
fromArrowField(Field) - Static method in class org.apache.spark.sql.util.ArrowUtils
 
fromArrowSchema(Schema) - Static method in class org.apache.spark.sql.util.ArrowUtils
 
fromArrowType(ArrowType) - Static method in class org.apache.spark.sql.util.ArrowUtils
 
fromCOO(int, int, Iterable<Tuple3<Object, Object, Object>>) - Static method in class org.apache.spark.ml.linalg.SparseMatrix
Generate a SparseMatrix from Coordinate List (COO) format.
fromCOO(int, int, Iterable<Tuple3<Object, Object, Object>>) - Static method in class org.apache.spark.mllib.linalg.SparseMatrix
Generate a SparseMatrix from Coordinate List (COO) format.
fromDDL(String) - Static method in class org.apache.spark.sql.types.DataType
 
fromDDL(String) - Static method in class org.apache.spark.sql.types.StructType
Creates StructType for a given DDL-formatted string, which is a comma separated list of field definitions, e.g., a INT, b STRING.
fromDecimal(Object) - Static method in class org.apache.spark.sql.types.Decimal
 
fromDStream(DStream<T>, ClassTag<T>) - Static method in class org.apache.spark.streaming.api.java.JavaDStream
Convert a scala DStream to a Java-friendly JavaDStream.
fromEdgePartitions(RDD<Tuple2<Object, EdgePartition<ED, VD>>>, VD, StorageLevel, StorageLevel, ClassTag<VD>, ClassTag<ED>) - Static method in class org.apache.spark.graphx.impl.GraphImpl
Create a graph from EdgePartitions, setting referenced vertices to defaultVertexAttr.
fromEdges(RDD<Edge<ED>>, ClassTag<ED>, ClassTag<VD>) - Static method in class org.apache.spark.graphx.EdgeRDD
Creates an EdgeRDD from a set of edges.
fromEdges(RDD<Edge<ED>>, VD, StorageLevel, StorageLevel, ClassTag<VD>, ClassTag<ED>) - Static method in class org.apache.spark.graphx.Graph
Construct a graph from a collection of edges.
fromEdges(EdgeRDD<?>, int, VD, ClassTag<VD>) - Static method in class org.apache.spark.graphx.VertexRDD
Constructs a VertexRDD containing all vertices referred to in edges.
fromEdgeTuples(RDD<Tuple2<Object, Object>>, VD, Option<PartitionStrategy>, StorageLevel, StorageLevel, ClassTag<VD>) - Static method in class org.apache.spark.graphx.Graph
Construct a graph from a collection of edges encoded as vertex id pairs.
fromExistingRDDs(VertexRDD<VD>, EdgeRDD<ED>, ClassTag<VD>, ClassTag<ED>) - Static method in class org.apache.spark.graphx.impl.GraphImpl
Create a graph from a VertexRDD and an EdgeRDD with the same replicated vertex type as the vertices.
fromFileStatus(FileStatus) - Static method in class org.apache.spark.paths.SparkPath
 
fromInputDStream(InputDStream<T>, ClassTag<T>) - Static method in class org.apache.spark.streaming.api.java.JavaInputDStream
Convert a scala InputDStream to a Java-friendly JavaInputDStream.
fromInputDStream(InputDStream<Tuple2<K, V>>, ClassTag<K>, ClassTag<V>) - Static method in class org.apache.spark.streaming.api.java.JavaPairInputDStream
Convert a scala InputDStream of pairs to a Java-friendly JavaPairInputDStream.
fromInt(int) - Static method in class org.apache.spark.sql.types.ByteExactNumeric
 
fromInt(int) - Method in interface org.apache.spark.sql.types.Decimal.DecimalIsConflicted
 
fromInt(int) - Static method in class org.apache.spark.sql.types.DecimalExactNumeric
 
fromInt(int) - Static method in class org.apache.spark.sql.types.DoubleExactNumeric
 
fromInt(int) - Method in interface org.apache.spark.sql.types.DoubleType.DoubleIsConflicted
 
fromInt(int) - Static method in class org.apache.spark.sql.types.FloatExactNumeric
 
fromInt(int) - Method in interface org.apache.spark.sql.types.FloatType.FloatIsConflicted
 
fromInt(int) - Static method in class org.apache.spark.sql.types.IntegerExactNumeric
 
fromInt(int) - Static method in class org.apache.spark.sql.types.LongExactNumeric
 
fromInt(int) - Static method in class org.apache.spark.sql.types.ShortExactNumeric
 
fromJavaDStream(JavaDStream<Tuple2<K, V>>) - Static method in class org.apache.spark.streaming.api.java.JavaPairDStream
 
fromJavaRDD(JavaRDD<Tuple2<K, V>>) - Static method in class org.apache.spark.api.java.JavaPairRDD
Convert a JavaRDD of key-value pairs to JavaPairRDD.
fromJson(String) - Static method in class org.apache.spark.ml.linalg.JsonMatrixConverter
Parses the JSON representation of a Matrix into a Matrix.
fromJson(String) - Static method in class org.apache.spark.ml.linalg.JsonVectorConverter
Parses the JSON representation of a vector into a Vector.
fromJson(String) - Static method in class org.apache.spark.mllib.linalg.Vectors
Parses the JSON representation of a vector into a Vector.
fromJson(String) - Static method in class org.apache.spark.sql.types.DataType
 
fromJson(String) - Static method in class org.apache.spark.sql.types.Metadata
Creates a Metadata instance from JSON.
fromKinesisInitialPosition(InitialPositionInStream) - Static method in class org.apache.spark.streaming.kinesis.KinesisInitialPositions
Returns instance of [[KinesisInitialPosition]] based on the passed [[InitialPositionInStream]].
fromMetadata(Metadata) - Method in interface org.apache.spark.ml.attribute.AttributeFactory
Creates an Attribute from a Metadata instance.
fromML(DenseMatrix) - Static method in class org.apache.spark.mllib.linalg.DenseMatrix
Convert new linalg type to spark.mllib type.
fromML(DenseVector) - Static method in class org.apache.spark.mllib.linalg.DenseVector
Convert new linalg type to spark.mllib type.
fromML(Matrix) - Static method in class org.apache.spark.mllib.linalg.Matrices
Convert new linalg type to spark.mllib type.
fromML(SparseMatrix) - Static method in class org.apache.spark.mllib.linalg.SparseMatrix
Convert new linalg type to spark.mllib type.
fromML(SparseVector) - Static method in class org.apache.spark.mllib.linalg.SparseVector
Convert new linalg type to spark.mllib type.
fromML(Vector) - Static method in class org.apache.spark.mllib.linalg.Vectors
Convert new linalg type to spark.mllib type.
fromName(String) - Static method in class org.apache.spark.ml.attribute.AttributeType
Gets the AttributeType object from its name.
fromNullable(T) - Static method in class org.apache.spark.api.java.Optional
 
fromOld(Node, Map<Object, Object>) - Static method in class org.apache.spark.ml.tree.Node
Create a new Node from the old Node format, recursively creating child nodes as needed.
fromPairDStream(DStream<Tuple2<K, V>>, ClassTag<K>, ClassTag<V>) - Static method in class org.apache.spark.streaming.api.java.JavaPairDStream
 
fromPairRDD(RDD<Tuple2<K, V>>, ClassTag<K>, ClassTag<V>) - Static method in class org.apache.spark.mllib.rdd.MLPairRDDFunctions
Implicit conversion from a pair RDD to MLPairRDDFunctions.
fromParams(GeneralizedLinearRegressionBase) - Method in class org.apache.spark.ml.regression.GeneralizedLinearRegression.Family$
Gets the Family object based on param family and variancePower.
fromParams(GeneralizedLinearRegressionBase) - Method in class org.apache.spark.ml.regression.GeneralizedLinearRegression.Link$
Gets the Link object based on param family, link and linkPower.
fromPath(Path) - Static method in class org.apache.spark.paths.SparkPath
 
fromPathString(String) - Static method in class org.apache.spark.paths.SparkPath
Creates a SparkPath from a hadoop Path string.
fromRDD(RDD<Object>) - Static method in class org.apache.spark.api.java.JavaDoubleRDD
 
fromRDD(RDD<Tuple2<K, V>>, ClassTag<K>, ClassTag<V>) - Static method in class org.apache.spark.api.java.JavaPairRDD
 
fromRDD(RDD<T>, ClassTag<T>) - Static method in class org.apache.spark.api.java.JavaRDD
 
fromRDD(RDD<T>, ClassTag<T>) - Static method in class org.apache.spark.mllib.rdd.RDDFunctions
Implicit conversion from an RDD to RDDFunctions.
fromRdd(RDD<?>) - Static method in class org.apache.spark.storage.RDDInfo
 
fromReceiverInputDStream(ReceiverInputDStream<Tuple2<K, V>>, ClassTag<K>, ClassTag<V>) - Static method in class org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
Convert a scala ReceiverInputDStream to a Java-friendly JavaReceiverInputDStream.
fromReceiverInputDStream(ReceiverInputDStream<T>, ClassTag<T>) - Static method in class org.apache.spark.streaming.api.java.JavaReceiverInputDStream
Convert a scala ReceiverInputDStream to a Java-friendly JavaReceiverInputDStream.
fromSparkContext(SparkContext) - Static method in class org.apache.spark.api.java.JavaSparkContext
 
fromStage(Stage, int, Option<Object>, TaskMetrics, Seq<Seq<TaskLocation>>, int) - Static method in class org.apache.spark.scheduler.StageInfo
Construct a StageInfo from a Stage.
fromString(String) - Static method in enum org.apache.spark.JobExecutionStatus
 
fromString(String) - Static method in class org.apache.spark.mllib.tree.impurity.Impurities
 
fromString(String) - Static method in class org.apache.spark.mllib.tree.loss.Losses
 
fromString(UTF8String) - Static method in class org.apache.spark.sql.types.Decimal
 
fromString(String) - Static method in enum org.apache.spark.status.api.v1.ApplicationStatus
 
fromString(String) - Static method in enum org.apache.spark.status.api.v1.StageStatus
 
fromString(String) - Static method in enum org.apache.spark.status.api.v1.streaming.BatchStatus
 
fromString(String) - Static method in enum org.apache.spark.status.api.v1.TaskSorting
 
fromString(String) - Static method in enum org.apache.spark.status.api.v1.TaskStatus
 
fromString(String) - Static method in class org.apache.spark.storage.StorageLevel
:: DeveloperApi :: Return the StorageLevel object with the specified name.
fromStringANSI(UTF8String, DecimalType, SQLQueryContext) - Static method in class org.apache.spark.sql.types.Decimal
 
fromStructField(StructField) - Static method in class org.apache.spark.ml.attribute.Attribute
 
fromStructField(StructField) - Method in interface org.apache.spark.ml.attribute.AttributeFactory
Creates an Attribute from a StructField instance.
fromStructField(StructField) - Static method in class org.apache.spark.ml.attribute.AttributeGroup
Creates an attribute group from a StructField instance.
fromStructField(StructField) - Static method in class org.apache.spark.ml.attribute.BinaryAttribute
 
fromStructField(StructField) - Static method in class org.apache.spark.ml.attribute.NominalAttribute
 
fromStructField(StructField) - Static method in class org.apache.spark.ml.attribute.NumericAttribute
 
fromToIntervalUnsupportedError(String, String, ParserRuleContext) - Static method in class org.apache.spark.sql.errors.QueryParsingErrors
 
fromUri(URI) - Static method in class org.apache.spark.paths.SparkPath
 
fromUrlString(String) - Static method in class org.apache.spark.paths.SparkPath
Creates a SparkPath from a url-encoded string.
fullOuterJoin(JavaPairRDD<K, W>, Partitioner) - Method in class org.apache.spark.api.java.JavaPairRDD
Perform a full outer join of this and other.
fullOuterJoin(JavaPairRDD<K, W>) - Method in class org.apache.spark.api.java.JavaPairRDD
Perform a full outer join of this and other.
fullOuterJoin(JavaPairRDD<K, W>, int) - Method in class org.apache.spark.api.java.JavaPairRDD
Perform a full outer join of this and other.
fullOuterJoin(RDD<Tuple2<K, W>>, Partitioner) - Method in class org.apache.spark.rdd.PairRDDFunctions
Perform a full outer join of this and other.
fullOuterJoin(RDD<Tuple2<K, W>>) - Method in class org.apache.spark.rdd.PairRDDFunctions
Perform a full outer join of this and other.
fullOuterJoin(RDD<Tuple2<K, W>>, int) - Method in class org.apache.spark.rdd.PairRDDFunctions
Perform a full outer join of this and other.
fullOuterJoin(JavaPairDStream<K, W>) - Method in class org.apache.spark.streaming.api.java.JavaPairDStream
Return a new DStream by applying 'full outer join' between RDDs of this DStream and other DStream.
fullOuterJoin(JavaPairDStream<K, W>, int) - Method in class org.apache.spark.streaming.api.java.JavaPairDStream
Return a new DStream by applying 'full outer join' between RDDs of this DStream and other DStream.
fullOuterJoin(JavaPairDStream<K, W>, Partitioner) - Method in class org.apache.spark.streaming.api.java.JavaPairDStream
Return a new DStream by applying 'full outer join' between RDDs of this DStream and other DStream.
fullOuterJoin(DStream<Tuple2<K, W>>, ClassTag<W>) - Method in class org.apache.spark.streaming.dstream.PairDStreamFunctions
Return a new DStream by applying 'full outer join' between RDDs of this DStream and other DStream.
fullOuterJoin(DStream<Tuple2<K, W>>, int, ClassTag<W>) - Method in class org.apache.spark.streaming.dstream.PairDStreamFunctions
Return a new DStream by applying 'full outer join' between RDDs of this DStream and other DStream.
fullOuterJoin(DStream<Tuple2<K, W>>, Partitioner, ClassTag<W>) - Method in class org.apache.spark.streaming.dstream.PairDStreamFunctions
Return a new DStream by applying 'full outer join' between RDDs of this DStream and other DStream.
fullStackTrace() - Method in class org.apache.spark.ExceptionFailure
 
funcBuildError(String, Exception) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
Function<T1,R> - Interface in org.apache.spark.api.java.function
Base interface for functions whose return types do not create special RDDs.
Function - Class in org.apache.spark.sql.catalog
A user-defined function in Spark, as returned by listFunctions method in Catalog.
Function(String, String, String[], String, String, boolean) - Constructor for class org.apache.spark.sql.catalog.Function
 
Function(String, String, String, String, boolean) - Constructor for class org.apache.spark.sql.catalog.Function
 
Function - Interface in org.apache.spark.sql.connector.catalog.functions
Base class for user-defined functions.
function(Function4<Time, KeyType, Option<ValueType>, State<StateType>, Option<MappedType>>) - Static method in class org.apache.spark.streaming.StateSpec
Create a StateSpec for setting all the specifications of the mapWithState operation on a pair DStream.
function(Function3<KeyType, Option<ValueType>, State<StateType>, MappedType>) - Static method in class org.apache.spark.streaming.StateSpec
Create a StateSpec for setting all the specifications of the mapWithState operation on a pair DStream.
function(Function4<Time, KeyType, Optional<ValueType>, State<StateType>, Optional<MappedType>>) - Static method in class org.apache.spark.streaming.StateSpec
Create a StateSpec for setting all the specifications of the mapWithState operation on a JavaPairDStream.
function(Function3<KeyType, Optional<ValueType>, State<StateType>, MappedType>) - Static method in class org.apache.spark.streaming.StateSpec
Create a StateSpec for setting all the specifications of the mapWithState operation on a JavaPairDStream.
Function0<R> - Interface in org.apache.spark.api.java.function
A zero-argument function that returns an R.
Function2<T1,T2,R> - Interface in org.apache.spark.api.java.function
A two-argument function that takes arguments of type T1 and T2 and returns an R.
Function3<T1,T2,T3,R> - Interface in org.apache.spark.api.java.function
A three-argument function that takes arguments of type T1, T2 and T3 and returns an R.
Function4<T1,T2,T3,T4,R> - Interface in org.apache.spark.api.java.function
A four-argument function that takes arguments of type T1, T2, T3 and T4 and returns an R.
functionAlreadyExistsError(FunctionIdentifier) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
functionCannotProcessInputError(UnboundFunction, Seq<Expression>, UnsupportedOperationException) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
FunctionCatalog - Interface in org.apache.spark.sql.connector.catalog
Catalog methods for working with Functions.
functionExists(String) - Method in class org.apache.spark.sql.catalog.Catalog
Check if the function with the specified name exists.
functionExists(String, String) - Method in class org.apache.spark.sql.catalog.Catalog
Check if the function with the specified name exists in the specified database under the Hive Metastore.
functionExists(Identifier) - Method in class org.apache.spark.sql.connector.catalog.DelegatingCatalogExtension
 
functionExists(Identifier) - Method in interface org.apache.spark.sql.connector.catalog.FunctionCatalog
Returns true if the function exists, false otherwise.
FunctionIdentifierHelper(FunctionIdentifier) - Constructor for class org.apache.spark.sql.connector.catalog.CatalogV2Implicits.FunctionIdentifierHelper
 
functionNameUnsupportedError(String, ParserRuleContext) - Static method in class org.apache.spark.sql.errors.QueryParsingErrors
 
functions - Class in org.apache.spark.ml
 
functions() - Constructor for class org.apache.spark.ml.functions
 
functions - Class in org.apache.spark.sql.avro
 
functions() - Constructor for class org.apache.spark.sql.avro.functions
 
functions - Class in org.apache.spark.sql
Commonly used functions available for DataFrame operations.
functions() - Constructor for class org.apache.spark.sql.functions
 
functions() - Static method in class org.apache.spark.sql.jdbc.DB2Dialect
 
functions() - Static method in class org.apache.spark.sql.jdbc.DerbyDialect
 
functions() - Static method in class org.apache.spark.sql.jdbc.H2Dialect
 
functions() - Method in class org.apache.spark.sql.jdbc.JdbcDialect
List the user-defined functions in jdbc dialect.
functions() - Static method in class org.apache.spark.sql.jdbc.MsSqlServerDialect
 
functions() - Static method in class org.apache.spark.sql.jdbc.MySQLDialect
 
functions() - Static method in class org.apache.spark.sql.jdbc.NoopDialect
 
functions() - Static method in class org.apache.spark.sql.jdbc.OracleDialect
 
functions() - Static method in class org.apache.spark.sql.jdbc.PostgresDialect
 
functions() - Static method in class org.apache.spark.sql.jdbc.TeradataDialect
 
functionWithUnsupportedSyntaxError(String, String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
FutureAction<T> - Interface in org.apache.spark
A future for the result of an action to support cancellation.
futureExecutionContext() - Static method in class org.apache.spark.rdd.AsyncRDDActions
 
FValueTest - Class in org.apache.spark.ml.stat
FValue test for continuous data.
FValueTest() - Constructor for class org.apache.spark.ml.stat.FValueTest
 
fwe() - Method in interface org.apache.spark.ml.feature.SelectorParams
The upper bound of the expected family-wise error rate.
fwe() - Method in class org.apache.spark.mllib.feature.ChiSqSelector
 

G

gain() - Method in class org.apache.spark.ml.tree.DecisionTreeModelReadWrite.NodeData
 
gain() - Method in class org.apache.spark.ml.tree.InternalNode
 
gain() - Method in class org.apache.spark.mllib.tree.model.InformationGainStats
 
Gamma$() - Constructor for class org.apache.spark.ml.regression.GeneralizedLinearRegression.Gamma$
 
gamma1() - Method in class org.apache.spark.graphx.lib.SVDPlusPlus.Conf
 
gamma2() - Method in class org.apache.spark.graphx.lib.SVDPlusPlus.Conf
 
gamma6() - Method in class org.apache.spark.graphx.lib.SVDPlusPlus.Conf
 
gamma7() - Method in class org.apache.spark.graphx.lib.SVDPlusPlus.Conf
 
GammaGenerator - Class in org.apache.spark.mllib.random
Generates i.i.d.
GammaGenerator(double, double) - Constructor for class org.apache.spark.mllib.random.GammaGenerator
 
gammaJavaRDD(JavaSparkContext, double, double, long, int, long) - Static method in class org.apache.spark.mllib.random.RandomRDDs
Java-friendly version of RandomRDDs.gammaRDD.
gammaJavaRDD(JavaSparkContext, double, double, long, int) - Static method in class org.apache.spark.mllib.random.RandomRDDs
RandomRDDs.gammaJavaRDD with the default seed.
gammaJavaRDD(JavaSparkContext, double, double, long) - Static method in class org.apache.spark.mllib.random.RandomRDDs
RandomRDDs.gammaJavaRDD with the default number of partitions and the default seed.
gammaJavaVectorRDD(JavaSparkContext, double, double, long, int, int, long) - Static method in class org.apache.spark.mllib.random.RandomRDDs
Java-friendly version of RandomRDDs.gammaVectorRDD.
gammaJavaVectorRDD(JavaSparkContext, double, double, long, int, int) - Static method in class org.apache.spark.mllib.random.RandomRDDs
RandomRDDs.gammaJavaVectorRDD with the default seed.
gammaJavaVectorRDD(JavaSparkContext, double, double, long, int) - Static method in class org.apache.spark.mllib.random.RandomRDDs
RandomRDDs.gammaJavaVectorRDD with the default number of partitions and the default seed.
gammaRDD(SparkContext, double, double, long, int, long) - Static method in class org.apache.spark.mllib.random.RandomRDDs
Generates an RDD comprised of i.i.d. samples from the gamma distribution with the input shape and scale.
gammaVectorRDD(SparkContext, double, double, long, int, int, long) - Static method in class org.apache.spark.mllib.random.RandomRDDs
Generates an RDD[Vector] with vectors containing i.i.d. samples drawn from the gamma distribution with the input shape and scale.
gapply(RelationalGroupedDataset, byte[], byte[], Object[], StructType) - Static method in class org.apache.spark.sql.api.r.SQLUtils
The helper function for gapply() on R side.
gaps() - Method in class org.apache.spark.ml.feature.RegexTokenizer
Indicates whether regex splits on gaps (true) or matches tokens (false).
GarbageCollectionMetrics - Class in org.apache.spark.metrics
 
GarbageCollectionMetrics() - Constructor for class org.apache.spark.metrics.GarbageCollectionMetrics
 
GAUGE() - Static method in class org.apache.spark.metrics.sink.StatsdMetricType
 
Gaussian$() - Constructor for class org.apache.spark.ml.regression.GeneralizedLinearRegression.Gaussian$
 
GaussianMixture - Class in org.apache.spark.ml.clustering
Gaussian Mixture clustering.
GaussianMixture(String) - Constructor for class org.apache.spark.ml.clustering.GaussianMixture
 
GaussianMixture() - Constructor for class org.apache.spark.ml.clustering.GaussianMixture
 
GaussianMixture - Class in org.apache.spark.mllib.clustering
This class performs expectation maximization for multivariate Gaussian Mixture Models (GMMs).
GaussianMixture() - Constructor for class org.apache.spark.mllib.clustering.GaussianMixture
Constructs a default instance.
GaussianMixtureModel - Class in org.apache.spark.ml.clustering
Multivariate Gaussian Mixture Model (GMM) consisting of k Gaussians, where points are drawn from each Gaussian i with probability weights(i).
GaussianMixtureModel - Class in org.apache.spark.mllib.clustering
Multivariate Gaussian Mixture Model (GMM) consisting of k Gaussians, where points are drawn from each Gaussian i=1..k with probability w(i); mu(i) and sigma(i) are the respective mean and covariance for each Gaussian distribution i=1..k.
GaussianMixtureModel(double[], MultivariateGaussian[]) - Constructor for class org.apache.spark.mllib.clustering.GaussianMixtureModel
 
GaussianMixtureParams - Interface in org.apache.spark.ml.clustering
Common params for GaussianMixture and GaussianMixtureModel
GaussianMixtureSummary - Class in org.apache.spark.ml.clustering
Summary of GaussianMixture.
gaussians() - Method in class org.apache.spark.ml.clustering.GaussianMixtureModel
 
gaussians() - Method in class org.apache.spark.mllib.clustering.GaussianMixtureModel
 
gaussiansDF() - Method in class org.apache.spark.ml.clustering.GaussianMixtureModel
Retrieve Gaussian distributions as a DataFrame.
GBTClassificationModel - Class in org.apache.spark.ml.classification
Gradient-Boosted Trees (GBTs) (http://en.wikipedia.org/wiki/Gradient_boosting) model for classification.
GBTClassificationModel(String, DecisionTreeRegressionModel[], double[]) - Constructor for class org.apache.spark.ml.classification.GBTClassificationModel
Construct a GBTClassificationModel
GBTClassifier - Class in org.apache.spark.ml.classification
Gradient-Boosted Trees (GBTs) (http://en.wikipedia.org/wiki/Gradient_boosting) learning algorithm for classification.
GBTClassifier(String) - Constructor for class org.apache.spark.ml.classification.GBTClassifier
 
GBTClassifier() - Constructor for class org.apache.spark.ml.classification.GBTClassifier
 
GBTClassifierParams - Interface in org.apache.spark.ml.tree
 
GBTParams - Interface in org.apache.spark.ml.tree
Parameters for Gradient-Boosted Tree algorithms.
GBTRegressionModel - Class in org.apache.spark.ml.regression
Gradient-Boosted Trees (GBTs) model for regression.
GBTRegressionModel(String, DecisionTreeRegressionModel[], double[]) - Constructor for class org.apache.spark.ml.regression.GBTRegressionModel
Construct a GBTRegressionModel
GBTRegressor - Class in org.apache.spark.ml.regression
Gradient-Boosted Trees (GBTs) learning algorithm for regression.
GBTRegressor(String) - Constructor for class org.apache.spark.ml.regression.GBTRegressor
 
GBTRegressor() - Constructor for class org.apache.spark.ml.regression.GBTRegressor
 
GBTRegressorParams - Interface in org.apache.spark.ml.tree
 
GC_TIME() - Static method in class org.apache.spark.status.TaskIndexNames
 
GC_TIME() - Static method in class org.apache.spark.ui.ToolTips
 
gemm(double, Matrix, DenseMatrix, double, DenseMatrix) - Static method in class org.apache.spark.ml.linalg.BLAS
C := alpha * A * B + beta * C
gemm(double, Matrix, DenseMatrix, double, double[]) - Static method in class org.apache.spark.ml.linalg.BLAS
CValues[0: A.numRows * B.numCols] := alpha * A * B + beta * CValues[0: A.numRows * B.numCols]
gemm(double, Matrix, DenseMatrix, double, DenseMatrix) - Static method in class org.apache.spark.mllib.linalg.BLAS
C := alpha * A * B + beta * C
gemv(double, Matrix, double[], double, double[]) - Static method in class org.apache.spark.ml.linalg.BLAS
y[0: A.numRows] := alpha * A * x[0: A.numCols] + beta * y[0: A.numRows]
gemv(double, Matrix, Vector, double, DenseVector) - Static method in class org.apache.spark.ml.linalg.BLAS
y := alpha * A * x + beta * y
gemv(double, Matrix, Vector, double, DenseVector) - Static method in class org.apache.spark.mllib.linalg.BLAS
y := alpha * A * x + beta * y
GeneralAggregateFunc - Class in org.apache.spark.sql.connector.expressions.aggregate
The general implementation of AggregateFunc, which contains the upper-cased function name, the `isDistinct` flag and all the inputs.
GeneralAggregateFunc(String, boolean, Expression[]) - Constructor for class org.apache.spark.sql.connector.expressions.aggregate.GeneralAggregateFunc
 
GeneralizedLinearAlgorithm<M extends GeneralizedLinearModel> - Class in org.apache.spark.mllib.regression
GeneralizedLinearAlgorithm implements methods to train a Generalized Linear Model (GLM).
GeneralizedLinearAlgorithm() - Constructor for class org.apache.spark.mllib.regression.GeneralizedLinearAlgorithm
 
GeneralizedLinearModel - Class in org.apache.spark.mllib.regression
GeneralizedLinearModel (GLM) represents a model trained using GeneralizedLinearAlgorithm.
GeneralizedLinearModel(Vector, double) - Constructor for class org.apache.spark.mllib.regression.GeneralizedLinearModel
 
GeneralizedLinearRegression - Class in org.apache.spark.ml.regression
Fit a Generalized Linear Model (see Generalized linear model (Wikipedia)) specified by giving a symbolic description of the linear predictor (link function) and a description of the error distribution (family).
GeneralizedLinearRegression(String) - Constructor for class org.apache.spark.ml.regression.GeneralizedLinearRegression
 
GeneralizedLinearRegression() - Constructor for class org.apache.spark.ml.regression.GeneralizedLinearRegression
 
GeneralizedLinearRegression.Binomial$ - Class in org.apache.spark.ml.regression
Binomial exponential family distribution.
GeneralizedLinearRegression.CLogLog$ - Class in org.apache.spark.ml.regression
 
GeneralizedLinearRegression.Family$ - Class in org.apache.spark.ml.regression
 
GeneralizedLinearRegression.FamilyAndLink$ - Class in org.apache.spark.ml.regression
 
GeneralizedLinearRegression.Gamma$ - Class in org.apache.spark.ml.regression
Gamma exponential family distribution.
GeneralizedLinearRegression.Gaussian$ - Class in org.apache.spark.ml.regression
Gaussian exponential family distribution.
GeneralizedLinearRegression.Identity$ - Class in org.apache.spark.ml.regression
 
GeneralizedLinearRegression.Inverse$ - Class in org.apache.spark.ml.regression
 
GeneralizedLinearRegression.Link$ - Class in org.apache.spark.ml.regression
 
GeneralizedLinearRegression.Log$ - Class in org.apache.spark.ml.regression
 
GeneralizedLinearRegression.Logit$ - Class in org.apache.spark.ml.regression
 
GeneralizedLinearRegression.Poisson$ - Class in org.apache.spark.ml.regression
Poisson exponential family distribution.
GeneralizedLinearRegression.Probit$ - Class in org.apache.spark.ml.regression
 
GeneralizedLinearRegression.Sqrt$ - Class in org.apache.spark.ml.regression
 
GeneralizedLinearRegression.Tweedie$ - Class in org.apache.spark.ml.regression
 
GeneralizedLinearRegressionBase - Interface in org.apache.spark.ml.regression
Params for Generalized Linear Regression.
GeneralizedLinearRegressionModel - Class in org.apache.spark.ml.regression
Model produced by GeneralizedLinearRegression.
GeneralizedLinearRegressionSummary - Class in org.apache.spark.ml.regression
Summary of GeneralizedLinearRegression model and predictions.
GeneralizedLinearRegressionTrainingSummary - Class in org.apache.spark.ml.regression
Summary of GeneralizedLinearRegression fitting and model.
GeneralMLWritable - Interface in org.apache.spark.ml.util
Trait for classes that provide GeneralMLWriter.
GeneralMLWriter - Class in org.apache.spark.ml.util
A ML Writer which delegates based on the requested format.
GeneralMLWriter(PipelineStage) - Constructor for class org.apache.spark.ml.util.GeneralMLWriter
 
GeneralScalarExpression - Class in org.apache.spark.sql.connector.expressions
The general representation of SQL scalar expressions, which contains the upper-cased expression name and all the children expressions.
GeneralScalarExpression(String, Expression[]) - Constructor for class org.apache.spark.sql.connector.expressions.GeneralScalarExpression
 
generateAssociationRules(double) - Method in class org.apache.spark.mllib.fpm.FPGrowthModel
Generates association rules for the Items in freqItemsets.
generateKMeansRDD(SparkContext, int, int, int, double, int) - Static method in class org.apache.spark.mllib.util.KMeansDataGenerator
Generate an RDD containing test data for KMeans.
generateLinearInput(double, double[], int, int, double) - Static method in class org.apache.spark.mllib.util.LinearDataGenerator
For compatibility, the generated data without specifying the mean and variance will have zero mean and variance of (1.0/3.0) since the original output range is [-1, 1] with uniform distribution, and the variance of uniform distribution is (b - a)^2^ / 12 which will be (1.0/3.0)
generateLinearInput(double, double[], double[], double[], int, int, double) - Static method in class org.apache.spark.mllib.util.LinearDataGenerator
 
generateLinearInput(double, double[], double[], double[], int, int, double, double) - Static method in class org.apache.spark.mllib.util.LinearDataGenerator
 
generateLinearInputAsList(double, double[], int, int, double) - Static method in class org.apache.spark.mllib.util.LinearDataGenerator
Return a Java List of synthetic data randomly generated according to a multi collinear model.
generateLinearRDD(SparkContext, int, int, double, int, double) - Static method in class org.apache.spark.mllib.util.LinearDataGenerator
Generate an RDD containing sample data for Linear Regression models - including Ridge, Lasso, and unregularized variants.
generateLogisticRDD(SparkContext, int, int, double, int, double) - Static method in class org.apache.spark.mllib.util.LogisticRegressionDataGenerator
Generate an RDD containing test data for LogisticRegression.
generateRandomEdges(int, int, int, long) - Static method in class org.apache.spark.graphx.util.GraphGenerators
 
generateRolledOverFileSuffix() - Method in interface org.apache.spark.util.logging.RollingPolicy
Get the desired name of the rollover file
generationExpression() - Method in interface org.apache.spark.sql.connector.catalog.Column
Returns the generation expression of this table column.
generatorNotExpectedError(FunctionIdentifier, String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
generatorOutsideSelectError(LogicalPlan) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
geq(Object) - Method in class org.apache.spark.sql.Column
Greater than or equal to an expression.
get(Object) - Method in class org.apache.spark.api.java.JavaUtils.SerializableMapWrapper
 
get() - Method in class org.apache.spark.api.java.Optional
 
get() - Static method in class org.apache.spark.BarrierTaskContext
:: Experimental :: Returns the currently active BarrierTaskContext.
get() - Method in interface org.apache.spark.FutureAction
Blocks and returns the result of this job.
get(Param<T>) - Method in class org.apache.spark.ml.param.ParamMap
Optionally returns the value associated with a param.
get(Param<T>) - Method in interface org.apache.spark.ml.param.Params
Optionally returns the user-supplied value of a param.
get(String) - Method in class org.apache.spark.SparkConf
Get a parameter; throws a NoSuchElementException if it's not set
get(String, String) - Method in class org.apache.spark.SparkConf
Get a parameter, falling back to a default if not set
get() - Static method in class org.apache.spark.SparkEnv
Returns the SparkEnv.
get(String) - Static method in class org.apache.spark.SparkFiles
Get the absolute path of a file added through SparkContext.addFile().
get() - Method in interface org.apache.spark.sql.connector.read.PartitionReader
Return the current record.
get(Column, Column) - Static method in class org.apache.spark.sql.functions
Returns element of array at given (0-based) index.
get(String) - Static method in class org.apache.spark.sql.jdbc.JdbcDialects
Fetch the JdbcDialect class corresponding to a given database url.
get() - Method in class org.apache.spark.sql.Observation
(Scala-specific) Get the observed metrics.
get(int) - Method in interface org.apache.spark.sql.Row
Returns the value at position i.
get(String) - Method in class org.apache.spark.sql.RuntimeConfig
Returns the value of Spark runtime configuration property for the given key.
get(String, String) - Method in class org.apache.spark.sql.RuntimeConfig
Returns the value of Spark runtime configuration property for the given key.
get() - Method in interface org.apache.spark.sql.streaming.GroupState
Get the state value if it exists, or throw NoSuchElementException.
get(UUID) - Method in class org.apache.spark.sql.streaming.StreamingQueryManager
Returns the query if there is an active query with the given id, or null.
get(String) - Method in class org.apache.spark.sql.streaming.StreamingQueryManager
Returns the query if there is an active query with the given id, or null.
get(Object) - Method in class org.apache.spark.sql.util.CaseInsensitiveStringMap
 
get(int, DataType) - Method in class org.apache.spark.sql.vectorized.ColumnarArray
 
get(int, DataType) - Method in class org.apache.spark.sql.vectorized.ColumnarBatchRow
 
get(int, DataType) - Method in class org.apache.spark.sql.vectorized.ColumnarRow
 
get() - Method in class org.apache.spark.streaming.State
Get the state if it exists, otherwise it will throw java.util.NoSuchElementException.
get() - Static method in class org.apache.spark.TaskContext
Return the currently active TaskContext.
get(long) - Static method in class org.apache.spark.util.AccumulatorContext
Returns the AccumulatorV2 registered with the given ID, if any.
get_json_object(Column, String) - Static method in class org.apache.spark.sql.functions
Extracts json object from a json string based on json path specified, and returns json string of the extracted json object.
getAbsolutePathFromExecutable(String) - Static method in class org.apache.spark.TestUtils
Get the absolute path from the executable.
getAcceptanceResults(RDD<Tuple2<K, V>>, boolean, Map<K, Object>, Option<Map<K, Object>>, long) - Static method in class org.apache.spark.util.random.StratifiedSamplingUtils
Count the number of items instantly accepted and generate the waitlist for each stratum.
getAccumulatorId() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetric.Builder
int64 accumulator_id = 2;
getAccumulatorId() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetric
int64 accumulator_id = 2;
getAccumulatorId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetricOrBuilder
int64 accumulator_id = 2;
getAccumulatorUpdates(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 44;
getAccumulatorUpdates(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 44;
getAccumulatorUpdates(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 44;
getAccumulatorUpdates(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
getAccumulatorUpdates(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
getAccumulatorUpdates(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataOrBuilder
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
getAccumulatorUpdates(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
getAccumulatorUpdates(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
getAccumulatorUpdates(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
getAccumulatorUpdatesBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 44;
getAccumulatorUpdatesBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
getAccumulatorUpdatesBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
getAccumulatorUpdatesBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 44;
getAccumulatorUpdatesBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
getAccumulatorUpdatesBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
getAccumulatorUpdatesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 44;
getAccumulatorUpdatesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 44;
getAccumulatorUpdatesCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 44;
getAccumulatorUpdatesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
getAccumulatorUpdatesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
getAccumulatorUpdatesCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataOrBuilder
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
getAccumulatorUpdatesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
getAccumulatorUpdatesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
getAccumulatorUpdatesCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
getAccumulatorUpdatesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 44;
getAccumulatorUpdatesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 44;
getAccumulatorUpdatesList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 44;
getAccumulatorUpdatesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
getAccumulatorUpdatesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
getAccumulatorUpdatesList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataOrBuilder
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
getAccumulatorUpdatesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
getAccumulatorUpdatesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
getAccumulatorUpdatesList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
getAccumulatorUpdatesOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 44;
getAccumulatorUpdatesOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 44;
getAccumulatorUpdatesOrBuilder(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 44;
getAccumulatorUpdatesOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
getAccumulatorUpdatesOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
getAccumulatorUpdatesOrBuilder(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataOrBuilder
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
getAccumulatorUpdatesOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
getAccumulatorUpdatesOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
getAccumulatorUpdatesOrBuilder(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
getAccumulatorUpdatesOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 44;
getAccumulatorUpdatesOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 44;
getAccumulatorUpdatesOrBuilderList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 44;
getAccumulatorUpdatesOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
getAccumulatorUpdatesOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
getAccumulatorUpdatesOrBuilderList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataOrBuilder
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
getAccumulatorUpdatesOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
getAccumulatorUpdatesOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
getAccumulatorUpdatesOrBuilderList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
getActive() - Static method in class org.apache.spark.streaming.StreamingContext
Deprecated.
Get the currently active context, if there is one.
getActiveJobIds() - Method in class org.apache.spark.api.java.JavaSparkStatusTracker
Returns an array containing the ids of all active jobs.
getActiveJobIds() - Method in class org.apache.spark.SparkStatusTracker
Returns an array containing the ids of all active jobs.
getActiveOrCreate(Function0<StreamingContext>) - Static method in class org.apache.spark.streaming.StreamingContext
Deprecated.
Either return the "active" StreamingContext (that is, started but not stopped), or create a new StreamingContext that is
getActiveOrCreate(String, Function0<StreamingContext>, Configuration, boolean) - Static method in class org.apache.spark.streaming.StreamingContext
Deprecated.
Either get the currently active StreamingContext (that is, started but not stopped), OR recreate a StreamingContext from checkpoint data in the given path.
getActiveSession() - Static method in class org.apache.spark.sql.SparkSession
Returns the active SparkSession for the current thread, returned by the builder.
getActiveStageIds() - Method in class org.apache.spark.api.java.JavaSparkStatusTracker
Returns an array containing the ids of all active stages.
getActiveStageIds() - Method in class org.apache.spark.SparkStatusTracker
Returns an array containing the ids of all active stages.
getActiveTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
int32 active_tasks = 9;
getActiveTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
int32 active_tasks = 9;
getActiveTasks() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
int32 active_tasks = 9;
getAddColumnQuery(String, String, String) - Static method in class org.apache.spark.sql.jdbc.DB2Dialect
 
getAddColumnQuery(String, String, String) - Static method in class org.apache.spark.sql.jdbc.DerbyDialect
 
getAddColumnQuery(String, String, String) - Static method in class org.apache.spark.sql.jdbc.H2Dialect
 
getAddColumnQuery(String, String, String) - Method in class org.apache.spark.sql.jdbc.JdbcDialect
 
getAddColumnQuery(String, String, String) - Static method in class org.apache.spark.sql.jdbc.MsSqlServerDialect
 
getAddColumnQuery(String, String, String) - Static method in class org.apache.spark.sql.jdbc.MySQLDialect
 
getAddColumnQuery(String, String, String) - Static method in class org.apache.spark.sql.jdbc.NoopDialect
 
getAddColumnQuery(String, String, String) - Static method in class org.apache.spark.sql.jdbc.OracleDialect
 
getAddColumnQuery(String, String, String) - Static method in class org.apache.spark.sql.jdbc.PostgresDialect
 
getAddColumnQuery(String, String, String) - Static method in class org.apache.spark.sql.jdbc.TeradataDialect
 
getAddress() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution.Builder
optional string address = 1;
getAddress() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution
optional string address = 1;
getAddress() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDDataDistributionOrBuilder
optional string address = 1;
getAddressBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution.Builder
optional string address = 1;
getAddressBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution
optional string address = 1;
getAddressBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDDataDistributionOrBuilder
optional string address = 1;
getAddresses(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceInformation.Builder
repeated string addresses = 2;
getAddresses(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceInformation
repeated string addresses = 2;
getAddresses(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ResourceInformationOrBuilder
repeated string addresses = 2;
getAddressesBytes(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceInformation.Builder
repeated string addresses = 2;
getAddressesBytes(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceInformation
repeated string addresses = 2;
getAddressesBytes(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ResourceInformationOrBuilder
repeated string addresses = 2;
getAddressesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceInformation.Builder
repeated string addresses = 2;
getAddressesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceInformation
repeated string addresses = 2;
getAddressesCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ResourceInformationOrBuilder
repeated string addresses = 2;
getAddressesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceInformation.Builder
repeated string addresses = 2;
getAddressesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceInformation
repeated string addresses = 2;
getAddressesList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ResourceInformationOrBuilder
repeated string addresses = 2;
getAddTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
int64 add_time = 20;
getAddTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
int64 add_time = 20;
getAddTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
int64 add_time = 20;
getAddTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder
int64 add_time = 5;
getAddTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary
int64 add_time = 5;
getAddTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryOrBuilder
int64 add_time = 5;
getAggregationDepth() - Method in interface org.apache.spark.ml.param.shared.HasAggregationDepth
 
getAlgo() - Method in class org.apache.spark.mllib.tree.configuration.Strategy
 
getAll() - Method in class org.apache.spark.SparkConf
Get all parameters as a list of pairs
getAll() - Method in class org.apache.spark.sql.RuntimeConfig
Returns all properties set in this conf.
getAllClusterConfigs(SparkConf) - Static method in class org.apache.spark.kafka010.KafkaTokenSparkConf
 
getAllConfs() - Method in class org.apache.spark.sql.SQLContext
Return all the configuration properties that have been set (i.e.
getAllPools() - Method in class org.apache.spark.SparkContext
:: DeveloperApi :: Return pools for fair scheduler
GetAllReceiverInfo - Class in org.apache.spark.streaming.scheduler
 
GetAllReceiverInfo() - Constructor for class org.apache.spark.streaming.scheduler.GetAllReceiverInfo
 
getAllRemovalsTimeMs() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
int64 all_removals_time_ms = 6;
getAllRemovalsTimeMs() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress
int64 all_removals_time_ms = 6;
getAllRemovalsTimeMs() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgressOrBuilder
int64 all_removals_time_ms = 6;
getAllUpdatesTimeMs() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
int64 all_updates_time_ms = 4;
getAllUpdatesTimeMs() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress
int64 all_updates_time_ms = 4;
getAllUpdatesTimeMs() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgressOrBuilder
int64 all_updates_time_ms = 4;
getAllWithPrefix(String) - Method in class org.apache.spark.SparkConf
Get all parameters that start with prefix
getAlpha() - Method in interface org.apache.spark.ml.recommendation.ALSParams
 
getAlpha() - Method in class org.apache.spark.mllib.clustering.LDA
Alias for getDocConcentration
getAmount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest.Builder
int64 amount = 2;
getAmount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest
int64 amount = 2;
getAmount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequestOrBuilder
int64 amount = 2;
getAmount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskResourceRequest.Builder
double amount = 2;
getAmount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskResourceRequest
double amount = 2;
getAmount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskResourceRequestOrBuilder
double amount = 2;
getAnyValAs(int) - Method in interface org.apache.spark.sql.Row
Returns the value at position i.
getAppId() - Method in interface org.apache.spark.launcher.SparkAppHandle
Returns the application ID, or null if not yet known.
getAppId() - Method in class org.apache.spark.SparkConf
Returns the Spark application id, valid in the Driver after TaskScheduler registration and from the start in the Executor.
getApplicationInfo(String) - Method in interface org.apache.spark.status.api.v1.UIRoot
 
getApplicationInfoList() - Method in interface org.apache.spark.status.api.v1.UIRoot
 
getAppSparkVersion() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo.Builder
optional string app_spark_version = 8;
getAppSparkVersion() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo
optional string app_spark_version = 8;
getAppSparkVersion() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfoOrBuilder
optional string app_spark_version = 8;
getAppSparkVersionBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo.Builder
optional string app_spark_version = 8;
getAppSparkVersionBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo
optional string app_spark_version = 8;
getAppSparkVersionBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfoOrBuilder
optional string app_spark_version = 8;
getArray(int) - Method in class org.apache.spark.sql.vectorized.ArrowColumnVector
 
getArray(int) - Method in class org.apache.spark.sql.vectorized.ColumnarArray
 
getArray(int) - Method in class org.apache.spark.sql.vectorized.ColumnarBatchRow
 
getArray(int) - Method in class org.apache.spark.sql.vectorized.ColumnarRow
 
getArray(int) - Method in class org.apache.spark.sql.vectorized.ColumnVector
Returns the array type value for rowId.
getAs(int) - Method in interface org.apache.spark.sql.Row
Returns the value at position i.
getAs(String) - Method in interface org.apache.spark.sql.Row
Returns the value of a given fieldName.
getAsJava() - Method in class org.apache.spark.sql.Observation
(Java-specific) Get the observed metrics.
getAssociationRulesFromFP(Dataset<?>, String, String, double, Map<T, Object>, long, ClassTag<T>) - Static method in class org.apache.spark.ml.fpm.AssociationRules
Computes the association rules with confidence above minConfidence.
getAsymmetricAlpha() - Method in class org.apache.spark.mllib.clustering.LDA
Alias for getAsymmetricDocConcentration
getAsymmetricDocConcentration() - Method in class org.apache.spark.mllib.clustering.LDA
Concentration parameter (commonly named "alpha") for the prior placed on documents' distributions over topics ("theta").
getAttempt() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
int32 attempt = 3;
getAttempt() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
int32 attempt = 3;
getAttempt() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataOrBuilder
int32 attempt = 3;
getAttempt() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int32 attempt = 3;
getAttempt() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
int32 attempt = 3;
getAttempt() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
int32 attempt = 3;
getAttemptId() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo.Builder
optional string attempt_id = 1;
getAttemptId() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo
optional string attempt_id = 1;
getAttemptId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfoOrBuilder
optional string attempt_id = 1;
getAttemptId() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int32 attempt_id = 3;
getAttemptId() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
int32 attempt_id = 3;
getAttemptId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
int32 attempt_id = 3;
getAttemptIdBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo.Builder
optional string attempt_id = 1;
getAttemptIdBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo
optional string attempt_id = 1;
getAttemptIdBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfoOrBuilder
optional string attempt_id = 1;
getAttempts(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
repeated .org.apache.spark.status.protobuf.ApplicationAttemptInfo attempts = 7;
getAttempts(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo
repeated .org.apache.spark.status.protobuf.ApplicationAttemptInfo attempts = 7;
getAttempts(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoOrBuilder
repeated .org.apache.spark.status.protobuf.ApplicationAttemptInfo attempts = 7;
getAttemptsBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
repeated .org.apache.spark.status.protobuf.ApplicationAttemptInfo attempts = 7;
getAttemptsBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
repeated .org.apache.spark.status.protobuf.ApplicationAttemptInfo attempts = 7;
getAttemptsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
repeated .org.apache.spark.status.protobuf.ApplicationAttemptInfo attempts = 7;
getAttemptsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo
repeated .org.apache.spark.status.protobuf.ApplicationAttemptInfo attempts = 7;
getAttemptsCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoOrBuilder
repeated .org.apache.spark.status.protobuf.ApplicationAttemptInfo attempts = 7;
getAttemptsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
repeated .org.apache.spark.status.protobuf.ApplicationAttemptInfo attempts = 7;
getAttemptsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo
repeated .org.apache.spark.status.protobuf.ApplicationAttemptInfo attempts = 7;
getAttemptsList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoOrBuilder
repeated .org.apache.spark.status.protobuf.ApplicationAttemptInfo attempts = 7;
getAttemptsOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
repeated .org.apache.spark.status.protobuf.ApplicationAttemptInfo attempts = 7;
getAttemptsOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo
repeated .org.apache.spark.status.protobuf.ApplicationAttemptInfo attempts = 7;
getAttemptsOrBuilder(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoOrBuilder
repeated .org.apache.spark.status.protobuf.ApplicationAttemptInfo attempts = 7;
getAttemptsOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
repeated .org.apache.spark.status.protobuf.ApplicationAttemptInfo attempts = 7;
getAttemptsOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo
repeated .org.apache.spark.status.protobuf.ApplicationAttemptInfo attempts = 7;
getAttemptsOrBuilderList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoOrBuilder
repeated .org.apache.spark.status.protobuf.ApplicationAttemptInfo attempts = 7;
getAttr(String) - Method in class org.apache.spark.ml.attribute.AttributeGroup
Gets an attribute by its name.
getAttr(int) - Method in class org.apache.spark.ml.attribute.AttributeGroup
Gets an attribute by its index.
getAttributes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
Deprecated.
getAttributes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
Deprecated.
getAttributes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
Deprecated.
getAttributesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
 
getAttributesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
 
getAttributesCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
map<string, string> attributes = 27;
getAttributesMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
map<string, string> attributes = 27;
getAttributesMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
map<string, string> attributes = 27;
getAttributesMap() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
map<string, string> attributes = 27;
getAttributesOrDefault(String, String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
map<string, string> attributes = 27;
getAttributesOrDefault(String, String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
map<string, string> attributes = 27;
getAttributesOrDefault(String, String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
map<string, string> attributes = 27;
getAttributesOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
map<string, string> attributes = 27;
getAttributesOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
map<string, string> attributes = 27;
getAttributesOrThrow(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
map<string, string> attributes = 27;
getAvroField(String, int) - Method in class org.apache.spark.sql.avro.AvroUtils.AvroSchemaHelper
Get the Avro field corresponding to the provided Catalyst field name/position, if any.
getAvroSchema() - Method in class org.apache.spark.SparkConf
Gets all the avro schemas in the configuration used in the generic Avro record serializer
getBarrier() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode.Builder
bool barrier = 4;
getBarrier() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode
bool barrier = 4;
getBarrier() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationNodeOrBuilder
bool barrier = 4;
getBatchDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
int64 batch_duration = 6;
getBatchDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
int64 batch_duration = 6;
getBatchDuration() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
int64 batch_duration = 6;
getBatchId() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
int64 batch_id = 5;
getBatchId() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
int64 batch_id = 5;
getBatchId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
int64 batch_id = 5;
getBatchingTimeout(SparkConf) - Static method in class org.apache.spark.streaming.util.WriteAheadLogUtils
How long we will wait for the wrappedLog in the BatchedWriteAheadLog to write the records before we fail the write attempt to unblock receivers.
getBernoulliSamplingFunction(RDD<Tuple2<K, V>>, Map<K, Object>, boolean, long) - Static method in class org.apache.spark.util.random.StratifiedSamplingUtils
Return the per partition sampling function used for sampling without replacement.
getBeta() - Method in class org.apache.spark.ml.evaluation.MulticlassClassificationEvaluator
 
getBeta() - Method in class org.apache.spark.mllib.clustering.LDA
Alias for getTopicConcentration
getBin(int) - Method in class org.apache.spark.sql.util.NumericHistogram
Returns a particular histogram bin.
getBinary() - Method in interface org.apache.spark.ml.feature.CountVectorizerParams
 
getBinary() - Method in class org.apache.spark.ml.feature.HashingTF
 
getBinary(int) - Method in class org.apache.spark.sql.vectorized.ArrowColumnVector
 
getBinary(int) - Method in class org.apache.spark.sql.vectorized.ColumnarArray
 
getBinary(int) - Method in class org.apache.spark.sql.vectorized.ColumnarBatchRow
 
getBinary(int) - Method in class org.apache.spark.sql.vectorized.ColumnarRow
 
getBinary(int) - Method in class org.apache.spark.sql.vectorized.ColumnVector
Returns the binary type value for rowId.
getBlacklistedInStages(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
repeated int64 blacklisted_in_stages = 25;
getBlacklistedInStages(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
repeated int64 blacklisted_in_stages = 25;
getBlacklistedInStages(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
repeated int64 blacklisted_in_stages = 25;
getBlacklistedInStagesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
repeated int64 blacklisted_in_stages = 25;
getBlacklistedInStagesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
repeated int64 blacklisted_in_stages = 25;
getBlacklistedInStagesCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
repeated int64 blacklisted_in_stages = 25;
getBlacklistedInStagesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
repeated int64 blacklisted_in_stages = 25;
getBlacklistedInStagesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
repeated int64 blacklisted_in_stages = 25;
getBlacklistedInStagesList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
repeated int64 blacklisted_in_stages = 25;
getBlockName() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo.Builder
optional string block_name = 1;
getBlockName() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo
optional string block_name = 1;
getBlockName() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfoOrBuilder
optional string block_name = 1;
getBlockNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo.Builder
optional string block_name = 1;
getBlockNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo
optional string block_name = 1;
getBlockNameBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfoOrBuilder
optional string block_name = 1;
getBlockSize() - Method in interface org.apache.spark.ml.param.shared.HasBlockSize
 
GetBlockStatus(BlockId, boolean) - Constructor for class org.apache.spark.storage.BlockManagerMessages.GetBlockStatus
 
GetBlockStatus$() - Constructor for class org.apache.spark.storage.BlockManagerMessages.GetBlockStatus$
 
getBoolean(String, boolean) - Method in class org.apache.spark.SparkConf
Get a parameter as a boolean, falling back to a default if not set
getBoolean(int) - Method in interface org.apache.spark.sql.Row
Returns the value at position i as a primitive boolean.
getBoolean(String) - Method in class org.apache.spark.sql.types.Metadata
Gets a Boolean.
getBoolean(String, boolean) - Method in class org.apache.spark.sql.util.CaseInsensitiveStringMap
Returns the boolean value to which the specified key is mapped, or defaultValue if there is no mapping for the key.
getBoolean(int) - Method in class org.apache.spark.sql.vectorized.ArrowColumnVector
 
getBoolean(int) - Method in class org.apache.spark.sql.vectorized.ColumnarArray
 
getBoolean(int) - Method in class org.apache.spark.sql.vectorized.ColumnarBatchRow
 
getBoolean(int) - Method in class org.apache.spark.sql.vectorized.ColumnarRow
 
getBoolean(int) - Method in class org.apache.spark.sql.vectorized.ColumnVector
Returns the boolean type value for rowId.
getBooleanArray(String) - Method in class org.apache.spark.sql.types.Metadata
Gets a Boolean array.
getBooleans(int, int) - Method in class org.apache.spark.sql.vectorized.ColumnVector
Gets boolean type values from [rowId, rowId + count).
getBootstrap() - Method in interface org.apache.spark.ml.tree.RandomForestParams
 
getBucketLength() - Method in interface org.apache.spark.ml.feature.BucketedRandomProjectionLSHParams
 
getBuilder() - Method in class org.apache.spark.storage.memory.DeserializedValuesHolder
 
getBuilder() - Method in class org.apache.spark.storage.memory.SerializedValuesHolder
 
getBuilder() - Method in interface org.apache.spark.storage.memory.ValuesHolder
Note: After this method is called, the ValuesHolder is invalid, we can't store data and get estimate size again.
getByte(int) - Method in interface org.apache.spark.sql.Row
Returns the value at position i as a primitive byte.
getByte(int) - Method in class org.apache.spark.sql.vectorized.ArrowColumnVector
 
getByte(int) - Method in class org.apache.spark.sql.vectorized.ColumnarArray
 
getByte(int) - Method in class org.apache.spark.sql.vectorized.ColumnarBatchRow
 
getByte(int) - Method in class org.apache.spark.sql.vectorized.ColumnarRow
 
getByte(int) - Method in class org.apache.spark.sql.vectorized.ColumnVector
Returns the byte type value for rowId.
getBytes(int, int) - Method in class org.apache.spark.sql.vectorized.ColumnVector
Gets byte type values from [rowId, rowId + count).
getBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double bytes_read = 18;
getBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
double bytes_read = 18;
getBytesRead() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
double bytes_read = 18;
getBytesRead(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributions.Builder
repeated double bytes_read = 1;
getBytesRead(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributions
repeated double bytes_read = 1;
getBytesRead(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributionsOrBuilder
repeated double bytes_read = 1;
getBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetrics.Builder
int64 bytes_read = 1;
getBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetrics
int64 bytes_read = 1;
getBytesRead() - Method in interface org.apache.spark.status.protobuf.StoreTypes.InputMetricsOrBuilder
int64 bytes_read = 1;
getBytesReadCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributions.Builder
repeated double bytes_read = 1;
getBytesReadCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributions
repeated double bytes_read = 1;
getBytesReadCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributionsOrBuilder
repeated double bytes_read = 1;
getBytesReadList() - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributions.Builder
repeated double bytes_read = 1;
getBytesReadList() - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributions
repeated double bytes_read = 1;
getBytesReadList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributionsOrBuilder
repeated double bytes_read = 1;
getBytesWritten() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double bytes_written = 20;
getBytesWritten() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
double bytes_written = 20;
getBytesWritten() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
double bytes_written = 20;
getBytesWritten(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributions.Builder
repeated double bytes_written = 1;
getBytesWritten(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributions
repeated double bytes_written = 1;
getBytesWritten(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributionsOrBuilder
repeated double bytes_written = 1;
getBytesWritten() - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetrics.Builder
int64 bytes_written = 1;
getBytesWritten() - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetrics
int64 bytes_written = 1;
getBytesWritten() - Method in interface org.apache.spark.status.protobuf.StoreTypes.OutputMetricsOrBuilder
int64 bytes_written = 1;
getBytesWritten() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetrics.Builder
int64 bytes_written = 1;
getBytesWritten() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetrics
int64 bytes_written = 1;
getBytesWritten() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricsOrBuilder
int64 bytes_written = 1;
getBytesWrittenCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributions.Builder
repeated double bytes_written = 1;
getBytesWrittenCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributions
repeated double bytes_written = 1;
getBytesWrittenCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributionsOrBuilder
repeated double bytes_written = 1;
getBytesWrittenList() - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributions.Builder
repeated double bytes_written = 1;
getBytesWrittenList() - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributions
repeated double bytes_written = 1;
getBytesWrittenList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributionsOrBuilder
repeated double bytes_written = 1;
getCached() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode.Builder
bool cached = 3;
getCached() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode
bool cached = 3;
getCached() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationNodeOrBuilder
bool cached = 3;
getCachedBlockManagerId(BlockManagerId) - Static method in class org.apache.spark.storage.BlockManagerId
 
getCachedMetadata(String) - Static method in class org.apache.spark.rdd.HadoopRDD
The three methods below are helpers for accessing the local map, a property of the SparkEnv of the local process.
getCacheNodeIds() - Method in interface org.apache.spark.ml.tree.DecisionTreeParams
 
getCallsite() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode.Builder
optional string callsite = 5;
getCallsite() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode
optional string callsite = 5;
getCallsite() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationNodeOrBuilder
optional string callsite = 5;
getCallSite(Function1<String, Object>) - Static method in class org.apache.spark.util.Utils
When called inside a class in the spark package, returns the name of the user code class (outside the spark package) that called into Spark, as well as which Spark method they called.
getCallsiteBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode.Builder
optional string callsite = 5;
getCallsiteBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode
optional string callsite = 5;
getCallsiteBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationNodeOrBuilder
optional string callsite = 5;
getCaseSensitive() - Method in class org.apache.spark.ml.feature.StopWordsRemover
 
getCatalystType(int, String, int, MetadataBuilder) - Method in class org.apache.spark.sql.jdbc.AggregatedDialect
 
getCatalystType(int, String, int, MetadataBuilder) - Static method in class org.apache.spark.sql.jdbc.DB2Dialect
 
getCatalystType(int, String, int, MetadataBuilder) - Static method in class org.apache.spark.sql.jdbc.DerbyDialect
 
getCatalystType(int, String, int, MetadataBuilder) - Static method in class org.apache.spark.sql.jdbc.H2Dialect
 
getCatalystType(int, String, int, MetadataBuilder) - Method in class org.apache.spark.sql.jdbc.JdbcDialect
Get the custom datatype mapping for the given jdbc meta information.
getCatalystType(int, String, int, MetadataBuilder) - Static method in class org.apache.spark.sql.jdbc.MsSqlServerDialect
 
getCatalystType(int, String, int, MetadataBuilder) - Static method in class org.apache.spark.sql.jdbc.MySQLDialect
 
getCatalystType(int, String, int, MetadataBuilder) - Static method in class org.apache.spark.sql.jdbc.NoopDialect
 
getCatalystType(int, String, int, MetadataBuilder) - Static method in class org.apache.spark.sql.jdbc.OracleDialect
 
getCatalystType(int, String, int, MetadataBuilder) - Static method in class org.apache.spark.sql.jdbc.PostgresDialect
 
getCatalystType(int, String, int, MetadataBuilder) - Static method in class org.apache.spark.sql.jdbc.TeradataDialect
 
getCategoricalCols() - Method in class org.apache.spark.ml.feature.FeatureHasher
 
getCategoricalFeatures(StructField) - Static method in class org.apache.spark.ml.util.MetadataUtils
Examine a schema to identify categorical (Binary and Nominal) features.
getCategoricalFeaturesInfo() - Method in class org.apache.spark.mllib.tree.configuration.Strategy
 
getCensorCol() - Method in interface org.apache.spark.ml.regression.AFTSurvivalRegressionParams
 
getCheckpointDir() - Method in class org.apache.spark.api.java.JavaSparkContext
 
getCheckpointDir() - Method in class org.apache.spark.SparkContext
 
getCheckpointFile() - Method in interface org.apache.spark.api.java.JavaRDDLike
Gets the name of the file to which this RDD was checkpointed
getCheckpointFile() - Method in class org.apache.spark.graphx.impl.EdgeRDDImpl
 
getCheckpointFile() - Method in class org.apache.spark.graphx.impl.VertexRDDImpl
 
getCheckpointFile() - Method in class org.apache.spark.rdd.RDD
Gets the name of the directory to which this RDD was checkpointed.
getCheckpointFiles() - Method in class org.apache.spark.graphx.Graph
Gets the name of the files to which this Graph was checkpointed.
getCheckpointFiles() - Method in class org.apache.spark.graphx.impl.GraphImpl
 
getCheckpointFiles() - Method in class org.apache.spark.ml.clustering.DistributedLDAModel
If using checkpointing and LDA.keepLastCheckpoint is set to true, then there may be saved checkpoint files.
getCheckpointInterval() - Method in interface org.apache.spark.ml.param.shared.HasCheckpointInterval
 
getCheckpointInterval() - Method in class org.apache.spark.mllib.clustering.LDA
Period (in iterations) between checkpoints.
getCheckpointInterval() - Method in class org.apache.spark.mllib.tree.configuration.Strategy
 
getChild(int) - Method in class org.apache.spark.sql.vectorized.ArrowColumnVector
 
getChild(int) - Method in class org.apache.spark.sql.vectorized.ColumnVector
 
getChildClusters(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationClusterWrapper child_clusters = 4;
getChildClusters(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper
repeated .org.apache.spark.status.protobuf.RDDOperationClusterWrapper child_clusters = 4;
getChildClusters(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapperOrBuilder
repeated .org.apache.spark.status.protobuf.RDDOperationClusterWrapper child_clusters = 4;
getChildClustersBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationClusterWrapper child_clusters = 4;
getChildClustersBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationClusterWrapper child_clusters = 4;
getChildClustersCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationClusterWrapper child_clusters = 4;
getChildClustersCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper
repeated .org.apache.spark.status.protobuf.RDDOperationClusterWrapper child_clusters = 4;
getChildClustersCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapperOrBuilder
repeated .org.apache.spark.status.protobuf.RDDOperationClusterWrapper child_clusters = 4;
getChildClustersList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationClusterWrapper child_clusters = 4;
getChildClustersList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper
repeated .org.apache.spark.status.protobuf.RDDOperationClusterWrapper child_clusters = 4;
getChildClustersList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapperOrBuilder
repeated .org.apache.spark.status.protobuf.RDDOperationClusterWrapper child_clusters = 4;
getChildClustersOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationClusterWrapper child_clusters = 4;
getChildClustersOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper
repeated .org.apache.spark.status.protobuf.RDDOperationClusterWrapper child_clusters = 4;
getChildClustersOrBuilder(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapperOrBuilder
repeated .org.apache.spark.status.protobuf.RDDOperationClusterWrapper child_clusters = 4;
getChildClustersOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationClusterWrapper child_clusters = 4;
getChildClustersOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper
repeated .org.apache.spark.status.protobuf.RDDOperationClusterWrapper child_clusters = 4;
getChildClustersOrBuilderList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapperOrBuilder
repeated .org.apache.spark.status.protobuf.RDDOperationClusterWrapper child_clusters = 4;
getChildNodes(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationNode child_nodes = 3;
getChildNodes(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper
repeated .org.apache.spark.status.protobuf.RDDOperationNode child_nodes = 3;
getChildNodes(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapperOrBuilder
repeated .org.apache.spark.status.protobuf.RDDOperationNode child_nodes = 3;
getChildNodesBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationNode child_nodes = 3;
getChildNodesBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationNode child_nodes = 3;
getChildNodesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationNode child_nodes = 3;
getChildNodesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper
repeated .org.apache.spark.status.protobuf.RDDOperationNode child_nodes = 3;
getChildNodesCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapperOrBuilder
repeated .org.apache.spark.status.protobuf.RDDOperationNode child_nodes = 3;
getChildNodesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationNode child_nodes = 3;
getChildNodesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper
repeated .org.apache.spark.status.protobuf.RDDOperationNode child_nodes = 3;
getChildNodesList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapperOrBuilder
repeated .org.apache.spark.status.protobuf.RDDOperationNode child_nodes = 3;
getChildNodesOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationNode child_nodes = 3;
getChildNodesOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper
repeated .org.apache.spark.status.protobuf.RDDOperationNode child_nodes = 3;
getChildNodesOrBuilder(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapperOrBuilder
repeated .org.apache.spark.status.protobuf.RDDOperationNode child_nodes = 3;
getChildNodesOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationNode child_nodes = 3;
getChildNodesOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper
repeated .org.apache.spark.status.protobuf.RDDOperationNode child_nodes = 3;
getChildNodesOrBuilderList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapperOrBuilder
repeated .org.apache.spark.status.protobuf.RDDOperationNode child_nodes = 3;
getClassifier() - Method in interface org.apache.spark.ml.classification.OneVsRestParams
 
getClasspathEntries(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings classpath_entries = 6;
getClasspathEntries(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
repeated .org.apache.spark.status.protobuf.PairStrings classpath_entries = 6;
getClasspathEntries(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoOrBuilder
repeated .org.apache.spark.status.protobuf.PairStrings classpath_entries = 6;
getClasspathEntriesBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings classpath_entries = 6;
getClasspathEntriesBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings classpath_entries = 6;
getClasspathEntriesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings classpath_entries = 6;
getClasspathEntriesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
repeated .org.apache.spark.status.protobuf.PairStrings classpath_entries = 6;
getClasspathEntriesCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoOrBuilder
repeated .org.apache.spark.status.protobuf.PairStrings classpath_entries = 6;
getClasspathEntriesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings classpath_entries = 6;
getClasspathEntriesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
repeated .org.apache.spark.status.protobuf.PairStrings classpath_entries = 6;
getClasspathEntriesList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoOrBuilder
repeated .org.apache.spark.status.protobuf.PairStrings classpath_entries = 6;
getClasspathEntriesOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings classpath_entries = 6;
getClasspathEntriesOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
repeated .org.apache.spark.status.protobuf.PairStrings classpath_entries = 6;
getClasspathEntriesOrBuilder(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoOrBuilder
repeated .org.apache.spark.status.protobuf.PairStrings classpath_entries = 6;
getClasspathEntriesOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings classpath_entries = 6;
getClasspathEntriesOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
repeated .org.apache.spark.status.protobuf.PairStrings classpath_entries = 6;
getClasspathEntriesOrBuilderList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoOrBuilder
repeated .org.apache.spark.status.protobuf.PairStrings classpath_entries = 6;
getCluster() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper.Builder
.org.apache.spark.status.protobuf.SparkPlanGraphClusterWrapper cluster = 2;
getCluster() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper
.org.apache.spark.status.protobuf.SparkPlanGraphClusterWrapper cluster = 2;
getCluster() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapperOrBuilder
.org.apache.spark.status.protobuf.SparkPlanGraphClusterWrapper cluster = 2;
getClusterBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper.Builder
.org.apache.spark.status.protobuf.SparkPlanGraphClusterWrapper cluster = 2;
getClusterConfig(SparkConf, String) - Static method in class org.apache.spark.kafka010.KafkaTokenSparkConf
 
getClusterOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper.Builder
.org.apache.spark.status.protobuf.SparkPlanGraphClusterWrapper cluster = 2;
getClusterOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper
.org.apache.spark.status.protobuf.SparkPlanGraphClusterWrapper cluster = 2;
getClusterOrBuilder() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapperOrBuilder
.org.apache.spark.status.protobuf.SparkPlanGraphClusterWrapper cluster = 2;
getColdStartStrategy() - Method in interface org.apache.spark.ml.recommendation.ALSModelParams
 
getCollectSubModels() - Method in interface org.apache.spark.ml.param.shared.HasCollectSubModels
 
getColumnName(Seq<Object>, StructType) - Static method in class org.apache.spark.sql.util.SchemaUtils
Gets the name of the column in the given position.
getCombOp() - Static method in class org.apache.spark.util.random.StratifiedSamplingUtils
Returns the function used combine results returned by seqOp from different partitions.
getComment() - Method in class org.apache.spark.sql.types.StructField
Return the comment of this StructField.
getCommitTimeMs() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
int64 commit_time_ms = 7;
getCommitTimeMs() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress
int64 commit_time_ms = 7;
getCommitTimeMs() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgressOrBuilder
int64 commit_time_ms = 7;
getCompleted() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo.Builder
bool completed = 7;
getCompleted() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo
bool completed = 7;
getCompleted() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfoOrBuilder
bool completed = 7;
getCompletedTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
int32 completed_tasks = 11;
getCompletedTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
int32 completed_tasks = 11;
getCompletedTasks() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
int32 completed_tasks = 11;
getCompletionTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
optional int64 completion_time = 5;
getCompletionTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
optional int64 completion_time = 5;
getCompletionTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataOrBuilder
optional int64 completion_time = 5;
getCompletionTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
optional int64 completion_time = 9;
getCompletionTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
optional int64 completion_time = 9;
getCompletionTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
optional int64 completion_time = 9;
getCompletionTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
optional int64 completion_time = 12;
getCompletionTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
optional int64 completion_time = 12;
getCompletionTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
optional int64 completion_time = 12;
getConf() - Method in class org.apache.spark.api.java.JavaSparkContext
Return a copy of this JavaSparkContext's configuration.
getConf() - Method in interface org.apache.spark.input.Configurable
 
getConf() - Method in class org.apache.spark.rdd.HadoopRDD
 
getConf() - Method in class org.apache.spark.rdd.NewHadoopRDD
 
getConf() - Method in class org.apache.spark.SparkContext
Return a copy of this SparkContext's configuration.
getConf(String) - Method in class org.apache.spark.sql.SQLContext
Return the value of Spark SQL configuration property for the given key.
getConf(String, String) - Method in class org.apache.spark.sql.SQLContext
Return the value of Spark SQL configuration property for the given key.
getConfiguration() - Method in class org.apache.spark.input.PortableDataStream
 
getConfiguredLocalDirs(SparkConf) - Static method in class org.apache.spark.util.Utils
Return the configured local directories where Spark can write files.
getConnection() - Method in interface org.apache.spark.rdd.JdbcRDD.ConnectionFactory
 
getConnection(Driver, Map<String, String>) - Method in class org.apache.spark.sql.jdbc.JdbcConnectionProvider
Opens connection to the database.
getContextOrSparkClassLoader() - Static method in class org.apache.spark.util.Utils
Get the Context ClassLoader on this thread or, if not present, the ClassLoader that loaded Spark.
getConvergenceTol() - Method in class org.apache.spark.mllib.clustering.GaussianMixture
Return the largest change in log-likelihood at which convergence is considered to have occurred.
getCoresGranted() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
optional int32 cores_granted = 3;
getCoresGranted() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo
optional int32 cores_granted = 3;
getCoresGranted() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoOrBuilder
optional int32 cores_granted = 3;
getCoresPerExecutor() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
optional int32 cores_per_executor = 5;
getCoresPerExecutor() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo
optional int32 cores_per_executor = 5;
getCoresPerExecutor() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoOrBuilder
optional int32 cores_per_executor = 5;
getCorrelationFromName(String) - Static method in class org.apache.spark.mllib.stat.correlation.Correlations
 
getCorruptMergedBlockChunks(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
repeated double corrupt_merged_block_chunks = 1;
getCorruptMergedBlockChunks(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions
repeated double corrupt_merged_block_chunks = 1;
getCorruptMergedBlockChunks(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributionsOrBuilder
repeated double corrupt_merged_block_chunks = 1;
getCorruptMergedBlockChunks() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics.Builder
int64 corrupt_merged_block_chunks = 1;
getCorruptMergedBlockChunks() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics
int64 corrupt_merged_block_chunks = 1;
getCorruptMergedBlockChunks() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricsOrBuilder
int64 corrupt_merged_block_chunks = 1;
getCorruptMergedBlockChunksCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
repeated double corrupt_merged_block_chunks = 1;
getCorruptMergedBlockChunksCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions
repeated double corrupt_merged_block_chunks = 1;
getCorruptMergedBlockChunksCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributionsOrBuilder
repeated double corrupt_merged_block_chunks = 1;
getCorruptMergedBlockChunksList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
repeated double corrupt_merged_block_chunks = 1;
getCorruptMergedBlockChunksList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions
repeated double corrupt_merged_block_chunks = 1;
getCorruptMergedBlockChunksList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributionsOrBuilder
repeated double corrupt_merged_block_chunks = 1;
getCount() - Method in class org.apache.spark.storage.CountingWritableChannel
 
getCurrentDefaultValue() - Method in class org.apache.spark.sql.types.StructField
Return the current default value of this StructField.
getCurrentProcessingTimeMs() - Method in interface org.apache.spark.sql.streaming.GroupState
Get the current processing time as milliseconds in epoch time.
getCurrentUserGroups(SparkConf, String) - Static method in class org.apache.spark.util.Utils
 
getCurrentUserName() - Static method in class org.apache.spark.util.Utils
Returns the current user name.
getCurrentWatermarkMs() - Method in interface org.apache.spark.sql.streaming.GroupState
Get the current event time watermark as milliseconds in epoch time.
getCustomMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
Deprecated.
getCustomMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress
Deprecated.
getCustomMetrics() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgressOrBuilder
Deprecated.
getCustomMetricsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
 
getCustomMetricsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress
 
getCustomMetricsCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgressOrBuilder
map<string, int64> custom_metrics = 12;
getCustomMetricsMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
map<string, int64> custom_metrics = 12;
getCustomMetricsMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress
map<string, int64> custom_metrics = 12;
getCustomMetricsMap() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgressOrBuilder
map<string, int64> custom_metrics = 12;
getCustomMetricsOrDefault(String, long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
map<string, int64> custom_metrics = 12;
getCustomMetricsOrDefault(String, long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress
map<string, int64> custom_metrics = 12;
getCustomMetricsOrDefault(String, long) - Method in interface org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgressOrBuilder
map<string, int64> custom_metrics = 12;
getCustomMetricsOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
map<string, int64> custom_metrics = 12;
getCustomMetricsOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress
map<string, int64> custom_metrics = 12;
getCustomMetricsOrThrow(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgressOrBuilder
map<string, int64> custom_metrics = 12;
getData(Row) - Static method in class org.apache.spark.ml.image.ImageSchema
Gets the image data
getDatabase(String) - Method in class org.apache.spark.sql.catalog.Catalog
Get the database (namespace) with the specified name (can be qualified with catalog).
getDataDistribution(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
repeated .org.apache.spark.status.protobuf.RDDDataDistribution data_distribution = 8;
getDataDistribution(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo
repeated .org.apache.spark.status.protobuf.RDDDataDistribution data_distribution = 8;
getDataDistribution(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoOrBuilder
repeated .org.apache.spark.status.protobuf.RDDDataDistribution data_distribution = 8;
getDataDistributionBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
repeated .org.apache.spark.status.protobuf.RDDDataDistribution data_distribution = 8;
getDataDistributionBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
repeated .org.apache.spark.status.protobuf.RDDDataDistribution data_distribution = 8;
getDataDistributionCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
repeated .org.apache.spark.status.protobuf.RDDDataDistribution data_distribution = 8;
getDataDistributionCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo
repeated .org.apache.spark.status.protobuf.RDDDataDistribution data_distribution = 8;
getDataDistributionCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoOrBuilder
repeated .org.apache.spark.status.protobuf.RDDDataDistribution data_distribution = 8;
getDataDistributionList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
repeated .org.apache.spark.status.protobuf.RDDDataDistribution data_distribution = 8;
getDataDistributionList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo
repeated .org.apache.spark.status.protobuf.RDDDataDistribution data_distribution = 8;
getDataDistributionList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoOrBuilder
repeated .org.apache.spark.status.protobuf.RDDDataDistribution data_distribution = 8;
getDataDistributionOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
repeated .org.apache.spark.status.protobuf.RDDDataDistribution data_distribution = 8;
getDataDistributionOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo
repeated .org.apache.spark.status.protobuf.RDDDataDistribution data_distribution = 8;
getDataDistributionOrBuilder(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoOrBuilder
repeated .org.apache.spark.status.protobuf.RDDDataDistribution data_distribution = 8;
getDataDistributionOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
repeated .org.apache.spark.status.protobuf.RDDDataDistribution data_distribution = 8;
getDataDistributionOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo
repeated .org.apache.spark.status.protobuf.RDDDataDistribution data_distribution = 8;
getDataDistributionOrBuilderList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoOrBuilder
repeated .org.apache.spark.status.protobuf.RDDDataDistribution data_distribution = 8;
getDate(int) - Method in interface org.apache.spark.sql.Row
Returns the value at position i of date type as java.sql.Date.
getDecimal(int) - Method in interface org.apache.spark.sql.Row
Returns the value at position i of decimal type as java.math.BigDecimal.
getDecimal(int, int, int) - Method in class org.apache.spark.sql.vectorized.ArrowColumnVector
 
getDecimal(int, int, int) - Method in class org.apache.spark.sql.vectorized.ColumnarArray
 
getDecimal(int, int, int) - Method in class org.apache.spark.sql.vectorized.ColumnarBatchRow
 
getDecimal(int, int, int) - Method in class org.apache.spark.sql.vectorized.ColumnarRow
 
getDecimal(int, int, int) - Method in class org.apache.spark.sql.vectorized.ColumnVector
Returns the decimal type value for rowId.
getDefault(Param<T>) - Method in interface org.apache.spark.ml.param.Params
Gets the default value of a parameter.
getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo
 
getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo
 
getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
 
getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoWrapper
 
getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo
 
getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoWrapper
 
getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.AppSummary
 
getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
 
getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetrics
 
getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
 
getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions
 
getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest
 
getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary
 
getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper
 
getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
 
getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryWrapper
 
getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributions
 
getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.InputMetrics
 
getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.JobData
 
getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper
 
getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.MemoryMetrics
 
getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributions
 
getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetrics
 
getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.PairStrings
 
getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.PoolData
 
getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary
 
getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryWrapper
 
getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution
 
getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper
 
getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationEdge
 
getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper
 
getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode
 
getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo
 
getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo
 
getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoWrapper
 
getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ResourceInformation
 
getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo
 
getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileWrapper
 
getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.RuntimeInfo
 
getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions
 
getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics
 
getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions
 
getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics
 
getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions
 
getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetrics
 
getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress
 
getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress
 
getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper
 
getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphEdge
 
getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode
 
getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper
 
getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper
 
getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummary
 
getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryWrapper
 
getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
 
getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetric
 
getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.StageData
 
getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper
 
getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress
 
getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData
 
getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData
 
getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
 
getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressWrapper
 
getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
 
getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
 
getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
 
getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics
 
getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.TaskResourceRequest
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo.Builder
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo.Builder
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoWrapper.Builder
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoWrapper
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoWrapper.Builder
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoWrapper
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.AppSummary.Builder
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.AppSummary
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetrics.Builder
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetrics
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions.Builder
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest.Builder
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper.Builder
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryWrapper.Builder
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryWrapper
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributions.Builder
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributions
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetrics.Builder
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetrics
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper.Builder
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.MemoryMetrics.Builder
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.MemoryMetrics
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributions.Builder
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributions
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetrics.Builder
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetrics
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.PairStrings.Builder
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.PairStrings
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.PoolData.Builder
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.PoolData
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryWrapper.Builder
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryWrapper
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution.Builder
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationEdge.Builder
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationEdge
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode.Builder
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo.Builder
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoWrapper.Builder
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoWrapper
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceInformation.Builder
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceInformation
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo.Builder
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileWrapper.Builder
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileWrapper
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.RuntimeInfo.Builder
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.RuntimeInfo
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics.Builder
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics.Builder
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions.Builder
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetrics.Builder
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetrics
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress.Builder
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphEdge.Builder
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphEdge
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper.Builder
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummary.Builder
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummary
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryWrapper.Builder
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryWrapper
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetric.Builder
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetric
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper.Builder
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData.Builder
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressWrapper.Builder
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressWrapper
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskResourceRequest.Builder
 
getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskResourceRequest
 
getDefaultPropertiesFile(Map<String, String>) - Static method in class org.apache.spark.util.Utils
Return the path of the default Spark properties file.
getDefaultReadLimit() - Method in interface org.apache.spark.sql.connector.read.streaming.SupportsAdmissionControl
Returns the read limits potentially passed to the data source through options when creating the data source.
getDefaultSession() - Static method in class org.apache.spark.sql.SparkSession
Returns the default SparkSession that is returned by the builder.
getDegree() - Method in class org.apache.spark.ml.feature.PolynomialExpansion
 
getDeleteColumnQuery(String, String) - Static method in class org.apache.spark.sql.jdbc.DB2Dialect
 
getDeleteColumnQuery(String, String) - Static method in class org.apache.spark.sql.jdbc.DerbyDialect
 
getDeleteColumnQuery(String, String) - Static method in class org.apache.spark.sql.jdbc.H2Dialect
 
getDeleteColumnQuery(String, String) - Method in class org.apache.spark.sql.jdbc.JdbcDialect
 
getDeleteColumnQuery(String, String) - Static method in class org.apache.spark.sql.jdbc.MsSqlServerDialect
 
getDeleteColumnQuery(String, String) - Static method in class org.apache.spark.sql.jdbc.MySQLDialect
 
getDeleteColumnQuery(String, String) - Static method in class org.apache.spark.sql.jdbc.NoopDialect
 
getDeleteColumnQuery(String, String) - Static method in class org.apache.spark.sql.jdbc.OracleDialect
 
getDeleteColumnQuery(String, String) - Static method in class org.apache.spark.sql.jdbc.PostgresDialect
 
getDeleteColumnQuery(String, String) - Static method in class org.apache.spark.sql.jdbc.TeradataDialect
 
getDenseSizeInBytes() - Method in interface org.apache.spark.ml.linalg.Matrix
Gets the size of the dense representation of this `Matrix`.
getDependencies() - Method in class org.apache.spark.rdd.CoGroupedRDD
 
getDependencies() - Method in class org.apache.spark.rdd.ShuffledRDD
 
getDependencies() - Method in class org.apache.spark.rdd.UnionRDD
 
getDeprecatedConfig(String, Map<String, String>) - Static method in class org.apache.spark.SparkConf
Looks for available deprecated keys for the given config option, and return the first value available.
getDesc() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
optional string desc = 3;
getDesc() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper
optional string desc = 3;
getDesc() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapperOrBuilder
optional string desc = 3;
getDesc() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
optional string desc = 3;
getDesc() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode
optional string desc = 3;
getDesc() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeOrBuilder
optional string desc = 3;
getDescBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
optional string desc = 3;
getDescBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper
optional string desc = 3;
getDescBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapperOrBuilder
optional string desc = 3;
getDescBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
optional string desc = 3;
getDescBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode
optional string desc = 3;
getDescBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeOrBuilder
optional string desc = 3;
getDescription() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
optional string description = 3;
getDescription() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
optional string description = 3;
getDescription() - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataOrBuilder
optional string description = 3;
getDescription() - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress.Builder
optional string description = 1;
getDescription() - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress
optional string description = 1;
getDescription() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SinkProgressOrBuilder
optional string description = 1;
getDescription() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
optional string description = 1;
getDescription() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress
optional string description = 1;
getDescription() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SourceProgressOrBuilder
optional string description = 1;
getDescription() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
optional string description = 3;
getDescription() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
optional string description = 3;
getDescription() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
optional string description = 3;
getDescription() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
optional string description = 40;
getDescription() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
optional string description = 40;
getDescription() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
optional string description = 40;
getDescriptionBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
optional string description = 3;
getDescriptionBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
optional string description = 3;
getDescriptionBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataOrBuilder
optional string description = 3;
getDescriptionBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress.Builder
optional string description = 1;
getDescriptionBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress
optional string description = 1;
getDescriptionBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SinkProgressOrBuilder
optional string description = 1;
getDescriptionBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
optional string description = 1;
getDescriptionBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress
optional string description = 1;
getDescriptionBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SourceProgressOrBuilder
optional string description = 1;
getDescriptionBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
optional string description = 3;
getDescriptionBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
optional string description = 3;
getDescriptionBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
optional string description = 3;
getDescriptionBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
optional string description = 40;
getDescriptionBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
optional string description = 40;
getDescriptionBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
optional string description = 40;
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo.Builder
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo.Builder
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoWrapper.Builder
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoWrapper
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoWrapper.Builder
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoWrapper
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.AppSummary.Builder
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.AppSummary
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
 
getDescriptor() - Static method in enum org.apache.spark.status.protobuf.StoreTypes.DeterministicLevel
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetrics.Builder
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetrics
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions.Builder
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest.Builder
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper.Builder
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryWrapper.Builder
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryWrapper
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributions.Builder
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributions
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.InputMetrics.Builder
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.InputMetrics
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.JobData
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper.Builder
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper
 
getDescriptor() - Static method in enum org.apache.spark.status.protobuf.StoreTypes.JobExecutionStatus
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.MemoryMetrics.Builder
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.MemoryMetrics
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributions.Builder
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributions
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetrics.Builder
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetrics
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.PairStrings.Builder
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.PairStrings
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.PoolData.Builder
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.PoolData
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryWrapper.Builder
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryWrapper
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution.Builder
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationEdge.Builder
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationEdge
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode.Builder
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo.Builder
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoWrapper.Builder
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoWrapper
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ResourceInformation.Builder
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ResourceInformation
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo.Builder
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileWrapper.Builder
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileWrapper
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.RuntimeInfo.Builder
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.RuntimeInfo
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics.Builder
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics.Builder
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions.Builder
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetrics.Builder
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetrics
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress.Builder
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphEdge.Builder
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphEdge
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper.Builder
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummary.Builder
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummary
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryWrapper.Builder
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryWrapper
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetric.Builder
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetric
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.StageData
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper.Builder
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper
 
getDescriptor() - Static method in enum org.apache.spark.status.protobuf.StoreTypes.StageStatus
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData.Builder
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressWrapper.Builder
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressWrapper
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.TaskResourceRequest.Builder
 
getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.TaskResourceRequest
 
getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo.Builder
 
getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo.Builder
 
getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
 
getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoWrapper.Builder
 
getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
 
getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoWrapper.Builder
 
getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.AppSummary.Builder
 
getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
 
getDescriptorForType() - Method in enum org.apache.spark.status.protobuf.StoreTypes.DeterministicLevel
 
getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetrics.Builder
 
getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
 
getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions.Builder
 
getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest.Builder
 
getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
 
getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper.Builder
 
getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
 
getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryWrapper.Builder
 
getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributions.Builder
 
getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetrics.Builder
 
getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
 
getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper.Builder
 
getDescriptorForType() - Method in enum org.apache.spark.status.protobuf.StoreTypes.JobExecutionStatus
 
getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.MemoryMetrics.Builder
 
getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributions.Builder
 
getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetrics.Builder
 
getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.PairStrings.Builder
 
getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.PoolData.Builder
 
getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder
 
getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryWrapper.Builder
 
getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution.Builder
 
getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
 
getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationEdge.Builder
 
getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
 
getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode.Builder
 
getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo.Builder
 
getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
 
getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoWrapper.Builder
 
getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceInformation.Builder
 
getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo.Builder
 
getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileWrapper.Builder
 
getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.RuntimeInfo.Builder
 
getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
 
getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics.Builder
 
getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
 
getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics.Builder
 
getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions.Builder
 
getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetrics.Builder
 
getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress.Builder
 
getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
 
getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
 
getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphEdge.Builder
 
getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
 
getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper.Builder
 
getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
 
getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummary.Builder
 
getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryWrapper.Builder
 
getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
 
getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetric.Builder
 
getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
 
getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper.Builder
 
getDescriptorForType() - Method in enum org.apache.spark.status.protobuf.StoreTypes.StageStatus
 
getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
 
getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
 
getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData.Builder
 
getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
 
getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressWrapper.Builder
 
getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
 
getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
 
getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
 
getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
 
getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskResourceRequest.Builder
 
getDeserialized() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
bool deserialized = 7;
getDeserialized() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData
bool deserialized = 7;
getDeserialized() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamBlockDataOrBuilder
bool deserialized = 7;
getDetails() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
optional string details = 4;
getDetails() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
optional string details = 4;
getDetails() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
optional string details = 4;
getDetails() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
optional string details = 41;
getDetails() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
optional string details = 41;
getDetails() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
optional string details = 41;
getDetailsBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
optional string details = 4;
getDetailsBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
optional string details = 4;
getDetailsBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
optional string details = 4;
getDetailsBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
optional string details = 41;
getDetailsBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
optional string details = 41;
getDetailsBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
optional string details = 41;
getDiscoveryScript() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest.Builder
optional string discoveryScript = 3;
getDiscoveryScript() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest
optional string discoveryScript = 3;
getDiscoveryScript() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequestOrBuilder
optional string discoveryScript = 3;
getDiscoveryScriptBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest.Builder
optional string discoveryScript = 3;
getDiscoveryScriptBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest
optional string discoveryScript = 3;
getDiscoveryScriptBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequestOrBuilder
optional string discoveryScript = 3;
getDiskBytesSpilled() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double disk_bytes_spilled = 17;
getDiskBytesSpilled() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
double disk_bytes_spilled = 17;
getDiskBytesSpilled() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
double disk_bytes_spilled = 17;
getDiskBytesSpilled(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double disk_bytes_spilled = 15;
getDiskBytesSpilled(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
repeated double disk_bytes_spilled = 15;
getDiskBytesSpilled(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
repeated double disk_bytes_spilled = 15;
getDiskBytesSpilled() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
int64 disk_bytes_spilled = 14;
getDiskBytesSpilled() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary
int64 disk_bytes_spilled = 14;
getDiskBytesSpilled() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryOrBuilder
int64 disk_bytes_spilled = 14;
getDiskBytesSpilled() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int64 disk_bytes_spilled = 22;
getDiskBytesSpilled() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
int64 disk_bytes_spilled = 22;
getDiskBytesSpilled() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
int64 disk_bytes_spilled = 22;
getDiskBytesSpilled() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 disk_bytes_spilled = 24;
getDiskBytesSpilled() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
int64 disk_bytes_spilled = 24;
getDiskBytesSpilled() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
int64 disk_bytes_spilled = 24;
getDiskBytesSpilled(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double disk_bytes_spilled = 14;
getDiskBytesSpilled(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
repeated double disk_bytes_spilled = 14;
getDiskBytesSpilled(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
repeated double disk_bytes_spilled = 14;
getDiskBytesSpilled() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
int64 disk_bytes_spilled = 9;
getDiskBytesSpilled() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics
int64 disk_bytes_spilled = 9;
getDiskBytesSpilled() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricsOrBuilder
int64 disk_bytes_spilled = 9;
getDiskBytesSpilledCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double disk_bytes_spilled = 15;
getDiskBytesSpilledCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
repeated double disk_bytes_spilled = 15;
getDiskBytesSpilledCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
repeated double disk_bytes_spilled = 15;
getDiskBytesSpilledCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double disk_bytes_spilled = 14;
getDiskBytesSpilledCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
repeated double disk_bytes_spilled = 14;
getDiskBytesSpilledCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
repeated double disk_bytes_spilled = 14;
getDiskBytesSpilledList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double disk_bytes_spilled = 15;
getDiskBytesSpilledList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
repeated double disk_bytes_spilled = 15;
getDiskBytesSpilledList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
repeated double disk_bytes_spilled = 15;
getDiskBytesSpilledList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double disk_bytes_spilled = 14;
getDiskBytesSpilledList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
repeated double disk_bytes_spilled = 14;
getDiskBytesSpilledList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
repeated double disk_bytes_spilled = 14;
getDiskSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
int64 disk_size = 9;
getDiskSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData
int64 disk_size = 9;
getDiskSize() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamBlockDataOrBuilder
int64 disk_size = 9;
getDiskUsed() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
int64 disk_used = 6;
getDiskUsed() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
int64 disk_used = 6;
getDiskUsed() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
int64 disk_used = 6;
getDiskUsed() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution.Builder
int64 disk_used = 4;
getDiskUsed() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution
int64 disk_used = 4;
getDiskUsed() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDDataDistributionOrBuilder
int64 disk_used = 4;
getDiskUsed() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo.Builder
int64 disk_used = 4;
getDiskUsed() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo
int64 disk_used = 4;
getDiskUsed() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfoOrBuilder
int64 disk_used = 4;
getDiskUsed() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
int64 disk_used = 7;
getDiskUsed() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo
int64 disk_used = 7;
getDiskUsed() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoOrBuilder
int64 disk_used = 7;
getDistanceMeasure() - Method in class org.apache.spark.ml.evaluation.ClusteringEvaluator
 
getDistanceMeasure() - Method in class org.apache.spark.ml.evaluation.ClusteringMetrics
 
getDistanceMeasure() - Method in interface org.apache.spark.ml.param.shared.HasDistanceMeasure
 
getDistanceMeasure() - Method in class org.apache.spark.mllib.clustering.BisectingKMeans
The distance suite used by the algorithm.
getDistanceMeasure() - Method in class org.apache.spark.mllib.clustering.KMeans
The distance suite used by the algorithm.
getDistributions() - Method in class org.apache.spark.status.LiveRDD
 
getDocConcentration() - Method in interface org.apache.spark.ml.clustering.LDAParams
 
getDocConcentration() - Method in class org.apache.spark.mllib.clustering.LDA
Concentration parameter (commonly named "alpha") for the prior placed on documents' distributions over topics ("theta").
getDouble(String, double) - Method in class org.apache.spark.SparkConf
Get a parameter as a double, falling back to a default if not ste
getDouble(int) - Method in interface org.apache.spark.sql.Row
Returns the value at position i as a primitive double.
getDouble(String) - Method in class org.apache.spark.sql.types.Metadata
Gets a Double.
getDouble(String, double) - Method in class org.apache.spark.sql.util.CaseInsensitiveStringMap
Returns the double value to which the specified key is mapped, or defaultValue if there is no mapping for the key.
getDouble(int) - Method in class org.apache.spark.sql.vectorized.ArrowColumnVector
 
getDouble(int) - Method in class org.apache.spark.sql.vectorized.ColumnarArray
 
getDouble(int) - Method in class org.apache.spark.sql.vectorized.ColumnarBatchRow
 
getDouble(int) - Method in class org.apache.spark.sql.vectorized.ColumnarRow
 
getDouble(int) - Method in class org.apache.spark.sql.vectorized.ColumnVector
Returns the double type value for rowId.
getDoubleArray(String) - Method in class org.apache.spark.sql.types.Metadata
Gets a Double array.
getDoubles(int, int) - Method in class org.apache.spark.sql.vectorized.ColumnVector
Gets double type values from [rowId, rowId + count).
getDriverAttributes() - Method in interface org.apache.spark.scheduler.SchedulerBackend
Get the attributes on driver.
getDriverLogUrls() - Method in interface org.apache.spark.scheduler.SchedulerBackend
Get the URLs for the driver logs.
getDropLast() - Method in interface org.apache.spark.ml.feature.OneHotEncoderBase
 
getDstCol() - Method in interface org.apache.spark.ml.clustering.PowerIterationClusteringParams
 
getDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo.Builder
int64 duration = 5;
getDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo
int64 duration = 5;
getDuration() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfoOrBuilder
int64 duration = 5;
getDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double duration = 5;
getDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
double duration = 5;
getDuration() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
double duration = 5;
getDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
optional int64 duration = 7;
getDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
optional int64 duration = 7;
getDuration() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataOrBuilder
optional int64 duration = 7;
getDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 duration = 7;
getDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
int64 duration = 7;
getDuration() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
int64 duration = 7;
getDuration(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double duration = 2;
getDuration(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
repeated double duration = 2;
getDuration(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
repeated double duration = 2;
getDuration(JobData) - Static method in class org.apache.spark.ui.jobs.JobDataUtil
 
getDurationCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double duration = 2;
getDurationCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
repeated double duration = 2;
getDurationCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
repeated double duration = 2;
getDurationList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double duration = 2;
getDurationList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
repeated double duration = 2;
getDurationList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
repeated double duration = 2;
getDurationMs() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
Deprecated.
getDurationMs() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
Deprecated.
getDurationMs() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
Deprecated.
getDurationMsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
 
getDurationMsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
 
getDurationMsCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
map<string, int64> duration_ms = 7;
getDurationMsMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
map<string, int64> duration_ms = 7;
getDurationMsMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
map<string, int64> duration_ms = 7;
getDurationMsMap() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
map<string, int64> duration_ms = 7;
getDurationMsOrDefault(String, long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
map<string, int64> duration_ms = 7;
getDurationMsOrDefault(String, long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
map<string, int64> duration_ms = 7;
getDurationMsOrDefault(String, long) - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
map<string, int64> duration_ms = 7;
getDurationMsOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
map<string, int64> duration_ms = 7;
getDurationMsOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
map<string, int64> duration_ms = 7;
getDurationMsOrThrow(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
map<string, int64> duration_ms = 7;
getDynamicAllocationInitialExecutors(SparkConf) - Static method in class org.apache.spark.util.Utils
Return the initial number of executors for dynamic allocation.
getEdges(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationEdge edges = 2;
getEdges(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper
repeated .org.apache.spark.status.protobuf.RDDOperationEdge edges = 2;
getEdges(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapperOrBuilder
repeated .org.apache.spark.status.protobuf.RDDOperationEdge edges = 2;
getEdges(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.SparkPlanGraphEdge edges = 3;
getEdges(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper
repeated .org.apache.spark.status.protobuf.SparkPlanGraphEdge edges = 3;
getEdges(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapperOrBuilder
repeated .org.apache.spark.status.protobuf.SparkPlanGraphEdge edges = 3;
getEdgesBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationEdge edges = 2;
getEdgesBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.SparkPlanGraphEdge edges = 3;
getEdgesBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationEdge edges = 2;
getEdgesBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.SparkPlanGraphEdge edges = 3;
getEdgesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationEdge edges = 2;
getEdgesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper
repeated .org.apache.spark.status.protobuf.RDDOperationEdge edges = 2;
getEdgesCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapperOrBuilder
repeated .org.apache.spark.status.protobuf.RDDOperationEdge edges = 2;
getEdgesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.SparkPlanGraphEdge edges = 3;
getEdgesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper
repeated .org.apache.spark.status.protobuf.SparkPlanGraphEdge edges = 3;
getEdgesCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapperOrBuilder
repeated .org.apache.spark.status.protobuf.SparkPlanGraphEdge edges = 3;
getEdgesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationEdge edges = 2;
getEdgesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper
repeated .org.apache.spark.status.protobuf.RDDOperationEdge edges = 2;
getEdgesList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapperOrBuilder
repeated .org.apache.spark.status.protobuf.RDDOperationEdge edges = 2;
getEdgesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.SparkPlanGraphEdge edges = 3;
getEdgesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper
repeated .org.apache.spark.status.protobuf.SparkPlanGraphEdge edges = 3;
getEdgesList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapperOrBuilder
repeated .org.apache.spark.status.protobuf.SparkPlanGraphEdge edges = 3;
getEdgesOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationEdge edges = 2;
getEdgesOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper
repeated .org.apache.spark.status.protobuf.RDDOperationEdge edges = 2;
getEdgesOrBuilder(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapperOrBuilder
repeated .org.apache.spark.status.protobuf.RDDOperationEdge edges = 2;
getEdgesOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.SparkPlanGraphEdge edges = 3;
getEdgesOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper
repeated .org.apache.spark.status.protobuf.SparkPlanGraphEdge edges = 3;
getEdgesOrBuilder(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapperOrBuilder
repeated .org.apache.spark.status.protobuf.SparkPlanGraphEdge edges = 3;
getEdgesOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationEdge edges = 2;
getEdgesOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper
repeated .org.apache.spark.status.protobuf.RDDOperationEdge edges = 2;
getEdgesOrBuilderList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapperOrBuilder
repeated .org.apache.spark.status.protobuf.RDDOperationEdge edges = 2;
getEdgesOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.SparkPlanGraphEdge edges = 3;
getEdgesOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper
repeated .org.apache.spark.status.protobuf.SparkPlanGraphEdge edges = 3;
getEdgesOrBuilderList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapperOrBuilder
repeated .org.apache.spark.status.protobuf.SparkPlanGraphEdge edges = 3;
getElasticNetParam() - Method in interface org.apache.spark.ml.param.shared.HasElasticNetParam
 
getEndOffset() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
optional string end_offset = 3;
getEndOffset() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress
optional string end_offset = 3;
getEndOffset() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SourceProgressOrBuilder
optional string end_offset = 3;
getEndOffsetBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
optional string end_offset = 3;
getEndOffsetBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress
optional string end_offset = 3;
getEndOffsetBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SourceProgressOrBuilder
optional string end_offset = 3;
getEndTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo.Builder
int64 end_time = 3;
getEndTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo
int64 end_time = 3;
getEndTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfoOrBuilder
int64 end_time = 3;
getEndTimeEpoch() - Method in class org.apache.spark.status.api.v1.ApplicationAttemptInfo
 
getEndTimestamp() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData.Builder
optional int64 end_timestamp = 7;
getEndTimestamp() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData
optional int64 end_timestamp = 7;
getEndTimestamp() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryDataOrBuilder
optional int64 end_timestamp = 7;
getEps() - Method in class org.apache.spark.ml.evaluation.MulticlassClassificationEvaluator
 
getEpsilon() - Method in interface org.apache.spark.ml.regression.LinearRegressionParams
 
getEpsilon() - Method in class org.apache.spark.mllib.clustering.KMeans
The distance threshold within which we've consider centers to have converged.
getError() - Method in interface org.apache.spark.launcher.SparkAppHandle
If the application failed due to an error, return the underlying error.
getErrorClass() - Method in exception org.apache.spark.SparkException
 
getErrorClass() - Method in interface org.apache.spark.SparkThrowable
 
getErrorClass() - Method in exception org.apache.spark.sql.AnalysisException
 
getErrorClass() - Method in exception org.apache.spark.sql.streaming.StreamingQueryException
 
getErrorMessage(String, Map<String, String>) - Method in class org.apache.spark.ErrorClassesJsonReader
 
getErrorMessage() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
optional string error_message = 10;
getErrorMessage() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
optional string error_message = 10;
getErrorMessage() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
optional string error_message = 10;
getErrorMessage() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
optional string error_message = 14;
getErrorMessage() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
optional string error_message = 14;
getErrorMessage() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataOrBuilder
optional string error_message = 14;
getErrorMessage() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
optional string error_message = 14;
getErrorMessage() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
optional string error_message = 14;
getErrorMessage() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
optional string error_message = 14;
getErrorMessageBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
optional string error_message = 10;
getErrorMessageBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
optional string error_message = 10;
getErrorMessageBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
optional string error_message = 10;
getErrorMessageBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
optional string error_message = 14;
getErrorMessageBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
optional string error_message = 14;
getErrorMessageBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataOrBuilder
optional string error_message = 14;
getErrorMessageBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
optional string error_message = 14;
getErrorMessageBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
optional string error_message = 14;
getErrorMessageBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
optional string error_message = 14;
getEstimator() - Method in interface org.apache.spark.ml.tuning.ValidatorParams
 
getEstimatorParamMaps() - Method in interface org.apache.spark.ml.tuning.ValidatorParams
 
getEvaluator() - Method in interface org.apache.spark.ml.tuning.ValidatorParams
 
getEventTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
Deprecated.
getEventTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
Deprecated.
getEventTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
Deprecated.
getEventTimeCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
 
getEventTimeCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
 
getEventTimeCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
map<string, string> event_time = 8;
getEventTimeMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
map<string, string> event_time = 8;
getEventTimeMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
map<string, string> event_time = 8;
getEventTimeMap() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
map<string, string> event_time = 8;
getEventTimeOrDefault(String, String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
map<string, string> event_time = 8;
getEventTimeOrDefault(String, String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
map<string, string> event_time = 8;
getEventTimeOrDefault(String, String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
map<string, string> event_time = 8;
getEventTimeOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
map<string, string> event_time = 8;
getEventTimeOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
map<string, string> event_time = 8;
getEventTimeOrThrow(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
map<string, string> event_time = 8;
getException() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData.Builder
optional string exception = 5;
getException() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData
optional string exception = 5;
getException() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryDataOrBuilder
optional string exception = 5;
getExceptionBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData.Builder
optional string exception = 5;
getExceptionBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData
optional string exception = 5;
getExceptionBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryDataOrBuilder
optional string exception = 5;
getExcludedInStages(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
repeated int64 excluded_in_stages = 31;
getExcludedInStages(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
repeated int64 excluded_in_stages = 31;
getExcludedInStages(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
repeated int64 excluded_in_stages = 31;
getExcludedInStagesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
repeated int64 excluded_in_stages = 31;
getExcludedInStagesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
repeated int64 excluded_in_stages = 31;
getExcludedInStagesCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
repeated int64 excluded_in_stages = 31;
getExcludedInStagesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
repeated int64 excluded_in_stages = 31;
getExcludedInStagesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
repeated int64 excluded_in_stages = 31;
getExcludedInStagesList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
repeated int64 excluded_in_stages = 31;
getExecutionContext() - Method in interface org.apache.spark.ml.param.shared.HasParallelism
Create a new execution context with a thread-pool that has a maximum number of threads set to the value of parallelism.
getExecutionId() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
int64 execution_id = 1;
getExecutionId() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper
int64 execution_id = 1;
getExecutionId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapperOrBuilder
int64 execution_id = 1;
getExecutionId() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
int64 execution_id = 1;
getExecutionId() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
int64 execution_id = 1;
getExecutionId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
int64 execution_id = 1;
getExecutorCpuTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double executor_cpu_time = 9;
getExecutorCpuTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
double executor_cpu_time = 9;
getExecutorCpuTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
double executor_cpu_time = 9;
getExecutorCpuTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int64 executor_cpu_time = 17;
getExecutorCpuTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
int64 executor_cpu_time = 17;
getExecutorCpuTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
int64 executor_cpu_time = 17;
getExecutorCpuTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 executor_cpu_time = 19;
getExecutorCpuTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
int64 executor_cpu_time = 19;
getExecutorCpuTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
int64 executor_cpu_time = 19;
getExecutorCpuTime(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double executor_cpu_time = 6;
getExecutorCpuTime(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
repeated double executor_cpu_time = 6;
getExecutorCpuTime(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
repeated double executor_cpu_time = 6;
getExecutorCpuTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
int64 executor_cpu_time = 4;
getExecutorCpuTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics
int64 executor_cpu_time = 4;
getExecutorCpuTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricsOrBuilder
int64 executor_cpu_time = 4;
getExecutorCpuTimeCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double executor_cpu_time = 6;
getExecutorCpuTimeCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
repeated double executor_cpu_time = 6;
getExecutorCpuTimeCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
repeated double executor_cpu_time = 6;
getExecutorCpuTimeList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double executor_cpu_time = 6;
getExecutorCpuTimeList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
repeated double executor_cpu_time = 6;
getExecutorCpuTimeList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
repeated double executor_cpu_time = 6;
getExecutorDecommissionState(String) - Method in interface org.apache.spark.scheduler.TaskScheduler
If an executor is decommissioned, return its corresponding decommission info
getExecutorDeserializeCpuTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double executor_deserialize_cpu_time = 7;
getExecutorDeserializeCpuTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
double executor_deserialize_cpu_time = 7;
getExecutorDeserializeCpuTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
double executor_deserialize_cpu_time = 7;
getExecutorDeserializeCpuTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int64 executor_deserialize_cpu_time = 15;
getExecutorDeserializeCpuTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
int64 executor_deserialize_cpu_time = 15;
getExecutorDeserializeCpuTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
int64 executor_deserialize_cpu_time = 15;
getExecutorDeserializeCpuTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 executor_deserialize_cpu_time = 17;
getExecutorDeserializeCpuTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
int64 executor_deserialize_cpu_time = 17;
getExecutorDeserializeCpuTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
int64 executor_deserialize_cpu_time = 17;
getExecutorDeserializeCpuTime(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double executor_deserialize_cpu_time = 4;
getExecutorDeserializeCpuTime(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
repeated double executor_deserialize_cpu_time = 4;
getExecutorDeserializeCpuTime(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
repeated double executor_deserialize_cpu_time = 4;
getExecutorDeserializeCpuTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
int64 executor_deserialize_cpu_time = 2;
getExecutorDeserializeCpuTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics
int64 executor_deserialize_cpu_time = 2;
getExecutorDeserializeCpuTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricsOrBuilder
int64 executor_deserialize_cpu_time = 2;
getExecutorDeserializeCpuTimeCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double executor_deserialize_cpu_time = 4;
getExecutorDeserializeCpuTimeCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
repeated double executor_deserialize_cpu_time = 4;
getExecutorDeserializeCpuTimeCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
repeated double executor_deserialize_cpu_time = 4;
getExecutorDeserializeCpuTimeList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double executor_deserialize_cpu_time = 4;
getExecutorDeserializeCpuTimeList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
repeated double executor_deserialize_cpu_time = 4;
getExecutorDeserializeCpuTimeList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
repeated double executor_deserialize_cpu_time = 4;
getExecutorDeserializeTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double executor_deserialize_time = 6;
getExecutorDeserializeTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
double executor_deserialize_time = 6;
getExecutorDeserializeTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
double executor_deserialize_time = 6;
getExecutorDeserializeTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int64 executor_deserialize_time = 14;
getExecutorDeserializeTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
int64 executor_deserialize_time = 14;
getExecutorDeserializeTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
int64 executor_deserialize_time = 14;
getExecutorDeserializeTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 executor_deserialize_time = 16;
getExecutorDeserializeTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
int64 executor_deserialize_time = 16;
getExecutorDeserializeTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
int64 executor_deserialize_time = 16;
getExecutorDeserializeTime(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double executor_deserialize_time = 3;
getExecutorDeserializeTime(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
repeated double executor_deserialize_time = 3;
getExecutorDeserializeTime(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
repeated double executor_deserialize_time = 3;
getExecutorDeserializeTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
int64 executor_deserialize_time = 1;
getExecutorDeserializeTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics
int64 executor_deserialize_time = 1;
getExecutorDeserializeTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricsOrBuilder
int64 executor_deserialize_time = 1;
getExecutorDeserializeTimeCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double executor_deserialize_time = 3;
getExecutorDeserializeTimeCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
repeated double executor_deserialize_time = 3;
getExecutorDeserializeTimeCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
repeated double executor_deserialize_time = 3;
getExecutorDeserializeTimeList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double executor_deserialize_time = 3;
getExecutorDeserializeTimeList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
repeated double executor_deserialize_time = 3;
getExecutorDeserializeTimeList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
repeated double executor_deserialize_time = 3;
GetExecutorEndpointRef(String) - Constructor for class org.apache.spark.storage.BlockManagerMessages.GetExecutorEndpointRef
 
GetExecutorEndpointRef$() - Constructor for class org.apache.spark.storage.BlockManagerMessages.GetExecutorEndpointRef$
 
getExecutorEnv() - Method in class org.apache.spark.SparkConf
Get all executor environment variables set on this SparkConf
getExecutorId() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper.Builder
optional string executor_id = 3;
getExecutorId() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper
optional string executor_id = 3;
getExecutorId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapperOrBuilder
optional string executor_id = 3;
getExecutorId() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
optional string executor_id = 2;
getExecutorId() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData
optional string executor_id = 2;
getExecutorId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamBlockDataOrBuilder
optional string executor_id = 2;
getExecutorId() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
optional string executor_id = 8;
getExecutorId() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
optional string executor_id = 8;
getExecutorId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataOrBuilder
optional string executor_id = 8;
getExecutorId() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
optional string executor_id = 8;
getExecutorId() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
optional string executor_id = 8;
getExecutorId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
optional string executor_id = 8;
getExecutorIdBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper.Builder
optional string executor_id = 3;
getExecutorIdBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper
optional string executor_id = 3;
getExecutorIdBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapperOrBuilder
optional string executor_id = 3;
getExecutorIdBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
optional string executor_id = 2;
getExecutorIdBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData
optional string executor_id = 2;
getExecutorIdBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamBlockDataOrBuilder
optional string executor_id = 2;
getExecutorIdBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
optional string executor_id = 8;
getExecutorIdBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
optional string executor_id = 8;
getExecutorIdBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataOrBuilder
optional string executor_id = 8;
getExecutorIdBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
optional string executor_id = 8;
getExecutorIdBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
optional string executor_id = 8;
getExecutorIdBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
optional string executor_id = 8;
getExecutorInfos() - Method in class org.apache.spark.SparkStatusTracker
Returns information of all known executors, including host, port, cacheSize, numRunningTasks and memory metrics.
getExecutorLogs() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
Deprecated.
getExecutorLogs() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
Deprecated.
getExecutorLogs() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
Deprecated.
getExecutorLogs() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
Deprecated.
getExecutorLogs() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
Deprecated.
getExecutorLogs() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataOrBuilder
Deprecated.
getExecutorLogsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
 
getExecutorLogsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
 
getExecutorLogsCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
map<string, string> executor_logs = 23;
getExecutorLogsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
 
getExecutorLogsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
 
getExecutorLogsCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataOrBuilder
map<string, string> executor_logs = 16;
getExecutorLogsMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
map<string, string> executor_logs = 23;
getExecutorLogsMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
map<string, string> executor_logs = 23;
getExecutorLogsMap() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
map<string, string> executor_logs = 23;
getExecutorLogsMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
map<string, string> executor_logs = 16;
getExecutorLogsMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
map<string, string> executor_logs = 16;
getExecutorLogsMap() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataOrBuilder
map<string, string> executor_logs = 16;
getExecutorLogsOrDefault(String, String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
map<string, string> executor_logs = 23;
getExecutorLogsOrDefault(String, String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
map<string, string> executor_logs = 23;
getExecutorLogsOrDefault(String, String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
map<string, string> executor_logs = 23;
getExecutorLogsOrDefault(String, String) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
map<string, string> executor_logs = 16;
getExecutorLogsOrDefault(String, String) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
map<string, string> executor_logs = 16;
getExecutorLogsOrDefault(String, String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataOrBuilder
map<string, string> executor_logs = 16;
getExecutorLogsOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
map<string, string> executor_logs = 23;
getExecutorLogsOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
map<string, string> executor_logs = 23;
getExecutorLogsOrThrow(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
map<string, string> executor_logs = 23;
getExecutorLogsOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
map<string, string> executor_logs = 16;
getExecutorLogsOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
map<string, string> executor_logs = 16;
getExecutorLogsOrThrow(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataOrBuilder
map<string, string> executor_logs = 16;
GetExecutorLossReason(String) - Constructor for class org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.GetExecutorLossReason
 
GetExecutorLossReason$() - Constructor for class org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.GetExecutorLossReason$
 
getExecutorMemoryStatus() - Method in class org.apache.spark.SparkContext
Return a map from the block manager to the max memory available for caching and the remaining memory available for caching.
getExecutorMetrics(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions.Builder
repeated .org.apache.spark.status.protobuf.ExecutorMetrics executor_metrics = 2;
getExecutorMetrics(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions
repeated .org.apache.spark.status.protobuf.ExecutorMetrics executor_metrics = 2;
getExecutorMetrics(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributionsOrBuilder
repeated .org.apache.spark.status.protobuf.ExecutorMetrics executor_metrics = 2;
getExecutorMetricsBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions.Builder
repeated .org.apache.spark.status.protobuf.ExecutorMetrics executor_metrics = 2;
getExecutorMetricsBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions.Builder
repeated .org.apache.spark.status.protobuf.ExecutorMetrics executor_metrics = 2;
getExecutorMetricsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions.Builder
repeated .org.apache.spark.status.protobuf.ExecutorMetrics executor_metrics = 2;
getExecutorMetricsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions
repeated .org.apache.spark.status.protobuf.ExecutorMetrics executor_metrics = 2;
getExecutorMetricsCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributionsOrBuilder
repeated .org.apache.spark.status.protobuf.ExecutorMetrics executor_metrics = 2;
getExecutorMetricsDistributions() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
optional .org.apache.spark.status.protobuf.ExecutorMetricsDistributions executor_metrics_distributions = 52;
getExecutorMetricsDistributions() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
optional .org.apache.spark.status.protobuf.ExecutorMetricsDistributions executor_metrics_distributions = 52;
getExecutorMetricsDistributions() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
optional .org.apache.spark.status.protobuf.ExecutorMetricsDistributions executor_metrics_distributions = 52;
getExecutorMetricsDistributionsBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
optional .org.apache.spark.status.protobuf.ExecutorMetricsDistributions executor_metrics_distributions = 52;
getExecutorMetricsDistributionsOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
optional .org.apache.spark.status.protobuf.ExecutorMetricsDistributions executor_metrics_distributions = 52;
getExecutorMetricsDistributionsOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
optional .org.apache.spark.status.protobuf.ExecutorMetricsDistributions executor_metrics_distributions = 52;
getExecutorMetricsDistributionsOrBuilder() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
optional .org.apache.spark.status.protobuf.ExecutorMetricsDistributions executor_metrics_distributions = 52;
getExecutorMetricsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions.Builder
repeated .org.apache.spark.status.protobuf.ExecutorMetrics executor_metrics = 2;
getExecutorMetricsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions
repeated .org.apache.spark.status.protobuf.ExecutorMetrics executor_metrics = 2;
getExecutorMetricsList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributionsOrBuilder
repeated .org.apache.spark.status.protobuf.ExecutorMetrics executor_metrics = 2;
getExecutorMetricsOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions.Builder
repeated .org.apache.spark.status.protobuf.ExecutorMetrics executor_metrics = 2;
getExecutorMetricsOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions
repeated .org.apache.spark.status.protobuf.ExecutorMetrics executor_metrics = 2;
getExecutorMetricsOrBuilder(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributionsOrBuilder
repeated .org.apache.spark.status.protobuf.ExecutorMetrics executor_metrics = 2;
getExecutorMetricsOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions.Builder
repeated .org.apache.spark.status.protobuf.ExecutorMetrics executor_metrics = 2;
getExecutorMetricsOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions
repeated .org.apache.spark.status.protobuf.ExecutorMetrics executor_metrics = 2;
getExecutorMetricsOrBuilderList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributionsOrBuilder
repeated .org.apache.spark.status.protobuf.ExecutorMetrics executor_metrics = 2;
getExecutorResources() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo.Builder
Deprecated.
getExecutorResources() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo
Deprecated.
getExecutorResources() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfoOrBuilder
Deprecated.
getExecutorResourcesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo.Builder
 
getExecutorResourcesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo
 
getExecutorResourcesCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfoOrBuilder
map<string, .org.apache.spark.status.protobuf.ExecutorResourceRequest> executor_resources = 2;
getExecutorResourcesMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo.Builder
map<string, .org.apache.spark.status.protobuf.ExecutorResourceRequest> executor_resources = 2;
getExecutorResourcesMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo
map<string, .org.apache.spark.status.protobuf.ExecutorResourceRequest> executor_resources = 2;
getExecutorResourcesMap() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfoOrBuilder
map<string, .org.apache.spark.status.protobuf.ExecutorResourceRequest> executor_resources = 2;
getExecutorResourcesOrDefault(String, StoreTypes.ExecutorResourceRequest) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo.Builder
map<string, .org.apache.spark.status.protobuf.ExecutorResourceRequest> executor_resources = 2;
getExecutorResourcesOrDefault(String, StoreTypes.ExecutorResourceRequest) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo
map<string, .org.apache.spark.status.protobuf.ExecutorResourceRequest> executor_resources = 2;
getExecutorResourcesOrDefault(String, StoreTypes.ExecutorResourceRequest) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfoOrBuilder
map<string, .org.apache.spark.status.protobuf.ExecutorResourceRequest> executor_resources = 2;
getExecutorResourcesOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo.Builder
map<string, .org.apache.spark.status.protobuf.ExecutorResourceRequest> executor_resources = 2;
getExecutorResourcesOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo
map<string, .org.apache.spark.status.protobuf.ExecutorResourceRequest> executor_resources = 2;
getExecutorResourcesOrThrow(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfoOrBuilder
map<string, .org.apache.spark.status.protobuf.ExecutorResourceRequest> executor_resources = 2;
getExecutorRunTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double executor_run_time = 8;
getExecutorRunTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
double executor_run_time = 8;
getExecutorRunTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
double executor_run_time = 8;
getExecutorRunTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int64 executor_run_time = 16;
getExecutorRunTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
int64 executor_run_time = 16;
getExecutorRunTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
int64 executor_run_time = 16;
getExecutorRunTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 executor_run_time = 18;
getExecutorRunTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
int64 executor_run_time = 18;
getExecutorRunTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
int64 executor_run_time = 18;
getExecutorRunTime(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double executor_run_time = 5;
getExecutorRunTime(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
repeated double executor_run_time = 5;
getExecutorRunTime(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
repeated double executor_run_time = 5;
getExecutorRunTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
int64 executor_run_time = 3;
getExecutorRunTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics
int64 executor_run_time = 3;
getExecutorRunTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricsOrBuilder
int64 executor_run_time = 3;
getExecutorRunTimeCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double executor_run_time = 5;
getExecutorRunTimeCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
repeated double executor_run_time = 5;
getExecutorRunTimeCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
repeated double executor_run_time = 5;
getExecutorRunTimeList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double executor_run_time = 5;
getExecutorRunTimeList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
repeated double executor_run_time = 5;
getExecutorRunTimeList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
repeated double executor_run_time = 5;
getExecutors(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo.Builder
repeated string executors = 5;
getExecutors(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo
repeated string executors = 5;
getExecutors(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfoOrBuilder
repeated string executors = 5;
getExecutorsBytes(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo.Builder
repeated string executors = 5;
getExecutorsBytes(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo
repeated string executors = 5;
getExecutorsBytes(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfoOrBuilder
repeated string executors = 5;
getExecutorsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo.Builder
repeated string executors = 5;
getExecutorsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo
repeated string executors = 5;
getExecutorsCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfoOrBuilder
repeated string executors = 5;
getExecutorsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo.Builder
repeated string executors = 5;
getExecutorsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo
repeated string executors = 5;
getExecutorsList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfoOrBuilder
repeated string executors = 5;
getExecutorSummary() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
Deprecated.
getExecutorSummary() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
Deprecated.
getExecutorSummary() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
Deprecated.
getExecutorSummaryCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
 
getExecutorSummaryCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
 
getExecutorSummaryCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
map<string, .org.apache.spark.status.protobuf.ExecutorStageSummary> executor_summary = 46;
getExecutorSummaryMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
map<string, .org.apache.spark.status.protobuf.ExecutorStageSummary> executor_summary = 46;
getExecutorSummaryMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
map<string, .org.apache.spark.status.protobuf.ExecutorStageSummary> executor_summary = 46;
getExecutorSummaryMap() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
map<string, .org.apache.spark.status.protobuf.ExecutorStageSummary> executor_summary = 46;
getExecutorSummaryOrDefault(String, StoreTypes.ExecutorStageSummary) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
map<string, .org.apache.spark.status.protobuf.ExecutorStageSummary> executor_summary = 46;
getExecutorSummaryOrDefault(String, StoreTypes.ExecutorStageSummary) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
map<string, .org.apache.spark.status.protobuf.ExecutorStageSummary> executor_summary = 46;
getExecutorSummaryOrDefault(String, StoreTypes.ExecutorStageSummary) - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
map<string, .org.apache.spark.status.protobuf.ExecutorStageSummary> executor_summary = 46;
getExecutorSummaryOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
map<string, .org.apache.spark.status.protobuf.ExecutorStageSummary> executor_summary = 46;
getExecutorSummaryOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
map<string, .org.apache.spark.status.protobuf.ExecutorStageSummary> executor_summary = 46;
getExecutorSummaryOrThrow(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
map<string, .org.apache.spark.status.protobuf.ExecutorStageSummary> executor_summary = 46;
getFactorSize() - Method in interface org.apache.spark.ml.regression.FactorizationMachinesParams
 
getFailedTasks(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double failed_tasks = 3;
getFailedTasks(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
repeated double failed_tasks = 3;
getFailedTasks(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
repeated double failed_tasks = 3;
getFailedTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
int32 failed_tasks = 2;
getFailedTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary
int32 failed_tasks = 2;
getFailedTasks() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryOrBuilder
int32 failed_tasks = 2;
getFailedTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
int32 failed_tasks = 10;
getFailedTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
int32 failed_tasks = 10;
getFailedTasks() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
int32 failed_tasks = 10;
getFailedTasksCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double failed_tasks = 3;
getFailedTasksCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
repeated double failed_tasks = 3;
getFailedTasksCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
repeated double failed_tasks = 3;
getFailedTasksList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double failed_tasks = 3;
getFailedTasksList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
repeated double failed_tasks = 3;
getFailedTasksList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
repeated double failed_tasks = 3;
getFailureReason() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
optional string failure_reason = 13;
getFailureReason() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
optional string failure_reason = 13;
getFailureReason() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
optional string failure_reason = 13;
getFailureReasonBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
optional string failure_reason = 13;
getFailureReasonBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
optional string failure_reason = 13;
getFailureReasonBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
optional string failure_reason = 13;
getFamily() - Method in interface org.apache.spark.ml.classification.LogisticRegressionParams
 
getFamily() - Method in interface org.apache.spark.ml.regression.GeneralizedLinearRegressionBase
 
getFdr() - Method in interface org.apache.spark.ml.feature.SelectorParams
 
getFeatureIndex() - Method in interface org.apache.spark.ml.regression.IsotonicRegressionBase
 
getFeatureIndicesFromNames(StructField, String[]) - Static method in class org.apache.spark.ml.util.MetadataUtils
Takes a Vector column and a list of feature names, and returns the corresponding list of feature indices in the column, in order.
getFeatures() - Method in class org.apache.spark.ml.feature.LabeledPoint
 
getFeatures() - Method in class org.apache.spark.mllib.regression.LabeledPoint
 
getFeaturesAndLabels(RFormulaModel, Dataset<?>) - Static method in class org.apache.spark.ml.r.RWrapperUtils
Get the feature names and original labels from the schema of DataFrame transformed by RFormulaModel.
getFeaturesCol() - Method in interface org.apache.spark.ml.param.shared.HasFeaturesCol
 
getFeatureSubsetStrategy() - Method in interface org.apache.spark.ml.tree.TreeEnsembleParams
 
getFeatureType() - Method in interface org.apache.spark.ml.feature.UnivariateFeatureSelectorParams
 
getFetchWaitTime(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
repeated double fetch_wait_time = 5;
getFetchWaitTime(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions
repeated double fetch_wait_time = 5;
getFetchWaitTime(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributionsOrBuilder
repeated double fetch_wait_time = 5;
getFetchWaitTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics.Builder
int64 fetch_wait_time = 3;
getFetchWaitTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics
int64 fetch_wait_time = 3;
getFetchWaitTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricsOrBuilder
int64 fetch_wait_time = 3;
getFetchWaitTimeCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
repeated double fetch_wait_time = 5;
getFetchWaitTimeCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions
repeated double fetch_wait_time = 5;
getFetchWaitTimeCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributionsOrBuilder
repeated double fetch_wait_time = 5;
getFetchWaitTimeList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
repeated double fetch_wait_time = 5;
getFetchWaitTimeList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions
repeated double fetch_wait_time = 5;
getFetchWaitTimeList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributionsOrBuilder
repeated double fetch_wait_time = 5;
getField(String) - Method in class org.apache.spark.sql.Column
An expression that gets a field by name in a StructType.
getFileLength(File, SparkConf) - Static method in class org.apache.spark.util.Utils
Return the file length, if the file is compressed it returns the uncompressed file length.
getFileSegmentLocations(String, long, long, Configuration) - Static method in class org.apache.spark.streaming.util.HdfsUtils
Get the locations of the HDFS blocks containing the given file segment.
getFileSystemForPath(Path, Configuration) - Static method in class org.apache.spark.streaming.util.HdfsUtils
 
getFinalStorageLevel() - Method in interface org.apache.spark.ml.recommendation.ALSParams
 
getFinalValue() - Method in class org.apache.spark.partial.PartialResult
Blocking method to wait for and return the final value.
getFirstTaskLaunchedTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
optional int64 first_task_launched_time = 11;
getFirstTaskLaunchedTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
optional int64 first_task_launched_time = 11;
getFirstTaskLaunchedTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
optional int64 first_task_launched_time = 11;
getFitIntercept() - Method in interface org.apache.spark.ml.param.shared.HasFitIntercept
 
getFitLinear() - Method in interface org.apache.spark.ml.regression.FactorizationMachinesParams
 
getFloat(int) - Method in interface org.apache.spark.sql.Row
Returns the value at position i as a primitive float.
getFloat(int) - Method in class org.apache.spark.sql.vectorized.ArrowColumnVector
 
getFloat(int) - Method in class org.apache.spark.sql.vectorized.ColumnarArray
 
getFloat(int) - Method in class org.apache.spark.sql.vectorized.ColumnarBatchRow
 
getFloat(int) - Method in class org.apache.spark.sql.vectorized.ColumnarRow
 
getFloat(int) - Method in class org.apache.spark.sql.vectorized.ColumnVector
Returns the float type value for rowId.
getFloats(int, int) - Method in class org.apache.spark.sql.vectorized.ColumnVector
Gets float type values from [rowId, rowId + count).
getFoldCol() - Method in interface org.apache.spark.ml.tuning.CrossValidatorParams
 
getForceIndexLabel() - Method in interface org.apache.spark.ml.feature.RFormulaBase
 
getFormattedClassName(Object) - Static method in class org.apache.spark.util.Utils
Return the class name of the given object, removing all dollar signs
getFormattedDuration(JobData) - Static method in class org.apache.spark.ui.jobs.JobDataUtil
 
getFormattedSubmissionTime(JobData) - Static method in class org.apache.spark.ui.jobs.JobDataUtil
 
getFormula() - Method in interface org.apache.spark.ml.feature.RFormulaBase
 
getFpr() - Method in interface org.apache.spark.ml.feature.SelectorParams
 
getFromId() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationEdge.Builder
int32 from_id = 1;
getFromId() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationEdge
int32 from_id = 1;
getFromId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationEdgeOrBuilder
int32 from_id = 1;
getFromId() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphEdge.Builder
int64 from_id = 1;
getFromId() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphEdge
int64 from_id = 1;
getFromId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphEdgeOrBuilder
int64 from_id = 1;
getFunction(String) - Method in class org.apache.spark.sql.catalog.Catalog
Get the function with the specified name.
getFunction(String, String) - Method in class org.apache.spark.sql.catalog.Catalog
Get the function with the specified name in the specified database under the Hive Metastore.
getFwe() - Method in interface org.apache.spark.ml.feature.SelectorParams
 
getGaps() - Method in class org.apache.spark.ml.feature.RegexTokenizer
 
getGettingResultTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double getting_result_time = 13;
getGettingResultTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
double getting_result_time = 13;
getGettingResultTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
double getting_result_time = 13;
getGettingResultTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
int64 getting_result_time = 18;
getGettingResultTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
int64 getting_result_time = 18;
getGettingResultTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataOrBuilder
int64 getting_result_time = 18;
getGettingResultTime(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double getting_result_time = 10;
getGettingResultTime(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
repeated double getting_result_time = 10;
getGettingResultTime(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
repeated double getting_result_time = 10;
getGettingResultTimeCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double getting_result_time = 10;
getGettingResultTimeCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
repeated double getting_result_time = 10;
getGettingResultTimeCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
repeated double getting_result_time = 10;
getGettingResultTimeList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double getting_result_time = 10;
getGettingResultTimeList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
repeated double getting_result_time = 10;
getGettingResultTimeList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
repeated double getting_result_time = 10;
getGroups(String) - Method in interface org.apache.spark.security.GroupMappingServiceProvider
Get the groups the user belongs to.
getHadoopFileSystem(URI, Configuration) - Static method in class org.apache.spark.util.Utils
Return a Hadoop FileSystem with the scheme encoded in the given path.
getHadoopFileSystem(String, Configuration) - Static method in class org.apache.spark.util.Utils
Return a Hadoop FileSystem with the scheme encoded in the given path.
getHadoopProperties(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings hadoop_properties = 3;
getHadoopProperties(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
repeated .org.apache.spark.status.protobuf.PairStrings hadoop_properties = 3;
getHadoopProperties(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoOrBuilder
repeated .org.apache.spark.status.protobuf.PairStrings hadoop_properties = 3;
getHadoopPropertiesBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings hadoop_properties = 3;
getHadoopPropertiesBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings hadoop_properties = 3;
getHadoopPropertiesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings hadoop_properties = 3;
getHadoopPropertiesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
repeated .org.apache.spark.status.protobuf.PairStrings hadoop_properties = 3;
getHadoopPropertiesCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoOrBuilder
repeated .org.apache.spark.status.protobuf.PairStrings hadoop_properties = 3;
getHadoopPropertiesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings hadoop_properties = 3;
getHadoopPropertiesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
repeated .org.apache.spark.status.protobuf.PairStrings hadoop_properties = 3;
getHadoopPropertiesList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoOrBuilder
repeated .org.apache.spark.status.protobuf.PairStrings hadoop_properties = 3;
getHadoopPropertiesOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings hadoop_properties = 3;
getHadoopPropertiesOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
repeated .org.apache.spark.status.protobuf.PairStrings hadoop_properties = 3;
getHadoopPropertiesOrBuilder(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoOrBuilder
repeated .org.apache.spark.status.protobuf.PairStrings hadoop_properties = 3;
getHadoopPropertiesOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings hadoop_properties = 3;
getHadoopPropertiesOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
repeated .org.apache.spark.status.protobuf.PairStrings hadoop_properties = 3;
getHadoopPropertiesOrBuilderList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoOrBuilder
repeated .org.apache.spark.status.protobuf.PairStrings hadoop_properties = 3;
getHandleInvalid() - Method in interface org.apache.spark.ml.param.shared.HasHandleInvalid
 
getHasMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
bool has_metrics = 15;
getHasMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
bool has_metrics = 15;
getHasMetrics() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
bool has_metrics = 15;
getHeight(Row) - Static method in class org.apache.spark.ml.image.ImageSchema
Gets the height of the image
getHost() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
optional string host = 9;
getHost() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
optional string host = 9;
getHost() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataOrBuilder
optional string host = 9;
getHost() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
optional string host = 9;
getHost() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
optional string host = 9;
getHost() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
optional string host = 9;
getHostBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
optional string host = 9;
getHostBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
optional string host = 9;
getHostBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataOrBuilder
optional string host = 9;
getHostBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
optional string host = 9;
getHostBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
optional string host = 9;
getHostBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
optional string host = 9;
getHostPort() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
optional string host_port = 2;
getHostPort() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
optional string host_port = 2;
getHostPort() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
optional string host_port = 2;
getHostPort() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder
optional string host_port = 2;
getHostPort() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary
optional string host_port = 2;
getHostPort() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryOrBuilder
optional string host_port = 2;
getHostPort() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
optional string host_port = 3;
getHostPort() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData
optional string host_port = 3;
getHostPort() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamBlockDataOrBuilder
optional string host_port = 3;
getHostPortBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
optional string host_port = 2;
getHostPortBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
optional string host_port = 2;
getHostPortBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
optional string host_port = 2;
getHostPortBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder
optional string host_port = 2;
getHostPortBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary
optional string host_port = 2;
getHostPortBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryOrBuilder
optional string host_port = 2;
getHostPortBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
optional string host_port = 3;
getHostPortBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData
optional string host_port = 3;
getHostPortBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamBlockDataOrBuilder
optional string host_port = 3;
getId() - Method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo.Builder
int64 id = 1;
getId() - Method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo
int64 id = 1;
getId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.AccumulableInfoOrBuilder
int64 id = 1;
getId() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
optional string id = 1;
getId() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo
optional string id = 1;
getId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoOrBuilder
optional string id = 1;
getId() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
optional string id = 1;
getId() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
optional string id = 1;
getId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
optional string id = 1;
getId() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder
optional string id = 1;
getId() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary
optional string id = 1;
getId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryOrBuilder
optional string id = 1;
getId() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
optional string id = 1;
getId() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper
optional string id = 1;
getId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapperOrBuilder
optional string id = 1;
getId() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode.Builder
int32 id = 1;
getId() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode
int32 id = 1;
getId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationNodeOrBuilder
int32 id = 1;
getId() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
int32 id = 1;
getId() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo
int32 id = 1;
getId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoOrBuilder
int32 id = 1;
getId() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo.Builder
int32 id = 1;
getId() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo
int32 id = 1;
getId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfoOrBuilder
int32 id = 1;
getId() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
int64 id = 1;
getId() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper
int64 id = 1;
getId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapperOrBuilder
int64 id = 1;
getId() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
int64 id = 1;
getId() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode
int64 id = 1;
getId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeOrBuilder
int64 id = 1;
getId() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData.Builder
optional string id = 2;
getId() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData
optional string id = 2;
getId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryDataOrBuilder
optional string id = 2;
getId() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
optional string id = 1;
getId() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
optional string id = 1;
getId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
optional string id = 1;
getIdBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
optional string id = 1;
getIdBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo
optional string id = 1;
getIdBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoOrBuilder
optional string id = 1;
getIdBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
optional string id = 1;
getIdBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
optional string id = 1;
getIdBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
optional string id = 1;
getIdBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder
optional string id = 1;
getIdBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary
optional string id = 1;
getIdBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryOrBuilder
optional string id = 1;
getIdBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
optional string id = 1;
getIdBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper
optional string id = 1;
getIdBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapperOrBuilder
optional string id = 1;
getIdBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData.Builder
optional string id = 2;
getIdBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData
optional string id = 2;
getIdBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryDataOrBuilder
optional string id = 2;
getIdBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
optional string id = 1;
getIdBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
optional string id = 1;
getIdBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
optional string id = 1;
getImplicitPrefs() - Method in interface org.apache.spark.ml.recommendation.ALSParams
 
getImpurity() - Method in interface org.apache.spark.ml.tree.HasVarianceImpurity
 
getImpurity() - Method in interface org.apache.spark.ml.tree.TreeClassifierParams
 
getImpurity() - Method in class org.apache.spark.mllib.tree.configuration.Strategy
 
getIncomingEdges(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationEdge incoming_edges = 4;
getIncomingEdges(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper
repeated .org.apache.spark.status.protobuf.RDDOperationEdge incoming_edges = 4;
getIncomingEdges(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapperOrBuilder
repeated .org.apache.spark.status.protobuf.RDDOperationEdge incoming_edges = 4;
getIncomingEdgesBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationEdge incoming_edges = 4;
getIncomingEdgesBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationEdge incoming_edges = 4;
getIncomingEdgesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationEdge incoming_edges = 4;
getIncomingEdgesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper
repeated .org.apache.spark.status.protobuf.RDDOperationEdge incoming_edges = 4;
getIncomingEdgesCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapperOrBuilder
repeated .org.apache.spark.status.protobuf.RDDOperationEdge incoming_edges = 4;
getIncomingEdgesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationEdge incoming_edges = 4;
getIncomingEdgesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper
repeated .org.apache.spark.status.protobuf.RDDOperationEdge incoming_edges = 4;
getIncomingEdgesList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapperOrBuilder
repeated .org.apache.spark.status.protobuf.RDDOperationEdge incoming_edges = 4;
getIncomingEdgesOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationEdge incoming_edges = 4;
getIncomingEdgesOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper
repeated .org.apache.spark.status.protobuf.RDDOperationEdge incoming_edges = 4;
getIncomingEdgesOrBuilder(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapperOrBuilder
repeated .org.apache.spark.status.protobuf.RDDOperationEdge incoming_edges = 4;
getIncomingEdgesOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationEdge incoming_edges = 4;
getIncomingEdgesOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper
repeated .org.apache.spark.status.protobuf.RDDOperationEdge incoming_edges = 4;
getIncomingEdgesOrBuilderList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapperOrBuilder
repeated .org.apache.spark.status.protobuf.RDDOperationEdge incoming_edges = 4;
getIndex() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
int32 index = 2;
getIndex() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
int32 index = 2;
getIndex() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataOrBuilder
int32 index = 2;
getIndex() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int32 index = 2;
getIndex() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
int32 index = 2;
getIndex() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
int32 index = 2;
getIndices() - Method in class org.apache.spark.ml.feature.VectorSlicer
 
getInfo() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoWrapper.Builder
.org.apache.spark.status.protobuf.ApplicationEnvironmentInfo info = 1;
getInfo() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoWrapper
.org.apache.spark.status.protobuf.ApplicationEnvironmentInfo info = 1;
getInfo() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoWrapperOrBuilder
.org.apache.spark.status.protobuf.ApplicationEnvironmentInfo info = 1;
getInfo() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoWrapper.Builder
.org.apache.spark.status.protobuf.ApplicationInfo info = 1;
getInfo() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoWrapper
.org.apache.spark.status.protobuf.ApplicationInfo info = 1;
getInfo() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoWrapperOrBuilder
.org.apache.spark.status.protobuf.ApplicationInfo info = 1;
getInfo() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper.Builder
.org.apache.spark.status.protobuf.ExecutorStageSummary info = 4;
getInfo() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper
.org.apache.spark.status.protobuf.ExecutorStageSummary info = 4;
getInfo() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapperOrBuilder
.org.apache.spark.status.protobuf.ExecutorStageSummary info = 4;
getInfo() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryWrapper.Builder
.org.apache.spark.status.protobuf.ExecutorSummary info = 1;
getInfo() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryWrapper
.org.apache.spark.status.protobuf.ExecutorSummary info = 1;
getInfo() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryWrapperOrBuilder
.org.apache.spark.status.protobuf.ExecutorSummary info = 1;
getInfo() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper.Builder
.org.apache.spark.status.protobuf.JobData info = 1;
getInfo() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper
.org.apache.spark.status.protobuf.JobData info = 1;
getInfo() - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataWrapperOrBuilder
.org.apache.spark.status.protobuf.JobData info = 1;
getInfo() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryWrapper.Builder
.org.apache.spark.status.protobuf.ProcessSummary info = 1;
getInfo() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryWrapper
.org.apache.spark.status.protobuf.ProcessSummary info = 1;
getInfo() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryWrapperOrBuilder
.org.apache.spark.status.protobuf.ProcessSummary info = 1;
getInfo() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoWrapper.Builder
.org.apache.spark.status.protobuf.RDDStorageInfo info = 1;
getInfo() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoWrapper
.org.apache.spark.status.protobuf.RDDStorageInfo info = 1;
getInfo() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoWrapperOrBuilder
.org.apache.spark.status.protobuf.RDDStorageInfo info = 1;
getInfo() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryWrapper.Builder
.org.apache.spark.status.protobuf.SpeculationStageSummary info = 3;
getInfo() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryWrapper
.org.apache.spark.status.protobuf.SpeculationStageSummary info = 3;
getInfo() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryWrapperOrBuilder
.org.apache.spark.status.protobuf.SpeculationStageSummary info = 3;
getInfo() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper.Builder
.org.apache.spark.status.protobuf.StageData info = 1;
getInfo() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper
.org.apache.spark.status.protobuf.StageData info = 1;
getInfo() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataWrapperOrBuilder
.org.apache.spark.status.protobuf.StageData info = 1;
getInfoBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoWrapper.Builder
.org.apache.spark.status.protobuf.ApplicationEnvironmentInfo info = 1;
getInfoBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoWrapper.Builder
.org.apache.spark.status.protobuf.ApplicationInfo info = 1;
getInfoBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper.Builder
.org.apache.spark.status.protobuf.ExecutorStageSummary info = 4;
getInfoBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryWrapper.Builder
.org.apache.spark.status.protobuf.ExecutorSummary info = 1;
getInfoBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper.Builder
.org.apache.spark.status.protobuf.JobData info = 1;
getInfoBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryWrapper.Builder
.org.apache.spark.status.protobuf.ProcessSummary info = 1;
getInfoBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoWrapper.Builder
.org.apache.spark.status.protobuf.RDDStorageInfo info = 1;
getInfoBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryWrapper.Builder
.org.apache.spark.status.protobuf.SpeculationStageSummary info = 3;
getInfoBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper.Builder
.org.apache.spark.status.protobuf.StageData info = 1;
getInfoOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoWrapper.Builder
.org.apache.spark.status.protobuf.ApplicationEnvironmentInfo info = 1;
getInfoOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoWrapper
.org.apache.spark.status.protobuf.ApplicationEnvironmentInfo info = 1;
getInfoOrBuilder() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoWrapperOrBuilder
.org.apache.spark.status.protobuf.ApplicationEnvironmentInfo info = 1;
getInfoOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoWrapper.Builder
.org.apache.spark.status.protobuf.ApplicationInfo info = 1;
getInfoOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoWrapper
.org.apache.spark.status.protobuf.ApplicationInfo info = 1;
getInfoOrBuilder() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoWrapperOrBuilder
.org.apache.spark.status.protobuf.ApplicationInfo info = 1;
getInfoOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper.Builder
.org.apache.spark.status.protobuf.ExecutorStageSummary info = 4;
getInfoOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper
.org.apache.spark.status.protobuf.ExecutorStageSummary info = 4;
getInfoOrBuilder() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapperOrBuilder
.org.apache.spark.status.protobuf.ExecutorStageSummary info = 4;
getInfoOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryWrapper.Builder
.org.apache.spark.status.protobuf.ExecutorSummary info = 1;
getInfoOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryWrapper
.org.apache.spark.status.protobuf.ExecutorSummary info = 1;
getInfoOrBuilder() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryWrapperOrBuilder
.org.apache.spark.status.protobuf.ExecutorSummary info = 1;
getInfoOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper.Builder
.org.apache.spark.status.protobuf.JobData info = 1;
getInfoOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper
.org.apache.spark.status.protobuf.JobData info = 1;
getInfoOrBuilder() - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataWrapperOrBuilder
.org.apache.spark.status.protobuf.JobData info = 1;
getInfoOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryWrapper.Builder
.org.apache.spark.status.protobuf.ProcessSummary info = 1;
getInfoOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryWrapper
.org.apache.spark.status.protobuf.ProcessSummary info = 1;
getInfoOrBuilder() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryWrapperOrBuilder
.org.apache.spark.status.protobuf.ProcessSummary info = 1;
getInfoOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoWrapper.Builder
.org.apache.spark.status.protobuf.RDDStorageInfo info = 1;
getInfoOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoWrapper
.org.apache.spark.status.protobuf.RDDStorageInfo info = 1;
getInfoOrBuilder() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoWrapperOrBuilder
.org.apache.spark.status.protobuf.RDDStorageInfo info = 1;
getInfoOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryWrapper.Builder
.org.apache.spark.status.protobuf.SpeculationStageSummary info = 3;
getInfoOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryWrapper
.org.apache.spark.status.protobuf.SpeculationStageSummary info = 3;
getInfoOrBuilder() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryWrapperOrBuilder
.org.apache.spark.status.protobuf.SpeculationStageSummary info = 3;
getInfoOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper.Builder
.org.apache.spark.status.protobuf.StageData info = 1;
getInfoOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper
.org.apache.spark.status.protobuf.StageData info = 1;
getInfoOrBuilder() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataWrapperOrBuilder
.org.apache.spark.status.protobuf.StageData info = 1;
getInitializationMode() - Method in class org.apache.spark.mllib.clustering.KMeans
The initialization algorithm.
getInitializationSteps() - Method in class org.apache.spark.mllib.clustering.KMeans
Number of steps for the k-means|| initialization mode
getInitialModel() - Method in class org.apache.spark.mllib.clustering.GaussianMixture
Return the user supplied initial GMM, if supplied
getInitialTargetExecutorNumber(SparkConf, int) - Static method in class org.apache.spark.scheduler.cluster.SchedulerBackendUtils
Getting the initial target number of executors depends on whether dynamic allocation is enabled.
getInitialWeights() - Method in interface org.apache.spark.ml.classification.MultilayerPerceptronParams
 
getInitMode() - Method in interface org.apache.spark.ml.clustering.KMeansParams
 
getInitMode() - Method in interface org.apache.spark.ml.clustering.PowerIterationClusteringParams
 
getInitStd() - Method in interface org.apache.spark.ml.regression.FactorizationMachinesParams
 
getInitSteps() - Method in interface org.apache.spark.ml.clustering.KMeansParams
 
getInOutCols() - Method in interface org.apache.spark.ml.feature.ImputerParams
Returns the input and output column names corresponding in pair.
getInOutCols() - Method in interface org.apache.spark.ml.feature.OneHotEncoderBase
Returns the input and output column names corresponding in pair.
getInOutCols() - Method in interface org.apache.spark.ml.feature.StringIndexerBase
Returns the input and output column names corresponding in pair.
getInputBytes(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double input_bytes = 6;
getInputBytes(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
repeated double input_bytes = 6;
getInputBytes(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
repeated double input_bytes = 6;
getInputBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
int64 input_bytes = 5;
getInputBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary
int64 input_bytes = 5;
getInputBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryOrBuilder
int64 input_bytes = 5;
getInputBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int64 input_bytes = 24;
getInputBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
int64 input_bytes = 24;
getInputBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
int64 input_bytes = 24;
getInputBytesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double input_bytes = 6;
getInputBytesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
repeated double input_bytes = 6;
getInputBytesCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
repeated double input_bytes = 6;
getInputBytesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double input_bytes = 6;
getInputBytesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
repeated double input_bytes = 6;
getInputBytesList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
repeated double input_bytes = 6;
getInputBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 input_bytes_read = 26;
getInputBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
int64 input_bytes_read = 26;
getInputBytesRead() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
int64 input_bytes_read = 26;
getInputCol() - Method in interface org.apache.spark.ml.param.shared.HasInputCol
 
getInputCols() - Method in interface org.apache.spark.ml.param.shared.HasInputCols
 
getInputFilePath() - Static method in class org.apache.spark.rdd.InputFileBlockHolder
Returns the holding file name or empty string if it is unknown.
getInputMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
.org.apache.spark.status.protobuf.InputMetricDistributions input_metrics = 15;
getInputMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
.org.apache.spark.status.protobuf.InputMetricDistributions input_metrics = 15;
getInputMetrics() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
.org.apache.spark.status.protobuf.InputMetricDistributions input_metrics = 15;
getInputMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
.org.apache.spark.status.protobuf.InputMetrics input_metrics = 11;
getInputMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics
.org.apache.spark.status.protobuf.InputMetrics input_metrics = 11;
getInputMetrics() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricsOrBuilder
.org.apache.spark.status.protobuf.InputMetrics input_metrics = 11;
getInputMetricsBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
.org.apache.spark.status.protobuf.InputMetricDistributions input_metrics = 15;
getInputMetricsBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
.org.apache.spark.status.protobuf.InputMetrics input_metrics = 11;
getInputMetricsOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
.org.apache.spark.status.protobuf.InputMetricDistributions input_metrics = 15;
getInputMetricsOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
.org.apache.spark.status.protobuf.InputMetricDistributions input_metrics = 15;
getInputMetricsOrBuilder() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
.org.apache.spark.status.protobuf.InputMetricDistributions input_metrics = 15;
getInputMetricsOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
.org.apache.spark.status.protobuf.InputMetrics input_metrics = 11;
getInputMetricsOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics
.org.apache.spark.status.protobuf.InputMetrics input_metrics = 11;
getInputMetricsOrBuilder() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricsOrBuilder
.org.apache.spark.status.protobuf.InputMetrics input_metrics = 11;
getInputRecords(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double input_records = 7;
getInputRecords(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
repeated double input_records = 7;
getInputRecords(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
repeated double input_records = 7;
getInputRecords() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
int64 input_records = 6;
getInputRecords() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary
int64 input_records = 6;
getInputRecords() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryOrBuilder
int64 input_records = 6;
getInputRecords() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int64 input_records = 25;
getInputRecords() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
int64 input_records = 25;
getInputRecords() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
int64 input_records = 25;
getInputRecordsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double input_records = 7;
getInputRecordsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
repeated double input_records = 7;
getInputRecordsCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
repeated double input_records = 7;
getInputRecordsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double input_records = 7;
getInputRecordsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
repeated double input_records = 7;
getInputRecordsList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
repeated double input_records = 7;
getInputRecordsRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 input_records_read = 27;
getInputRecordsRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
int64 input_records_read = 27;
getInputRecordsRead() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
int64 input_records_read = 27;
getInputRowsPerSecond() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
double input_rows_per_second = 6;
getInputRowsPerSecond() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress
double input_rows_per_second = 6;
getInputRowsPerSecond() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SourceProgressOrBuilder
double input_rows_per_second = 6;
getInputStream(String, Configuration) - Static method in class org.apache.spark.streaming.util.HdfsUtils
 
getInstant(int) - Method in interface org.apache.spark.sql.Row
Returns the value at position i of date type as java.time.Instant.
getInt(String, int) - Method in class org.apache.spark.SparkConf
Get a parameter as an integer, falling back to a default if not set
getInt(int) - Method in interface org.apache.spark.sql.Row
Returns the value at position i as a primitive int.
getInt(String, int) - Method in class org.apache.spark.sql.util.CaseInsensitiveStringMap
Returns the integer value to which the specified key is mapped, or defaultValue if there is no mapping for the key.
getInt(int) - Method in class org.apache.spark.sql.vectorized.ArrowColumnVector
 
getInt(int) - Method in class org.apache.spark.sql.vectorized.ColumnarArray
 
getInt(int) - Method in class org.apache.spark.sql.vectorized.ColumnarBatchRow
 
getInt(int) - Method in class org.apache.spark.sql.vectorized.ColumnarRow
 
getInt(int) - Method in class org.apache.spark.sql.vectorized.ColumnVector
Returns the int type value for rowId.
getIntermediateStorageLevel() - Method in interface org.apache.spark.ml.recommendation.ALSParams
 
getInterval(int) - Method in class org.apache.spark.sql.vectorized.ColumnarArray
 
getInterval(int) - Method in class org.apache.spark.sql.vectorized.ColumnarBatchRow
 
getInterval(int) - Method in class org.apache.spark.sql.vectorized.ColumnarRow
 
getInterval(int) - Method in class org.apache.spark.sql.vectorized.ColumnVector
Returns the calendar interval type value for rowId.
getInts(int, int) - Method in class org.apache.spark.sql.vectorized.ColumnVector
Gets int type values from [rowId, rowId + count).
getInverse() - Method in class org.apache.spark.ml.feature.DCT
 
getIsActive() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
bool is_active = 3;
getIsActive() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
bool is_active = 3;
getIsActive() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
bool is_active = 3;
getIsActive() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder
bool is_active = 3;
getIsActive() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary
bool is_active = 3;
getIsActive() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryOrBuilder
bool is_active = 3;
getIsActive() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData.Builder
bool is_active = 4;
getIsActive() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData
bool is_active = 4;
getIsActive() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryDataOrBuilder
bool is_active = 4;
getIsBlacklisted() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
bool is_blacklisted = 18;
getIsBlacklisted() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
bool is_blacklisted = 18;
getIsBlacklisted() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
bool is_blacklisted = 18;
getIsBlacklistedForStage() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
bool is_blacklisted_for_stage = 15;
getIsBlacklistedForStage() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary
bool is_blacklisted_for_stage = 15;
getIsBlacklistedForStage() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryOrBuilder
bool is_blacklisted_for_stage = 15;
getIsExcluded() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
bool is_excluded = 30;
getIsExcluded() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
bool is_excluded = 30;
getIsExcluded() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
bool is_excluded = 30;
getIsExcludedForStage() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
bool is_excluded_for_stage = 17;
getIsExcludedForStage() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary
bool is_excluded_for_stage = 17;
getIsExcludedForStage() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryOrBuilder
bool is_excluded_for_stage = 17;
getIsExperiment() - Method in class org.apache.spark.mllib.stat.test.BinarySample
 
getIsotonic() - Method in interface org.apache.spark.ml.regression.IsotonicRegressionBase
 
getIsShufflePushEnabled() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
bool is_shuffle_push_enabled = 63;
getIsShufflePushEnabled() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
bool is_shuffle_push_enabled = 63;
getIsShufflePushEnabled() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
bool is_shuffle_push_enabled = 63;
getItem(Object) - Method in class org.apache.spark.sql.Column
An expression that gets an item at position ordinal out of an array, or gets a value by key key in a MapType.
getItemCol() - Method in interface org.apache.spark.ml.recommendation.ALSModelParams
 
getItemsCol() - Method in interface org.apache.spark.ml.fpm.FPGrowthParams
 
getIteratorSize(Iterator<?>) - Static method in class org.apache.spark.util.Utils
Counts the number of elements of an iterator.
getIteratorZipWithIndex(Iterator<T>, long) - Static method in class org.apache.spark.util.Utils
Generate a zipWithIndex iterator, avoid index value overflowing problem in scala's zipWithIndex
getIvyProperties() - Static method in class org.apache.spark.util.DependencyUtils
 
getJavaHome() - Method in class org.apache.spark.status.protobuf.StoreTypes.RuntimeInfo.Builder
optional string java_home = 2;
getJavaHome() - Method in class org.apache.spark.status.protobuf.StoreTypes.RuntimeInfo
optional string java_home = 2;
getJavaHome() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RuntimeInfoOrBuilder
optional string java_home = 2;
getJavaHomeBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.RuntimeInfo.Builder
optional string java_home = 2;
getJavaHomeBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.RuntimeInfo
optional string java_home = 2;
getJavaHomeBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RuntimeInfoOrBuilder
optional string java_home = 2;
getJavaMap(int) - Method in interface org.apache.spark.sql.Row
Returns the value at position i of array type as a java.util.Map.
getJavaSparkContext(SparkSession) - Static method in class org.apache.spark.sql.api.r.SQLUtils
 
getJavaVersion() - Method in class org.apache.spark.status.protobuf.StoreTypes.RuntimeInfo.Builder
optional string java_version = 1;
getJavaVersion() - Method in class org.apache.spark.status.protobuf.StoreTypes.RuntimeInfo
optional string java_version = 1;
getJavaVersion() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RuntimeInfoOrBuilder
optional string java_version = 1;
getJavaVersionBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.RuntimeInfo.Builder
optional string java_version = 1;
getJavaVersionBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.RuntimeInfo
optional string java_version = 1;
getJavaVersionBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RuntimeInfoOrBuilder
optional string java_version = 1;
getJDBCType(DataType) - Method in class org.apache.spark.sql.jdbc.AggregatedDialect
 
getJDBCType(DataType) - Static method in class org.apache.spark.sql.jdbc.DB2Dialect
 
getJDBCType(DataType) - Static method in class org.apache.spark.sql.jdbc.DerbyDialect
 
getJDBCType(DataType) - Static method in class org.apache.spark.sql.jdbc.H2Dialect
 
getJDBCType(DataType) - Method in class org.apache.spark.sql.jdbc.JdbcDialect
Retrieve the jdbc / sql type for a given datatype.
getJDBCType(DataType) - Static method in class org.apache.spark.sql.jdbc.MsSqlServerDialect
 
getJDBCType(DataType) - Static method in class org.apache.spark.sql.jdbc.MySQLDialect
 
getJDBCType(DataType) - Static method in class org.apache.spark.sql.jdbc.NoopDialect
 
getJDBCType(DataType) - Static method in class org.apache.spark.sql.jdbc.OracleDialect
 
getJDBCType(DataType) - Static method in class org.apache.spark.sql.jdbc.PostgresDialect
 
getJDBCType(DataType) - Static method in class org.apache.spark.sql.jdbc.TeradataDialect
 
getJobGroup() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
optional string job_group = 7;
getJobGroup() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
optional string job_group = 7;
getJobGroup() - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataOrBuilder
optional string job_group = 7;
getJobGroupBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
optional string job_group = 7;
getJobGroupBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
optional string job_group = 7;
getJobGroupBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataOrBuilder
optional string job_group = 7;
getJobId() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
All IDs are int64 for extendability, even when they are currently int32 in Spark.
getJobId() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
All IDs are int64 for extendability, even when they are currently int32 in Spark.
getJobId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataOrBuilder
All IDs are int64 for extendability, even when they are currently int32 in Spark.
getJobIds(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper.Builder
repeated int64 job_ids = 2;
getJobIds(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper
repeated int64 job_ids = 2;
getJobIds(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataWrapperOrBuilder
repeated int64 job_ids = 2;
getJobIdsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper.Builder
repeated int64 job_ids = 2;
getJobIdsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper
repeated int64 job_ids = 2;
getJobIdsCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataWrapperOrBuilder
repeated int64 job_ids = 2;
getJobIdsForGroup(String) - Method in class org.apache.spark.api.java.JavaSparkStatusTracker
Return a list of all known jobs in a particular job group.
getJobIdsForGroup(String) - Method in class org.apache.spark.SparkStatusTracker
Return a list of all known jobs in a particular job group.
getJobIdsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper.Builder
repeated int64 job_ids = 2;
getJobIdsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper
repeated int64 job_ids = 2;
getJobIdsList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataWrapperOrBuilder
repeated int64 job_ids = 2;
getJobInfo(int) - Method in class org.apache.spark.api.java.JavaSparkStatusTracker
Returns job information, or null if the job info could not be found or was garbage collected.
getJobInfo(int) - Method in class org.apache.spark.SparkStatusTracker
Returns job information, or None if the job info could not be found or was garbage collected.
getJobs() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
Deprecated.
getJobs() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
Deprecated.
getJobs() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
Deprecated.
getJobsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
 
getJobsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
 
getJobsCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
map<int64, .org.apache.spark.status.protobuf.JobExecutionStatus> jobs = 11;
getJobsMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
map<int64, .org.apache.spark.status.protobuf.JobExecutionStatus> jobs = 11;
getJobsMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
map<int64, .org.apache.spark.status.protobuf.JobExecutionStatus> jobs = 11;
getJobsMap() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
map<int64, .org.apache.spark.status.protobuf.JobExecutionStatus> jobs = 11;
getJobsOrDefault(long, StoreTypes.JobExecutionStatus) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
map<int64, .org.apache.spark.status.protobuf.JobExecutionStatus> jobs = 11;
getJobsOrDefault(long, StoreTypes.JobExecutionStatus) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
map<int64, .org.apache.spark.status.protobuf.JobExecutionStatus> jobs = 11;
getJobsOrDefault(long, StoreTypes.JobExecutionStatus) - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
map<int64, .org.apache.spark.status.protobuf.JobExecutionStatus> jobs = 11;
getJobsOrThrow(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
map<int64, .org.apache.spark.status.protobuf.JobExecutionStatus> jobs = 11;
getJobsOrThrow(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
map<int64, .org.apache.spark.status.protobuf.JobExecutionStatus> jobs = 11;
getJobsOrThrow(long) - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
map<int64, .org.apache.spark.status.protobuf.JobExecutionStatus> jobs = 11;
getJobsValue() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
Deprecated.
getJobsValue() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
Deprecated.
getJobsValue() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
Deprecated.
getJobsValueMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
map<int64, .org.apache.spark.status.protobuf.JobExecutionStatus> jobs = 11;
getJobsValueMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
map<int64, .org.apache.spark.status.protobuf.JobExecutionStatus> jobs = 11;
getJobsValueMap() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
map<int64, .org.apache.spark.status.protobuf.JobExecutionStatus> jobs = 11;
getJobsValueOrDefault(long, int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
map<int64, .org.apache.spark.status.protobuf.JobExecutionStatus> jobs = 11;
getJobsValueOrDefault(long, int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
map<int64, .org.apache.spark.status.protobuf.JobExecutionStatus> jobs = 11;
getJobsValueOrDefault(long, int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
map<int64, .org.apache.spark.status.protobuf.JobExecutionStatus> jobs = 11;
getJobsValueOrThrow(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
map<int64, .org.apache.spark.status.protobuf.JobExecutionStatus> jobs = 11;
getJobsValueOrThrow(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
map<int64, .org.apache.spark.status.protobuf.JobExecutionStatus> jobs = 11;
getJobsValueOrThrow(long) - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
map<int64, .org.apache.spark.status.protobuf.JobExecutionStatus> jobs = 11;
getJvmGcTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double jvm_gc_time = 11;
getJvmGcTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
double jvm_gc_time = 11;
getJvmGcTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
double jvm_gc_time = 11;
getJvmGcTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int64 jvm_gc_time = 19;
getJvmGcTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
int64 jvm_gc_time = 19;
getJvmGcTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
int64 jvm_gc_time = 19;
getJvmGcTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 jvm_gc_time = 21;
getJvmGcTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
int64 jvm_gc_time = 21;
getJvmGcTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
int64 jvm_gc_time = 21;
getJvmGcTime(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double jvm_gc_time = 8;
getJvmGcTime(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
repeated double jvm_gc_time = 8;
getJvmGcTime(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
repeated double jvm_gc_time = 8;
getJvmGcTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
int64 jvm_gc_time = 6;
getJvmGcTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics
int64 jvm_gc_time = 6;
getJvmGcTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricsOrBuilder
int64 jvm_gc_time = 6;
getJvmGcTimeCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double jvm_gc_time = 8;
getJvmGcTimeCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
repeated double jvm_gc_time = 8;
getJvmGcTimeCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
repeated double jvm_gc_time = 8;
getJvmGcTimeList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double jvm_gc_time = 8;
getJvmGcTimeList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
repeated double jvm_gc_time = 8;
getJvmGcTimeList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
repeated double jvm_gc_time = 8;
getK() - Method in interface org.apache.spark.ml.clustering.BisectingKMeansParams
 
getK() - Method in interface org.apache.spark.ml.clustering.GaussianMixtureParams
 
getK() - Method in interface org.apache.spark.ml.clustering.KMeansParams
 
getK() - Method in interface org.apache.spark.ml.clustering.LDAParams
 
getK() - Method in interface org.apache.spark.ml.clustering.PowerIterationClusteringParams
 
getK() - Method in class org.apache.spark.ml.evaluation.RankingEvaluator
 
getK() - Method in interface org.apache.spark.ml.feature.PCAParams
 
getK() - Method in class org.apache.spark.mllib.clustering.BisectingKMeans
Gets the desired number of leaf clusters.
getK() - Method in class org.apache.spark.mllib.clustering.GaussianMixture
Return the number of Gaussians in the mixture model
getK() - Method in class org.apache.spark.mllib.clustering.KMeans
Number of clusters to create (k).
getK() - Method in class org.apache.spark.mllib.clustering.LDA
Number of topics to infer, i.e., the number of soft cluster centers.
getKappa() - Method in class org.apache.spark.mllib.clustering.OnlineLDAOptimizer
Learning rate: exponential decay rate
getKeepLastCheckpoint() - Method in interface org.apache.spark.ml.clustering.LDAParams
 
getKeepLastCheckpoint() - Method in class org.apache.spark.mllib.clustering.EMLDAOptimizer
If using checkpointing, this indicates whether to keep the last checkpoint (vs clean up).
getKeytabJaasParams(String, String, String) - Static method in class org.apache.spark.kafka010.KafkaTokenUtil
 
getKilledTasks(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double killed_tasks = 5;
getKilledTasks(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
repeated double killed_tasks = 5;
getKilledTasks(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
repeated double killed_tasks = 5;
getKilledTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
int32 killed_tasks = 4;
getKilledTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary
int32 killed_tasks = 4;
getKilledTasks() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryOrBuilder
int32 killed_tasks = 4;
getKilledTasksCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double killed_tasks = 5;
getKilledTasksCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
repeated double killed_tasks = 5;
getKilledTasksCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
repeated double killed_tasks = 5;
getKilledTasksList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double killed_tasks = 5;
getKilledTasksList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
repeated double killed_tasks = 5;
getKilledTasksList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
repeated double killed_tasks = 5;
getKilledTasksSummary() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
Deprecated.
getKilledTasksSummary() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
Deprecated.
getKilledTasksSummary() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
Deprecated.
getKilledTasksSummaryCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
 
getKilledTasksSummaryCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
 
getKilledTasksSummaryCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
map<string, int32> killed_tasks_summary = 48;
getKilledTasksSummaryMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
map<string, int32> killed_tasks_summary = 48;
getKilledTasksSummaryMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
map<string, int32> killed_tasks_summary = 48;
getKilledTasksSummaryMap() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
map<string, int32> killed_tasks_summary = 48;
getKilledTasksSummaryOrDefault(String, int) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
map<string, int32> killed_tasks_summary = 48;
getKilledTasksSummaryOrDefault(String, int) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
map<string, int32> killed_tasks_summary = 48;
getKilledTasksSummaryOrDefault(String, int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
map<string, int32> killed_tasks_summary = 48;
getKilledTasksSummaryOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
map<string, int32> killed_tasks_summary = 48;
getKilledTasksSummaryOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
map<string, int32> killed_tasks_summary = 48;
getKilledTasksSummaryOrThrow(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
map<string, int32> killed_tasks_summary = 48;
getKillTasksSummary() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
Deprecated.
getKillTasksSummary() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
Deprecated.
getKillTasksSummary() - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataOrBuilder
Deprecated.
getKillTasksSummaryCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
 
getKillTasksSummaryCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
 
getKillTasksSummaryCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataOrBuilder
map<string, int32> kill_tasks_summary = 20;
getKillTasksSummaryMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
map<string, int32> kill_tasks_summary = 20;
getKillTasksSummaryMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
map<string, int32> kill_tasks_summary = 20;
getKillTasksSummaryMap() - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataOrBuilder
map<string, int32> kill_tasks_summary = 20;
getKillTasksSummaryOrDefault(String, int) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
map<string, int32> kill_tasks_summary = 20;
getKillTasksSummaryOrDefault(String, int) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
map<string, int32> kill_tasks_summary = 20;
getKillTasksSummaryOrDefault(String, int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataOrBuilder
map<string, int32> kill_tasks_summary = 20;
getKillTasksSummaryOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
map<string, int32> kill_tasks_summary = 20;
getKillTasksSummaryOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
map<string, int32> kill_tasks_summary = 20;
getKillTasksSummaryOrThrow(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataOrBuilder
map<string, int32> kill_tasks_summary = 20;
getKrb5LoginModuleName() - Static method in class org.apache.spark.util.SecurityUtils
Krb5LoginModule package varies in different JVMs.
getLabel() - Method in class org.apache.spark.ml.feature.LabeledPoint
 
getLabel() - Method in class org.apache.spark.mllib.regression.LabeledPoint
 
getLabelCol() - Method in interface org.apache.spark.ml.param.shared.HasLabelCol
 
getLabels() - Method in class org.apache.spark.ml.feature.IndexToString
 
getLabelType() - Method in interface org.apache.spark.ml.feature.UnivariateFeatureSelectorParams
 
getLambda() - Method in class org.apache.spark.mllib.classification.NaiveBayes
Get the smoothing parameter.
getLastUpdated() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo.Builder
int64 last_updated = 4;
getLastUpdated() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo
int64 last_updated = 4;
getLastUpdated() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfoOrBuilder
int64 last_updated = 4;
getLastUpdatedEpoch() - Method in class org.apache.spark.status.api.v1.ApplicationAttemptInfo
 
getLatestOffset() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
optional string latest_offset = 4;
getLatestOffset() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress
optional string latest_offset = 4;
getLatestOffset() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SourceProgressOrBuilder
optional string latest_offset = 4;
getLatestOffsetBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
optional string latest_offset = 4;
getLatestOffsetBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress
optional string latest_offset = 4;
getLatestOffsetBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SourceProgressOrBuilder
optional string latest_offset = 4;
getLaunchTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
int64 launch_time = 5;
getLaunchTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
int64 launch_time = 5;
getLaunchTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataOrBuilder
int64 launch_time = 5;
getLaunchTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 launch_time = 5;
getLaunchTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
int64 launch_time = 5;
getLaunchTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
int64 launch_time = 5;
getLayers() - Method in interface org.apache.spark.ml.classification.MultilayerPerceptronParams
 
getLDAModel(double[]) - Method in interface org.apache.spark.mllib.clustering.LDAOptimizer
 
getLeafCol() - Method in interface org.apache.spark.ml.tree.DecisionTreeParams
 
getLeafField(String) - Method in interface org.apache.spark.ml.tree.DecisionTreeModel
 
getLeafField(String) - Method in interface org.apache.spark.ml.tree.TreeEnsembleModel
 
getLearningDecay() - Method in interface org.apache.spark.ml.clustering.LDAParams
 
getLearningOffset() - Method in interface org.apache.spark.ml.clustering.LDAParams
 
getLearningRate() - Method in class org.apache.spark.mllib.tree.configuration.BoostingStrategy
 
getLeastGroupHash(String) - Method in class org.apache.spark.rdd.DefaultPartitionCoalescer
Gets the least element of the list associated with key in groupHash The returned PartitionGroup is the least loaded of all groups that represent the machine "key"
getLength() - Static method in class org.apache.spark.rdd.InputFileBlockHolder
Returns the length of the block being read, or -1 if it is unknown.
getLimitClause(Integer) - Static method in class org.apache.spark.sql.jdbc.DB2Dialect
 
getLimitClause(Integer) - Static method in class org.apache.spark.sql.jdbc.DerbyDialect
 
getLimitClause(Integer) - Static method in class org.apache.spark.sql.jdbc.H2Dialect
 
getLimitClause(Integer) - Method in class org.apache.spark.sql.jdbc.JdbcDialect
returns the LIMIT clause for the SELECT statement
getLimitClause(Integer) - Static method in class org.apache.spark.sql.jdbc.MsSqlServerDialect
 
getLimitClause(Integer) - Static method in class org.apache.spark.sql.jdbc.MySQLDialect
 
getLimitClause(Integer) - Static method in class org.apache.spark.sql.jdbc.NoopDialect
 
getLimitClause(Integer) - Static method in class org.apache.spark.sql.jdbc.OracleDialect
 
getLimitClause(Integer) - Static method in class org.apache.spark.sql.jdbc.PostgresDialect
 
getLimitClause(Integer) - Static method in class org.apache.spark.sql.jdbc.TeradataDialect
 
getLink() - Method in interface org.apache.spark.ml.regression.GeneralizedLinearRegressionBase
 
getLinkPower() - Method in interface org.apache.spark.ml.regression.GeneralizedLinearRegressionBase
 
getLinkPredictionCol() - Method in interface org.apache.spark.ml.regression.GeneralizedLinearRegressionBase
 
getList(int) - Method in interface org.apache.spark.sql.Row
Returns the value at position i of array type as java.util.List.
getLocalBlocksFetched(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
repeated double local_blocks_fetched = 4;
getLocalBlocksFetched(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions
repeated double local_blocks_fetched = 4;
getLocalBlocksFetched(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributionsOrBuilder
repeated double local_blocks_fetched = 4;
getLocalBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics.Builder
int64 local_blocks_fetched = 2;
getLocalBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics
int64 local_blocks_fetched = 2;
getLocalBlocksFetched() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricsOrBuilder
int64 local_blocks_fetched = 2;
getLocalBlocksFetchedCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
repeated double local_blocks_fetched = 4;
getLocalBlocksFetchedCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions
repeated double local_blocks_fetched = 4;
getLocalBlocksFetchedCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributionsOrBuilder
repeated double local_blocks_fetched = 4;
getLocalBlocksFetchedList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
repeated double local_blocks_fetched = 4;
getLocalBlocksFetchedList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions
repeated double local_blocks_fetched = 4;
getLocalBlocksFetchedList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributionsOrBuilder
repeated double local_blocks_fetched = 4;
getLocalBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics.Builder
int64 local_bytes_read = 6;
getLocalBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics
int64 local_bytes_read = 6;
getLocalBytesRead() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricsOrBuilder
int64 local_bytes_read = 6;
getLocalDate(int) - Method in interface org.apache.spark.sql.Row
Returns the value at position i of date type as java.time.LocalDate.
getLocalDir(SparkConf) - Static method in class org.apache.spark.util.Utils
Get the path of a temporary directory.
getLocale() - Method in class org.apache.spark.ml.feature.StopWordsRemover
 
getLocality() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper.Builder
Deprecated.
getLocality() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper
Deprecated.
getLocality() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataWrapperOrBuilder
Deprecated.
getLocalityCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper.Builder
 
getLocalityCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper
 
getLocalityCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataWrapperOrBuilder
map<string, int64> locality = 3;
getLocalityMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper.Builder
map<string, int64> locality = 3;
getLocalityMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper
map<string, int64> locality = 3;
getLocalityMap() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataWrapperOrBuilder
map<string, int64> locality = 3;
getLocalityOrDefault(String, long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper.Builder
map<string, int64> locality = 3;
getLocalityOrDefault(String, long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper
map<string, int64> locality = 3;
getLocalityOrDefault(String, long) - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataWrapperOrBuilder
map<string, int64> locality = 3;
getLocalityOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper.Builder
map<string, int64> locality = 3;
getLocalityOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper
map<string, int64> locality = 3;
getLocalityOrThrow(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataWrapperOrBuilder
map<string, int64> locality = 3;
getLocalMergedBlocksFetched(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
repeated double local_merged_blocks_fetched = 4;
getLocalMergedBlocksFetched(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions
repeated double local_merged_blocks_fetched = 4;
getLocalMergedBlocksFetched(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributionsOrBuilder
repeated double local_merged_blocks_fetched = 4;
getLocalMergedBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics.Builder
int64 local_merged_blocks_fetched = 4;
getLocalMergedBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics
int64 local_merged_blocks_fetched = 4;
getLocalMergedBlocksFetched() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricsOrBuilder
int64 local_merged_blocks_fetched = 4;
getLocalMergedBlocksFetchedCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
repeated double local_merged_blocks_fetched = 4;
getLocalMergedBlocksFetchedCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions
repeated double local_merged_blocks_fetched = 4;
getLocalMergedBlocksFetchedCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributionsOrBuilder
repeated double local_merged_blocks_fetched = 4;
getLocalMergedBlocksFetchedList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
repeated double local_merged_blocks_fetched = 4;
getLocalMergedBlocksFetchedList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions
repeated double local_merged_blocks_fetched = 4;
getLocalMergedBlocksFetchedList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributionsOrBuilder
repeated double local_merged_blocks_fetched = 4;
getLocalMergedBytesRead(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
repeated double local_merged_bytes_read = 8;
getLocalMergedBytesRead(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions
repeated double local_merged_bytes_read = 8;
getLocalMergedBytesRead(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributionsOrBuilder
repeated double local_merged_bytes_read = 8;
getLocalMergedBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics.Builder
int64 local_merged_bytes_read = 8;
getLocalMergedBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics
int64 local_merged_bytes_read = 8;
getLocalMergedBytesRead() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricsOrBuilder
int64 local_merged_bytes_read = 8;
getLocalMergedBytesReadCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
repeated double local_merged_bytes_read = 8;
getLocalMergedBytesReadCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions
repeated double local_merged_bytes_read = 8;
getLocalMergedBytesReadCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributionsOrBuilder
repeated double local_merged_bytes_read = 8;
getLocalMergedBytesReadList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
repeated double local_merged_bytes_read = 8;
getLocalMergedBytesReadList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions
repeated double local_merged_bytes_read = 8;
getLocalMergedBytesReadList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributionsOrBuilder
repeated double local_merged_bytes_read = 8;
getLocalMergedChunksFetched(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
repeated double local_merged_chunks_fetched = 6;
getLocalMergedChunksFetched(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions
repeated double local_merged_chunks_fetched = 6;
getLocalMergedChunksFetched(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributionsOrBuilder
repeated double local_merged_chunks_fetched = 6;
getLocalMergedChunksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics.Builder
int64 local_merged_chunks_fetched = 6;
getLocalMergedChunksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics
int64 local_merged_chunks_fetched = 6;
getLocalMergedChunksFetched() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricsOrBuilder
int64 local_merged_chunks_fetched = 6;
getLocalMergedChunksFetchedCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
repeated double local_merged_chunks_fetched = 6;
getLocalMergedChunksFetchedCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions
repeated double local_merged_chunks_fetched = 6;
getLocalMergedChunksFetchedCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributionsOrBuilder
repeated double local_merged_chunks_fetched = 6;
getLocalMergedChunksFetchedList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
repeated double local_merged_chunks_fetched = 6;
getLocalMergedChunksFetchedList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions
repeated double local_merged_chunks_fetched = 6;
getLocalMergedChunksFetchedList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributionsOrBuilder
repeated double local_merged_chunks_fetched = 6;
getLocalProperty(String) - Method in class org.apache.spark.api.java.JavaSparkContext
Get a local property set in this thread, or null if it is missing.
getLocalProperty(String) - Method in class org.apache.spark.BarrierTaskContext
 
getLocalProperty(String) - Method in class org.apache.spark.SparkContext
Get a local property set in this thread, or null if it is missing.
getLocalProperty(String) - Method in class org.apache.spark.TaskContext
Get a local property set upstream in the driver, or null if it is missing.
getLocalUserJarsForShell(SparkConf) - Static method in class org.apache.spark.util.Utils
Return the local jar files which will be added to REPL's classpath.
GetLocations(BlockId) - Constructor for class org.apache.spark.storage.BlockManagerMessages.GetLocations
 
GetLocations$() - Constructor for class org.apache.spark.storage.BlockManagerMessages.GetLocations$
 
GetLocationsAndStatus(BlockId, String) - Constructor for class org.apache.spark.storage.BlockManagerMessages.GetLocationsAndStatus
 
GetLocationsAndStatus$() - Constructor for class org.apache.spark.storage.BlockManagerMessages.GetLocationsAndStatus$
 
GetLocationsMultipleBlockIds(BlockId[]) - Constructor for class org.apache.spark.storage.BlockManagerMessages.GetLocationsMultipleBlockIds
 
GetLocationsMultipleBlockIds$() - Constructor for class org.apache.spark.storage.BlockManagerMessages.GetLocationsMultipleBlockIds$
 
getLong(String, long) - Method in class org.apache.spark.SparkConf
Get a parameter as a long, falling back to a default if not set
getLong(int) - Method in interface org.apache.spark.sql.Row
Returns the value at position i as a primitive long.
getLong(String) - Method in class org.apache.spark.sql.types.Metadata
Gets a Long.
getLong(String, long) - Method in class org.apache.spark.sql.util.CaseInsensitiveStringMap
Returns the long value to which the specified key is mapped, or defaultValue if there is no mapping for the key.
getLong(int) - Method in class org.apache.spark.sql.vectorized.ArrowColumnVector
 
getLong(int) - Method in class org.apache.spark.sql.vectorized.ColumnarArray
 
getLong(int) - Method in class org.apache.spark.sql.vectorized.ColumnarBatchRow
 
getLong(int) - Method in class org.apache.spark.sql.vectorized.ColumnarRow
 
getLong(int) - Method in class org.apache.spark.sql.vectorized.ColumnVector
Returns the long type value for rowId.
getLongArray(String) - Method in class org.apache.spark.sql.types.Metadata
Gets a Long array.
getLongs(int, int) - Method in class org.apache.spark.sql.vectorized.ColumnVector
Gets long type values from [rowId, rowId + count).
getLoss() - Method in interface org.apache.spark.ml.param.shared.HasLoss
 
getLoss() - Method in class org.apache.spark.mllib.tree.configuration.BoostingStrategy
 
getLossType() - Method in interface org.apache.spark.ml.tree.GBTClassifierParams
 
getLossType() - Method in interface org.apache.spark.ml.tree.GBTRegressorParams
 
getLower() - Method in interface org.apache.spark.ml.feature.RobustScalerParams
 
getLowerBound(double, long, double) - Static method in class org.apache.spark.util.random.BinomialBounds
Returns a threshold p such that if we conduct n Bernoulli trials with success rate = p, it is very unlikely to have more than fraction * n successes.
getLowerBound(double) - Static method in class org.apache.spark.util.random.PoissonBounds
Returns a lambda such that Pr[X > s] is very small, where X ~ Pois(lambda).
getLowerBoundsOnCoefficients() - Method in interface org.apache.spark.ml.classification.LogisticRegressionParams
 
getLowerBoundsOnIntercepts() - Method in interface org.apache.spark.ml.classification.LogisticRegressionParams
 
getMap(int) - Method in interface org.apache.spark.sql.Row
Returns the value at position i of map type as a Scala Map.
getMap(int) - Method in class org.apache.spark.sql.vectorized.ArrowColumnVector
 
getMap(int) - Method in class org.apache.spark.sql.vectorized.ColumnarArray
 
getMap(int) - Method in class org.apache.spark.sql.vectorized.ColumnarBatchRow
 
getMap(int) - Method in class org.apache.spark.sql.vectorized.ColumnarRow
 
getMap(int) - Method in class org.apache.spark.sql.vectorized.ColumnVector
Returns the map type value for rowId.
getMapOutputMetadata() - Method in class org.apache.spark.shuffle.api.metadata.MapOutputCommitMessage
 
GetMatchingBlockIds(Function1<BlockId, Object>, boolean) - Constructor for class org.apache.spark.storage.BlockManagerMessages.GetMatchingBlockIds
 
GetMatchingBlockIds$() - Constructor for class org.apache.spark.storage.BlockManagerMessages.GetMatchingBlockIds$
 
getMax() - Method in interface org.apache.spark.ml.feature.MinMaxScalerParams
 
getMaxBins() - Method in interface org.apache.spark.ml.tree.DecisionTreeParams
 
getMaxBins() - Method in class org.apache.spark.mllib.tree.configuration.Strategy
 
getMaxBlockSizeInMB() - Method in interface org.apache.spark.ml.param.shared.HasMaxBlockSizeInMB
 
getMaxCategories() - Method in interface org.apache.spark.ml.feature.VectorIndexerParams
 
getMaxCores() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
optional int32 max_cores = 4;
getMaxCores() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo
optional int32 max_cores = 4;
getMaxCores() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoOrBuilder
optional int32 max_cores = 4;
getMaxDepth() - Method in interface org.apache.spark.ml.tree.DecisionTreeParams
 
getMaxDepth() - Method in class org.apache.spark.mllib.tree.configuration.Strategy
 
getMaxDF() - Method in interface org.apache.spark.ml.feature.CountVectorizerParams
 
getMaxFailures(SparkConf, boolean) - Static method in class org.apache.spark.streaming.util.WriteAheadLogUtils
 
getMaxIter() - Method in interface org.apache.spark.ml.param.shared.HasMaxIter
 
getMaxIterations() - Method in class org.apache.spark.mllib.clustering.BisectingKMeans
Gets the max number of k-means iterations to split clusters.
getMaxIterations() - Method in class org.apache.spark.mllib.clustering.GaussianMixture
Return the maximum number of iterations allowed
getMaxIterations() - Method in class org.apache.spark.mllib.clustering.KMeans
Maximum number of iterations allowed.
getMaxIterations() - Method in class org.apache.spark.mllib.clustering.LDA
Maximum number of iterations allowed.
getMaxLocalProjDBSize() - Method in class org.apache.spark.ml.fpm.PrefixSpan
 
getMaxLocalProjDBSize() - Method in class org.apache.spark.mllib.fpm.PrefixSpan
Gets the maximum number of items allowed in a projected database before local processing.
getMaxMemory() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
int64 max_memory = 19;
getMaxMemory() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
int64 max_memory = 19;
getMaxMemory() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
int64 max_memory = 19;
getMaxMemoryInMB() - Method in interface org.apache.spark.ml.tree.DecisionTreeParams
 
getMaxMemoryInMB() - Method in class org.apache.spark.mllib.tree.configuration.Strategy
 
getMaxPatternLength() - Method in class org.apache.spark.ml.fpm.PrefixSpan
 
getMaxPatternLength() - Method in class org.apache.spark.mllib.fpm.PrefixSpan
Gets the maximal pattern length (i.e.
getMaxSentenceLength() - Method in interface org.apache.spark.ml.feature.Word2VecBase
 
getMaxTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
int32 max_tasks = 8;
getMaxTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
int32 max_tasks = 8;
getMaxTasks() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
int32 max_tasks = 8;
getMemoryBytesSpilled() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double memory_bytes_spilled = 16;
getMemoryBytesSpilled() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
double memory_bytes_spilled = 16;
getMemoryBytesSpilled() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
double memory_bytes_spilled = 16;
getMemoryBytesSpilled(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double memory_bytes_spilled = 14;
getMemoryBytesSpilled(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
repeated double memory_bytes_spilled = 14;
getMemoryBytesSpilled(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
repeated double memory_bytes_spilled = 14;
getMemoryBytesSpilled() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
int64 memory_bytes_spilled = 13;
getMemoryBytesSpilled() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary
int64 memory_bytes_spilled = 13;
getMemoryBytesSpilled() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryOrBuilder
int64 memory_bytes_spilled = 13;
getMemoryBytesSpilled() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int64 memory_bytes_spilled = 21;
getMemoryBytesSpilled() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
int64 memory_bytes_spilled = 21;
getMemoryBytesSpilled() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
int64 memory_bytes_spilled = 21;
getMemoryBytesSpilled() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 memory_bytes_spilled = 23;
getMemoryBytesSpilled() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
int64 memory_bytes_spilled = 23;
getMemoryBytesSpilled() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
int64 memory_bytes_spilled = 23;
getMemoryBytesSpilled(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double memory_bytes_spilled = 13;
getMemoryBytesSpilled(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
repeated double memory_bytes_spilled = 13;
getMemoryBytesSpilled(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
repeated double memory_bytes_spilled = 13;
getMemoryBytesSpilled() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
int64 memory_bytes_spilled = 8;
getMemoryBytesSpilled() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics
int64 memory_bytes_spilled = 8;
getMemoryBytesSpilled() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricsOrBuilder
int64 memory_bytes_spilled = 8;
getMemoryBytesSpilledCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double memory_bytes_spilled = 14;
getMemoryBytesSpilledCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
repeated double memory_bytes_spilled = 14;
getMemoryBytesSpilledCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
repeated double memory_bytes_spilled = 14;
getMemoryBytesSpilledCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double memory_bytes_spilled = 13;
getMemoryBytesSpilledCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
repeated double memory_bytes_spilled = 13;
getMemoryBytesSpilledCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
repeated double memory_bytes_spilled = 13;
getMemoryBytesSpilledList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double memory_bytes_spilled = 14;
getMemoryBytesSpilledList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
repeated double memory_bytes_spilled = 14;
getMemoryBytesSpilledList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
repeated double memory_bytes_spilled = 14;
getMemoryBytesSpilledList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double memory_bytes_spilled = 13;
getMemoryBytesSpilledList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
repeated double memory_bytes_spilled = 13;
getMemoryBytesSpilledList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
repeated double memory_bytes_spilled = 13;
getMemoryMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
optional .org.apache.spark.status.protobuf.MemoryMetrics memory_metrics = 24;
getMemoryMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
optional .org.apache.spark.status.protobuf.MemoryMetrics memory_metrics = 24;
getMemoryMetrics() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
optional .org.apache.spark.status.protobuf.MemoryMetrics memory_metrics = 24;
getMemoryMetricsBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
optional .org.apache.spark.status.protobuf.MemoryMetrics memory_metrics = 24;
getMemoryMetricsOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
optional .org.apache.spark.status.protobuf.MemoryMetrics memory_metrics = 24;
getMemoryMetricsOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
optional .org.apache.spark.status.protobuf.MemoryMetrics memory_metrics = 24;
getMemoryMetricsOrBuilder() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
optional .org.apache.spark.status.protobuf.MemoryMetrics memory_metrics = 24;
getMemoryPerExecutorMb() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
optional int32 memory_per_executor_mb = 6;
getMemoryPerExecutorMb() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo
optional int32 memory_per_executor_mb = 6;
getMemoryPerExecutorMb() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoOrBuilder
optional int32 memory_per_executor_mb = 6;
getMemoryRemaining() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution.Builder
int64 memory_remaining = 3;
getMemoryRemaining() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution
int64 memory_remaining = 3;
getMemoryRemaining() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDDataDistributionOrBuilder
int64 memory_remaining = 3;
GetMemoryStatus$() - Constructor for class org.apache.spark.storage.BlockManagerMessages.GetMemoryStatus$
 
getMemoryUsed() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
int64 memory_used = 5;
getMemoryUsed() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
int64 memory_used = 5;
getMemoryUsed() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
int64 memory_used = 5;
getMemoryUsed() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution.Builder
int64 memory_used = 2;
getMemoryUsed() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution
int64 memory_used = 2;
getMemoryUsed() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDDataDistributionOrBuilder
int64 memory_used = 2;
getMemoryUsed() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo.Builder
int64 memory_used = 3;
getMemoryUsed() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo
int64 memory_used = 3;
getMemoryUsed() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfoOrBuilder
int64 memory_used = 3;
getMemoryUsed() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
int64 memory_used = 6;
getMemoryUsed() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo
int64 memory_used = 6;
getMemoryUsed() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoOrBuilder
int64 memory_used = 6;
getMemoryUsedBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
int64 memory_used_bytes = 8;
getMemoryUsedBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress
int64 memory_used_bytes = 8;
getMemoryUsedBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgressOrBuilder
int64 memory_used_bytes = 8;
getMemSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
int64 mem_size = 8;
getMemSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData
int64 mem_size = 8;
getMemSize() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamBlockDataOrBuilder
int64 mem_size = 8;
getMergedFetchFallbackCount(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
repeated double merged_fetch_fallback_count = 2;
getMergedFetchFallbackCount(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions
repeated double merged_fetch_fallback_count = 2;
getMergedFetchFallbackCount(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributionsOrBuilder
repeated double merged_fetch_fallback_count = 2;
getMergedFetchFallbackCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics.Builder
int64 merged_fetch_fallback_count = 2;
getMergedFetchFallbackCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics
int64 merged_fetch_fallback_count = 2;
getMergedFetchFallbackCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricsOrBuilder
int64 merged_fetch_fallback_count = 2;
getMergedFetchFallbackCountCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
repeated double merged_fetch_fallback_count = 2;
getMergedFetchFallbackCountCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions
repeated double merged_fetch_fallback_count = 2;
getMergedFetchFallbackCountCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributionsOrBuilder
repeated double merged_fetch_fallback_count = 2;
getMergedFetchFallbackCountList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
repeated double merged_fetch_fallback_count = 2;
getMergedFetchFallbackCountList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions
repeated double merged_fetch_fallback_count = 2;
getMergedFetchFallbackCountList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributionsOrBuilder
repeated double merged_fetch_fallback_count = 2;
getMergerLocs() - Method in class org.apache.spark.ShuffleDependency
 
getMessage(String, Map<String, String>) - Static method in class org.apache.spark.SparkThrowableHelper
 
getMessage(String, Map<String, String>) - Static method in class org.apache.spark.SparkThrowableHelper
 
getMessage(String, Map<String, String>, String) - Static method in class org.apache.spark.SparkThrowableHelper
 
getMessage(SparkThrowable, Enumeration.Value) - Static method in class org.apache.spark.SparkThrowableHelper
 
getMessage(Throwable) - Static method in class org.apache.spark.SparkThrowableHelper
 
getMessage() - Method in exception org.apache.spark.sql.AnalysisException
 
getMessageParameters() - Method in exception org.apache.spark.SparkException
 
getMessageParameters() - Method in interface org.apache.spark.SparkThrowable
 
getMessageParameters() - Method in exception org.apache.spark.sql.AnalysisException
 
getMessageParameters() - Method in exception org.apache.spark.sql.streaming.StreamingQueryException
 
getMessageTemplate(String) - Method in class org.apache.spark.ErrorClassesJsonReader
 
getMetadata(String) - Method in class org.apache.spark.sql.types.Metadata
Gets a Metadata.
getMetadataArray(String) - Method in class org.apache.spark.sql.types.Metadata
Gets a Metadata array.
getMetricDistribution(String) - Method in class org.apache.spark.status.api.v1.ExecutorPeakMetricsDistributions
Returns the distributions for the specified metric.
getMetricLabel() - Method in class org.apache.spark.ml.evaluation.MulticlassClassificationEvaluator
 
getMetricLabel() - Method in class org.apache.spark.ml.evaluation.MultilabelClassificationEvaluator
 
getMetricName() - Method in class org.apache.spark.ml.evaluation.BinaryClassificationEvaluator
 
getMetricName() - Method in class org.apache.spark.ml.evaluation.ClusteringEvaluator
 
getMetricName() - Method in class org.apache.spark.ml.evaluation.MulticlassClassificationEvaluator
 
getMetricName() - Method in class org.apache.spark.ml.evaluation.MultilabelClassificationEvaluator
 
getMetricName() - Method in class org.apache.spark.ml.evaluation.RankingEvaluator
 
getMetricName() - Method in class org.apache.spark.ml.evaluation.RegressionEvaluator
 
getMetrics(Dataset<?>) - Method in class org.apache.spark.ml.evaluation.BinaryClassificationEvaluator
Get a BinaryClassificationMetrics, which can be used to get binary classification metrics such as areaUnderROC and areaUnderPR.
getMetrics(Dataset<?>) - Method in class org.apache.spark.ml.evaluation.ClusteringEvaluator
Get a ClusteringMetrics, which can be used to get clustering metrics such as silhouette score.
getMetrics(Dataset<?>) - Method in class org.apache.spark.ml.evaluation.MulticlassClassificationEvaluator
Get a MulticlassMetrics, which can be used to get multiclass classification metrics such as accuracy, weightedPrecision, etc.
getMetrics(Dataset<?>) - Method in class org.apache.spark.ml.evaluation.MultilabelClassificationEvaluator
Get a MultilabelMetrics, which can be used to get multilabel classification metrics such as accuracy, precision, precisionByLabel, etc.
getMetrics(Dataset<?>) - Method in class org.apache.spark.ml.evaluation.RankingEvaluator
Get a RankingMetrics, which can be used to get ranking metrics such as meanAveragePrecision, meanAveragePrecisionAtK, etc.
getMetrics(Dataset<?>) - Method in class org.apache.spark.ml.evaluation.RegressionEvaluator
Get a RegressionMetrics, which can be used to get regression metrics such as rootMeanSquaredError, meanSquaredError, etc.
getMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetrics.Builder
Deprecated.
getMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetrics
Deprecated.
getMetrics() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsOrBuilder
Deprecated.
getMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress.Builder
Deprecated.
getMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress
Deprecated.
getMetrics() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SinkProgressOrBuilder
Deprecated.
getMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
Deprecated.
getMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress
Deprecated.
getMetrics() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SourceProgressOrBuilder
Deprecated.
getMetrics(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 5;
getMetrics(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 5;
getMetrics(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapperOrBuilder
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 5;
getMetrics(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 4;
getMetrics(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 4;
getMetrics(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeOrBuilder
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 4;
getMetrics(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 7;
getMetrics(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 7;
getMetrics(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 7;
getMetricsBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 5;
getMetricsBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 4;
getMetricsBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 7;
getMetricsBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 5;
getMetricsBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 4;
getMetricsBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 7;
getMetricsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetrics.Builder
 
getMetricsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetrics
 
getMetricsCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsOrBuilder
map<string, int64> metrics = 1;
getMetricsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress.Builder
 
getMetricsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress
 
getMetricsCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SinkProgressOrBuilder
map<string, string> metrics = 3;
getMetricsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
 
getMetricsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress
 
getMetricsCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SourceProgressOrBuilder
map<string, string> metrics = 8;
getMetricsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 5;
getMetricsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 5;
getMetricsCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapperOrBuilder
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 5;
getMetricsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 4;
getMetricsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 4;
getMetricsCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeOrBuilder
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 4;
getMetricsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 7;
getMetricsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 7;
getMetricsCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 7;
getMetricsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 5;
getMetricsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 5;
getMetricsList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapperOrBuilder
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 5;
getMetricsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 4;
getMetricsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 4;
getMetricsList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeOrBuilder
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 4;
getMetricsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 7;
getMetricsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 7;
getMetricsList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 7;
getMetricsMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetrics.Builder
map<string, int64> metrics = 1;
getMetricsMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetrics
map<string, int64> metrics = 1;
getMetricsMap() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsOrBuilder
map<string, int64> metrics = 1;
getMetricsMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress.Builder
map<string, string> metrics = 3;
getMetricsMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress
map<string, string> metrics = 3;
getMetricsMap() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SinkProgressOrBuilder
map<string, string> metrics = 3;
getMetricsMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
map<string, string> metrics = 8;
getMetricsMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress
map<string, string> metrics = 8;
getMetricsMap() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SourceProgressOrBuilder
map<string, string> metrics = 8;
getMetricsOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 5;
getMetricsOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 5;
getMetricsOrBuilder(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapperOrBuilder
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 5;
getMetricsOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 4;
getMetricsOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 4;
getMetricsOrBuilder(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeOrBuilder
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 4;
getMetricsOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 7;
getMetricsOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 7;
getMetricsOrBuilder(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 7;
getMetricsOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 5;
getMetricsOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 5;
getMetricsOrBuilderList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapperOrBuilder
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 5;
getMetricsOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 4;
getMetricsOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 4;
getMetricsOrBuilderList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeOrBuilder
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 4;
getMetricsOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 7;
getMetricsOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 7;
getMetricsOrBuilderList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 7;
getMetricsOrDefault(String, long) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetrics.Builder
map<string, int64> metrics = 1;
getMetricsOrDefault(String, long) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetrics
map<string, int64> metrics = 1;
getMetricsOrDefault(String, long) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsOrBuilder
map<string, int64> metrics = 1;
getMetricsOrDefault(String, String) - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress.Builder
map<string, string> metrics = 3;
getMetricsOrDefault(String, String) - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress
map<string, string> metrics = 3;
getMetricsOrDefault(String, String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.SinkProgressOrBuilder
map<string, string> metrics = 3;
getMetricsOrDefault(String, String) - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
map<string, string> metrics = 8;
getMetricsOrDefault(String, String) - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress
map<string, string> metrics = 8;
getMetricsOrDefault(String, String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.SourceProgressOrBuilder
map<string, string> metrics = 8;
getMetricsOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetrics.Builder
map<string, int64> metrics = 1;
getMetricsOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetrics
map<string, int64> metrics = 1;
getMetricsOrThrow(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsOrBuilder
map<string, int64> metrics = 1;
getMetricsOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress.Builder
map<string, string> metrics = 3;
getMetricsOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress
map<string, string> metrics = 3;
getMetricsOrThrow(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.SinkProgressOrBuilder
map<string, string> metrics = 3;
getMetricsOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
map<string, string> metrics = 8;
getMetricsOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress
map<string, string> metrics = 8;
getMetricsOrThrow(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.SourceProgressOrBuilder
map<string, string> metrics = 8;
getMetricsProperties(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings metrics_properties = 5;
getMetricsProperties(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
repeated .org.apache.spark.status.protobuf.PairStrings metrics_properties = 5;
getMetricsProperties(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoOrBuilder
repeated .org.apache.spark.status.protobuf.PairStrings metrics_properties = 5;
getMetricsPropertiesBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings metrics_properties = 5;
getMetricsPropertiesBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings metrics_properties = 5;
getMetricsPropertiesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings metrics_properties = 5;
getMetricsPropertiesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
repeated .org.apache.spark.status.protobuf.PairStrings metrics_properties = 5;
getMetricsPropertiesCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoOrBuilder
repeated .org.apache.spark.status.protobuf.PairStrings metrics_properties = 5;
getMetricsPropertiesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings metrics_properties = 5;
getMetricsPropertiesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
repeated .org.apache.spark.status.protobuf.PairStrings metrics_properties = 5;
getMetricsPropertiesList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoOrBuilder
repeated .org.apache.spark.status.protobuf.PairStrings metrics_properties = 5;
getMetricsPropertiesOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings metrics_properties = 5;
getMetricsPropertiesOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
repeated .org.apache.spark.status.protobuf.PairStrings metrics_properties = 5;
getMetricsPropertiesOrBuilder(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoOrBuilder
repeated .org.apache.spark.status.protobuf.PairStrings metrics_properties = 5;
getMetricsPropertiesOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings metrics_properties = 5;
getMetricsPropertiesOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
repeated .org.apache.spark.status.protobuf.PairStrings metrics_properties = 5;
getMetricsPropertiesOrBuilderList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoOrBuilder
repeated .org.apache.spark.status.protobuf.PairStrings metrics_properties = 5;
getMetricsSources(String) - Method in class org.apache.spark.BarrierTaskContext
 
getMetricsSources(String) - Method in class org.apache.spark.TaskContext
::DeveloperApi:: Returns all metrics sources with the given name which are associated with the instance which runs the task.
getMetricType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetric.Builder
optional string metric_type = 3;
getMetricType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetric
optional string metric_type = 3;
getMetricType() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetricOrBuilder
optional string metric_type = 3;
getMetricTypeBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetric.Builder
optional string metric_type = 3;
getMetricTypeBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetric
optional string metric_type = 3;
getMetricTypeBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetricOrBuilder
optional string metric_type = 3;
getMetricValue(MemoryManager) - Method in interface org.apache.spark.metrics.SingleValueExecutorMetricType
 
getMetricValues(MemoryManager) - Method in interface org.apache.spark.metrics.ExecutorMetricType
 
getMetricValues(MemoryManager) - Method in interface org.apache.spark.metrics.SingleValueExecutorMetricType
 
getMetricValues() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
Deprecated.
getMetricValues() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
Deprecated.
getMetricValues() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
Deprecated.
getMetricValuesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
 
getMetricValuesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
 
getMetricValuesCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
map<int64, string> metric_values = 14;
getMetricValuesIsNull() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
bool metric_values_is_null = 13;
getMetricValuesIsNull() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
bool metric_values_is_null = 13;
getMetricValuesIsNull() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
bool metric_values_is_null = 13;
getMetricValuesMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
map<int64, string> metric_values = 14;
getMetricValuesMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
map<int64, string> metric_values = 14;
getMetricValuesMap() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
map<int64, string> metric_values = 14;
getMetricValuesOrDefault(long, String) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
map<int64, string> metric_values = 14;
getMetricValuesOrDefault(long, String) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
map<int64, string> metric_values = 14;
getMetricValuesOrDefault(long, String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
map<int64, string> metric_values = 14;
getMetricValuesOrThrow(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
map<int64, string> metric_values = 14;
getMetricValuesOrThrow(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
map<int64, string> metric_values = 14;
getMetricValuesOrThrow(long) - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
map<int64, string> metric_values = 14;
getMin() - Method in interface org.apache.spark.ml.feature.MinMaxScalerParams
 
getMinConfidence() - Method in interface org.apache.spark.ml.fpm.FPGrowthParams
 
getMinCount() - Method in interface org.apache.spark.ml.feature.Word2VecBase
 
getMinDF() - Method in interface org.apache.spark.ml.feature.CountVectorizerParams
 
getMinDivisibleClusterSize() - Method in interface org.apache.spark.ml.clustering.BisectingKMeansParams
 
getMinDivisibleClusterSize() - Method in class org.apache.spark.mllib.clustering.BisectingKMeans
Gets the minimum number of points (if greater than or equal to 1.0) or the minimum proportion of points (if less than 1.0) of a divisible cluster.
getMinDocFreq() - Method in interface org.apache.spark.ml.feature.IDFBase
 
getMiniBatchFraction() - Method in interface org.apache.spark.ml.regression.FactorizationMachinesParams
 
getMiniBatchFraction() - Method in class org.apache.spark.mllib.clustering.OnlineLDAOptimizer
Mini-batch fraction, which sets the fraction of document sampled and used in each iteration
getMinInfoGain() - Method in interface org.apache.spark.ml.tree.DecisionTreeParams
 
getMinInfoGain() - Method in class org.apache.spark.mllib.tree.configuration.Strategy
 
getMinInstancesPerNode() - Method in interface org.apache.spark.ml.tree.DecisionTreeParams
 
getMinInstancesPerNode() - Method in class org.apache.spark.mllib.tree.configuration.Strategy
 
getMinSupport() - Method in interface org.apache.spark.ml.fpm.FPGrowthParams
 
getMinSupport() - Method in class org.apache.spark.ml.fpm.PrefixSpan
 
getMinSupport() - Method in class org.apache.spark.mllib.fpm.PrefixSpan
Get the minimal support (i.e.
getMinTF() - Method in interface org.apache.spark.ml.feature.CountVectorizerParams
 
getMinTokenLength() - Method in class org.apache.spark.ml.feature.RegexTokenizer
 
getMinWeightFractionPerNode() - Method in interface org.apache.spark.ml.tree.DecisionTreeParams
 
getMinWeightFractionPerNode() - Method in class org.apache.spark.mllib.tree.configuration.Strategy
 
getMissingValue() - Method in interface org.apache.spark.ml.feature.ImputerParams
 
getMode(Row) - Static method in class org.apache.spark.ml.image.ImageSchema
Gets the OpenCV representation as an int
getModelType() - Method in interface org.apache.spark.ml.classification.NaiveBayesParams
 
getModelType() - Method in class org.apache.spark.mllib.classification.NaiveBayes
Get the model type.
getModifiedConfigs() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
Deprecated.
getModifiedConfigs() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
Deprecated.
getModifiedConfigs() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
Deprecated.
getModifiedConfigsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
 
getModifiedConfigsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
 
getModifiedConfigsCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
map<string, string> modified_configs = 6;
getModifiedConfigsMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
map<string, string> modified_configs = 6;
getModifiedConfigsMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
map<string, string> modified_configs = 6;
getModifiedConfigsMap() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
map<string, string> modified_configs = 6;
getModifiedConfigsOrDefault(String, String) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
map<string, string> modified_configs = 6;
getModifiedConfigsOrDefault(String, String) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
map<string, string> modified_configs = 6;
getModifiedConfigsOrDefault(String, String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
map<string, string> modified_configs = 6;
getModifiedConfigsOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
map<string, string> modified_configs = 6;
getModifiedConfigsOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
map<string, string> modified_configs = 6;
getModifiedConfigsOrThrow(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
map<string, string> modified_configs = 6;
getMutableAttributes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
Deprecated.
getMutableCustomMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
Deprecated.
getMutableDurationMs() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
Deprecated.
getMutableEventTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
Deprecated.
getMutableExecutorLogs() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
Deprecated.
getMutableExecutorLogs() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
Deprecated.
getMutableExecutorResources() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo.Builder
Deprecated.
getMutableExecutorSummary() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
Deprecated.
getMutableJobs() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
Deprecated.
getMutableJobsValue() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
Deprecated.
getMutableKilledTasksSummary() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
Deprecated.
getMutableKillTasksSummary() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
Deprecated.
getMutableLocality() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper.Builder
Deprecated.
getMutableMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetrics.Builder
Deprecated.
getMutableMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress.Builder
Deprecated.
getMutableMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
Deprecated.
getMutableMetricValues() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
Deprecated.
getMutableModifiedConfigs() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
Deprecated.
getMutableObservedMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
Deprecated.
getMutableProcessLogs() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder
Deprecated.
getMutableResources() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
Deprecated.
getMutableTaskResources() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo.Builder
Deprecated.
getMutableTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
Deprecated.
getN() - Method in class org.apache.spark.ml.feature.NGram
 
getName() - Method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo.Builder
optional string name = 2;
getName() - Method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo
optional string name = 2;
getName() - Method in interface org.apache.spark.status.protobuf.StoreTypes.AccumulableInfoOrBuilder
optional string name = 2;
getName() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
optional string name = 2;
getName() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo
optional string name = 2;
getName() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoOrBuilder
optional string name = 2;
getName() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
optional string name = 2;
getName() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
optional string name = 2;
getName() - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataOrBuilder
optional string name = 2;
getName() - Method in class org.apache.spark.status.protobuf.StoreTypes.PoolData.Builder
optional string name = 1;
getName() - Method in class org.apache.spark.status.protobuf.StoreTypes.PoolData
optional string name = 1;
getName() - Method in interface org.apache.spark.status.protobuf.StoreTypes.PoolDataOrBuilder
optional string name = 1;
getName() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
optional string name = 2;
getName() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper
optional string name = 2;
getName() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapperOrBuilder
optional string name = 2;
getName() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode.Builder
optional string name = 2;
getName() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode
optional string name = 2;
getName() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationNodeOrBuilder
optional string name = 2;
getName() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
optional string name = 2;
getName() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo
optional string name = 2;
getName() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoOrBuilder
optional string name = 2;
getName() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceInformation.Builder
optional string name = 1;
getName() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceInformation
optional string name = 1;
getName() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ResourceInformationOrBuilder
optional string name = 1;
getName() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
optional string name = 2;
getName() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper
optional string name = 2;
getName() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapperOrBuilder
optional string name = 2;
getName() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
optional string name = 2;
getName() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode
optional string name = 2;
getName() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeOrBuilder
optional string name = 2;
getName() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetric.Builder
optional string name = 1;
getName() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetric
optional string name = 1;
getName() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetricOrBuilder
optional string name = 1;
getName() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
optional string name = 39;
getName() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
optional string name = 39;
getName() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
optional string name = 39;
getName() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
optional string name = 1;
getName() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData
optional string name = 1;
getName() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamBlockDataOrBuilder
optional string name = 1;
getName() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData.Builder
optional string name = 1;
getName() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData
optional string name = 1;
getName() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryDataOrBuilder
optional string name = 1;
getName() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
optional string name = 3;
getName() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
optional string name = 3;
getName() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
optional string name = 3;
getNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo.Builder
optional string name = 2;
getNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo
optional string name = 2;
getNameBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.AccumulableInfoOrBuilder
optional string name = 2;
getNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
optional string name = 2;
getNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo
optional string name = 2;
getNameBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoOrBuilder
optional string name = 2;
getNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
optional string name = 2;
getNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
optional string name = 2;
getNameBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataOrBuilder
optional string name = 2;
getNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.PoolData.Builder
optional string name = 1;
getNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.PoolData
optional string name = 1;
getNameBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.PoolDataOrBuilder
optional string name = 1;
getNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
optional string name = 2;
getNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper
optional string name = 2;
getNameBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapperOrBuilder
optional string name = 2;
getNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode.Builder
optional string name = 2;
getNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode
optional string name = 2;
getNameBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationNodeOrBuilder
optional string name = 2;
getNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
optional string name = 2;
getNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo
optional string name = 2;
getNameBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoOrBuilder
optional string name = 2;
getNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceInformation.Builder
optional string name = 1;
getNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceInformation
optional string name = 1;
getNameBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ResourceInformationOrBuilder
optional string name = 1;
getNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
optional string name = 2;
getNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper
optional string name = 2;
getNameBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapperOrBuilder
optional string name = 2;
getNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
optional string name = 2;
getNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode
optional string name = 2;
getNameBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeOrBuilder
optional string name = 2;
getNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetric.Builder
optional string name = 1;
getNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetric
optional string name = 1;
getNameBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetricOrBuilder
optional string name = 1;
getNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
optional string name = 39;
getNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
optional string name = 39;
getNameBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
optional string name = 39;
getNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
optional string name = 1;
getNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData
optional string name = 1;
getNameBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamBlockDataOrBuilder
optional string name = 1;
getNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData.Builder
optional string name = 1;
getNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData
optional string name = 1;
getNameBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryDataOrBuilder
optional string name = 1;
getNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
optional string name = 3;
getNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
optional string name = 3;
getNameBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
optional string name = 3;
getNames() - Method in class org.apache.spark.ml.feature.VectorSlicer
 
getNChannels(Row) - Static method in class org.apache.spark.ml.image.ImageSchema
Gets the number of channels in the image
getNode(int, Node) - Static method in class org.apache.spark.mllib.tree.model.Node
Traces down from a root node to get the node with the given node index.
getNode() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper.Builder
.org.apache.spark.status.protobuf.SparkPlanGraphNode node = 1;
getNode() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper
.org.apache.spark.status.protobuf.SparkPlanGraphNode node = 1;
getNode() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapperOrBuilder
.org.apache.spark.status.protobuf.SparkPlanGraphNode node = 1;
getNodeBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper.Builder
.org.apache.spark.status.protobuf.SparkPlanGraphNode node = 1;
getNodeOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper.Builder
.org.apache.spark.status.protobuf.SparkPlanGraphNode node = 1;
getNodeOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper
.org.apache.spark.status.protobuf.SparkPlanGraphNode node = 1;
getNodeOrBuilder() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapperOrBuilder
.org.apache.spark.status.protobuf.SparkPlanGraphNode node = 1;
getNodes(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 4;
getNodes(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 4;
getNodes(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapperOrBuilder
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 4;
getNodes(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 2;
getNodes(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 2;
getNodes(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapperOrBuilder
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 2;
getNodesBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 4;
getNodesBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 2;
getNodesBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 4;
getNodesBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 2;
getNodesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 4;
getNodesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 4;
getNodesCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapperOrBuilder
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 4;
getNodesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 2;
getNodesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 2;
getNodesCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapperOrBuilder
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 2;
getNodesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 4;
getNodesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 4;
getNodesList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapperOrBuilder
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 4;
getNodesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 2;
getNodesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 2;
getNodesList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapperOrBuilder
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 2;
getNodesOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 4;
getNodesOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 4;
getNodesOrBuilder(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapperOrBuilder
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 4;
getNodesOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 2;
getNodesOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 2;
getNodesOrBuilder(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapperOrBuilder
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 2;
getNodesOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 4;
getNodesOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 4;
getNodesOrBuilderList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapperOrBuilder
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 4;
getNodesOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 2;
getNodesOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 2;
getNodesOrBuilderList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapperOrBuilder
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 2;
getNonnegative() - Method in interface org.apache.spark.ml.recommendation.ALSParams
 
getNumActiveStages() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
int32 num_active_stages = 16;
getNumActiveStages() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
int32 num_active_stages = 16;
getNumActiveStages() - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataOrBuilder
int32 num_active_stages = 16;
getNumActiveTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
int32 num_active_tasks = 10;
getNumActiveTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
int32 num_active_tasks = 10;
getNumActiveTasks() - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataOrBuilder
int32 num_active_tasks = 10;
getNumActiveTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummary.Builder
int32 num_active_tasks = 2;
getNumActiveTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummary
int32 num_active_tasks = 2;
getNumActiveTasks() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryOrBuilder
int32 num_active_tasks = 2;
getNumActiveTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int32 num_active_tasks = 5;
getNumActiveTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
int32 num_active_tasks = 5;
getNumActiveTasks() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
int32 num_active_tasks = 5;
getNumber() - Method in enum org.apache.spark.status.protobuf.StoreTypes.DeterministicLevel
 
getNumber() - Method in enum org.apache.spark.status.protobuf.StoreTypes.JobExecutionStatus
 
getNumber() - Method in enum org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper.WrapperCase
 
getNumber() - Method in enum org.apache.spark.status.protobuf.StoreTypes.StageStatus
 
getNumBins() - Method in class org.apache.spark.ml.evaluation.BinaryClassificationEvaluator
 
getNumBins() - Method in class org.apache.spark.sql.util.NumericHistogram
Returns the number of bins.
getNumBuckets() - Method in interface org.apache.spark.ml.feature.QuantileDiscretizerBase
 
getNumBucketsArray() - Method in interface org.apache.spark.ml.feature.QuantileDiscretizerBase
 
getNumBytesWritten() - Method in interface org.apache.spark.shuffle.api.ShufflePartitionWriter
Returns the number of bytes written either by this writer's output stream opened by ShufflePartitionWriter.openStream() or the byte channel opened by ShufflePartitionWriter.openChannelWrapper().
getNumCachedPartitions() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
int32 num_cached_partitions = 4;
getNumCachedPartitions() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo
int32 num_cached_partitions = 4;
getNumCachedPartitions() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoOrBuilder
int32 num_cached_partitions = 4;
getNumClasses(StructField) - Static method in class org.apache.spark.ml.util.MetadataUtils
Examine a schema to identify the number of classes in a label column.
getNumClasses() - Method in class org.apache.spark.mllib.tree.configuration.Strategy
 
getNumCompletedIndices() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
int32 num_completed_indices = 15;
getNumCompletedIndices() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
int32 num_completed_indices = 15;
getNumCompletedIndices() - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataOrBuilder
int32 num_completed_indices = 15;
getNumCompletedIndices() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int32 num_completed_indices = 9;
getNumCompletedIndices() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
int32 num_completed_indices = 9;
getNumCompletedIndices() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
int32 num_completed_indices = 9;
getNumCompletedJobs() - Method in class org.apache.spark.status.protobuf.StoreTypes.AppSummary.Builder
int32 num_completed_jobs = 1;
getNumCompletedJobs() - Method in class org.apache.spark.status.protobuf.StoreTypes.AppSummary
int32 num_completed_jobs = 1;
getNumCompletedJobs() - Method in interface org.apache.spark.status.protobuf.StoreTypes.AppSummaryOrBuilder
int32 num_completed_jobs = 1;
getNumCompletedStages() - Method in class org.apache.spark.status.protobuf.StoreTypes.AppSummary.Builder
int32 num_completed_stages = 2;
getNumCompletedStages() - Method in class org.apache.spark.status.protobuf.StoreTypes.AppSummary
int32 num_completed_stages = 2;
getNumCompletedStages() - Method in interface org.apache.spark.status.protobuf.StoreTypes.AppSummaryOrBuilder
int32 num_completed_stages = 2;
getNumCompletedStages() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
int32 num_completed_stages = 17;
getNumCompletedStages() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
int32 num_completed_stages = 17;
getNumCompletedStages() - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataOrBuilder
int32 num_completed_stages = 17;
getNumCompletedTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
int32 num_completed_tasks = 11;
getNumCompletedTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
int32 num_completed_tasks = 11;
getNumCompletedTasks() - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataOrBuilder
int32 num_completed_tasks = 11;
getNumCompletedTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummary.Builder
int32 num_completed_tasks = 3;
getNumCompletedTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummary
int32 num_completed_tasks = 3;
getNumCompletedTasks() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryOrBuilder
int32 num_completed_tasks = 3;
getNumCompleteTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int32 num_complete_tasks = 6;
getNumCompleteTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
int32 num_complete_tasks = 6;
getNumCompleteTasks() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
int32 num_complete_tasks = 6;
getNumFailedStages() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
int32 num_failed_stages = 19;
getNumFailedStages() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
int32 num_failed_stages = 19;
getNumFailedStages() - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataOrBuilder
int32 num_failed_stages = 19;
getNumFailedTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
int32 num_failed_tasks = 13;
getNumFailedTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
int32 num_failed_tasks = 13;
getNumFailedTasks() - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataOrBuilder
int32 num_failed_tasks = 13;
getNumFailedTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummary.Builder
int32 num_failed_tasks = 4;
getNumFailedTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummary
int32 num_failed_tasks = 4;
getNumFailedTasks() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryOrBuilder
int32 num_failed_tasks = 4;
getNumFailedTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int32 num_failed_tasks = 7;
getNumFailedTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
int32 num_failed_tasks = 7;
getNumFailedTasks() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
int32 num_failed_tasks = 7;
getNumFeatures() - Method in interface org.apache.spark.ml.param.shared.HasNumFeatures
 
getNumFeatures(StructField) - Static method in class org.apache.spark.ml.util.MetadataUtils
Examine a schema to identify the number of features in a vector column.
getNumFeatures() - Method in class org.apache.spark.mllib.regression.GeneralizedLinearAlgorithm
The dimension of training features.
getNumFolds() - Method in interface org.apache.spark.ml.tuning.CrossValidatorParams
 
getNumHashTables() - Method in interface org.apache.spark.ml.feature.LSHParams
 
getNumInputRows() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
int64 num_input_rows = 5;
getNumInputRows() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress
int64 num_input_rows = 5;
getNumInputRows() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SourceProgressOrBuilder
int64 num_input_rows = 5;
getNumItemBlocks() - Method in interface org.apache.spark.ml.recommendation.ALSParams
 
getNumIterations() - Method in class org.apache.spark.mllib.tree.configuration.BoostingStrategy
 
getNumKilledTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
int32 num_killed_tasks = 14;
getNumKilledTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
int32 num_killed_tasks = 14;
getNumKilledTasks() - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataOrBuilder
int32 num_killed_tasks = 14;
getNumKilledTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummary.Builder
int32 num_killed_tasks = 5;
getNumKilledTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummary
int32 num_killed_tasks = 5;
getNumKilledTasks() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryOrBuilder
int32 num_killed_tasks = 5;
getNumKilledTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int32 num_killed_tasks = 8;
getNumKilledTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
int32 num_killed_tasks = 8;
getNumKilledTasks() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
int32 num_killed_tasks = 8;
getNumObjFields() - Method in class org.apache.spark.serializer.SerializationDebugger.ObjectStreamClassMethods
 
getNumOutputRows() - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress.Builder
int64 num_output_rows = 2;
getNumOutputRows() - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress
int64 num_output_rows = 2;
getNumOutputRows() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SinkProgressOrBuilder
int64 num_output_rows = 2;
getNumPartitions() - Method in interface org.apache.spark.api.java.JavaRDDLike
Return the number of partitions in this RDD.
getNumPartitions() - Method in interface org.apache.spark.ml.feature.Word2VecBase
 
getNumPartitions() - Method in interface org.apache.spark.ml.fpm.FPGrowthParams
 
getNumPartitions() - Method in class org.apache.spark.rdd.RDD
Returns the number of partitions of this RDD.
getNumPartitions() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
int32 num_partitions = 3;
getNumPartitions() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo
int32 num_partitions = 3;
getNumPartitions() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoOrBuilder
int32 num_partitions = 3;
getNumRowsDroppedByWatermark() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
int64 num_rows_dropped_by_watermark = 9;
getNumRowsDroppedByWatermark() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress
int64 num_rows_dropped_by_watermark = 9;
getNumRowsDroppedByWatermark() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgressOrBuilder
int64 num_rows_dropped_by_watermark = 9;
getNumRowsRemoved() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
int64 num_rows_removed = 5;
getNumRowsRemoved() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress
int64 num_rows_removed = 5;
getNumRowsRemoved() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgressOrBuilder
int64 num_rows_removed = 5;
getNumRowsTotal() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
int64 num_rows_total = 2;
getNumRowsTotal() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress
int64 num_rows_total = 2;
getNumRowsTotal() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgressOrBuilder
int64 num_rows_total = 2;
getNumRowsUpdated() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
int64 num_rows_updated = 3;
getNumRowsUpdated() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress
int64 num_rows_updated = 3;
getNumRowsUpdated() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgressOrBuilder
int64 num_rows_updated = 3;
getNumShufflePartitions() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
int64 num_shuffle_partitions = 10;
getNumShufflePartitions() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress
int64 num_shuffle_partitions = 10;
getNumShufflePartitions() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgressOrBuilder
int64 num_shuffle_partitions = 10;
getNumSkippedStages() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
int32 num_skipped_stages = 18;
getNumSkippedStages() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
int32 num_skipped_stages = 18;
getNumSkippedStages() - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataOrBuilder
int32 num_skipped_stages = 18;
getNumSkippedTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
int32 num_skipped_tasks = 12;
getNumSkippedTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
int32 num_skipped_tasks = 12;
getNumSkippedTasks() - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataOrBuilder
int32 num_skipped_tasks = 12;
getNumStateStoreInstances() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
int64 num_state_store_instances = 11;
getNumStateStoreInstances() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress
int64 num_state_store_instances = 11;
getNumStateStoreInstances() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgressOrBuilder
int64 num_state_store_instances = 11;
getNumTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
int32 num_tasks = 9;
getNumTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
int32 num_tasks = 9;
getNumTasks() - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataOrBuilder
int32 num_tasks = 9;
getNumTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummary.Builder
int32 num_tasks = 1;
getNumTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummary
int32 num_tasks = 1;
getNumTasks() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryOrBuilder
int32 num_tasks = 1;
getNumTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int32 num_tasks = 4;
getNumTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
int32 num_tasks = 4;
getNumTasks() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
int32 num_tasks = 4;
getNumTopFeatures() - Method in interface org.apache.spark.ml.feature.SelectorParams
 
getNumTrees() - Method in class org.apache.spark.ml.classification.GBTClassificationModel
Number of trees in ensemble
getNumTrees() - Method in class org.apache.spark.ml.regression.GBTRegressionModel
Number of trees in ensemble
getNumTrees() - Method in interface org.apache.spark.ml.tree.RandomForestParams
 
getNumUserBlocks() - Method in interface org.apache.spark.ml.recommendation.ALSParams
 
getNumValues() - Method in class org.apache.spark.ml.attribute.NominalAttribute
Get the number of values, either from numValues or from values.
getObjFieldValues(Object, Object[]) - Method in class org.apache.spark.serializer.SerializationDebugger.ObjectStreamClassMethods
 
getObservedMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
Deprecated.
getObservedMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
Deprecated.
getObservedMetrics() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
Deprecated.
getObservedMetricsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
 
getObservedMetricsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
 
getObservedMetricsCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
map<string, string> observed_metrics = 12;
getObservedMetricsMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
map<string, string> observed_metrics = 12;
getObservedMetricsMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
map<string, string> observed_metrics = 12;
getObservedMetricsMap() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
map<string, string> observed_metrics = 12;
getObservedMetricsOrDefault(String, String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
map<string, string> observed_metrics = 12;
getObservedMetricsOrDefault(String, String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
map<string, string> observed_metrics = 12;
getObservedMetricsOrDefault(String, String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
map<string, string> observed_metrics = 12;
getObservedMetricsOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
map<string, string> observed_metrics = 12;
getObservedMetricsOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
map<string, string> observed_metrics = 12;
getObservedMetricsOrThrow(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
map<string, string> observed_metrics = 12;
getOffHeapMemoryRemaining() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution.Builder
optional int64 off_heap_memory_remaining = 8;
getOffHeapMemoryRemaining() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution
optional int64 off_heap_memory_remaining = 8;
getOffHeapMemoryRemaining() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDDataDistributionOrBuilder
optional int64 off_heap_memory_remaining = 8;
getOffHeapMemoryUsed() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution.Builder
optional int64 off_heap_memory_used = 6;
getOffHeapMemoryUsed() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution
optional int64 off_heap_memory_used = 6;
getOffHeapMemoryUsed() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDDataDistributionOrBuilder
optional int64 off_heap_memory_used = 6;
getOffset() - Method in interface org.apache.spark.sql.connector.read.streaming.ContinuousPartitionReader
Get the offset of the current record, or the start offset if no records have been read.
getOffsetClause(Integer) - Static method in class org.apache.spark.sql.jdbc.DB2Dialect
 
getOffsetClause(Integer) - Static method in class org.apache.spark.sql.jdbc.DerbyDialect
 
getOffsetClause(Integer) - Static method in class org.apache.spark.sql.jdbc.H2Dialect
 
getOffsetClause(Integer) - Method in class org.apache.spark.sql.jdbc.JdbcDialect
returns the OFFSET clause for the SELECT statement
getOffsetClause(Integer) - Static method in class org.apache.spark.sql.jdbc.MsSqlServerDialect
 
getOffsetClause(Integer) - Static method in class org.apache.spark.sql.jdbc.MySQLDialect
 
getOffsetClause(Integer) - Static method in class org.apache.spark.sql.jdbc.NoopDialect
 
getOffsetClause(Integer) - Static method in class org.apache.spark.sql.jdbc.OracleDialect
 
getOffsetClause(Integer) - Static method in class org.apache.spark.sql.jdbc.PostgresDialect
 
getOffsetClause(Integer) - Static method in class org.apache.spark.sql.jdbc.TeradataDialect
 
getOffsetCol() - Method in interface org.apache.spark.ml.regression.GeneralizedLinearRegressionBase
 
getOldBoostingStrategy(Map<Object, Object>, Enumeration.Value) - Method in interface org.apache.spark.ml.tree.GBTParams
(private[ml]) Create a BoostingStrategy instance to use with the old API.
getOldDocConcentration() - Method in interface org.apache.spark.ml.clustering.LDAParams
Get docConcentration used by spark.mllib LDA
getOldImpurity() - Method in interface org.apache.spark.ml.tree.HasVarianceImpurity
Convert new impurity to old impurity.
getOldImpurity() - Method in interface org.apache.spark.ml.tree.TreeClassifierParams
Convert new impurity to old impurity.
getOldLossType() - Method in interface org.apache.spark.ml.tree.GBTClassifierParams
(private[ml]) Convert new loss to old loss.
getOldLossType() - Method in interface org.apache.spark.ml.tree.GBTParams
Get old Gradient Boosting Loss type
getOldLossType() - Method in interface org.apache.spark.ml.tree.GBTRegressorParams
(private[ml]) Convert new loss to old loss.
getOldOptimizer() - Method in interface org.apache.spark.ml.clustering.LDAParams
 
getOldStrategy(Map<Object, Object>, int, Enumeration.Value, Impurity, double) - Method in interface org.apache.spark.ml.tree.DecisionTreeParams
(private[ml]) Create a Strategy instance to use with the old API.
getOldStrategy(Map<Object, Object>, int, Enumeration.Value, Impurity) - Method in interface org.apache.spark.ml.tree.TreeEnsembleParams
Create a Strategy instance to use with the old API.
getOldTopicConcentration() - Method in interface org.apache.spark.ml.clustering.LDAParams
Get topicConcentration used by spark.mllib LDA
getOnHeapMemoryRemaining() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution.Builder
optional int64 on_heap_memory_remaining = 7;
getOnHeapMemoryRemaining() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution
optional int64 on_heap_memory_remaining = 7;
getOnHeapMemoryRemaining() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDDataDistributionOrBuilder
optional int64 on_heap_memory_remaining = 7;
getOnHeapMemoryUsed() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution.Builder
optional int64 on_heap_memory_used = 5;
getOnHeapMemoryUsed() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution
optional int64 on_heap_memory_used = 5;
getOnHeapMemoryUsed() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDDataDistributionOrBuilder
optional int64 on_heap_memory_used = 5;
getOperatorName() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
optional string operator_name = 1;
getOperatorName() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress
optional string operator_name = 1;
getOperatorName() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgressOrBuilder
optional string operator_name = 1;
getOperatorNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
optional string operator_name = 1;
getOperatorNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress
optional string operator_name = 1;
getOperatorNameBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgressOrBuilder
optional string operator_name = 1;
getOptimizeDocConcentration() - Method in interface org.apache.spark.ml.clustering.LDAParams
 
getOptimizeDocConcentration() - Method in class org.apache.spark.mllib.clustering.OnlineLDAOptimizer
Optimize docConcentration, indicates whether docConcentration (Dirichlet parameter for document-topic distribution) will be optimized during training.
getOptimizer() - Method in interface org.apache.spark.ml.clustering.LDAParams
 
getOptimizer() - Method in class org.apache.spark.mllib.clustering.LDA
LDAOptimizer used to perform the actual calculation
getOption(String) - Method in class org.apache.spark.SparkConf
Get a parameter as an Option
getOption(String) - Method in class org.apache.spark.sql.RuntimeConfig
Returns the value of Spark runtime configuration property for the given key.
getOption() - Method in interface org.apache.spark.sql.streaming.GroupState
Get the state value as a scala Option.
getOption() - Method in class org.apache.spark.streaming.State
Get the state as a scala.Option.
getOptional(boolean, Function0<T>) - Static method in class org.apache.spark.status.protobuf.Utils
 
getOrCreate(SparkConf) - Static method in class org.apache.spark.SparkContext
This function may be used to get or instantiate a SparkContext and register it as a singleton object.
getOrCreate() - Static method in class org.apache.spark.SparkContext
This function may be used to get or instantiate a SparkContext and register it as a singleton object.
getOrCreate() - Method in class org.apache.spark.sql.SparkSession.Builder
Gets an existing SparkSession or, if there is no existing one, creates a new one based on the options set in this builder.
getOrCreate(SparkContext) - Static method in class org.apache.spark.sql.SQLContext
Deprecated.
Use SparkSession.builder instead. Since 2.0.0.
getOrCreate(String, Function0<JavaStreamingContext>) - Static method in class org.apache.spark.streaming.api.java.JavaStreamingContext
Deprecated.
Either recreate a StreamingContext from checkpoint data or create a new StreamingContext.
getOrCreate(String, Function0<JavaStreamingContext>, Configuration) - Static method in class org.apache.spark.streaming.api.java.JavaStreamingContext
Deprecated.
Either recreate a StreamingContext from checkpoint data or create a new StreamingContext.
getOrCreate(String, Function0<JavaStreamingContext>, Configuration, boolean) - Static method in class org.apache.spark.streaming.api.java.JavaStreamingContext
Deprecated.
Either recreate a StreamingContext from checkpoint data or create a new StreamingContext.
getOrCreate(String, Function0<StreamingContext>, Configuration, boolean) - Static method in class org.apache.spark.streaming.StreamingContext
Deprecated.
Either recreate a StreamingContext from checkpoint data or create a new StreamingContext.
getOrCreateSparkSession(JavaSparkContext, Map<Object, Object>, boolean) - Static method in class org.apache.spark.sql.api.r.SQLUtils
 
getOrDefault(Param<T>) - Method in interface org.apache.spark.ml.param.Params
Gets the value of a param in the embedded param map or its default value.
getOrDiscoverAllResources(SparkConf, String, Option<String>) - Static method in class org.apache.spark.resource.ResourceUtils
Gets all allocated resource information for the input component from input resources file and the application level Spark configs.
getOrDiscoverAllResourcesForResourceProfile(Option<String>, String, ResourceProfile, SparkConf) - Static method in class org.apache.spark.resource.ResourceUtils
This function is similar to getOrDiscoverallResources, except for it uses the ResourceProfile information instead of the application level configs.
getOrElse(Param<T>, T) - Method in class org.apache.spark.ml.param.ParamMap
Returns the value associated with a param or a default value.
getOrigin(Row) - Static method in class org.apache.spark.ml.image.ImageSchema
Gets the origin of the image
getOutgoingEdges(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationEdge outgoing_edges = 3;
getOutgoingEdges(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper
repeated .org.apache.spark.status.protobuf.RDDOperationEdge outgoing_edges = 3;
getOutgoingEdges(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapperOrBuilder
repeated .org.apache.spark.status.protobuf.RDDOperationEdge outgoing_edges = 3;
getOutgoingEdgesBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationEdge outgoing_edges = 3;
getOutgoingEdgesBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationEdge outgoing_edges = 3;
getOutgoingEdgesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationEdge outgoing_edges = 3;
getOutgoingEdgesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper
repeated .org.apache.spark.status.protobuf.RDDOperationEdge outgoing_edges = 3;
getOutgoingEdgesCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapperOrBuilder
repeated .org.apache.spark.status.protobuf.RDDOperationEdge outgoing_edges = 3;
getOutgoingEdgesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationEdge outgoing_edges = 3;
getOutgoingEdgesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper
repeated .org.apache.spark.status.protobuf.RDDOperationEdge outgoing_edges = 3;
getOutgoingEdgesList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapperOrBuilder
repeated .org.apache.spark.status.protobuf.RDDOperationEdge outgoing_edges = 3;
getOutgoingEdgesOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationEdge outgoing_edges = 3;
getOutgoingEdgesOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper
repeated .org.apache.spark.status.protobuf.RDDOperationEdge outgoing_edges = 3;
getOutgoingEdgesOrBuilder(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapperOrBuilder
repeated .org.apache.spark.status.protobuf.RDDOperationEdge outgoing_edges = 3;
getOutgoingEdgesOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
repeated .org.apache.spark.status.protobuf.RDDOperationEdge outgoing_edges = 3;
getOutgoingEdgesOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper
repeated .org.apache.spark.status.protobuf.RDDOperationEdge outgoing_edges = 3;
getOutgoingEdgesOrBuilderList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapperOrBuilder
repeated .org.apache.spark.status.protobuf.RDDOperationEdge outgoing_edges = 3;
getOutputAttrGroupFromData(Dataset<?>, Seq<String>, Seq<String>, boolean) - Static method in class org.apache.spark.ml.feature.OneHotEncoderCommon
This method is called when we want to generate AttributeGroup from actual data for one-hot encoder.
getOutputBytes(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double output_bytes = 8;
getOutputBytes(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
repeated double output_bytes = 8;
getOutputBytes(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
repeated double output_bytes = 8;
getOutputBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
int64 output_bytes = 7;
getOutputBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary
int64 output_bytes = 7;
getOutputBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryOrBuilder
int64 output_bytes = 7;
getOutputBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int64 output_bytes = 26;
getOutputBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
int64 output_bytes = 26;
getOutputBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
int64 output_bytes = 26;
getOutputBytesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double output_bytes = 8;
getOutputBytesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
repeated double output_bytes = 8;
getOutputBytesCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
repeated double output_bytes = 8;
getOutputBytesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double output_bytes = 8;
getOutputBytesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
repeated double output_bytes = 8;
getOutputBytesList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
repeated double output_bytes = 8;
getOutputBytesWritten() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 output_bytes_written = 28;
getOutputBytesWritten() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
int64 output_bytes_written = 28;
getOutputBytesWritten() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
int64 output_bytes_written = 28;
getOutputCol() - Method in interface org.apache.spark.ml.param.shared.HasOutputCol
 
getOutputCols() - Method in interface org.apache.spark.ml.param.shared.HasOutputCols
 
getOutputDeterministicLevel() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode.Builder
.org.apache.spark.status.protobuf.DeterministicLevel output_deterministic_level = 6;
getOutputDeterministicLevel() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode
.org.apache.spark.status.protobuf.DeterministicLevel output_deterministic_level = 6;
getOutputDeterministicLevel() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationNodeOrBuilder
.org.apache.spark.status.protobuf.DeterministicLevel output_deterministic_level = 6;
getOutputDeterministicLevelValue() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode.Builder
.org.apache.spark.status.protobuf.DeterministicLevel output_deterministic_level = 6;
getOutputDeterministicLevelValue() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode
.org.apache.spark.status.protobuf.DeterministicLevel output_deterministic_level = 6;
getOutputDeterministicLevelValue() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationNodeOrBuilder
.org.apache.spark.status.protobuf.DeterministicLevel output_deterministic_level = 6;
getOutputMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
.org.apache.spark.status.protobuf.OutputMetricDistributions output_metrics = 16;
getOutputMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
.org.apache.spark.status.protobuf.OutputMetricDistributions output_metrics = 16;
getOutputMetrics() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
.org.apache.spark.status.protobuf.OutputMetricDistributions output_metrics = 16;
getOutputMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
.org.apache.spark.status.protobuf.OutputMetrics output_metrics = 12;
getOutputMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics
.org.apache.spark.status.protobuf.OutputMetrics output_metrics = 12;
getOutputMetrics() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricsOrBuilder
.org.apache.spark.status.protobuf.OutputMetrics output_metrics = 12;
getOutputMetricsBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
.org.apache.spark.status.protobuf.OutputMetricDistributions output_metrics = 16;
getOutputMetricsBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
.org.apache.spark.status.protobuf.OutputMetrics output_metrics = 12;
getOutputMetricsOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
.org.apache.spark.status.protobuf.OutputMetricDistributions output_metrics = 16;
getOutputMetricsOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
.org.apache.spark.status.protobuf.OutputMetricDistributions output_metrics = 16;
getOutputMetricsOrBuilder() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
.org.apache.spark.status.protobuf.OutputMetricDistributions output_metrics = 16;
getOutputMetricsOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
.org.apache.spark.status.protobuf.OutputMetrics output_metrics = 12;
getOutputMetricsOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics
.org.apache.spark.status.protobuf.OutputMetrics output_metrics = 12;
getOutputMetricsOrBuilder() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricsOrBuilder
.org.apache.spark.status.protobuf.OutputMetrics output_metrics = 12;
getOutputRecords(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double output_records = 9;
getOutputRecords(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
repeated double output_records = 9;
getOutputRecords(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
repeated double output_records = 9;
getOutputRecords() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
int64 output_records = 8;
getOutputRecords() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary
int64 output_records = 8;
getOutputRecords() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryOrBuilder
int64 output_records = 8;
getOutputRecords() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int64 output_records = 27;
getOutputRecords() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
int64 output_records = 27;
getOutputRecords() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
int64 output_records = 27;
getOutputRecordsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double output_records = 9;
getOutputRecordsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
repeated double output_records = 9;
getOutputRecordsCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
repeated double output_records = 9;
getOutputRecordsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double output_records = 9;
getOutputRecordsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
repeated double output_records = 9;
getOutputRecordsList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
repeated double output_records = 9;
getOutputRecordsWritten() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 output_records_written = 29;
getOutputRecordsWritten() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
int64 output_records_written = 29;
getOutputRecordsWritten() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
int64 output_records_written = 29;
getOutputSize(int) - Method in interface org.apache.spark.ml.ann.Layer
Returns the output size given the input size (not counting the stack size).
getOutputStream(String, Configuration) - Static method in class org.apache.spark.streaming.util.HdfsUtils
 
getP() - Method in class org.apache.spark.ml.feature.Normalizer
 
getParallelism() - Method in interface org.apache.spark.ml.param.shared.HasParallelism
 
getParam(String) - Method in interface org.apache.spark.ml.param.Params
Gets a param by its name.
getParameter(String) - Method in class org.apache.spark.ui.XssSafeRequest
 
getParameterMap() - Method in class org.apache.spark.ui.XssSafeRequest
 
getParameterNames() - Method in class org.apache.spark.ui.XssSafeRequest
 
getParameterOtherTable(HttpServletRequest, String) - Method in interface org.apache.spark.ui.PagedTable
Returns parameters of other tables in the page.
getParameterValues(String) - Method in class org.apache.spark.ui.XssSafeRequest
 
getParentLoggerNotImplementedError(String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
getParents(int) - Method in class org.apache.spark.NarrowDependency
Get the parent partitions for a child partition.
getParents(int) - Method in class org.apache.spark.OneToOneDependency
 
getParents(int) - Method in class org.apache.spark.RangeDependency
 
getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo
 
getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo
 
getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
 
getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoWrapper
 
getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo
 
getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoWrapper
 
getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.AppSummary
 
getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
 
getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetrics
 
getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
 
getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions
 
getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest
 
getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary
 
getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper
 
getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
 
getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryWrapper
 
getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributions
 
getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetrics
 
getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
 
getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper
 
getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.MemoryMetrics
 
getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributions
 
getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetrics
 
getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.PairStrings
 
getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.PoolData
 
getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary
 
getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryWrapper
 
getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution
 
getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper
 
getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationEdge
 
getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper
 
getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode
 
getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo
 
getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo
 
getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoWrapper
 
getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceInformation
 
getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo
 
getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileWrapper
 
getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.RuntimeInfo
 
getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions
 
getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics
 
getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions
 
getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics
 
getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions
 
getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetrics
 
getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress
 
getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress
 
getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper
 
getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphEdge
 
getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode
 
getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper
 
getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper
 
getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummary
 
getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryWrapper
 
getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
 
getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetric
 
getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
 
getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper
 
getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress
 
getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData
 
getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData
 
getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
 
getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressWrapper
 
getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
 
getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
 
getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
 
getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics
 
getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskResourceRequest
 
getPartition(long, long, int) - Method in class org.apache.spark.graphx.PartitionStrategy.CanonicalRandomVertexCut$
 
getPartition(long, long, int) - Method in class org.apache.spark.graphx.PartitionStrategy.EdgePartition1D$
 
getPartition(long, long, int) - Method in class org.apache.spark.graphx.PartitionStrategy.EdgePartition2D$
 
getPartition(long, long, int) - Method in interface org.apache.spark.graphx.PartitionStrategy
Returns the partition number for a given edge.
getPartition(long, long, int) - Method in class org.apache.spark.graphx.PartitionStrategy.RandomVertexCut$
 
getPartition(Object) - Method in class org.apache.spark.HashPartitioner
 
getPartition(Object) - Method in class org.apache.spark.Partitioner
 
getPartition(Object) - Method in class org.apache.spark.RangePartitioner
 
getPartitionId() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
int32 partition_id = 4;
getPartitionId() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
int32 partition_id = 4;
getPartitionId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataOrBuilder
int32 partition_id = 4;
getPartitionId() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int32 partition_id = 4;
getPartitionId() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
int32 partition_id = 4;
getPartitionId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
int32 partition_id = 4;
getPartitionId() - Static method in class org.apache.spark.TaskContext
Returns the partition id of currently active TaskContext.
getPartitionLengths() - Method in class org.apache.spark.shuffle.api.metadata.MapOutputCommitMessage
 
getPartitionMetadataByFilterError(InvocationTargetException) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
getPartitions() - Method in class org.apache.spark.api.r.BaseRRDD
 
getPartitions() - Method in class org.apache.spark.rdd.CoGroupedRDD
 
getPartitions() - Method in class org.apache.spark.rdd.DefaultPartitionCoalescer
 
getPartitions() - Method in class org.apache.spark.rdd.HadoopRDD
 
getPartitions() - Method in class org.apache.spark.rdd.JdbcRDD
 
getPartitions() - Method in class org.apache.spark.rdd.NewHadoopRDD
 
getPartitions() - Method in class org.apache.spark.rdd.ShuffledRDD
 
getPartitions() - Method in class org.apache.spark.rdd.UnionRDD
 
getPartitions() - Method in class org.apache.spark.status.LiveRDD
 
getPartitions(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
repeated .org.apache.spark.status.protobuf.RDDPartitionInfo partitions = 9;
getPartitions(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo
repeated .org.apache.spark.status.protobuf.RDDPartitionInfo partitions = 9;
getPartitions(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoOrBuilder
repeated .org.apache.spark.status.protobuf.RDDPartitionInfo partitions = 9;
getPartitionsBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
repeated .org.apache.spark.status.protobuf.RDDPartitionInfo partitions = 9;
getPartitionsBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
repeated .org.apache.spark.status.protobuf.RDDPartitionInfo partitions = 9;
getPartitionsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
repeated .org.apache.spark.status.protobuf.RDDPartitionInfo partitions = 9;
getPartitionsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo
repeated .org.apache.spark.status.protobuf.RDDPartitionInfo partitions = 9;
getPartitionsCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoOrBuilder
repeated .org.apache.spark.status.protobuf.RDDPartitionInfo partitions = 9;
getPartitionsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
repeated .org.apache.spark.status.protobuf.RDDPartitionInfo partitions = 9;
getPartitionsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo
repeated .org.apache.spark.status.protobuf.RDDPartitionInfo partitions = 9;
getPartitionsList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoOrBuilder
repeated .org.apache.spark.status.protobuf.RDDPartitionInfo partitions = 9;
getPartitionsOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
repeated .org.apache.spark.status.protobuf.RDDPartitionInfo partitions = 9;
getPartitionsOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo
repeated .org.apache.spark.status.protobuf.RDDPartitionInfo partitions = 9;
getPartitionsOrBuilder(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoOrBuilder
repeated .org.apache.spark.status.protobuf.RDDPartitionInfo partitions = 9;
getPartitionsOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
repeated .org.apache.spark.status.protobuf.RDDPartitionInfo partitions = 9;
getPartitionsOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo
repeated .org.apache.spark.status.protobuf.RDDPartitionInfo partitions = 9;
getPartitionsOrBuilderList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoOrBuilder
repeated .org.apache.spark.status.protobuf.RDDPartitionInfo partitions = 9;
getPartitionWriter(int) - Method in interface org.apache.spark.shuffle.api.ShuffleMapOutputWriter
Creates a writer that can open an output stream to persist bytes targeted for a given reduce partition id.
getPath() - Method in class org.apache.spark.input.PortableDataStream
 
getPattern() - Method in class org.apache.spark.ml.feature.RegexTokenizer
 
getPeakExecutionMemory() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double peak_execution_memory = 15;
getPeakExecutionMemory() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
double peak_execution_memory = 15;
getPeakExecutionMemory() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
double peak_execution_memory = 15;
getPeakExecutionMemory() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int64 peak_execution_memory = 23;
getPeakExecutionMemory() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
int64 peak_execution_memory = 23;
getPeakExecutionMemory() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
int64 peak_execution_memory = 23;
getPeakExecutionMemory() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 peak_execution_memory = 25;
getPeakExecutionMemory() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
int64 peak_execution_memory = 25;
getPeakExecutionMemory() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
int64 peak_execution_memory = 25;
getPeakExecutionMemory(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double peak_execution_memory = 12;
getPeakExecutionMemory(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
repeated double peak_execution_memory = 12;
getPeakExecutionMemory(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
repeated double peak_execution_memory = 12;
getPeakExecutionMemory() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
int64 peak_execution_memory = 10;
getPeakExecutionMemory() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics
int64 peak_execution_memory = 10;
getPeakExecutionMemory() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricsOrBuilder
int64 peak_execution_memory = 10;
getPeakExecutionMemoryCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double peak_execution_memory = 12;
getPeakExecutionMemoryCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
repeated double peak_execution_memory = 12;
getPeakExecutionMemoryCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
repeated double peak_execution_memory = 12;
getPeakExecutionMemoryList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double peak_execution_memory = 12;
getPeakExecutionMemoryList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
repeated double peak_execution_memory = 12;
getPeakExecutionMemoryList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
repeated double peak_execution_memory = 12;
getPeakExecutorMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
optional .org.apache.spark.status.protobuf.ExecutorMetrics peak_executor_metrics = 50;
getPeakExecutorMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
optional .org.apache.spark.status.protobuf.ExecutorMetrics peak_executor_metrics = 50;
getPeakExecutorMetrics() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
optional .org.apache.spark.status.protobuf.ExecutorMetrics peak_executor_metrics = 50;
getPeakExecutorMetricsBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
optional .org.apache.spark.status.protobuf.ExecutorMetrics peak_executor_metrics = 50;
getPeakExecutorMetricsOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
optional .org.apache.spark.status.protobuf.ExecutorMetrics peak_executor_metrics = 50;
getPeakExecutorMetricsOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
optional .org.apache.spark.status.protobuf.ExecutorMetrics peak_executor_metrics = 50;
getPeakExecutorMetricsOrBuilder() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
optional .org.apache.spark.status.protobuf.ExecutorMetrics peak_executor_metrics = 50;
getPeakMemoryMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
.org.apache.spark.status.protobuf.ExecutorPeakMetricsDistributions peak_memory_metrics = 16;
getPeakMemoryMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
.org.apache.spark.status.protobuf.ExecutorPeakMetricsDistributions peak_memory_metrics = 16;
getPeakMemoryMetrics() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
.org.apache.spark.status.protobuf.ExecutorPeakMetricsDistributions peak_memory_metrics = 16;
getPeakMemoryMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
optional .org.apache.spark.status.protobuf.ExecutorMetrics peak_memory_metrics = 16;
getPeakMemoryMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary
optional .org.apache.spark.status.protobuf.ExecutorMetrics peak_memory_metrics = 16;
getPeakMemoryMetrics() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryOrBuilder
optional .org.apache.spark.status.protobuf.ExecutorMetrics peak_memory_metrics = 16;
getPeakMemoryMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
optional .org.apache.spark.status.protobuf.ExecutorMetrics peak_memory_metrics = 26;
getPeakMemoryMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
optional .org.apache.spark.status.protobuf.ExecutorMetrics peak_memory_metrics = 26;
getPeakMemoryMetrics() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
optional .org.apache.spark.status.protobuf.ExecutorMetrics peak_memory_metrics = 26;
getPeakMemoryMetricsBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
.org.apache.spark.status.protobuf.ExecutorPeakMetricsDistributions peak_memory_metrics = 16;
getPeakMemoryMetricsBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
optional .org.apache.spark.status.protobuf.ExecutorMetrics peak_memory_metrics = 16;
getPeakMemoryMetricsBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
optional .org.apache.spark.status.protobuf.ExecutorMetrics peak_memory_metrics = 26;
getPeakMemoryMetricsOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
.org.apache.spark.status.protobuf.ExecutorPeakMetricsDistributions peak_memory_metrics = 16;
getPeakMemoryMetricsOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
.org.apache.spark.status.protobuf.ExecutorPeakMetricsDistributions peak_memory_metrics = 16;
getPeakMemoryMetricsOrBuilder() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
.org.apache.spark.status.protobuf.ExecutorPeakMetricsDistributions peak_memory_metrics = 16;
getPeakMemoryMetricsOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
optional .org.apache.spark.status.protobuf.ExecutorMetrics peak_memory_metrics = 16;
getPeakMemoryMetricsOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary
optional .org.apache.spark.status.protobuf.ExecutorMetrics peak_memory_metrics = 16;
getPeakMemoryMetricsOrBuilder() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryOrBuilder
optional .org.apache.spark.status.protobuf.ExecutorMetrics peak_memory_metrics = 16;
getPeakMemoryMetricsOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
optional .org.apache.spark.status.protobuf.ExecutorMetrics peak_memory_metrics = 26;
getPeakMemoryMetricsOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
optional .org.apache.spark.status.protobuf.ExecutorMetrics peak_memory_metrics = 26;
getPeakMemoryMetricsOrBuilder() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
optional .org.apache.spark.status.protobuf.ExecutorMetrics peak_memory_metrics = 26;
GetPeers(BlockManagerId) - Constructor for class org.apache.spark.storage.BlockManagerMessages.GetPeers
 
GetPeers$() - Constructor for class org.apache.spark.storage.BlockManagerMessages.GetPeers$
 
getPercentile() - Method in interface org.apache.spark.ml.feature.SelectorParams
 
getPersistentRDDs() - Method in class org.apache.spark.api.java.JavaSparkContext
Returns a Java map of JavaRDDs that have marked themselves as persistent via cache() call.
getPersistentRDDs() - Method in class org.apache.spark.SparkContext
Returns an immutable map of RDDs that have marked themselves as persistent via cache() call.
getPhysicalPlanDescription() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
optional string physical_plan_description = 5;
getPhysicalPlanDescription() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
optional string physical_plan_description = 5;
getPhysicalPlanDescription() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
optional string physical_plan_description = 5;
getPhysicalPlanDescriptionBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
optional string physical_plan_description = 5;
getPhysicalPlanDescriptionBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
optional string physical_plan_description = 5;
getPhysicalPlanDescriptionBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
optional string physical_plan_description = 5;
getPmml() - Method in interface org.apache.spark.mllib.pmml.export.PMMLModelExport
 
getPoissonSamplingFunction(RDD<Tuple2<K, V>>, Map<K, Object>, boolean, long, ClassTag<K>, ClassTag<V>) - Static method in class org.apache.spark.util.random.StratifiedSamplingUtils
Return the per partition sampling function used for sampling with replacement.
getPoolForName(String) - Method in class org.apache.spark.SparkContext
:: DeveloperApi :: Return the pool associated with the given name, if one exists
getPosition() - Method in class org.apache.spark.streaming.kinesis.KinesisInitialPositions.AtTimestamp
 
getPosition() - Method in class org.apache.spark.streaming.kinesis.KinesisInitialPositions.Latest
 
getPosition() - Method in class org.apache.spark.streaming.kinesis.KinesisInitialPositions.TrimHorizon
 
getPowerIterationClustering(int, String, int, String, String, String) - Static method in class org.apache.spark.ml.r.PowerIterationClusteringWrapper
 
getPredictionCol() - Method in interface org.apache.spark.ml.param.shared.HasPredictionCol
 
getPreferredLocations(Partition) - Method in class org.apache.spark.rdd.HadoopRDD
 
getPreferredLocations(Partition) - Method in class org.apache.spark.rdd.NewHadoopRDD
 
getPreferredLocations(Partition) - Method in class org.apache.spark.rdd.UnionRDD
 
getPrefixSpan(double, int, double, String) - Static method in class org.apache.spark.ml.r.PrefixSpanWrapper
 
getProbabilityCol() - Method in interface org.apache.spark.ml.param.shared.HasProbabilityCol
 
getProcessedRowsPerSecond() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
double processed_rows_per_second = 7;
getProcessedRowsPerSecond() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress
double processed_rows_per_second = 7;
getProcessedRowsPerSecond() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SourceProgressOrBuilder
double processed_rows_per_second = 7;
getProcessLogs() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder
Deprecated.
getProcessLogs() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary
Deprecated.
getProcessLogs() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryOrBuilder
Deprecated.
getProcessLogsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder
 
getProcessLogsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary
 
getProcessLogsCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryOrBuilder
map<string, string> process_logs = 7;
getProcessLogsMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder
map<string, string> process_logs = 7;
getProcessLogsMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary
map<string, string> process_logs = 7;
getProcessLogsMap() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryOrBuilder
map<string, string> process_logs = 7;
getProcessLogsOrDefault(String, String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder
map<string, string> process_logs = 7;
getProcessLogsOrDefault(String, String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary
map<string, string> process_logs = 7;
getProcessLogsOrDefault(String, String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryOrBuilder
map<string, string> process_logs = 7;
getProcessLogsOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder
map<string, string> process_logs = 7;
getProcessLogsOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary
map<string, string> process_logs = 7;
getProcessLogsOrThrow(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryOrBuilder
map<string, string> process_logs = 7;
getProcessName() - Static method in class org.apache.spark.util.Utils
Returns the name of this JVM process.
getProgress() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressWrapper.Builder
.org.apache.spark.status.protobuf.StreamingQueryProgress progress = 1;
getProgress() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressWrapper
.org.apache.spark.status.protobuf.StreamingQueryProgress progress = 1;
getProgress() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressWrapperOrBuilder
.org.apache.spark.status.protobuf.StreamingQueryProgress progress = 1;
getProgressBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressWrapper.Builder
.org.apache.spark.status.protobuf.StreamingQueryProgress progress = 1;
getProgressOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressWrapper.Builder
.org.apache.spark.status.protobuf.StreamingQueryProgress progress = 1;
getProgressOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressWrapper
.org.apache.spark.status.protobuf.StreamingQueryProgress progress = 1;
getProgressOrBuilder() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressWrapperOrBuilder
.org.apache.spark.status.protobuf.StreamingQueryProgress progress = 1;
getPropertiesFromFile(String) - Static method in class org.apache.spark.util.Utils
Load properties present in the given file.
getPythonRunnerConfMap(SQLConf) - Static method in class org.apache.spark.sql.util.ArrowUtils
Return Map with conf settings to be used in ArrowPythonRunner
getQuantile() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
optional string quantile = 3;
getQuantile() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
optional string quantile = 3;
getQuantile() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
optional string quantile = 3;
getQuantileBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
optional string quantile = 3;
getQuantileBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
optional string quantile = 3;
getQuantileBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
optional string quantile = 3;
getQuantileCalculationStrategy() - Method in class org.apache.spark.mllib.tree.configuration.Strategy
 
getQuantileProbabilities() - Method in interface org.apache.spark.ml.regression.AFTSurvivalRegressionParams
 
getQuantiles(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double quantiles = 1;
getQuantiles(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
repeated double quantiles = 1;
getQuantiles(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
repeated double quantiles = 1;
getQuantiles(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions.Builder
repeated double quantiles = 1;
getQuantiles(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions
repeated double quantiles = 1;
getQuantiles(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributionsOrBuilder
repeated double quantiles = 1;
getQuantiles(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double quantiles = 1;
getQuantiles(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
repeated double quantiles = 1;
getQuantiles(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
repeated double quantiles = 1;
getQuantilesCol() - Method in interface org.apache.spark.ml.regression.AFTSurvivalRegressionParams
 
getQuantilesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double quantiles = 1;
getQuantilesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
repeated double quantiles = 1;
getQuantilesCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
repeated double quantiles = 1;
getQuantilesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions.Builder
repeated double quantiles = 1;
getQuantilesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions
repeated double quantiles = 1;
getQuantilesCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributionsOrBuilder
repeated double quantiles = 1;
getQuantilesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double quantiles = 1;
getQuantilesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
repeated double quantiles = 1;
getQuantilesCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
repeated double quantiles = 1;
getQuantilesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double quantiles = 1;
getQuantilesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
repeated double quantiles = 1;
getQuantilesList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
repeated double quantiles = 1;
getQuantilesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions.Builder
repeated double quantiles = 1;
getQuantilesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions
repeated double quantiles = 1;
getQuantilesList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributionsOrBuilder
repeated double quantiles = 1;
getQuantilesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double quantiles = 1;
getQuantilesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
repeated double quantiles = 1;
getQuantilesList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
repeated double quantiles = 1;
getQueryContext() - Method in exception org.apache.spark.SparkException
 
getQueryContext() - Method in interface org.apache.spark.SparkThrowable
 
getQueryContext() - Method in exception org.apache.spark.sql.AnalysisException
 
getQueryContext(SQLQueryContext) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
getQueryContext(SQLQueryContext) - Method in interface org.apache.spark.sql.errors.QueryErrorsBase
 
getQueryContext(SQLQueryContext) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
getQueryContext(SQLQueryContext) - Static method in class org.apache.spark.sql.errors.QueryParsingErrors
 
getQueryName(StreamingQueryUIData) - Static method in class org.apache.spark.sql.streaming.ui.UIUtils
 
getQueryStatus(StreamingQueryUIData) - Static method in class org.apache.spark.sql.streaming.ui.UIUtils
 
getRandomSample(Seq<T>, int, Random) - Static method in class org.apache.spark.storage.BlockReplicationUtils
Get a random sample of size m from the elems
getRank() - Method in interface org.apache.spark.ml.recommendation.ALSParams
 
getRatingCol() - Method in interface org.apache.spark.ml.recommendation.ALSParams
 
getRawPredictionCol() - Method in interface org.apache.spark.ml.param.shared.HasRawPredictionCol
 
getRddBlocks() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
int32 rdd_blocks = 4;
getRddBlocks() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
int32 rdd_blocks = 4;
getRddBlocks() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
int32 rdd_blocks = 4;
getRddIds(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
repeated int64 rdd_ids = 43;
getRddIds(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
repeated int64 rdd_ids = 43;
getRddIds(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
repeated int64 rdd_ids = 43;
getRddIdsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
repeated int64 rdd_ids = 43;
getRddIdsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
repeated int64 rdd_ids = 43;
getRddIdsCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
repeated int64 rdd_ids = 43;
getRddIdsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
repeated int64 rdd_ids = 43;
getRddIdsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
repeated int64 rdd_ids = 43;
getRddIdsList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
repeated int64 rdd_ids = 43;
getRDDStorageInfo() - Method in class org.apache.spark.SparkContext
:: DeveloperApi :: Return information about what RDDs are cached, if they are in mem or on disk, how much space they take, etc.
getReadBytes(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
repeated double read_bytes = 1;
getReadBytes(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions
repeated double read_bytes = 1;
getReadBytes(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributionsOrBuilder
repeated double read_bytes = 1;
getReadBytesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
repeated double read_bytes = 1;
getReadBytesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions
repeated double read_bytes = 1;
getReadBytesCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributionsOrBuilder
repeated double read_bytes = 1;
getReadBytesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
repeated double read_bytes = 1;
getReadBytesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions
repeated double read_bytes = 1;
getReadBytesList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributionsOrBuilder
repeated double read_bytes = 1;
getReadLimits() - Method in class org.apache.spark.sql.connector.read.streaming.CompositeReadLimit
 
getReadRecords(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
repeated double read_records = 2;
getReadRecords(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions
repeated double read_records = 2;
getReadRecords(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributionsOrBuilder
repeated double read_records = 2;
getReadRecordsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
repeated double read_records = 2;
getReadRecordsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions
repeated double read_records = 2;
getReadRecordsCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributionsOrBuilder
repeated double read_records = 2;
getReadRecordsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
repeated double read_records = 2;
getReadRecordsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions
repeated double read_records = 2;
getReadRecordsList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributionsOrBuilder
repeated double read_records = 2;
getReceiver() - Method in class org.apache.spark.streaming.dstream.ReceiverInputDStream
Gets the receiver object that will be sent to the worker nodes to receive data.
getRecordsRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double records_read = 19;
getRecordsRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
double records_read = 19;
getRecordsRead() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
double records_read = 19;
getRecordsRead(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributions.Builder
repeated double records_read = 2;
getRecordsRead(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributions
repeated double records_read = 2;
getRecordsRead(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributionsOrBuilder
repeated double records_read = 2;
getRecordsRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetrics.Builder
int64 records_read = 2;
getRecordsRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetrics
int64 records_read = 2;
getRecordsRead() - Method in interface org.apache.spark.status.protobuf.StoreTypes.InputMetricsOrBuilder
int64 records_read = 2;
getRecordsRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics.Builder
int64 records_read = 7;
getRecordsRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics
int64 records_read = 7;
getRecordsRead() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricsOrBuilder
int64 records_read = 7;
getRecordsReadCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributions.Builder
repeated double records_read = 2;
getRecordsReadCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributions
repeated double records_read = 2;
getRecordsReadCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributionsOrBuilder
repeated double records_read = 2;
getRecordsReadList() - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributions.Builder
repeated double records_read = 2;
getRecordsReadList() - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributions
repeated double records_read = 2;
getRecordsReadList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributionsOrBuilder
repeated double records_read = 2;
getRecordsWritten() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double records_written = 21;
getRecordsWritten() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
double records_written = 21;
getRecordsWritten() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
double records_written = 21;
getRecordsWritten(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributions.Builder
repeated double records_written = 2;
getRecordsWritten(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributions
repeated double records_written = 2;
getRecordsWritten(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributionsOrBuilder
repeated double records_written = 2;
getRecordsWritten() - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetrics.Builder
int64 records_written = 2;
getRecordsWritten() - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetrics
int64 records_written = 2;
getRecordsWritten() - Method in interface org.apache.spark.status.protobuf.StoreTypes.OutputMetricsOrBuilder
int64 records_written = 2;
getRecordsWritten() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetrics.Builder
int64 records_written = 3;
getRecordsWritten() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetrics
int64 records_written = 3;
getRecordsWritten() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricsOrBuilder
int64 records_written = 3;
getRecordsWrittenCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributions.Builder
repeated double records_written = 2;
getRecordsWrittenCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributions
repeated double records_written = 2;
getRecordsWrittenCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributionsOrBuilder
repeated double records_written = 2;
getRecordsWrittenList() - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributions.Builder
repeated double records_written = 2;
getRecordsWrittenList() - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributions
repeated double records_written = 2;
getRecordsWrittenList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributionsOrBuilder
repeated double records_written = 2;
getRegParam() - Method in interface org.apache.spark.ml.param.shared.HasRegParam
 
getRelativeError() - Method in interface org.apache.spark.ml.param.shared.HasRelativeError
 
getRemoteBlocksFetched(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
repeated double remote_blocks_fetched = 3;
getRemoteBlocksFetched(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions
repeated double remote_blocks_fetched = 3;
getRemoteBlocksFetched(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributionsOrBuilder
repeated double remote_blocks_fetched = 3;
getRemoteBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics.Builder
int64 remote_blocks_fetched = 1;
getRemoteBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics
int64 remote_blocks_fetched = 1;
getRemoteBlocksFetched() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricsOrBuilder
int64 remote_blocks_fetched = 1;
getRemoteBlocksFetchedCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
repeated double remote_blocks_fetched = 3;
getRemoteBlocksFetchedCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions
repeated double remote_blocks_fetched = 3;
getRemoteBlocksFetchedCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributionsOrBuilder
repeated double remote_blocks_fetched = 3;
getRemoteBlocksFetchedList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
repeated double remote_blocks_fetched = 3;
getRemoteBlocksFetchedList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions
repeated double remote_blocks_fetched = 3;
getRemoteBlocksFetchedList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributionsOrBuilder
repeated double remote_blocks_fetched = 3;
getRemoteBytesRead(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
repeated double remote_bytes_read = 6;
getRemoteBytesRead(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions
repeated double remote_bytes_read = 6;
getRemoteBytesRead(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributionsOrBuilder
repeated double remote_bytes_read = 6;
getRemoteBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics.Builder
int64 remote_bytes_read = 4;
getRemoteBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics
int64 remote_bytes_read = 4;
getRemoteBytesRead() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricsOrBuilder
int64 remote_bytes_read = 4;
getRemoteBytesReadCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
repeated double remote_bytes_read = 6;
getRemoteBytesReadCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions
repeated double remote_bytes_read = 6;
getRemoteBytesReadCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributionsOrBuilder
repeated double remote_bytes_read = 6;
getRemoteBytesReadList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
repeated double remote_bytes_read = 6;
getRemoteBytesReadList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions
repeated double remote_bytes_read = 6;
getRemoteBytesReadList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributionsOrBuilder
repeated double remote_bytes_read = 6;
getRemoteBytesReadToDisk(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
repeated double remote_bytes_read_to_disk = 7;
getRemoteBytesReadToDisk(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions
repeated double remote_bytes_read_to_disk = 7;
getRemoteBytesReadToDisk(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributionsOrBuilder
repeated double remote_bytes_read_to_disk = 7;
getRemoteBytesReadToDisk() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics.Builder
int64 remote_bytes_read_to_disk = 5;
getRemoteBytesReadToDisk() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics
int64 remote_bytes_read_to_disk = 5;
getRemoteBytesReadToDisk() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricsOrBuilder
int64 remote_bytes_read_to_disk = 5;
getRemoteBytesReadToDiskCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
repeated double remote_bytes_read_to_disk = 7;
getRemoteBytesReadToDiskCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions
repeated double remote_bytes_read_to_disk = 7;
getRemoteBytesReadToDiskCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributionsOrBuilder
repeated double remote_bytes_read_to_disk = 7;
getRemoteBytesReadToDiskList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
repeated double remote_bytes_read_to_disk = 7;
getRemoteBytesReadToDiskList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions
repeated double remote_bytes_read_to_disk = 7;
getRemoteBytesReadToDiskList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributionsOrBuilder
repeated double remote_bytes_read_to_disk = 7;
getRemoteMergedBlocksFetched(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
repeated double remote_merged_blocks_fetched = 3;
getRemoteMergedBlocksFetched(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions
repeated double remote_merged_blocks_fetched = 3;
getRemoteMergedBlocksFetched(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributionsOrBuilder
repeated double remote_merged_blocks_fetched = 3;
getRemoteMergedBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics.Builder
int64 remote_merged_blocks_fetched = 3;
getRemoteMergedBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics
int64 remote_merged_blocks_fetched = 3;
getRemoteMergedBlocksFetched() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricsOrBuilder
int64 remote_merged_blocks_fetched = 3;
getRemoteMergedBlocksFetchedCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
repeated double remote_merged_blocks_fetched = 3;
getRemoteMergedBlocksFetchedCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions
repeated double remote_merged_blocks_fetched = 3;
getRemoteMergedBlocksFetchedCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributionsOrBuilder
repeated double remote_merged_blocks_fetched = 3;
getRemoteMergedBlocksFetchedList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
repeated double remote_merged_blocks_fetched = 3;
getRemoteMergedBlocksFetchedList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions
repeated double remote_merged_blocks_fetched = 3;
getRemoteMergedBlocksFetchedList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributionsOrBuilder
repeated double remote_merged_blocks_fetched = 3;
getRemoteMergedBytesRead(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
repeated double remote_merged_bytes_read = 7;
getRemoteMergedBytesRead(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions
repeated double remote_merged_bytes_read = 7;
getRemoteMergedBytesRead(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributionsOrBuilder
repeated double remote_merged_bytes_read = 7;
getRemoteMergedBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics.Builder
int64 remote_merged_bytes_read = 7;
getRemoteMergedBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics
int64 remote_merged_bytes_read = 7;
getRemoteMergedBytesRead() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricsOrBuilder
int64 remote_merged_bytes_read = 7;
getRemoteMergedBytesReadCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
repeated double remote_merged_bytes_read = 7;
getRemoteMergedBytesReadCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions
repeated double remote_merged_bytes_read = 7;
getRemoteMergedBytesReadCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributionsOrBuilder
repeated double remote_merged_bytes_read = 7;
getRemoteMergedBytesReadList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
repeated double remote_merged_bytes_read = 7;
getRemoteMergedBytesReadList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions
repeated double remote_merged_bytes_read = 7;
getRemoteMergedBytesReadList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributionsOrBuilder
repeated double remote_merged_bytes_read = 7;
getRemoteMergedChunksFetched(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
repeated double remote_merged_chunks_fetched = 5;
getRemoteMergedChunksFetched(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions
repeated double remote_merged_chunks_fetched = 5;
getRemoteMergedChunksFetched(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributionsOrBuilder
repeated double remote_merged_chunks_fetched = 5;
getRemoteMergedChunksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics.Builder
int64 remote_merged_chunks_fetched = 5;
getRemoteMergedChunksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics
int64 remote_merged_chunks_fetched = 5;
getRemoteMergedChunksFetched() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricsOrBuilder
int64 remote_merged_chunks_fetched = 5;
getRemoteMergedChunksFetchedCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
repeated double remote_merged_chunks_fetched = 5;
getRemoteMergedChunksFetchedCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions
repeated double remote_merged_chunks_fetched = 5;
getRemoteMergedChunksFetchedCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributionsOrBuilder
repeated double remote_merged_chunks_fetched = 5;
getRemoteMergedChunksFetchedList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
repeated double remote_merged_chunks_fetched = 5;
getRemoteMergedChunksFetchedList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions
repeated double remote_merged_chunks_fetched = 5;
getRemoteMergedChunksFetchedList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributionsOrBuilder
repeated double remote_merged_chunks_fetched = 5;
getRemoteMergedReqsDuration(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
repeated double remote_merged_reqs_duration = 9;
getRemoteMergedReqsDuration(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions
repeated double remote_merged_reqs_duration = 9;
getRemoteMergedReqsDuration(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributionsOrBuilder
repeated double remote_merged_reqs_duration = 9;
getRemoteMergedReqsDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics.Builder
int64 remote_merged_reqs_duration = 9;
getRemoteMergedReqsDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics
int64 remote_merged_reqs_duration = 9;
getRemoteMergedReqsDuration() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricsOrBuilder
int64 remote_merged_reqs_duration = 9;
getRemoteMergedReqsDurationCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
repeated double remote_merged_reqs_duration = 9;
getRemoteMergedReqsDurationCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions
repeated double remote_merged_reqs_duration = 9;
getRemoteMergedReqsDurationCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributionsOrBuilder
repeated double remote_merged_reqs_duration = 9;
getRemoteMergedReqsDurationList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
repeated double remote_merged_reqs_duration = 9;
getRemoteMergedReqsDurationList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions
repeated double remote_merged_reqs_duration = 9;
getRemoteMergedReqsDurationList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributionsOrBuilder
repeated double remote_merged_reqs_duration = 9;
getRemoteReqsDuration(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
repeated double remote_reqs_duration = 9;
getRemoteReqsDuration(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions
repeated double remote_reqs_duration = 9;
getRemoteReqsDuration(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributionsOrBuilder
repeated double remote_reqs_duration = 9;
getRemoteReqsDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics.Builder
int64 remote_reqs_duration = 8;
getRemoteReqsDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics
int64 remote_reqs_duration = 8;
getRemoteReqsDuration() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricsOrBuilder
int64 remote_reqs_duration = 8;
getRemoteReqsDurationCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
repeated double remote_reqs_duration = 9;
getRemoteReqsDurationCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions
repeated double remote_reqs_duration = 9;
getRemoteReqsDurationCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributionsOrBuilder
repeated double remote_reqs_duration = 9;
getRemoteReqsDurationList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
repeated double remote_reqs_duration = 9;
getRemoteReqsDurationList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions
repeated double remote_reqs_duration = 9;
getRemoteReqsDurationList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributionsOrBuilder
repeated double remote_reqs_duration = 9;
getRemoteUser() - Method in class org.apache.spark.ui.XssSafeRequest
 
getRemoveReason() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
optional string remove_reason = 22;
getRemoveReason() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
optional string remove_reason = 22;
getRemoveReason() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
optional string remove_reason = 22;
getRemoveReasonBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
optional string remove_reason = 22;
getRemoveReasonBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
optional string remove_reason = 22;
getRemoveReasonBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
optional string remove_reason = 22;
getRemoveTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
optional int64 remove_time = 21;
getRemoveTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
optional int64 remove_time = 21;
getRemoveTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
optional int64 remove_time = 21;
getRemoveTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder
optional int64 remove_time = 6;
getRemoveTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary
optional int64 remove_time = 6;
getRemoveTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryOrBuilder
optional int64 remove_time = 6;
getRenameColumnQuery(String, String, String, int) - Static method in class org.apache.spark.sql.jdbc.DB2Dialect
 
getRenameColumnQuery(String, String, String, int) - Static method in class org.apache.spark.sql.jdbc.DerbyDialect
 
getRenameColumnQuery(String, String, String, int) - Static method in class org.apache.spark.sql.jdbc.H2Dialect
 
getRenameColumnQuery(String, String, String, int) - Method in class org.apache.spark.sql.jdbc.JdbcDialect
 
getRenameColumnQuery(String, String, String, int) - Static method in class org.apache.spark.sql.jdbc.MsSqlServerDialect
 
getRenameColumnQuery(String, String, String, int) - Static method in class org.apache.spark.sql.jdbc.MySQLDialect
 
getRenameColumnQuery(String, String, String, int) - Static method in class org.apache.spark.sql.jdbc.NoopDialect
 
getRenameColumnQuery(String, String, String, int) - Static method in class org.apache.spark.sql.jdbc.OracleDialect
 
getRenameColumnQuery(String, String, String, int) - Static method in class org.apache.spark.sql.jdbc.PostgresDialect
 
getRenameColumnQuery(String, String, String, int) - Static method in class org.apache.spark.sql.jdbc.TeradataDialect
 
GetReplicateInfoForRDDBlocks(BlockManagerId) - Constructor for class org.apache.spark.storage.BlockManagerMessages.GetReplicateInfoForRDDBlocks
 
GetReplicateInfoForRDDBlocks$() - Constructor for class org.apache.spark.storage.BlockManagerMessages.GetReplicateInfoForRDDBlocks$
 
getResource(String) - Method in class org.apache.spark.util.ChildFirstURLClassLoader
 
getResourceName() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest.Builder
optional string resource_name = 1;
getResourceName() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest
optional string resource_name = 1;
getResourceName() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequestOrBuilder
optional string resource_name = 1;
getResourceName() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskResourceRequest.Builder
optional string resource_name = 1;
getResourceName() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskResourceRequest
optional string resource_name = 1;
getResourceName() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskResourceRequestOrBuilder
optional string resource_name = 1;
getResourceNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest.Builder
optional string resource_name = 1;
getResourceNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest
optional string resource_name = 1;
getResourceNameBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequestOrBuilder
optional string resource_name = 1;
getResourceNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskResourceRequest.Builder
optional string resource_name = 1;
getResourceNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskResourceRequest
optional string resource_name = 1;
getResourceNameBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskResourceRequestOrBuilder
optional string resource_name = 1;
getResourceProfile() - Method in class org.apache.spark.api.java.JavaRDD
Get the ResourceProfile specified with this RDD or None if it wasn't specified.
getResourceProfile() - Method in class org.apache.spark.rdd.RDD
Get the ResourceProfile specified with this RDD or null if it wasn't specified.
getResourceProfileId() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
int32 resource_profile_id = 29;
getResourceProfileId() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
int32 resource_profile_id = 29;
getResourceProfileId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
int32 resource_profile_id = 29;
getResourceProfileId() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int32 resource_profile_id = 49;
getResourceProfileId() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
int32 resource_profile_id = 49;
getResourceProfileId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
int32 resource_profile_id = 49;
getResourceProfiles(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.ResourceProfileInfo resource_profiles = 7;
getResourceProfiles(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
repeated .org.apache.spark.status.protobuf.ResourceProfileInfo resource_profiles = 7;
getResourceProfiles(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoOrBuilder
repeated .org.apache.spark.status.protobuf.ResourceProfileInfo resource_profiles = 7;
getResourceProfilesBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.ResourceProfileInfo resource_profiles = 7;
getResourceProfilesBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.ResourceProfileInfo resource_profiles = 7;
getResourceProfilesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.ResourceProfileInfo resource_profiles = 7;
getResourceProfilesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
repeated .org.apache.spark.status.protobuf.ResourceProfileInfo resource_profiles = 7;
getResourceProfilesCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoOrBuilder
repeated .org.apache.spark.status.protobuf.ResourceProfileInfo resource_profiles = 7;
getResourceProfilesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.ResourceProfileInfo resource_profiles = 7;
getResourceProfilesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
repeated .org.apache.spark.status.protobuf.ResourceProfileInfo resource_profiles = 7;
getResourceProfilesList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoOrBuilder
repeated .org.apache.spark.status.protobuf.ResourceProfileInfo resource_profiles = 7;
getResourceProfilesOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.ResourceProfileInfo resource_profiles = 7;
getResourceProfilesOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
repeated .org.apache.spark.status.protobuf.ResourceProfileInfo resource_profiles = 7;
getResourceProfilesOrBuilder(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoOrBuilder
repeated .org.apache.spark.status.protobuf.ResourceProfileInfo resource_profiles = 7;
getResourceProfilesOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.ResourceProfileInfo resource_profiles = 7;
getResourceProfilesOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
repeated .org.apache.spark.status.protobuf.ResourceProfileInfo resource_profiles = 7;
getResourceProfilesOrBuilderList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoOrBuilder
repeated .org.apache.spark.status.protobuf.ResourceProfileInfo resource_profiles = 7;
getResources() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
Deprecated.
getResources() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
Deprecated.
getResources() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
Deprecated.
getResources(String) - Method in class org.apache.spark.util.ChildFirstURLClassLoader
 
getResourcesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
 
getResourcesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
 
getResourcesCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
map<string, .org.apache.spark.status.protobuf.ResourceInformation> resources = 28;
getResourcesMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
map<string, .org.apache.spark.status.protobuf.ResourceInformation> resources = 28;
getResourcesMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
map<string, .org.apache.spark.status.protobuf.ResourceInformation> resources = 28;
getResourcesMap() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
map<string, .org.apache.spark.status.protobuf.ResourceInformation> resources = 28;
getResourcesOrDefault(String, StoreTypes.ResourceInformation) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
map<string, .org.apache.spark.status.protobuf.ResourceInformation> resources = 28;
getResourcesOrDefault(String, StoreTypes.ResourceInformation) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
map<string, .org.apache.spark.status.protobuf.ResourceInformation> resources = 28;
getResourcesOrDefault(String, StoreTypes.ResourceInformation) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
map<string, .org.apache.spark.status.protobuf.ResourceInformation> resources = 28;
getResourcesOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
map<string, .org.apache.spark.status.protobuf.ResourceInformation> resources = 28;
getResourcesOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
map<string, .org.apache.spark.status.protobuf.ResourceInformation> resources = 28;
getResourcesOrThrow(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
map<string, .org.apache.spark.status.protobuf.ResourceInformation> resources = 28;
getResultFetchStart() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
optional int64 result_fetch_start = 6;
getResultFetchStart() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
optional int64 result_fetch_start = 6;
getResultFetchStart() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataOrBuilder
optional int64 result_fetch_start = 6;
getResultFetchStart() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 result_fetch_start = 6;
getResultFetchStart() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
int64 result_fetch_start = 6;
getResultFetchStart() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
int64 result_fetch_start = 6;
getResultSerializationTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double result_serialization_time = 12;
getResultSerializationTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
double result_serialization_time = 12;
getResultSerializationTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
double result_serialization_time = 12;
getResultSerializationTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int64 result_serialization_time = 20;
getResultSerializationTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
int64 result_serialization_time = 20;
getResultSerializationTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
int64 result_serialization_time = 20;
getResultSerializationTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 result_serialization_time = 22;
getResultSerializationTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
int64 result_serialization_time = 22;
getResultSerializationTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
int64 result_serialization_time = 22;
getResultSerializationTime(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double result_serialization_time = 9;
getResultSerializationTime(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
repeated double result_serialization_time = 9;
getResultSerializationTime(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
repeated double result_serialization_time = 9;
getResultSerializationTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
int64 result_serialization_time = 7;
getResultSerializationTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics
int64 result_serialization_time = 7;
getResultSerializationTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricsOrBuilder
int64 result_serialization_time = 7;
getResultSerializationTimeCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double result_serialization_time = 9;
getResultSerializationTimeCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
repeated double result_serialization_time = 9;
getResultSerializationTimeCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
repeated double result_serialization_time = 9;
getResultSerializationTimeList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double result_serialization_time = 9;
getResultSerializationTimeList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
repeated double result_serialization_time = 9;
getResultSerializationTimeList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
repeated double result_serialization_time = 9;
getResultSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double result_size = 10;
getResultSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
double result_size = 10;
getResultSize() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
double result_size = 10;
getResultSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int64 result_size = 18;
getResultSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
int64 result_size = 18;
getResultSize() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
int64 result_size = 18;
getResultSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 result_size = 20;
getResultSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
int64 result_size = 20;
getResultSize() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
int64 result_size = 20;
getResultSize(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double result_size = 7;
getResultSize(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
repeated double result_size = 7;
getResultSize(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
repeated double result_size = 7;
getResultSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
int64 result_size = 5;
getResultSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics
int64 result_size = 5;
getResultSize() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricsOrBuilder
int64 result_size = 5;
getResultSizeCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double result_size = 7;
getResultSizeCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
repeated double result_size = 7;
getResultSizeCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
repeated double result_size = 7;
getResultSizeList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double result_size = 7;
getResultSizeList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
repeated double result_size = 7;
getResultSizeList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
repeated double result_size = 7;
getRoaringBitMap(ShuffleBlockChunkId) - Method in class org.apache.spark.storage.PushBasedFetchHelper
Get the RoaringBitMap for a specific ShuffleBlockChunkId
getRollingIntervalSecs(SparkConf, boolean) - Static method in class org.apache.spark.streaming.util.WriteAheadLogUtils
 
getRootCluster() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
.org.apache.spark.status.protobuf.RDDOperationClusterWrapper root_cluster = 5;
getRootCluster() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper
.org.apache.spark.status.protobuf.RDDOperationClusterWrapper root_cluster = 5;
getRootCluster() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapperOrBuilder
.org.apache.spark.status.protobuf.RDDOperationClusterWrapper root_cluster = 5;
getRootClusterBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
.org.apache.spark.status.protobuf.RDDOperationClusterWrapper root_cluster = 5;
getRootClusterOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
.org.apache.spark.status.protobuf.RDDOperationClusterWrapper root_cluster = 5;
getRootClusterOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper
.org.apache.spark.status.protobuf.RDDOperationClusterWrapper root_cluster = 5;
getRootClusterOrBuilder() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapperOrBuilder
.org.apache.spark.status.protobuf.RDDOperationClusterWrapper root_cluster = 5;
getRootDirectory() - Static method in class org.apache.spark.SparkFiles
Get the root directory that contains files added through SparkContext.addFile().
getRootExecutionId() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
int64 root_execution_id = 2;
getRootExecutionId() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
int64 root_execution_id = 2;
getRootExecutionId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
int64 root_execution_id = 2;
getRow(int) - Method in class org.apache.spark.sql.vectorized.ColumnarBatch
Returns the row in this batch at `rowId`.
getRpInfo() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileWrapper.Builder
.org.apache.spark.status.protobuf.ResourceProfileInfo rpInfo = 1;
getRpInfo() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileWrapper
.org.apache.spark.status.protobuf.ResourceProfileInfo rpInfo = 1;
getRpInfo() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ResourceProfileWrapperOrBuilder
.org.apache.spark.status.protobuf.ResourceProfileInfo rpInfo = 1;
getRpInfoBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileWrapper.Builder
.org.apache.spark.status.protobuf.ResourceProfileInfo rpInfo = 1;
getRpInfoOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileWrapper.Builder
.org.apache.spark.status.protobuf.ResourceProfileInfo rpInfo = 1;
getRpInfoOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileWrapper
.org.apache.spark.status.protobuf.ResourceProfileInfo rpInfo = 1;
getRpInfoOrBuilder() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ResourceProfileWrapperOrBuilder
.org.apache.spark.status.protobuf.ResourceProfileInfo rpInfo = 1;
getRunId() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData.Builder
optional string run_id = 3;
getRunId() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData
optional string run_id = 3;
getRunId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryDataOrBuilder
optional string run_id = 3;
getRunId() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
optional string run_id = 2;
getRunId() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
optional string run_id = 2;
getRunId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
optional string run_id = 2;
getRunIdBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData.Builder
optional string run_id = 3;
getRunIdBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData
optional string run_id = 3;
getRunIdBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryDataOrBuilder
optional string run_id = 3;
getRunIdBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
optional string run_id = 2;
getRunIdBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
optional string run_id = 2;
getRunIdBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
optional string run_id = 2;
getRuntime() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
.org.apache.spark.status.protobuf.RuntimeInfo runtime = 1;
getRuntime() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
.org.apache.spark.status.protobuf.RuntimeInfo runtime = 1;
getRuntime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoOrBuilder
.org.apache.spark.status.protobuf.RuntimeInfo runtime = 1;
getRuntimeBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
.org.apache.spark.status.protobuf.RuntimeInfo runtime = 1;
getRuntimeOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
.org.apache.spark.status.protobuf.RuntimeInfo runtime = 1;
getRuntimeOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
.org.apache.spark.status.protobuf.RuntimeInfo runtime = 1;
getRuntimeOrBuilder() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoOrBuilder
.org.apache.spark.status.protobuf.RuntimeInfo runtime = 1;
getScalaVersion() - Method in class org.apache.spark.status.protobuf.StoreTypes.RuntimeInfo.Builder
optional string scala_version = 3;
getScalaVersion() - Method in class org.apache.spark.status.protobuf.StoreTypes.RuntimeInfo
optional string scala_version = 3;
getScalaVersion() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RuntimeInfoOrBuilder
optional string scala_version = 3;
getScalaVersionBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.RuntimeInfo.Builder
optional string scala_version = 3;
getScalaVersionBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.RuntimeInfo
optional string scala_version = 3;
getScalaVersionBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RuntimeInfoOrBuilder
optional string scala_version = 3;
getScalingVec() - Method in class org.apache.spark.ml.feature.ElementwiseProduct
 
getSchedulableByName(String) - Method in interface org.apache.spark.scheduler.Schedulable
 
getSchedulerDelay() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double scheduler_delay = 14;
getSchedulerDelay() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
double scheduler_delay = 14;
getSchedulerDelay() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
double scheduler_delay = 14;
getSchedulerDelay() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
int64 scheduler_delay = 17;
getSchedulerDelay() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
int64 scheduler_delay = 17;
getSchedulerDelay() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataOrBuilder
int64 scheduler_delay = 17;
getSchedulerDelay(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double scheduler_delay = 11;
getSchedulerDelay(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
repeated double scheduler_delay = 11;
getSchedulerDelay(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
repeated double scheduler_delay = 11;
getSchedulerDelayCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double scheduler_delay = 11;
getSchedulerDelayCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
repeated double scheduler_delay = 11;
getSchedulerDelayCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
repeated double scheduler_delay = 11;
getSchedulerDelayList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
repeated double scheduler_delay = 11;
getSchedulerDelayList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
repeated double scheduler_delay = 11;
getSchedulerDelayList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
repeated double scheduler_delay = 11;
getSchedulingMode() - Method in class org.apache.spark.SparkContext
Return current scheduling mode
getSchedulingPool() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
optional string scheduling_pool = 42;
getSchedulingPool() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
optional string scheduling_pool = 42;
getSchedulingPool() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
optional string scheduling_pool = 42;
getSchedulingPoolBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
optional string scheduling_pool = 42;
getSchedulingPoolBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
optional string scheduling_pool = 42;
getSchedulingPoolBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
optional string scheduling_pool = 42;
getSchemaCommentQuery(String, String) - Static method in class org.apache.spark.sql.jdbc.DB2Dialect
 
getSchemaCommentQuery(String, String) - Static method in class org.apache.spark.sql.jdbc.DerbyDialect
 
getSchemaCommentQuery(String, String) - Static method in class org.apache.spark.sql.jdbc.H2Dialect
 
getSchemaCommentQuery(String, String) - Method in class org.apache.spark.sql.jdbc.JdbcDialect
 
getSchemaCommentQuery(String, String) - Static method in class org.apache.spark.sql.jdbc.MsSqlServerDialect
 
getSchemaCommentQuery(String, String) - Static method in class org.apache.spark.sql.jdbc.MySQLDialect
 
getSchemaCommentQuery(String, String) - Static method in class org.apache.spark.sql.jdbc.NoopDialect
 
getSchemaCommentQuery(String, String) - Static method in class org.apache.spark.sql.jdbc.OracleDialect
 
getSchemaCommentQuery(String, String) - Static method in class org.apache.spark.sql.jdbc.PostgresDialect
 
getSchemaCommentQuery(String, String) - Static method in class org.apache.spark.sql.jdbc.TeradataDialect
 
getSchemaQuery(String) - Method in class org.apache.spark.sql.jdbc.AggregatedDialect
 
getSchemaQuery(String) - Static method in class org.apache.spark.sql.jdbc.DB2Dialect
 
getSchemaQuery(String) - Static method in class org.apache.spark.sql.jdbc.DerbyDialect
 
getSchemaQuery(String) - Static method in class org.apache.spark.sql.jdbc.H2Dialect
 
getSchemaQuery(String) - Method in class org.apache.spark.sql.jdbc.JdbcDialect
The SQL query that should be used to discover the schema of a table.
getSchemaQuery(String) - Static method in class org.apache.spark.sql.jdbc.MsSqlServerDialect
 
getSchemaQuery(String) - Static method in class org.apache.spark.sql.jdbc.MySQLDialect
 
getSchemaQuery(String) - Static method in class org.apache.spark.sql.jdbc.NoopDialect
 
getSchemaQuery(String) - Static method in class org.apache.spark.sql.jdbc.OracleDialect
 
getSchemaQuery(String) - Static method in class org.apache.spark.sql.jdbc.PostgresDialect
 
getSchemaQuery(String) - Static method in class org.apache.spark.sql.jdbc.TeradataDialect
 
getSeed() - Method in interface org.apache.spark.ml.param.shared.HasSeed
 
getSeed() - Method in class org.apache.spark.mllib.clustering.BisectingKMeans
Gets the random seed.
getSeed() - Method in class org.apache.spark.mllib.clustering.GaussianMixture
Return the random seed
getSeed() - Method in class org.apache.spark.mllib.clustering.KMeans
The random seed for cluster initialization.
getSeed() - Method in class org.apache.spark.mllib.clustering.LDA
Random seed for cluster initialization.
getSeed() - Method in class org.apache.spark.mllib.clustering.LocalLDAModel
Random seed for cluster initialization.
getSelectionMode() - Method in interface org.apache.spark.ml.feature.UnivariateFeatureSelectorParams
 
getSelectionThreshold() - Method in interface org.apache.spark.ml.feature.UnivariateFeatureSelectorParams
 
getSelectorType() - Method in interface org.apache.spark.ml.feature.SelectorParams
 
getSeq(int) - Method in interface org.apache.spark.sql.Row
Returns the value at position i of array type as a Scala Seq.
getSeqOp(boolean, Map<K, Object>, org.apache.spark.util.random.StratifiedSamplingUtils.RandomDataGenerator, Option<Map<K, Object>>) - Static method in class org.apache.spark.util.random.StratifiedSamplingUtils
Returns the function used by aggregate to collect sampling statistics for each partition.
getSequenceCol() - Method in class org.apache.spark.ml.fpm.PrefixSpan
 
getSerializationProxy(Object) - Static method in class org.apache.spark.util.IndylambdaScalaClosures
Check if the given reference is a indylambda style Scala closure.
getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo
 
getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo
 
getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
 
getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoWrapper
 
getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo
 
getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoWrapper
 
getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.AppSummary
 
getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
 
getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetrics
 
getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
 
getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions
 
getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest
 
getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary
 
getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper
 
getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
 
getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryWrapper
 
getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributions
 
getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetrics
 
getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
 
getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper
 
getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.MemoryMetrics
 
getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributions
 
getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetrics
 
getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.PairStrings
 
getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.PoolData
 
getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary
 
getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryWrapper
 
getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution
 
getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper
 
getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationEdge
 
getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper
 
getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode
 
getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo
 
getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo
 
getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoWrapper
 
getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceInformation
 
getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo
 
getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileWrapper
 
getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.RuntimeInfo
 
getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions
 
getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics
 
getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions
 
getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics
 
getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions
 
getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetrics
 
getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress
 
getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress
 
getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper
 
getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphEdge
 
getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode
 
getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper
 
getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper
 
getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummary
 
getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryWrapper
 
getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
 
getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetric
 
getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
 
getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper
 
getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress
 
getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData
 
getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData
 
getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
 
getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressWrapper
 
getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
 
getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
 
getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
 
getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics
 
getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskResourceRequest
 
getSessionConf(SparkSession) - Static method in class org.apache.spark.sql.api.r.SQLUtils
 
getShort(int) - Method in interface org.apache.spark.sql.Row
Returns the value at position i as a primitive short.
getShort(int) - Method in class org.apache.spark.sql.vectorized.ArrowColumnVector
 
getShort(int) - Method in class org.apache.spark.sql.vectorized.ColumnarArray
 
getShort(int) - Method in class org.apache.spark.sql.vectorized.ColumnarBatchRow
 
getShort(int) - Method in class org.apache.spark.sql.vectorized.ColumnarRow
 
getShort(int) - Method in class org.apache.spark.sql.vectorized.ColumnVector
Returns the short type value for rowId.
getShorts(int, int) - Method in class org.apache.spark.sql.vectorized.ColumnVector
Gets short type values from [rowId, rowId + count).
getShuffleBytesWritten() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 shuffle_bytes_written = 37;
getShuffleBytesWritten() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
int64 shuffle_bytes_written = 37;
getShuffleBytesWritten() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
int64 shuffle_bytes_written = 37;
getShuffleChunkCardinality(ShuffleBlockChunkId) - Method in class org.apache.spark.storage.PushBasedFetchHelper
Get the number of map blocks in a ShuffleBlockChunk
getShuffleCorruptMergedBlockChunks() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double shuffle_corrupt_merged_block_chunks = 33;
getShuffleCorruptMergedBlockChunks() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
double shuffle_corrupt_merged_block_chunks = 33;
getShuffleCorruptMergedBlockChunks() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
double shuffle_corrupt_merged_block_chunks = 33;
getShuffleCorruptMergedBlockChunks() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int64 shuffle_corrupt_merged_block_chunks = 53;
getShuffleCorruptMergedBlockChunks() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
int64 shuffle_corrupt_merged_block_chunks = 53;
getShuffleCorruptMergedBlockChunks() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
int64 shuffle_corrupt_merged_block_chunks = 53;
getShuffleCorruptMergedBlockChunks() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 shuffle_corrupt_merged_block_chunks = 42;
getShuffleCorruptMergedBlockChunks() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
int64 shuffle_corrupt_merged_block_chunks = 42;
getShuffleCorruptMergedBlockChunks() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
int64 shuffle_corrupt_merged_block_chunks = 42;
getShuffleFetchWaitTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double shuffle_fetch_wait_time = 26;
getShuffleFetchWaitTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
double shuffle_fetch_wait_time = 26;
getShuffleFetchWaitTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
double shuffle_fetch_wait_time = 26;
getShuffleFetchWaitTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int64 shuffle_fetch_wait_time = 30;
getShuffleFetchWaitTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
int64 shuffle_fetch_wait_time = 30;
getShuffleFetchWaitTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
int64 shuffle_fetch_wait_time = 30;
getShuffleFetchWaitTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 shuffle_fetch_wait_time = 32;
getShuffleFetchWaitTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
int64 shuffle_fetch_wait_time = 32;
getShuffleFetchWaitTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
int64 shuffle_fetch_wait_time = 32;
getShuffleLocalBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double shuffle_local_blocks_fetched = 25;
getShuffleLocalBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
double shuffle_local_blocks_fetched = 25;
getShuffleLocalBlocksFetched() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
double shuffle_local_blocks_fetched = 25;
getShuffleLocalBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int64 shuffle_local_blocks_fetched = 29;
getShuffleLocalBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
int64 shuffle_local_blocks_fetched = 29;
getShuffleLocalBlocksFetched() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
int64 shuffle_local_blocks_fetched = 29;
getShuffleLocalBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 shuffle_local_blocks_fetched = 31;
getShuffleLocalBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
int64 shuffle_local_blocks_fetched = 31;
getShuffleLocalBlocksFetched() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
int64 shuffle_local_blocks_fetched = 31;
getShuffleLocalBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int64 shuffle_local_bytes_read = 33;
getShuffleLocalBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
int64 shuffle_local_bytes_read = 33;
getShuffleLocalBytesRead() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
int64 shuffle_local_bytes_read = 33;
getShuffleLocalBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 shuffle_local_bytes_read = 35;
getShuffleLocalBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
int64 shuffle_local_bytes_read = 35;
getShuffleLocalBytesRead() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
int64 shuffle_local_bytes_read = 35;
getShuffleMergedFetchFallbackCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double shuffle_merged_fetch_fallback_count = 34;
getShuffleMergedFetchFallbackCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
double shuffle_merged_fetch_fallback_count = 34;
getShuffleMergedFetchFallbackCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
double shuffle_merged_fetch_fallback_count = 34;
getShuffleMergedFetchFallbackCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int64 shuffle_merged_fetch_fallback_count = 54;
getShuffleMergedFetchFallbackCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
int64 shuffle_merged_fetch_fallback_count = 54;
getShuffleMergedFetchFallbackCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
int64 shuffle_merged_fetch_fallback_count = 54;
getShuffleMergedFetchFallbackCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 shuffle_merged_fetch_fallback_count = 43;
getShuffleMergedFetchFallbackCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
int64 shuffle_merged_fetch_fallback_count = 43;
getShuffleMergedFetchFallbackCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
int64 shuffle_merged_fetch_fallback_count = 43;
getShuffleMergedLocalBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double shuffle_merged_local_blocks_fetched = 36;
getShuffleMergedLocalBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
double shuffle_merged_local_blocks_fetched = 36;
getShuffleMergedLocalBlocksFetched() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
double shuffle_merged_local_blocks_fetched = 36;
getShuffleMergedLocalBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int64 shuffle_merged_local_blocks_fetched = 56;
getShuffleMergedLocalBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
int64 shuffle_merged_local_blocks_fetched = 56;
getShuffleMergedLocalBlocksFetched() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
int64 shuffle_merged_local_blocks_fetched = 56;
getShuffleMergedLocalBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 shuffle_merged_local_blocks_fetched = 45;
getShuffleMergedLocalBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
int64 shuffle_merged_local_blocks_fetched = 45;
getShuffleMergedLocalBlocksFetched() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
int64 shuffle_merged_local_blocks_fetched = 45;
getShuffleMergedLocalBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double shuffle_merged_local_bytes_read = 40;
getShuffleMergedLocalBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
double shuffle_merged_local_bytes_read = 40;
getShuffleMergedLocalBytesRead() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
double shuffle_merged_local_bytes_read = 40;
getShuffleMergedLocalBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int64 shuffle_merged_local_bytes_read = 60;
getShuffleMergedLocalBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
int64 shuffle_merged_local_bytes_read = 60;
getShuffleMergedLocalBytesRead() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
int64 shuffle_merged_local_bytes_read = 60;
getShuffleMergedLocalBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 shuffle_merged_local_bytes_read = 49;
getShuffleMergedLocalBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
int64 shuffle_merged_local_bytes_read = 49;
getShuffleMergedLocalBytesRead() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
int64 shuffle_merged_local_bytes_read = 49;
getShuffleMergedLocalChunksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double shuffle_merged_local_chunks_fetched = 38;
getShuffleMergedLocalChunksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
double shuffle_merged_local_chunks_fetched = 38;
getShuffleMergedLocalChunksFetched() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
double shuffle_merged_local_chunks_fetched = 38;
getShuffleMergedLocalChunksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int64 shuffle_merged_local_chunks_fetched = 58;
getShuffleMergedLocalChunksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
int64 shuffle_merged_local_chunks_fetched = 58;
getShuffleMergedLocalChunksFetched() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
int64 shuffle_merged_local_chunks_fetched = 58;
getShuffleMergedLocalChunksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 shuffle_merged_local_chunks_fetched = 47;
getShuffleMergedLocalChunksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
int64 shuffle_merged_local_chunks_fetched = 47;
getShuffleMergedLocalChunksFetched() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
int64 shuffle_merged_local_chunks_fetched = 47;
getShuffleMergedRemoteBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double shuffle_merged_remote_blocks_fetched = 35;
getShuffleMergedRemoteBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
double shuffle_merged_remote_blocks_fetched = 35;
getShuffleMergedRemoteBlocksFetched() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
double shuffle_merged_remote_blocks_fetched = 35;
getShuffleMergedRemoteBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int64 shuffle_merged_remote_blocks_fetched = 55;
getShuffleMergedRemoteBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
int64 shuffle_merged_remote_blocks_fetched = 55;
getShuffleMergedRemoteBlocksFetched() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
int64 shuffle_merged_remote_blocks_fetched = 55;
getShuffleMergedRemoteBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 shuffle_merged_remote_blocks_fetched = 44;
getShuffleMergedRemoteBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
int64 shuffle_merged_remote_blocks_fetched = 44;
getShuffleMergedRemoteBlocksFetched() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
int64 shuffle_merged_remote_blocks_fetched = 44;
getShuffleMergedRemoteBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double shuffle_merged_remote_bytes_read = 39;
getShuffleMergedRemoteBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
double shuffle_merged_remote_bytes_read = 39;
getShuffleMergedRemoteBytesRead() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
double shuffle_merged_remote_bytes_read = 39;
getShuffleMergedRemoteBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int64 shuffle_merged_remote_bytes_read = 59;
getShuffleMergedRemoteBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
int64 shuffle_merged_remote_bytes_read = 59;
getShuffleMergedRemoteBytesRead() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
int64 shuffle_merged_remote_bytes_read = 59;
getShuffleMergedRemoteBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 shuffle_merged_remote_bytes_read = 48;
getShuffleMergedRemoteBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
int64 shuffle_merged_remote_bytes_read = 48;
getShuffleMergedRemoteBytesRead() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
int64 shuffle_merged_remote_bytes_read = 48;
getShuffleMergedRemoteChunksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double shuffle_merged_remote_chunks_fetched = 37;
getShuffleMergedRemoteChunksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
double shuffle_merged_remote_chunks_fetched = 37;
getShuffleMergedRemoteChunksFetched() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
double shuffle_merged_remote_chunks_fetched = 37;
getShuffleMergedRemoteChunksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int64 shuffle_merged_remote_chunks_fetched = 57;
getShuffleMergedRemoteChunksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
int64 shuffle_merged_remote_chunks_fetched = 57;
getShuffleMergedRemoteChunksFetched() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
int64 shuffle_merged_remote_chunks_fetched = 57;
getShuffleMergedRemoteChunksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 shuffle_merged_remote_chunks_fetched = 46;
getShuffleMergedRemoteChunksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
int64 shuffle_merged_remote_chunks_fetched = 46;
getShuffleMergedRemoteChunksFetched() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
int64 shuffle_merged_remote_chunks_fetched = 46;
getShuffleMergedRemoteReqDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 shuffle_merged_remote_req_duration = 51;
getShuffleMergedRemoteReqDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
int64 shuffle_merged_remote_req_duration = 51;
getShuffleMergedRemoteReqDuration() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
int64 shuffle_merged_remote_req_duration = 51;
getShuffleMergedRemoteReqsDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double shuffle_merged_remote_reqs_duration = 42;
getShuffleMergedRemoteReqsDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
double shuffle_merged_remote_reqs_duration = 42;
getShuffleMergedRemoteReqsDuration() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
double shuffle_merged_remote_reqs_duration = 42;
getShuffleMergedRemoteReqsDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int64 shuffle_merged_remote_reqs_duration = 62;
getShuffleMergedRemoteReqsDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
int64 shuffle_merged_remote_reqs_duration = 62;
getShuffleMergedRemoteReqsDuration() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
int64 shuffle_merged_remote_reqs_duration = 62;
getShuffleMergersCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int32 shuffle_mergers_count = 64;
getShuffleMergersCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
int32 shuffle_mergers_count = 64;
getShuffleMergersCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
int32 shuffle_mergers_count = 64;
getShufflePushMergerLocations(int, int) - Method in interface org.apache.spark.scheduler.SchedulerBackend
Get the list of host locations for push based shuffle
getShufflePushMergerLocations() - Method in class org.apache.spark.ShuffleStatus
 
GetShufflePushMergerLocations(int, Set<String>) - Constructor for class org.apache.spark.storage.BlockManagerMessages.GetShufflePushMergerLocations
 
GetShufflePushMergerLocations$() - Constructor for class org.apache.spark.storage.BlockManagerMessages.GetShufflePushMergerLocations$
 
getShufflePushReadMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics.Builder
.org.apache.spark.status.protobuf.ShufflePushReadMetrics shuffle_push_read_metrics = 9;
getShufflePushReadMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics
.org.apache.spark.status.protobuf.ShufflePushReadMetrics shuffle_push_read_metrics = 9;
getShufflePushReadMetrics() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricsOrBuilder
.org.apache.spark.status.protobuf.ShufflePushReadMetrics shuffle_push_read_metrics = 9;
getShufflePushReadMetricsBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics.Builder
.org.apache.spark.status.protobuf.ShufflePushReadMetrics shuffle_push_read_metrics = 9;
getShufflePushReadMetricsDist() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
.org.apache.spark.status.protobuf.ShufflePushReadMetricDistributions shuffle_push_read_metrics_dist = 10;
getShufflePushReadMetricsDist() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions
.org.apache.spark.status.protobuf.ShufflePushReadMetricDistributions shuffle_push_read_metrics_dist = 10;
getShufflePushReadMetricsDist() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributionsOrBuilder
.org.apache.spark.status.protobuf.ShufflePushReadMetricDistributions shuffle_push_read_metrics_dist = 10;
getShufflePushReadMetricsDistBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
.org.apache.spark.status.protobuf.ShufflePushReadMetricDistributions shuffle_push_read_metrics_dist = 10;
getShufflePushReadMetricsDistOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
.org.apache.spark.status.protobuf.ShufflePushReadMetricDistributions shuffle_push_read_metrics_dist = 10;
getShufflePushReadMetricsDistOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions
.org.apache.spark.status.protobuf.ShufflePushReadMetricDistributions shuffle_push_read_metrics_dist = 10;
getShufflePushReadMetricsDistOrBuilder() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributionsOrBuilder
.org.apache.spark.status.protobuf.ShufflePushReadMetricDistributions shuffle_push_read_metrics_dist = 10;
getShufflePushReadMetricsOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics.Builder
.org.apache.spark.status.protobuf.ShufflePushReadMetrics shuffle_push_read_metrics = 9;
getShufflePushReadMetricsOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics
.org.apache.spark.status.protobuf.ShufflePushReadMetrics shuffle_push_read_metrics = 9;
getShufflePushReadMetricsOrBuilder() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricsOrBuilder
.org.apache.spark.status.protobuf.ShufflePushReadMetrics shuffle_push_read_metrics = 9;
getShuffleRead(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double shuffle_read = 10;
getShuffleRead(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
repeated double shuffle_read = 10;
getShuffleRead(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
repeated double shuffle_read = 10;
getShuffleRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
int64 shuffle_read = 9;
getShuffleRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary
int64 shuffle_read = 9;
getShuffleRead() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryOrBuilder
int64 shuffle_read = 9;
getShuffleReadBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double shuffle_read_bytes = 22;
getShuffleReadBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
double shuffle_read_bytes = 22;
getShuffleReadBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
double shuffle_read_bytes = 22;
getShuffleReadBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int64 shuffle_read_bytes = 34;
getShuffleReadBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
int64 shuffle_read_bytes = 34;
getShuffleReadBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
int64 shuffle_read_bytes = 34;
getShuffleReadCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double shuffle_read = 10;
getShuffleReadCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
repeated double shuffle_read = 10;
getShuffleReadCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
repeated double shuffle_read = 10;
getShuffleReadList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double shuffle_read = 10;
getShuffleReadList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
repeated double shuffle_read = 10;
getShuffleReadList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
repeated double shuffle_read = 10;
getShuffleReadMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
.org.apache.spark.status.protobuf.ShuffleReadMetricDistributions shuffle_read_metrics = 17;
getShuffleReadMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
.org.apache.spark.status.protobuf.ShuffleReadMetricDistributions shuffle_read_metrics = 17;
getShuffleReadMetrics() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
.org.apache.spark.status.protobuf.ShuffleReadMetricDistributions shuffle_read_metrics = 17;
getShuffleReadMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
.org.apache.spark.status.protobuf.ShuffleReadMetrics shuffle_read_metrics = 13;
getShuffleReadMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics
.org.apache.spark.status.protobuf.ShuffleReadMetrics shuffle_read_metrics = 13;
getShuffleReadMetrics() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricsOrBuilder
.org.apache.spark.status.protobuf.ShuffleReadMetrics shuffle_read_metrics = 13;
getShuffleReadMetricsBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
.org.apache.spark.status.protobuf.ShuffleReadMetricDistributions shuffle_read_metrics = 17;
getShuffleReadMetricsBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
.org.apache.spark.status.protobuf.ShuffleReadMetrics shuffle_read_metrics = 13;
getShuffleReadMetricsOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
.org.apache.spark.status.protobuf.ShuffleReadMetricDistributions shuffle_read_metrics = 17;
getShuffleReadMetricsOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
.org.apache.spark.status.protobuf.ShuffleReadMetricDistributions shuffle_read_metrics = 17;
getShuffleReadMetricsOrBuilder() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
.org.apache.spark.status.protobuf.ShuffleReadMetricDistributions shuffle_read_metrics = 17;
getShuffleReadMetricsOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
.org.apache.spark.status.protobuf.ShuffleReadMetrics shuffle_read_metrics = 13;
getShuffleReadMetricsOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics
.org.apache.spark.status.protobuf.ShuffleReadMetrics shuffle_read_metrics = 13;
getShuffleReadMetricsOrBuilder() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricsOrBuilder
.org.apache.spark.status.protobuf.ShuffleReadMetrics shuffle_read_metrics = 13;
getShuffleReadRecords(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double shuffle_read_records = 11;
getShuffleReadRecords(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
repeated double shuffle_read_records = 11;
getShuffleReadRecords(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
repeated double shuffle_read_records = 11;
getShuffleReadRecords() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
int64 shuffle_read_records = 10;
getShuffleReadRecords() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary
int64 shuffle_read_records = 10;
getShuffleReadRecords() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryOrBuilder
int64 shuffle_read_records = 10;
getShuffleReadRecords() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int64 shuffle_read_records = 35;
getShuffleReadRecords() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
int64 shuffle_read_records = 35;
getShuffleReadRecords() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
int64 shuffle_read_records = 35;
getShuffleReadRecordsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double shuffle_read_records = 11;
getShuffleReadRecordsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
repeated double shuffle_read_records = 11;
getShuffleReadRecordsCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
repeated double shuffle_read_records = 11;
getShuffleReadRecordsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double shuffle_read_records = 11;
getShuffleReadRecordsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
repeated double shuffle_read_records = 11;
getShuffleReadRecordsList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
repeated double shuffle_read_records = 11;
getShuffleRecordsRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double shuffle_records_read = 23;
getShuffleRecordsRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
double shuffle_records_read = 23;
getShuffleRecordsRead() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
double shuffle_records_read = 23;
getShuffleRecordsRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 shuffle_records_read = 36;
getShuffleRecordsRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
int64 shuffle_records_read = 36;
getShuffleRecordsRead() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
int64 shuffle_records_read = 36;
getShuffleRecordsWritten() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 shuffle_records_written = 39;
getShuffleRecordsWritten() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
int64 shuffle_records_written = 39;
getShuffleRecordsWritten() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
int64 shuffle_records_written = 39;
getShuffleRemoteBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double shuffle_remote_blocks_fetched = 24;
getShuffleRemoteBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
double shuffle_remote_blocks_fetched = 24;
getShuffleRemoteBlocksFetched() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
double shuffle_remote_blocks_fetched = 24;
getShuffleRemoteBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int64 shuffle_remote_blocks_fetched = 28;
getShuffleRemoteBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
int64 shuffle_remote_blocks_fetched = 28;
getShuffleRemoteBlocksFetched() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
int64 shuffle_remote_blocks_fetched = 28;
getShuffleRemoteBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 shuffle_remote_blocks_fetched = 30;
getShuffleRemoteBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
int64 shuffle_remote_blocks_fetched = 30;
getShuffleRemoteBlocksFetched() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
int64 shuffle_remote_blocks_fetched = 30;
getShuffleRemoteBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double shuffle_remote_bytes_read = 27;
getShuffleRemoteBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
double shuffle_remote_bytes_read = 27;
getShuffleRemoteBytesRead() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
double shuffle_remote_bytes_read = 27;
getShuffleRemoteBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int64 shuffle_remote_bytes_read = 31;
getShuffleRemoteBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
int64 shuffle_remote_bytes_read = 31;
getShuffleRemoteBytesRead() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
int64 shuffle_remote_bytes_read = 31;
getShuffleRemoteBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 shuffle_remote_bytes_read = 33;
getShuffleRemoteBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
int64 shuffle_remote_bytes_read = 33;
getShuffleRemoteBytesRead() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
int64 shuffle_remote_bytes_read = 33;
getShuffleRemoteBytesReadToDisk() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double shuffle_remote_bytes_read_to_disk = 28;
getShuffleRemoteBytesReadToDisk() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
double shuffle_remote_bytes_read_to_disk = 28;
getShuffleRemoteBytesReadToDisk() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
double shuffle_remote_bytes_read_to_disk = 28;
getShuffleRemoteBytesReadToDisk() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int64 shuffle_remote_bytes_read_to_disk = 32;
getShuffleRemoteBytesReadToDisk() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
int64 shuffle_remote_bytes_read_to_disk = 32;
getShuffleRemoteBytesReadToDisk() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
int64 shuffle_remote_bytes_read_to_disk = 32;
getShuffleRemoteBytesReadToDisk() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 shuffle_remote_bytes_read_to_disk = 34;
getShuffleRemoteBytesReadToDisk() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
int64 shuffle_remote_bytes_read_to_disk = 34;
getShuffleRemoteBytesReadToDisk() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
int64 shuffle_remote_bytes_read_to_disk = 34;
getShuffleRemoteReqsDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double shuffle_remote_reqs_duration = 41;
getShuffleRemoteReqsDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
double shuffle_remote_reqs_duration = 41;
getShuffleRemoteReqsDuration() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
double shuffle_remote_reqs_duration = 41;
getShuffleRemoteReqsDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int64 shuffle_remote_reqs_duration = 61;
getShuffleRemoteReqsDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
int64 shuffle_remote_reqs_duration = 61;
getShuffleRemoteReqsDuration() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
int64 shuffle_remote_reqs_duration = 61;
getShuffleRemoteReqsDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 shuffle_remote_reqs_duration = 50;
getShuffleRemoteReqsDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
int64 shuffle_remote_reqs_duration = 50;
getShuffleRemoteReqsDuration() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
int64 shuffle_remote_reqs_duration = 50;
getShuffleTotalBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double shuffle_total_blocks_fetched = 29;
getShuffleTotalBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
double shuffle_total_blocks_fetched = 29;
getShuffleTotalBlocksFetched() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
double shuffle_total_blocks_fetched = 29;
getShuffleWrite(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double shuffle_write = 12;
getShuffleWrite(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
repeated double shuffle_write = 12;
getShuffleWrite(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
repeated double shuffle_write = 12;
getShuffleWrite() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
int64 shuffle_write = 11;
getShuffleWrite() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary
int64 shuffle_write = 11;
getShuffleWrite() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryOrBuilder
int64 shuffle_write = 11;
getShuffleWriteBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double shuffle_write_bytes = 30;
getShuffleWriteBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
double shuffle_write_bytes = 30;
getShuffleWriteBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
double shuffle_write_bytes = 30;
getShuffleWriteBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int64 shuffle_write_bytes = 36;
getShuffleWriteBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
int64 shuffle_write_bytes = 36;
getShuffleWriteBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
int64 shuffle_write_bytes = 36;
getShuffleWriteCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double shuffle_write = 12;
getShuffleWriteCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
repeated double shuffle_write = 12;
getShuffleWriteCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
repeated double shuffle_write = 12;
getShuffleWriteList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double shuffle_write = 12;
getShuffleWriteList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
repeated double shuffle_write = 12;
getShuffleWriteList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
repeated double shuffle_write = 12;
getShuffleWriteMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
.org.apache.spark.status.protobuf.ShuffleWriteMetricDistributions shuffle_write_metrics = 18;
getShuffleWriteMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
.org.apache.spark.status.protobuf.ShuffleWriteMetricDistributions shuffle_write_metrics = 18;
getShuffleWriteMetrics() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
.org.apache.spark.status.protobuf.ShuffleWriteMetricDistributions shuffle_write_metrics = 18;
getShuffleWriteMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
.org.apache.spark.status.protobuf.ShuffleWriteMetrics shuffle_write_metrics = 14;
getShuffleWriteMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics
.org.apache.spark.status.protobuf.ShuffleWriteMetrics shuffle_write_metrics = 14;
getShuffleWriteMetrics() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricsOrBuilder
.org.apache.spark.status.protobuf.ShuffleWriteMetrics shuffle_write_metrics = 14;
getShuffleWriteMetricsBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
.org.apache.spark.status.protobuf.ShuffleWriteMetricDistributions shuffle_write_metrics = 18;
getShuffleWriteMetricsBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
.org.apache.spark.status.protobuf.ShuffleWriteMetrics shuffle_write_metrics = 14;
getShuffleWriteMetricsOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
.org.apache.spark.status.protobuf.ShuffleWriteMetricDistributions shuffle_write_metrics = 18;
getShuffleWriteMetricsOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
.org.apache.spark.status.protobuf.ShuffleWriteMetricDistributions shuffle_write_metrics = 18;
getShuffleWriteMetricsOrBuilder() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
.org.apache.spark.status.protobuf.ShuffleWriteMetricDistributions shuffle_write_metrics = 18;
getShuffleWriteMetricsOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
.org.apache.spark.status.protobuf.ShuffleWriteMetrics shuffle_write_metrics = 14;
getShuffleWriteMetricsOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics
.org.apache.spark.status.protobuf.ShuffleWriteMetrics shuffle_write_metrics = 14;
getShuffleWriteMetricsOrBuilder() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricsOrBuilder
.org.apache.spark.status.protobuf.ShuffleWriteMetrics shuffle_write_metrics = 14;
getShuffleWriteRecords() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double shuffle_write_records = 31;
getShuffleWriteRecords() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
double shuffle_write_records = 31;
getShuffleWriteRecords() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
double shuffle_write_records = 31;
getShuffleWriteRecords(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double shuffle_write_records = 13;
getShuffleWriteRecords(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
repeated double shuffle_write_records = 13;
getShuffleWriteRecords(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
repeated double shuffle_write_records = 13;
getShuffleWriteRecords() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
int64 shuffle_write_records = 12;
getShuffleWriteRecords() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary
int64 shuffle_write_records = 12;
getShuffleWriteRecords() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryOrBuilder
int64 shuffle_write_records = 12;
getShuffleWriteRecords() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int64 shuffle_write_records = 38;
getShuffleWriteRecords() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
int64 shuffle_write_records = 38;
getShuffleWriteRecords() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
int64 shuffle_write_records = 38;
getShuffleWriteRecordsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double shuffle_write_records = 13;
getShuffleWriteRecordsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
repeated double shuffle_write_records = 13;
getShuffleWriteRecordsCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
repeated double shuffle_write_records = 13;
getShuffleWriteRecordsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double shuffle_write_records = 13;
getShuffleWriteRecordsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
repeated double shuffle_write_records = 13;
getShuffleWriteRecordsList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
repeated double shuffle_write_records = 13;
getShuffleWriteTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
double shuffle_write_time = 32;
getShuffleWriteTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
double shuffle_write_time = 32;
getShuffleWriteTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
double shuffle_write_time = 32;
getShuffleWriteTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int64 shuffle_write_time = 37;
getShuffleWriteTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
int64 shuffle_write_time = 37;
getShuffleWriteTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
int64 shuffle_write_time = 37;
getShuffleWriteTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 shuffle_write_time = 38;
getShuffleWriteTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
int64 shuffle_write_time = 38;
getShuffleWriteTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
int64 shuffle_write_time = 38;
getSimpleMessage() - Method in exception org.apache.spark.sql.AnalysisException
 
getSimpleName(Class<?>) - Static method in class org.apache.spark.util.Utils
Safer than Class obj's getSimpleName which may throw Malformed class name error in scala.
getSink() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
.org.apache.spark.status.protobuf.SinkProgress sink = 11;
getSink() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
.org.apache.spark.status.protobuf.SinkProgress sink = 11;
getSink() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
.org.apache.spark.status.protobuf.SinkProgress sink = 11;
getSinkBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
.org.apache.spark.status.protobuf.SinkProgress sink = 11;
getSinkOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
.org.apache.spark.status.protobuf.SinkProgress sink = 11;
getSinkOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
.org.apache.spark.status.protobuf.SinkProgress sink = 11;
getSinkOrBuilder() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
.org.apache.spark.status.protobuf.SinkProgress sink = 11;
getSize() - Method in class org.apache.spark.ml.feature.VectorSizeHint
group getParam
getSizeAsBytes(String) - Method in class org.apache.spark.SparkConf
Get a size parameter as bytes; throws a NoSuchElementException if it's not set.
getSizeAsBytes(String, String) - Method in class org.apache.spark.SparkConf
Get a size parameter as bytes, falling back to a default if not set.
getSizeAsBytes(String, long) - Method in class org.apache.spark.SparkConf
Get a size parameter as bytes, falling back to a default if not set.
getSizeAsGb(String) - Method in class org.apache.spark.SparkConf
Get a size parameter as Gibibytes; throws a NoSuchElementException if it's not set.
getSizeAsGb(String, String) - Method in class org.apache.spark.SparkConf
Get a size parameter as Gibibytes, falling back to a default if not set.
getSizeAsKb(String) - Method in class org.apache.spark.SparkConf
Get a size parameter as Kibibytes; throws a NoSuchElementException if it's not set.
getSizeAsKb(String, String) - Method in class org.apache.spark.SparkConf
Get a size parameter as Kibibytes, falling back to a default if not set.
getSizeAsMb(String) - Method in class org.apache.spark.SparkConf
Get a size parameter as Mebibytes; throws a NoSuchElementException if it's not set.
getSizeAsMb(String, String) - Method in class org.apache.spark.SparkConf
Get a size parameter as Mebibytes, falling back to a default if not set.
getSizeForBlock(int) - Method in interface org.apache.spark.scheduler.MapStatus
Estimated size for the reduce block, in bytes.
getSizeInBytes() - Method in interface org.apache.spark.ml.linalg.Matrix
Gets the current size in bytes of this `Matrix`.
getSkippedStages(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper.Builder
repeated int32 skipped_stages = 2;
getSkippedStages(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper
repeated int32 skipped_stages = 2;
getSkippedStages(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataWrapperOrBuilder
repeated int32 skipped_stages = 2;
getSkippedStagesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper.Builder
repeated int32 skipped_stages = 2;
getSkippedStagesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper
repeated int32 skipped_stages = 2;
getSkippedStagesCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataWrapperOrBuilder
repeated int32 skipped_stages = 2;
getSkippedStagesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper.Builder
repeated int32 skipped_stages = 2;
getSkippedStagesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper
repeated int32 skipped_stages = 2;
getSkippedStagesList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataWrapperOrBuilder
repeated int32 skipped_stages = 2;
getSlotDescs() - Method in class org.apache.spark.serializer.SerializationDebugger.ObjectStreamClassMethods
 
getSmoothing() - Method in interface org.apache.spark.ml.classification.NaiveBayesParams
 
getSolver() - Method in interface org.apache.spark.ml.param.shared.HasSolver
 
getSortedTaskSetQueue() - Method in interface org.apache.spark.scheduler.Schedulable
 
getSources(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
repeated .org.apache.spark.status.protobuf.SourceProgress sources = 10;
getSources(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
repeated .org.apache.spark.status.protobuf.SourceProgress sources = 10;
getSources(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
repeated .org.apache.spark.status.protobuf.SourceProgress sources = 10;
getSourcesBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
repeated .org.apache.spark.status.protobuf.SourceProgress sources = 10;
getSourcesBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
repeated .org.apache.spark.status.protobuf.SourceProgress sources = 10;
getSourcesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
repeated .org.apache.spark.status.protobuf.SourceProgress sources = 10;
getSourcesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
repeated .org.apache.spark.status.protobuf.SourceProgress sources = 10;
getSourcesCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
repeated .org.apache.spark.status.protobuf.SourceProgress sources = 10;
getSourcesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
repeated .org.apache.spark.status.protobuf.SourceProgress sources = 10;
getSourcesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
repeated .org.apache.spark.status.protobuf.SourceProgress sources = 10;
getSourcesList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
repeated .org.apache.spark.status.protobuf.SourceProgress sources = 10;
getSourcesOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
repeated .org.apache.spark.status.protobuf.SourceProgress sources = 10;
getSourcesOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
repeated .org.apache.spark.status.protobuf.SourceProgress sources = 10;
getSourcesOrBuilder(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
repeated .org.apache.spark.status.protobuf.SourceProgress sources = 10;
getSourcesOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
repeated .org.apache.spark.status.protobuf.SourceProgress sources = 10;
getSourcesOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
repeated .org.apache.spark.status.protobuf.SourceProgress sources = 10;
getSourcesOrBuilderList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
repeated .org.apache.spark.status.protobuf.SourceProgress sources = 10;
getSparkClassLoader() - Static method in class org.apache.spark.util.Utils
Get the ClassLoader which loaded Spark.
getSparkHome() - Method in class org.apache.spark.api.java.JavaSparkContext
Get Spark's home location from either a value set through the constructor, or the spark.home Java property, or the SPARK_HOME environment variable (in that order of preference).
getSparkOrYarnConfig(SparkConf, String, String) - Static method in class org.apache.spark.util.Utils
Return the value of a config either through the SparkConf or the Hadoop configuration.
getSparkProperties(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings spark_properties = 2;
getSparkProperties(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
repeated .org.apache.spark.status.protobuf.PairStrings spark_properties = 2;
getSparkProperties(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoOrBuilder
repeated .org.apache.spark.status.protobuf.PairStrings spark_properties = 2;
getSparkPropertiesBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings spark_properties = 2;
getSparkPropertiesBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings spark_properties = 2;
getSparkPropertiesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings spark_properties = 2;
getSparkPropertiesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
repeated .org.apache.spark.status.protobuf.PairStrings spark_properties = 2;
getSparkPropertiesCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoOrBuilder
repeated .org.apache.spark.status.protobuf.PairStrings spark_properties = 2;
getSparkPropertiesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings spark_properties = 2;
getSparkPropertiesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
repeated .org.apache.spark.status.protobuf.PairStrings spark_properties = 2;
getSparkPropertiesList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoOrBuilder
repeated .org.apache.spark.status.protobuf.PairStrings spark_properties = 2;
getSparkPropertiesOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings spark_properties = 2;
getSparkPropertiesOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
repeated .org.apache.spark.status.protobuf.PairStrings spark_properties = 2;
getSparkPropertiesOrBuilder(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoOrBuilder
repeated .org.apache.spark.status.protobuf.PairStrings spark_properties = 2;
getSparkPropertiesOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings spark_properties = 2;
getSparkPropertiesOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
repeated .org.apache.spark.status.protobuf.PairStrings spark_properties = 2;
getSparkPropertiesOrBuilderList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoOrBuilder
repeated .org.apache.spark.status.protobuf.PairStrings spark_properties = 2;
getSparkUser() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo.Builder
optional string spark_user = 6;
getSparkUser() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo
optional string spark_user = 6;
getSparkUser() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfoOrBuilder
optional string spark_user = 6;
getSparkUserBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo.Builder
optional string spark_user = 6;
getSparkUserBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo
optional string spark_user = 6;
getSparkUserBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfoOrBuilder
optional string spark_user = 6;
getSparseSizeInBytes(boolean) - Method in interface org.apache.spark.ml.linalg.Matrix
Gets the size of the minimal sparse representation of this `Matrix`.
getSpeculationSummary() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
optional .org.apache.spark.status.protobuf.SpeculationStageSummary speculation_summary = 47;
getSpeculationSummary() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
optional .org.apache.spark.status.protobuf.SpeculationStageSummary speculation_summary = 47;
getSpeculationSummary() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
optional .org.apache.spark.status.protobuf.SpeculationStageSummary speculation_summary = 47;
getSpeculationSummaryBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
optional .org.apache.spark.status.protobuf.SpeculationStageSummary speculation_summary = 47;
getSpeculationSummaryOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
optional .org.apache.spark.status.protobuf.SpeculationStageSummary speculation_summary = 47;
getSpeculationSummaryOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
optional .org.apache.spark.status.protobuf.SpeculationStageSummary speculation_summary = 47;
getSpeculationSummaryOrBuilder() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
optional .org.apache.spark.status.protobuf.SpeculationStageSummary speculation_summary = 47;
getSpeculative() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
bool speculative = 12;
getSpeculative() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
bool speculative = 12;
getSpeculative() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataOrBuilder
bool speculative = 12;
getSpeculative() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
bool speculative = 12;
getSpeculative() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
bool speculative = 12;
getSpeculative() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
bool speculative = 12;
getSplit() - Method in class org.apache.spark.ml.tree.DecisionTreeModelReadWrite.SplitData
 
getSplits() - Method in class org.apache.spark.ml.feature.Bucketizer
 
getSplitsArray() - Method in class org.apache.spark.ml.feature.Bucketizer
 
getSql() - Method in class org.apache.spark.sql.connector.catalog.ColumnDefaultValue
Returns the SQL string (Spark SQL dialect) of the default value expression.
getSqlExecutionId() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper.Builder
optional int64 sql_execution_id = 3;
getSqlExecutionId() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper
optional int64 sql_execution_id = 3;
getSqlExecutionId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataWrapperOrBuilder
optional int64 sql_execution_id = 3;
getSqlState(String) - Method in class org.apache.spark.ErrorClassesJsonReader
 
getSqlState() - Method in interface org.apache.spark.SparkThrowable
 
getSqlState(String) - Static method in class org.apache.spark.SparkThrowableHelper
 
getSrcCol() - Method in interface org.apache.spark.ml.clustering.PowerIterationClusteringParams
 
getStageAttemptId() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
int32 stage_attempt_id = 2;
getStageAttemptId() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
int32 stage_attempt_id = 2;
getStageAttemptId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
int32 stage_attempt_id = 2;
getStageAttemptId() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper.Builder
int32 stage_attempt_id = 2;
getStageAttemptId() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper
int32 stage_attempt_id = 2;
getStageAttemptId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapperOrBuilder
int32 stage_attempt_id = 2;
getStageAttemptId() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryWrapper.Builder
int32 stage_attempt_id = 2;
getStageAttemptId() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryWrapper
int32 stage_attempt_id = 2;
getStageAttemptId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryWrapperOrBuilder
int32 stage_attempt_id = 2;
getStageAttemptId() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int32 stage_attempt_id = 41;
getStageAttemptId() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
int32 stage_attempt_id = 41;
getStageAttemptId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
int32 stage_attempt_id = 41;
getStageId() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
int64 stage_id = 1;
getStageId() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
int64 stage_id = 1;
getStageId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
int64 stage_id = 1;
getStageId() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper.Builder
int64 stage_id = 1;
getStageId() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper
int64 stage_id = 1;
getStageId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapperOrBuilder
int64 stage_id = 1;
getStageId() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
int64 stage_id = 1;
getStageId() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper
int64 stage_id = 1;
getStageId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapperOrBuilder
int64 stage_id = 1;
getStageId() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryWrapper.Builder
int64 stage_id = 1;
getStageId() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryWrapper
int64 stage_id = 1;
getStageId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryWrapperOrBuilder
int64 stage_id = 1;
getStageId() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
int64 stage_id = 2;
getStageId() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
int64 stage_id = 2;
getStageId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
int64 stage_id = 2;
getStageId() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 stage_id = 40;
getStageId() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
int64 stage_id = 40;
getStageId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
int64 stage_id = 40;
getStageIds(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
repeated int64 stage_ids = 6;
getStageIds(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
repeated int64 stage_ids = 6;
getStageIds(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataOrBuilder
repeated int64 stage_ids = 6;
getStageIds(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.PoolData.Builder
repeated int64 stage_ids = 2;
getStageIds(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.PoolData
repeated int64 stage_ids = 2;
getStageIds(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.PoolDataOrBuilder
repeated int64 stage_ids = 2;
getStageIdsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
repeated int64 stage_ids = 6;
getStageIdsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
repeated int64 stage_ids = 6;
getStageIdsCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataOrBuilder
repeated int64 stage_ids = 6;
getStageIdsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.PoolData.Builder
repeated int64 stage_ids = 2;
getStageIdsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.PoolData
repeated int64 stage_ids = 2;
getStageIdsCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.PoolDataOrBuilder
repeated int64 stage_ids = 2;
getStageIdsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
repeated int64 stage_ids = 6;
getStageIdsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
repeated int64 stage_ids = 6;
getStageIdsList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataOrBuilder
repeated int64 stage_ids = 6;
getStageIdsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.PoolData.Builder
repeated int64 stage_ids = 2;
getStageIdsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.PoolData
repeated int64 stage_ids = 2;
getStageIdsList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.PoolDataOrBuilder
repeated int64 stage_ids = 2;
getStageInfo(int) - Method in class org.apache.spark.api.java.JavaSparkStatusTracker
Returns stage information, or null if the stage info could not be found or was garbage collected.
getStageInfo(int) - Method in class org.apache.spark.SparkStatusTracker
Returns stage information, or None if the stage info could not be found or was garbage collected.
getStagePath(String, int, int, String) - Method in class org.apache.spark.ml.Pipeline.SharedReadWrite$
Get path for saving the given stage.
getStages() - Method in class org.apache.spark.ml.Pipeline
 
getStages(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
repeated int64 stages = 12;
getStages(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
repeated int64 stages = 12;
getStages(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
repeated int64 stages = 12;
getStagesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
repeated int64 stages = 12;
getStagesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
repeated int64 stages = 12;
getStagesCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
repeated int64 stages = 12;
getStagesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
repeated int64 stages = 12;
getStagesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
repeated int64 stages = 12;
getStagesList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
repeated int64 stages = 12;
getStandardization() - Method in interface org.apache.spark.ml.param.shared.HasStandardization
 
getStartOffset() - Static method in class org.apache.spark.rdd.InputFileBlockHolder
Returns the starting offset of the block currently being read, or -1 if it is unknown.
getStartOffset() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
optional string start_offset = 2;
getStartOffset() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress
optional string start_offset = 2;
getStartOffset() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SourceProgressOrBuilder
optional string start_offset = 2;
getStartOffsetBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
optional string start_offset = 2;
getStartOffsetBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress
optional string start_offset = 2;
getStartOffsetBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SourceProgressOrBuilder
optional string start_offset = 2;
getStartTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo.Builder
int64 start_time = 2;
getStartTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo
int64 start_time = 2;
getStartTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfoOrBuilder
int64 start_time = 2;
getStartTimeEpoch() - Method in class org.apache.spark.status.api.v1.ApplicationAttemptInfo
 
getStartTimestamp() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData.Builder
int64 start_timestamp = 6;
getStartTimestamp() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData
int64 start_timestamp = 6;
getStartTimestamp() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryDataOrBuilder
int64 start_timestamp = 6;
getState() - Method in interface org.apache.spark.launcher.SparkAppHandle
Returns the current application state.
getState() - Method in class org.apache.spark.streaming.api.java.JavaStreamingContext
Deprecated.
:: DeveloperApi ::
getState() - Method in class org.apache.spark.streaming.StreamingContext
Deprecated.
:: DeveloperApi ::
getStatement() - Method in class org.apache.spark.ml.feature.SQLTransformer
 
getStateOperators(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
repeated .org.apache.spark.status.protobuf.StateOperatorProgress state_operators = 9;
getStateOperators(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
repeated .org.apache.spark.status.protobuf.StateOperatorProgress state_operators = 9;
getStateOperators(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
repeated .org.apache.spark.status.protobuf.StateOperatorProgress state_operators = 9;
getStateOperatorsBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
repeated .org.apache.spark.status.protobuf.StateOperatorProgress state_operators = 9;
getStateOperatorsBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
repeated .org.apache.spark.status.protobuf.StateOperatorProgress state_operators = 9;
getStateOperatorsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
repeated .org.apache.spark.status.protobuf.StateOperatorProgress state_operators = 9;
getStateOperatorsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
repeated .org.apache.spark.status.protobuf.StateOperatorProgress state_operators = 9;
getStateOperatorsCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
repeated .org.apache.spark.status.protobuf.StateOperatorProgress state_operators = 9;
getStateOperatorsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
repeated .org.apache.spark.status.protobuf.StateOperatorProgress state_operators = 9;
getStateOperatorsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
repeated .org.apache.spark.status.protobuf.StateOperatorProgress state_operators = 9;
getStateOperatorsList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
repeated .org.apache.spark.status.protobuf.StateOperatorProgress state_operators = 9;
getStateOperatorsOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
repeated .org.apache.spark.status.protobuf.StateOperatorProgress state_operators = 9;
getStateOperatorsOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
repeated .org.apache.spark.status.protobuf.StateOperatorProgress state_operators = 9;
getStateOperatorsOrBuilder(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
repeated .org.apache.spark.status.protobuf.StateOperatorProgress state_operators = 9;
getStateOperatorsOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
repeated .org.apache.spark.status.protobuf.StateOperatorProgress state_operators = 9;
getStateOperatorsOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
repeated .org.apache.spark.status.protobuf.StateOperatorProgress state_operators = 9;
getStateOperatorsOrBuilderList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
repeated .org.apache.spark.status.protobuf.StateOperatorProgress state_operators = 9;
getStatus() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
.org.apache.spark.status.protobuf.JobExecutionStatus status = 8;
getStatus() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
.org.apache.spark.status.protobuf.JobExecutionStatus status = 8;
getStatus() - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataOrBuilder
.org.apache.spark.status.protobuf.JobExecutionStatus status = 8;
getStatus() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
.org.apache.spark.status.protobuf.StageStatus status = 1;
getStatus() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
.org.apache.spark.status.protobuf.StageStatus status = 1;
getStatus() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
.org.apache.spark.status.protobuf.StageStatus status = 1;
getStatus() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
optional string status = 10;
getStatus() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
optional string status = 10;
getStatus() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataOrBuilder
optional string status = 10;
getStatus() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
optional string status = 10;
getStatus() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
optional string status = 10;
getStatus() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
optional string status = 10;
getStatusBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
optional string status = 10;
getStatusBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
optional string status = 10;
getStatusBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataOrBuilder
optional string status = 10;
getStatusBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
optional string status = 10;
getStatusBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
optional string status = 10;
getStatusBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
optional string status = 10;
getStatusValue() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
.org.apache.spark.status.protobuf.JobExecutionStatus status = 8;
getStatusValue() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
.org.apache.spark.status.protobuf.JobExecutionStatus status = 8;
getStatusValue() - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataOrBuilder
.org.apache.spark.status.protobuf.JobExecutionStatus status = 8;
getStatusValue() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
.org.apache.spark.status.protobuf.StageStatus status = 1;
getStatusValue() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
.org.apache.spark.status.protobuf.StageStatus status = 1;
getStatusValue() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
.org.apache.spark.status.protobuf.StageStatus status = 1;
getStderr(Process, long) - Static method in class org.apache.spark.util.Utils
Return the stderr of a process after waiting for the process to terminate.
getStepSize() - Method in interface org.apache.spark.ml.param.shared.HasStepSize
 
getStopWords() - Method in class org.apache.spark.ml.feature.StopWordsRemover
 
getStorageLevel() - Method in interface org.apache.spark.api.java.JavaRDDLike
Get the RDD's current storage level, or StorageLevel.NONE if none is set.
getStorageLevel() - Method in class org.apache.spark.graphx.impl.EdgeRDDImpl
 
getStorageLevel() - Method in class org.apache.spark.graphx.impl.VertexRDDImpl
 
getStorageLevel() - Method in class org.apache.spark.rdd.RDD
Get the RDD's current storage level, or StorageLevel.NONE if none is set.
getStorageLevel() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo.Builder
optional string storage_level = 2;
getStorageLevel() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo
optional string storage_level = 2;
getStorageLevel() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfoOrBuilder
optional string storage_level = 2;
getStorageLevel() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
optional string storage_level = 5;
getStorageLevel() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo
optional string storage_level = 5;
getStorageLevel() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoOrBuilder
optional string storage_level = 5;
getStorageLevel() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
optional string storage_level = 4;
getStorageLevel() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData
optional string storage_level = 4;
getStorageLevel() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamBlockDataOrBuilder
optional string storage_level = 4;
getStorageLevelBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo.Builder
optional string storage_level = 2;
getStorageLevelBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo
optional string storage_level = 2;
getStorageLevelBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfoOrBuilder
optional string storage_level = 2;
getStorageLevelBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
optional string storage_level = 5;
getStorageLevelBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo
optional string storage_level = 5;
getStorageLevelBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoOrBuilder
optional string storage_level = 5;
getStorageLevelBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
optional string storage_level = 4;
getStorageLevelBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData
optional string storage_level = 4;
getStorageLevelBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamBlockDataOrBuilder
optional string storage_level = 4;
GetStorageStatus$() - Constructor for class org.apache.spark.storage.BlockManagerMessages.GetStorageStatus$
 
getStrategy() - Method in interface org.apache.spark.ml.feature.ImputerParams
 
getString(int) - Method in interface org.apache.spark.sql.Row
Returns the value at position i as a String object.
getString(String) - Method in class org.apache.spark.sql.types.Metadata
Gets a String.
getStringArray(String) - Method in class org.apache.spark.sql.types.Metadata
Gets a String array.
getStringField(boolean, Function0<String>) - Static method in class org.apache.spark.status.protobuf.Utils
 
getStringIndexerOrderType() - Method in interface org.apache.spark.ml.feature.RFormulaBase
 
getStringOrderType() - Method in interface org.apache.spark.ml.feature.StringIndexerBase
 
getStruct(int) - Method in interface org.apache.spark.sql.Row
Returns the value at position i of struct type as a Row object.
getStruct(int, int) - Method in class org.apache.spark.sql.vectorized.ColumnarArray
 
getStruct(int, int) - Method in class org.apache.spark.sql.vectorized.ColumnarBatchRow
 
getStruct(int, int) - Method in class org.apache.spark.sql.vectorized.ColumnarRow
 
getStruct(int) - Method in class org.apache.spark.sql.vectorized.ColumnVector
Returns the struct type value for rowId.
getSubmissionTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
optional int64 submission_time = 4;
getSubmissionTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
optional int64 submission_time = 4;
getSubmissionTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataOrBuilder
optional int64 submission_time = 4;
getSubmissionTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
int64 submission_time = 8;
getSubmissionTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
int64 submission_time = 8;
getSubmissionTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
int64 submission_time = 8;
getSubmissionTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
optional int64 submission_time = 10;
getSubmissionTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
optional int64 submission_time = 10;
getSubmissionTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
optional int64 submission_time = 10;
getSubsamplingRate() - Method in interface org.apache.spark.ml.clustering.LDAParams
 
getSubsamplingRate() - Method in interface org.apache.spark.ml.tree.TreeEnsembleParams
 
getSubsamplingRate() - Method in class org.apache.spark.mllib.tree.configuration.Strategy
 
getSucceededTasks(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double succeeded_tasks = 4;
getSucceededTasks(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
repeated double succeeded_tasks = 4;
getSucceededTasks(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
repeated double succeeded_tasks = 4;
getSucceededTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
int32 succeeded_tasks = 3;
getSucceededTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary
int32 succeeded_tasks = 3;
getSucceededTasks() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryOrBuilder
int32 succeeded_tasks = 3;
getSucceededTasksCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double succeeded_tasks = 4;
getSucceededTasksCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
repeated double succeeded_tasks = 4;
getSucceededTasksCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
repeated double succeeded_tasks = 4;
getSucceededTasksList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double succeeded_tasks = 4;
getSucceededTasksList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
repeated double succeeded_tasks = 4;
getSucceededTasksList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
repeated double succeeded_tasks = 4;
getSummary(SQLQueryContext) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
getSummary(SQLQueryContext) - Method in interface org.apache.spark.sql.errors.QueryErrorsBase
 
getSummary(SQLQueryContext) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
getSummary(SQLQueryContext) - Static method in class org.apache.spark.sql.errors.QueryParsingErrors
 
getSystemProperties(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings system_properties = 4;
getSystemProperties(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
repeated .org.apache.spark.status.protobuf.PairStrings system_properties = 4;
getSystemProperties(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoOrBuilder
repeated .org.apache.spark.status.protobuf.PairStrings system_properties = 4;
getSystemProperties() - Static method in class org.apache.spark.util.Utils
Returns the system properties map that is thread-safe to iterator over.
getSystemPropertiesBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings system_properties = 4;
getSystemPropertiesBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings system_properties = 4;
getSystemPropertiesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings system_properties = 4;
getSystemPropertiesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
repeated .org.apache.spark.status.protobuf.PairStrings system_properties = 4;
getSystemPropertiesCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoOrBuilder
repeated .org.apache.spark.status.protobuf.PairStrings system_properties = 4;
getSystemPropertiesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings system_properties = 4;
getSystemPropertiesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
repeated .org.apache.spark.status.protobuf.PairStrings system_properties = 4;
getSystemPropertiesList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoOrBuilder
repeated .org.apache.spark.status.protobuf.PairStrings system_properties = 4;
getSystemPropertiesOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings system_properties = 4;
getSystemPropertiesOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
repeated .org.apache.spark.status.protobuf.PairStrings system_properties = 4;
getSystemPropertiesOrBuilder(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoOrBuilder
repeated .org.apache.spark.status.protobuf.PairStrings system_properties = 4;
getSystemPropertiesOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
repeated .org.apache.spark.status.protobuf.PairStrings system_properties = 4;
getSystemPropertiesOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
repeated .org.apache.spark.status.protobuf.PairStrings system_properties = 4;
getSystemPropertiesOrBuilderList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoOrBuilder
repeated .org.apache.spark.status.protobuf.PairStrings system_properties = 4;
getTable(String) - Method in class org.apache.spark.sql.catalog.Catalog
Get the table or view with the specified name.
getTable(String, String) - Method in class org.apache.spark.sql.catalog.Catalog
Get the table or view with the specified name in the specified database under the Hive Metastore.
getTable(CatalogPlugin, Identifier, Option<TimeTravelSpec>) - Static method in class org.apache.spark.sql.connector.catalog.CatalogV2Util
 
getTable(StructType, Transform[], Map<String, String>) - Method in interface org.apache.spark.sql.connector.catalog.TableProvider
Return a Table instance with the specified table schema, partitioning and properties to do read/write.
getTableCommentQuery(String, String) - Static method in class org.apache.spark.sql.jdbc.DB2Dialect
 
getTableCommentQuery(String, String) - Static method in class org.apache.spark.sql.jdbc.DerbyDialect
 
getTableCommentQuery(String, String) - Static method in class org.apache.spark.sql.jdbc.H2Dialect
 
getTableCommentQuery(String, String) - Method in class org.apache.spark.sql.jdbc.JdbcDialect
 
getTableCommentQuery(String, String) - Static method in class org.apache.spark.sql.jdbc.MsSqlServerDialect
 
getTableCommentQuery(String, String) - Static method in class org.apache.spark.sql.jdbc.MySQLDialect
 
getTableCommentQuery(String, String) - Static method in class org.apache.spark.sql.jdbc.NoopDialect
 
getTableCommentQuery(String, String) - Static method in class org.apache.spark.sql.jdbc.OracleDialect
 
getTableCommentQuery(String, String) - Static method in class org.apache.spark.sql.jdbc.PostgresDialect
 
getTableCommentQuery(String, String) - Static method in class org.apache.spark.sql.jdbc.TeradataDialect
 
getTableExistsQuery(String) - Method in class org.apache.spark.sql.jdbc.AggregatedDialect
 
getTableExistsQuery(String) - Static method in class org.apache.spark.sql.jdbc.DB2Dialect
 
getTableExistsQuery(String) - Static method in class org.apache.spark.sql.jdbc.DerbyDialect
 
getTableExistsQuery(String) - Static method in class org.apache.spark.sql.jdbc.H2Dialect
 
getTableExistsQuery(String) - Method in class org.apache.spark.sql.jdbc.JdbcDialect
Get the SQL query that should be used to find if the given table exists.
getTableExistsQuery(String) - Static method in class org.apache.spark.sql.jdbc.MsSqlServerDialect
 
getTableExistsQuery(String) - Static method in class org.apache.spark.sql.jdbc.MySQLDialect
 
getTableExistsQuery(String) - Static method in class org.apache.spark.sql.jdbc.NoopDialect
 
getTableExistsQuery(String) - Static method in class org.apache.spark.sql.jdbc.OracleDialect
 
getTableExistsQuery(String) - Static method in class org.apache.spark.sql.jdbc.PostgresDialect
 
getTableExistsQuery(String) - Static method in class org.apache.spark.sql.jdbc.TeradataDialect
 
getTableNames(SparkSession, String) - Static method in class org.apache.spark.sql.api.r.SQLUtils
 
getTableParameters(HttpServletRequest, String, String) - Method in interface org.apache.spark.ui.PagedTable
Returns parameter of this table.
getTableProviderCatalog(SupportsCatalogOptions, CatalogManager, CaseInsensitiveStringMap) - Static method in class org.apache.spark.sql.connector.catalog.CatalogV2Util
 
getTableSample(TableSampleInfo) - Static method in class org.apache.spark.sql.jdbc.DB2Dialect
 
getTableSample(TableSampleInfo) - Static method in class org.apache.spark.sql.jdbc.DerbyDialect
 
getTableSample(TableSampleInfo) - Static method in class org.apache.spark.sql.jdbc.H2Dialect
 
getTableSample(TableSampleInfo) - Method in class org.apache.spark.sql.jdbc.JdbcDialect
 
getTableSample(TableSampleInfo) - Static method in class org.apache.spark.sql.jdbc.MsSqlServerDialect
 
getTableSample(TableSampleInfo) - Static method in class org.apache.spark.sql.jdbc.MySQLDialect
 
getTableSample(TableSampleInfo) - Static method in class org.apache.spark.sql.jdbc.NoopDialect
 
getTableSample(TableSampleInfo) - Static method in class org.apache.spark.sql.jdbc.OracleDialect
 
getTableSample(TableSampleInfo) - Static method in class org.apache.spark.sql.jdbc.PostgresDialect
 
getTableSample(TableSampleInfo) - Static method in class org.apache.spark.sql.jdbc.TeradataDialect
 
getTablesByTypeUnsupportedByHiveVersionError() - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
 
getTaskCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
int64 task_count = 4;
getTaskCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
int64 task_count = 4;
getTaskCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
int64 task_count = 4;
getTaskId() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
int64 task_id = 1;
getTaskId() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
int64 task_id = 1;
getTaskId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataOrBuilder
int64 task_id = 1;
getTaskId() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
int64 task_id = 1;
getTaskId() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
int64 task_id = 1;
getTaskId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
int64 task_id = 1;
getTaskInfos() - Method in class org.apache.spark.BarrierTaskContext
:: Experimental :: Returns BarrierTaskInfo for all tasks in this barrier stage, ordered by partition ID.
getTaskLocality() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
optional string task_locality = 11;
getTaskLocality() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
optional string task_locality = 11;
getTaskLocality() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataOrBuilder
optional string task_locality = 11;
getTaskLocality() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
optional string task_locality = 11;
getTaskLocality() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
optional string task_locality = 11;
getTaskLocality() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
optional string task_locality = 11;
getTaskLocalityBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
optional string task_locality = 11;
getTaskLocalityBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
optional string task_locality = 11;
getTaskLocalityBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataOrBuilder
optional string task_locality = 11;
getTaskLocalityBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
optional string task_locality = 11;
getTaskLocalityBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
optional string task_locality = 11;
getTaskLocalityBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
optional string task_locality = 11;
getTaskMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
optional .org.apache.spark.status.protobuf.TaskMetrics task_metrics = 15;
getTaskMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
optional .org.apache.spark.status.protobuf.TaskMetrics task_metrics = 15;
getTaskMetrics() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataOrBuilder
optional .org.apache.spark.status.protobuf.TaskMetrics task_metrics = 15;
getTaskMetricsBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
optional .org.apache.spark.status.protobuf.TaskMetrics task_metrics = 15;
getTaskMetricsDistributions() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
optional .org.apache.spark.status.protobuf.TaskMetricDistributions task_metrics_distributions = 51;
getTaskMetricsDistributions() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
optional .org.apache.spark.status.protobuf.TaskMetricDistributions task_metrics_distributions = 51;
getTaskMetricsDistributions() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
optional .org.apache.spark.status.protobuf.TaskMetricDistributions task_metrics_distributions = 51;
getTaskMetricsDistributionsBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
optional .org.apache.spark.status.protobuf.TaskMetricDistributions task_metrics_distributions = 51;
getTaskMetricsDistributionsOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
optional .org.apache.spark.status.protobuf.TaskMetricDistributions task_metrics_distributions = 51;
getTaskMetricsDistributionsOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
optional .org.apache.spark.status.protobuf.TaskMetricDistributions task_metrics_distributions = 51;
getTaskMetricsDistributionsOrBuilder() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
optional .org.apache.spark.status.protobuf.TaskMetricDistributions task_metrics_distributions = 51;
getTaskMetricsOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
optional .org.apache.spark.status.protobuf.TaskMetrics task_metrics = 15;
getTaskMetricsOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
optional .org.apache.spark.status.protobuf.TaskMetrics task_metrics = 15;
getTaskMetricsOrBuilder() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataOrBuilder
optional .org.apache.spark.status.protobuf.TaskMetrics task_metrics = 15;
getTaskResources() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo.Builder
Deprecated.
getTaskResources() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo
Deprecated.
getTaskResources() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfoOrBuilder
Deprecated.
getTaskResourcesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo.Builder
 
getTaskResourcesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo
 
getTaskResourcesCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfoOrBuilder
map<string, .org.apache.spark.status.protobuf.TaskResourceRequest> task_resources = 3;
getTaskResourcesMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo.Builder
map<string, .org.apache.spark.status.protobuf.TaskResourceRequest> task_resources = 3;
getTaskResourcesMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo
map<string, .org.apache.spark.status.protobuf.TaskResourceRequest> task_resources = 3;
getTaskResourcesMap() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfoOrBuilder
map<string, .org.apache.spark.status.protobuf.TaskResourceRequest> task_resources = 3;
getTaskResourcesOrDefault(String, StoreTypes.TaskResourceRequest) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo.Builder
map<string, .org.apache.spark.status.protobuf.TaskResourceRequest> task_resources = 3;
getTaskResourcesOrDefault(String, StoreTypes.TaskResourceRequest) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo
map<string, .org.apache.spark.status.protobuf.TaskResourceRequest> task_resources = 3;
getTaskResourcesOrDefault(String, StoreTypes.TaskResourceRequest) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfoOrBuilder
map<string, .org.apache.spark.status.protobuf.TaskResourceRequest> task_resources = 3;
getTaskResourcesOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo.Builder
map<string, .org.apache.spark.status.protobuf.TaskResourceRequest> task_resources = 3;
getTaskResourcesOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo
map<string, .org.apache.spark.status.protobuf.TaskResourceRequest> task_resources = 3;
getTaskResourcesOrThrow(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfoOrBuilder
map<string, .org.apache.spark.status.protobuf.TaskResourceRequest> task_resources = 3;
getTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
Deprecated.
getTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
Deprecated.
getTasks() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
Deprecated.
getTasksCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
 
getTasksCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
 
getTasksCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
map<int64, .org.apache.spark.status.protobuf.TaskData> tasks = 45;
getTasksMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
map<int64, .org.apache.spark.status.protobuf.TaskData> tasks = 45;
getTasksMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
map<int64, .org.apache.spark.status.protobuf.TaskData> tasks = 45;
getTasksMap() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
map<int64, .org.apache.spark.status.protobuf.TaskData> tasks = 45;
getTasksOrDefault(long, StoreTypes.TaskData) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
map<int64, .org.apache.spark.status.protobuf.TaskData> tasks = 45;
getTasksOrDefault(long, StoreTypes.TaskData) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
map<int64, .org.apache.spark.status.protobuf.TaskData> tasks = 45;
getTasksOrDefault(long, StoreTypes.TaskData) - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
map<int64, .org.apache.spark.status.protobuf.TaskData> tasks = 45;
getTasksOrThrow(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
map<int64, .org.apache.spark.status.protobuf.TaskData> tasks = 45;
getTasksOrThrow(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
map<int64, .org.apache.spark.status.protobuf.TaskData> tasks = 45;
getTasksOrThrow(long) - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
map<int64, .org.apache.spark.status.protobuf.TaskData> tasks = 45;
getTaskTime(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double task_time = 2;
getTaskTime(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
repeated double task_time = 2;
getTaskTime(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
repeated double task_time = 2;
getTaskTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
int64 task_time = 1;
getTaskTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary
int64 task_time = 1;
getTaskTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryOrBuilder
int64 task_time = 1;
getTaskTimeCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double task_time = 2;
getTaskTimeCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
repeated double task_time = 2;
getTaskTimeCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
repeated double task_time = 2;
getTaskTimeList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
repeated double task_time = 2;
getTaskTimeList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
repeated double task_time = 2;
getTaskTimeList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
repeated double task_time = 2;
getTau0() - Method in class org.apache.spark.mllib.clustering.OnlineLDAOptimizer
A (positive) learning parameter that downweights early iterations.
getThreadDump() - Static method in class org.apache.spark.util.Utils
Return a thread dump of all threads' stacktraces.
getThreadDumpForThread(long) - Static method in class org.apache.spark.util.Utils
 
getThreshold() - Method in class org.apache.spark.ml.classification.LogisticRegression
 
getThreshold() - Method in class org.apache.spark.ml.classification.LogisticRegressionModel
 
getThreshold() - Method in interface org.apache.spark.ml.classification.LogisticRegressionParams
Get threshold for binary classification.
getThreshold() - Method in interface org.apache.spark.ml.param.shared.HasThreshold
 
getThreshold() - Method in class org.apache.spark.mllib.classification.LogisticRegressionModel
Returns the threshold (if any) used for converting raw prediction scores into 0/1 predictions.
getThreshold() - Method in class org.apache.spark.mllib.classification.SVMModel
Returns the threshold (if any) used for converting raw prediction scores into 0/1 predictions.
getThresholds() - Method in class org.apache.spark.ml.classification.LogisticRegression
 
getThresholds() - Method in class org.apache.spark.ml.classification.LogisticRegressionModel
 
getThresholds() - Method in interface org.apache.spark.ml.classification.LogisticRegressionParams
Get thresholds for binary or multiclass classification.
getThresholds() - Method in interface org.apache.spark.ml.param.shared.HasThresholds
 
getThroughOrigin() - Method in class org.apache.spark.ml.evaluation.RegressionEvaluator
 
getTimeAsMs(String) - Method in class org.apache.spark.SparkConf
Get a time parameter as milliseconds; throws a NoSuchElementException if it's not set.
getTimeAsMs(String, String) - Method in class org.apache.spark.SparkConf
Get a time parameter as milliseconds, falling back to a default if not set.
getTimeAsSeconds(String) - Method in class org.apache.spark.SparkConf
Get a time parameter as seconds; throws a NoSuchElementException if it's not set.
getTimeAsSeconds(String, String) - Method in class org.apache.spark.SparkConf
Get a time parameter as seconds, falling back to a default if not set.
getTimeMillis() - Method in interface org.apache.spark.util.Clock
 
getTimeoutTimestampMs() - Method in interface org.apache.spark.sql.streaming.TestGroupState
Returns the timestamp if setTimeoutTimestamp() is called.
getTimer(L) - Method in interface org.apache.spark.util.ListenerBus
Returns a CodaHale metrics Timer for measuring the listener's event processing time.
getTimestamp(int) - Method in interface org.apache.spark.sql.Row
Returns the value at position i of date type as java.sql.Timestamp.
getTimestamp() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
optional string timestamp = 4;
getTimestamp() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
optional string timestamp = 4;
getTimestamp() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
optional string timestamp = 4;
getTimestamp() - Method in class org.apache.spark.streaming.kinesis.KinesisInitialPositions.AtTimestamp
 
getTimestampBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
optional string timestamp = 4;
getTimestampBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
optional string timestamp = 4;
getTimestampBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
optional string timestamp = 4;
getTimeZoneOffset() - Static method in class org.apache.spark.ui.UIUtils
 
GETTING_RESULT_TIME() - Static method in class org.apache.spark.status.TaskIndexNames
 
GETTING_RESULT_TIME() - Static method in class org.apache.spark.ui.jobs.TaskDetailsClassNames
 
GETTING_RESULT_TIME() - Static method in class org.apache.spark.ui.ToolTips
 
GETTING_RESULT_TIME_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
 
GETTING_RESULT_TIME_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.TaskData
 
GETTING_RESULT_TIME_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
 
gettingResult() - Method in class org.apache.spark.scheduler.TaskInfo
 
gettingResultTime() - Method in class org.apache.spark.scheduler.TaskInfo
The time when the task started remotely getting the result.
gettingResultTime() - Method in class org.apache.spark.status.api.v1.TaskData
 
gettingResultTime() - Method in class org.apache.spark.status.api.v1.TaskMetricDistributions
 
gettingResultTime(TaskData) - Static method in class org.apache.spark.status.AppStatusUtils
 
gettingResultTime(long, long, long) - Static method in class org.apache.spark.status.AppStatusUtils
 
getToId() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationEdge.Builder
int32 to_id = 2;
getToId() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationEdge
int32 to_id = 2;
getToId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationEdgeOrBuilder
int32 to_id = 2;
getToId() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphEdge.Builder
int64 to_id = 2;
getToId() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphEdge
int64 to_id = 2;
getToId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphEdgeOrBuilder
int64 to_id = 2;
getTokenJaasParams(KafkaTokenClusterConf) - Static method in class org.apache.spark.kafka010.KafkaTokenUtil
 
getTol() - Method in interface org.apache.spark.ml.param.shared.HasTol
 
getToLowercase() - Method in class org.apache.spark.ml.feature.RegexTokenizer
 
getTopicConcentration() - Method in interface org.apache.spark.ml.clustering.LDAParams
 
getTopicConcentration() - Method in class org.apache.spark.mllib.clustering.LDA
Concentration parameter (commonly named "beta" or "eta") for the prior placed on topics' distributions over terms.
getTopicDistributionCol() - Method in interface org.apache.spark.ml.clustering.LDAParams
 
getTopologyForHost(String) - Method in class org.apache.spark.storage.DefaultTopologyMapper
 
getTopologyForHost(String) - Method in class org.apache.spark.storage.FileBasedTopologyMapper
 
getTopologyForHost(String) - Method in class org.apache.spark.storage.TopologyMapper
Gets the topology information given the host name
getTotalBlocksFetched(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
repeated double total_blocks_fetched = 8;
getTotalBlocksFetched(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions
repeated double total_blocks_fetched = 8;
getTotalBlocksFetched(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributionsOrBuilder
repeated double total_blocks_fetched = 8;
getTotalBlocksFetchedCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
repeated double total_blocks_fetched = 8;
getTotalBlocksFetchedCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions
repeated double total_blocks_fetched = 8;
getTotalBlocksFetchedCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributionsOrBuilder
repeated double total_blocks_fetched = 8;
getTotalBlocksFetchedList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
repeated double total_blocks_fetched = 8;
getTotalBlocksFetchedList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions
repeated double total_blocks_fetched = 8;
getTotalBlocksFetchedList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributionsOrBuilder
repeated double total_blocks_fetched = 8;
getTotalCores() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
int32 total_cores = 7;
getTotalCores() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
int32 total_cores = 7;
getTotalCores() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
int32 total_cores = 7;
getTotalCores() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder
int32 total_cores = 4;
getTotalCores() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary
int32 total_cores = 4;
getTotalCores() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryOrBuilder
int32 total_cores = 4;
getTotalDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
int64 total_duration = 13;
getTotalDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
int64 total_duration = 13;
getTotalDuration() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
int64 total_duration = 13;
getTotalGcTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
int64 total_gc_time = 14;
getTotalGcTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
int64 total_gc_time = 14;
getTotalGcTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
int64 total_gc_time = 14;
getTotalInputBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
int64 total_input_bytes = 15;
getTotalInputBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
int64 total_input_bytes = 15;
getTotalInputBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
int64 total_input_bytes = 15;
getTotalOffHeapStorageMemory() - Method in class org.apache.spark.status.protobuf.StoreTypes.MemoryMetrics.Builder
int64 total_off_heap_storage_memory = 4;
getTotalOffHeapStorageMemory() - Method in class org.apache.spark.status.protobuf.StoreTypes.MemoryMetrics
int64 total_off_heap_storage_memory = 4;
getTotalOffHeapStorageMemory() - Method in interface org.apache.spark.status.protobuf.StoreTypes.MemoryMetricsOrBuilder
int64 total_off_heap_storage_memory = 4;
getTotalOnHeapStorageMemory() - Method in class org.apache.spark.status.protobuf.StoreTypes.MemoryMetrics.Builder
int64 total_on_heap_storage_memory = 3;
getTotalOnHeapStorageMemory() - Method in class org.apache.spark.status.protobuf.StoreTypes.MemoryMetrics
int64 total_on_heap_storage_memory = 3;
getTotalOnHeapStorageMemory() - Method in interface org.apache.spark.status.protobuf.StoreTypes.MemoryMetricsOrBuilder
int64 total_on_heap_storage_memory = 3;
getTotalShuffleRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
int64 total_shuffle_read = 16;
getTotalShuffleRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
int64 total_shuffle_read = 16;
getTotalShuffleRead() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
int64 total_shuffle_read = 16;
getTotalShuffleWrite() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
int64 total_shuffle_write = 17;
getTotalShuffleWrite() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
int64 total_shuffle_write = 17;
getTotalShuffleWrite() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
int64 total_shuffle_write = 17;
getTotalTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
int32 total_tasks = 12;
getTotalTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
int32 total_tasks = 12;
getTotalTasks() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
int32 total_tasks = 12;
getTrainRatio() - Method in interface org.apache.spark.ml.tuning.TrainValidationSplitParams
 
getTreeStrategy() - Method in class org.apache.spark.mllib.tree.configuration.BoostingStrategy
 
getTruncateQuery(String, Option<Object>) - Method in class org.apache.spark.sql.jdbc.AggregatedDialect
The SQL query used to truncate a table.
getTruncateQuery(String, Option<Object>) - Static method in class org.apache.spark.sql.jdbc.DB2Dialect
 
getTruncateQuery(String) - Static method in class org.apache.spark.sql.jdbc.DerbyDialect
 
getTruncateQuery(String, Option<Object>) - Static method in class org.apache.spark.sql.jdbc.DerbyDialect
 
getTruncateQuery(String) - Static method in class org.apache.spark.sql.jdbc.H2Dialect
 
getTruncateQuery(String, Option<Object>) - Static method in class org.apache.spark.sql.jdbc.H2Dialect
 
getTruncateQuery(String) - Method in class org.apache.spark.sql.jdbc.JdbcDialect
The SQL query that should be used to truncate a table.
getTruncateQuery(String, Option<Object>) - Method in class org.apache.spark.sql.jdbc.JdbcDialect
The SQL query that should be used to truncate a table.
getTruncateQuery(String) - Static method in class org.apache.spark.sql.jdbc.MsSqlServerDialect
 
getTruncateQuery(String, Option<Object>) - Static method in class org.apache.spark.sql.jdbc.MsSqlServerDialect
 
getTruncateQuery(String) - Static method in class org.apache.spark.sql.jdbc.MySQLDialect
 
getTruncateQuery(String, Option<Object>) - Static method in class org.apache.spark.sql.jdbc.MySQLDialect
 
getTruncateQuery(String) - Static method in class org.apache.spark.sql.jdbc.NoopDialect
 
getTruncateQuery(String, Option<Object>) - Static method in class org.apache.spark.sql.jdbc.NoopDialect
 
getTruncateQuery(String, Option<Object>) - Static method in class org.apache.spark.sql.jdbc.OracleDialect
The SQL query used to truncate a table.
getTruncateQuery(String, Option<Object>) - Static method in class org.apache.spark.sql.jdbc.PostgresDialect
The SQL query used to truncate a table.
getTruncateQuery(String, Option<Object>) - Static method in class org.apache.spark.sql.jdbc.TeradataDialect
The SQL query used to truncate a table.
getTruncateQuery$default$2() - Static method in class org.apache.spark.sql.jdbc.DB2Dialect
 
getTruncateQuery$default$2() - Static method in class org.apache.spark.sql.jdbc.DerbyDialect
 
getTruncateQuery$default$2() - Static method in class org.apache.spark.sql.jdbc.H2Dialect
 
getTruncateQuery$default$2() - Static method in class org.apache.spark.sql.jdbc.MsSqlServerDialect
 
getTruncateQuery$default$2() - Static method in class org.apache.spark.sql.jdbc.MySQLDialect
 
getTruncateQuery$default$2() - Static method in class org.apache.spark.sql.jdbc.NoopDialect
 
getTruncateQuery$default$2() - Static method in class org.apache.spark.sql.jdbc.OracleDialect
 
getTruncateQuery$default$2() - Static method in class org.apache.spark.sql.jdbc.PostgresDialect
 
getTruncateQuery$default$2() - Static method in class org.apache.spark.sql.jdbc.TeradataDialect
 
getUDTFor(String) - Static method in class org.apache.spark.sql.types.UDTRegistration
Returns the Class of UserDefinedType for the name of a given user class.
getUidMap(Params) - Static method in class org.apache.spark.ml.util.MetaAlgorithmReadWrite
Examine the given estimator (which may be a compound estimator) and extract a mapping from UIDs to corresponding Params instances.
getUiRoot(ServletContext) - Static method in class org.apache.spark.status.api.v1.UIRootFromServletContext
 
getUnknownFields() - Method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo
 
getUnknownFields() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo
 
getUnknownFields() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
 
getUnknownFields() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoWrapper
 
getUnknownFields() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo
 
getUnknownFields() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoWrapper
 
getUnknownFields() - Method in class org.apache.spark.status.protobuf.StoreTypes.AppSummary
 
getUnknownFields() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
 
getUnknownFields() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetrics
 
getUnknownFields() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
 
getUnknownFields() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions
 
getUnknownFields() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest
 
getUnknownFields() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary
 
getUnknownFields() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper
 
getUnknownFields() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
 
getUnknownFields() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryWrapper
 
getUnknownFields() - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributions
 
getUnknownFields() - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetrics
 
getUnknownFields() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
 
getUnknownFields() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper
 
getUnknownFields() - Method in class org.apache.spark.status.protobuf.StoreTypes.MemoryMetrics
 
getUnknownFields() - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributions
 
getUnknownFields() - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetrics
 
getUnknownFields() - Method in class org.apache.spark.status.protobuf.StoreTypes.PairStrings
 
getUnknownFields() - Method in class org.apache.spark.status.protobuf.StoreTypes.PoolData
 
getUnknownFields() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary
 
getUnknownFields() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryWrapper
 
getUnknownFields() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution
 
getUnknownFields() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper
 
getUnknownFields() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationEdge
 
getUnknownFields() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper
 
getUnknownFields() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode
 
getUnknownFields() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo
 
getUnknownFields() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo
 
getUnknownFields() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoWrapper
 
getUnknownFields() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceInformation
 
getUnknownFields() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo
 
getUnknownFields() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileWrapper
 
getUnknownFields() - Method in class org.apache.spark.status.protobuf.StoreTypes.RuntimeInfo
 
getUnknownFields() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions
 
getUnknownFields() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics
 
getUnknownFields() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions
 
getUnknownFields() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics
 
getUnknownFields() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions
 
getUnknownFields() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetrics
 
getUnknownFields() - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress
 
getUnknownFields() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress
 
getUnknownFields() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper
 
getUnknownFields() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphEdge
 
getUnknownFields() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode
 
getUnknownFields() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper
 
getUnknownFields() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper
 
getUnknownFields() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummary
 
getUnknownFields() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryWrapper
 
getUnknownFields() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
 
getUnknownFields() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetric
 
getUnknownFields() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
 
getUnknownFields() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper
 
getUnknownFields() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress
 
getUnknownFields() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData
 
getUnknownFields() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData
 
getUnknownFields() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
 
getUnknownFields() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressWrapper
 
getUnknownFields() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
 
getUnknownFields() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
 
getUnknownFields() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
 
getUnknownFields() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics
 
getUnknownFields() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskResourceRequest
 
getUpdate() - Method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo.Builder
optional string update = 3;
getUpdate() - Method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo
optional string update = 3;
getUpdate() - Method in interface org.apache.spark.status.protobuf.StoreTypes.AccumulableInfoOrBuilder
optional string update = 3;
getUpdateBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo.Builder
optional string update = 3;
getUpdateBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo
optional string update = 3;
getUpdateBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.AccumulableInfoOrBuilder
optional string update = 3;
getUpdateColumnNullabilityQuery(String, String, boolean) - Static method in class org.apache.spark.sql.jdbc.DB2Dialect
 
getUpdateColumnNullabilityQuery(String, String, boolean) - Static method in class org.apache.spark.sql.jdbc.DerbyDialect
 
getUpdateColumnNullabilityQuery(String, String, boolean) - Static method in class org.apache.spark.sql.jdbc.H2Dialect
 
getUpdateColumnNullabilityQuery(String, String, boolean) - Method in class org.apache.spark.sql.jdbc.JdbcDialect
 
getUpdateColumnNullabilityQuery(String, String, boolean) - Static method in class org.apache.spark.sql.jdbc.MsSqlServerDialect
 
getUpdateColumnNullabilityQuery(String, String, boolean) - Static method in class org.apache.spark.sql.jdbc.MySQLDialect
 
getUpdateColumnNullabilityQuery(String, String, boolean) - Static method in class org.apache.spark.sql.jdbc.NoopDialect
 
getUpdateColumnNullabilityQuery(String, String, boolean) - Static method in class org.apache.spark.sql.jdbc.OracleDialect
 
getUpdateColumnNullabilityQuery(String, String, boolean) - Static method in class org.apache.spark.sql.jdbc.PostgresDialect
 
getUpdateColumnNullabilityQuery(String, String, boolean) - Static method in class org.apache.spark.sql.jdbc.TeradataDialect
 
getUpdateColumnTypeQuery(String, String, String) - Static method in class org.apache.spark.sql.jdbc.DB2Dialect
 
getUpdateColumnTypeQuery(String, String, String) - Static method in class org.apache.spark.sql.jdbc.DerbyDialect
 
getUpdateColumnTypeQuery(String, String, String) - Static method in class org.apache.spark.sql.jdbc.H2Dialect
 
getUpdateColumnTypeQuery(String, String, String) - Method in class org.apache.spark.sql.jdbc.JdbcDialect
 
getUpdateColumnTypeQuery(String, String, String) - Static method in class org.apache.spark.sql.jdbc.MsSqlServerDialect
 
getUpdateColumnTypeQuery(String, String, String) - Static method in class org.apache.spark.sql.jdbc.MySQLDialect
 
getUpdateColumnTypeQuery(String, String, String) - Static method in class org.apache.spark.sql.jdbc.NoopDialect
 
getUpdateColumnTypeQuery(String, String, String) - Static method in class org.apache.spark.sql.jdbc.OracleDialect
 
getUpdateColumnTypeQuery(String, String, String) - Static method in class org.apache.spark.sql.jdbc.PostgresDialect
 
getUpdateColumnTypeQuery(String, String, String) - Static method in class org.apache.spark.sql.jdbc.TeradataDialect
 
getUpper() - Method in interface org.apache.spark.ml.feature.RobustScalerParams
 
getUpperBound(double, long, double) - Static method in class org.apache.spark.util.random.BinomialBounds
Returns a threshold p such that if we conduct n Bernoulli trials with success rate = p, it is very unlikely to have less than fraction * n successes.
getUpperBound(double) - Static method in class org.apache.spark.util.random.PoissonBounds
Returns a lambda such that Pr[X < s] is very small, where X ~ Pois(lambda).
getUpperBoundsOnCoefficients() - Method in interface org.apache.spark.ml.classification.LogisticRegressionParams
 
getUpperBoundsOnIntercepts() - Method in interface org.apache.spark.ml.classification.LogisticRegressionParams
 
getUsedBins() - Method in class org.apache.spark.sql.util.NumericHistogram
Returns the number of bins currently being used by the histogram.
getUseDisk() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
bool use_disk = 6;
getUseDisk() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData
bool use_disk = 6;
getUseDisk() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamBlockDataOrBuilder
bool use_disk = 6;
getUsedOffHeapStorageMemory() - Method in class org.apache.spark.status.protobuf.StoreTypes.MemoryMetrics.Builder
int64 used_off_heap_storage_memory = 2;
getUsedOffHeapStorageMemory() - Method in class org.apache.spark.status.protobuf.StoreTypes.MemoryMetrics
int64 used_off_heap_storage_memory = 2;
getUsedOffHeapStorageMemory() - Method in interface org.apache.spark.status.protobuf.StoreTypes.MemoryMetricsOrBuilder
int64 used_off_heap_storage_memory = 2;
getUsedOnHeapStorageMemory() - Method in class org.apache.spark.status.protobuf.StoreTypes.MemoryMetrics.Builder
int64 used_on_heap_storage_memory = 1;
getUsedOnHeapStorageMemory() - Method in class org.apache.spark.status.protobuf.StoreTypes.MemoryMetrics
int64 used_on_heap_storage_memory = 1;
getUsedOnHeapStorageMemory() - Method in interface org.apache.spark.status.protobuf.StoreTypes.MemoryMetricsOrBuilder
int64 used_on_heap_storage_memory = 1;
getUsedTimeNs(long) - Static method in class org.apache.spark.util.Utils
Return the string to tell how long has passed in milliseconds.
getUseMemory() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
bool use_memory = 5;
getUseMemory() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData
bool use_memory = 5;
getUseMemory() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamBlockDataOrBuilder
bool use_memory = 5;
getUseNodeIdCache() - Method in class org.apache.spark.mllib.tree.configuration.Strategy
 
getUserCol() - Method in interface org.apache.spark.ml.recommendation.ALSModelParams
 
getUserJars(SparkConf) - Static method in class org.apache.spark.util.Utils
Return the jar files pointed by the "spark.jars" property.
getUTF8String(int) - Method in class org.apache.spark.sql.vectorized.ArrowColumnVector
 
getUTF8String(int) - Method in class org.apache.spark.sql.vectorized.ColumnarArray
 
getUTF8String(int) - Method in class org.apache.spark.sql.vectorized.ColumnarBatchRow
 
getUTF8String(int) - Method in class org.apache.spark.sql.vectorized.ColumnarRow
 
getUTF8String(int) - Method in class org.apache.spark.sql.vectorized.ColumnVector
Returns the string type value for rowId.
getValidationIndicatorCol() - Method in interface org.apache.spark.ml.param.shared.HasValidationIndicatorCol
 
getValidationTol() - Method in interface org.apache.spark.ml.tree.GBTParams
 
getValidationTol() - Method in class org.apache.spark.mllib.tree.configuration.BoostingStrategy
 
getValue(int) - Method in class org.apache.spark.ml.attribute.NominalAttribute
Gets a value given its index.
getValue() - Method in class org.apache.spark.mllib.stat.test.BinarySample
 
getValue() - Method in class org.apache.spark.sql.connector.catalog.ColumnDefaultValue
Returns the default value literal.
getValue() - Method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo.Builder
optional string value = 4;
getValue() - Method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo
optional string value = 4;
getValue() - Method in interface org.apache.spark.status.protobuf.StoreTypes.AccumulableInfoOrBuilder
optional string value = 4;
getValue1() - Method in class org.apache.spark.status.protobuf.StoreTypes.PairStrings.Builder
optional string value1 = 1;
getValue1() - Method in class org.apache.spark.status.protobuf.StoreTypes.PairStrings
optional string value1 = 1;
getValue1() - Method in interface org.apache.spark.status.protobuf.StoreTypes.PairStringsOrBuilder
optional string value1 = 1;
getValue1Bytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.PairStrings.Builder
optional string value1 = 1;
getValue1Bytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.PairStrings
optional string value1 = 1;
getValue1Bytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.PairStringsOrBuilder
optional string value1 = 1;
getValue2() - Method in class org.apache.spark.status.protobuf.StoreTypes.PairStrings.Builder
optional string value2 = 2;
getValue2() - Method in class org.apache.spark.status.protobuf.StoreTypes.PairStrings
optional string value2 = 2;
getValue2() - Method in interface org.apache.spark.status.protobuf.StoreTypes.PairStringsOrBuilder
optional string value2 = 2;
getValue2Bytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.PairStrings.Builder
optional string value2 = 2;
getValue2Bytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.PairStrings
optional string value2 = 2;
getValue2Bytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.PairStringsOrBuilder
optional string value2 = 2;
getValueBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo.Builder
optional string value = 4;
getValueBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo
optional string value = 4;
getValueBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.AccumulableInfoOrBuilder
optional string value = 4;
getValueDescriptor() - Method in enum org.apache.spark.status.protobuf.StoreTypes.DeterministicLevel
 
getValueDescriptor() - Method in enum org.apache.spark.status.protobuf.StoreTypes.JobExecutionStatus
 
getValueDescriptor() - Method in enum org.apache.spark.status.protobuf.StoreTypes.StageStatus
 
getValuesMap(Seq<String>) - Method in interface org.apache.spark.sql.Row
Returns a Map consisting of names and values for the requested fieldNames For primitive types if value is null it returns 'zero value' specific for primitive i.e.
getValueVector() - Method in class org.apache.spark.sql.vectorized.ArrowColumnVector
 
getVarianceCol() - Method in interface org.apache.spark.ml.param.shared.HasVarianceCol
 
getVariancePower() - Method in interface org.apache.spark.ml.regression.GeneralizedLinearRegressionBase
 
getVarianceThreshold() - Method in interface org.apache.spark.ml.feature.VarianceThresholdSelectorParams
 
getVectors() - Method in class org.apache.spark.ml.feature.Word2VecModel
 
getVectors() - Method in class org.apache.spark.mllib.feature.Word2VecModel
Returns a map of words to their vector representations.
getVectorSize() - Method in interface org.apache.spark.ml.feature.Word2VecBase
 
getVendor() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest.Builder
optional string vendor = 4;
getVendor() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest
optional string vendor = 4;
getVendor() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequestOrBuilder
optional string vendor = 4;
getVendorBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest.Builder
optional string vendor = 4;
getVendorBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest
optional string vendor = 4;
getVendorBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequestOrBuilder
optional string vendor = 4;
getVocabSize() - Method in interface org.apache.spark.ml.feature.CountVectorizerParams
 
getWeightCol() - Method in interface org.apache.spark.ml.param.shared.HasWeightCol
 
getWidth(Row) - Static method in class org.apache.spark.ml.image.ImageSchema
Gets the width of the image
getWindowSize() - Method in interface org.apache.spark.ml.feature.Word2VecBase
 
getWithCentering() - Method in interface org.apache.spark.ml.feature.RobustScalerParams
 
getWithMean() - Method in interface org.apache.spark.ml.feature.StandardScalerParams
 
getWithScaling() - Method in interface org.apache.spark.ml.feature.RobustScalerParams
 
getWithStd() - Method in interface org.apache.spark.ml.feature.StandardScalerParams
 
getWrapperCase() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper.Builder
 
getWrapperCase() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper
 
getWrapperCase() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapperOrBuilder
 
getWriteBytes(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions.Builder
repeated double write_bytes = 1;
getWriteBytes(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions
repeated double write_bytes = 1;
getWriteBytes(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributionsOrBuilder
repeated double write_bytes = 1;
getWriteBytesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions.Builder
repeated double write_bytes = 1;
getWriteBytesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions
repeated double write_bytes = 1;
getWriteBytesCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributionsOrBuilder
repeated double write_bytes = 1;
getWriteBytesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions.Builder
repeated double write_bytes = 1;
getWriteBytesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions
repeated double write_bytes = 1;
getWriteBytesList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributionsOrBuilder
repeated double write_bytes = 1;
getWriteRecords(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions.Builder
repeated double write_records = 2;
getWriteRecords(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions
repeated double write_records = 2;
getWriteRecords(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributionsOrBuilder
repeated double write_records = 2;
getWriteRecordsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions.Builder
repeated double write_records = 2;
getWriteRecordsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions
repeated double write_records = 2;
getWriteRecordsCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributionsOrBuilder
repeated double write_records = 2;
getWriteRecordsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions.Builder
repeated double write_records = 2;
getWriteRecordsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions
repeated double write_records = 2;
getWriteRecordsList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributionsOrBuilder
repeated double write_records = 2;
getWriteTime(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions.Builder
repeated double write_time = 3;
getWriteTime(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions
repeated double write_time = 3;
getWriteTime(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributionsOrBuilder
repeated double write_time = 3;
getWriteTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetrics.Builder
int64 write_time = 2;
getWriteTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetrics
int64 write_time = 2;
getWriteTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricsOrBuilder
int64 write_time = 2;
getWriteTimeCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions.Builder
repeated double write_time = 3;
getWriteTimeCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions
repeated double write_time = 3;
getWriteTimeCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributionsOrBuilder
repeated double write_time = 3;
getWriteTimeList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions.Builder
repeated double write_time = 3;
getWriteTimeList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions
repeated double write_time = 3;
getWriteTimeList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributionsOrBuilder
repeated double write_time = 3;
Gini - Class in org.apache.spark.mllib.tree.impurity
Class for calculating the Gini impurity (http://en.wikipedia.org/wiki/Decision_tree_learning#Gini_impurity) during multiclass classification.
Gini() - Constructor for class org.apache.spark.mllib.tree.impurity.Gini
 
GLMClassificationModel - Class in org.apache.spark.mllib.classification.impl
Helper class for import/export of GLM classification models.
GLMClassificationModel() - Constructor for class org.apache.spark.mllib.classification.impl.GLMClassificationModel
 
GLMClassificationModel.SaveLoadV1_0$ - Class in org.apache.spark.mllib.classification.impl
 
GLMClassificationModel.SaveLoadV1_0$.Data - Class in org.apache.spark.mllib.classification.impl
Model data for import/export
GLMClassificationModel.SaveLoadV1_0$.Data$ - Class in org.apache.spark.mllib.classification.impl
 
GLMRegressionModel - Class in org.apache.spark.mllib.regression.impl
Helper methods for import/export of GLM regression models.
GLMRegressionModel() - Constructor for class org.apache.spark.mllib.regression.impl.GLMRegressionModel
 
GLMRegressionModel.SaveLoadV1_0$ - Class in org.apache.spark.mllib.regression.impl
 
GLMRegressionModel.SaveLoadV1_0$.Data - Class in org.apache.spark.mllib.regression.impl
Model data for model import/export
GLMRegressionModel.SaveLoadV1_0$.Data$ - Class in org.apache.spark.mllib.regression.impl
 
glom() - Method in interface org.apache.spark.api.java.JavaRDDLike
Return an RDD created by coalescing all elements within each partition into an array.
glom() - Method in class org.apache.spark.rdd.RDD
Return an RDD created by coalescing all elements within each partition into an array.
glom() - Method in interface org.apache.spark.streaming.api.java.JavaDStreamLike
Return a new DStream in which each RDD is generated by applying glom() to each RDD of this DStream.
glom() - Method in class org.apache.spark.streaming.dstream.DStream
Return a new DStream in which each RDD is generated by applying glom() to each RDD of this DStream.
goButtonFormPath() - Method in interface org.apache.spark.ui.PagedTable
Returns the submission path for the "go to page #" form.
goodnessOfFit() - Method in class org.apache.spark.mllib.stat.test.ChiSqTest.NullHypothesis$
 
GPU() - Static method in class org.apache.spark.resource.ResourceUtils
 
grad(DenseMatrix<Object>, DenseMatrix<Object>, DenseVector<Object>) - Method in interface org.apache.spark.ml.ann.LayerModel
Computes the gradient.
grad() - Method in class org.apache.spark.mllib.optimization.NNLS.Workspace
 
gradient() - Method in interface org.apache.spark.ml.optim.aggregator.DifferentiableLossAggregator
The current weighted averaged gradient.
Gradient - Class in org.apache.spark.mllib.optimization
Class used to compute the gradient for a loss function, given a single data point.
Gradient() - Constructor for class org.apache.spark.mllib.optimization.Gradient
 
gradient(double, double) - Static method in class org.apache.spark.mllib.tree.loss.AbsoluteError
Method to calculate the gradients for the gradient boosting calculation for least absolute error calculation.
gradient(double, double) - Static method in class org.apache.spark.mllib.tree.loss.LogLoss
Method to calculate the loss gradients for the gradient boosting calculation for binary classification The gradient with respect to F(x) is: - 4 y / (1 + exp(2 y F(x)))
gradient(double, double) - Method in interface org.apache.spark.mllib.tree.loss.Loss
Method to calculate the gradients for the gradient boosting calculation.
gradient(double, double) - Static method in class org.apache.spark.mllib.tree.loss.SquaredError
Method to calculate the gradients for the gradient boosting calculation for least squares error calculation.
GradientBoostedTrees - Class in org.apache.spark.ml.tree.impl
 
GradientBoostedTrees() - Constructor for class org.apache.spark.ml.tree.impl.GradientBoostedTrees
 
GradientBoostedTrees - Class in org.apache.spark.mllib.tree
A class that implements Stochastic Gradient Boosting for regression and binary classification.
GradientBoostedTrees(BoostingStrategy) - Constructor for class org.apache.spark.mllib.tree.GradientBoostedTrees
 
GradientBoostedTreesModel - Class in org.apache.spark.mllib.tree.model
Represents a gradient boosted trees model.
GradientBoostedTreesModel(Enumeration.Value, DecisionTreeModel[], double[]) - Constructor for class org.apache.spark.mllib.tree.model.GradientBoostedTreesModel
 
GradientDescent - Class in org.apache.spark.mllib.optimization
Class used to solve an optimization problem using Gradient Descent.
gradientSumArray() - Method in interface org.apache.spark.ml.optim.aggregator.DifferentiableLossAggregator
Array of gradient values that are mutated when new instances are added to the aggregator.
Graph<VD,ED> - Class in org.apache.spark.graphx
The Graph abstractly represents a graph with arbitrary objects associated with vertices and edges.
GraphGenerators - Class in org.apache.spark.graphx.util
A collection of graph generating functions.
GraphGenerators() - Constructor for class org.apache.spark.graphx.util.GraphGenerators
 
GraphImpl<VD,ED> - Class in org.apache.spark.graphx.impl
An implementation of Graph to support computation on graphs.
graphiteSinkInvalidProtocolError(String) - Static method in class org.apache.spark.errors.SparkCoreErrors
 
graphiteSinkPropertyMissingError(String) - Static method in class org.apache.spark.errors.SparkCoreErrors
 
GraphLoader - Class in org.apache.spark.graphx
Provides utilities for loading Graphs from files.
GraphLoader() - Constructor for class org.apache.spark.graphx.GraphLoader
 
GraphOps<VD,ED> - Class in org.apache.spark.graphx
Contains additional functionality for Graph.
GraphOps(Graph<VD, ED>, ClassTag<VD>, ClassTag<ED>) - Constructor for class org.apache.spark.graphx.GraphOps
 
graphToGraphOps(Graph<VD, ED>, ClassTag<VD>, ClassTag<ED>) - Static method in class org.apache.spark.graphx.Graph
Implicitly extracts the GraphOps member from a graph.
GraphXUtils - Class in org.apache.spark.graphx
 
GraphXUtils() - Constructor for class org.apache.spark.graphx.GraphXUtils
 
greater(Duration) - Method in class org.apache.spark.streaming.Duration
 
greater(Time) - Method in class org.apache.spark.streaming.Time
 
greaterEq(Duration) - Method in class org.apache.spark.streaming.Duration
 
greaterEq(Time) - Method in class org.apache.spark.streaming.Time
 
GreaterThan - Class in org.apache.spark.sql.sources
A filter that evaluates to true iff the attribute evaluates to a value greater than value.
GreaterThan(String, Object) - Constructor for class org.apache.spark.sql.sources.GreaterThan
 
GreaterThanOrEqual - Class in org.apache.spark.sql.sources
A filter that evaluates to true iff the attribute evaluates to a value greater than or equal to value.
GreaterThanOrEqual(String, Object) - Constructor for class org.apache.spark.sql.sources.GreaterThanOrEqual
 
greatest(Column...) - Static method in class org.apache.spark.sql.functions
Returns the greatest value of the list of values, skipping null values.
greatest(String, String...) - Static method in class org.apache.spark.sql.functions
Returns the greatest value of the list of column names, skipping null values.
greatest(Seq<Column>) - Static method in class org.apache.spark.sql.functions
Returns the greatest value of the list of values, skipping null values.
greatest(String, Seq<String>) - Static method in class org.apache.spark.sql.functions
Returns the greatest value of the list of column names, skipping null values.
gridGraph(SparkContext, int, int) - Static method in class org.apache.spark.graphx.util.GraphGenerators
Create rows by cols grid graph with each vertex connected to its row+1 and col+1 neighbors.
groupArr() - Method in class org.apache.spark.rdd.DefaultPartitionCoalescer
 
groupBy(Function<T, U>) - Method in interface org.apache.spark.api.java.JavaRDDLike
Return an RDD of grouped elements.
groupBy(Function<T, U>, int) - Method in interface org.apache.spark.api.java.JavaRDDLike
Return an RDD of grouped elements.
groupBy(Function1<T, K>, ClassTag<K>) - Method in class org.apache.spark.rdd.RDD
Return an RDD of grouped items.
groupBy(Function1<T, K>, int, ClassTag<K>) - Method in class org.apache.spark.rdd.RDD
Return an RDD of grouped elements.
groupBy(Function1<T, K>, Partitioner, ClassTag<K>, Ordering<K>) - Method in class org.apache.spark.rdd.RDD
Return an RDD of grouped items.
groupBy(Column...) - Method in class org.apache.spark.sql.Dataset
Groups the Dataset using the specified columns, so we can run aggregation on them.
groupBy(String, String...) - Method in class org.apache.spark.sql.Dataset
Groups the Dataset using the specified columns, so that we can run aggregation on them.
groupBy(Seq<Column>) - Method in class org.apache.spark.sql.Dataset
Groups the Dataset using the specified columns, so we can run aggregation on them.
groupBy(String, Seq<String>) - Method in class org.apache.spark.sql.Dataset
Groups the Dataset using the specified columns, so that we can run aggregation on them.
groupByExpressions() - Method in class org.apache.spark.sql.connector.expressions.aggregate.Aggregation
 
groupByKey(Partitioner) - Method in class org.apache.spark.api.java.JavaPairRDD
Group the values for each key in the RDD into a single sequence.
groupByKey(int) - Method in class org.apache.spark.api.java.JavaPairRDD
Group the values for each key in the RDD into a single sequence.
groupByKey() - Method in class org.apache.spark.api.java.JavaPairRDD
Group the values for each key in the RDD into a single sequence.
groupByKey(Partitioner) - Method in class org.apache.spark.rdd.PairRDDFunctions
Group the values for each key in the RDD into a single sequence.
groupByKey(int) - Method in class org.apache.spark.rdd.PairRDDFunctions
Group the values for each key in the RDD into a single sequence.
groupByKey() - Method in class org.apache.spark.rdd.PairRDDFunctions
Group the values for each key in the RDD into a single sequence.
groupByKey(Function1<T, K>, Encoder<K>) - Method in class org.apache.spark.sql.Dataset
(Scala-specific) Returns a KeyValueGroupedDataset where the data is grouped by the given key func.
groupByKey(MapFunction<T, K>, Encoder<K>) - Method in class org.apache.spark.sql.Dataset
(Java-specific) Returns a KeyValueGroupedDataset where the data is grouped by the given key func.
groupByKey() - Method in class org.apache.spark.streaming.api.java.JavaPairDStream
Return a new DStream by applying groupByKey to each RDD.
groupByKey(int) - Method in class org.apache.spark.streaming.api.java.JavaPairDStream
Return a new DStream by applying groupByKey to each RDD.
groupByKey(Partitioner) - Method in class org.apache.spark.streaming.api.java.JavaPairDStream
Return a new DStream by applying groupByKey on each RDD of this DStream.
groupByKey() - Method in class org.apache.spark.streaming.dstream.PairDStreamFunctions
Return a new DStream by applying groupByKey to each RDD.
groupByKey(int) - Method in class org.apache.spark.streaming.dstream.PairDStreamFunctions
Return a new DStream by applying groupByKey to each RDD.
groupByKey(Partitioner) - Method in class org.apache.spark.streaming.dstream.PairDStreamFunctions
Return a new DStream by applying groupByKey on each RDD.
groupByKeyAndWindow(Duration) - Method in class org.apache.spark.streaming.api.java.JavaPairDStream
Return a new DStream by applying groupByKey over a sliding window.
groupByKeyAndWindow(Duration, Duration) - Method in class org.apache.spark.streaming.api.java.JavaPairDStream
Return a new DStream by applying groupByKey over a sliding window.
groupByKeyAndWindow(Duration, Duration, int) - Method in class org.apache.spark.streaming.api.java.JavaPairDStream
Return a new DStream by applying groupByKey over a sliding window on this DStream.
groupByKeyAndWindow(Duration, Duration, Partitioner) - Method in class org.apache.spark.streaming.api.java.JavaPairDStream
Return a new DStream by applying groupByKey over a sliding window on this DStream.
groupByKeyAndWindow(Duration) - Method in class org.apache.spark.streaming.dstream.PairDStreamFunctions
Return a new DStream by applying groupByKey over a sliding window.
groupByKeyAndWindow(Duration, Duration) - Method in class org.apache.spark.streaming.dstream.PairDStreamFunctions
Return a new DStream by applying groupByKey over a sliding window.
groupByKeyAndWindow(Duration, Duration, int) - Method in class org.apache.spark.streaming.dstream.PairDStreamFunctions
Return a new DStream by applying groupByKey over a sliding window on this DStream.
groupByKeyAndWindow(Duration, Duration, Partitioner) - Method in class org.apache.spark.streaming.dstream.PairDStreamFunctions
Create a new DStream by applying groupByKey over a sliding window on this DStream.
groupByPositionRangeError(int, int) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
groupByPositionRefersToAggregateFunctionError(int, Expression) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
GroupByType$() - Constructor for class org.apache.spark.sql.RelationalGroupedDataset.GroupByType$
 
groupEdges(Function2<ED, ED, ED>) - Method in class org.apache.spark.graphx.Graph
Merges multiple edges between two vertices into a single edge.
groupEdges(Function2<ED, ED, ED>) - Method in class org.apache.spark.graphx.impl.GraphImpl
 
groupHash() - Method in class org.apache.spark.rdd.DefaultPartitionCoalescer
 
grouping(Column) - Static method in class org.apache.spark.sql.functions
Aggregate function: indicates whether a specified column in a GROUP BY list is aggregated or not, returns 1 for aggregated or 0 for not aggregated in the result set.
grouping(String) - Static method in class org.apache.spark.sql.functions
Aggregate function: indicates whether a specified column in a GROUP BY list is aggregated or not, returns 1 for aggregated or 0 for not aggregated in the result set.
grouping_id(Seq<Column>) - Static method in class org.apache.spark.sql.functions
Aggregate function: returns the level of grouping, equals to
grouping_id(String, Seq<String>) - Static method in class org.apache.spark.sql.functions
Aggregate function: returns the level of grouping, equals to
groupingColInvalidError(Expression, Seq<Expression>) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
groupingIDMismatchError(GroupingID, Seq<Expression>) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
groupingMustWithGroupingSetsOrCubeOrRollupError() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
groupingSizeTooLargeError(int) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
 
GroupMappingServiceProvider - Interface in org.apache.spark.security
This Spark trait is used for mapping a given userName to a set of groups which it belongs to.
GroupState<S> - Interface in org.apache.spark.sql.streaming
:: Experimental ::
GroupStateTimeout - Class in org.apache.spark.sql.streaming
Represents the type of timeouts possible for the Dataset operations mapGroupsWithState and flatMapGroupsWithState.
GroupStateTimeout() - Constructor for class org.apache.spark.sql.streaming.GroupStateTimeout
 
groupWith(JavaPairRDD<K, W>) - Method in class org.apache.spark.api.java.JavaPairRDD
Alias for cogroup.
groupWith(JavaPairRDD<K, W1>, JavaPairRDD<K, W2>) - Method in class org.apache.spark.api.java.JavaPairRDD
Alias for cogroup.
groupWith(JavaPairRDD<K, W1>, JavaPairRDD<K, W2>, JavaPairRDD<K, W3>) - Method in class org.apache.spark.api.java.JavaPairRDD
Alias for cogroup.
groupWith(RDD<Tuple2<K, W>>) - Method in class org.apache.spark.rdd.PairRDDFunctions
Alias for cogroup.
groupWith(RDD<Tuple2<K, W1>>, RDD<Tuple2<K, W2>>) - Method in class org.apache.spark.rdd.PairRDDFunctions
Alias for cogroup.
groupWith(RDD<Tuple2<K, W1>>, RDD<Tuple2<K, W2>>, RDD<Tuple2<K, W3>>) - Method in class org.apache.spark.rdd.PairRDDFunctions
Alias for cogroup.
gt(double) - Static method in class org.apache.spark.ml.param.ParamValidators
Check if value is greater than lowerBound
gt(Object) - Method in class org.apache.spark.sql.Column
Greater than.
gt(T, T) - Static method in class org.apache.spark.sql.types.ByteExactNumeric
 
gt(T, T) - Static method in class org.apache.spark.sql.types.DecimalExactNumeric
 
gt(T, T) - Static method in class org.apache.spark.sql.types.DoubleExactNumeric
 
gt(T, T) - Static method in class org.apache.spark.sql.types.FloatExactNumeric
 
gt(T, T) - Static method in class org.apache.spark.sql.types.IntegerExactNumeric
 
gt(T, T) - Static method in class org.apache.spark.sql.types.LongExactNumeric
 
gt(T, T) - Static method in class org.apache.spark.sql.types.ShortExactNumeric
 
gtEq(double) - Static method in class org.apache.spark.ml.param.ParamValidators
Check if value is greater than or equal to lowerBound
gteq(T, T) - Static method in class org.apache.spark.sql.types.ByteExactNumeric
 
gteq(T, T) - Static method in class org.apache.spark.sql.types.DecimalExactNumeric
 
gteq(T, T) - Static method in class org.apache.spark.sql.types.DoubleExactNumeric
 
gteq(T, T) - Static method in class org.apache.spark.sql.types.FloatExactNumeric
 
gteq(T, T) - Static method in class org.apache.spark.sql.types.IntegerExactNumeric
 
gteq(T, T) - Static method in class org.apache.spark.sql.types.LongExactNumeric
 
gteq(T, T) - Static method in class org.apache.spark.sql.types.ShortExactNumeric
 
guard(Function0<Parsers.Parser<T>>) - Static method in class org.apache.spark.ml.feature.RFormulaParser
 

H

H2Dialect - Class in org.apache.spark.sql.jdbc
 
H2Dialect() - Constructor for class org.apache.spark.sql.jdbc.H2Dialect
 
H2Dialect.H2SQLBuilder - Class in org.apache.spark.sql.jdbc
 
H2SQLBuilder() - Constructor for class org.apache.spark.sql.jdbc.H2Dialect.H2SQLBuilder
 
HADOOP_PROPERTIES_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
 
hadoopConfiguration() - Method in class org.apache.spark.api.java.JavaSparkContext
Returns the Hadoop configuration used for the Hadoop code (e.g.
hadoopConfiguration() - Method in class org.apache.spark.SparkContext
A default Hadoop Configuration for the Hadoop code (e.g.
hadoopDelegationCreds() - Method in class org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.SparkAppConfig
 
HadoopDelegationTokenProvider - Interface in org.apache.spark.security
::DeveloperApi:: Hadoop delegation token provider.
hadoopFile(String, Class<F>, Class<K>, Class<V>, int) - Method in class org.apache.spark.api.java.JavaSparkContext
Get an RDD for a Hadoop file with an arbitrary InputFormat.
hadoopFile(String, Class<F>, Class<K>, Class<V>) - Method in class org.apache.spark.api.java.JavaSparkContext
Get an RDD for a Hadoop file with an arbitrary InputFormat
hadoopFile(String, Class<? extends InputFormat<K, V>>, Class<K>, Class<V>, int) - Method in class org.apache.spark.SparkContext
Get an RDD for a Hadoop file with an arbitrary InputFormat
hadoopFile(String, int, ClassTag<K>, ClassTag<V>, ClassTag<F>) - Method in class org.apache.spark.SparkContext
Smarter version of hadoopFile() that uses class tags to figure out the classes of keys, values and the InputFormat so that users don't need to pass them directly.
hadoopFile(String, ClassTag<K>, ClassTag<V>, ClassTag<F>) - Method in class org.apache.spark.SparkContext
Smarter version of hadoopFile() that uses class tags to figure out the classes of keys, values and the InputFormat so that users don't need to pass them directly.
HadoopFSUtils - Class in org.apache.spark.util
Utility functions to simplify and speed-up file listing.
HadoopFSUtils() - Constructor for class org.apache.spark.util.HadoopFSUtils
 
HadoopMapPartitionsWithSplitRDD$() - Constructor for class org.apache.spark.rdd.HadoopRDD.HadoopMapPartitionsWithSplitRDD$
 
hadoopProperties() - Method in class org.apache.spark.status.api.v1.ApplicationEnvironmentInfo
 
hadoopRDD(JobConf, Class<F>, Class<K>, Class<V>, int) - Method in class org.apache.spark.api.java.JavaSparkContext
Get an RDD for a Hadoop-readable dataset from a Hadoop JobConf giving its InputFormat and any other necessary info (e.g.
hadoopRDD(JobConf, Class<F>, Class<K>, Class<V>) - Method in class org.apache.spark.api.java.JavaSparkContext
Get an RDD for a Hadoop-readable dataset from a Hadoop JobConf giving its InputFormat and any other necessary info (e.g.
HadoopRDD<K,V> - Class in org.apache.spark.rdd
:: DeveloperApi :: An RDD that provides core functionality for reading data stored in Hadoop (e.g., files in HDFS, sources in HBase, or S3), using the older MapReduce API (org.apache.hadoop.mapred).
HadoopRDD(SparkContext, Broadcast<SerializableConfiguration>, Option<Function1<JobConf, BoxedUnit>>, Class<? extends InputFormat<K, V>>, Class<K>, Class<V>, int) - Constructor for class org.apache.spark.rdd.HadoopRDD
 
HadoopRDD(SparkContext, JobConf, Class<? extends InputFormat<K, V>>, Class<K>, Class<V>, int) - Constructor for class org.apache.spark.rdd.HadoopRDD
 
hadoopRDD(JobConf, Class<? extends InputFormat<K, V>>, Class<K>, Class<V>, int) - Method in class org.apache.spark.SparkContext
Get an RDD for a Hadoop-readable dataset from a Hadoop JobConf given its InputFormat and other necessary info (e.g.
HadoopRDD.HadoopMapPartitionsWithSplitRDD$ - Class in org.apache.spark.rdd
 
hammingLoss() - Method in class org.apache.spark.mllib.evaluation.MulticlassMetrics
 
hammingLoss() - Method in class org.apache.spark.mllib.evaluation.MultilabelMetrics
Returns Hamming-loss
handle(String, Request, HttpServletRequest, HttpServletResponse) - Method in class org.apache.spark.ui.ProxyRedirectHandler
 
handleInvalid() - Method in class org.apache.spark.ml.feature.Bucketizer
Param for how to handle invalid entries containing NaN values.
handleInvalid() - Method in class org.apache.spark.ml.feature.OneHotEncoder
 
handleInvalid() - Method in interface org.apache.spark.ml.feature.OneHotEncoderBase
Param for how to handle invalid data during transform().
handleInvalid() - Method in class org.apache.spark.ml.feature.OneHotEncoderModel
 
handleInvalid() - Method in class org.apache.spark.ml.feature.QuantileDiscretizer
 
handleInvalid() - Method in interface org.apache.spark.ml.feature.QuantileDiscretizerBase
Param for how to handle invalid entries.
handleInvalid() - Method in class org.apache.spark.ml.feature.RFormula
 
handleInvalid() - Method in interface org.apache.spark.ml.feature.RFormulaBase
Param for how to handle invalid data (unseen or NULL values) in features and label column of string type.
handleInvalid() - Method in class org.apache.spark.ml.feature.RFormulaModel
 
handleInvalid() - Method in class org.apache.spark.ml.feature.StringIndexer
 
handleInvalid() - Method in interface org.apache.spark.ml.feature.StringIndexerBase
Param for how to handle invalid data (unseen labels or NULL values).
handleInvalid() - Method in class org.apache.spark.ml.feature.StringIndexerModel
 
handleInvalid() - Method in class org.apache.spark.ml.feature.VectorAssembler
Param for how to handle invalid data (NULL values).
handleInvalid() - Method in class org.apache.spark.ml.feature.VectorIndexer
 
handleInvalid() - Method in class org.apache.spark.ml.feature.VectorIndexerModel
 
handleInvalid() - Method in interface org.apache.spark.ml.feature.VectorIndexerParams
Param for how to handle invalid data (unseen labels or NULL values).
handleInvalid() - Method in class org.apache.spark.ml.feature.VectorSizeHint
Param for how to handle invalid entries.
handleInvalid() - Method in interface org.apache.spark.ml.param.shared.HasHandleInvalid
Param for how to handle invalid entries.
HAS_METRICS_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
 
hasAccumulators(StageData) - Static method in class org.apache.spark.ui.jobs.ApiHelper
 
hasAddress() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution.Builder
optional string address = 1;
hasAddress() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution
optional string address = 1;
hasAddress() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDDataDistributionOrBuilder
optional string address = 1;
HasAggregationDepth - Interface in org.apache.spark.ml.param.shared
Trait for shared param aggregationDepth (default: 2).
hasAppSparkVersion() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo.Builder
optional string app_spark_version = 8;
hasAppSparkVersion() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo
optional string app_spark_version = 8;
hasAppSparkVersion() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfoOrBuilder
optional string app_spark_version = 8;
hasAttemptId() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo.Builder
optional string attempt_id = 1;
hasAttemptId() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo
optional string attempt_id = 1;
hasAttemptId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfoOrBuilder
optional string attempt_id = 1;
hasAttr(String) - Method in class org.apache.spark.ml.attribute.AttributeGroup
Test whether this attribute group contains a specific attribute.
hasBlockName() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo.Builder
optional string block_name = 1;
hasBlockName() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo
optional string block_name = 1;
hasBlockName() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfoOrBuilder
optional string block_name = 1;
HasBlockSize - Interface in org.apache.spark.ml.param.shared
Trait for shared param blockSize.
hasBytesSpilled(StageData) - Static method in class org.apache.spark.ui.jobs.ApiHelper
 
hasCachedSerializedBroadcast() - Method in class org.apache.spark.ShuffleStatus
 
hasCallsite() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode.Builder
optional string callsite = 5;
hasCallsite() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode
optional string callsite = 5;
hasCallsite() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationNodeOrBuilder
optional string callsite = 5;
HasCheckpointInterval - Interface in org.apache.spark.ml.param.shared
Trait for shared param checkpointInterval.
hasCluster() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper.Builder
.org.apache.spark.status.protobuf.SparkPlanGraphClusterWrapper cluster = 2;
hasCluster() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper
.org.apache.spark.status.protobuf.SparkPlanGraphClusterWrapper cluster = 2;
hasCluster() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapperOrBuilder
.org.apache.spark.status.protobuf.SparkPlanGraphClusterWrapper cluster = 2;
HasCollectSubModels - Interface in org.apache.spark.ml.param.shared
Trait for shared param collectSubModels (default: false).
hasCompletionTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
optional int64 completion_time = 5;
hasCompletionTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
optional int64 completion_time = 5;
hasCompletionTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataOrBuilder
optional int64 completion_time = 5;
hasCompletionTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
optional int64 completion_time = 9;
hasCompletionTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
optional int64 completion_time = 9;
hasCompletionTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
optional int64 completion_time = 9;
hasCompletionTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
optional int64 completion_time = 12;
hasCompletionTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
optional int64 completion_time = 12;
hasCompletionTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
optional int64 completion_time = 12;
hasCoresGranted() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
optional int32 cores_granted = 3;
hasCoresGranted() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo
optional int32 cores_granted = 3;
hasCoresGranted() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoOrBuilder
optional int32 cores_granted = 3;
hasCoresPerExecutor() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
optional int32 cores_per_executor = 5;
hasCoresPerExecutor() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo
optional int32 cores_per_executor = 5;
hasCoresPerExecutor() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoOrBuilder
optional int32 cores_per_executor = 5;
hasDefault(Param<T>) - Method in interface org.apache.spark.ml.param.Params
Tests whether the input param has a default value set.
hasDesc() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
optional string desc = 3;
hasDesc() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper
optional string desc = 3;
hasDesc() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapperOrBuilder
optional string desc = 3;
hasDesc() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
optional string desc = 3;
hasDesc() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode
optional string desc = 3;
hasDesc() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeOrBuilder
optional string desc = 3;
hasDescription() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
optional string description = 3;
hasDescription() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
optional string description = 3;
hasDescription() - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataOrBuilder
optional string description = 3;
hasDescription() - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress.Builder
optional string description = 1;
hasDescription() - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress
optional string description = 1;
hasDescription() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SinkProgressOrBuilder
optional string description = 1;
hasDescription() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
optional string description = 1;
hasDescription() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress
optional string description = 1;
hasDescription() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SourceProgressOrBuilder
optional string description = 1;
hasDescription() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
optional string description = 3;
hasDescription() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
optional string description = 3;
hasDescription() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
optional string description = 3;
hasDescription() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
optional string description = 40;
hasDescription() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
optional string description = 40;
hasDescription() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
optional string description = 40;
hasDetails() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
optional string details = 4;
hasDetails() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
optional string details = 4;
hasDetails() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
optional string details = 4;
hasDetails() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
optional string details = 41;
hasDetails() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
optional string details = 41;
hasDetails() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
optional string details = 41;
hasDiscoveryScript() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest.Builder
optional string discoveryScript = 3;
hasDiscoveryScript() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest
optional string discoveryScript = 3;
hasDiscoveryScript() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequestOrBuilder
optional string discoveryScript = 3;
HasDistanceMeasure - Interface in org.apache.spark.ml.param.shared
Trait for shared param distanceMeasure (default: "euclidean").
hasDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
optional int64 duration = 7;
hasDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
optional int64 duration = 7;
hasDuration() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataOrBuilder
optional int64 duration = 7;
HasElasticNetParam - Interface in org.apache.spark.ml.param.shared
Trait for shared param elasticNetParam.
hasEndOffset() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
optional string end_offset = 3;
hasEndOffset() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress
optional string end_offset = 3;
hasEndOffset() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SourceProgressOrBuilder
optional string end_offset = 3;
hasEndTimestamp() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData.Builder
optional int64 end_timestamp = 7;
hasEndTimestamp() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData
optional int64 end_timestamp = 7;
hasEndTimestamp() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryDataOrBuilder
optional int64 end_timestamp = 7;
hasErrorMessage() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
optional string error_message = 10;
hasErrorMessage() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
optional string error_message = 10;
hasErrorMessage() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
optional string error_message = 10;
hasErrorMessage() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
optional string error_message = 14;
hasErrorMessage() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
optional string error_message = 14;
hasErrorMessage() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataOrBuilder
optional string error_message = 14;
hasErrorMessage() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
optional string error_message = 14;
hasErrorMessage() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
optional string error_message = 14;
hasErrorMessage() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
optional string error_message = 14;
hasException() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData.Builder
optional string exception = 5;
hasException() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData
optional string exception = 5;
hasException() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryDataOrBuilder
optional string exception = 5;
hasExecutorId() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper.Builder
optional string executor_id = 3;
hasExecutorId() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper
optional string executor_id = 3;
hasExecutorId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapperOrBuilder
optional string executor_id = 3;
hasExecutorId() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
optional string executor_id = 2;
hasExecutorId() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData
optional string executor_id = 2;
hasExecutorId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamBlockDataOrBuilder
optional string executor_id = 2;
hasExecutorId() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
optional string executor_id = 8;
hasExecutorId() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
optional string executor_id = 8;
hasExecutorId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataOrBuilder
optional string executor_id = 8;
hasExecutorId() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
optional string executor_id = 8;
hasExecutorId() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
optional string executor_id = 8;
hasExecutorId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
optional string executor_id = 8;
hasExecutorMetricsDistributions() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
optional .org.apache.spark.status.protobuf.ExecutorMetricsDistributions executor_metrics_distributions = 52;
hasExecutorMetricsDistributions() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
optional .org.apache.spark.status.protobuf.ExecutorMetricsDistributions executor_metrics_distributions = 52;
hasExecutorMetricsDistributions() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
optional .org.apache.spark.status.protobuf.ExecutorMetricsDistributions executor_metrics_distributions = 52;
hasFailureReason() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
optional string failure_reason = 13;
hasFailureReason() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
optional string failure_reason = 13;
hasFailureReason() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
optional string failure_reason = 13;
HasFeaturesCol - Interface in org.apache.spark.ml.param.shared
Trait for shared param featuresCol (default: "features").
hasFirstTaskLaunchedTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
optional int64 first_task_launched_time = 11;
hasFirstTaskLaunchedTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
optional int64 first_task_launched_time = 11;
hasFirstTaskLaunchedTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
optional int64 first_task_launched_time = 11;