| org.apache.spark.sql.SQLContext.applySchema(JavaRDD<?>, Class<?>) | 
| org.apache.spark.sql.SQLContext.applySchema(JavaRDD<Row>, StructType) | 
| org.apache.spark.sql.SQLContext.applySchema(RDD<?>, Class<?>) | 
| org.apache.spark.sql.SQLContext.applySchema(RDD<Row>, StructType) | 
| org.apache.spark.sql.functions.approxCountDistinct(Column) | 
| org.apache.spark.sql.functions.approxCountDistinct(Column, double) | 
| org.apache.spark.sql.functions.approxCountDistinct(String) | 
| org.apache.spark.sql.functions.approxCountDistinct(String, double) | 
| org.apache.spark.sql.functions.bitwiseNOT(Column) | 
| org.apache.spark.status.api.v1.ExecutorSummary.blacklistedInStages() | 
| org.apache.spark.sql.connector.write.WriteBuilder.buildForBatch() | 
| org.apache.spark.sql.connector.write.WriteBuilder.buildForStreaming() | 
| org.apache.spark.sql.functions.callUDF(String, Seq<Column>) | 
| org.apache.spark.sql.SQLContext.clearActive() | 
| org.apache.spark.sql.jdbc.JdbcDialect.compileAggregate(AggregateFunc) | 
| org.apache.spark.ml.clustering.BisectingKMeansModel.computeCost(Dataset<?>) | 
| org.apache.spark.sql.SQLContext.createExternalTable(String, String) | 
| org.apache.spark.sql.catalog.Catalog.createExternalTable(String, String) | 
| org.apache.spark.sql.SQLContext.createExternalTable(String, String, Map<String, String>) | 
| org.apache.spark.sql.SQLContext.createExternalTable(String, String, Map<String, String>) | 
| org.apache.spark.sql.catalog.Catalog.createExternalTable(String, String, Map<String, String>) | 
| org.apache.spark.sql.catalog.Catalog.createExternalTable(String, String, Map<String, String>) | 
| org.apache.spark.sql.SQLContext.createExternalTable(String, String, String) | 
| org.apache.spark.sql.catalog.Catalog.createExternalTable(String, String, String) | 
| org.apache.spark.sql.SQLContext.createExternalTable(String, String, StructType, Map<String, String>) | 
| org.apache.spark.sql.SQLContext.createExternalTable(String, String, StructType, Map<String, String>) | 
| org.apache.spark.sql.catalog.Catalog.createExternalTable(String, String, StructType, Map<String, String>) | 
| org.apache.spark.sql.catalog.Catalog.createExternalTable(String, String, StructType, Map<String, String>) | 
| org.apache.spark.sql.connector.catalog.TableCatalog.createTable(Identifier, StructType, Transform[], Map<String, String>) | 
| org.apache.spark.sql.Dataset.explode(Seq<Column>, Function1<Row, TraversableOnce<A>>, TypeTags.TypeTag<A>) | 
| org.apache.spark.sql.Dataset.explode(String, String, Function1<A, TraversableOnce<B>>, TypeTags.TypeTag<B>) | 
| org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder.getAttributes() | 
| org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.getAttributes() | 
| org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder.getAttributes() | 
| org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgressOrBuilder.getCustomMetrics() | 
| org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.getCustomMetrics() | 
| org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder.getCustomMetrics() | 
| org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder.getDurationMs() | 
| org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.getDurationMs() | 
| org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder.getDurationMs() | 
| org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder.getEventTime() | 
| org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.getEventTime() | 
| org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder.getEventTime() | 
| org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder.getExecutorLogs() | 
| org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.getExecutorLogs() | 
| org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder.getExecutorLogs() | 
| org.apache.spark.status.protobuf.StoreTypes.TaskDataOrBuilder.getExecutorLogs() | 
| org.apache.spark.status.protobuf.StoreTypes.TaskData.getExecutorLogs() | 
| org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder.getExecutorLogs() | 
| org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfoOrBuilder.getExecutorResources() | 
| org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo.getExecutorResources() | 
| org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo.Builder.getExecutorResources() | 
| org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder.getExecutorSummary() | 
| org.apache.spark.status.protobuf.StoreTypes.StageData.getExecutorSummary() | 
| org.apache.spark.status.protobuf.StoreTypes.StageData.Builder.getExecutorSummary() | 
| org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder.getJobs() | 
| org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.getJobs() | 
| org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder.getJobs() | 
| org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder.getJobsValue() | 
| org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.getJobsValue() | 
| org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder.getJobsValue() | 
| org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder.getKilledTasksSummary() | 
| org.apache.spark.status.protobuf.StoreTypes.StageData.getKilledTasksSummary() | 
| org.apache.spark.status.protobuf.StoreTypes.StageData.Builder.getKilledTasksSummary() | 
| org.apache.spark.status.protobuf.StoreTypes.JobDataOrBuilder.getKillTasksSummary() | 
| org.apache.spark.status.protobuf.StoreTypes.JobData.getKillTasksSummary() | 
| org.apache.spark.status.protobuf.StoreTypes.JobData.Builder.getKillTasksSummary() | 
| org.apache.spark.status.protobuf.StoreTypes.StageDataWrapperOrBuilder.getLocality() | 
| org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper.getLocality() | 
| org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper.Builder.getLocality() | 
| org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsOrBuilder.getMetrics() | 
| org.apache.spark.status.protobuf.StoreTypes.ExecutorMetrics.getMetrics() | 
| org.apache.spark.status.protobuf.StoreTypes.ExecutorMetrics.Builder.getMetrics() | 
| org.apache.spark.status.protobuf.StoreTypes.SourceProgressOrBuilder.getMetrics() | 
| org.apache.spark.status.protobuf.StoreTypes.SourceProgress.getMetrics() | 
| org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder.getMetrics() | 
| org.apache.spark.status.protobuf.StoreTypes.SinkProgressOrBuilder.getMetrics() | 
| org.apache.spark.status.protobuf.StoreTypes.SinkProgress.getMetrics() | 
| org.apache.spark.status.protobuf.StoreTypes.SinkProgress.Builder.getMetrics() | 
| org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder.getMetricValues() | 
| org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.getMetricValues() | 
| org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder.getMetricValues() | 
| org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder.getModifiedConfigs() | 
| org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.getModifiedConfigs() | 
| org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder.getModifiedConfigs() | 
| org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder.getMutableAttributes() | 
| org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder.getMutableCustomMetrics() | 
| org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder.getMutableDurationMs() | 
| org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder.getMutableEventTime() | 
| org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder.getMutableExecutorLogs() | 
| org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder.getMutableExecutorLogs() | 
| org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo.Builder.getMutableExecutorResources() | 
| org.apache.spark.status.protobuf.StoreTypes.StageData.Builder.getMutableExecutorSummary() | 
| org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder.getMutableJobs() | 
| org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder.getMutableJobsValue() | 
| org.apache.spark.status.protobuf.StoreTypes.StageData.Builder.getMutableKilledTasksSummary() | 
| org.apache.spark.status.protobuf.StoreTypes.JobData.Builder.getMutableKillTasksSummary() | 
| org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper.Builder.getMutableLocality() | 
| org.apache.spark.status.protobuf.StoreTypes.ExecutorMetrics.Builder.getMutableMetrics() | 
| org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder.getMutableMetrics() | 
| org.apache.spark.status.protobuf.StoreTypes.SinkProgress.Builder.getMutableMetrics() | 
| org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder.getMutableMetricValues() | 
| org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder.getMutableModifiedConfigs() | 
| org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder.getMutableObservedMetrics() | 
| org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder.getMutableProcessLogs() | 
| org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder.getMutableResources() | 
| org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo.Builder.getMutableTaskResources() | 
| org.apache.spark.status.protobuf.StoreTypes.StageData.Builder.getMutableTasks() | 
| org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder.getObservedMetrics() | 
| org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.getObservedMetrics() | 
| org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder.getObservedMetrics() | 
| org.apache.spark.sql.SQLContext.getOrCreate(SparkContext) | 
| org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryOrBuilder.getProcessLogs() | 
| org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.getProcessLogs() | 
| org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder.getProcessLogs() | 
| org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder.getResources() | 
| org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.getResources() | 
| org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder.getResources() | 
| org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfoOrBuilder.getTaskResources() | 
| org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo.getTaskResources() | 
| org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo.Builder.getTaskResources() | 
| org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder.getTasks() | 
| org.apache.spark.status.protobuf.StoreTypes.StageData.getTasks() | 
| org.apache.spark.status.protobuf.StoreTypes.StageData.Builder.getTasks() | 
| org.apache.spark.status.api.v1.ExecutorSummary.isBlacklisted() | 
| org.apache.spark.status.api.v1.ExecutorStageSummary.isBlacklistedForStage() | 
| org.apache.spark.sql.SQLContext.jdbc(String, String) | 
| org.apache.spark.sql.SQLContext.jdbc(String, String, String[]) | 
| org.apache.spark.sql.SQLContext.jdbc(String, String, String, long, long, int) | 
| org.apache.spark.sql.DataFrameReader.json(JavaRDD<String>) | 
| org.apache.spark.sql.DataFrameReader.json(RDD<String>) | 
| org.apache.spark.sql.SQLContext.jsonFile(String) | 
| org.apache.spark.sql.SQLContext.jsonFile(String, double) | 
| org.apache.spark.sql.SQLContext.jsonFile(String, StructType) | 
| org.apache.spark.sql.SQLContext.jsonRDD(JavaRDD<String>) | 
| org.apache.spark.sql.SQLContext.jsonRDD(JavaRDD<String>, double) | 
| org.apache.spark.sql.SQLContext.jsonRDD(JavaRDD<String>, StructType) | 
| org.apache.spark.sql.SQLContext.jsonRDD(RDD<String>) | 
| org.apache.spark.sql.SQLContext.jsonRDD(RDD<String>, double) | 
| org.apache.spark.sql.SQLContext.jsonRDD(RDD<String>, StructType) | 
| org.apache.spark.ml.feature.StringIndexerModel.labels() | 
| org.apache.spark.sql.SQLContext.load(String) | 
| org.apache.spark.sql.SQLContext.load(String, Map<String, String>) | 
| org.apache.spark.sql.SQLContext.load(String, Map<String, String>) | 
| org.apache.spark.sql.SQLContext.load(String, String) | 
| org.apache.spark.sql.SQLContext.load(String, StructType, Map<String, String>) | 
| org.apache.spark.sql.SQLContext.load(String, StructType, Map<String, String>) | 
| org.apache.spark.sql.functions.monotonicallyIncreasingId() | 
| org.apache.spark.sql.SQLImplicits.newBooleanSeqEncoder() | 
| org.apache.spark.sql.SQLImplicits.newByteSeqEncoder() | 
| org.apache.spark.sql.SQLImplicits.newDoubleSeqEncoder() | 
| org.apache.spark.sql.SQLImplicits.newFloatSeqEncoder() | 
| org.apache.spark.sql.SQLImplicits.newIntSeqEncoder() | 
| org.apache.spark.sql.SQLImplicits.newLongSeqEncoder() | 
| org.apache.spark.sql.SQLImplicits.newProductSeqEncoder(TypeTags.TypeTag<A>) | 
| org.apache.spark.sql.SQLImplicits.newShortSeqEncoder() | 
| org.apache.spark.sql.SQLImplicits.newStringSeqEncoder() | 
| org.apache.spark.sql.streaming.Trigger.Once() | 
| org.apache.spark.scheduler.SparkListenerInterface.onExecutorBlacklisted(SparkListenerExecutorBlacklisted) | 
| org.apache.spark.scheduler.SparkListenerInterface.onExecutorBlacklistedForStage(SparkListenerExecutorBlacklistedForStage) | 
| org.apache.spark.scheduler.SparkListenerInterface.onExecutorUnblacklisted(SparkListenerExecutorUnblacklisted) | 
| org.apache.spark.scheduler.SparkListenerInterface.onNodeBlacklisted(SparkListenerNodeBlacklisted) | 
| org.apache.spark.scheduler.SparkListenerInterface.onNodeBlacklistedForStage(SparkListenerNodeBlacklistedForStage) | 
| org.apache.spark.scheduler.SparkListenerInterface.onNodeUnblacklisted(SparkListenerNodeUnblacklisted) | 
| org.apache.spark.sql.SQLContext.parquetFile(Seq<String>) | 
| org.apache.spark.sql.SQLContext.parquetFile(String...) | 
| org.apache.spark.sql.UDFRegistration.register(String, UserDefinedAggregateFunction) | 
| org.apache.spark.sql.Dataset.registerTempTable(String) | 
| org.apache.spark.sql.connector.catalog.Table.schema() | 
| org.apache.spark.mllib.evaluation.BinaryClassificationMetrics.scoreLabelsWeight() | 
| org.apache.spark.sql.SQLContext.setActive(SQLContext) | 
| org.apache.spark.sql.functions.shiftLeft(Column, int) | 
| org.apache.spark.sql.functions.shiftRight(Column, int) | 
| org.apache.spark.sql.functions.shiftRightUnsigned(Column, int) | 
| org.apache.spark.sql.connector.catalog.StagingTableCatalog.stageCreate(Identifier, StructType, Transform[], Map<String, String>) | 
| org.apache.spark.sql.functions.sumDistinct(Column) | 
| org.apache.spark.sql.functions.sumDistinct(String) | 
| org.apache.spark.sql.functions.toDegrees(Column) | 
| org.apache.spark.sql.functions.toDegrees(String) | 
| org.apache.spark.sql.functions.toRadians(Column) | 
| org.apache.spark.sql.functions.toRadians(String) | 
| org.apache.spark.sql.functions.udf(Object, DataType) | 
| org.apache.spark.status.protobuf.StoreTypes.JobExecutionStatus.valueOf(int) | 
| org.apache.spark.status.protobuf.StoreTypes.DeterministicLevel.valueOf(int) | 
| org.apache.spark.status.protobuf.StoreTypes.StageStatus.valueOf(int) | 
| org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper.WrapperCase.valueOf(int) |