org.apache.hadoop.hive.ql.exec.vector.VectorizationContext vContext
org.apache.hadoop.hive.ql.exec.vector.VectorizationContext vContext
org.apache.hadoop.hive.ql.exec.vector.expressions.VectorExpression conditionEvaluator
org.apache.hadoop.hive.ql.plan.VectorGroupByDesc vectorDesc
org.apache.hadoop.hive.ql.exec.vector.expressions.aggregates.VectorAggregateExpression[] aggregators
org.apache.hadoop.hive.ql.exec.vector.expressions.VectorExpression[] keyExpressions
int outputKeyLength
boolean isVectorOutput
org.apache.hadoop.hive.ql.exec.vector.VectorizationContext vOutContext
float memoryThreshold
org.apache.hadoop.hive.ql.exec.vector.VectorizationContext vOutContext
org.apache.hadoop.hive.ql.exec.vector.expressions.VectorExpression[] keyExpressions
org.apache.hadoop.hive.ql.exec.vector.expressions.VectorExpression[] bigTableFilterExpressions
org.apache.hadoop.hive.ql.exec.vector.expressions.VectorExpression[] bigTableValueExpressions
org.apache.hadoop.hive.ql.exec.vector.expressions.VectorExpressionWriter[] rowWriters
org.apache.hadoop.hive.ql.exec.vector.VectorizationContext vContext
org.apache.hadoop.hive.serde2.Deserializer currentPartDeserializer
org.apache.hadoop.hive.serde2.objectinspector.StructObjectInspector currentPartRawRowObjectInspector
org.apache.hadoop.hive.ql.exec.vector.VectorAssignRow currentVectorAssign
org.apache.hadoop.hive.ql.exec.vector.VectorizationContext vContext
org.apache.hadoop.hive.ql.plan.VectorSelectDesc vectorDesc
org.apache.hadoop.hive.ql.exec.vector.expressions.VectorExpression[] vExpressions
int[] projectedOutputColumns
org.apache.hadoop.hive.ql.exec.vector.VectorizationContext vOutContext
org.apache.hadoop.hive.ql.exec.vector.expressions.VectorExpression[] bigTableValueExpressions
org.apache.hadoop.hive.ql.exec.vector.expressions.VectorExpression[] bigTableFilterExpressions
org.apache.hadoop.hive.ql.exec.vector.expressions.VectorExpression[] keyExpressions
org.apache.hadoop.hive.ql.exec.vector.expressions.VectorExpressionWriter[] keyOutputWriters
org.apache.hadoop.hive.ql.exec.vector.VectorizationContext vOutContext
org.apache.hadoop.hive.ql.exec.vector.VectorizationContext vContext
org.apache.hadoop.hive.ql.exec.vector.VectorizationContext vContext
FileScanTask[] tasks
DeleteFile[] deletes
final java.lang.Object writeReplace()
PartitionSpec spec
SortOrder sortOrder
BaseTable table
java.lang.String name
StructLike groupingKey
java.lang.Object[] tasks
java.lang.Object writeReplace()
TableOperations ops
java.lang.String name
MetricsReporter reporter
java.lang.Object writeReplace()
int[] fromProjectionPos
InputFile file
java.lang.String manifestPath
java.lang.Long length
int specId
ManifestContent content
long sequenceNumber
long minSequenceNumber
java.lang.Long snapshotId
java.lang.Integer addedFilesCount
java.lang.Integer existingFilesCount
java.lang.Integer deletedFilesCount
java.lang.Long addedRowsCount
java.lang.Long existingRowsCount
java.lang.Long deletedRowsCount
ManifestFile.PartitionFieldSummary[] partitions
byte[] keyMetadata
int[] fromProjectionPos
boolean containsNull
java.lang.Boolean containsNaN
byte[] lowerBound
byte[] upperBound
java.lang.String partition
UnboundPartitionSpec spec
Schema schema
int lastColumnId
Snapshot snapshot
UnboundSortOrder sortOrder
java.lang.String uuid
java.util.Set<E> removed
long snapshotId
java.lang.String refName
long snapshotId
int schemaId
int specId
int sortOrderId
java.lang.String location
java.util.Map<K,V> updated
java.lang.String refName
java.lang.Long snapshotId
org.apache.iceberg.SnapshotRefType type
java.lang.Integer minSnapshotsToKeep
java.lang.Long maxSnapshotAgeMs
java.lang.Long maxRefAgeMs
long snapshotId
StatisticsFile statisticsFile
int formatVersion
private void readObject(java.io.ObjectInputStream in) throws java.io.IOException, java.lang.ClassNotFoundException
java.io.IOException
- On serialization errorjava.lang.ClassNotFoundException
- If the class is not foundprivate void writeObject(java.io.ObjectOutputStream out) throws java.io.IOException
java.io.IOException
- On serialization errorjava.lang.Long rowCount
java.util.Map<K,V> columnSizes
java.util.Map<K,V> valueCounts
java.util.Map<K,V> nullValueCounts
java.util.Map<K,V> nanValueCounts
java.util.Map<K,V> lowerBounds
java.util.Map<K,V> upperBounds
java.util.Map<K,V> columnModes
MetricsModes.MetricsMode defaultMode
int length
Types.StructType partitionType
int size
java.lang.Object[] data
java.lang.String stringSchema
PartitionSpec spec
int size
java.lang.Object[] partitionTuple
SerializableFunction<S,T>[] transforms
Accessor<T>[] accessors
Schema schema
int specId
PartitionField[] fields
int lastAssignedFieldId
Schema schema
Schema schema
int defaultSpecId
java.util.Map<K,V> specs
Types.StructType struct
int schemaId
int[] identifierFieldIds
int highestFieldId
java.lang.String name
java.lang.String location
java.lang.String metadataFileLocation
java.util.Map<K,V> properties
java.lang.String schemaAsJson
int defaultSpecId
java.util.Map<K,V> specAsJsonMap
java.lang.String sortOrderAsJson
FileIO io
EncryptionManager encryption
LocationProvider locationProvider
java.util.Map<K,V> refs
MetadataTableType type
java.lang.String baseTableName
long snapshotId
org.apache.iceberg.SnapshotRefType type
java.lang.Integer minSnapshotsToKeep
java.lang.Long maxSnapshotAgeMs
java.lang.Long maxRefAgeMs
Transform<S,T> transform
int sourceId
SortDirection direction
NullOrder nullOrder
java.lang.String metadataFileLocation
int formatVersion
java.lang.String uuid
java.lang.String location
long lastSequenceNumber
long lastUpdatedMillis
int lastColumnId
int currentSchemaId
java.util.List<E> schemas
int defaultSpecId
java.util.List<E> specs
int lastAssignedPartitionId
int defaultSortOrderId
java.util.List<E> sortOrders
java.util.Map<K,V> properties
long currentSnapshotId
java.util.Map<K,V> schemasById
java.util.Map<K,V> specsById
java.util.Map<K,V> sortOrdersById
java.util.List<E> snapshotLog
java.util.List<E> previousFiles
java.util.List<E> statisticsFiles
java.util.List<E> changes
SerializableSupplier<T> snapshotsSupplier
java.util.List<E> snapshots
java.util.Map<K,V> snapshotsById
java.util.Map<K,V> refs
boolean snapshotsLoaded
long timestampMillis
long snapshotId
int minInputFiles
int deleteFileThreshold
long minFileSize
long maxFileSize
long targetFileSize
long maxGroupSize
boolean rewriteAll
SortOrder sortOrder
java.lang.String ossEndpoint
java.lang.String accessKeyId
java.lang.String accessKeySecret
java.lang.String ossStagingDirectory
SerializableSupplier<T> oss
AliyunProperties aliyunProperties
MetricsContext metrics
java.util.concurrent.atomic.AtomicBoolean isResourceClosed
AwsProperties awsProperties
java.lang.String roleSessionName
java.lang.String httpClientType
java.util.Map<K,V> httpClientProperties
java.util.Set<E> stsClientAssumeRoleTags
java.lang.String clientAssumeRoleArn
java.lang.String clientAssumeRoleExternalId
int clientAssumeRoleTimeoutSec
java.lang.String clientAssumeRoleRegion
java.lang.String clientAssumeRoleSessionName
java.lang.String clientRegion
java.lang.String clientCredentialsProvider
java.util.Map<K,V> clientCredentialsProviderProperties
java.lang.String s3FileIoSseType
java.lang.String s3FileIoSseKey
java.lang.String s3FileIoSseMd5
java.lang.String s3AccessKeyId
java.lang.String s3SecretAccessKey
java.lang.String s3SessionToken
int s3FileIoMultipartUploadThreads
int s3FileIoMultiPartSize
int s3FileIoDeleteBatchSize
double s3FileIoMultipartThresholdFactor
java.lang.String s3fileIoStagingDirectory
software.amazon.awssdk.services.s3.model.ObjectCannedACL s3FileIoAcl
boolean isS3ChecksumEnabled
java.util.Set<E> s3WriteTags
boolean s3WriteTableTagEnabled
boolean s3WriteNamespaceTagEnabled
java.util.Set<E> s3DeleteTags
int s3FileIoDeleteThreads
boolean isS3DeleteEnabled
java.util.Map<K,V> s3BucketToAccessPointMapping
boolean s3PreloadClientEnabled
boolean s3DualStackEnabled
boolean s3PathStyleAccess
boolean s3UseArnRegionEnabled
boolean s3AccelerationEnabled
java.lang.String s3Endpoint
java.lang.String glueEndpoint
java.lang.String glueCatalogId
boolean glueCatalogSkipArchive
boolean glueCatalogSkipNameValidation
boolean glueLakeFormationEnabled
java.lang.String dynamoDbTableName
java.lang.String dynamoDbEndpoint
boolean s3RemoteSigningEnabled
java.util.Map<K,V> allProperties
java.lang.String restSigningRegion
java.lang.String restSigningName
java.lang.String restAccessKeyId
java.lang.String restSecretAccessKey
java.lang.String restSessionToken
java.lang.String dbName
java.lang.String tableName
java.lang.String glueCatalogId
java.lang.String glueAccountId
java.lang.String credential
SerializableSupplier<T> s3
AwsProperties awsProperties
SerializableMap<K,V> properties
MetricsContext metrics
java.util.concurrent.atomic.AtomicBoolean isResourceClosed
java.lang.String ecsS3Endpoint
java.lang.String ecsS3AccessKeyId
java.lang.String ecsS3SecretAccessKey
SerializableSupplier<T> s3
DellProperties dellProperties
DellClientFactory dellClientFactory
java.util.concurrent.atomic.AtomicBoolean isResourceClosed
MetricsContext metrics
Expression.Operation op
Term term
Expression left
Expression right
Types.NestedField field
Accessor<T> accessor
java.lang.String name
java.util.Set<E> literalSet
BoundReference<T> ref
Transform<S,T> transform
SerializableFunction<S,T> func
int fieldId
Types.NestedField field
Expression expr
java.lang.Object writeReplace() throws java.io.ObjectStreamException
java.io.ObjectStreamException
int fieldId
Type.PrimitiveType type
java.util.Comparator<T> comparator
int fieldId
Type.PrimitiveType type
java.util.Comparator<T> comparator
java.lang.String name
Expression child
Expression left
Expression right
Expression.Operation op
Term term
PartitionSpec spec
Expression expr
boolean caseSensitive
java.lang.Object writeReplace() throws java.io.ObjectStreamException
java.io.ObjectStreamException
java.util.List<E> literals
NamedReference<T> ref
Transform<S,T> transform
NamedReference<T>[] refs
SerializableConfiguration hadoopConf
java.util.Map<K,V> properties
java.lang.String name
java.lang.String impl
java.lang.String catalogName
SerializableConfiguration hadoopConf
java.lang.String warehouseLocation
java.util.Map<K,V> properties
java.lang.String catalogName
SerializableConfiguration hadoopConf
java.lang.String uri
java.lang.String warehouse
int clientPoolSize
java.util.Map<K,V> properties
java.lang.String catalogName
SerializableConfiguration hadoopConf
java.util.Map<K,V> properties
CatalogLoader catalogLoader
java.lang.String identifier
java.lang.String location
SerializableConfiguration hadoopConf
org.apache.flink.formats.avro.AvroToRowDataConverters.AvroToRowDataConverter converter
Schema schema
org.apache.flink.table.types.logical.RowType flinkSchema
java.util.Map<K,V> props
PartitionSpec spec
int[] equalityFieldIds
Schema eqDeleteRowSchema
Schema posDeleteRowSchema
Table table
org.apache.flink.table.types.logical.RowType eqDeleteFlinkSchema
org.apache.flink.table.types.logical.RowType posDeleteFlinkSchema
Table table
Schema schema
org.apache.flink.table.types.logical.RowType flinkSchema
PartitionSpec spec
FileIO io
long targetFileSizeBytes
FileFormat format
java.util.List<E> equalityFieldIds
boolean upsert
FileAppenderFactory<T> appenderFactory
RowDataFileScanTaskReader rowDataReader
RowDataToAvroGenericRecordConverter converter
Schema readSchema
TableLoader tableLoader
FileIO io
EncryptionManager encryption
ScanContext context
FileScanTaskReader<T> rowDataReader
CombinedScanTask task
TableLoader tableLoader
ScanContext scanContext
ReaderFunction<T> readerFunction
SplitAssignerFactory assignerFactory
TaskWriter<T> writer
int subTaskId
int attemptId
Schema schema
java.lang.String nameMapping
FileIO io
boolean caseSensitive
EncryptionManager encryptionManager
TaskWriterFactory<T> taskWriterFactory
RowDataFileScanTaskReader rowDataReader
boolean caseSensitive
boolean exposeLocality
java.lang.Long snapshotId
java.lang.String branch
java.lang.String tag
StreamingStartingStrategy startingStrategy
java.lang.Long startSnapshotId
java.lang.Long startSnapshotTimestamp
java.lang.Long endSnapshotId
java.lang.Long asOfTimestamp
java.lang.String startTag
java.lang.String endTag
java.lang.Long splitSize
java.lang.Integer splitLookback
java.lang.Long splitOpenFileCost
boolean isStreaming
java.time.Duration monitorInterval
java.lang.String nameMapping
Schema schema
java.util.List<E> filters
long limit
boolean includeColumnStats
java.lang.Integer planParallelism
int maxPlanningSnapshotCount
TableLoader tableLoader
ScanContext scanContext
boolean isRunning
long lastSnapshotId
org.apache.flink.api.common.operators.MailboxExecutor executor
FlinkInputFormat format
org.apache.iceberg.flink.source.enumerator.IcebergEnumeratorPosition lastEnumeratedPosition
java.util.Collection<E> pendingSplits
int[] enumerationSplitCountHistory
java.lang.String tableName
Schema readSchema
FileIO io
EncryptionManager encryption
RowDataFileScanTaskReader rowDataReader
DataIteratorBatcher<T> batcher
Schema readSchema
FileIO io
EncryptionManager encryption
Schema tableSchema
Schema readSchema
java.lang.String nameMapping
boolean caseSensitive
FileIO io
EncryptionManager encryption
CombinedScanTask task
int fileOffset
long recordOffset
java.util.Collection<E> finishedSplitIds
java.lang.String requesterHostname
java.lang.String projectId
java.lang.String clientLibToken
java.lang.String serviceHost
java.lang.String gcsDecryptionKey
java.lang.String gcsEncryptionKey
java.lang.String gcsUserProject
java.lang.Integer gcsChannelReadChunkSize
java.lang.Integer gcsChannelWriteChunkSize
SerializableSupplier<T> storageSupplier
GCPProperties gcpProperties
MetricsContext metrics
java.util.concurrent.atomic.AtomicBoolean isResourceClosed
SerializableMap<K,V> properties
SerializableSupplier<T> hadoopConf
SerializableMap<K,V> properties
java.lang.String scheme
private void readObject(java.io.ObjectInputStream in) throws java.lang.ClassNotFoundException, java.io.IOException
java.lang.ClassNotFoundException
java.io.IOException
private void writeObject(java.io.ObjectOutputStream out) throws java.io.IOException
java.io.IOException
int numberFailedObjects
java.util.Map<K,V> ioInstances
SerializableMap<K,V> properties
SerializableSupplier<T> hadoopConf
DataFile[] dataFiles
DeleteFile[] deleteFiles
java.lang.CharSequence[] referencedDataFiles
java.util.Set<E> names
java.lang.Integer id
MappedFields nestedMapping
java.util.List<E> fields
MappedFields mapping
java.util.Map<K,V> values
java.lang.String uri
java.lang.String format
java.util.Set<E> set
Table table
org.apache.spark.sql.SparkSession spark
SparkTableCache tableCache
FileScanTaskSetManager manager
FileRewriteCoordinator rewriteCoordinator
Table table
org.apache.spark.sql.SparkSession spark
SparkTableCache tableCache
FileScanTaskSetManager manager
FileRewriteCoordinator rewriteCoordinator
double sizeEstimateMultiple
java.util.List<E> zOrderColNames
int maxOutputSize
int varLengthContribution
org.apache.spark.sql.types.DataType sqlType
int precision
int scale
org.apache.spark.sql.types.DataType sqlType
org.apache.spark.sql.types.DataType sqlType
int precision
int scale
DataFile[] dataFiles
DeleteFile[] deleteFiles
java.lang.CharSequence[] referencedDataFiles
DataFile[] taskFiles
java.lang.Object writeReplace() throws java.io.ObjectStreamException
java.io.ObjectStreamException
java.lang.Object writeReplace() throws java.io.ObjectStreamException
java.io.ObjectStreamException
java.lang.Object writeReplace() throws java.io.ObjectStreamException
java.io.ObjectStreamException
java.lang.String transform
java.lang.Object writeReplace() throws java.io.ObjectStreamException
java.io.ObjectStreamException
int scale
int precision
int length
Types.NestedField elementField
Types.NestedField keyField
Types.NestedField valueField
boolean isOptional
int id
java.lang.String name
Type type
java.lang.String doc
Types.NestedField[] fields
boolean adjustToUTC
java.util.Set<E> wrapperSet
java.lang.CharSequence wrapped
org.apache.avro.Schema schema
java.lang.Object first
java.lang.Object second
java.util.Map<K,V> copiedMap
scala.collection.Seq<A> catalogAndNamespace
int nestedViewDepth
int maxNestedViewDepth
scala.collection.mutable.Map<K,V> relationCache
scala.collection.Seq<A> referredTempViewNames
scala.collection.mutable.Set<A> referredTempFunctionNames
scala.Option<A> outerPlan
private java.lang.Object readResolve()
java.lang.ThreadLocal<T> value
org.apache.spark.sql.connector.catalog.CatalogManager catalogManager
org.apache.spark.sql.connector.catalog.LookupCatalog.CatalogAndMultipartIdentifier$ CatalogAndMultipartIdentifier$module
org.apache.spark.sql.connector.catalog.LookupCatalog.SessionCatalogAndIdentifier$ SessionCatalogAndIdentifier$module
org.apache.spark.sql.connector.catalog.LookupCatalog.NonSessionCatalogAndIdentifier$ NonSessionCatalogAndIdentifier$module
org.apache.spark.sql.connector.catalog.LookupCatalog.CatalogAndNamespace$ CatalogAndNamespace$module
org.apache.spark.sql.connector.catalog.LookupCatalog.CatalogAndIdentifier$ CatalogAndIdentifier$module
org.apache.spark.sql.connector.catalog.LookupCatalog.AsTableIdentifier$ AsTableIdentifier$module
long timestamp
private java.lang.Object readResolve()
java.lang.String version
private java.lang.Object readResolve()
private java.lang.Object readResolve()
boolean resolved
int ordinal
org.apache.spark.sql.types.DataType dataType
boolean bitmap$0
private java.lang.Object readResolve()
boolean resolved
java.lang.String viewName
java.lang.String colName
int ordinal
int expectedNumCandidates
scala.Option<A> viewDDL
boolean bitmap$0
private java.lang.Object readResolve()
private java.lang.Object readResolve()
boolean resolved
org.apache.spark.sql.catalyst.expressions.Expression child
scala.collection.Seq<A> names
scala.collection.Seq<A> nodePatterns
boolean bitmap$0
private java.lang.Object readResolve()
java.lang.String message
scala.Option<A> cause
private java.lang.Object readResolve()
java.lang.String db
private java.lang.Object readResolve()
java.lang.String message
scala.Option<A> cause
private java.lang.Object readResolve()
private java.lang.Object readResolve()
java.lang.String message
scala.Option<A> cause
private java.lang.Object readResolve()
private java.lang.Object readResolve()
private java.lang.Object readResolve()
java.lang.String db
java.lang.String func
private java.lang.Object readResolve()
java.lang.String message
scala.Option<A> cause
private java.lang.Object readResolve()
java.lang.String func
private java.lang.Object readResolve()
org.apache.spark.sql.catalyst.plans.logical.LogicalPlan plan
int _hashCode
private java.lang.Object readResolve()
boolean resolved
org.apache.spark.sql.catalyst.plans.logical.LogicalPlan relation
scala.Option<A> timestamp
scala.Option<A> version
scala.collection.Seq<A> nodePatterns
boolean bitmap$0
private java.lang.Object readResolve()
org.apache.spark.sql.connector.catalog.CatalogPlugin catalog
scala.collection.Seq<A> nameParts
private java.lang.Object readResolve()
scala.collection.Seq<A> path
org.apache.spark.sql.types.StructField field
private java.lang.Object readResolve()
org.apache.spark.sql.connector.catalog.TableChange.ColumnPosition position
private java.lang.Object readResolve()
org.apache.spark.sql.connector.catalog.CatalogPlugin catalog
scala.collection.Seq<A> namespace
private java.lang.Object readResolve()
java.lang.String name
org.apache.spark.sql.connector.catalog.functions.UnboundFunction func
private java.lang.Object readResolve()
scala.collection.Seq<A> names
org.apache.spark.sql.catalyst.InternalRow ident
scala.Option<A> location
private java.lang.Object readResolve()
org.apache.spark.sql.connector.catalog.FunctionCatalog catalog
org.apache.spark.sql.connector.catalog.Identifier identifier
org.apache.spark.sql.connector.catalog.functions.UnboundFunction func
private java.lang.Object readResolve()
scala.collection.Seq<A> expressions
private java.lang.Object readResolve()
org.apache.spark.sql.connector.catalog.TableCatalog catalog
org.apache.spark.sql.connector.catalog.Identifier identifier
org.apache.spark.sql.connector.catalog.Table table
scala.collection.Seq<A> outputAttributes
private java.lang.Object readResolve()
org.apache.spark.sql.connector.catalog.Identifier identifier
boolean isTemp
private java.lang.Object readResolve()
org.apache.spark.sql.catalyst.analysis.Analyzer analyzer
org.apache.spark.sql.SparkSession spark
boolean bitmap$0
private java.lang.Object readResolve()
org.apache.spark.sql.connector.catalog.CatalogManager catalogManager
org.apache.spark.sql.SparkSession spark
org.apache.spark.sql.connector.catalog.LookupCatalog.CatalogAndMultipartIdentifier$ CatalogAndMultipartIdentifier$module
org.apache.spark.sql.connector.catalog.LookupCatalog.SessionCatalogAndIdentifier$ SessionCatalogAndIdentifier$module
org.apache.spark.sql.connector.catalog.LookupCatalog.NonSessionCatalogAndIdentifier$ NonSessionCatalogAndIdentifier$module
org.apache.spark.sql.connector.catalog.LookupCatalog.CatalogAndNamespace$ CatalogAndNamespace$module
org.apache.spark.sql.connector.catalog.LookupCatalog.CatalogAndIdentifier$ CatalogAndIdentifier$module
org.apache.spark.sql.connector.catalog.LookupCatalog.AsTableIdentifier$ AsTableIdentifier$module
boolean bitmap$0
private java.lang.Object readResolve()
org.apache.spark.sql.catalyst.catalog.SessionCatalog catalog
private java.lang.Object readResolve()
private java.lang.Object readResolve()
org.apache.spark.sql.catalyst.expressions.Expression canonicalized
org.apache.spark.sql.catalyst.expressions.Expression child
scala.collection.Seq<A> nameParts
boolean bitmap$0
private java.lang.Object readResolve()
java.lang.String message
private java.lang.Object readResolve()
boolean resolved
org.apache.spark.sql.catalyst.expressions.Expression child
scala.Option<A> aliasFunc
scala.collection.Seq<A> nodePatterns
boolean bitmap$0
private java.lang.Object readResolve()
boolean resolved
scala.collection.Seq<A> nameParts
scala.collection.Seq<A> nodePatterns
boolean bitmap$0
private java.lang.Object readResolve()
boolean resolved
scala.collection.Seq<A> nameParts
boolean isNamespace
boolean bitmap$0
private java.lang.Object readResolve()
boolean resolved
org.apache.spark.sql.catalyst.expressions.Expression deserializer
scala.collection.Seq<A> inputAttributes
scala.collection.Seq<A> nodePatterns
boolean bitmap$0
private java.lang.Object readResolve()
boolean resolved
org.apache.spark.sql.catalyst.expressions.Expression child
org.apache.spark.sql.catalyst.expressions.Expression extraction
boolean bitmap$0
private java.lang.Object readResolve()
boolean resolved
scala.collection.Seq<A> name
boolean bitmap$0
private java.lang.Object readResolve()
boolean resolved
org.apache.spark.sql.connector.catalog.TableChange.ColumnPosition position
boolean bitmap$0
private java.lang.Object readResolve()
boolean resolved
scala.collection.Seq<A> multipartIdentifier
java.lang.String commandName
boolean requirePersistent
scala.Option<A> funcTypeMismatchHint
scala.Option<A> possibleQualifiedName
scala.collection.Seq<A> nodePatterns
boolean bitmap$0
private java.lang.Object readResolve()
boolean resolved
scala.collection.Seq<A> nameParts
scala.collection.Seq<A> arguments
boolean isDistinct
scala.Option<A> filter
boolean ignoreNulls
scala.collection.Seq<A> nodePatterns
boolean bitmap$0
private java.lang.Object readResolve()
boolean resolved
org.apache.spark.sql.catalyst.FunctionIdentifier name
scala.collection.Seq<A> children
scala.collection.Seq<A> nodePatterns
boolean bitmap$0
private java.lang.Object readResolve()
boolean resolved
org.apache.spark.sql.catalyst.expressions.Expression havingCondition
org.apache.spark.sql.catalyst.plans.logical.LogicalPlan child
org.apache.spark.sql.catalyst.expressions.ExpressionSet validConstraints
byte bitmap$0
private java.lang.Object readResolve()
boolean expressionsResolved
boolean resolved
scala.collection.Seq<A> names
scala.collection.Seq<A> rows
byte bitmap$0
private java.lang.Object readResolve()
boolean resolved
scala.collection.Seq<A> multipartIdentifier
boolean bitmap$0
private java.lang.Object readResolve()
boolean resolved
int ordinal
scala.collection.Seq<A> nodePatterns
boolean bitmap$0
private java.lang.Object readResolve()
boolean resolved
scala.collection.immutable.Map<K,V> spec
scala.Option<A> location
boolean bitmap$0
private java.lang.Object readResolve()
java.lang.String regexPattern
scala.Option<A> table
boolean caseSensitive
private java.lang.Object readResolve()
boolean resolved
scala.collection.Seq<A> multipartIdentifier
org.apache.spark.sql.util.CaseInsensitiveStringMap options
boolean isStreaming
scala.collection.Seq<A> nodePatterns
boolean bitmap$0
private java.lang.Object readResolve()
private java.lang.Object readResolve()
boolean resolved
boolean bitmap$0
scala.Option<A> target
private java.lang.Object readResolve()
boolean resolved
scala.collection.Seq<A> outputColumnNames
org.apache.spark.sql.catalyst.plans.logical.LogicalPlan child
scala.collection.Seq<A> nodePatterns
org.apache.spark.sql.catalyst.expressions.ExpressionSet validConstraints
byte bitmap$0
private java.lang.Object readResolve()
boolean resolved
scala.collection.Seq<A> multipartIdentifier
java.lang.String commandName
scala.Option<A> relationTypeMismatchHint
boolean bitmap$0
private java.lang.Object readResolve()
boolean resolved
scala.collection.Seq<A> multipartIdentifier
java.lang.String commandName
boolean allowTempView
boolean bitmap$0
private java.lang.Object readResolve()
boolean resolved
org.apache.spark.sql.catalyst.FunctionIdentifier name
scala.collection.Seq<A> functionArgs
scala.collection.Seq<A> outputNames
boolean bitmap$0
private java.lang.Object readResolve()
boolean resolved
scala.collection.Seq<A> multipartIdentifier
java.lang.String commandName
boolean allowTemp
scala.Option<A> relationTypeMismatchHint
boolean bitmap$0
private java.lang.Object readResolve()
private java.lang.Object readResolve()
scala.Option<A> command
org.apache.spark.sql.catalyst.trees.Origin start
org.apache.spark.sql.catalyst.trees.Origin stop
private java.lang.Object readResolve()