object ParquetMetadataStoreManager extends MetadataStoreManager with Logging
A MetadataBackend backed by Parquet objects
- Alphabetic
- By Inheritance
- ParquetMetadataStoreManager
- Logging
- MetadataStoreManager
- AnyRef
- Any
- Hide All
- Show All
- Public
- All
Value Members
-
final
def
!=(arg0: Any): Boolean
- Definition Classes
- AnyRef → Any
-
final
def
##(): Int
- Definition Classes
- AnyRef → Any
-
final
def
==(arg0: Any): Boolean
- Definition Classes
- AnyRef → Any
-
final
def
asInstanceOf[T0]: T0
- Definition Classes
- Any
-
def
clearActiveMetadataHandles(): Unit
Clear all active MetadataHandles managed by this MetadataStoreManager
Clear all active MetadataHandles managed by this MetadataStoreManager
- Definition Classes
- MetadataStoreManager
-
def
clearStats(): Unit
Clears the stats for all active MetadataHandle instances.
Clears the stats for all active MetadataHandle instances. Should be called before each query (or call clearActiveMetadataHandles) to make sure the aggregated stats are cleared
- Definition Classes
- MetadataStoreManager
-
def
clone(): AnyRef
- Attributes
- protected[java.lang]
- Definition Classes
- AnyRef
- Annotations
- @native() @throws( ... )
-
def
createMetadataHandle(sparkSession: SparkSession, tableIdentifier: String): MetadataHandle
Creates a new instance MetadataHandle Used by getOrCreateMetadataHandle in order to create new instances
Creates a new instance MetadataHandle Used by getOrCreateMetadataHandle in order to create new instances
- Definition Classes
- ParquetMetadataStoreManager → MetadataStoreManager
-
final
def
eq(arg0: AnyRef): Boolean
- Definition Classes
- AnyRef
-
def
equals(arg0: Any): Boolean
- Definition Classes
- AnyRef → Any
-
def
finalize(): Unit
- Attributes
- protected[java.lang]
- Definition Classes
- AnyRef
- Annotations
- @throws( classOf[java.lang.Throwable] )
-
def
getActiveMetadataHandles(): Map[String, MetadataHandle]
Returns all active MetadataHandles managed by this MetadataStoreManager
Returns all active MetadataHandles managed by this MetadataStoreManager
- Definition Classes
- MetadataStoreManager
-
final
def
getClass(): Class[_]
- Definition Classes
- AnyRef → Any
- Annotations
- @native()
-
def
getDataSkippingFileFilterParams(tid: String, sparkSession: SparkSession, fileIndex: FileIndex): Map[String, String]
Returns a map of parameters to be set by the io.xskipper.search.DataSkippingFileFilter on the MetadataHandle when using this MetadataStoreManager.
Returns a map of parameters to be set by the io.xskipper.search.DataSkippingFileFilter on the MetadataHandle when using this MetadataStoreManager. This map can be used to pass specific parameters depending on the type of FileIndex that is being replaced. For example - in ParquetMetadataStoreManager we can specify that the metadata location should be inferred from the table/default db properties
- tid
the table identifier for which the DataSkippingFileFilter will be created
- sparkSession
the sparks session
- fileIndex
the fileIndex that will be replaced by io.xskipper.search.DataSkippingFileFilter
- returns
a map of parameters to be set on the MetadataStore when the DataSkippingFileFilter is created
- Definition Classes
- ParquetMetadataStoreManager → MetadataStoreManager
-
def
getLatestQueryAggregatedStats(): QueryIndexStatsResult
Gets the aggregated latest query skipping stats for all active MetadataHandle instances.
Gets the aggregated latest query skipping stats for all active MetadataHandle instances. In order to get reliable results it is assumed that either clearStats or clearActiveMetadataHandles was called before running the query.
This is needed since the way we aggregate the skipping stats is by going over all active MetadataHandles of the MetadataStoreManager and aggregating their stats. When running multiple queries there could be a scenario in which the first query used dataset
a
and the second query didn't use it, therefore, when calling aggregatedStats for the second query the MetadataHandle for dataseta
will be present as an active MetadataHandle therefore we need its stats to be cleared.- returns
QueryIndexStatsResult instance with the latest query aggregated stats
- Definition Classes
- MetadataStoreManager
-
def
getOrCreateMetadataHandle(sparkSession: SparkSession, tableIdentifier: String): MetadataHandle
Gets an existing MetadataHandle or, if there is no existing one, creates a new one.
Gets an existing MetadataHandle or, if there is no existing one, creates a new one.
- sparkSession
the sparks session
- tableIdentifier
the table identifier of the dataset for which the MetadataHandle instance is created
- returns
a new instance of the MetadataHandle associated with this backend
- Definition Classes
- MetadataStoreManager
-
def
getParams(sparkSession: SparkSession): Map[String, String]
- returns
a map representing relevant MetadataStoreManager parameters
- Definition Classes
- ParquetMetadataStoreManager → MetadataStoreManager
-
def
getType: MetadataStoreManagerType
- returns
the MetadataStoreManagerType associated with this backend
- Definition Classes
- ParquetMetadataStoreManager → MetadataStoreManager
-
def
hashCode(): Int
- Definition Classes
- AnyRef → Any
- Annotations
- @native()
-
def
init(): Unit
A unit function for setting up the MetadataStoreManager.
A unit function for setting up the MetadataStoreManager. The setup is called once when the MetadataStoreManager is registered
- Definition Classes
- ParquetMetadataStoreManager → MetadataStoreManager
-
def
initializeLogIfNecessary(isInterpreter: Boolean, silent: Boolean): Boolean
- Attributes
- protected
- Definition Classes
- Logging
-
def
initializeLogIfNecessary(isInterpreter: Boolean): Unit
- Attributes
- protected
- Definition Classes
- Logging
-
final
def
isInstanceOf[T0]: Boolean
- Definition Classes
- Any
-
def
isTraceEnabled(): Boolean
- Attributes
- protected
- Definition Classes
- Logging
-
def
listIndexedDatasets(sparkSession: SparkSession): Map[String, (MetadataVersionStatus.MetadataVersionStatus, Seq[Index])]
- returns
Map of indexed datasets to their respective md version status and indexes
- Definition Classes
- ParquetMetadataStoreManager → MetadataStoreManager
-
def
log: Logger
- Attributes
- protected
- Definition Classes
- Logging
-
def
logDebug(msg: ⇒ String, throwable: Throwable): Unit
- Attributes
- protected
- Definition Classes
- Logging
-
def
logDebug(msg: ⇒ String): Unit
- Attributes
- protected
- Definition Classes
- Logging
-
def
logError(msg: ⇒ String, throwable: Throwable): Unit
- Attributes
- protected
- Definition Classes
- Logging
-
def
logError(msg: ⇒ String): Unit
- Attributes
- protected
- Definition Classes
- Logging
-
def
logInfo(msg: ⇒ String, throwable: Throwable): Unit
- Attributes
- protected
- Definition Classes
- Logging
-
def
logInfo(msg: ⇒ String): Unit
- Attributes
- protected
- Definition Classes
- Logging
-
def
logName: String
- Attributes
- protected
- Definition Classes
- Logging
-
def
logTrace(msg: ⇒ String, throwable: Throwable): Unit
- Attributes
- protected
- Definition Classes
- Logging
-
def
logTrace(msg: ⇒ String): Unit
- Attributes
- protected
- Definition Classes
- Logging
-
def
logWarning(msg: ⇒ String, throwable: Throwable): Unit
- Attributes
- protected
- Definition Classes
- Logging
-
def
logWarning(msg: ⇒ String): Unit
- Attributes
- protected
- Definition Classes
- Logging
-
val
metadataHandlesInstanceMap: ConcurrentHashMap[String, MetadataHandle]
- Definition Classes
- MetadataStoreManager
-
final
def
ne(arg0: AnyRef): Boolean
- Definition Classes
- AnyRef
-
final
def
notify(): Unit
- Definition Classes
- AnyRef
- Annotations
- @native()
-
final
def
notifyAll(): Unit
- Definition Classes
- AnyRef
- Annotations
- @native()
-
final
def
synchronized[T0](arg0: ⇒ T0): T0
- Definition Classes
- AnyRef
-
def
toString(): String
- Definition Classes
- AnyRef → Any
-
final
def
wait(): Unit
- Definition Classes
- AnyRef
- Annotations
- @throws( ... )
-
final
def
wait(arg0: Long, arg1: Int): Unit
- Definition Classes
- AnyRef
- Annotations
- @throws( ... )
-
final
def
wait(arg0: Long): Unit
- Definition Classes
- AnyRef
- Annotations
- @native() @throws( ... )