public class Db2TableSource extends Object implements org.apache.flink.table.connector.source.ScanTableSource, org.apache.flink.table.connector.source.abilities.SupportsReadingMetadata
| Modifier and Type | Field and Description |
|---|---|
protected List<String> |
metadataKeys
Metadata that is appended at the end of a physical source row.
|
protected org.apache.flink.table.types.DataType |
producedDataType
Data type that describes the final output of the source.
|
| Constructor and Description |
|---|
Db2TableSource(org.apache.flink.table.catalog.ResolvedSchema physicalSchema,
int port,
String hostname,
String database,
String tableName,
String username,
String password,
java.time.ZoneId serverTimeZone,
Properties dbzProperties,
org.apache.flink.cdc.connectors.base.options.StartupOptions startupOptions,
boolean enableParallelRead,
int splitSize,
int splitMetaGroupSize,
int fetchSize,
java.time.Duration connectTimeout,
int connectMaxRetries,
int connectionPoolSize,
double distributionFactorUpper,
double distributionFactorLower,
String chunkKeyColumn,
boolean closeIdleReaders,
boolean skipSnapshotBackfill) |
| Modifier and Type | Method and Description |
|---|---|
void |
applyReadableMetadata(List<String> metadataKeys,
org.apache.flink.table.types.DataType producedDataType) |
String |
asSummaryString() |
org.apache.flink.table.connector.source.DynamicTableSource |
copy() |
boolean |
equals(Object o) |
org.apache.flink.table.connector.ChangelogMode |
getChangelogMode() |
org.apache.flink.table.connector.source.ScanTableSource.ScanRuntimeProvider |
getScanRuntimeProvider(org.apache.flink.table.connector.source.ScanTableSource.ScanContext scanContext) |
int |
hashCode() |
Map<String,org.apache.flink.table.types.DataType> |
listReadableMetadata() |
protected org.apache.flink.table.types.DataType producedDataType
public Db2TableSource(org.apache.flink.table.catalog.ResolvedSchema physicalSchema,
int port,
String hostname,
String database,
String tableName,
String username,
String password,
java.time.ZoneId serverTimeZone,
Properties dbzProperties,
org.apache.flink.cdc.connectors.base.options.StartupOptions startupOptions,
boolean enableParallelRead,
int splitSize,
int splitMetaGroupSize,
int fetchSize,
java.time.Duration connectTimeout,
int connectMaxRetries,
int connectionPoolSize,
double distributionFactorUpper,
double distributionFactorLower,
@Nullable
String chunkKeyColumn,
boolean closeIdleReaders,
boolean skipSnapshotBackfill)
public org.apache.flink.table.connector.ChangelogMode getChangelogMode()
getChangelogMode in interface org.apache.flink.table.connector.source.ScanTableSourcepublic org.apache.flink.table.connector.source.ScanTableSource.ScanRuntimeProvider getScanRuntimeProvider(org.apache.flink.table.connector.source.ScanTableSource.ScanContext scanContext)
getScanRuntimeProvider in interface org.apache.flink.table.connector.source.ScanTableSourcepublic org.apache.flink.table.connector.source.DynamicTableSource copy()
copy in interface org.apache.flink.table.connector.source.DynamicTableSourcepublic String asSummaryString()
asSummaryString in interface org.apache.flink.table.connector.source.DynamicTableSourcepublic Map<String,org.apache.flink.table.types.DataType> listReadableMetadata()
listReadableMetadata in interface org.apache.flink.table.connector.source.abilities.SupportsReadingMetadataCopyright © 2024 The Apache Software Foundation. All rights reserved.