Interface DataLakeDatasetSchema.Builder

  • Method Details

    • name

      The name of the dataset schema.

      Parameters:
      name - The name of the dataset schema.
      Returns:
      Returns a reference to this object so that method calls can be chained together.
    • fields

      The list of field details of the dataset schema.

      Parameters:
      fields - The list of field details of the dataset schema.
      Returns:
      Returns a reference to this object so that method calls can be chained together.
    • fields

      The list of field details of the dataset schema.

      Parameters:
      fields - The list of field details of the dataset schema.
      Returns:
      Returns a reference to this object so that method calls can be chained together.
    • fields

      The list of field details of the dataset schema.

      This is a convenience method that creates an instance of the DataLakeDatasetSchemaField.Builder avoiding the need to create one manually via DataLakeDatasetSchemaField.builder().

      When the Consumer completes, SdkBuilder.build() is called immediately and its result is passed to fields(List<DataLakeDatasetSchemaField>).

      Parameters:
      fields - a consumer that will call methods on DataLakeDatasetSchemaField.Builder
      Returns:
      Returns a reference to this object so that method calls can be chained together.
      See Also:
    • primaryKeys

      The list of primary key fields for the dataset. Primary keys defined can help data ingestion methods to ensure data uniqueness: CreateDataIntegrationFlow's dedupe strategy will leverage primary keys to perform records deduplication before write to dataset; SendDataIntegrationEvent's UPSERT and DELETE can only work with dataset with primary keys. For more details, refer to those data ingestion documentations.

      Note that defining primary keys does not necessarily mean the dataset cannot have duplicate records, duplicate records can still be ingested if CreateDataIntegrationFlow's dedupe disabled or through SendDataIntegrationEvent's APPEND operation.

      Parameters:
      primaryKeys - The list of primary key fields for the dataset. Primary keys defined can help data ingestion methods to ensure data uniqueness: CreateDataIntegrationFlow's dedupe strategy will leverage primary keys to perform records deduplication before write to dataset; SendDataIntegrationEvent's UPSERT and DELETE can only work with dataset with primary keys. For more details, refer to those data ingestion documentations.

      Note that defining primary keys does not necessarily mean the dataset cannot have duplicate records, duplicate records can still be ingested if CreateDataIntegrationFlow's dedupe disabled or through SendDataIntegrationEvent's APPEND operation.

      Returns:
      Returns a reference to this object so that method calls can be chained together.
    • primaryKeys

      The list of primary key fields for the dataset. Primary keys defined can help data ingestion methods to ensure data uniqueness: CreateDataIntegrationFlow's dedupe strategy will leverage primary keys to perform records deduplication before write to dataset; SendDataIntegrationEvent's UPSERT and DELETE can only work with dataset with primary keys. For more details, refer to those data ingestion documentations.

      Note that defining primary keys does not necessarily mean the dataset cannot have duplicate records, duplicate records can still be ingested if CreateDataIntegrationFlow's dedupe disabled or through SendDataIntegrationEvent's APPEND operation.

      Parameters:
      primaryKeys - The list of primary key fields for the dataset. Primary keys defined can help data ingestion methods to ensure data uniqueness: CreateDataIntegrationFlow's dedupe strategy will leverage primary keys to perform records deduplication before write to dataset; SendDataIntegrationEvent's UPSERT and DELETE can only work with dataset with primary keys. For more details, refer to those data ingestion documentations.

      Note that defining primary keys does not necessarily mean the dataset cannot have duplicate records, duplicate records can still be ingested if CreateDataIntegrationFlow's dedupe disabled or through SendDataIntegrationEvent's APPEND operation.

      Returns:
      Returns a reference to this object so that method calls can be chained together.
    • primaryKeys

      The list of primary key fields for the dataset. Primary keys defined can help data ingestion methods to ensure data uniqueness: CreateDataIntegrationFlow's dedupe strategy will leverage primary keys to perform records deduplication before write to dataset; SendDataIntegrationEvent's UPSERT and DELETE can only work with dataset with primary keys. For more details, refer to those data ingestion documentations.

      Note that defining primary keys does not necessarily mean the dataset cannot have duplicate records, duplicate records can still be ingested if CreateDataIntegrationFlow's dedupe disabled or through SendDataIntegrationEvent's APPEND operation.

      This is a convenience method that creates an instance of the DataLakeDatasetPrimaryKeyField.Builder avoiding the need to create one manually via DataLakeDatasetPrimaryKeyField.builder().

      When the Consumer completes, SdkBuilder.build() is called immediately and its result is passed to primaryKeys(List<DataLakeDatasetPrimaryKeyField>).

      Parameters:
      primaryKeys - a consumer that will call methods on DataLakeDatasetPrimaryKeyField.Builder
      Returns:
      Returns a reference to this object so that method calls can be chained together.
      See Also: