airbyte.SourceAzureBlobStorage
Explore with Pulumi AI
SourceAzureBlobStorage Resource
Example Usage
Coming soon!
Coming soon!
Coming soon!
Coming soon!
package generated_program;
import com.pulumi.Context;
import com.pulumi.Pulumi;
import com.pulumi.core.Output;
import com.pulumi.airbyte.SourceAzureBlobStorage;
import com.pulumi.airbyte.SourceAzureBlobStorageArgs;
import com.pulumi.airbyte.inputs.SourceAzureBlobStorageConfigurationArgs;
import com.pulumi.airbyte.inputs.SourceAzureBlobStorageConfigurationCredentialsArgs;
import com.pulumi.airbyte.inputs.SourceAzureBlobStorageConfigurationCredentialsAuthenticateViaClientCredentialsArgs;
import com.pulumi.airbyte.inputs.SourceAzureBlobStorageConfigurationCredentialsAuthenticateViaOauth2Args;
import com.pulumi.airbyte.inputs.SourceAzureBlobStorageConfigurationCredentialsAuthenticateViaStorageAccountKeyArgs;
import java.util.List;
import java.util.ArrayList;
import java.util.Map;
import java.io.File;
import java.nio.file.Files;
import java.nio.file.Paths;
public class App {
    public static void main(String[] args) {
        Pulumi.run(App::stack);
    }
    public static void stack(Context ctx) {
        var mySourceAzureblobstorage = new SourceAzureBlobStorage("mySourceAzureblobstorage", SourceAzureBlobStorageArgs.builder()
            .configuration(SourceAzureBlobStorageConfigurationArgs.builder()
                .azure_blob_storage_account_name("airbyte5storage")
                .azure_blob_storage_container_name("airbytetescontainername")
                .azure_blob_storage_endpoint("blob.core.windows.net")
                .credentials(SourceAzureBlobStorageConfigurationCredentialsArgs.builder()
                    .authenticateViaClientCredentials(SourceAzureBlobStorageConfigurationCredentialsAuthenticateViaClientCredentialsArgs.builder()
                        .appClientId("...my_app_client_id...")
                        .appClientSecret("...my_app_client_secret...")
                        .appTenantId("...my_app_tenant_id...")
                        .build())
                    .authenticateViaOauth2(SourceAzureBlobStorageConfigurationCredentialsAuthenticateViaOauth2Args.builder()
                        .clientId("...my_client_id...")
                        .clientSecret("...my_client_secret...")
                        .refreshToken("...my_refresh_token...")
                        .tenantId("...my_tenant_id...")
                        .build())
                    .authenticateViaStorageAccountKey(SourceAzureBlobStorageConfigurationCredentialsAuthenticateViaStorageAccountKeyArgs.builder()
                        .azureBlobStorageAccountKey("Z8ZkZpteggFx394vm+PJHnGTvdRncaYS+JhLKdj789YNmD+iyGTnG+PV+POiuYNhBg/ACS+LKjd%4FG3FHGN12Nd==")
                        .build())
                    .build())
                .start_date("2021-01-01T00:00:00.000000Z")
                .streams(SourceAzureBlobStorageConfigurationStreamArgs.builder()
                    .daysToSyncIfHistoryIsFull(1)
                    .format(SourceAzureBlobStorageConfigurationStreamFormatArgs.builder()
                        .avroFormat(SourceAzureBlobStorageConfigurationStreamFormatAvroFormatArgs.builder()
                            .doubleAsString(true)
                            .build())
                        .csvFormat(SourceAzureBlobStorageConfigurationStreamFormatCsvFormatArgs.builder()
                            .delimiter("...my_delimiter...")
                            .doubleQuote(true)
                            .encoding("...my_encoding...")
                            .escapeChar("...my_escape_char...")
                            .falseValues("...")
                            .headerDefinition(SourceAzureBlobStorageConfigurationStreamFormatCsvFormatHeaderDefinitionArgs.builder()
                                .autogenerated()
                                .fromCsv()
                                .userProvided(SourceAzureBlobStorageConfigurationStreamFormatCsvFormatHeaderDefinitionUserProvidedArgs.builder()
                                    .columnNames("...")
                                    .build())
                                .build())
                            .ignoreErrorsOnFieldsMismatch(false)
                            .nullValues("...")
                            .quoteChar("...my_quote_char...")
                            .skipRowsAfterHeader(5)
                            .skipRowsBeforeHeader(0)
                            .stringsCanBeNull(true)
                            .trueValues("...")
                            .build())
                        .documentFileTypeFormatExperimental(SourceAzureBlobStorageConfigurationStreamFormatDocumentFileTypeFormatExperimentalArgs.builder()
                            .processing(SourceAzureBlobStorageConfigurationStreamFormatDocumentFileTypeFormatExperimentalProcessingArgs.builder()
                                .local()
                                .build())
                            .skipUnprocessableFiles(false)
                            .strategy("hi_res")
                            .build())
                        .jsonlFormat()
                        .parquetFormat(SourceAzureBlobStorageConfigurationStreamFormatParquetFormatArgs.builder()
                            .decimalAsFloat(false)
                            .build())
                        .build())
                    .globs("...")
                    .inputSchema("...my_input_schema...")
                    .name("...my_name...")
                    .schemaless(true)
                    .validationPolicy("Wait for Discover")
                    .build())
                .build())
            .definitionId("3385920f-d837-42e0-b72d-7927f28bf9f2")
            .secretId("...my_secret_id...")
            .workspaceId("2c3aeaad-c70f-44a8-a981-aca12752c864")
            .build());
    }
}
resources:
  mySourceAzureblobstorage:
    type: airbyte:SourceAzureBlobStorage
    properties:
      configuration:
        azure_blob_storage_account_name: airbyte5storage
        azure_blob_storage_container_name: airbytetescontainername
        azure_blob_storage_endpoint: blob.core.windows.net
        credentials:
          authenticateViaClientCredentials:
            appClientId: '...my_app_client_id...'
            appClientSecret: '...my_app_client_secret...'
            appTenantId: '...my_app_tenant_id...'
          authenticateViaOauth2:
            clientId: '...my_client_id...'
            clientSecret: '...my_client_secret...'
            refreshToken: '...my_refresh_token...'
            tenantId: '...my_tenant_id...'
          authenticateViaStorageAccountKey:
            azureBlobStorageAccountKey: Z8ZkZpteggFx394vm+PJHnGTvdRncaYS+JhLKdj789YNmD+iyGTnG+PV+POiuYNhBg/ACS+LKjd%4FG3FHGN12Nd==
        start_date: 2021-01-01T00:00:00.000000Z
        streams:
          - daysToSyncIfHistoryIsFull: 1
            format:
              avroFormat:
                doubleAsString: true
              csvFormat:
                delimiter: '...my_delimiter...'
                doubleQuote: true
                encoding: '...my_encoding...'
                escapeChar: '...my_escape_char...'
                falseValues:
                  - '...'
                headerDefinition:
                  autogenerated: {}
                  fromCsv: {}
                  userProvided:
                    columnNames:
                      - '...'
                ignoreErrorsOnFieldsMismatch: false
                nullValues:
                  - '...'
                quoteChar: '...my_quote_char...'
                skipRowsAfterHeader: 5
                skipRowsBeforeHeader: 0
                stringsCanBeNull: true
                trueValues:
                  - '...'
              documentFileTypeFormatExperimental:
                processing:
                  local: {}
                skipUnprocessableFiles: false
                strategy: hi_res
              jsonlFormat: {}
              parquetFormat:
                decimalAsFloat: false
            globs:
              - '...'
            inputSchema: '...my_input_schema...'
            name: '...my_name...'
            schemaless: true
            validationPolicy: Wait for Discover
      definitionId: 3385920f-d837-42e0-b72d-7927f28bf9f2
      secretId: '...my_secret_id...'
      workspaceId: 2c3aeaad-c70f-44a8-a981-aca12752c864
Create SourceAzureBlobStorage Resource
Resources are created with functions called constructors. To learn more about declaring and configuring resources, see Resources.
Constructor syntax
new SourceAzureBlobStorage(name: string, args: SourceAzureBlobStorageArgs, opts?: CustomResourceOptions);@overload
def SourceAzureBlobStorage(resource_name: str,
                           args: SourceAzureBlobStorageArgs,
                           opts: Optional[ResourceOptions] = None)
@overload
def SourceAzureBlobStorage(resource_name: str,
                           opts: Optional[ResourceOptions] = None,
                           configuration: Optional[SourceAzureBlobStorageConfigurationArgs] = None,
                           workspace_id: Optional[str] = None,
                           definition_id: Optional[str] = None,
                           name: Optional[str] = None,
                           secret_id: Optional[str] = None)func NewSourceAzureBlobStorage(ctx *Context, name string, args SourceAzureBlobStorageArgs, opts ...ResourceOption) (*SourceAzureBlobStorage, error)public SourceAzureBlobStorage(string name, SourceAzureBlobStorageArgs args, CustomResourceOptions? opts = null)
public SourceAzureBlobStorage(String name, SourceAzureBlobStorageArgs args)
public SourceAzureBlobStorage(String name, SourceAzureBlobStorageArgs args, CustomResourceOptions options)
type: airbyte:SourceAzureBlobStorage
properties: # The arguments to resource properties.
options: # Bag of options to control resource's behavior.
Parameters
- name string
- The unique name of the resource.
- args SourceAzureBlobStorageArgs
- The arguments to resource properties.
- opts CustomResourceOptions
- Bag of options to control resource's behavior.
- resource_name str
- The unique name of the resource.
- args SourceAzureBlobStorageArgs
- The arguments to resource properties.
- opts ResourceOptions
- Bag of options to control resource's behavior.
- ctx Context
- Context object for the current deployment.
- name string
- The unique name of the resource.
- args SourceAzureBlobStorageArgs
- The arguments to resource properties.
- opts ResourceOption
- Bag of options to control resource's behavior.
- name string
- The unique name of the resource.
- args SourceAzureBlobStorageArgs
- The arguments to resource properties.
- opts CustomResourceOptions
- Bag of options to control resource's behavior.
- name String
- The unique name of the resource.
- args SourceAzureBlobStorageArgs
- The arguments to resource properties.
- options CustomResourceOptions
- Bag of options to control resource's behavior.
Constructor example
The following reference example uses placeholder values for all input properties.
var sourceAzureBlobStorageResource = new Airbyte.SourceAzureBlobStorage("sourceAzureBlobStorageResource", new()
{
    Configuration = new Airbyte.Inputs.SourceAzureBlobStorageConfigurationArgs
    {
        AzureBlobStorageAccountName = "string",
        AzureBlobStorageContainerName = "string",
        Credentials = new Airbyte.Inputs.SourceAzureBlobStorageConfigurationCredentialsArgs
        {
            AuthenticateViaClientCredentials = new Airbyte.Inputs.SourceAzureBlobStorageConfigurationCredentialsAuthenticateViaClientCredentialsArgs
            {
                AppClientId = "string",
                AppClientSecret = "string",
                AppTenantId = "string",
            },
            AuthenticateViaOauth2 = new Airbyte.Inputs.SourceAzureBlobStorageConfigurationCredentialsAuthenticateViaOauth2Args
            {
                ClientId = "string",
                ClientSecret = "string",
                RefreshToken = "string",
                TenantId = "string",
            },
            AuthenticateViaStorageAccountKey = new Airbyte.Inputs.SourceAzureBlobStorageConfigurationCredentialsAuthenticateViaStorageAccountKeyArgs
            {
                AzureBlobStorageAccountKey = "string",
            },
        },
        Streams = new[]
        {
            new Airbyte.Inputs.SourceAzureBlobStorageConfigurationStreamArgs
            {
                Format = new Airbyte.Inputs.SourceAzureBlobStorageConfigurationStreamFormatArgs
                {
                    AvroFormat = new Airbyte.Inputs.SourceAzureBlobStorageConfigurationStreamFormatAvroFormatArgs
                    {
                        DoubleAsString = false,
                    },
                    CsvFormat = new Airbyte.Inputs.SourceAzureBlobStorageConfigurationStreamFormatCsvFormatArgs
                    {
                        Delimiter = "string",
                        DoubleQuote = false,
                        Encoding = "string",
                        EscapeChar = "string",
                        FalseValues = new[]
                        {
                            "string",
                        },
                        HeaderDefinition = new Airbyte.Inputs.SourceAzureBlobStorageConfigurationStreamFormatCsvFormatHeaderDefinitionArgs
                        {
                            Autogenerated = null,
                            FromCsv = null,
                            UserProvided = new Airbyte.Inputs.SourceAzureBlobStorageConfigurationStreamFormatCsvFormatHeaderDefinitionUserProvidedArgs
                            {
                                ColumnNames = new[]
                                {
                                    "string",
                                },
                            },
                        },
                        IgnoreErrorsOnFieldsMismatch = false,
                        NullValues = new[]
                        {
                            "string",
                        },
                        QuoteChar = "string",
                        SkipRowsAfterHeader = 0,
                        SkipRowsBeforeHeader = 0,
                        StringsCanBeNull = false,
                        TrueValues = new[]
                        {
                            "string",
                        },
                    },
                    DocumentFileTypeFormatExperimental = new Airbyte.Inputs.SourceAzureBlobStorageConfigurationStreamFormatDocumentFileTypeFormatExperimentalArgs
                    {
                        Processing = new Airbyte.Inputs.SourceAzureBlobStorageConfigurationStreamFormatDocumentFileTypeFormatExperimentalProcessingArgs
                        {
                            Local = null,
                        },
                        SkipUnprocessableFiles = false,
                        Strategy = "string",
                    },
                    JsonlFormat = null,
                    ParquetFormat = new Airbyte.Inputs.SourceAzureBlobStorageConfigurationStreamFormatParquetFormatArgs
                    {
                        DecimalAsFloat = false,
                    },
                },
                Name = "string",
                DaysToSyncIfHistoryIsFull = 0,
                Globs = new[]
                {
                    "string",
                },
                InputSchema = "string",
                Schemaless = false,
                ValidationPolicy = "string",
            },
        },
        AzureBlobStorageEndpoint = "string",
        StartDate = "string",
    },
    WorkspaceId = "string",
    DefinitionId = "string",
    Name = "string",
    SecretId = "string",
});
example, err := airbyte.NewSourceAzureBlobStorage(ctx, "sourceAzureBlobStorageResource", &airbyte.SourceAzureBlobStorageArgs{
Configuration: &.SourceAzureBlobStorageConfigurationArgs{
AzureBlobStorageAccountName: pulumi.String("string"),
AzureBlobStorageContainerName: pulumi.String("string"),
Credentials: &.SourceAzureBlobStorageConfigurationCredentialsArgs{
AuthenticateViaClientCredentials: &.SourceAzureBlobStorageConfigurationCredentialsAuthenticateViaClientCredentialsArgs{
AppClientId: pulumi.String("string"),
AppClientSecret: pulumi.String("string"),
AppTenantId: pulumi.String("string"),
},
AuthenticateViaOauth2: &.SourceAzureBlobStorageConfigurationCredentialsAuthenticateViaOauth2Args{
ClientId: pulumi.String("string"),
ClientSecret: pulumi.String("string"),
RefreshToken: pulumi.String("string"),
TenantId: pulumi.String("string"),
},
AuthenticateViaStorageAccountKey: &.SourceAzureBlobStorageConfigurationCredentialsAuthenticateViaStorageAccountKeyArgs{
AzureBlobStorageAccountKey: pulumi.String("string"),
},
},
Streams: .SourceAzureBlobStorageConfigurationStreamArray{
&.SourceAzureBlobStorageConfigurationStreamArgs{
Format: &.SourceAzureBlobStorageConfigurationStreamFormatArgs{
AvroFormat: &.SourceAzureBlobStorageConfigurationStreamFormatAvroFormatArgs{
DoubleAsString: pulumi.Bool(false),
},
CsvFormat: &.SourceAzureBlobStorageConfigurationStreamFormatCsvFormatArgs{
Delimiter: pulumi.String("string"),
DoubleQuote: pulumi.Bool(false),
Encoding: pulumi.String("string"),
EscapeChar: pulumi.String("string"),
FalseValues: pulumi.StringArray{
pulumi.String("string"),
},
HeaderDefinition: &.SourceAzureBlobStorageConfigurationStreamFormatCsvFormatHeaderDefinitionArgs{
Autogenerated: &.SourceAzureBlobStorageConfigurationStreamFormatCsvFormatHeaderDefinitionAutogeneratedArgs{
},
FromCsv: &.SourceAzureBlobStorageConfigurationStreamFormatCsvFormatHeaderDefinitionFromCsvArgs{
},
UserProvided: &.SourceAzureBlobStorageConfigurationStreamFormatCsvFormatHeaderDefinitionUserProvidedArgs{
ColumnNames: pulumi.StringArray{
pulumi.String("string"),
},
},
},
IgnoreErrorsOnFieldsMismatch: pulumi.Bool(false),
NullValues: pulumi.StringArray{
pulumi.String("string"),
},
QuoteChar: pulumi.String("string"),
SkipRowsAfterHeader: pulumi.Float64(0),
SkipRowsBeforeHeader: pulumi.Float64(0),
StringsCanBeNull: pulumi.Bool(false),
TrueValues: pulumi.StringArray{
pulumi.String("string"),
},
},
DocumentFileTypeFormatExperimental: &.SourceAzureBlobStorageConfigurationStreamFormatDocumentFileTypeFormatExperimentalArgs{
Processing: &.SourceAzureBlobStorageConfigurationStreamFormatDocumentFileTypeFormatExperimentalProcessingArgs{
Local: &.SourceAzureBlobStorageConfigurationStreamFormatDocumentFileTypeFormatExperimentalProcessingLocalArgs{
},
},
SkipUnprocessableFiles: pulumi.Bool(false),
Strategy: pulumi.String("string"),
},
JsonlFormat: &.SourceAzureBlobStorageConfigurationStreamFormatJsonlFormatArgs{
},
ParquetFormat: &.SourceAzureBlobStorageConfigurationStreamFormatParquetFormatArgs{
DecimalAsFloat: pulumi.Bool(false),
},
},
Name: pulumi.String("string"),
DaysToSyncIfHistoryIsFull: pulumi.Float64(0),
Globs: pulumi.StringArray{
pulumi.String("string"),
},
InputSchema: pulumi.String("string"),
Schemaless: pulumi.Bool(false),
ValidationPolicy: pulumi.String("string"),
},
},
AzureBlobStorageEndpoint: pulumi.String("string"),
StartDate: pulumi.String("string"),
},
WorkspaceId: pulumi.String("string"),
DefinitionId: pulumi.String("string"),
Name: pulumi.String("string"),
SecretId: pulumi.String("string"),
})
var sourceAzureBlobStorageResource = new SourceAzureBlobStorage("sourceAzureBlobStorageResource", SourceAzureBlobStorageArgs.builder()
    .configuration(SourceAzureBlobStorageConfigurationArgs.builder()
        .azureBlobStorageAccountName("string")
        .azureBlobStorageContainerName("string")
        .credentials(SourceAzureBlobStorageConfigurationCredentialsArgs.builder()
            .authenticateViaClientCredentials(SourceAzureBlobStorageConfigurationCredentialsAuthenticateViaClientCredentialsArgs.builder()
                .appClientId("string")
                .appClientSecret("string")
                .appTenantId("string")
                .build())
            .authenticateViaOauth2(SourceAzureBlobStorageConfigurationCredentialsAuthenticateViaOauth2Args.builder()
                .clientId("string")
                .clientSecret("string")
                .refreshToken("string")
                .tenantId("string")
                .build())
            .authenticateViaStorageAccountKey(SourceAzureBlobStorageConfigurationCredentialsAuthenticateViaStorageAccountKeyArgs.builder()
                .azureBlobStorageAccountKey("string")
                .build())
            .build())
        .streams(SourceAzureBlobStorageConfigurationStreamArgs.builder()
            .format(SourceAzureBlobStorageConfigurationStreamFormatArgs.builder()
                .avroFormat(SourceAzureBlobStorageConfigurationStreamFormatAvroFormatArgs.builder()
                    .doubleAsString(false)
                    .build())
                .csvFormat(SourceAzureBlobStorageConfigurationStreamFormatCsvFormatArgs.builder()
                    .delimiter("string")
                    .doubleQuote(false)
                    .encoding("string")
                    .escapeChar("string")
                    .falseValues("string")
                    .headerDefinition(SourceAzureBlobStorageConfigurationStreamFormatCsvFormatHeaderDefinitionArgs.builder()
                        .autogenerated()
                        .fromCsv()
                        .userProvided(SourceAzureBlobStorageConfigurationStreamFormatCsvFormatHeaderDefinitionUserProvidedArgs.builder()
                            .columnNames("string")
                            .build())
                        .build())
                    .ignoreErrorsOnFieldsMismatch(false)
                    .nullValues("string")
                    .quoteChar("string")
                    .skipRowsAfterHeader(0)
                    .skipRowsBeforeHeader(0)
                    .stringsCanBeNull(false)
                    .trueValues("string")
                    .build())
                .documentFileTypeFormatExperimental(SourceAzureBlobStorageConfigurationStreamFormatDocumentFileTypeFormatExperimentalArgs.builder()
                    .processing(SourceAzureBlobStorageConfigurationStreamFormatDocumentFileTypeFormatExperimentalProcessingArgs.builder()
                        .local()
                        .build())
                    .skipUnprocessableFiles(false)
                    .strategy("string")
                    .build())
                .jsonlFormat()
                .parquetFormat(SourceAzureBlobStorageConfigurationStreamFormatParquetFormatArgs.builder()
                    .decimalAsFloat(false)
                    .build())
                .build())
            .name("string")
            .daysToSyncIfHistoryIsFull(0)
            .globs("string")
            .inputSchema("string")
            .schemaless(false)
            .validationPolicy("string")
            .build())
        .azureBlobStorageEndpoint("string")
        .startDate("string")
        .build())
    .workspaceId("string")
    .definitionId("string")
    .name("string")
    .secretId("string")
    .build());
source_azure_blob_storage_resource = airbyte.SourceAzureBlobStorage("sourceAzureBlobStorageResource",
    configuration={
        "azure_blob_storage_account_name": "string",
        "azure_blob_storage_container_name": "string",
        "credentials": {
            "authenticate_via_client_credentials": {
                "app_client_id": "string",
                "app_client_secret": "string",
                "app_tenant_id": "string",
            },
            "authenticate_via_oauth2": {
                "client_id": "string",
                "client_secret": "string",
                "refresh_token": "string",
                "tenant_id": "string",
            },
            "authenticate_via_storage_account_key": {
                "azure_blob_storage_account_key": "string",
            },
        },
        "streams": [{
            "format": {
                "avro_format": {
                    "double_as_string": False,
                },
                "csv_format": {
                    "delimiter": "string",
                    "double_quote": False,
                    "encoding": "string",
                    "escape_char": "string",
                    "false_values": ["string"],
                    "header_definition": {
                        "autogenerated": {},
                        "from_csv": {},
                        "user_provided": {
                            "column_names": ["string"],
                        },
                    },
                    "ignore_errors_on_fields_mismatch": False,
                    "null_values": ["string"],
                    "quote_char": "string",
                    "skip_rows_after_header": 0,
                    "skip_rows_before_header": 0,
                    "strings_can_be_null": False,
                    "true_values": ["string"],
                },
                "document_file_type_format_experimental": {
                    "processing": {
                        "local": {},
                    },
                    "skip_unprocessable_files": False,
                    "strategy": "string",
                },
                "jsonl_format": {},
                "parquet_format": {
                    "decimal_as_float": False,
                },
            },
            "name": "string",
            "days_to_sync_if_history_is_full": 0,
            "globs": ["string"],
            "input_schema": "string",
            "schemaless": False,
            "validation_policy": "string",
        }],
        "azure_blob_storage_endpoint": "string",
        "start_date": "string",
    },
    workspace_id="string",
    definition_id="string",
    name="string",
    secret_id="string")
const sourceAzureBlobStorageResource = new airbyte.SourceAzureBlobStorage("sourceAzureBlobStorageResource", {
    configuration: {
        azureBlobStorageAccountName: "string",
        azureBlobStorageContainerName: "string",
        credentials: {
            authenticateViaClientCredentials: {
                appClientId: "string",
                appClientSecret: "string",
                appTenantId: "string",
            },
            authenticateViaOauth2: {
                clientId: "string",
                clientSecret: "string",
                refreshToken: "string",
                tenantId: "string",
            },
            authenticateViaStorageAccountKey: {
                azureBlobStorageAccountKey: "string",
            },
        },
        streams: [{
            format: {
                avroFormat: {
                    doubleAsString: false,
                },
                csvFormat: {
                    delimiter: "string",
                    doubleQuote: false,
                    encoding: "string",
                    escapeChar: "string",
                    falseValues: ["string"],
                    headerDefinition: {
                        autogenerated: {},
                        fromCsv: {},
                        userProvided: {
                            columnNames: ["string"],
                        },
                    },
                    ignoreErrorsOnFieldsMismatch: false,
                    nullValues: ["string"],
                    quoteChar: "string",
                    skipRowsAfterHeader: 0,
                    skipRowsBeforeHeader: 0,
                    stringsCanBeNull: false,
                    trueValues: ["string"],
                },
                documentFileTypeFormatExperimental: {
                    processing: {
                        local: {},
                    },
                    skipUnprocessableFiles: false,
                    strategy: "string",
                },
                jsonlFormat: {},
                parquetFormat: {
                    decimalAsFloat: false,
                },
            },
            name: "string",
            daysToSyncIfHistoryIsFull: 0,
            globs: ["string"],
            inputSchema: "string",
            schemaless: false,
            validationPolicy: "string",
        }],
        azureBlobStorageEndpoint: "string",
        startDate: "string",
    },
    workspaceId: "string",
    definitionId: "string",
    name: "string",
    secretId: "string",
});
type: airbyte:SourceAzureBlobStorage
properties:
    configuration:
        azureBlobStorageAccountName: string
        azureBlobStorageContainerName: string
        azureBlobStorageEndpoint: string
        credentials:
            authenticateViaClientCredentials:
                appClientId: string
                appClientSecret: string
                appTenantId: string
            authenticateViaOauth2:
                clientId: string
                clientSecret: string
                refreshToken: string
                tenantId: string
            authenticateViaStorageAccountKey:
                azureBlobStorageAccountKey: string
        startDate: string
        streams:
            - daysToSyncIfHistoryIsFull: 0
              format:
                avroFormat:
                    doubleAsString: false
                csvFormat:
                    delimiter: string
                    doubleQuote: false
                    encoding: string
                    escapeChar: string
                    falseValues:
                        - string
                    headerDefinition:
                        autogenerated: {}
                        fromCsv: {}
                        userProvided:
                            columnNames:
                                - string
                    ignoreErrorsOnFieldsMismatch: false
                    nullValues:
                        - string
                    quoteChar: string
                    skipRowsAfterHeader: 0
                    skipRowsBeforeHeader: 0
                    stringsCanBeNull: false
                    trueValues:
                        - string
                documentFileTypeFormatExperimental:
                    processing:
                        local: {}
                    skipUnprocessableFiles: false
                    strategy: string
                jsonlFormat: {}
                parquetFormat:
                    decimalAsFloat: false
              globs:
                - string
              inputSchema: string
              name: string
              schemaless: false
              validationPolicy: string
    definitionId: string
    name: string
    secretId: string
    workspaceId: string
SourceAzureBlobStorage Resource Properties
To learn more about resource properties and how to use them, see Inputs and Outputs in the Architecture and Concepts docs.
Inputs
In Python, inputs that are objects can be passed either as argument classes or as dictionary literals.
The SourceAzureBlobStorage resource accepts the following input properties:
- Configuration
SourceAzure Blob Storage Configuration 
- NOTE: When this Spec is changed, legacyconfigtransformer.py must also be modified to uptake the changes because it is responsible for converting legacy Azure Blob Storage v0 configs into v1 configs using the File-Based CDK.
- WorkspaceId string
- DefinitionId string
- The UUID of the connector definition. One of configuration.sourceType or definitionId must be provided. Requires replacement if changed.
- Name string
- Name of the source e.g. dev-mysql-instance.
- SecretId string
- Optional secretID obtained through the public API OAuth redirect flow. Requires replacement if changed.
- Configuration
SourceAzure Blob Storage Configuration Args 
- NOTE: When this Spec is changed, legacyconfigtransformer.py must also be modified to uptake the changes because it is responsible for converting legacy Azure Blob Storage v0 configs into v1 configs using the File-Based CDK.
- WorkspaceId string
- DefinitionId string
- The UUID of the connector definition. One of configuration.sourceType or definitionId must be provided. Requires replacement if changed.
- Name string
- Name of the source e.g. dev-mysql-instance.
- SecretId string
- Optional secretID obtained through the public API OAuth redirect flow. Requires replacement if changed.
- configuration
SourceAzure Blob Storage Configuration 
- NOTE: When this Spec is changed, legacyconfigtransformer.py must also be modified to uptake the changes because it is responsible for converting legacy Azure Blob Storage v0 configs into v1 configs using the File-Based CDK.
- workspaceId String
- definitionId String
- The UUID of the connector definition. One of configuration.sourceType or definitionId must be provided. Requires replacement if changed.
- name String
- Name of the source e.g. dev-mysql-instance.
- secretId String
- Optional secretID obtained through the public API OAuth redirect flow. Requires replacement if changed.
- configuration
SourceAzure Blob Storage Configuration 
- NOTE: When this Spec is changed, legacyconfigtransformer.py must also be modified to uptake the changes because it is responsible for converting legacy Azure Blob Storage v0 configs into v1 configs using the File-Based CDK.
- workspaceId string
- definitionId string
- The UUID of the connector definition. One of configuration.sourceType or definitionId must be provided. Requires replacement if changed.
- name string
- Name of the source e.g. dev-mysql-instance.
- secretId string
- Optional secretID obtained through the public API OAuth redirect flow. Requires replacement if changed.
- configuration
SourceAzure Blob Storage Configuration Args 
- NOTE: When this Spec is changed, legacyconfigtransformer.py must also be modified to uptake the changes because it is responsible for converting legacy Azure Blob Storage v0 configs into v1 configs using the File-Based CDK.
- workspace_id str
- definition_id str
- The UUID of the connector definition. One of configuration.sourceType or definitionId must be provided. Requires replacement if changed.
- name str
- Name of the source e.g. dev-mysql-instance.
- secret_id str
- Optional secretID obtained through the public API OAuth redirect flow. Requires replacement if changed.
- configuration Property Map
- NOTE: When this Spec is changed, legacyconfigtransformer.py must also be modified to uptake the changes because it is responsible for converting legacy Azure Blob Storage v0 configs into v1 configs using the File-Based CDK.
- workspaceId String
- definitionId String
- The UUID of the connector definition. One of configuration.sourceType or definitionId must be provided. Requires replacement if changed.
- name String
- Name of the source e.g. dev-mysql-instance.
- secretId String
- Optional secretID obtained through the public API OAuth redirect flow. Requires replacement if changed.
Outputs
All input properties are implicitly available as output properties. Additionally, the SourceAzureBlobStorage resource produces the following output properties:
- CreatedAt double
- Id string
- The provider-assigned unique ID for this managed resource.
- SourceId string
- SourceType string
- CreatedAt float64
- Id string
- The provider-assigned unique ID for this managed resource.
- SourceId string
- SourceType string
- createdAt Double
- id String
- The provider-assigned unique ID for this managed resource.
- sourceId String
- sourceType String
- createdAt number
- id string
- The provider-assigned unique ID for this managed resource.
- sourceId string
- sourceType string
- created_at float
- id str
- The provider-assigned unique ID for this managed resource.
- source_id str
- source_type str
- createdAt Number
- id String
- The provider-assigned unique ID for this managed resource.
- sourceId String
- sourceType String
Look up Existing SourceAzureBlobStorage Resource
Get an existing SourceAzureBlobStorage resource’s state with the given name, ID, and optional extra properties used to qualify the lookup.
public static get(name: string, id: Input<ID>, state?: SourceAzureBlobStorageState, opts?: CustomResourceOptions): SourceAzureBlobStorage@staticmethod
def get(resource_name: str,
        id: str,
        opts: Optional[ResourceOptions] = None,
        configuration: Optional[SourceAzureBlobStorageConfigurationArgs] = None,
        created_at: Optional[float] = None,
        definition_id: Optional[str] = None,
        name: Optional[str] = None,
        secret_id: Optional[str] = None,
        source_id: Optional[str] = None,
        source_type: Optional[str] = None,
        workspace_id: Optional[str] = None) -> SourceAzureBlobStoragefunc GetSourceAzureBlobStorage(ctx *Context, name string, id IDInput, state *SourceAzureBlobStorageState, opts ...ResourceOption) (*SourceAzureBlobStorage, error)public static SourceAzureBlobStorage Get(string name, Input<string> id, SourceAzureBlobStorageState? state, CustomResourceOptions? opts = null)public static SourceAzureBlobStorage get(String name, Output<String> id, SourceAzureBlobStorageState state, CustomResourceOptions options)resources:  _:    type: airbyte:SourceAzureBlobStorage    get:      id: ${id}- name
- The unique name of the resulting resource.
- id
- The unique provider ID of the resource to lookup.
- state
- Any extra arguments used during the lookup.
- opts
- A bag of options that control this resource's behavior.
- resource_name
- The unique name of the resulting resource.
- id
- The unique provider ID of the resource to lookup.
- name
- The unique name of the resulting resource.
- id
- The unique provider ID of the resource to lookup.
- state
- Any extra arguments used during the lookup.
- opts
- A bag of options that control this resource's behavior.
- name
- The unique name of the resulting resource.
- id
- The unique provider ID of the resource to lookup.
- state
- Any extra arguments used during the lookup.
- opts
- A bag of options that control this resource's behavior.
- name
- The unique name of the resulting resource.
- id
- The unique provider ID of the resource to lookup.
- state
- Any extra arguments used during the lookup.
- opts
- A bag of options that control this resource's behavior.
- Configuration
SourceAzure Blob Storage Configuration 
- NOTE: When this Spec is changed, legacyconfigtransformer.py must also be modified to uptake the changes because it is responsible for converting legacy Azure Blob Storage v0 configs into v1 configs using the File-Based CDK.
- CreatedAt double
- DefinitionId string
- The UUID of the connector definition. One of configuration.sourceType or definitionId must be provided. Requires replacement if changed.
- Name string
- Name of the source e.g. dev-mysql-instance.
- SecretId string
- Optional secretID obtained through the public API OAuth redirect flow. Requires replacement if changed.
- SourceId string
- SourceType string
- WorkspaceId string
- Configuration
SourceAzure Blob Storage Configuration Args 
- NOTE: When this Spec is changed, legacyconfigtransformer.py must also be modified to uptake the changes because it is responsible for converting legacy Azure Blob Storage v0 configs into v1 configs using the File-Based CDK.
- CreatedAt float64
- DefinitionId string
- The UUID of the connector definition. One of configuration.sourceType or definitionId must be provided. Requires replacement if changed.
- Name string
- Name of the source e.g. dev-mysql-instance.
- SecretId string
- Optional secretID obtained through the public API OAuth redirect flow. Requires replacement if changed.
- SourceId string
- SourceType string
- WorkspaceId string
- configuration
SourceAzure Blob Storage Configuration 
- NOTE: When this Spec is changed, legacyconfigtransformer.py must also be modified to uptake the changes because it is responsible for converting legacy Azure Blob Storage v0 configs into v1 configs using the File-Based CDK.
- createdAt Double
- definitionId String
- The UUID of the connector definition. One of configuration.sourceType or definitionId must be provided. Requires replacement if changed.
- name String
- Name of the source e.g. dev-mysql-instance.
- secretId String
- Optional secretID obtained through the public API OAuth redirect flow. Requires replacement if changed.
- sourceId String
- sourceType String
- workspaceId String
- configuration
SourceAzure Blob Storage Configuration 
- NOTE: When this Spec is changed, legacyconfigtransformer.py must also be modified to uptake the changes because it is responsible for converting legacy Azure Blob Storage v0 configs into v1 configs using the File-Based CDK.
- createdAt number
- definitionId string
- The UUID of the connector definition. One of configuration.sourceType or definitionId must be provided. Requires replacement if changed.
- name string
- Name of the source e.g. dev-mysql-instance.
- secretId string
- Optional secretID obtained through the public API OAuth redirect flow. Requires replacement if changed.
- sourceId string
- sourceType string
- workspaceId string
- configuration
SourceAzure Blob Storage Configuration Args 
- NOTE: When this Spec is changed, legacyconfigtransformer.py must also be modified to uptake the changes because it is responsible for converting legacy Azure Blob Storage v0 configs into v1 configs using the File-Based CDK.
- created_at float
- definition_id str
- The UUID of the connector definition. One of configuration.sourceType or definitionId must be provided. Requires replacement if changed.
- name str
- Name of the source e.g. dev-mysql-instance.
- secret_id str
- Optional secretID obtained through the public API OAuth redirect flow. Requires replacement if changed.
- source_id str
- source_type str
- workspace_id str
- configuration Property Map
- NOTE: When this Spec is changed, legacyconfigtransformer.py must also be modified to uptake the changes because it is responsible for converting legacy Azure Blob Storage v0 configs into v1 configs using the File-Based CDK.
- createdAt Number
- definitionId String
- The UUID of the connector definition. One of configuration.sourceType or definitionId must be provided. Requires replacement if changed.
- name String
- Name of the source e.g. dev-mysql-instance.
- secretId String
- Optional secretID obtained through the public API OAuth redirect flow. Requires replacement if changed.
- sourceId String
- sourceType String
- workspaceId String
Supporting Types
SourceAzureBlobStorageConfiguration, SourceAzureBlobStorageConfigurationArgs          
- AzureBlob stringStorage Account Name 
- The account's name of the Azure Blob Storage.
- AzureBlob stringStorage Container Name 
- The name of the Azure blob storage container.
- Credentials
SourceAzure Blob Storage Configuration Credentials 
- Credentials for connecting to the Azure Blob Storage
- Streams
List<SourceAzure Blob Storage Configuration Stream> 
- Each instance of this configuration defines a \n\nstream\n\n. Use this to define which files belong in the stream, their format, and how they should be parsed and validated. When sending data to warehouse destination such as Snowflake or BigQuery, each stream is a separate table.
- AzureBlob stringStorage Endpoint 
- This is Azure Blob Storage endpoint domain name. Leave default value (or leave it empty if run container from command line) to use Microsoft native from example.
- StartDate string
- UTC date and time in the format 2017-01-25T00:00:00.000000Z. Any file modified before this date will not be replicated.
- AzureBlob stringStorage Account Name 
- The account's name of the Azure Blob Storage.
- AzureBlob stringStorage Container Name 
- The name of the Azure blob storage container.
- Credentials
SourceAzure Blob Storage Configuration Credentials 
- Credentials for connecting to the Azure Blob Storage
- Streams
[]SourceAzure Blob Storage Configuration Stream 
- Each instance of this configuration defines a \n\nstream\n\n. Use this to define which files belong in the stream, their format, and how they should be parsed and validated. When sending data to warehouse destination such as Snowflake or BigQuery, each stream is a separate table.
- AzureBlob stringStorage Endpoint 
- This is Azure Blob Storage endpoint domain name. Leave default value (or leave it empty if run container from command line) to use Microsoft native from example.
- StartDate string
- UTC date and time in the format 2017-01-25T00:00:00.000000Z. Any file modified before this date will not be replicated.
- azureBlob StringStorage Account Name 
- The account's name of the Azure Blob Storage.
- azureBlob StringStorage Container Name 
- The name of the Azure blob storage container.
- credentials
SourceAzure Blob Storage Configuration Credentials 
- Credentials for connecting to the Azure Blob Storage
- streams
List<SourceAzure Blob Storage Configuration Stream> 
- Each instance of this configuration defines a \n\nstream\n\n. Use this to define which files belong in the stream, their format, and how they should be parsed and validated. When sending data to warehouse destination such as Snowflake or BigQuery, each stream is a separate table.
- azureBlob StringStorage Endpoint 
- This is Azure Blob Storage endpoint domain name. Leave default value (or leave it empty if run container from command line) to use Microsoft native from example.
- startDate String
- UTC date and time in the format 2017-01-25T00:00:00.000000Z. Any file modified before this date will not be replicated.
- azureBlob stringStorage Account Name 
- The account's name of the Azure Blob Storage.
- azureBlob stringStorage Container Name 
- The name of the Azure blob storage container.
- credentials
SourceAzure Blob Storage Configuration Credentials 
- Credentials for connecting to the Azure Blob Storage
- streams
SourceAzure Blob Storage Configuration Stream[] 
- Each instance of this configuration defines a \n\nstream\n\n. Use this to define which files belong in the stream, their format, and how they should be parsed and validated. When sending data to warehouse destination such as Snowflake or BigQuery, each stream is a separate table.
- azureBlob stringStorage Endpoint 
- This is Azure Blob Storage endpoint domain name. Leave default value (or leave it empty if run container from command line) to use Microsoft native from example.
- startDate string
- UTC date and time in the format 2017-01-25T00:00:00.000000Z. Any file modified before this date will not be replicated.
- azure_blob_ strstorage_ account_ name 
- The account's name of the Azure Blob Storage.
- azure_blob_ strstorage_ container_ name 
- The name of the Azure blob storage container.
- credentials
SourceAzure Blob Storage Configuration Credentials 
- Credentials for connecting to the Azure Blob Storage
- streams
Sequence[SourceAzure Blob Storage Configuration Stream] 
- Each instance of this configuration defines a \n\nstream\n\n. Use this to define which files belong in the stream, their format, and how they should be parsed and validated. When sending data to warehouse destination such as Snowflake or BigQuery, each stream is a separate table.
- azure_blob_ strstorage_ endpoint 
- This is Azure Blob Storage endpoint domain name. Leave default value (or leave it empty if run container from command line) to use Microsoft native from example.
- start_date str
- UTC date and time in the format 2017-01-25T00:00:00.000000Z. Any file modified before this date will not be replicated.
- azureBlob StringStorage Account Name 
- The account's name of the Azure Blob Storage.
- azureBlob StringStorage Container Name 
- The name of the Azure blob storage container.
- credentials Property Map
- Credentials for connecting to the Azure Blob Storage
- streams List<Property Map>
- Each instance of this configuration defines a \n\nstream\n\n. Use this to define which files belong in the stream, their format, and how they should be parsed and validated. When sending data to warehouse destination such as Snowflake or BigQuery, each stream is a separate table.
- azureBlob StringStorage Endpoint 
- This is Azure Blob Storage endpoint domain name. Leave default value (or leave it empty if run container from command line) to use Microsoft native from example.
- startDate String
- UTC date and time in the format 2017-01-25T00:00:00.000000Z. Any file modified before this date will not be replicated.
SourceAzureBlobStorageConfigurationCredentials, SourceAzureBlobStorageConfigurationCredentialsArgs            
- AuthenticateVia SourceClient Credentials Azure Blob Storage Configuration Credentials Authenticate Via Client Credentials 
- AuthenticateVia SourceOauth2 Azure Blob Storage Configuration Credentials Authenticate Via Oauth2 
- AuthenticateVia SourceStorage Account Key Azure Blob Storage Configuration Credentials Authenticate Via Storage Account Key 
- AuthenticateVia SourceClient Credentials Azure Blob Storage Configuration Credentials Authenticate Via Client Credentials 
- AuthenticateVia SourceOauth2 Azure Blob Storage Configuration Credentials Authenticate Via Oauth2 
- AuthenticateVia SourceStorage Account Key Azure Blob Storage Configuration Credentials Authenticate Via Storage Account Key 
- authenticateVia SourceClient Credentials Azure Blob Storage Configuration Credentials Authenticate Via Client Credentials 
- authenticateVia SourceOauth2 Azure Blob Storage Configuration Credentials Authenticate Via Oauth2 
- authenticateVia SourceStorage Account Key Azure Blob Storage Configuration Credentials Authenticate Via Storage Account Key 
- authenticateVia SourceClient Credentials Azure Blob Storage Configuration Credentials Authenticate Via Client Credentials 
- authenticateVia SourceOauth2 Azure Blob Storage Configuration Credentials Authenticate Via Oauth2 
- authenticateVia SourceStorage Account Key Azure Blob Storage Configuration Credentials Authenticate Via Storage Account Key 
- authenticate_via_ Sourceclient_ credentials Azure Blob Storage Configuration Credentials Authenticate Via Client Credentials 
- authenticate_via_ Sourceoauth2 Azure Blob Storage Configuration Credentials Authenticate Via Oauth2 
- authenticate_via_ Sourcestorage_ account_ key Azure Blob Storage Configuration Credentials Authenticate Via Storage Account Key 
SourceAzureBlobStorageConfigurationCredentialsAuthenticateViaClientCredentials, SourceAzureBlobStorageConfigurationCredentialsAuthenticateViaClientCredentialsArgs                    
- AppClient stringId 
- Client ID of your Microsoft developer application
- AppClient stringSecret 
- Client Secret of your Microsoft developer application
- AppTenant stringId 
- Tenant ID of the Microsoft Azure Application
- AppClient stringId 
- Client ID of your Microsoft developer application
- AppClient stringSecret 
- Client Secret of your Microsoft developer application
- AppTenant stringId 
- Tenant ID of the Microsoft Azure Application
- appClient StringId 
- Client ID of your Microsoft developer application
- appClient StringSecret 
- Client Secret of your Microsoft developer application
- appTenant StringId 
- Tenant ID of the Microsoft Azure Application
- appClient stringId 
- Client ID of your Microsoft developer application
- appClient stringSecret 
- Client Secret of your Microsoft developer application
- appTenant stringId 
- Tenant ID of the Microsoft Azure Application
- app_client_ strid 
- Client ID of your Microsoft developer application
- app_client_ strsecret 
- Client Secret of your Microsoft developer application
- app_tenant_ strid 
- Tenant ID of the Microsoft Azure Application
- appClient StringId 
- Client ID of your Microsoft developer application
- appClient StringSecret 
- Client Secret of your Microsoft developer application
- appTenant StringId 
- Tenant ID of the Microsoft Azure Application
SourceAzureBlobStorageConfigurationCredentialsAuthenticateViaOauth2, SourceAzureBlobStorageConfigurationCredentialsAuthenticateViaOauth2Args                  
- ClientId string
- Client ID of your Microsoft developer application
- ClientSecret string
- Client Secret of your Microsoft developer application
- RefreshToken string
- Refresh Token of your Microsoft developer application
- TenantId string
- Tenant ID of the Microsoft Azure Application user
- ClientId string
- Client ID of your Microsoft developer application
- ClientSecret string
- Client Secret of your Microsoft developer application
- RefreshToken string
- Refresh Token of your Microsoft developer application
- TenantId string
- Tenant ID of the Microsoft Azure Application user
- clientId String
- Client ID of your Microsoft developer application
- clientSecret String
- Client Secret of your Microsoft developer application
- refreshToken String
- Refresh Token of your Microsoft developer application
- tenantId String
- Tenant ID of the Microsoft Azure Application user
- clientId string
- Client ID of your Microsoft developer application
- clientSecret string
- Client Secret of your Microsoft developer application
- refreshToken string
- Refresh Token of your Microsoft developer application
- tenantId string
- Tenant ID of the Microsoft Azure Application user
- client_id str
- Client ID of your Microsoft developer application
- client_secret str
- Client Secret of your Microsoft developer application
- refresh_token str
- Refresh Token of your Microsoft developer application
- tenant_id str
- Tenant ID of the Microsoft Azure Application user
- clientId String
- Client ID of your Microsoft developer application
- clientSecret String
- Client Secret of your Microsoft developer application
- refreshToken String
- Refresh Token of your Microsoft developer application
- tenantId String
- Tenant ID of the Microsoft Azure Application user
SourceAzureBlobStorageConfigurationCredentialsAuthenticateViaStorageAccountKey, SourceAzureBlobStorageConfigurationCredentialsAuthenticateViaStorageAccountKeyArgs                      
- AzureBlob stringStorage Account Key 
- The Azure blob storage account key.
- AzureBlob stringStorage Account Key 
- The Azure blob storage account key.
- azureBlob StringStorage Account Key 
- The Azure blob storage account key.
- azureBlob stringStorage Account Key 
- The Azure blob storage account key.
- azure_blob_ strstorage_ account_ key 
- The Azure blob storage account key.
- azureBlob StringStorage Account Key 
- The Azure blob storage account key.
SourceAzureBlobStorageConfigurationStream, SourceAzureBlobStorageConfigurationStreamArgs            
- Format
SourceAzure Blob Storage Configuration Stream Format 
- The configuration options that are used to alter how to read incoming files that deviate from the standard formatting.
- Name string
- The name of the stream.
- DaysTo doubleSync If History Is Full 
- When the state history of the file store is full, syncs will only read files that were last modified in the provided day range. Default: 3
- Globs List<string>
- The pattern used to specify which files should be selected from the file system. For more information on glob pattern matching look \n\nhere\n\n.
- InputSchema string
- The schema that will be used to validate records extracted from the file. This will override the stream schema that is auto-detected from incoming files.
- Schemaless bool
- When enabled, syncs will not validate or structure records against the stream's schema. Default: false
- ValidationPolicy string
- The name of the validation policy that dictates sync behavior when a record does not adhere to the stream schema. Default: "Emit Record"; must be one of ["Emit Record", "Skip Record", "Wait for Discover"]
- Format
SourceAzure Blob Storage Configuration Stream Format 
- The configuration options that are used to alter how to read incoming files that deviate from the standard formatting.
- Name string
- The name of the stream.
- DaysTo float64Sync If History Is Full 
- When the state history of the file store is full, syncs will only read files that were last modified in the provided day range. Default: 3
- Globs []string
- The pattern used to specify which files should be selected from the file system. For more information on glob pattern matching look \n\nhere\n\n.
- InputSchema string
- The schema that will be used to validate records extracted from the file. This will override the stream schema that is auto-detected from incoming files.
- Schemaless bool
- When enabled, syncs will not validate or structure records against the stream's schema. Default: false
- ValidationPolicy string
- The name of the validation policy that dictates sync behavior when a record does not adhere to the stream schema. Default: "Emit Record"; must be one of ["Emit Record", "Skip Record", "Wait for Discover"]
- format
SourceAzure Blob Storage Configuration Stream Format 
- The configuration options that are used to alter how to read incoming files that deviate from the standard formatting.
- name String
- The name of the stream.
- daysTo DoubleSync If History Is Full 
- When the state history of the file store is full, syncs will only read files that were last modified in the provided day range. Default: 3
- globs List<String>
- The pattern used to specify which files should be selected from the file system. For more information on glob pattern matching look \n\nhere\n\n.
- inputSchema String
- The schema that will be used to validate records extracted from the file. This will override the stream schema that is auto-detected from incoming files.
- schemaless Boolean
- When enabled, syncs will not validate or structure records against the stream's schema. Default: false
- validationPolicy String
- The name of the validation policy that dictates sync behavior when a record does not adhere to the stream schema. Default: "Emit Record"; must be one of ["Emit Record", "Skip Record", "Wait for Discover"]
- format
SourceAzure Blob Storage Configuration Stream Format 
- The configuration options that are used to alter how to read incoming files that deviate from the standard formatting.
- name string
- The name of the stream.
- daysTo numberSync If History Is Full 
- When the state history of the file store is full, syncs will only read files that were last modified in the provided day range. Default: 3
- globs string[]
- The pattern used to specify which files should be selected from the file system. For more information on glob pattern matching look \n\nhere\n\n.
- inputSchema string
- The schema that will be used to validate records extracted from the file. This will override the stream schema that is auto-detected from incoming files.
- schemaless boolean
- When enabled, syncs will not validate or structure records against the stream's schema. Default: false
- validationPolicy string
- The name of the validation policy that dictates sync behavior when a record does not adhere to the stream schema. Default: "Emit Record"; must be one of ["Emit Record", "Skip Record", "Wait for Discover"]
- format
SourceAzure Blob Storage Configuration Stream Format 
- The configuration options that are used to alter how to read incoming files that deviate from the standard formatting.
- name str
- The name of the stream.
- days_to_ floatsync_ if_ history_ is_ full 
- When the state history of the file store is full, syncs will only read files that were last modified in the provided day range. Default: 3
- globs Sequence[str]
- The pattern used to specify which files should be selected from the file system. For more information on glob pattern matching look \n\nhere\n\n.
- input_schema str
- The schema that will be used to validate records extracted from the file. This will override the stream schema that is auto-detected from incoming files.
- schemaless bool
- When enabled, syncs will not validate or structure records against the stream's schema. Default: false
- validation_policy str
- The name of the validation policy that dictates sync behavior when a record does not adhere to the stream schema. Default: "Emit Record"; must be one of ["Emit Record", "Skip Record", "Wait for Discover"]
- format Property Map
- The configuration options that are used to alter how to read incoming files that deviate from the standard formatting.
- name String
- The name of the stream.
- daysTo NumberSync If History Is Full 
- When the state history of the file store is full, syncs will only read files that were last modified in the provided day range. Default: 3
- globs List<String>
- The pattern used to specify which files should be selected from the file system. For more information on glob pattern matching look \n\nhere\n\n.
- inputSchema String
- The schema that will be used to validate records extracted from the file. This will override the stream schema that is auto-detected from incoming files.
- schemaless Boolean
- When enabled, syncs will not validate or structure records against the stream's schema. Default: false
- validationPolicy String
- The name of the validation policy that dictates sync behavior when a record does not adhere to the stream schema. Default: "Emit Record"; must be one of ["Emit Record", "Skip Record", "Wait for Discover"]
SourceAzureBlobStorageConfigurationStreamFormat, SourceAzureBlobStorageConfigurationStreamFormatArgs              
- AvroFormat SourceAzure Blob Storage Configuration Stream Format Avro Format 
- CsvFormat SourceAzure Blob Storage Configuration Stream Format Csv Format 
- DocumentFile SourceType Format Experimental Azure Blob Storage Configuration Stream Format Document File Type Format Experimental 
- Extract text from document formats (.pdf, .docx, .md, .pptx) and emit as one record per file.
- JsonlFormat SourceAzure Blob Storage Configuration Stream Format Jsonl Format 
- ParquetFormat SourceAzure Blob Storage Configuration Stream Format Parquet Format 
- AvroFormat SourceAzure Blob Storage Configuration Stream Format Avro Format 
- CsvFormat SourceAzure Blob Storage Configuration Stream Format Csv Format 
- DocumentFile SourceType Format Experimental Azure Blob Storage Configuration Stream Format Document File Type Format Experimental 
- Extract text from document formats (.pdf, .docx, .md, .pptx) and emit as one record per file.
- JsonlFormat SourceAzure Blob Storage Configuration Stream Format Jsonl Format 
- ParquetFormat SourceAzure Blob Storage Configuration Stream Format Parquet Format 
- avroFormat SourceAzure Blob Storage Configuration Stream Format Avro Format 
- csvFormat SourceAzure Blob Storage Configuration Stream Format Csv Format 
- documentFile SourceType Format Experimental Azure Blob Storage Configuration Stream Format Document File Type Format Experimental 
- Extract text from document formats (.pdf, .docx, .md, .pptx) and emit as one record per file.
- jsonlFormat SourceAzure Blob Storage Configuration Stream Format Jsonl Format 
- parquetFormat SourceAzure Blob Storage Configuration Stream Format Parquet Format 
- avroFormat SourceAzure Blob Storage Configuration Stream Format Avro Format 
- csvFormat SourceAzure Blob Storage Configuration Stream Format Csv Format 
- documentFile SourceType Format Experimental Azure Blob Storage Configuration Stream Format Document File Type Format Experimental 
- Extract text from document formats (.pdf, .docx, .md, .pptx) and emit as one record per file.
- jsonlFormat SourceAzure Blob Storage Configuration Stream Format Jsonl Format 
- parquetFormat SourceAzure Blob Storage Configuration Stream Format Parquet Format 
- avro_format SourceAzure Blob Storage Configuration Stream Format Avro Format 
- csv_format SourceAzure Blob Storage Configuration Stream Format Csv Format 
- document_file_ Sourcetype_ format_ experimental Azure Blob Storage Configuration Stream Format Document File Type Format Experimental 
- Extract text from document formats (.pdf, .docx, .md, .pptx) and emit as one record per file.
- jsonl_format SourceAzure Blob Storage Configuration Stream Format Jsonl Format 
- parquet_format SourceAzure Blob Storage Configuration Stream Format Parquet Format 
- avroFormat Property Map
- csvFormat Property Map
- documentFile Property MapType Format Experimental 
- Extract text from document formats (.pdf, .docx, .md, .pptx) and emit as one record per file.
- jsonlFormat Property Map
- parquetFormat Property Map
SourceAzureBlobStorageConfigurationStreamFormatAvroFormat, SourceAzureBlobStorageConfigurationStreamFormatAvroFormatArgs                  
- DoubleAs boolString 
- Whether to convert double fields to strings. This is recommended if you have decimal numbers with a high degree of precision because there can be a loss precision when handling floating point numbers. Default: false
- DoubleAs boolString 
- Whether to convert double fields to strings. This is recommended if you have decimal numbers with a high degree of precision because there can be a loss precision when handling floating point numbers. Default: false
- doubleAs BooleanString 
- Whether to convert double fields to strings. This is recommended if you have decimal numbers with a high degree of precision because there can be a loss precision when handling floating point numbers. Default: false
- doubleAs booleanString 
- Whether to convert double fields to strings. This is recommended if you have decimal numbers with a high degree of precision because there can be a loss precision when handling floating point numbers. Default: false
- double_as_ boolstring 
- Whether to convert double fields to strings. This is recommended if you have decimal numbers with a high degree of precision because there can be a loss precision when handling floating point numbers. Default: false
- doubleAs BooleanString 
- Whether to convert double fields to strings. This is recommended if you have decimal numbers with a high degree of precision because there can be a loss precision when handling floating point numbers. Default: false
SourceAzureBlobStorageConfigurationStreamFormatCsvFormat, SourceAzureBlobStorageConfigurationStreamFormatCsvFormatArgs                  
- Delimiter string
- The character delimiting individual cells in the CSV data. This may only be a 1-character string. For tab-delimited data enter '\t'. Default: ","
- DoubleQuote bool
- Whether two quotes in a quoted CSV value denote a single quote in the data. Default: true
- Encoding string
- The character encoding of the CSV data. Leave blank to default to \n\nUTF8\n\n. See \n\nlist of python encodings\n\n for allowable options. Default: "utf8"
- EscapeChar string
- The character used for escaping special characters. To disallow escaping, leave this field blank.
- FalseValues List<string>
- A set of case-sensitive strings that should be interpreted as false values.
- HeaderDefinition SourceAzure Blob Storage Configuration Stream Format Csv Format Header Definition 
- How headers will be defined. User Providedassumes the CSV does not have a header row and uses the headers provided andAutogeneratedassumes the CSV does not have a header row and the CDK will generate headers using forf{i}whereiis the index starting from 0. Else, the default behavior is to use the header from the CSV file. If a user wants to autogenerate or provide column names for a CSV having headers, they can skip rows.
- IgnoreErrors boolOn Fields Mismatch 
- Whether to ignore errors that occur when the number of fields in the CSV does not match the number of columns in the schema. Default: false
- NullValues List<string>
- A set of case-sensitive strings that should be interpreted as null values. For example, if the value 'NA' should be interpreted as null, enter 'NA' in this field.
- QuoteChar string
- The character used for quoting CSV values. To disallow quoting, make this field blank. Default: """
- SkipRows doubleAfter Header 
- The number of rows to skip after the header row. Default: 0
- SkipRows doubleBefore Header 
- The number of rows to skip before the header row. For example, if the header row is on the 3rd row, enter 2 in this field. Default: 0
- StringsCan boolBe Null 
- Whether strings can be interpreted as null values. If true, strings that match the nullvalues set will be interpreted as null. If false, strings that match the nullvalues set will be interpreted as the string itself. Default: true
- TrueValues List<string>
- A set of case-sensitive strings that should be interpreted as true values.
- Delimiter string
- The character delimiting individual cells in the CSV data. This may only be a 1-character string. For tab-delimited data enter '\t'. Default: ","
- DoubleQuote bool
- Whether two quotes in a quoted CSV value denote a single quote in the data. Default: true
- Encoding string
- The character encoding of the CSV data. Leave blank to default to \n\nUTF8\n\n. See \n\nlist of python encodings\n\n for allowable options. Default: "utf8"
- EscapeChar string
- The character used for escaping special characters. To disallow escaping, leave this field blank.
- FalseValues []string
- A set of case-sensitive strings that should be interpreted as false values.
- HeaderDefinition SourceAzure Blob Storage Configuration Stream Format Csv Format Header Definition 
- How headers will be defined. User Providedassumes the CSV does not have a header row and uses the headers provided andAutogeneratedassumes the CSV does not have a header row and the CDK will generate headers using forf{i}whereiis the index starting from 0. Else, the default behavior is to use the header from the CSV file. If a user wants to autogenerate or provide column names for a CSV having headers, they can skip rows.
- IgnoreErrors boolOn Fields Mismatch 
- Whether to ignore errors that occur when the number of fields in the CSV does not match the number of columns in the schema. Default: false
- NullValues []string
- A set of case-sensitive strings that should be interpreted as null values. For example, if the value 'NA' should be interpreted as null, enter 'NA' in this field.
- QuoteChar string
- The character used for quoting CSV values. To disallow quoting, make this field blank. Default: """
- SkipRows float64After Header 
- The number of rows to skip after the header row. Default: 0
- SkipRows float64Before Header 
- The number of rows to skip before the header row. For example, if the header row is on the 3rd row, enter 2 in this field. Default: 0
- StringsCan boolBe Null 
- Whether strings can be interpreted as null values. If true, strings that match the nullvalues set will be interpreted as null. If false, strings that match the nullvalues set will be interpreted as the string itself. Default: true
- TrueValues []string
- A set of case-sensitive strings that should be interpreted as true values.
- delimiter String
- The character delimiting individual cells in the CSV data. This may only be a 1-character string. For tab-delimited data enter '\t'. Default: ","
- doubleQuote Boolean
- Whether two quotes in a quoted CSV value denote a single quote in the data. Default: true
- encoding String
- The character encoding of the CSV data. Leave blank to default to \n\nUTF8\n\n. See \n\nlist of python encodings\n\n for allowable options. Default: "utf8"
- escapeChar String
- The character used for escaping special characters. To disallow escaping, leave this field blank.
- falseValues List<String>
- A set of case-sensitive strings that should be interpreted as false values.
- headerDefinition SourceAzure Blob Storage Configuration Stream Format Csv Format Header Definition 
- How headers will be defined. User Providedassumes the CSV does not have a header row and uses the headers provided andAutogeneratedassumes the CSV does not have a header row and the CDK will generate headers using forf{i}whereiis the index starting from 0. Else, the default behavior is to use the header from the CSV file. If a user wants to autogenerate or provide column names for a CSV having headers, they can skip rows.
- ignoreErrors BooleanOn Fields Mismatch 
- Whether to ignore errors that occur when the number of fields in the CSV does not match the number of columns in the schema. Default: false
- nullValues List<String>
- A set of case-sensitive strings that should be interpreted as null values. For example, if the value 'NA' should be interpreted as null, enter 'NA' in this field.
- quoteChar String
- The character used for quoting CSV values. To disallow quoting, make this field blank. Default: """
- skipRows DoubleAfter Header 
- The number of rows to skip after the header row. Default: 0
- skipRows DoubleBefore Header 
- The number of rows to skip before the header row. For example, if the header row is on the 3rd row, enter 2 in this field. Default: 0
- stringsCan BooleanBe Null 
- Whether strings can be interpreted as null values. If true, strings that match the nullvalues set will be interpreted as null. If false, strings that match the nullvalues set will be interpreted as the string itself. Default: true
- trueValues List<String>
- A set of case-sensitive strings that should be interpreted as true values.
- delimiter string
- The character delimiting individual cells in the CSV data. This may only be a 1-character string. For tab-delimited data enter '\t'. Default: ","
- doubleQuote boolean
- Whether two quotes in a quoted CSV value denote a single quote in the data. Default: true
- encoding string
- The character encoding of the CSV data. Leave blank to default to \n\nUTF8\n\n. See \n\nlist of python encodings\n\n for allowable options. Default: "utf8"
- escapeChar string
- The character used for escaping special characters. To disallow escaping, leave this field blank.
- falseValues string[]
- A set of case-sensitive strings that should be interpreted as false values.
- headerDefinition SourceAzure Blob Storage Configuration Stream Format Csv Format Header Definition 
- How headers will be defined. User Providedassumes the CSV does not have a header row and uses the headers provided andAutogeneratedassumes the CSV does not have a header row and the CDK will generate headers using forf{i}whereiis the index starting from 0. Else, the default behavior is to use the header from the CSV file. If a user wants to autogenerate or provide column names for a CSV having headers, they can skip rows.
- ignoreErrors booleanOn Fields Mismatch 
- Whether to ignore errors that occur when the number of fields in the CSV does not match the number of columns in the schema. Default: false
- nullValues string[]
- A set of case-sensitive strings that should be interpreted as null values. For example, if the value 'NA' should be interpreted as null, enter 'NA' in this field.
- quoteChar string
- The character used for quoting CSV values. To disallow quoting, make this field blank. Default: """
- skipRows numberAfter Header 
- The number of rows to skip after the header row. Default: 0
- skipRows numberBefore Header 
- The number of rows to skip before the header row. For example, if the header row is on the 3rd row, enter 2 in this field. Default: 0
- stringsCan booleanBe Null 
- Whether strings can be interpreted as null values. If true, strings that match the nullvalues set will be interpreted as null. If false, strings that match the nullvalues set will be interpreted as the string itself. Default: true
- trueValues string[]
- A set of case-sensitive strings that should be interpreted as true values.
- delimiter str
- The character delimiting individual cells in the CSV data. This may only be a 1-character string. For tab-delimited data enter '\t'. Default: ","
- double_quote bool
- Whether two quotes in a quoted CSV value denote a single quote in the data. Default: true
- encoding str
- The character encoding of the CSV data. Leave blank to default to \n\nUTF8\n\n. See \n\nlist of python encodings\n\n for allowable options. Default: "utf8"
- escape_char str
- The character used for escaping special characters. To disallow escaping, leave this field blank.
- false_values Sequence[str]
- A set of case-sensitive strings that should be interpreted as false values.
- header_definition SourceAzure Blob Storage Configuration Stream Format Csv Format Header Definition 
- How headers will be defined. User Providedassumes the CSV does not have a header row and uses the headers provided andAutogeneratedassumes the CSV does not have a header row and the CDK will generate headers using forf{i}whereiis the index starting from 0. Else, the default behavior is to use the header from the CSV file. If a user wants to autogenerate or provide column names for a CSV having headers, they can skip rows.
- ignore_errors_ boolon_ fields_ mismatch 
- Whether to ignore errors that occur when the number of fields in the CSV does not match the number of columns in the schema. Default: false
- null_values Sequence[str]
- A set of case-sensitive strings that should be interpreted as null values. For example, if the value 'NA' should be interpreted as null, enter 'NA' in this field.
- quote_char str
- The character used for quoting CSV values. To disallow quoting, make this field blank. Default: """
- skip_rows_ floatafter_ header 
- The number of rows to skip after the header row. Default: 0
- skip_rows_ floatbefore_ header 
- The number of rows to skip before the header row. For example, if the header row is on the 3rd row, enter 2 in this field. Default: 0
- strings_can_ boolbe_ null 
- Whether strings can be interpreted as null values. If true, strings that match the nullvalues set will be interpreted as null. If false, strings that match the nullvalues set will be interpreted as the string itself. Default: true
- true_values Sequence[str]
- A set of case-sensitive strings that should be interpreted as true values.
- delimiter String
- The character delimiting individual cells in the CSV data. This may only be a 1-character string. For tab-delimited data enter '\t'. Default: ","
- doubleQuote Boolean
- Whether two quotes in a quoted CSV value denote a single quote in the data. Default: true
- encoding String
- The character encoding of the CSV data. Leave blank to default to \n\nUTF8\n\n. See \n\nlist of python encodings\n\n for allowable options. Default: "utf8"
- escapeChar String
- The character used for escaping special characters. To disallow escaping, leave this field blank.
- falseValues List<String>
- A set of case-sensitive strings that should be interpreted as false values.
- headerDefinition Property Map
- How headers will be defined. User Providedassumes the CSV does not have a header row and uses the headers provided andAutogeneratedassumes the CSV does not have a header row and the CDK will generate headers using forf{i}whereiis the index starting from 0. Else, the default behavior is to use the header from the CSV file. If a user wants to autogenerate or provide column names for a CSV having headers, they can skip rows.
- ignoreErrors BooleanOn Fields Mismatch 
- Whether to ignore errors that occur when the number of fields in the CSV does not match the number of columns in the schema. Default: false
- nullValues List<String>
- A set of case-sensitive strings that should be interpreted as null values. For example, if the value 'NA' should be interpreted as null, enter 'NA' in this field.
- quoteChar String
- The character used for quoting CSV values. To disallow quoting, make this field blank. Default: """
- skipRows NumberAfter Header 
- The number of rows to skip after the header row. Default: 0
- skipRows NumberBefore Header 
- The number of rows to skip before the header row. For example, if the header row is on the 3rd row, enter 2 in this field. Default: 0
- stringsCan BooleanBe Null 
- Whether strings can be interpreted as null values. If true, strings that match the nullvalues set will be interpreted as null. If false, strings that match the nullvalues set will be interpreted as the string itself. Default: true
- trueValues List<String>
- A set of case-sensitive strings that should be interpreted as true values.
SourceAzureBlobStorageConfigurationStreamFormatCsvFormatHeaderDefinition, SourceAzureBlobStorageConfigurationStreamFormatCsvFormatHeaderDefinitionArgs                      
SourceAzureBlobStorageConfigurationStreamFormatCsvFormatHeaderDefinitionUserProvided, SourceAzureBlobStorageConfigurationStreamFormatCsvFormatHeaderDefinitionUserProvidedArgs                          
- ColumnNames List<string>
- The column names that will be used while emitting the CSV records
- ColumnNames []string
- The column names that will be used while emitting the CSV records
- columnNames List<String>
- The column names that will be used while emitting the CSV records
- columnNames string[]
- The column names that will be used while emitting the CSV records
- column_names Sequence[str]
- The column names that will be used while emitting the CSV records
- columnNames List<String>
- The column names that will be used while emitting the CSV records
SourceAzureBlobStorageConfigurationStreamFormatDocumentFileTypeFormatExperimental, SourceAzureBlobStorageConfigurationStreamFormatDocumentFileTypeFormatExperimentalArgs                        
- Processing
SourceAzure Blob Storage Configuration Stream Format Document File Type Format Experimental Processing 
- Processing configuration
- SkipUnprocessable boolFiles 
- If true, skip files that cannot be parsed and pass the error message along as the absourcefileparse_error field. If false, fail the sync. Default: true
- Strategy string
- The strategy used to parse documents. fastextracts text directly from the document which doesn't work for all files.ocr_onlyis more reliable, but slower.hi_resis the most reliable, but requires an API key and a hosted instance of unstructured and can't be used with local mode. See the unstructured.io documentation for more details: https://unstructured-io.github.io/unstructured/core/partition.html#partition-pdf. Default: "auto"; must be one of ["auto", "fast", "ocronly", "hires"]
- Processing
SourceAzure Blob Storage Configuration Stream Format Document File Type Format Experimental Processing 
- Processing configuration
- SkipUnprocessable boolFiles 
- If true, skip files that cannot be parsed and pass the error message along as the absourcefileparse_error field. If false, fail the sync. Default: true
- Strategy string
- The strategy used to parse documents. fastextracts text directly from the document which doesn't work for all files.ocr_onlyis more reliable, but slower.hi_resis the most reliable, but requires an API key and a hosted instance of unstructured and can't be used with local mode. See the unstructured.io documentation for more details: https://unstructured-io.github.io/unstructured/core/partition.html#partition-pdf. Default: "auto"; must be one of ["auto", "fast", "ocronly", "hires"]
- processing
SourceAzure Blob Storage Configuration Stream Format Document File Type Format Experimental Processing 
- Processing configuration
- skipUnprocessable BooleanFiles 
- If true, skip files that cannot be parsed and pass the error message along as the absourcefileparse_error field. If false, fail the sync. Default: true
- strategy String
- The strategy used to parse documents. fastextracts text directly from the document which doesn't work for all files.ocr_onlyis more reliable, but slower.hi_resis the most reliable, but requires an API key and a hosted instance of unstructured and can't be used with local mode. See the unstructured.io documentation for more details: https://unstructured-io.github.io/unstructured/core/partition.html#partition-pdf. Default: "auto"; must be one of ["auto", "fast", "ocronly", "hires"]
- processing
SourceAzure Blob Storage Configuration Stream Format Document File Type Format Experimental Processing 
- Processing configuration
- skipUnprocessable booleanFiles 
- If true, skip files that cannot be parsed and pass the error message along as the absourcefileparse_error field. If false, fail the sync. Default: true
- strategy string
- The strategy used to parse documents. fastextracts text directly from the document which doesn't work for all files.ocr_onlyis more reliable, but slower.hi_resis the most reliable, but requires an API key and a hosted instance of unstructured and can't be used with local mode. See the unstructured.io documentation for more details: https://unstructured-io.github.io/unstructured/core/partition.html#partition-pdf. Default: "auto"; must be one of ["auto", "fast", "ocronly", "hires"]
- processing
SourceAzure Blob Storage Configuration Stream Format Document File Type Format Experimental Processing 
- Processing configuration
- skip_unprocessable_ boolfiles 
- If true, skip files that cannot be parsed and pass the error message along as the absourcefileparse_error field. If false, fail the sync. Default: true
- strategy str
- The strategy used to parse documents. fastextracts text directly from the document which doesn't work for all files.ocr_onlyis more reliable, but slower.hi_resis the most reliable, but requires an API key and a hosted instance of unstructured and can't be used with local mode. See the unstructured.io documentation for more details: https://unstructured-io.github.io/unstructured/core/partition.html#partition-pdf. Default: "auto"; must be one of ["auto", "fast", "ocronly", "hires"]
- processing Property Map
- Processing configuration
- skipUnprocessable BooleanFiles 
- If true, skip files that cannot be parsed and pass the error message along as the absourcefileparse_error field. If false, fail the sync. Default: true
- strategy String
- The strategy used to parse documents. fastextracts text directly from the document which doesn't work for all files.ocr_onlyis more reliable, but slower.hi_resis the most reliable, but requires an API key and a hosted instance of unstructured and can't be used with local mode. See the unstructured.io documentation for more details: https://unstructured-io.github.io/unstructured/core/partition.html#partition-pdf. Default: "auto"; must be one of ["auto", "fast", "ocronly", "hires"]
SourceAzureBlobStorageConfigurationStreamFormatDocumentFileTypeFormatExperimentalProcessing, SourceAzureBlobStorageConfigurationStreamFormatDocumentFileTypeFormatExperimentalProcessingArgs                          
- Local
SourceAzure Blob Storage Configuration Stream Format Document File Type Format Experimental Processing Local 
- Process files locally, supporting fastandocrmodes. This is the default option.
- Local
SourceAzure Blob Storage Configuration Stream Format Document File Type Format Experimental Processing Local 
- Process files locally, supporting fastandocrmodes. This is the default option.
- local
SourceAzure Blob Storage Configuration Stream Format Document File Type Format Experimental Processing Local 
- Process files locally, supporting fastandocrmodes. This is the default option.
- local
SourceAzure Blob Storage Configuration Stream Format Document File Type Format Experimental Processing Local 
- Process files locally, supporting fastandocrmodes. This is the default option.
- local
SourceAzure Blob Storage Configuration Stream Format Document File Type Format Experimental Processing Local 
- Process files locally, supporting fastandocrmodes. This is the default option.
- local Property Map
- Process files locally, supporting fastandocrmodes. This is the default option.
SourceAzureBlobStorageConfigurationStreamFormatParquetFormat, SourceAzureBlobStorageConfigurationStreamFormatParquetFormatArgs                  
- DecimalAs boolFloat 
- Whether to convert decimal fields to floats. There is a loss of precision when converting decimals to floats, so this is not recommended. Default: false
- DecimalAs boolFloat 
- Whether to convert decimal fields to floats. There is a loss of precision when converting decimals to floats, so this is not recommended. Default: false
- decimalAs BooleanFloat 
- Whether to convert decimal fields to floats. There is a loss of precision when converting decimals to floats, so this is not recommended. Default: false
- decimalAs booleanFloat 
- Whether to convert decimal fields to floats. There is a loss of precision when converting decimals to floats, so this is not recommended. Default: false
- decimal_as_ boolfloat 
- Whether to convert decimal fields to floats. There is a loss of precision when converting decimals to floats, so this is not recommended. Default: false
- decimalAs BooleanFloat 
- Whether to convert decimal fields to floats. There is a loss of precision when converting decimals to floats, so this is not recommended. Default: false
Import
$ pulumi import airbyte:index/sourceAzureBlobStorage:SourceAzureBlobStorage my_airbyte_source_azure_blob_storage ""
To learn more about importing existing cloud resources, see Importing resources.
Package Details
- Repository
- airbyte airbytehq/terraform-provider-airbyte
- License
- Notes
- This Pulumi package is based on the airbyteTerraform Provider.