Select a staging area for the data. Staging areas can be created through Snowflake using the CREATE STAGE command. Internal stages can be set up this way to store staged data within Snowflake. Selecting [Custom] will avail the user of properties to specify a custom staging area on S3. Users can add a fully qualified stage by typing the stage name. This should follow the format databaseName.schemaName.stageName Select the authentication method. Users can choose either:
- Credentials: Uses AWS security credentials.
- Storage Integration: Use a Snowflake storage integration. A storage integration is a Snowflake object that stores a generated identity and access management (IAM) entity for your external cloud storage, along with an optional set of permitted or blocked storage locations (Amazon S3, Google Cloud Storage, or Microsoft Azure Blob Storage). More information can be found at CREATE STORAGE INTEGRATION.
Select the storage integration. Storage integrations are required to permit Snowflake to read data from and write to a cloud storage location. Integrations must be set up in advance of selecting them. Storage integrations can be configured to support Amazon S3, Google Cloud Storage, or Microsoft Azure Blob Storage, regardless of the cloud provider that hosts your Snowflake account.
To retrieve the intended files, use the file explorer to enter the container path where the S3 bucket is located, or select from the list of S3 buckets.This must have the format S3://<bucket>/<path>.
A string that will partially match all file paths and names that are to be included in the load. Defaults to .* indicating all files within the S3 Object Prefix. This property is a pattern on the complete path of the file, and is not just relative to the directory configured in the S3 Object Prefix property.The subfolder containing the object to load must be included here.
Decide how the files are encrypted inside the S3 bucket. This property is available when using an existing Amazon S3 location for staging. The ID of the KMS encryption key you have chosen to use in the Encryption property.
The Snowflake warehouse used to run the queries. The special value [Environment Default] uses the warehouse defined in the environment. Read Overview of Warehouses to learn more. The Snowflake schema. The special value [Environment Default] uses the schema defined in the environment. Read Database, Schema, and Share DDL to learn more. Select an existing table to load data into. The tables available for selection depend on the chosen schema.
Choose the columns to load. If you leave this parameter empty, all columns will be loaded.
Select a pre-made file format that will automatically set many of the S3 Load component properties. These formats can be created through the Create File Format component.Users can add a fully qualified format by typing the format name. This should read as databaseName.schemaName.formatName Select the type of data to load. Available data types are: AVRO, CSV, JSON, ORC, PARQUET, and XML. Some file types may require additional formatting—this is explained in the Snowflake documentation.Component properties will change to reflect the selected file type. Click one of the tabs below for properties applicable to that file type. AVRO
CSV
JSON
ORC
PARQUET
XML
Select the compression method if you wish to compress your data. If you do not wish to compress at all, select NONE. The default setting is AUTO.
Specify one or more strings (one string per row of the table) to convert to NULL values. When one of these strings is encountered in the file, it is replaced with an SQL NULL value for that field in the loaded table. Click + to add a string.
When True, removes whitespace from fields. Default setting is False.
Select the compression method if you wish to compress your data. If you do not wish to compress at all, select NONE. The default setting is AUTO.
Input a delimiter for records. This can be one or more single-byte or multibyte characters that separate records in an input file.Accepted values include: leaving the field empty; a newline character \ or its hex equivalent 0x0a; a carriage return \\r or its hex equivalent 0x0d. Also accepts a value of NONE.If you set the Skip Header to a value such as 1, then you should use a record delimiter that includes a line feed or carriage return, such as \ or \\r. Otherwise, your entire file will be interpreted as the header row, and no data will be loaded.The specified delimiter must be a valid UTF-8 character and not a random sequence of bytes.Do not specify characters used for other file type options such as Escape or Escape Unenclosed Field.The default (if the field is left blank) is a newline character.
Input a delimiter for fields. This can be one or more single-byte or multibyte characters that separate fields in an input file.Accepted characters include common escape sequences, octal values (prefixed by \), or hex values (prefixed by 0x). Also accepts a value of NONE.This delimiter is limited to a maximum of 20 characters.While multi-character delimiters are supported, the field delimiter cannot be a substring of the record delimiter, and vice versa. For example, if the field delimiter is “aa”, the record delimiter cannot be “aabb”.The specified delimiter must be a valid UTF-8 character and not a random sequence of bytes.Do not specify characters used for other file type options such as Escape or Escape Unenclosed Field.The Default setting is a comma: ,.
Specify the number of rows to skip. The default is 0.If Skip Header is used, the value of the record delimiter will not be used to determine where the header line is. Instead, the specified number of CRLF will be skipped. For example, if the value of Skip Header = 1, skips to the first CRLF that it finds. If you have set the Field Delimiter property to be a single character without a CRLF, then skips to the end of the file (treating the entire file as a header).
When True, ignores blank lines that only contain a line feed in a data file and does not try to load them. Default setting is False.
Define the format of date values in the data files to be loaded. If a value is not specified or is AUTO, the value for the DATE_INPUT_FORMAT session parameter is used. The default setting is AUTO.
Define the format of time values in the data files to be loaded. If a value is not specified or is AUTO, the value for the TIME_INPUT_FORMAT session parameter is used. The default setting is AUTO.
Define the format of timestamp values in the data files to be loaded. If a value is not specified or is AUTO, the value for the TIMESTAMP_INPUT_FORMAT session parameter is used.
Specify a single character to be used as the escape character for field values that are enclosed. Default is NONE.
Specify a single character to be used as the escape character for unenclosed field values only. Default is \\. If you have set a value in the property Field Optionally Enclosed, all fields will become enclosed, rendering the Escape Unenclosed Field property redundant, in which case, it will be ignored.
When True, removes whitespace from fields. Default setting is False.
Field Optionally Enclosed
Specify a character used to enclose strings. The value can be NONE, single quote character ', or double quote character ". To use the single quote character, use the octal or hex representation 0x27 or the double single-quoted escape ''. Default is NONE.When a field contains one of these characters, escape the field using the same character. For example, to escape a string like this: 1 “2” 3, use double quotation to escape, like this: 1 ""2"" 3.
Specify one or more strings (one string per row of the table) to convert to NULL values. When one of these strings is encountered in the file, it is replaced with an SQL NULL value for that field in the loaded table. Click + to add a string.
Error On Column Count Mismatch
When True, generates an error if the number of delimited columns in an input file does not match the number of columns in the corresponding table. When False (default), an error is not generated and the load continues. If the file is successfully loaded in this case:
- Where the input file contains records with more fields than columns in the table, the matching fields are loaded in order of occurrence in the file, and the remaining fields are not loaded.
- Where the input file contains records with fewer fields than columns in the table, the non-matching columns in the table are loaded with NULL values.
When True, inserts NULL values for empty fields in an input file. This is the default setting.
Replace Invalid Characters
Snowflake replaces invalid UTF-8 characters with the Unicode replacement character. When False (default), the load operation produces an error when invalid UTF-8 character encoding is detected.
Select the string that specifies the character set of the source data when loading data into a table. Refer to the Snowflake documentation for more information.
Select the compression method if you wish to compress your data. If you do not wish to compress at all, select NONE. The default setting is AUTO.
When True, enables the parsing of octal values. Default setting is False.
When True, allows duplicate object field names. Default setting is False.
When True, instructs the JSON parser to remove outer brackets. Default setting is False.
When True, instructs the JSON parser to remove any object fields or array elements containing NULL values. Default setting is False.
When True, replaces any invalid UTF-8 sequences with Unicode characters. When False (default), UTF-8 errors will not produce an error in the pipeline run.
Specify one or more strings (one string per row of the table) to convert to NULL values. When one of these strings is encountered in the file, it is replaced with an SQL NULL value for that field in the loaded table. Click + to add a string.
When True, removes whitespace from fields. Default setting is False.
Specify one or more strings (one string per row of the table) to convert to NULL values. When one of these strings is encountered in the file, it is replaced with an SQL NULL value for that field in the loaded table. Click + to add a string.
When True, removes whitespace from fields. Default setting is False.
Select the compression method if you wish to compress your data. If you do not wish to compress at all, select NONE. The default setting is AUTO.
Specify one or more strings (one string per row of the table) to convert to NULL values. When one of these strings is encountered in the file, it is replaced with an SQL NULL value for that field in the loaded table. Click + to add a string.
When True, removes whitespace from fields. Default setting is False.
Select the compression method if you wish to compress your data. If you do not wish to compress at all, select NONE. The default setting is AUTO.
When True, replaces any invalid UTF-8 sequences with Unicode characters. When False (default), UTF-8 errors will not produce an error in the pipeline run.
When True, the XML parser preserves leading and trailing spaces in element content. Default setting is False.
When True, the XML parser strips out any outer XML elements, exposing second-level elements as separate documents. Default setting is False.
When True, the XML parser will not recognize Snowflake semi-structured data tags. Default setting is False.
When True, the XML parser will disable automatic conversion of numeric and Boolean values from text to native representations. Default setting is False.
Decide how to proceed upon an error.
- Abort Statement: Aborts the load if any error is encountered. This is the default setting.
- Continue: Continue loading the file.
- Skip File: Skip file if any errors are encountered in the file.
- Skip File When n Errors: Skip file when the number of errors in the file is equal to or greater than the specified number in the next property, n.
- Skip File When n% Errors: Skip file when the percentage of errors in the file exceeds the specified percentage of n.
Specify the number of errors or the percentage of errors required to skip the file. This parameter only accepts integer characters. % is not accepted. Specify percentages as a number only.
Specify the maximum size, in bytes, of data to be loaded for a given COPY statement. If the maximum is exceeded, the COPY operation discontinues loading files. For more information, refer to the Snowflake documentation.
When True, purges data files after the data is successfully loaded. Default setting is False.
Specify whether to load semi-structured data into columns in the target table that match corresponding columns represented in the data.
- Case Insensitive: Load semi-structured data into columns in the target table that match corresponding columns represented in the data. Column names should be case-insensitive.
- Case Sensitive: Load semi-structured data into columns in the target table that match corresponding columns represented in the data. Column names should be case-sensitive.
- None: The COPY operation loads the semi-structured data into a variant column or, if a query is included in the COPY statement, transforms the data.
When True, strings are automatically truncated to the target column length. When False (default), the COPY statement produces an error if a loaded string exceeds the target column length.
When True, loads all files, regardless of whether they have been loaded previously and haven’t changed since they were loaded. Default setting is False.When set to True, this option reloads files and can lead to duplicated data in a table.
Snowflake metadata columns available to include in the load. For more information, read Querying Metadata for Staged Files. This property is only available when an external stage is selected. To retrieve the intended files, use the file explorer to enter the container path where the S3 bucket is located, or select from the list of S3 buckets.This must have the format S3://<bucket>/<path>.
Files in the specified Location will only be loaded if their names match the pattern you specify here. You can use wildcards in the pattern. Enter .* to match all files in the location. This property is mandatory.
Select a Databricks Unity Catalog. The special value [Environment Default] uses the catalog defined in the environment. Selecting a catalog will determine which databases are available in the next parameter. The Databricks schema. The special value [Environment Default] uses the schema defined in the environment. Read Create and manage schemas to learn more. Select an existing table to load data into. The tables available for selection depend on the chosen schema (database).
Choose the columns to load. If you leave this parameter empty, all columns will be loaded.
Select the type of data to load. Available data types are: CSV, JSON, PARQUET, and AVRO.Component properties will change to reflect the selected file type. Click one of the tabs below for properties applicable to that file type.Select Yes to use the first line of the file as column names. If not specified, the default is No.
Enter the delimiter character used to separate fields in the CSV file. This can be one or more single-byte or multibyte characters that separate fields in an input file. If none is specified, the default is a comma.Accepted characters include common escape sequences, octal values (prefixed by \\), or hex values (prefixed by 0x). This delimiter is limited to a maximum of 20 characters. The specified delimiter must be a valid UTF-8 character and not a random sequence of bytes.
Describe the format of date values in the data files to be loaded. If none is specified, the default is yyyy-MM-dd.
Describe the format of timestamp values in the data files to be loaded. If none is specified, the default is yyyy-MM-dd'T'HH:mm:ss.[SSS][XXX].
The encoding type to use when decoding the input files. If none is specified, the default is UTF-8.
Ignore Leading Whitespace
When Yes, removes whitespace from the beginning of fields. If not specified, the default is No.
Ignore Trailing Whitespace
When Yes, removes whitespace from the end of fields. If not specified, the default is No.
If Yes, will attempt to determine the input schema automatically from the data contained in the CSV file. If not specified, the default is No.
If Yes, will parse records which may span multiple lines. If not specified, the default is No.
Sets the string representation of a null value. If not specified, the default value is an empty string.
Sets the string representation of an empty value. If not specified, the default value is an empty string.
If Yes, partition inference is disabled. If not specified, the default is No. To control which files are loaded, use the Pattern property instead.
If Yes, files are loaded regardless of whether they’ve been loaded before. If not specified, the default is No.
Describe the format of date values in the data files to be loaded. If none is specified, the default is yyyy-MM-dd.
Describe the format of timestamp values in the data files to be loaded. If none is specified, the default is yyyy-MM-dd'T'HH:mm:ss.[SSS][XXX].
The encoding type to use when decoding the input files. If none is specified, the default is UTF-8.
If Yes, will parse records which may span multiple lines. If not specified, the default is No.
If Yes, primitive data types are interpreted as strings in JSON files. If not specified, the default is No.
If Yes, all floating-point values will be treated as a decimal type in JSON files. If not specified, the default is No.
If Yes, will allow JAVA/C++ comments in JSON records. If not specified, the default is No.
Allow Unquoted Field Names
If Yes, will allow unquoted JSON field names. If not specified, the default is No.
If Yes, will allow single quotes in addition to double quotes in JSON records. If not specified, the default is No.
Allow Numeric Leading Zeros
If Yes, will allow leading zeros in numbers in JSON records. If not specified, the default is No.
Allow Backslash Escaping Any Character
If Yes, will allow the quoting of all characters using the backslash quoting mechanism in JSON records. If not specified, the default is No.
Allow Unquoted Control Chars
If Yes, will allow JSON strings to include unquoted control characters in JSON records. If not specified, the default is No.
If Yes, will ignore columns that contain only null values in JSON records. If not specified, the default is No.
If Yes, partition inference is disabled. If not specified, the default is No. To control which files are loaded, use the Pattern property instead.
If Yes, files are loaded regardless of whether they’ve been loaded before. If not specified, the default is No.
If Yes, will merge schema from all PARQUET and AVRO part-files. If not specified, the default is No.
If Yes, partition inference is disabled. If not specified, the default is No. To control which files are loaded, use the Pattern property instead.
If Yes, files are loaded regardless of whether they’ve been loaded before. If not specified, the default is No.
If Yes, will merge schema from all PARQUET and AVRO part-files. If not specified, the default is No.
If Yes, partition inference is disabled. If not specified, the default is No. To control which files are loaded, use the Pattern property instead.
If Yes, files are loaded regardless of whether they’ve been loaded before. If not specified, the default is No.
Select the table schema. The special value [Environment Default] uses the schema defined in the environment. For more information on using multiple schemas, read Schemas. Select an existing table to load data into. The selected target table is where data will be loaded into from the S3 object.When selecting a target table with the Redshift SUPER data type, the S3 Load component will only offer the following data file types: JSON, ORC, and PARQUET. However, if you select to include columns of data file types other than SUPER in the Load Columns property, all data file types will be available for selection.
Select which of the target table’s columns should be loaded.
To retrieve the intended files, use the file explorer to enter the container path where the S3 bucket is located, or select from the list of S3 buckets.This must have the format S3://<bucket>/<path>.
All files that begin with this prefix will be included in the load into the target table.
Select an IAM role Amazon Resource Name (ARN) that is already attached to your Redshift cluster, and that has the necessary permissions to access S3.This setting is optional, since without this style of setup, the credentials of the environment will be used.Read the Redshift documentation for more information about using a role ARN with Redshift. The type of expected data to load. Some may require additional formatting, explained in the Amazon documentation.Available options are: Avro, CSV, Delimited, Fixed Width, JSON, ORC, PARQUET.Component properties will change to reflect the choice made here and give options based on the specific data file type. Fixed width only. Loads the data from a file where each column width is a fixed length, rather than columns being separated by a delimiter. For more information, read the AWS documentation.To use variables in this field, type the name of the variable prefixed by the dollar symbol and surrounded by { } brackets, as follows: ${variable}. Once you type ${, a drop-down list of autocompleted suggested variables will appear. This list updates as you type; for example, if you type ${date, functions and variables containing date will be listed. (JSON only) Defaults to “auto”, which will work for the majority of JSON files if the fields match the table field names. More information about JSON file types can be found in the AWS documentation. (CSV only) Specify a delimiting character to separate columns. The default character is a comma: ,A [TAB] character can be specified using ”\ ”.
(CSV only) Specify the character to be used as the quote character when using the CSV file type.
Select whether to load data from the S3 objects into an IDENTITY column. Read the Redshift documentation for more information. Select the Amazon S3 region that hosts the selected S3 bucket. This setting is optional and can be left as “None” if the bucket is in the same region as your Redshift cluster.
Specify whether the input is compressed in any of the following formats: BZip2, gzip, LZop, Zstd, or not at all (None).
Select the encoding format the data is in. The default for this setting is UTF-8.
Replace Invalid Characters
Specify a single character that will replace any invalid unicode characters in the data. The default is: ?.
(Delimited, Fixed Width only) When “Yes”, removes surrounding quotation marks from strings in the incoming data. All characters within the quotation marks, including delimiters, are retained.If a string has a beginning single or double quotation mark but no corresponding ending mark, the COPY command fails to load that row and returns an error.The default setting is “No”.For more information, read the AWS documentation. The maximum number of individual parsing errors permitted before the load will fail. Values up to this will be substituted as NULL values. The Amazon default is 1000.
Defaults to ‘auto’. Users can specify their preferred date format manually if they wish. For more information on date formats, read the AWS documentation. Defaults to ‘auto’. Users can specify their preferred time format manually if they wish. For more information on time formats, read the AWS documentation. Specify the number of rows at the top of the file to ignore. The default is 0 (no rows ignored).
- Yes: Invalid dates such as ‘45-65-2020’ are not considered an error, but will be loaded as NULL values. This is the default setting.
- No: Invalid dates will return an error.
(CSV only) When “Yes”, ignores blank lines that only contain a line feed in a data file and does not try to load them. This is the default setting.
- Yes: any instance of data in the input file that is too long to fit into the specified target column width will be truncated to fit, instead of causing an error.
- No: any data in the input file that is too long to fit into the specified target column width will cause an error. This is the default setting.
When “Yes”, allows data files to be loaded when contiguous columns are missing at the end of some records. The missing columns are filled with either zero-length strings or NULLs, as appropriate for the data types of the columns in question. This is the default setting.
- Yes: Removes the trailing white space characters from a VARCHAR string. This property only applies to columns with a VARCHAR data type.
- No: does not remove trailing white space characters from the input data.
Loads fields that match the specified NULL string. The default is \\N with an additional \ at the start to escape. Case-sensitive.For more information, read the AWS documentation. When “Yes”, empty columns in the input file will become NULL values. This is the default setting.
When “Yes”, loads blank columns, which consist of only whitespace characters, as NULL. This is the default setting.This option applies only to CHAR and VARCHAR columns. Blank fields for other data types, such as INT, are always loaded with NULL. For example, a string that contains three space characters in succession (and no other characters) is loaded as a NULL. The default behavior, without this option, is to load the space characters as is.
When “On”, compression encodings are automatically applied during a COPY command. This is the default setting.This is usually a good idea to optimize the compression used when storing the data.
When “On”, governs automatic computation and refreshing of optimizer statistics at the end of a successful COPY command. This is the default setting.
(Delimited only) When “Yes”, the backslash character \ in input data is treated as an escape character. The character that immediately follows the backslash character is loaded into the table as part of the current column value, even if it is a character that normally serves a special purpose. For example, you can use this parameter to escape the delimiter character, a quotation mark, an embedded newline character, or the escape character itself when any of these characters is a legitimate part of a column value.The default setting is “No”.
When “Yes”, any decimals are rounded to fit into the column in any instance where the number of decimal places in the input is larger than defined for the target column. This is the default setting.
When “Yes”, the given object prefix is that of a manifest file. The default setting is “No”.For more information, read the AWS documentation.