Skip to main content
The Volume to Delta Table component lets users transfer data from a pre-existing volume in Databricks into a Delta Lake table without replacing or deleting any existing data.
If you have selected hive_metastore as your default catalog, limitations can’t be used, volumes can’t be created, and the feature won’t be available to you. This component only supports Unity Catalog.
You must use a SQL data warehouse or a cluster running Databricks Runtime 13.3 LTS or above, otherwise unexpected behavior in the component can occur.

Properties

Name
string
required
A human-readable name for the component.
Source
drop-down
required
The file location to load the data from.Files in this location must have the specified FILEFORMAT. Accepted encryption options are TYPE = 'AWS_SSE_C', and MASTER_KEY for AWS S3.
Pattern
string
required
Files in the specified location will only be loaded if their names match the pattern you specify here. You can use wildcards in the pattern. Enter .* to match all files in the location.
Catalog
drop-down
required
Select a Databricks Unity Catalog. The special value [Environment Default] uses the catalog defined in the environment. Selecting a catalog will determine which databases are available in the next parameter.
If you have selected hive_metastore as your default catalog, limitations can’t be used, volumes can’t be created, and the feature won’t be available to you. This component only supports Unity Catalog.
You must have appropriate permissions and access rights to the source volume and destination Delta table.
Schema (Database)
drop-down
required
The Databricks schema. The special value [Environment Default] uses the schema defined in the environment. Read Create and manage schemas to learn more.
Table
string
required
The name of the Delta table. This table will be recreated and will drop any existing table of the same name.
Load Columns
dual listbox
Choose the columns to load. If you leave this parameter empty, all columns will be loaded.
File Type
drop-down
required
The format of the source files to load. Available file types are CSV, JSON, PARQUET, and AVRO.Component properties will change to reflect the selected file type. Click one of the tabs below for properties applicable to that file type.
Header
boolean
Select Yes to use the first line of the file as column names. If not specified, the default is No.
Field Delimiter
string
required
Enter the delimiter character used to separate fields in the CSV file. This can be one or more single-byte or multibyte characters that separate fields in an input file. If none is specified, the default is a comma.Accepted characters include common escape sequences, octal values (prefixed by \\), or hex values (prefixed by 0x). This delimiter is limited to a maximum of 20 characters. The specified delimiter must be a valid UTF-8 character and not a random sequence of bytes.
A TAB character can be specified as "".
Date Format
string
Manually set a date format in the data files to be loaded. If none is specified, the default is yyyy-MM-dd.
Timestamp Format
string
Manually set a timestamp format in the CSV files to be loaded. If none is specified, the default is yyyy-MM-dd'T'HH:mm:ss.[SSS][XXX].
Encoding Type
string
The encoding type to use when decoding the CSV files. If none is specified, the default is UTF-8.
Ignore Leading Whitespace
boolean
When Yes, skips any leading whitespaces. If not specified, the default is No.
Ignore Trailing Whitespace
boolean
When Yes, skips any trailing whitespaces. If not specified, the default is No.
Infer Schema
boolean
If Yes, will attempt to determine the input schema automatically from the data contained in the CSV file. If not specified, the default is No.
Multi Line
boolean
If Yes, will parse records which may span multiple lines. If not specified, the default is No.
Null Value
string
Sets the string representation of a null value. If not specified, the default value is an empty string.
Empty Value
string
Sets the string representation of an empty value. If not specified, the default value is an empty string.
Recursive File Lookup
boolean
If Yes, partition inference is disabled. If not specified, the default is No. To control which files are loaded, use the Pattern property instead.
Force Load
boolean
If Yes, files are loaded regardless of whether they’ve been loaded before. If not specified, the default is No.