Skip to main content
This page describes how to configure the connector component as part of a data pipeline within . The component uses the Connect and Configure parameters to create a table of data, which is then stored in your preferred storage location (Snowflake, Databricks, Amazon Redshift, or cloud storage). You do not need to use the Create Table component when using this connector, as the component will create a new table or replace an existing table for you using the Destination parameters you define. The connector is a Flex connector. In , Flex connectors let you connect to a curated set of endpoints to load data. You can use connector in its preconfigured state, or you can edit the connector by adding or amending available endpoints as per your use case. You can edit Flex connectors in the Custom Connector user interface. For detailed information about authentication, specific endpoint parameters, pagination, and more aspects of the API, read the API documentation.

Video example

Watch our video about using Flex connector: YouTube.

Properties

Reference material is provided below for the Connect, Configure, Destination, and Advanced Settings properties.
Name
string
required
A human-readable name for the component.

Connect

Data Source
drop-down
required
The data source to load data from in this pipeline. The drop-down menu lists the API endpoints available in the connector. For detailed information about specific endpoints, read the API documentation.
EndpointMethodReference
List All ProjectsGETRetrieves a list of projects
List All EnvironmentsGETRetrieves a list of environments
List All Published PipelinesGETRetrieves a list of published pipelines
Execute Published PipelinePOSTExecute a pipeline
Get Pipeline StatusGETRetrieves the status of an executed pipeline
Get Pipeline Steps StatusGETRetrieves a single execution of a pipeline triggered from a step of another pipeline
Get ConsumptionGETRetrieves a breakdown of your credit usage for flat-rated products
Get Matillion ETL Users ConsumptionGETRetrieves a breakdown of your credit usage for Matillion ETL users
List All SchedulesGETRetrieves all schedules for a project
Create SchedulePOSTCreate a new schedule for a specific project, pipeline and environment
Get ScheduleGETRetrieves schedule details for a given schedule id
List ArtifactsGETRetrieves a list of artifacts in an environment
Get ArtifactGETRetrieves an artifact by a version name
Promote ArtifactPOSTPromote artifacts from one environment to another
Pipeline ExecutionsGETRetrieves a list of summarized pipeline executions
List Custom ConnectorsGETRetrieves custom connector profiles for the requesting account
List Flex ConnectorsGETRetrieves flex connector profiles
List All Secret ReferencesGETRetrieves a list of all secret references
Create Secret ReferencePOSTCreate a secret reference
List All AgentsGETLists all agents, including summary details for each agent
Create AgentPOSTCreates a new agent with the specified configuration details
Get Agent DetailsGETRetrieves details for a specific agent
Trigger Agent CommandPOSTTriggers a specific command on an agent
Get Agent Client CredentialsGETRetrieves the Client ID and Client secret for the agent
Perform Action On Agent CredentialsPOSTPerform actions on agent credentials, such as refreshing them
Query Audit EventsGETRetrieves audit events for a given time range
Get Lineage EventsGETRetrieves OpenLineage events for a specified time period
Authentication Type
drop-down
required
The authentication method to authorize access to your data. Currently supports OAuth 2.0 Client Credentials. Read Authenticating to the API to learn more.
Authentication
string
required
Select your authentication profile.To create a new profile, read OAuth client credentials.

Configure

URI Parameters
column editor
required
  • Parameter Name: The name of a URI parameter.
  • Parameter Value: The value of the corresponding parameter.
Required parameterEndpointsDescription
serverAll endpointsEnter eu1 or us1, depending on the region of your account. To find your account’s region, click the Profile & Account icon on the left side of the page.
api_versionAll endpointsEnter v1.
projectIdList All Environments, List All Published Pipelines, Execute Published Pipeline, Get Pipeline Status, Get Pipeline Steps Status, List All Schedules, Create Schedule, List Artifacts, Get Artifact, Promote Artifact, List All Secret References, Create Secret ReferenceprojectId is unique to every project. Retrieve this value by using the List All Projects endpoint.
pipelineExecutionIdGet Pipeline Status, Get Pipeline Steps StatuspipelineExecutionId is unique to every pipeline. Retrieve this value by using the Execute Published Pipeline endpoint.
secretReferenceNameCreate Secret ReferenceThe name of secret reference. This can be found in the Secret definitions tab in , under the Name column.
agentIdGet Agent Details, Trigger Agent Command, Get Agent Client Credentials, Perform Action On Agent CredentialsThe ID of the agent to retrieve. This can be found in the left navigation, under Agents & Instances, then click Agents. Select the intended agent, and click the Parameters tab.
Query Parameters
column editor
required
  • Parameter Name: The name of a query parameter.
  • Parameter Value: The value of the corresponding parameter.
Required parameterEndpointsDescription
sizeList All Projects, List All Environments, List All Published Pipelines, Get Pipeline Steps Status, List All Schedules, List Artifacts, List All Secret References, List All Agents, Query Audit Events, Get Lineage EventsEnter the number of records per page, ranging from 1 to 100.
environmentNameList All Published Pipelines, List Artifacts, Get ArtifactEnter the environment name. For example, test-environment-1.
consumedFromGet ConsumptionEnter the start date for the results. This value is inclusive, meaning results from this date onward are included. For example, 2024-11-01.
consumedBeforeGet ConsumptionEnter the end date for the results. This value is exclusive, meaning it includes only results occurring before (but not on) this date. For example, 2024-12-01.
consumedFromGet Matillion ETL Users ConsumptionEnter the start date and time for the results. This value is inclusive, meaning results from this date and time onward are included. For example, 2024-07-01T00:00:00.123Z.
consumedBeforeGet Matillion ETL Users ConsumptionEnter the end date and time for the results. This value is exclusive, meaning it includes only results occurring before (but not on) this date and time. For example,2024-07-31T00:00:00.123Z.
versionNameGet ArtifactThe Version name when you Push local changes to the remote repository in . For more information, read Git push local changes.
limitPipeline ExecutionsEnter the maximum number of results to return. The default value is set to 25.
fromQuery Audit EventsEnter the earliest date and time of audit events to retrieve. The date time format must be in ISO 8601 format, for example: 2025-02-20T07:15:15.000-01:00.
toQuery Audit EventsEnter the latest date and time of audit events to retrieve. The date time format must be in ISO 8601 format, for example: 2025-02-21T07:15:15.000-01:00.
generatedFromGet Lineage EventsInclude events generated on or after this date time. The value must be earlier than generatedBefore.
generatedBeforeGet Lineage EventsInclude events generated up to, but not including, this date time. The value must be later than generatedFrom.
pageGet Lineage EventsThe page number to use for pagination, starting at 0. Must be 0 or greater.
The Get Matillion ETL Users endpoint provides information about the number of credits charged for Matillion ETL users, and identifies which users contributed to those charges. Matillion ETL users are billed based on monthly active unique users, so ensure that the consumedFrom and consumedBefore parameters correspond to the timeframe of a single monthly invoice.
Header Parameters
column editor
required
  • Parameter Name: The name of a header parameter.
  • Parameter Value: The value of the corresponding parameter.
Required parameterEndpointsDescription
Content-TypeExecute Published Pipeline, Get Pipeline Status, Get Pipeline Steps Status, Create Schedule, Promote Artifact, Create Secret Reference, List All Agents, Create Agent, Get Agent Details, Trigger Agent Command, Get Agent Client Credentials, Perform Action On Agent Credentials, Query Audit Events, Get Lineage EventsEnter application/json.
acceptExecute Published Pipeline, Get Pipeline Status, Get Pipeline Steps StatusEnter application/json.
Post Body
JSON
A JSON body to include as part of a POST request. Use Custom Connector to test your endpoints work as expected before moving to pipelines.You should also consult the developer documentation for the API you’re connecting to—as the developer portal may provide additional information about endpoints and requests.For the Execute Published Pipeline endpoint, include the following POST Body. This example demonstrates a POST Body used to execute a pipeline:
{
  "pipelineName": "test-pipeline",
  "environmentName": "test-environment"
}
For the Create Schedule endpoint, include the following POST Body. This example demonstrates a POST Body used to create a schedule:
{
  "pipeline": {
    "pipelineName": "pipeline-name",
    "environmentName": "environment-name"
  },
  "schedule": {
    "cronExpression": "0 * * ? * * *",
    "cronTimezone": "Europe/Dublin",
    "effectiveFrom": "2022-05-19T12:37:44Z",
    "name": "schedule-name",
    "scheduleEnabled": false
  }
}
For the Promote Artifact endpoint, include the following POST Body. This example demonstrates a POST Body used to promote an artifact:
{
  "sourceEnvironmentName": "source-environment-name",
  "targetEnvironmentName": "environment-name",
  "versionName": "version-name"
}
For the Create Secret Reference endpoint, include the following POST Body. This example demonstrates a POST Body used to create a secret for an AWS agent:
{
  "agentId": "agent-id",
  "agentType": "AWS",
  "description": "My secret reference",
  "secretReferenceType": "PASSWORD",
  "secretKey": "aws-secret-key",
  "secretName": "aws-secret-name"
}
For the Create Agent endpoint, include the following POST Body. This example demonstrates a POST Body used to create a new AWS agent:
{
  "agentType": "data_productivity_cloud",
  "cloudProvider": "aws",
  "deployment": "fargate",
  "description": "An AWS Agent",
  "enableAutoUpdates": true,
  "name": "AWS Agent",
  "restrictedAccess": true,
  "trackName": "current"
}
For the Trigger Agent Command endpoint, include the following POST Body. This example demonstrates a POST Body used to trigger the RESTART agent command:
{
  "command": "RESTART"
}
Other available agent commands are:
  • PAUSE
  • RESUME
For the Perform Action On Agent Credentials endpoint, include the following POST Body. This example demonstrates a POST Body used to refresh agent credentials:
{
  "action": "SECRET_REFRESH"
}
For other agents, the POST body will vary. For example, in Azure, you must specify a value for vaultName.
Page Limit
integer
A numeric value to limit the maximum number of records per page.

Destination

Select your cloud data warehouse.
Destination
drop-down
required
  • Snowflake: Load your data into Snowflake. You’ll need to set a cloud storage location for temporary staging of the data.
  • Cloud Storage: Load your data directly into your preferred cloud storage location.
Click either the Snowflake or Cloud Storage tab on this page for documentation applicable to that destination type.
Warehouse
drop-down
required
The Snowflake warehouse used to run the queries. The special value [Environment Default] uses the warehouse defined in the environment. Read Overview of Warehouses to learn more.
Database
drop-down
required
The Snowflake database. The special value [Environment Default] uses the database defined in the environment. Read Databases, Tables and Views - Overview to learn more.
Schema
drop-down
required
The Snowflake schema. The special value [Environment Default] uses the schema defined in the environment. Read Database, Schema, and Share DDL to learn more.
Table Name
string
required
The name of the table to be created.
Load Strategy
drop-down
required
  • Replace: If the specified table name already exists, that table will be destroyed and replaced by the table created during this pipeline run.
  • Truncate and Insert: If the specified table name already exists, all rows within the table will be removed and new rows will be inserted per the next run of this pipeline.
  • Fail if Exists: If the specified table name already exists, this pipeline will fail to run.
  • Append: If the specified table name already exists, then the data is inserted without altering or deleting the existing data in the table. It’s appended onto the end of the existing data in the table. If the specified table name doesn’t exist, then the table will be created, and your data will be inserted into the table.
Clean Staged files
boolean
required
  • Yes: Staged files will be destroyed after data is loaded. This is the default setting.
  • No: Staged files are retained in the staging area after data is loaded.
Stage Access Strategy
drop-down
Select the stage access strategy. The strategies available depend on the cloud platform you select in Stage Platform.
  • Credentials: Connects to the external stage (AWS, Azure) using your configured cloud provider credentials. Not available for Google Cloud Storage.
  • Storage Integration: Use a Snowflake storage integration to grant access to Snowflake to read data from and write to a cloud storage location. This will reveal the Storage Integration property, through which you can select any of your existing Snowflake storage integrations.
Stage Platform
drop-down
required
Choose a data staging platform using the drop-down menu.
  • Amazon S3: Stage your data on an AWS S3 bucket.
  • Snowflake: Stage your data on a Snowflake internal stage.
  • Azure Storage: Stage your data in an Azure Blob Storage container.
  • Google Cloud Storage: Stage your data in a Google Cloud Storage bucket.
Click one of the tabs below for documentation applicable to that staging platform.
Storage Integration
drop-down
required
Select the storage integration. Storage integrations are required to permit Snowflake to read data from and write to a cloud storage location. Integrations must be set up in advance of selecting them. Storage integrations can be configured to support Amazon S3, Google Cloud Storage, or Microsoft Azure Blob Storage, regardless of the cloud provider that hosts your Snowflake account.
Amazon S3 Bucket
drop-down
required
An AWS S3 bucket to stage data into. The drop-down menu will include buckets tied to the cloud provider credentials that you have associated with your environment.

Advanced Settings

Log Level
drop-down
Set the severity level of logging. Choose from Error, Warn, Info, Debug, or Trace. Logs can be found in the Message field of the task details after the pipeline has been run.
Load Selected Data
boolean
Choose whether to return the entire payload or only selected data objects. Read Structure to learn how to select which data objects to include in your API response.
  • No: Will return the entire payload. This is the default setting.
  • Yes: Will return only the objects in Custom Connector that are marked as Selected Data in the Structure setting.

Deactivate soft delete for Azure blobs (Databricks)

If you intend to set your destination as Databricks and your stage platform as Azure Storage, you must turn off the “Enable soft delete for blobs” setting in your Azure account for your pipeline to run successfully. To do this:
  1. In the Azure portal, navigate to your storage account.
  2. In the menu, under Data management, click Data protection.
  3. Clear the Enable soft delete for blobs checkbox. For more information, read Soft delete for blobs.