Configuration settings
To configure the DB extractor, you must create a configuration file. The file must be in YAML format.
You can set up extraction pipelines to use versioned extractor configuration files stored in the cloud.
Using values from environment variables
The configuration file allows substitutions with environment variables. For example:
cognite:
secret: ${COGNITE_CLIENT_SECRET}
will load the value from the COGNITE_CLIENT_SECRET
environment variable into the cognite/secret
parameter. You can also do string interpolation with environment variables, for example:
url: http://my-host.com/api/endpoint?secret=${MY_SECRET_TOKEN}
Implicit substitutions only work for unquoted value strings. For quoted strings, use the !env
tag to activate environment substitution:
url: !env 'http://my-host.com/api/endpoint?secret=${MY_SECRET_TOKEN}'
Using values from Azure Key Vault
The DB extractor also supports loading values from Azure Key Vault. To load a configuration value from Azure Key Vault, use the !keyvault
tag followed by the name of the secret you want to load. For example, to load the value of the my-secret-name
secret in Key Vault into a password
parameter, configure your extractor like this:
password: !keyvault my-secret-name
To use Key Vault, you also need to include the azure-keyvault
section in your configuration, with the following parameters:
Parameter | Description |
---|---|
keyvault-name | Name of Key Vault to load secrets from |
authentication-method | How to authenticate to Azure. Either default or client-secret . For default , the extractor will look at the user running the extractor, and look for pre-configured Azure logins from tools like the Azure CLI. For client-secret , the extractor will authenticate with a configured client ID/secret pair. |
client-id | Required for using the client-secret authentication method. The client ID to use when authenticating to Azure. |
secret | Required for using the client-secret authentication method. The client secret to use when authenticating to Azure. |
tenant-id | Required for using the client-secret authentication method. The tenant ID of the Key Vault in Azure. |
Example:
azure-keyvault:
keyvault-name: my-keyvault-name
authentication-method: client-secret
tenant-id: 6f3f324e-5bfc-4f12-9abe-22ac56e2e648
client-id: 6b4cc73e-ee58-4b61-ba43-83c4ba639be6
secret: 1234abcd
Base configuration object
Parameter | Type | Description |
---|---|---|
version | either string or integer | Configuration file version |
type | either local or remote | Configuration file type. Either local , meaning the full config is loaded from this file, or remote , which means that only the cognite section is loaded from this file, and the rest is loaded from extraction pipelines. Default value is local . |
cognite | object | The cognite section describes which CDF project the extractor will load data into and how to connect to the project. |
logger | object | The optional logger section sets up logging to a console and files. |
metrics | object | The metrics section describes where to send metrics on extractor performance for remote monitoring of the extractor. We recommend sending metrics to a Prometheus pushgateway, but you can also send metrics as time series in the CDF project. |
queries | list | List of queries to execute |
databases | list | List of databases to connect to |
extractor | object | General extractor configuration |
cognite
Global parameter.
The cognite section describes which CDF project the extractor will load data into and how to connect to the project.
Parameter | Type | Description |
---|---|---|
project | string | Insert the CDF project name. |
idp-authentication | object | The idp-authentication section enables the extractor to authenticate to CDF using an external identity provider (IdP), such as Microsoft Entra ID (formerly Azure Active Directory). |
data-set | object | Enter a data set the extractor should write data into |
extraction-pipeline | object | Enter the extraction pipeline used for remote config and reporting statuses |
host | string | Insert the base URL of the CDF project. Default value is https://api.cognitedata.com . |
timeout | integer | Enter the timeout on requests to CDF, in seconds. Default value is 30 . |
external-id-prefix | string | Prefix on external ID used when creating CDF resources |
connection | object | Configure network connection details |
idp-authentication
Part of cognite
configuration.
The idp-authentication
section enables the extractor to authenticate to CDF using an external identity provider (IdP), such as Microsoft Entra ID (formerly Azure Active Directory).
Parameter | Type | Description |
---|---|---|
authority | string | Insert the authority together with tenant to authenticate against Azure tenants. Default value is https://login.microsoftonline.com/ . |
client-id | string | Required. Enter the service principal client id from the IdP. |
tenant | string | Enter the Azure tenant. |
token-url | string | Insert the URL to fetch tokens from. |
secret | string | Enter the service principal client secret from the IdP. |
resource | string | Resource parameter passed along with token requests. |
audience | string | Audience parameter passed along with token requests. |
scopes | list | Enter a list of scopes requested for the token |
min-ttl | integer | Insert the minimum time in seconds a token will be valid. If the cached token expires in less than min-ttl seconds, it will be refreshed even if it is still valid. Default value is 30 . |
certificate | object | Authenticate with a client certificate |
scopes
Part of idp-authentication
configuration.
Enter a list of scopes requested for the token
Each element of this list should be a string.
certificate
Part of idp-authentication
configuration.
Authenticate with a client certificate
Parameter | Type | Description |
---|---|---|
authority-url | string | Authentication authority URL |
path | string | Required. Enter the path to the .pem or .pfx certificate to be used for authentication |
password | string | Enter the password for the key file, if it is encrypted. |
data-set
Part of cognite
configuration.
Enter a data set the extractor should write data into
Parameter | Type | Description |
---|---|---|
id | integer | Resource internal id |
external-id | string | Resource external id |
extraction-pipeline
Part of cognite
configuration.
Enter the extraction pipeline used for remote config and reporting statuses
Parameter | Type | Description |
---|---|---|
id | integer | Resource internal id |
external-id | string | Resource external id |
connection
Part of cognite
configuration.
Configure network connection details
Parameter | Type | Description |
---|---|---|
disable-gzip | boolean | Whether or not to disable gzipping of json bodies. |
status-forcelist | string | HTTP status codes to retry. Defaults to 429, 502, 503 and 504 |
max-retries | integer | Max number of retries on a given http request. Default value is 10 . |
max-retries-connect | integer | Max number of retries on connection errors. Default value is 3 . |
max-retry-backoff | integer | Retry strategy employs exponential backoff. This parameter sets a max on the amount of backoff after any request failure. Default value is 30 . |
max-connection-pool-size | integer | The maximum number of connections which will be kept in the SDKs connection pool. Default value is 50 . |
disable-ssl | boolean | Whether or not to disable SSL verification. |
proxies | object | Dictionary mapping from protocol to url. |
proxies
Part of connection
configuration.
Dictionary mapping from protocol to url.
logger
Global parameter.
The optional logger
section sets up logging to a console and files.
Parameter | Type | Description |
---|---|---|
console | object | Include the console section to enable logging to a standard output, such as a terminal window. |
file | object | Include the file section to enable logging to a file. The files are rotated daily. |
metrics | boolean | Enables metrics on the number of log messages recorded per logger and level. This requires metrics to be configured as well |
console
Part of logger
configuration.
Include the console section to enable logging to a standard output, such as a terminal window.
Parameter | Type | Description |
---|---|---|
level | either DEBUG , INFO , WARNING , ERROR or CRITICAL | Select the verbosity level for console logging. Valid options, in decreasing verbosity levels, are DEBUG , INFO , WARNING , ERROR , and CRITICAL . Default value is INFO . |
file
Part of logger
configuration.
Include the file section to enable logging to a file. The files are rotated daily.
Parameter | Type | Description |
---|---|---|
level | either DEBUG , INFO , WARNING , ERROR or CRITICAL | Select the verbosity level for file logging. Valid options, in decreasing verbosity levels, are DEBUG , INFO , WARNING , ERROR , and CRITICAL . Default value is INFO . |
path | string | Required. Insert the path to the log file. |
retention | integer | Specify the number of days to keep logs for. Default value is 7 . |
metrics
Global parameter.
The metrics
section describes where to send metrics on extractor performance for remote monitoring of the extractor. We recommend sending metrics to a Prometheus pushgateway, but you can also send metrics as time series in the CDF project.
Parameter | Type | Description |
---|---|---|
push-gateways | list | List of prometheus pushgateway configurations |
cognite | object | Push metrics to CDF timeseries. Requires CDF credentials to be configured |
server | object | The extractor can also be configured to expose a HTTP server with prometheus metrics for scraping |
push-gateways
Part of metrics
configuration.
List of prometheus pushgateway configurations
Each element of this list should be a the push-gateways sections contain a list of metric destinations.
Parameter | Type | Description |
---|---|---|
host | string | Enter the address of the host to push metrics to. |
job-name | string | Enter the value of the exported_job label to associate metrics with. This separates several deployments on a single pushgateway, and should be unique. |
username | string | Enter the credentials for the pushgateway. |
password | string | Enter the credentials for the pushgateway. |
clear-after | either null or integer | Enter the number of seconds to wait before clearing the pushgateway. When this parameter is present, the extractor will stall after the run is complete before deleting all metrics from the pushgateway. The recommended value is at least twice that of the scrape interval on the pushgateway. This is to ensure that the last metrics are gathered before the deletion. Default is disabled. |
push-interval | integer | Enter the interval in seconds between each push. Default value is 30 . |
cognite
Part of metrics
configuration.
Push metrics to CDF timeseries. Requires CDF credentials to be configured
Parameter | Type | Description |
---|---|---|
external-id-prefix | string | Required. Prefix on external ID used when creating CDF time series to store metrics. |
asset-name | string | Enter the name for a CDF asset that will have all the metrics time series attached to it. |
asset-external-id | string | Enter the external ID for a CDF asset that will have all the metrics time series attached to it. |
push-interval | integer | Enter the interval in seconds between each push to CDF. Default value is 30 . |
data-set | object | Data set the metrics will be created under |
data-set
Part of cognite
configuration.
Data set the metrics will be created under
Parameter | Type | Description |
---|---|---|
id | integer | Resource internal id |
external-id | string | Resource external id |
server
Part of metrics
configuration.
The extractor can also be configured to expose a HTTP server with prometheus metrics for scraping
Parameter | Type | Description |
---|---|---|
host | string | Host to run the prometheus server on. Default value is 0.0.0.0 . |
port | integer | Local port to expose the prometheus server on. Default value is 9000 . |
queries
Global parameter.
List of queries to execute
Each element of this list should be a description of a SQL query against a database
Parameter | Type | Description |
---|---|---|
database | string | Required. Enter the name of the database to connect to. This must be one of the database names configured in the databases section. |
name | string | Required. Enter a name of this query that will be used for logging and tagging metrics. The name must be unique for each query in the configuration file. |
query | string | Required. SQL query to execute. Supports interpolation with {incremental_field} and {start_at} |
destination | configuration for either RAW, Events, Assets, Time series, Sequence or Files | Required. The destination of the data in CDF. Examples: {'destination': {'type': 'raw', 'database': 'my-database', 'table': 'my-table'}} {'destination': {'type': 'events'}} |
primary-key | string | Insert the format of the row key in CDF RAW. This parameter supports case-sensitive substitutions with values from the table columns. For example, if there's a column called index, setting primary-key: row_{index} will result in rows with keys row_0 , row_1 , etc. This is a required value if the destination is a raw type.Example: row_{index} |
incremental-field | string | Insert the table column that holds the incremental field. Include to enable incremental loading, otherwise the extractor will default to a full run every time. To use incremental load, a state store is required |
freshness-field | string | Which column to use for freshness metric. Must be specified along with freshness-field-timezone |
freshness-field-timezone | string | Timezone to use for freshness metric |
initial-start | either string, number or integer | Enter the {start_at} for the first run. Later runs will use the value stored in the state store. Will only be used on the initial run, subsequent runs will use the stored state. Required when incremental-field is set. |
schedule | configuration for either Fixed interval or CRON expression | Enter the schedule for when this query should run. Make sure not to schedule runs too often, but leave some room for the previous execution to be done. Required when running in continuous mode, ignored otherwise. Examples: {'schedule': {'type': 'interval', 'expression': '1h'}} {'schedule': {'type': 'cron', 'expression': '0 7-17 * * 1-5'}} |
collection | string | Specify the collection on which the query will be executed. This parameter is mandatory when connecting to mongodb databases. |
container | string | Specify the container on which the query will be executed. This parameter is mandatory when connecting to cosmosdb databases. |
sheet | string | Specify the sheet on which the query will be executed. This parameter is mandatory when connecting to spreadsheet files. |
skip_rows | string | Specify the number of rows to be skipped when reading a spreadsheet. This parameter is optional when connecting to spreadsheet files. |
has_header | string | Specify if the extractor should skip the file header while reading a spreadsheet. This parameter is optional when connecting to spreadsheet files. |
parameters | string | Specify the parameters to be used when querying to AWS DynamoDB. This parameter is mandatory when connectong to dynamodb databases. |
destination
Part of queries
configuration.
The destination of the data in CDF.
Either one of the following options:
Examples:
destination:
type: raw
database: my-database
table: my-table
destination:
type: events
raw
Part of destination
configuration.
The raw destination writes data to the CDF staging area (RAW). The raw destination requires the primary-key
parameter in the query configuration.
Parameter | Type | Description |
---|---|---|
type | always raw | Type of CDF destination, set to raw to write data to RAW. |
database | string | Required. Enter the CDF RAW database to upload data into. This will be created if it doesn't exist. |
table | string | Required. Enter the CDF RAW table to upload data into. This will be created if it doesn't exist. |
events
Part of destination
configuration.
The events
destination inserts the resulting data as CDF events. The events destination is configured by setting the type
parameter to events
. No other parameters are required.
To ingest data into a events, the query must produce columns named
externalId
In addition, columns named
startTime
endTime
description
source
type
subType
may be included and will be mapped to corresponding fields in CDF events. Any other columns returned by the query will be mapped to key/value pairs in the metadata
field for events.
Parameter | Type | Description |
---|---|---|
type | always events | Type of CDF destination, set to events to write data to events. |
assets
Part of destination
configuration.
The assets
destination inserts the resulting data as CDF assets. The assets destination is configured by setting the type
parameter to assets
. No other parameters are required.
To ingest data into a assets, the query must produce columns named
name
In addition, columns named
externalId
parentExternalId
description
source
may be included and will be mapped to corresponding fields in CDF assets. Any other columns returned by the query will be mapped to key/value pairs in the metadata
field for assets.
Parameter | Type | Description |
---|---|---|
type | always assets | Type of CDF destination, set to assets to write data to assets. |
time_series
Part of destination
configuration.
The time_series
destination inserts the resulting data as data points in time series. The time series destination is configured by setting the type
parameter to time_series
. No other parameters are required.
To ingest data into a time series, the query must produce columns named
externalId
timestamp
value
In addition, include a column called status
to give the datapoint a status code. Statuses include a category, and an optional comma-separated list of modifyer flags. You can read more about status codes here. Some examples for status codes include Good
(which is assumed if status is omitted), UNCERTAIN, HIGH
and bad
.
The extractor will insert data points into time series identified by the externalId
column. If a time series does not exist, the extractor will create a minimal time series with only an external ID and the isString
property inferred from the type of first data point processed for that time series. All other time series attributes need to be added separately.
Parameter | Type | Description |
---|---|---|
type | always time_series | Type of CDF destination, set to time_series to write data to time series. |
sequence
Part of destination
configuration.
The sequence
destination writes data to a CDF sequence.
The column set of the query result will determine the columns of the sequence.
The result must include a column named row_number
, which must include an integer indicating which row number in the sequence to ingest the row into.
Parameter | Type | Description |
---|---|---|
type | always sequence | Type of CDF destination, set to sequence to write data to a sequence. |
external-id | string | Required. Configured sequence external ID |
value-types | either convert , drop or assert | How types are converted into the expected types in CDF. Convert attempts to make a conversion, which may fail. Drop drops the row if there is a mismatch. Assert fails the query if the types do not match. Default value is convert . |
files
Part of destination
configuration.
The files
destination inserts the resulting data as CDF files. The files destination is configured by setting the type
parameter to files
. No other parameters are required.
To ingest data into a files, the query must produce columns named
name
externalId
content
content
will be treated as binary data and uploaded to CDF files as the content of the file
In addition, columns named
source
mimeType
directory
sourceCreatedTime
sourceModifiedTime
asset_ids
may be included and will be mapped to corresponding fields in CDF files. Any other columns returned by the query will be mapped to key/value pairs in the metadata
field for files.
Parameter | Type | Description |
---|---|---|
type | always files | Type of CDF destination, set to files to write data to CDF files. |
content-column | string | Column used as file content. Default value is content . |
schedule
Part of queries
configuration.
Enter the schedule for when this query should run. Make sure not to schedule runs too often, but leave some room for the previous execution to be done. Required when running in continuous mode, ignored otherwise.
Either one of the following options:
Examples:
schedule:
type: interval
expression: 1h
schedule:
type: cron
expression: 0 7-17 * * 1-5
fixed_interval
Part of schedule
configuration.
Parameter | Type | Description |
---|---|---|
type | always interval | Required. Type of time interval configuration. Use interval to configure a fixed interval. |
expression | string | Required. Enter a time interval, with a unit. Avaiable units are s (seconds), m (minutes), h (hours) and d (days).Examples: 45s 15m 2h |
cron_expression
Part of schedule
configuration.
Parameter | Type | Description |
---|---|---|
type | always cron | Required. Type of time interval configuration. Use cron to configure CRON schedule. |
expression | string | Required. Enter a CRON expression. See crontab.guru for a guide on writing CRON expressions. Example: */15 8-16 * * * |
databases
Global parameter.
List of databases to connect to
Each element of this list should be a configuration for a database the extractor will connect to
Either one of the following options:
- ODBC
- PostgreSQL
- Oracle DB
- Snowflake
- MongoDB
- Azure Cosmos DB
- Local spreadsheet files
- Amazon Dynamo DB
- Amazon Redshift
- Google BigQuery
Example:
databases:
- type: odbc
name: my-odbc-database
connection-string: DRIVER={Oracle 19.3};DBQ=localhost:1521/XE;UID=SYSTEM;PWD=oracle
- type: postgres
name: postgres-db
host: pg.company.com
user: postgres
password: secret123Pas$word
odbc
Part of databases
configuration.
Open Database Connectivity (ODBC) is a generic protocol for querying databases. To connect to a database using ODBC, you must first download and install an ODBC driver for your database system on the machine running the extractor. Consult the documentation or contact the vendor of your database system to find its driver.
Example:
type: odbc
name: asset-database
connection-string: Driver={ODBC Driver 17 for SQL Server};Server=10.24.5.162;Database=assets;UID=extractorUser;PWD=myPassword;
Parameter | Type | Description |
---|---|---|
type | always odbc | Select the type of database connection. Set to odbc for ODBC databases. |
connection-string | string | Required. Enter the ODBC connection string. This will differ between database vendors. Examples: DRIVER={Oracle 19.3};DBQ=localhost:1521/XE;UID=SYSTEM;PWD=oracle DSN={MyDatabaseDsn} |
response-encoding | string | Override the encoding to expect on database responses if the driver does not adhere to the ODBC standard. Default is to follow the ODBC standard. Examples: utf8 iso-8859-1 |
query-encoding | string | Override the encoding to use on database queries if the driver does not adhere to the ODBC standard. Default is to follow the ODBC standard. Examples: utf8 iso-8859-1 |
timeout | integer | Enter the timeout in seconds for the ODBC connection and queries. The default value is no timeout. Some ODBC drivers don't accept either the SQL_ATTR_CONNECTION_TIMEOUT or the SQL_ATTR_QUERY_TIMEOUT option. The extractor will log an exception with the message Could not set timeout on the ODBC driver - timeouts might not work properly . Extractions will continue regardless but without timeouts. To avoid this logline, you can disable timeouts for the database causing these problems. |
batch-size | integer | Enter the number of rows to fetch from the database at a time. You can decrease this number if the machine with the extractor runs out of memory. Note that this will increase the run time. Default value is 1000 . |
name | string | Enter a name for the database that will be used throughout the queries section and for logging. The name must be unique for each database in the configuration file. |
timezone | configuration for either local time zone, universal coordinated time or offset from UTC | Specify how the extractor should handle timestamps from the source when timezone data is absent. Either local for the local timezone on the machine the extractor is running on, utc for UTC, or a number for a numerical offset from UTC. Default value is local .Examples: utc -8 5.5 |
postgresql
Part of databases
configuration.
Example:
type: postgres
name: my-database
host: 10.42.39.12
user: extractor-user
password: mySecretPassword
Parameter | Type | Description |
---|---|---|
type | always postgres | Required. Type of database connection, set to postgres for PostgreSQL databases. |
host | string | Required. Enter the hostname or address of postgres database Examples: 123.234.123.234 postgres.my-domain.com localhost |
user | string | Required. Enter the username for postgres database |
password | string | Required. Enter the password for postgres database |
database | string | Enter the database name to use. The default is to use the user name. |
port | integer | Enter the port to connect to. Default value is 5432 . |
timeout | integer | Enter the timeout in seconds for the database connection and queries. The default value is no timeout. |
batch-size | integer | Enter the number of rows to fetch from the database at a time. You can decrease this number if the machine with the extractor runs out of memory. Note that this will increase the run time. Default value is 1000 . |
name | string | Enter a name for the database that will be used throughout the queries section and for logging. The name must be unique for each database in the configuration file. |
timezone | configuration for either local time zone, universal coordinated time or offset from UTC | Specify how the extractor should handle timestamps from the source when timezone data is absent. Either local for the local timezone on the machine the extractor is running on, utc for UTC, or a number for a numerical offset from UTC. Default value is local .Examples: utc -8 5.5 |
oracle_db
Part of databases
configuration.
The Cognite DB Extractor can connect directly to an Oracle Database version 12.1 or later.
Example:
type: oracle
name: my-database
host: 10.42.39.12
user: extractor-user
password: mySecretPassword
Parameter | Type | Description |
---|---|---|
type | always oracle | Type of database connection, set to oracle for Oracle databases. |
host | string | Required. Enter the hostname or address of oracle database Examples: 123.234.123.234 database.my-domain.com localhost |
user | string | Required. Enter the user name |
password | string | Required. Enter the user password |
port | integer | Enter the port to connect to. Default value is 1521 . |
service-name | string | Optionally specify the service name of the database to connect to |
timeout | integer | Timeout for statements to the database |
batch-size | integer | Enter the number of rows to fetch from the database at a time. You can decrease this number if the machine with the extractor runs out of memory. Note that this will increase the run time. Default value is 1000 . |
name | string | Enter a name for the database that will be used throughout the queries section and for logging. The name must be unique for each database in the configuration file. |
timezone | configuration for either local time zone, universal coordinated time or offset from UTC | Specify how the extractor should handle timestamps from the source when timezone data is absent. Either local for the local timezone on the machine the extractor is running on, utc for UTC, or a number for a numerical offset from UTC. Default value is local .Examples: utc -8 5.5 |
snowflake
Part of databases
configuration.
Parameter | Type | Description |
---|---|---|
type | always snowflake | Type of database connection, set to snowflake for Snowflake data warehouses. |
user | string | Required. User name for Snowflake |
password | string | Required. Password for Snowflake |
account | string | Required. Snowflake account ID |
organization | string | Required. Snowflake organzation name |
database | string | Required. Snowflake database to use |
schema | string | Required. Snowflake schema to use |
name | string | Enter a name for the database that will be used throughout the queries section and for logging. The name must be unique for each database in the configuration file. |
timezone | configuration for either local time zone, universal coordinated time or offset from UTC | Specify how the extractor should handle timestamps from the source when timezone data is absent. Either local for the local timezone on the machine the extractor is running on, utc for UTC, or a number for a numerical offset from UTC. Default value is local .Examples: utc -8 5.5 |
mongodb
Part of databases
configuration.
Parameter | Type | Description |
---|---|---|
type | always mongodb | Type of database connection, set to mongodb for MongoDB databases. |
uri | string | Required. Adress and authentication data for the database as a Uniform Resource Identifier (URI). You can read more about MongoDB URIs here. Example: mongodb://mymongo:port/?retryWrites=true&connectTimeoutMS=10000 |
database | string | Required. Name of the related MongoDB database to use. |
name | string | Enter a name for the database that will be used throughout the queries section and for logging. The name must be unique for each database in the configuration file. |
timezone | configuration for either local time zone, universal coordinated time or offset from UTC | Specify how the extractor should handle timestamps from the source when timezone data is absent. Either local for the local timezone on the machine the extractor is running on, utc for UTC, or a number for a numerical offset from UTC. Default value is local .Examples: utc -8 5.5 |
azure_cosmos_db
Part of databases
configuration.
Parameter | Type | Description |
---|---|---|
type | always cosmosdb | Type of database connection, set to cosmosdb for Cosmos DB databases. |
host | string | Required. Host address for the database Example: https://my-cosmos-db.documents.azure.com |
key | string | Required. Azure Key used to connect to the Cosms DB instance |
database | string | Required. Database name to use |
name | string | Enter a name for the database that will be used throughout the queries section and for logging. The name must be unique for each database in the configuration file. |
timezone | configuration for either local time zone, universal coordinated time or offset from UTC | Specify how the extractor should handle timestamps from the source when timezone data is absent. Either local for the local timezone on the machine the extractor is running on, utc for UTC, or a number for a numerical offset from UTC. Default value is local .Examples: utc -8 5.5 |
local_spreadsheet_files
Part of databases
configuration.
The Cognite DB extractor can run against excel spreadsheets and other files containting tabular data. The currently supported file types are
- xlsx, xlsm and xlsb (modern Excel files)
- xls (legacy excel files)
- odf, ods and odt (OpenDocument Format, used by e.g. Libre Office and Open Office)
- csv (Comma separated values)
When using Excel or OpenDocument Format spreadsheets, you need to provide an additional sheet
parameter in the associated query configuration.
Parameter | Type | Description |
---|---|---|
type | always spreadsheet | Type of connection, set to spreadsheet for local spreadsheet files. |
path | string | Required. Path to a single spreadsheet file Examples: /path/to/my/excel/file.xlsx ./relative/path/file.csv C:\\Users\\Robert\\Documents\\spreadsheet.xls |
name | string | Enter a name for the database that will be used throughout the queries section and for logging. The name must be unique for each database in the configuration file. |
timezone | configuration for either local time zone, universal coordinated time or offset from UTC | Specify how the extractor should handle timestamps from the source when timezone data is absent. Either local for the local timezone on the machine the extractor is running on, utc for UTC, or a number for a numerical offset from UTC. Default value is local .Examples: utc -8 5.5 |
amazon_dynamo_db
Part of databases
configuration.
Parameter | Type | Description |
---|---|---|
type | always dynamodb | Type of database connection, set to dynamodb for DynamoDB databases. |
aws-access-key-id | string | Required. AWS authentication access key ID |
aws-secret-access-key | string | Required. AWS authentication access key secret |
region-name | string | Required. AWS region where your database is located. Example: us-east-1 |
name | string | Enter a name for the database that will be used throughout the queries section and for logging. The name must be unique for each database in the configuration file. |
timezone | configuration for either local time zone, universal coordinated time or offset from UTC | Specify how the extractor should handle timestamps from the source when timezone data is absent. Either local for the local timezone on the machine the extractor is running on, utc for UTC, or a number for a numerical offset from UTC. Default value is local .Examples: utc -8 5.5 |
amazon_redshift
Part of databases
configuration.
Parameter | Type | Description |
---|---|---|
type | always redshift | Type of database connection, set to redshift for Redshift databases. |
aws-access-key-id | string | Required. AWS authentication access key ID |
aws-secret-access-key | string | Required. AWS authentication access key secret |
region-name | string | Required. AWS region where your database is located. Example: us-east-1 |
database | string | Required. Redshift database |
secret-arn | string | AWS Secret ARN |
cluster-identifier | string | Name of the Redshift cluster to connect. This parameter is required when connecting to a managed Redshift cluster. |
workgroup-name | string | Name of the Redshift workgroup to connect. This parameter is mandatory when connecting to a Redshift Serverless database. |
name | string | Enter a name for the database that will be used throughout the queries section and for logging. The name must be unique for each database in the configuration file. |
timezone | configuration for either local time zone, universal coordinated time or offset from UTC | Specify how the extractor should handle timestamps from the source when timezone data is absent. Either local for the local timezone on the machine the extractor is running on, utc for UTC, or a number for a numerical offset from UTC. Default value is local .Examples: utc -8 5.5 |
google_bigquery
Part of databases
configuration.
The Cognite DB Extractor can run against Google BigQuery using Google SQL(like) query.
Because this extends the Google SDK, you also authenticate with the Google suggested authentication methods by setting the GOOGLE_APPLICATION_CREDENTIALS
environment variable to the path of your authentication key
Parameter | Type | Description |
---|---|---|
type | always bigquery | Type of database connection, set to bigquery for Google BigQuery |
name | string | Enter a name for the database that will be used throughout the queries section and for logging. The name must be unique for each database in the configuration file. |
timezone | configuration for either local time zone, universal coordinated time or offset from UTC | Specify how the extractor should handle timestamps from the source when timezone data is absent. Either local for the local timezone on the machine the extractor is running on, utc for UTC, or a number for a numerical offset from UTC. Default value is local .Examples: utc -8 5.5 |
extractor
Global parameter.
General extractor configuration
Parameter | Type | Description |
---|---|---|
state-store | object | Include the state store section to save extraction states between runs. Use this if data is loaded incrementally. We support multiple state stores, but you can only configure one at a time. |
upload-queue-size | integer | Maximum size of upload queue. Upload to CDF will be triggered once this limit is reached. Default value is 100000 . |
parallelism | integer | Maximum number of parallel queries. Default value is 4 . |
mode | either continuous or single | Extractor mode. Continuous runs the configured queries using the schedules configured per query. Single runs the queries once each. |
state-store
Part of extractor
configuration.
Include the state store section to save extraction states between runs. Use this if data is loaded incrementally. We support multiple state stores, but you can only configure one at a time.
Parameter | Type | Description |
---|---|---|
raw | object | A RAW state store stores the extraction state in a table in CDF RAW. |
local | object | A local state store stores the extraction state in a JSON file on the local machine. |
raw
Part of state-store
configuration.
A RAW state store stores the extraction state in a table in CDF RAW.
Parameter | Type | Description |
---|---|---|
database | string | Required. Enter the database name in CDF RAW. |
table | string | Required. Enter the table name in CDF RAW. |
upload-interval | integer | Enter the interval in seconds between each upload to CDF RAW. Default value is 30 . |
local
Part of state-store
configuration.
A local state store stores the extraction state in a JSON file on the local machine.
Parameter | Type | Description |
---|---|---|
path | string | Required. Insert the file path to a JSON file. |
save-interval | integer | Enter the interval in seconds between each save. Default value is 30 . |