Set up the DB extractor
Follow the steps below to set up the extractor.
Before you start
- If the database you're connecting to requires the extractor to use ODBC, make sure to download and install the ODBC drivers for your database .
Tipp
Navigate to Data management > Integrate > Extractors > Cognite DB extractor in CDF to see all supported sources and the recommended approach.
- Check the server requirements for the extractor.
- Create a configuration file according to the configuration settings. The file must be in YAML format.
Connect to a database
Native
The extractor has native support for some databases, and don't require any additional drivers.
ODBC
To connect via ODBC, you must install an ODBC driver for your database system on the machine where you're running the extractor.
Here's a list of links to ODBC drivers for some source systems:
Consult the documentation for your database or contact the vendor if you need help finding ODBC drivers.
ODBC connection strings
ODBC uses connection strings to reference databases. The connection strings contain information about which ODBC driver to use, where to find the database, sign-in credentials, etc. See Connection strings for examples.
Set up a Data Source Name (DSN)
We recommend setting up a DSN for the database if you're running the extractor against an ODBC source on Windows. Then, the Windows DSN system handles password storage instead of keeping it in the configuration file or as an environment variable. In addition, the connection strings will be less complex.
To set up a DSN for your database:
- Open the ODBC Data Sources tool on the machine you're running the extractor from.
- Select Add and the ODBC driver you want to use. In this example, we're configuring a PostgreSQL database:
-
Select Finish.
-
Enter the connection information and the database name. Note that the image below may differ depending on which database type you are configuring a DSN for.
-
Select Test to verify that the information is correct, then Save.
-
Use the simplified connection string in your configuration file:
databases:
- name: my-postgres
connection-string: 'DSN=MyPostgreSQL'
Run as a Windows standalone executable file
- Download the dbextractor-standalone-{VERSION}-win32.exe file via the download links available from the Cognite DB extractor section on the Extract data page in the Cognite Data Fusion (CDF) user interface.
- Save the file in a folder.
- Open a command line window and run the file with a configuration file as an argument.
In this example, the configuration file is named config.yml and saved in the same folder as the executable file:
dbextractor-standalone-<VERSION>-win32.exe ./config.yml
Run as a Windows service
-
Download the dbextractor-service-{VERSION}-win32.exe file via the download links available from the Cognite DB extractor section on the Extract data page in the CDF user interface.
-
Save the file to the same directory as a configuration file named config.yml. This must be the exact name of the file.
-
As an administrator, open up the command line window in the folder you placed the executable file and the configuration file and run the following command:
> .\dbextractor-service-<VERSION>-win32.exe install -
Open the Services app in Windows and find the Cognite DB Extractor Executor service.
-
Right-click the service and select Properties.
-
Configure the service according to your requirements.
Run as a Docker container
Use a Docker container to run the DB extractor on Mac OS or Linux systems. This article describes how to create file-specific bindings with the container. You should consider sharing a folder if a deployment uses more than one file, for example, logs and local state files.
Step 1: Set up the Docker container
-
Create a Docker image containing the ODBC drivers for the database systems you want to extract data from. This example uses a PostgreSQL database.
-
Create a Docker file:
FROM cognite/db-extractor-base:2.5.0
- Install your ODBC drivers. For example, for PostgreSQL:
RUN apt-get install odbc-postgresql
-
Save the Docker file with a descriptive name. In this example, we use postgres.Dockerfile.
-
Build the Docker image:
docker build -f postgres.Dockerfile -t db-extractor-postgres .
Replace
postgres.Dockerfile
with the name of your Docker file.Replace
db-extractor-postgres
with the tag you want to refer to this Docker image as.
-
-
Create a Docker container running this image:
docker run <other options in here, such as volume sharing and network config> db-extractor-postgres
Replace
db-extractor-postgres
with the tag you created in step 1.
When a Docker container runs, the localhost
address points to the container, not the host. This creates an issue if the container runs against a local database. For development and test environments, you can use the host.docker.internal
DNS name to resolve to the host. Don't use this approach for production environments.
You also need the local database to accept connections from outside the localhost for the container to access the database connection.
Step 2: Run the Docker container
Step 2a: Using docker run
- Enter this
docker run
statement:
`$ docker run --network=host --name db-extractor -v /path/to/config.yaml:/config.yaml eu.gcr.io/cognite-registry/db-extractor:<version> /config.yaml`
Where:
-
docker run
starts a new container. -
--name db-extractor
names the container db-extractor. This is to recognize the extractor when runningdocker ps
and to stop the extractor withdocker stop db-extractor
. -
--network=host
grants the container full network access. This is required for pushing metrics and connecting to CDF. -
-v /path/to/config.yaml:/config.yaml
shares the file at/path/to/config.yaml
with the container, putting it at/config.yaml
. -
eu.gcr.io/cognite-registry/db-extractor:<version>
is the name of the image. Replace<version>
with the version you'll run. -
/config.yaml
is sent to the extractor as the configuration file to be opened.
-
Share log files similarly to configuration files if you want the log files to persist when a run stops.
TipRemember to share the extractor state JSON file if you're using the local
state store
option. -
If you're using environment variables to hold configuration values, you must share the environment variable with the container. Use the
-e
parameter, and follow this syntax:
-e CONTAINER_VAR_NAME=value
Load data incrementally
If the database table has a column containing an incremental field, you can set up the extractor to only process new or updated rows since the last extraction. An incremental field is a field that increases for new entries, such as time stamps for the latest update or insertions if the rows never change, or a numerical index.
To load data incrementally for a query:
- Include the
state-store
parameter in theextractor
section of the configuration. - Include the
incremental-field
andinitial-start
parameters in the query configuration. - Update the SQL query with a
WHERE
statement using{incremental-field}
and{start-at}
. For example:
SELECT * FROM table WHERE {incremental-field} >= '{start-at}' ORDER BY {incremental-field} ASC
If the query returns many rows, consider paging the results with the LIMIT
statement. Be aware that this demands a well-defined paging strategy. Consider this example:
SELECT * FROM data WHERE {incremental-field} >= '{start-at}' ORDER BY {incremental-field} ASC LIMIT 10
If 10 or more rows have the same time, the start-at
field won't be updated since the largest value of the incremental field didn't increase in the run. The next run will start with the same value, and the extractor is stuck in a loop. However, changing the query from >=
to >
can cause data loss if many rows have the same value for the incremental field.
Therefore, set a sufficiently high limit if there are duplicates in the incremental field and you're using result paging.
There might be a slight boost in the extractor runtime if the database has an index on the incremental field. However, for a standard run of the DB extractor, network delays are often the main bottleneck for performance, so usually, the difference is minimal.