Zum Hauptinhalt wechseln

Set up the DB extractor

Follow the steps below to set up the extractor.

Before you start

  1. If the database you're connecting to requires the extractor to use ODBC, make sure to download and install the ODBC drivers for your database .
    Tipp

    Navigate to Data management > Integrate > Extractors > Cognite DB extractor in CDF to see all supported sources and the recommended approach.

  2. Check the server requirements for the extractor.
  3. Create a configuration file according to the configuration settings. The file must be in YAML format.

Connect to a database

Native

The extractor has native support for some databases, and don't require any additional drivers.

ODBC

To connect via ODBC, you must install an ODBC driver for your database system on the machine where you're running the extractor.

Here's a list of links to ODBC drivers for some source systems:

Consult the documentation for your database or contact the vendor if you need help finding ODBC drivers.

ODBC connection strings

ODBC uses connection strings to reference databases. The connection strings contain information about which ODBC driver to use, where to find the database, sign-in credentials, etc. See Connection strings for examples.

Set up a Data Source Name (DSN)

We recommend setting up a DSN for the database if you're running the extractor against an ODBC source on Windows. Then, the Windows DSN system handles password storage instead of keeping it in the configuration file or as an environment variable. In addition, the connection strings will be less complex.

To set up a DSN for your database:

  1. Open the ODBC Data Sources tool on the machine you're running the extractor from.
  2. Select Add and the ODBC driver you want to use. In this example, we're configuring a PostgreSQL database:
Add ODBC driver
  1. Select Finish.

  2. Enter the connection information and the database name. Note that the image below may differ depending on which database type you are configuring a DSN for.

Add ODBC driver
  1. Select Test to verify that the information is correct, then Save.

  2. Use the simplified connection string in your configuration file:

    databases:
    - name: my-postgres
    connection-string: 'DSN=MyPostgreSQL'

Run as a Windows standalone executable file

  1. Download the dbextractor-standalone-{VERSION}-win32.exe file via the download links available from the Cognite DB extractor section on the Extract data page in the Cognite Data Fusion (CDF) user interface.
  2. Save the file in a folder.
  3. Open a command line window and run the file with a configuration file as an argument.

In this example, the configuration file is named config.yml and saved in the same folder as the executable file:

dbextractor-standalone-<VERSION>-win32.exe ./config.yml

Run as a Windows service

  1. Download the dbextractor-service-{VERSION}-win32.exe file via the download links available from the Cognite DB extractor section on the Extract data page in the CDF user interface.

  2. Save the file to the same directory as a configuration file named config.yml. This must be the exact name of the file.

  3. As an administrator, open up the command line window in the folder you placed the executable file and the configuration file and run the following command:


    > .\dbextractor-winservice-<VERSION>-win32.exe install

  4. Open the Services app in Windows and find the Cognite DB Extractor Executor service.

  5. Right-click the service and select Properties.

  6. Configure the service according to your requirements.

Run as a Docker container

Use a Docker container to run the DB extractor on Mac OS or Linux systems. This article describes how to create file-specific bindings with the container. You should consider sharing a folder if a deployment uses more than one file, for example, logs and local state files.

ODBC only: Create a custom docker image

If you're using ODBC to connect to your database, you first need to create a docker image that contains the ODBC driver for your database. Cognite ships a base image for the DB extractor that contains the extractor and ODBC libraries, but not specific drivers.

This example uses a PostgreSQL database.

Info

This example uses PostgreSQL since the driver is easily available in Debian's package repository. When you're connecting to a PostgreSQL database, use the native PostgreSQL support in the DB extractor.

  1. Create a Dockerfile and extend the cognite/db-extractor-base image. We recommend locking the version to a major release. Then, install your ODBC drivers. For example, for PostgreSQL:

    FROM cognite/db-extractor-base:3
    RUN apt-get update \
    && apt-get install -y odbc-postgresql \
    && apt-get clean -y \
    && rm -rf /var/lib/apt/lists/* \
    && rm -rf /tmp/*
  2. Save the Docker file with a descriptive name. In this example, we use postgres.Dockerfile.

  3. Build the Docker image. Give the image a tag with the -t argument to make it easy to refer to the image.

    docker build -f postgres.Dockerfile -t db-extractor-postgres .

    Replace postgres.Dockerfile with the name of your Docker file.

    Replace db-extractor-postgres with the tag you want to assign to this Docker image.

  4. Run this image to create a Docker container:

    docker run <other options in here, such as volume sharing and network config> db-extractor-postgres

    Replace db-extractor-postgres with the tag you created above.

Run on a local host on Docker

When a Docker container runs, the localhost address points to the container, not the host. This creates an issue if the container runs against a local database. For development and test environments, you can use the host.docker.internal DNS name to resolve to the host. Don't use this approach for production environments.

You also need the local database to accept connections from outside the localhost for the container to access the database connection.

Load data incrementally

If the database table has a column containing an incremental field, you can set up the extractor to only process new or updated rows since the last extraction. An incremental field is a field that increases for new entries, such as time stamps for the latest update or insertions if the rows never change, or a numerical index.

To load data incrementally for a query:

  1. Include the state-store parameter in the extractor section of the configuration.
  2. Include the incremental-field and initial-start parameters in the query configuration.
  3. Update the SQL query with a WHERE statement using {incremental-field} and {start-at}. For example:
SELECT * FROM table WHERE {incremental-field} >= '{start-at}' ORDER BY {incremental-field} ASC

If the query returns many rows, consider paging the results with the LIMIT statement. Be aware that this demands a well-defined paging strategy. Consider this example:

SELECT * FROM data WHERE {incremental-field} >= '{start-at}' ORDER BY {incremental-field} ASC LIMIT 10
Multiple rows with identical values in the incremental field

If 10 or more rows have the same time, the start-at field won't be updated since the largest value of the incremental field didn't increase in the run. The next run will start with the same value, and the extractor is stuck in a loop. However, changing the query from >= to > can cause data loss if many rows have the same value for the incremental field.

Therefore, set a sufficiently high limit if there are duplicates in the incremental field and you're using result paging.

There might be a slight boost in the extractor runtime if the database has an index on the incremental field. However, for a standard run of the DB extractor, network delays are often the main bottleneck for performance, so usually, the difference is minimal.