Integrate data from Mimer SQL to Amazon Redshift using Matillion

The Mimer SQL to Redshift connector seamlessly transfers your data to Redshift within minutes, ensuring it remains current without requiring manual coding or complex ETL script management.

Mimer SQL
Amazon Redshift
Mimer SQL to Amazon Redshift banner

Extracting data from Mimer SQL to Amazon Redshift

Extracting data from Mimer SQL is an important step for organizations looking to integrate or migrate their data into cloud-based platforms such as Amazon Redshift. Whether you are consolidating data for analytics or establishing a new data pipeline, careful planning is essential to ensure a smooth and secure transfer. In this article, we will guide you through the fundamental steps required to extract data from Mimer SQL and load it into Redshift. We will begin by explaining how to create a suitable identity in Mimer SQL, ensuring your data extraction process is both secure and compliant. Next, for users of Matillion, we will demonstrate how to check for—or acquire—the necessary JDBC driver to connect Matillion to Mimer SQL. We will also provide guidance on setting up the required network connectivity, ensuring uninterrupted communication between your source and target systems. Finally, we will cover best practices for querying data, both for initial full extracts and for ongoing incremental loads, to support efficient and reliable data transfers. Whether you are new to working with Mimer SQL or an experienced database administrator, this article will provide a structured, practical approach to extracting your data and leveraging the power of Redshift for your organization’s analytics needs.


What is Mimer SQL?

Mimer SQL is a high-performance, standards-based relational database management system (RDBMS) renowned for its stable, zero-maintenance operation and robust support for the ANSI/ISO SQL standard. Engineered initially for mission-critical embedded and real-time systems, Mimer SQL excels in environments where reliability, scalability, and minimal administration are paramount. Its lock-free, multi-version concurrency control ensures smooth multi-user access without bottlenecks, while its cross-platform compatibility spans from mainstream server architectures to mobile and IoT devices. Furthermore, Mimer SQL integrates seamlessly with ODBC, JDBC, ADO.NET, and a variety of programming languages, making it a flexible choice for developers seeking a lightweight yet powerful database solution for enterprise or embedded applications.

matillion logo x Mimer SQL

What is Amazon Redshift?

Amazon Redshift is a fully managed, petabyte-scale data warehouse on AWS, optimized for high-performance analytics and large-scale data processing. It uses columnar storage, compression, and parallel processing for fast queries, supports standard SQL, and integrates with AWS services like S3 and Glue for efficient data ingestion and transformation. Redshift Spectrum allows direct querying of exabyte-scale S3 data without moving it. Security features include encryption and access control. With on-demand and concurrency scaling, Redshift adapts to dynamic workloads, making it ideal for modern, data-driven enterprises.

Why Move Data from Mimer SQL into Amazon Redshift

Leveraging Redshift Analytics by Replicating Data from Mimer SQL

A data engineer or architect might opt to copy data from Mimer SQL into Amazon Redshift for several strategic reasons. Firstly, Mimer SQL may host data that is potentially valuable for analytics, decision-making, or regulatory reporting. By itself, this data is limited in scope, but when integrated with information from other sources—such as operational databases, customer relationship systems, or web analytics platforms—its full value emerges through comprehensive analysis and insight generation. Furthermore, conducting complex analytical queries and data integration tasks directly on the Mimer SQL source can impose additional workloads, potentially degrading performance for transactional operations. By transferring the data into Redshift, a platform specifically designed for scalable analytics, organizations can leverage robust query capabilities and advanced integration features, all while ensuring Mimer SQL remains optimized for its primary functions.

Creating an Identity in Mimer SQL

To create an identity (user) in a Mimer SQL database, you need sufficient privileges—typically, this operation is performed by a user with DBA (Database Administrator) rights. The process involves using the

CREATE IDENT
SQL statement, which defines a database identity and assigns an authentication string (password).

Below are step-by-step instructions:

1. Connect to the Database as a DBA

Before creating an identity, connect to Mimer SQL using a privileged account, such as

SYSADM
.

-- Example connection from a suitable SQL client:
CONNECT SYSADM IDENTIFIED BY 'your_dba_password';

2. Create a New Identity

Run the following SQL statement to create a new identity. Replace

username
and
password
with your desired values.

CREATE IDENT username USING 'password';

Example:

CREATE IDENT johndoe USING 'Strong_Passw0rd';

  • username
    : The name of the identity (user) to create.
  • password
    : The authentication string. Enclose it in single quotes.

Notes

  • Identity names (
    username
    ) are not case-sensitive and are stored in uppercase.
  • The
    USING
    clause specifies the authentication string; password rules may apply according to database settings.

3. Grant Privileges to the Identity

A newly created identity does not have any privileges. Grant the required rights, such as the ability to connect to the database:

GRANT CONNECT TO johndoe;

You may also grant additional privileges (like

RESOURCE
,
DBA
, or object-specific permissions) as necessary. For example:

GRANT SELECT, INSERT ON mytable TO johndoe;

4. Verify the New Identity

Test logging into the database with the new identity:

CONNECT johndoe IDENTIFIED BY 'Strong_Passw0rd';

If the connection is successful, the identity was created and granted connect rights correctly.

Installing the JDBC Driver

At the time of writing, the JDBC driver for Mimer SQL is not bundled by default with Matillion Data Productivity Cloud. This is due to licensing or redistribution restrictions associated with the Mimer SQL JDBC driver. To use a Mimer SQL data source with Matillion, you must manually download and install the necessary driver.

Step 1: Download the JDBC Driver

  1. Visit the official Mimer SQL Database products page.
  2. Locate and download the latest Type 4 JDBC driver for your operating system and platform. A Type 4 driver (also called a “pure Java” driver) is preferred, as it operates entirely in Java and provides optimal compatibility for modern applications and cloud environments.
  3. Ensure you comply with any license agreements during download, keeping in mind any distribution or usage restrictions.

Step 2: Install the JDBC Driver into Matillion

To make the JDBC driver available to Matillion Data Productivity Cloud, you'll need to upload it as an external driver. Detailed installation instructions are maintained by Matillion and can be found in the documentation: Uploading External Drivers.

Follow these instructions carefully to ensure the driver is placed in the correct location and is accessible to Matillion's runtime environment. Typically, the installation involves:

  • Uploading the downloaded
    .jar
    file containing the JDBC driver.
  • Associating it with your Matillion Agent or relevant service.
  • Restarting or refreshing your agent as instructed, so the new driver takes effect.

Step 3: Configure Usage in Matillion

Once the driver has been successfully installed, you can connect to Mimer SQL databases from Matillion components such as database queries. Review the Usage Instructions provided in the official documentation to properly configure connections, specifying the correct JDBC driver class name and connection URL format for Mimer SQL.

By following these steps, you will enable secure and functional connections between Matillion Data Productivity Cloud and your Mimer SQL database, ensuring seamless data integration workflows.

Checking network connectivity

To ensure successful connectivity between the Matillion Data Productivity Cloud and your Mimer SQL database, verify that your Mimer SQL database is configured to allow incoming network connections according to your deployment type:

  • Full SaaS Agent Configuration:
    If you are using a Full SaaS agent, your Mimer SQL instance must permit incoming connections from the IP addresses published by Matillion. Consult the latest list of required source IP addresses here: Matillion IP Addresses.

  • Hybrid SaaS Deployment:
    If you have deployed a Hybrid SaaS agent within your own infrastructure, configure your Mimer SQL database security rules to allow inbound connections from the addresses or CIDR blocks associated with your own Virtual Private Cloud (VPC). To help test and troubleshoot connectivity, you can use the utilities provided at Matillion Network Access Utilities.

Additionally, if your Mimer SQL database is referenced using a DNS hostname rather than a direct IP address, ensure that the agent (whether Full SaaS or Hybrid SaaS) can successfully resolve the DNS name to reach your database server. Failure to resolve the DNS name will prevent Matillion from connecting to your Mimer SQL instance.

Querying Data from Mimer SQL Database

This guide provides technical instructions for querying data from a Mimer SQL database, including example SQL statements, notes on datatype conversion (especially when transferring data to Amazon Redshift), and a recommended approach for performing both initial and incremental data loads.


Example Mimer SQL SELECT Statements

Here are basic examples of querying data from a Mimer SQL database:

``` -- Select all columns from the "employees" table SELECT * FROM employees;

-- Select specific columns with a WHERE condition SELECT employee_id, first_name, last_name FROM employees WHERE department_id = 100;

-- Aggregate data with GROUP BY SELECT department_id, COUNT(*) AS num_employees FROM employees GROUP BY department_id; ```


Considerations for Datatype Conversion (Mimer SQL to Redshift)

When extracting data from Mimer SQL to load into Amazon Redshift, be aware that datatype conversion may occur. Here are a few common mappings:

Mimer SQL Datatype Amazon Redshift Datatype
INTEGER INTEGER
DECIMAL(p, s) NUMERIC(p, s)
FLOAT / DOUBLE DOUBLE PRECISION
CHAR(n) CHAR(n)
VARCHAR(n) VARCHAR(n)
DATE DATE
TIMESTAMP TIMESTAMP
BLOB Not directly supported (convert)

Handle non-compatible datatypes before transferring data, e.g., by casting BLOBs to base64 encoded strings or excluding them from your load process.


Loading Data: Initial and Incremental Loads

Best Pattern: Initial Load Followed by Incremental Loads

For efficiency and reliable synchronization, it’s recommended to perform a one-time initial load followed by incremental loads using the same Database Query component (e.g., in Matillion ETL).

Initial Load
  • Approach: Run an unfiltered SELECT statement to capture the entire table's data.
  • Example Statement:
    SELECT * FROM employees;
  • Usage: Intended for bootstrapping or full data refreshes, performed only once or as needed.
Incremental Load
  • Approach: Add a WHERE clause to the SELECT statement to capture only new or changed records since the last load.
    • The filter is often based on a timestamp or an incremental numeric ID.
  • Example Statement:
    SELECT * FROM employees
        WHERE last_modified > '2024-01-01 00:00:00';
  • Usage: Scheduled periodically (e.g., hourly, daily) to replicate only changed data for efficiency.

Note: For more information and patterns using Matillion, refer to the Matillion Knowledge Base: Incremental Load Data Replication Strategy.


Summary Table: Load Patterns with Database Query Component

Load Type Query Pattern Example Filter Included?
Initial Load
SELECT * FROM tablename;
No
Incremental
SELECT * FROM tablename WHERE updated_at > 'YYYY-MM-DD HH:MM:SS';
Yes (filter included)

Data Integration Architecture

Loading data in advance of integration is a fundamental strategy that divides and conquers data engineering challenges by splitting the process into two distinct steps: first, loading the raw data, and second, performing the integration. This approach is central to the Extract-Load-Transform (ELT) architecture, which offers several distinct advantages. Data integration inherently requires that data be transformed—for instance, through cleaning, joining, or aggregation—and the most effective way to manage these transformations is by building dedicated data transformation pipelines. One significant benefit of ELT is that both transformation and integration operations are executed directly within the target Redshift database. This not only ensures the transformations are fast, scalable, and available on-demand, but also eliminates the need to invest in—and manage—separate data processing infrastructure, resulting in simplified operations and cost savings.

Get started today

Matillion's comprehensive data pipeline platform offers more than point solutions.