Connectors → MySQL
About MySQL
MySQL is one of the most popular open-source relational database management systems. MySQL has a proven track record of performance, reliability, and ease-of-use. A relational database organizes data into tables. These tables contain data types that can have relations. These relationships help structure the data. Programmers use SQL to create, extract, modify, and control access to data in a relational database.
MySQL Connector Updates
This section is to explore the updates in the newer versions of the MySQL connector available on the Incorta connectors marketplace.
In order to get the newer version of the connector, please update the connector using the marketplace.
Version | Updates |
---|---|
2.0.2.0 | Introduce log-based incremental load support with a known limitation that in case of deleting records and performing full load, you must reconfigure the Debezium connector to avoid retrieving deleted records from the snapshot. |
2.0.2.2 | Fixed the known limitation in 2.0.2.0. |
2.2.0.2 | A new extra option, sql.chunks.limit, is now available to specify the maximum number of chunks that can be created by a data set when chunking is enabled. The default is 100, and the maximum is 300. If you specify a limit that exceeds 300, the connector resets it to 300. |
Keep your connector up-to-date with the latest connector version released to get all introduced fixes and enhancements.
MySQL Connector
Incorta loads data from a MySQL database using the MySQL connector. The MySQL Connector uses the 5.1.48 MySQL driver.
The MySQL connector supports the following Incorta specific functionality:
Feature | Supported |
---|---|
Chunking | ✔ |
Data Agent | ✔ |
Encryption at Ingest | |
Incremental Load | ✔ |
Multi-Source | ✔ |
OAuth | |
Performance Optimized | ✔ |
Remote | |
Single-Source | ✔ |
Spark Extraction | ✔ |
Webhook Callbacks | ✔ |
Pre-SQL support | ✔ |
The MySQL Connector supports two types of incremental load, including support for using a numeric column. To learn more see Types of Incremental Load
Steps to connect MySQL and Incorta
To connect MySQL and Incorta, here are the high-level steps, tools, and procedures:
- Create an external data source
- Create a schema with the Schema Wizard
- or, Create a schema with the Schema Designer
- Load the schema
- Explore the schema
Create an external data source
Here are the steps to create an external data source with the MySQL connector:
- Sign in to the Incorta Direct Data Platform.
- In the Navigation bar, select Data.
- In the Action bar, select + New → Add Data Source.
- In the Choose a Data Source dialog, in Database, select MySQL.
- In the New Data Source dialog, specify the applicable connector properties.
- To test, select Test Connection.
- Select Ok to save your changes.
MySQL connector properties
Here are the properties for the MySQL connector:
Property | Control | Description |
---|---|---|
Data Source Name | text box | Enter the name of the data source |
Username | text box | Enter the database username |
Password | text box | Enter the database password |
Connection Pool | text box | Enter the connection pool. The default is 30. |
Connection String | text box | Enter the connection string for the MySQL JDBC driver. The format is: jdbc:mysql://<HOST>:<PORT>/<DATABASE_NAME> The default JDBC port for a MySQL Database is: 3306. Verify port numbers with your datasource configurations. |
Connection Properties | text box | Customized string input of key/value pair properties. See below for details on connection properties. |
Extra Options | text box | Enter supported extra options in the form of key=value . |
Use Data Agent | toggle | Enable using a data agent to securely ingest data from an external data source that is behind a firewall. For more information, please review Tools → Data Agent and Tools → Data Manager. |
Data Agent | drop down list | Enable Use Data Agent to configure this property. Select from the data agents created in the tenant, if any. |
A data agent is a service that runs on a remote host. It is also a data agent object in the Data Manager for a given tenant. An authentication file shared between the data agent object and the data agent service enables an authorized connection without using a VPN or SSH tunnel. With a data agent, you can securely extract data from one or more databases behind a firewall to an Incorta cluster. Your Incorta cluster can reside on-premises or in the cloud.
You must have the Incorta cluster enabled and configured to support the use of Data Agents. To learn more, see Concepts → Data Agent and Tools → Data Agent.
Connection Properties
The connection properties allow for customized connection to your MySQL database. The properties are accepted in a key=value
format. The type of connector properties can range from username and password to SSL settings. The 5.1.48 MySQL driver determines the available connection properties. For a list of connection properties available with Incorta and a MySQL database refer to Configuration Properties for Connector.
Create a schema with the Schema Wizard
Here are the steps to create an MySQL schema with the Schema Wizard:
- Sign in to the Incorta Direct Data Platform.
- In the Navigation bar, select Schema.
- In the Action bar, select + New → Schema Wizard.
- In (1) Choose a Source, specify the following:
- For Enter a name, enter the schema name.
- For Select a Datasource, select the MySQL external data source.
- Optionally create a description.
- In the Schema Wizard footer, select Next.
- In (2) Manage Tables, in the Data Panel, first select the name of the Data Source, and then check the Select All checkbox.
- In the Schema Wizard footer, select Next.
- In (3) Finalize, in the Schema Wizard footer, select Create Schema.
Create a schema with the Schema Designer
Here are the steps to create an MySQL schema using the Schema Designer:
- Sign in to the Incorta Direct Data Platform.
- In the Navigation bar, select Schema.
- In the Action bar, select + New → Create Schema.
- In Name, specify the schema name, and select Save.
- In Start adding tables to your schema, select SQL Database.
- In the Data Source dialog, specify the MySQL table data source properties.
- Select Add.
- In the Table Editor, in the Table Summary section, enter the table name.
- To save your changes, select Done in the Action bar.
MySQL table data source properties
For a schema table in Incorta, you can define the following MySQL specific data source properties as follows:
Property | Control | Description |
---|---|---|
Type | drop down list | Default is SQL Database |
Data Source | drop down list | Select the MySQL external data source |
Incremental | toggle | Enable the incremental load configuration for the schema table. See Types of Incremental Load. |
Incremental Extract Using | drop down list | Enable Incremental to configure this property. Select between Last Successful Extract Time and Maximum Value of a Column. See Types of Incremental Load. |
Incremental Column | drop down list | Enable Incremental and select Maximum Value of a Column to configure this property. Select the column to be used for Maximum Value of a Column. The Loader will track and use the greatest value or most recent timestamp for each load operation. |
Query | text box | Enter the SQL query to retrieve data from the MySQL database |
Update Query | text box | Enable Incremental to configure this property. Define the SQL update query to use during an incremental load. |
Incremental Field Type | drop down list | Enable Incremental to configure this property. Select the format of the incremental field: ● Timestamp ● Unix Epoch (seconds) ● Unix Epoch (milliseconds) |
Pre-SQL | toggle | Enable running SQL statements or calling stored procedures before executing the original extraction query or incremental query during a load job |
Pre-SQL statement | text box | Enter the SQL statements or call the stored procedures that you want to execute before the query or incremental query during a load job |
Fetch Size | text box | Used for performance improvement, fetch size defines the number of records that will be retrieved from the database in each batch until all records are retrieved. The default is 5000. |
Chunking Method | drop down list | Chunking methods allow for parallel extraction of large tables. The default is No Chunking. There are two chunking methods: ● By Size of Chunking (Single Table) ● By Date/Timestamp |
Chunk Size | text box | Select By Size of Chunking for the Chunking Method to set this property. Enter the number of records to extract in each chunk in relation to the Fetch Size. The default is 3 times the Fetch Size. |
Order Column | drop down list | Select By Size of Chunking for the Chunking Method to set this property. Select a column in the source table you want to order by before chunking. It's typically an ID column and it must be numeric. |
Upper Bound for Order Column | text box | Optional. Enter the maximum value for the order column. |
Lower Bound for Order Column | text box | Optional. Enter the minimum value for the order column. |
Order Column [Date/Timestamp] | drop down list | Select By Date/Timestamp for the Chunking Method to set this property. Select a column in the source table you want to order by before chunking. It should be a Date/Timestamp column. |
Chunk Period | drop down list | Select the chunk period that will be used in dividing chunks: ● Daily ● Weekly (default) ● Monthly ● Yearly ● Custom |
Number of days | text box | Select Custom for the Chunk Period to set this property. Enter the chunking period in days |
Set Extraction Timeout | toggle | Enable the termination of extraction (loading) jobs for the SQL query for this table if the extraction time exceeds the configured time |
Stop after (Minutes) | text box | Enable Set Extraction Timeout to configure this property. Enter the maximum number of minutes allowed for the loading job to complete before terminating it. |
Enable Spark Based Extraction | toggle | Enable the configuration of a Spark job to parallelize the data ingest. |
Max Number of Parallel Queries | text box | Enable Spark based extraction to configure this property. In a Spark job, specify the maximum number of parallel queries. |
Column to Parallelize Queries on | drop down list | Enable Spark based extraction to configure this property. Select the numerical column in the source table that Spark will parallelize the extraction queries from. |
Memory Per Extractor | text box | Enable Spark based extraction to configure this property. In a Spark job, allocate the amount of memory per extractor. Units are in gigabytes. |
Callback | toggle | Enable post extraction callback, that is, enable callback on the data source data set(s) by invoking a certain callback URL with parameters containing details about the load job. |
Callback URL | text box | Enable Callback to configure this property. Specify the callback URL. |
View the schema diagram with the Schema Diagram Viewer
Here are the steps to view the schema diagram using the Schema Diagram Viewer:
- Sign in to the Incorta Direct Data Platform.
- In the Navigation bar, select Schema.
- In the list of schemas, select the MySQL schema.
- In the Schema Designer, in the Action bar, select Diagram.
Load the schema
Here are the steps to perform a Full Load of the MySQL schema using the Schema Designer:
- Sign in to the Incorta Direct Data Platform.
- In the Navigation bar, select Schema.
- In the list of schemas, select the MySQL schema.
- In the Schema Designer, in the Action bar, select Load → Full Load.
- To review the load status, in Last Load Status, select the date.
Explore the schema
With the full load of the MySQL schema complete, you can use the Analyzer to explore the schema, create your first insight, and save the insight to a new dashboard.
To open the Analyzer from the schema, follow these steps:
- In the Navigation bar, select Schema.
- In the Schema Manager, in the List view, select the MySQL schema.
- In the Schema Designer, in the Action bar, select Explore Data.
Additional Considerations
Types of Incremental Load
You can enable Incremental Load for a MySQL table data source. There are two types of incremental extracts:
Last Successful Extract Time
Fetch updates since the last time the tables were loaded. Determined by the difference between the current time and the database timestamp.
Maximum Value of a Column
The column-based strategy depends on an extra column called "Incremental Column" in each table. The MySQL connector supports both timestamp and numeric columns. A timestamp column is of the type date or timestamp. A numeric column is of the type int or long.
Changing the incremental load strategy requires a full load to ensure data integrity.
Log-based incremental load
Starting the connector version 2.0.2.0, Incorta supports the log-based incremental load using the change data capture (CDC).
CDC is the process of identifying and capturing changes made to data in a database using logs and then delivering those changes in real time to a downstream process or system.
Currently, the log-based incremental load is a preview feature.
Prerequisites
To be able to use the log-based incremental load, you need to be aware of and apply the following:
- Install and configure Apache Kafka and Kafka Connect.
- Configure Debezium connector, knowing that Incorta recommends using Debezium version 2.4.1.
- Disable snapshot while configuring Debezium.
- Make sure Debezium connector is configured to send data types to Incorta by adding the propagate property.
- Log-based incremental load only supports database physical tables.
- Tables must have primary keys.
Known limitations
- For the time being, Incorta does not track deletion updates through the log-based incremental load (CDC).
- Minimal mismatch in column INTERVAL data types.
- This feature supports Kafka topics that use a single partition only.
Incremental Load Example
In this example, the invoices
table must contain a column of the type Date or Timestamp in order to load the table incrementally with a last successful extract time strategy. In this case, the name of the date column is ModifiedDate and the format of the column is Timestamp.
Here are the data source property values for this example:
- Incremental is enabled
- Query contains
SELECT * FROM `invoices`
- Update Query contains
SELECT * FROM `invoices` WHERE `ModifiedDate` > ?
?
is a variable in the update query that contains the last schema refresh date.
- Incremental Field Type = Timestamp
Incorta replaces the ?
reference character with a value.
Valid Query Types
When creating a query for your MySQL Connector, please note that only SELECT
statements are valid.
Pre-SQL support
This feature allows executing SQL statements before executing the original extraction query or incremental query during a load job. This helps in different scenarios, for example, when you need to do one of the following:
- Run stored procedures
- Update the data source before extraction
- Create a temporary table before executing the extraction query
- Delete old records in the source so that the source includes only the latest records
The data source database management system determines the accepted statements and supported syntax. if the Pre-SQL statement fails during a load job, the object data extraction and loading fail and an error message appears. Logs also will contain the failure reason.