How to quickly import data into MySQL

How to quickly import data into MySQL

Preface:

In daily study and work, we often encounter the need to export data. For example, data migration, data recovery, creating a new slave database, etc. These operations may involve the import of a large amount of data. Sometimes the import progress is slow and the computer fan is spinning wildly, which is really frustrating. In fact, there are some tips that can make the import faster. In this article, the author will talk about how to import data quickly.

Note: This article only discusses how to quickly import SQL scripts generated by logical backups, and other file formats are not discussed for the time being.

1. Try to reduce the size of the imported file

First of all, I would like to give you a suggestion. When exporting and importing data, try to use the command line tools that come with MySQL, and do not use graphical tools such as Navicat and workbench. Especially when dealing with large amounts of data, exporting and importing with MySQL's built-in command line tools is several times faster than using graphical tools such as Navicat. Moreover, graphical tools such as Navicat can easily get stuck when performing operations on large amounts of data. The following is a brief introduction on how to use the command line tool that comes with MySQL for import and export.

# Export the entire instance mysqldump -uroot -pxxxxxx --all-databases > all_database.sql

# Export the specified library mysqldump -uroot -pxxxxxx --databases testdb > testdb.sql

# Export the specified table mysqldump -uroot -pxxxxxx testdb test_tb > test_tb.sql

#Import the specified SQL file (specify import into the testdb library)
mysql -uroot -pxxxxxx testdb < testdb.sql

Most of the imported SQL scripts create databases and tables first, and then insert data. The longest time-consuming part is probably inserting data. To reduce the file size, it is recommended to use the extended insert method, that is, batch insert of multiple rows together, similar to this: insert into table_name values ​​(),(),(),...,(); . Using extended insert will result in a much smaller file size and several times faster insertion speed than inserting one by one. Files exported using mysqldump use the batch insert method by default. When exporting, you can use the --skip-extended-insert parameter to change to inserting one by one. Below, we take a data table with 10 million data points as an example to test the speed of inserting files exported in different ways.

As can be seen from the figure above, it takes about 10 minutes to import SQL scripts using extended insert, while it takes too long to import SQL scripts one by one, and it is still not finished after about an hour. The import of a text of more than 2G has not been completed for more than an hour, and the author can't wait and cancels it manually? However, it can be seen that inserting multiple entries together saves several times more time than inserting data one by one.

2. Try to modify the parameters to speed up the import

In MySQL, there is a pair of famous "double one" parameters, namely innodb_flush_log_at_trx_commit and sync_binlog. For security reasons, the default values ​​of these two parameters are 1. In order to quickly import the script, we can temporarily modify these two parameters. The following is a brief introduction to these two parameters:

The default value of innodb_flush_log_at_trx_commit is 1 and can be set to 0, 1, or 2

If innodb_flush_log_at_trx_commit is set to 0, the log buffer will be written to the log file once per second, and the flush operation of the log file will be performed simultaneously. In this mode, when the transaction is committed, the write operation to disk will not be triggered actively.
If innodb_flush_log_at_trx_commit is set to 1, MySQL will write the data in the log buffer to the log file and flush it to disk each time a transaction is committed.
If innodb_flush_log_at_trx_commit is set to 2, MySQL will write the data in the log buffer to the log file each time a transaction is committed. However, the flush operation will not be performed at the same time. In this mode, MySQL will perform a flush operation once per second.

The default value of sync_binlog is 1 and can be set to [0,N)
When sync_binlog = 0, like the operating system's mechanism for flushing other files, MySQL will not synchronize to disk but rely on the operating system to refresh the binary log.
When sync_binlog = N (N>0), MySQL will use the fdatasync() function to synchronize its binary log to disk every time it writes the binary log N times.

These two parameters can be modified online. If you want to import quickly, you can follow the steps below:

# 1. Enter the MySQL command line to temporarily modify these two parameters set global innodb_flush_log_at_trx_commit = 2;
set global sync_binlog = 2000;

# 2. Execute the SQL script to import mysql -uroot -pxxxxxx testdb < testdb.sql

# 3. After the import is complete, change the parameters back to set global innodb_flush_log_at_trx_commit = 1;
set global sync_binlog = 1;

There is another scenario where you need to create a new slave database or do not need to generate binlog. In this case, you can set it not to record binlog temporarily when importing SQL scripts. You can add set sql_log_bin=0; at the beginning of the script and then execute the import, which will further speed up the process. If binlog is not enabled on your MySQL instance, you do not need to execute this statement.

Summarize:

This article mainly introduces methods for quickly importing data. There may be other methods to quickly import data, such as loading data or writing programs for multi-threaded insertion. The method described in this article is only suitable for manually importing SQL scripts. The following is a summary of the methods mentioned in this article.

  1. Use the command line tool that comes with MySQL to export and import.
  2. Use the extended insert method to insert multiple values ​​at once.
  3. Temporarily modify the innodb_flush_log_at_trx_commit and sync_binlog parameters.
  4. Close binlog or temporarily stop recording binlog.

In fact, there are some other solutions, such as not creating the index first, and then adding the index after inserting the data; or changing the table to MyISAM or MEMORY engine first, and then changing it to InnoDB engine after the import is complete. However, these two methods are troublesome to implement and their effects are unknown. The above methods are just the author’s summary based on personal experience, which may not be comprehensive. You are welcome to add to them.

The above is the details of how to quickly import data into MySQL. For more information about importing data into MySQL, please pay attention to other related articles on 123WORDPRESS.COM!

You may also be interested in:
  • Solution for importing more data from MySQL into Hive
  • MySQL 4 methods to import data
  • How to use Navicat to export and import mysql database
  • Java uses MYSQL LOAD DATA LOCAL INFILE to import large amounts of data into MySQL
  • Detailed explanation of MySQL command line export and import database instance
  • Steps to import mysql database under wampserver
  • MySQL command line export and import database
  • Use mysqldump to import data and mysqldump incremental backup (mysqldump usage)
  • Several different ways to import data into MYSQL
  • Modify php.ini to implement the maximum limit of Mysql import database file modification method
  • Steps for importing tens of millions of data into MySQL using .Net Core

<<:  jQuery implements the function of adding and deleting employee information

>>:  Windows 10 is too difficult to use. How to customize your Ubuntu?

Recommend

Solve MySQL deadlock routine by updating different indexes

The previous articles introduced how to debug loc...

MySQL query optimization: causes and solutions for slow queries

Friends who are doing development, especially tho...

Detailed explanation of WeChat Mini Program official face verification

The mini program collected user personal informat...

How to automatically import Vue components on demand

Table of contents Global Registration Partial Reg...

Detailed explanation of linux nslookup command usage

[Who is nslookup?] 】 The nslookup command is a ve...

MySQL database operations (create, select, delete)

MySQL Create Database After logging into the MySQ...

Native js to realize bouncing ball

On a whim, I wrote a case study of a small ball b...

Use IISMonitor to monitor web pages and automatically restart IIS

Table of contents 1. Tool Introduction 2. Workflo...

Docker starts in Exited state

After docker run, the status is always Exited Sol...

Detailed steps to install docker in 5 minutes

Installing Docker on CentOS requires the operatin...

SQL implementation of LeetCode (177. Nth highest salary)

[LeetCode] 177.Nth Highest Salary Write a SQL que...

Data storage implementation method in WeChat applet

Table of contents Global variable globalData Page...

Vue2.x - Example of using anti-shake and throttling

Table of contents utils: Use in vue: explain: Ima...