In-depth analysis of MySQL query interception

In-depth analysis of MySQL query interception

1. Query Optimization

1. MySQL tuning outline

  • Enable and capture slow queries
  • explain+slow SQL analysis
  • show profile queries the execution details and life cycle of SQL in the MySQL server
  • Parameter tuning of SQL database server

2. Small table drives large table

The implementation principle of MySQL join is to use the data of the driving table as the basis and use "nested loops" to match records in the driven table. The index of the driving table will be invalid, while the index of the driven table is valid.

#Assume that table a has 10,000 data and table b has 20 data select * from a join b on a.bid =b.id
Table a drives table b as follows:
for 20 data to match 10000 data (based on the join condition on a.bid=b.id, perform a B+ tree search)
The number of searches is: 20+ log10000
Table b drives table a to match 20 data for 10,000 data (based on the join condition on a.bid=b.id, a B+ tree search is performed). The number of searches is: 10,000+ log20

3. in and exists

Use of exists

  • EXISTS syntax: EXISTS(subquery) only returns TRUE or FALSE, so the SELECT * in the subquery can also be SELECT 1 or other. The official statement is that the SELECT list will be ignored during actual execution, so there is no difference
    • SELECT ... FROM table WHERE EXISTS(subquery)
    • This syntax can be understood as: putting the queried data into the subquery for conditional verification, and deciding whether the data results of the main query are retained based on the verification result (TRUE or FALSE).
  • The actual execution process of the EXISTS subquery may be optimized rather than a one-by-one comparison as we understand it. If you are concerned about efficiency issues, you can perform actual tests to determine whether there are any efficiency issues.
  • EXISTS subqueries can often be replaced by conditional expressions, other subqueries, or JOIN. The best solution requires specific analysis of specific issues.
#If in is used, the internal table B drives the external table A
select * from A where id in (select id from B)
#Using exists, external table A drives internal table B
select * from A where exists(select 1 from B where B.id = A.id)

in conclusion:

Always remember that when the B table data set is smaller than the A table data set, use in
When the data set of table A is smaller than that of table B, use exist

4. Create table by orderby

create table tblA(
 #id int primary key not null auto_increment,
 age int,
 birth timestamp not null
);

insert into tblA(age, birth) values(22, now());
insert into tblA(age, birth) values(23, now());
insert into tblA(age, birth) values(24, now());
#Create a composite index create index idx_A_ageBirth on tblA(age, birth); 

Orderby hits the index

Orderby does not hit the index

  • MySQL supports two sorting methods: Using index and Using filesort. Filesort is less efficient, and to use index sorting, two conditions must be met: sorting is done on the index column as much as possible, and the best left prefix of the index is followed.
    • The order by statement itself uses the leftmost column of the index
    • Use the where clause and order by clause condition column combination to meet the leftmost column
  • If the order by is not on the index column, the filesort algorithm will be used: two-way sorting and one-way sorting
    • Prior to MySQL 4.1, two-way sorting was used, which literally means scanning the disk twice to finally get the data. Read the row pointer and order by column, sort them, and then scan the sorted list and re-read the corresponding data transfer from the list according to the value in the list
    • Read all the columns needed for the query from disk, sort them in the buffer according to the order by column, and then scan the sorted list for output. It is faster, avoids reading data a second time, and turns random IO into sequential IO, but it uses more space because it keeps each row in memory.
select * from user where name = "zs" order by age
#Dual sort 1) Find the first primary key id that satisfies name = 'zs' from name
2) Take the entire row based on the primary key id, and put the sort fields age and primary key id into the sort buffer. 3) Take the next primary key id from name that satisfies the record with name = 'zs'.
4) Repeat 2 and 3 until name = 'zs' is no longer satisfied
5) Sort the age field and primary key id in sort_buffer according to the age field 6) Traverse the sorted id and age field, return to the original table according to the id value, retrieve the values ​​of all fields and return them to the client #Single-way sorting 1) Find the first primary key id from name that meets the condition name = 'zs'
2) Take the entire row based on the primary key id, take out the values ​​of all fields, and store them in the sort_buffer (sort buffer) 3) Find the next primary key id that satisfies the condition name = 'zs' from the index name
4) Repeat steps 2 and 3 until name = 'zs' is no longer satisfied
5) Sort the data in sort_buffer by the field age and return the result to the client

Problems and optimization of single-path sorting

question:
 Since the single-pass algorithm is an improved one, it is generally better than the dual-pass algorithm in the sort_buffer. Method B takes up much more space than method A because method B takes out all fields. Therefore, the total size of the data taken out may exceed the capacity of the sort_buffer, resulting in only taking out data up to the capacity of the sort_buffer each time for sorting (creating a tmp file, merging multiple passes), taking out data up to the capacity of the sort_buffer, and sorting again... This will result in multiple I/Os.
Optimization strategy:
 Increase the setting of the sort_buffer_size parameter Increase the setting of the max_length_for_sort_data parameter Notes:
  Selecting * when ordering by is a big taboo. Only query the required fields. Because the more fields there are, the more data must be stored in memory, which results in fewer data columns being loaded per I/O.

5. Groupby optimization

1) Group by actually sorts first and then groups, following the best left prefix of the index
2) When the index column cannot be used, increase the max_length_for_sort_data parameter setting + increase the sort_buffer_size parameter setting
3) where is higher than having. If the conditions can be written in where, do not use having.
4) The rest of the rules are the same as order by

2. Slow query log

1. What is the slow query log?

  1. The MySQL slow query log is a type of log record provided by MySQL. It is used to record statements in MySQL whose response time exceeds the threshold. Specifically, SQL statements whose running time exceeds the long_query_time value will be recorded in the slow query log.
  2. The default value of long_query_time is 10, which means that SQL statements that run for more than 10 seconds will be recorded.
  3. It is used to check which SQL statements exceed our maximum tolerance time value. For example, if a SQL statement takes more than 5 seconds to execute, we will consider it a slow SQL statement. We hope to collect SQL statements that take more than 5 seconds and conduct a comprehensive analysis based on the previous explain.

2. Enable slow query log

By default, the MySQL slow query log is not enabled. If it is not necessary for tuning, it is generally not recommended to enable this parameter, because enabling the slow query log will affect performance. The slow query log supports writing log records to files.

a) Enable slow query log

#Check whether slow log is enabled show variables like 'slow_query_log%';
# Enable the slow query log. To make it permanent, set global slow_query_log = 1 in my.cnf; 

b) Set the threshold for slow query logs

#The default threshold time for viewing slow query logs is 10s
show variables like 'long_query_time%';
#Set to 3s. Restart will fail. If you want it to be permanent, set global long_query_time = 3 in my.cnf
#To view again, you need to switch windows to view show variables like 'long_query_time%'; 

c) Persistent slow query logs and time thresholds

[mysqld]
#Persistent slow query log slow_query_log=1;
slow_query_log_file=/var/lib/mysql/hadoop102-slow.log
long_query_time = 3;
log_output=FILE

d) Slow query case

#Query wait 4s
select sleep(4);
#In Linux system, view the slow query log cat /var/lib/mysql/hadoop102-slow.log

e) Check the number of slow query logs in the current system

show global status like '%Slow_queries%';

3. Log analysis command mysqldumpslow

a) Parameter explanation

-s: indicates the sorting method
c: Number of visits
l: Lock time
r: Return records
t: query time
al: average lock time
ar: average number of returned records
at: average query time
-t: how many records are returned
-g: followed by a regular matching pattern, case-insensitive

b) Common methods

#Get the 10 SQL statements that return the most records
mysqldumpslow -sr -t 10 /var/lib/mysql/hadoop102-slow.log
#Get the 10 most visited SQLs
mysqldumpslow -sc -t 10 /var/lib/mysql/hadoop102-slow.log
#Get the first 10 queries sorted by time that contain left joins mysqldumpslow -st -t 10 -g "left join" /var/lib/mysql/hadoop102-slow.log
#These commands are combined with | and more to use mysqldumpslow -sr -t 10 /var/lib/mysql/hadoop102-slow.log | more

3. Batch writing data scripts

1. Create a table

CREATE TABLE dept
(
 deptno int unsigned primary key auto_increment,
 dname varchar(20) not null default '',
 loc varchar(8) not null default ''
)ENGINE=INNODB DEFAULT CHARSET=utf8;

CREATE TABLE emp
(
 id int unsigned primary key auto_increment,
 empno mediumint unsigned not null default 0,
 ename varchar(20) not null default '',
 job varchar(9) not null default '',
 mgr mediumint unsigned not null default 0,
 hiredate date not null,
 sal decimal(7,2) not null,
 comm decimal(7,2) not null,
 deptno mediumint unsigned not null default 0
)ENGINE=INNODB DEFAULT CHARSET=utf8;

2. Set whether to trust the creator of the stored function

#View binlog status show variables like 'log_bin%';
#Add trusted storage function creators set global log_bin_trust_function_creators = 1; 

3. Create a function

Function to generate random string

# Define two $$ to indicate the end (replace the original ;)
delimiter $$ 
create function rand_string(n int) returns varchar(255)
begin
 declare chars_str varchar(100) default 'abcdefghijklmnopqrstuvwxyzABCDEFGHIJKLMNOPQRSTUVWXYZ';
 declare return_str varchar(255) default '';
 declare i int default 0;
 while i < n do
 set return_str = concat(return_str,substring(chars_str,floor(1+rand()*52),1));
 set i=i+1;
 end while;
 return return_str;
end $$

Function to randomly generate department numbers

delimiter $$
create function rand_num() returns int(5)
begin
 declare i int default 0;
 set i=floor(100+rand()*10);
 return i;
end $$

4. Create a stored procedure

Create a stored procedure to insert data into the emp table

delimiter $$
create procedure insert_emp(in start int(10),in max_num int(10))
begin
 declare i int default 0;
 set autocommit = 0;
 repeat
 set i = i+1;
 insert into emp(empno,ename,job,mgr,hiredate,sal,comm,deptno) values((start+i),rand_string(6),'salesman',0001,curdate(),2000,400,rand_num());
 until i=max_num
 end repeat;
 commit;
end $$

Create a stored procedure to insert data into the dept table

delimiter $$
create procedure insert_dept(in start int(10),in max_num int(10))
begin
 declare i int default 0;
 set autocommit = 0;
 repeat
 set i = i+1;
 insert into dept(deptno,dname,loc) values((start+i),rand_string(10),rand_string(8));
 until i=max_num
 end repeat;
 commit;
end $$

5. Call stored procedures to generate data

#Insert 10 records into the department table DELIMITER;
CALL insert_dept(100, 10);
#Insert 500,000 records into the employee table CALL insert_emp(100001, 500000);

4. show profiles

1. Introduction

  • SHOW PROFILE is a command provided by MySQL that can be used to analyze the resource consumption of statement execution in the current session. Can be used for SQL tuning measurements.
  • By default, the parameter is turned off and the results of the last 15 runs are saved.

2. Open

#Check whether Show Profile is enabled show variables like 'profiling%';
#Open Show Profile
set profiling=on;

3. Use show profiles

Creating Test Data

select * from emp group by id%10 limit 150000;
select * from emp group by id%10 limit 150000;
select * from emp group by id%10 order by 5;
select * from emp
select * from dept
select * from emp left join dept on emp.deptno = dept.deptno

Execute show profiles

Execute show profile cpu, block io for query Query_ID;

Search Parameters

ALL: Display all cost information
BLOCK IO: Displays block IO related overhead
CONTEXT SWITCHES: Context switch related overhead
CPU: Display CPU related overhead information
IPC: Displays information about sending and receiving related overhead
MEMORY: Display memory-related overhead information
PAGE FAULTS: Displays page fault related overhead information
SOURCE: Displays the overhead information related to Source_function, Source_file, and Source_line
SWAPS: Displays information about the number of swaps related to overhead

Return results

converting HEAP to MyISAM: The query results are too large to fit in memory and have to be moved to disk.
Creating tmp table: Create a temporary table. MySQL will copy the data to the temporary table first, and then delete the temporary table after use.
Copying to tmp table on disk: Copying the temporary table in memory to disk, dangerous! ! !
locked: locked table

5. Global Query Log

Do not enable this feature in a production environment

Configure in my.cnf

# Enable general_log=1
# Record the path of the log file general_log_file=/path/logfile
# Output format log_output=FILE

Encoding Enabled

set global general_log=1;
set global log_output='TABLE';

After the configuration is completed, it will be recorded in the general_log table in the MySQL database

select * from mysql.general_log;

Summarize

This is the end of this article about MySQL query interception. For more relevant MySQL query interception content, please search 123WORDPRESS.COM’s previous articles or continue to browse the following related articles. I hope everyone will support 123WORDPRESS.COM in the future!

You may also be interested in:
  • Usage instructions for the Mysql string interception function SUBSTRING
  • MySql uses DATE_FORMAT to intercept the date value of the DateTime field
  • MySQL interception and split string function usage examples
  • Mysql string interception and obtaining data in the specified string
  • Usage of MySQL intercepted string function substring_index
  • mysql intercepts the content between two specified strings
  • MySQL intercepts the sql statement of the string function
  • Instructions for using mysql interception functions
  • Summary of MySQL string interception related functions
  • Loop through user information in MySQL and insert it into the corresponding field of the target table

<<:  WeChat applet custom bottom navigation bar component

>>:  Detailed steps to enable SourceGuardian (sg11) encryption component on Linux virtual hosts

Recommend

Suggestions on creating business HTML emails

Through permission-based email marketing, not onl...

Solve the problem of not finding NULL from set operation to mysql not like

An interesting discovery: There is a table with a...

Vue uses echarts to draw an organizational chart

Yesterday, I wrote a blog about the circular prog...

CSS achieves colorful and smart shadow effects

background Ever wondered how to create a shadow e...

Detailed implementation plan of Vue front-end exporting Excel files

Table of contents 1. Technology Selection 2. Tech...

Using JS to implement a rotating Christmas tree in HTML

<!DOCTYPE HEML PUBLIC> <html> <hea...

How to disable ads in the terminal welcome message in Ubuntu Server

If you are using the latest Ubuntu Server version...

Nginx Linux installation and deployment detailed tutorial

1. Introduction to Nginx Nginx is a web server th...

Summary of common knowledge points required for MySQL

Table of contents Primary key constraint Unique p...

Docker image management common operation code examples

Mirroring is also one of the core components of D...

IE8 uses multi-compatibility mode to display web pages normally

IE8 will have multiple compatibility modes . IE pl...

linux exa command (better file display experience than ls)

Install Follow the README to install The document...

Implementation of Docker container connection and communication

Port mapping is not the only way to connect Docke...