Demand background Part of the data in the business table tb_image is as follows, where id is unique but image_no is not unique. image_no represents the number of each file. Each file will generate several files in the business system. The unique ID of each file is the field id: Some information about the business table tb_image is as follows:
Solution Selection Based on the above analysis of the business, it is completely unnecessary to separate the database and tables. If a single database is sharded, since queries need to be based on image_no and id, one solution is to use redundant sharding (that is, one copy of the data is saved with image_no as the sharding key, and the other copy of the data is saved with id as the sharding key); another solution is to use only image_no as the sharding key, and based on the query requirements of id, the business layer merges the results or introduces third-party middleware. Considering that it is more complicated to divide a single database into tables, we decided to use the partition feature. In addition, the capacity evaluation partition table solution with 128 partitions (each partition has a kw-level data volume) can fully ensure that the business will run stably for at least 15 years (the orange part in the figure is more in line with the actual growth of our business): In addition, since RANGE, LIST, and HASH partitions do not support VARCHAR columns, it is decided to use KEY partitions. The official introduction of its principle is to use the MySQL built-in hash algorithm and then take the modulus of the partition number. Performance Testing After selecting the shard key as image_no and deciding the number of partitions to be 128, it is time to load data for feasibility and performance testing. The reason why the number of partitions is chosen as 128 is: 1.1 billion/1kw=110≈128. In addition, programmers like to use the power of 2, you know. However, the partition number 128 is where all the nightmares begin. I tried to insert 100,000 data into 128 partitions. After the insertion, a surprising phenomenon occurred: there was no data in all odd-numbered partitions (p1, p3, p5, …, p2n-1). At the same time, any even-numbered partition had a lot of data, and it was not very even. As shown in the following figure: Note: The ibd file size of odd-numbered partitions is 112k. This is the initialization size when the partition table is created. There is actually no data. We can verify it through SQL: select partition_name, partition_expression, table_rows from information_schema.partitions where table_schema = schema() and table_name='image_subpart' ;, and some of the results are shown in the following figure: Aren’t 100,000 pieces of data enough to illustrate the problem? On average, each partition has nearly 800 pieces of data! Ok, let’s do something drastic: I will insert another 990w pieces of data, for a total of 1kw of data. The result is still the same, odd-numbered partitions have no data, and even-numbered partitions have partitions. Questions to think about Let's recall the principle of KEY partitioning: Calculate the hash value of the shard key through the MySQL built-in hash algorithm and then take the modulus of the partition number. This principle can also be found on the MySQL official website. Please click on the link: 22.2.5 KEY Partitioning: https://dev.mysql.com/doc/refman/5.7/en/partitioning-key.html. The original text is as follows:
**There can’t be such a crappy hash algorithm in this world, right? **No matter what algorithm you write, it won’t be so uneven, right? At this time, I wonder if there is some configuration causing it. But there are no partition-related variables in show variables. At this time, ten thousand horses galloped past. Could it be that the document and source code are out of sync? Well, let's look at the MySQL source code. After all, the source code is the closest to the truth. The source code related to KEY partition is in the file sql_partition.cc. The author intercepted some key source codes as shown below. At first glance, there is nothing wrong. First calculate the hash value of the partition field and then take the modulus of the partition number: /** Calculate part_id for (SUB)PARTITION BY KEY @param file Handler to storage engine @param field_array Array of fields for PARTTION KEY @param num_parts Number of KEY partitions @param func_value[out] Returns calculated hash value @return Calculated partition id */ inline static uint32 get_part_id_key(handler *file, Field **field_array, uint num_parts, long long *func_value) { DBUG_ENTER("get_part_id_key"); // Calculate the hash value of the partition field *func_value = file->calculate_key_hash_value(field_array); // Modulo the number of partitions DBUG_RETURN((uint32) (*func_value % num_parts)); } In despair, please search for "uneven KEY partition data" on the search engine. In the CSDN forum (https://bbs.csdn.net/topics/390857704) in the search results, there is a folk master Hua Xia Xiao Zu who answered as follows: A colleague analyzed and measured the password function and found that for key partitioning, the number of partitions can only be specified as a prime number to ensure that each partition has data. I tested it, from 11 partitions to 17 partitions. Only the data in partitions 11, 13, and 17 are basically evenly distributed. At this time, another ten thousand horses galloped past. But while I was wondering WHAT THE F**K, I was also a little excited because I might have found a solution (although I still don’t know why the built-in hash algorithm of MySQL behaves like this). Finally, I tested the KEY partition again and concluded as follows:
As shown in the figure below, this is the situation after the author adjusted the number of partitions to 127 and inserted 1 million data. SQL proves that the amount of data in each partition is almost the same: Summary I was shocked to learn that there is no official explanation for the huge pitfalls of using MySQL's KEY partition. In addition there is MySQL bug: Bug #72428 Partition by KEY() results in uneven data distribution Students who are reading this article and have a strong interest can try to delve deeper into this issue. The author will also find some time to dig deeper into the MySQL source code to find out why the implementation of its hash algorithm is so sensitive to the number of partitions. This is the end of this article about the bloody case caused by KEY partition of MySQL learning records. For more relevant MySQL KEY partition bloody case content, please search 123WORDPRESS.COM's previous articles or continue to browse the following related articles. I hope everyone will support 123WORDPRESS.COM in the future! You may also be interested in:
|
<<: JavaScript anti-shake and throttling explained
>>: Docker Compose installation methods in different environments
1. In Windows system, many software installations...
1: nginx server solution, modify the .conf config...
1. Download MySQL 1. Log in to the official websi...
Effect picture: 1. Introduction Your own applet n...
Preface Backup is the basis of disaster recovery....
If the field storing the name uses the GBK charac...
1. Log4j for log segmentation 1) Prepare three pa...
The props of the component (props is an object) F...
Navigation, small amount of data table, centered &...
Preface Transactional data dictionary and atomic ...
Table of contents 1. Array.at() 2. Array.copyWith...
Recently, due to business reasons, I need to acce...
When working on a recent project, I found that th...
In CSS3, the transform function can be used to im...
This article shares the specific code of JavaScri...