Industry: Knowledge Sharing
Author: Mengyu Hu (Platform Engineer at Zhihu)
Transcreator: Ran Huang; Editor: Tom Dewan
Zhihu is China’s largest knowledge sharing platform, with over 220 million registered users and more than 310 million answers on our website. In March 2021, we went public on the New York Stock Exchange and were valued at about $5.3 billion.
As our business grew, we wanted to improve our infrastructure. We horizontally scaled Apache Hive Metastore by migrating from MySQL to TiDB, a distributed SQL database, and decided to put TiDB into other big data applications as well. After migration, the execution time of large queries decreased from 30–40 seconds to 6–7 seconds. We were thrilled by the power of TiDB.
In this article, I’ll show you two examples of how TiDB boosts the performance of our big data architecture. You’ll see how we use TiDB to build a real-time alert system for large queries in Hive and to accelerate NameNode RPC requests in HDFS.
Alert for large queries in Hive
Apache Hive is a data warehouse that allows you to query and analyze data. At Zhihu, we use Hive for extract, transform and load (ETL) tasks and ad-hoc queries.
Our pain points
For ETL tasks, the SQL statements are fairly stable, while ad-hoc queries often contain erratic SQL statements. It’s important to optimize ad-hoc queries; otherwise, the MapReduce job might scan too much data, which slows down task execution and exerts huge pressure on the Hadoop Distributed File System (HDFS). The whole cluster is thus unstable.
This issue frequently comes up when users try to query quarterly or annual data. When Hive serves such queries, the cluster often runs short of resources. Without sufficient resources, ETL tasks get stuck, and the reports are delayed.
A large query alerting system
To solve this problem, we developed a system to alert users about large queries in real time. When a user sends an SQL statement to Hive, the system:
- Parses the execution plan and transforms the plan into the table paths and partition paths that Hive needs to scan.
- Aggregates all the partition paths and calculates the size of data to be scanned.
- Determines whether the data size exceeds the threshold. If it does, the system alerts the user.
Obtain the HDFS paths
After each SQL statement is parsed, we use Hive server’s hook mechanism to output an audit log to Kafka. The log format is as follows:
"time": "2021-07-12 15:43:16.022",
"sql": "select count(*) from test_table where pdate in ('2021-07-01','2021-07-02')",
Note the following fields:
Calculate the partition size
To calculate the partition size, we needed to know the size of each directory in
inputPaths. We decided to parse HDFS
fsimage every day and calculate the size of each Hive directory, and then save the results in TiDB.
Because other applications also use
fsimage, we saved not only the Hive directory, but also the whole HDFS directory, totaling tens of billions of records. TiDB is an excellent choice for storing and indexing large amounts of data.
After the system calculates the data size, it determines whether to send a real-time alert to the user and let them know the possible risk. In the alerting process:
- Hive server sends the audit logs to Kafka in real time.
- Flink consumes data from Kafka and treats Kafka as a streaming table via Kafka Table Source.
- Flink uses JDBC Lookup Table Source to treat TiDB as a dimensional table.
- Flink calculates the data size scanned by each SQL query and decides whether to send alerts.
In the end, if a user writes a SQL statement that scans too much data, they receive an alert like this:
Accelerate NameNode RPC requests
In addition to sending alerts for large queries, we also used TiDB to speed up NameNode performance.
NameNode performance issue
In the past, many users complained that Hive queries took minutes or even hours. We investigated the issue and found the problem lies in the remote procedure call (RPC). When Hive calls the
getInputSummary method, it obtains a global lock. No other query can access the data until the lock is released. If Hive executes a large query, calling this method might take a long time. As a result, other query threads must wait.
We read the Hive source code and found that Hive can execute multiple
getInputSummary methods concurrently. (Internally, this method called HDFS's
getContentSummary method.) We removed the global lock and replaced it with a thread pool-like approach. Now,
getContentSummary could be executed with high concurrency.
However, another issue arose. HDFS’s
getContentSummary method is similar to a filesystem disk usage (
du) operation. When
getContentSummaryis executed in high concurrency, NameNode performance may significantly slow. Because other compute engines also use this method, it's important to optimize it.
Early in 2019, Zhihu split the filesystem into HDFS Federation by using the Router-based Federation. In this architecture, we introduced a new component called Router that forwards requests to NameNode.
Our resolution to the NameNode performance issue is adding cache for HDFS content summary in the Router layer. Specifically, we generate the content summary of all directories every day using
fsimage and store the cache in TiDB.
When the client sends a request, HDFS tries to look for data in the cache. If it doesn’t find the requested data, it fetches data from NameNode and updates the cache. A downside is that because a parent directory cannot detect the changes in a child directory, only leaf directories can be cached. In this implementation, very few requests go to NameNode, so NameNode performance is unaffected.
This solution is convenient because we already generated the content summary when we built the Hive large query alerting system. After we saved the cache in TiDB and created indexes for the requested paths, the latency for ordinary
getContentSummary requests dropped from several minutes to less than 10 ms.
Our future plans
Thanks to TiDB’s horizontal scalability to store tremendous amounts of data and its indexing capabilities to accelerate queries, we successfully stored the HDFS metadata to meet Zhihu’s business requirements.
As a hybrid transactional and analytical (HTAP) database, TiDB has a lot of potential that we can tap into. For example, we can cache HDFS file metadata in real time and subscribe to file changes from EditLog. By merging the file changes with
fsimage in TiDB, we can generate low-latency snapshots of NameNode and use them for online analytics.
Going forward, TiDB will continue to play an important part in Zhihu’s infrastructure and support our hyper growth. We look forward to applying it in more of our big data architecture.
Originally published at www.pingcap.com on Aug 17, 2021