MITB Banner

AI Researchers Develop New NLU Benchmark For Chinese Language

Share

Over the last few years, there have been significant advancements in the research of Chinese natural language understanding (NLU) and natural language processing (NLP). In order to make robust Chinese NLP models, various high-quality corpora of speech, text and others have been built by the researchers. 

The General Language Understanding Evaluation (GLUE) benchmark which was introduced in 2018 is a collection of resources for training, evaluating, and analyzing natural language understanding systems. These systems consist of a benchmark of sentence-pair language understanding tasks, a diagnostic dataset, and a public leaderboard for tracking performance along with a dashboard for visualising the performance of models on the diagnostic sets.  

Taking these two points into consideration — Chinese natural language processing (CNLP) and General Language Understanding Evaluation (GLUE) — researchers at Tsinghua University, Zhejiang University, and Peking University developed a model, known as ChineseGLUE

Currently, the ChineseGLUE includes datasets for Natural Language Inference, Sentiment Analysis for Internet News, Short Text Classification for News, Reading Comprehension for Traditional Chinese, etc. such as LCQMC colloquial description semantic similarity task, XNLI language inference task, TNEWS Today’s headline Chinese news (short text) classification, Delta Reading Comprehension Dataset (DRCD) Traditional Reading Comprehension and INEWS Internet sentiment analysis task.  

This Language Understanding Evaluation benchmark for Chinese (ChineseGLUE) consists of some specific features which bridge the gap between the popular GLUE and the Chinese language.

  • Benchmarking of Chinese Tasks, Covering Multiple Language Tasks of Different Degrees: This model includes a benchmark of several sentences or sentence pair language understanding tasks. Currently, the datasets used in these tasks are collected from the public and further, the researchers will include datasets with a private test set.
  • Open Leaderboard: ChineseGLUE consists of a public leaderboard for tracking performance. One can easily submit the prediction files on these tasks where each task will be evaluated and scored.
  • Baseline Model, Including The Starting Code, Pre-Training Model: The baseline models of ChineseGLUE tasks will be available in TensorFlow, PyTorch, Keras, and PaddlePaddle.
  • Corpus For Language Modeling, Pre-Training Or Generative Tasks: The researchers aim to include a large volume of the raw corpus for pre-train or language modelling research purposes. Also, it has been said that they will include at least 30G raw corpus by the end of 2020. One can use it for general purpose or domain adaption, or even for text generating purposes.

Why ChineseGLUE

There are several crucial reasons which the researchers put forward as to why this model is necessary. They are mentioned below

  • The Chinese language is a large language with its own specific and numerous applications.
  • Comparing to the English dataset, there are relatively few publicly available datasets in the Chinese language.
  • Language understanding has evolved to the current stage, and the pre-training model has greatly promoted natural language understanding. In this case, if there is a benchmark for Chinese tasks, it includes a set of data sets that can be widely used and evaluated by the public, the characteristics of Chinese-language tasks, and the development of current world technologies.

Wrapping Up

GLUE is a popular benchmark for evaluating natural language processing but it does not work for the evaluation of Chinese NLP models. Thus, the researchers developed this model in order to evaluate natural language understand for Chinese. ChineseGLUE has similar properties as GLUE, i.e. datasets, baselines, pre-trained models, corpus and leaderboard. 

The motives behind this model are to serve better Chinese language understanding, tasks, and industry, as a supplement to the common language model assessment and to promote the development of the Chinese language model by improving the way the Chinese language understands the infrastructure.

Share
Picture of Ambika Choudhury

Ambika Choudhury

A Technical Journalist who loves writing about Machine Learning and Artificial Intelligence. A lover of music, writing and learning something out of the box.
Related Posts

CORPORATE TRAINING PROGRAMS ON GENERATIVE AI

Generative AI Skilling for Enterprises

Our customized corporate training program on Generative AI provides a unique opportunity to empower, retain, and advance your talent.

Upcoming Large format Conference

May 30 and 31, 2024 | 📍 Bangalore, India

Download the easiest way to
stay informed

Subscribe to The Belamy: Our Weekly Newsletter

Biggest AI stories, delivered to your inbox every week.

AI Courses & Careers

Become a Certified Generative AI Engineer

AI Forum for India

Our Discord Community for AI Ecosystem, In collaboration with NVIDIA. 

Flagship Events

Rising 2024 | DE&I in Tech Summit

April 4 and 5, 2024 | 📍 Hilton Convention Center, Manyata Tech Park, Bangalore

MachineCon GCC Summit 2024

June 28 2024 | 📍Bangalore, India

MachineCon USA 2024

26 July 2024 | 583 Park Avenue, New York

Cypher India 2024

September 25-27, 2024 | 📍Bangalore, India

Cypher USA 2024

Nov 21-22 2024 | 📍Santa Clara Convention Center, California, USA

Data Engineering Summit 2024

May 30 and 31, 2024 | 📍 Bangalore, India

Subscribe to Our Newsletter

The Belamy, our weekly Newsletter is a rage. Just enter your email below.