HomeAI Tools
ChatGLM2-6B

ChatGLM2-6B

1 liked
entry-slick
About ChatGLM2-6B

ChatGLM 2 -6B is the second-generation version of the open source Chinese-English bilingual dialogue model ChatGLM-6B . On the basis of retaining many excellent features of the first generation model, such as smooth dialogue and low deployment threshold, ChatGLM 2 -6B introduces the following new features :

  1. More powerful performance : Based on the development experience of the first generation model of ChatGLM, we have fully upgraded the base model of ChatGLM2-6B. ChatGLM2-6B uses the mixed objective function of GLM , and has undergone pre-training of 1.4T Chinese-English identifiers and human preference alignment training. The evaluation results show that compared with the original model, ChatGLM2-6B is better at MMLU (+23%), CEval (+33%), GSM8K (+571%), BBH (+60%) and other data sets have achieved substantial improvement in performance, and have strong competitiveness in open source models of the same size.

    1. Longer context : Based on FlashAttention technology, we extended the context length (Context Length) of the base model from 2K of ChatGLM-6B to 32K, and used 8K context length training in the dialogue stage. For a longer context, we release the ChatGLM2-6B-32K model. The LongBench evaluation results show that among the open source models of the same magnitude, ChatGLM2-6B-32K has a more obvious competitive advantage.
    2. More efficient reasoning : Based on Multi-Query Attention technology, ChatGLM2-6B has more efficient reasoning speed and lower memory usage: under the official model implementation, the reasoning speed has increased by 42% compared with the first generation , under INT4 quantization, the dialogue length supported by 6G video memory is increased from 1K to 8K.
    3. A more open protocol : ChatGLM2-6B weights are completely open to academic research, and free commercial use is also allowed after filling out the questionnaire for registration.

    Welcome to chatglm.cn to experience a larger scale ChatGLM model.

—————

ChatGLM2-6B open source model aims to promote the development of large model technology with the open source community. Developers and everyone are urged to abide by the open source agreement and not to use the open source model And code and derivatives based on open source projects are used for any purposes that may cause harm to the country and society, and for any services that have not undergone security assessment and filing. Currently, the project team has not developed any applications based on ChatGLM2-6B, including web, Android, Apple iOS and Windows App applications.

Although the model tries its best to ensure the compliance and accuracy of the data at all stages of training, the output content cannot be guaranteed due to the small size of the ChatGLM2-6B model and the model is affected by probabilistic randomness factors accuracy, and the model is easily misleading. This project does not assume the risks and responsibilities of data security and public opinion risks caused by open source models and codes, or any risks and responsibilities arising from misleading, misusing, spreading, and improper use of any models.

Update

[2023/07/31] released the ChatGLM2-6B-32K model to improve the ability to understand long texts.

[2023/07/25] released the CodeGeeX2 model, based on ChatGLM2-6B adding code pre-training implementation, the coding ability has been comprehensively improved.

[2023/07/04] released P-Tuning v2 and full parameter fine-tuning script, see P-Tuning .

Friendship link

An open source project that accelerates ChatGLM2:

  • fastllm : full-platform accelerated reasoning solution, single GPU batch reasoning can reach 10000+ tokens per second, the lowest on mobile 3G memory running in real time (about 4~5 token/s on Snapdragon 865)
  • chatglm.cpp : CPU quantization acceleration reasoning scheme similar to llama.cpp, realizing real-time conversation on Mac notebook
  • ChatGLM2- TPU : Using the TPU acceleration inference solution, running about 3 token/s in real time on the end-side computing chip BM1684X (16T@FP16, 16G memory) Project:

  • ChatGLM2-6B deployment and fine-tuning tutorial

Visit Official Website

https://github.com/THUDM/ChatGLM2-6B

Community Posts
no data
Nothing to display