| QwenLM/Qwen |
20,293 |
|
0 |
0 |
2 months ago |
0 |
|
139 |
apache-2.0 |
Python |
| The official repo of Qwen (通义千问) chat & pretrained large language model proposed by Alibaba Cloud. |
| InternLM/InternLM |
4,412 |
|
0 |
0 |
about 2 years ago |
0 |
|
18 |
apache-2.0 |
Python |
| Official release of InternLM2 7B and 20B base and chat models. 200K context support |
| lonePatient/awesome-pretrained-chinese-nlp-models |
3,738 |
|
0 |
0 |
about 2 years ago |
0 |
|
1 |
mit |
Python |
| Awesome Pretrained Chinese NLP Models,高质量中文预训练模型&大模型&多模态模型&大语言模型集合 |
| brightmart/albert_zh |
3,723 |
|
0 |
0 |
over 3 years ago |
0 |
|
97 |
|
Python |
| A LITE BERT FOR SELF-SUPERVISED LEARNING OF LANGUAGE REPRESENTATIONS, 海量中文预训练ALBERT模型 |
| CLUEbenchmark/CLUE |
3,345 |
|
0 |
0 |
almost 3 years ago |
0 |
|
73 |
|
Python |
| 中文语言理解测评基准 Chinese Language Understanding Evaluation Benchmark: datasets, baselines, pre-trained models, corpus and leaderboard |
| huawei-noah/Pretrained-Language-Model |
2,912 |
|
0 |
0 |
about 2 years ago |
0 |
|
108 |
|
Python |
| Pretrained language model and its related optimization techniques developed by Huawei Noah's Ark Lab. |
| OFA-Sys/Chinese-CLIP |
2,816 |
|
0 |
1 |
over 2 years ago |
9 |
January 13, 2023 |
57 |
mit |
Python |
| Chinese version of CLIP which achieves Chinese cross-modal retrieval and representation generation. |
| OFA-Sys/OFA |
2,142 |
|
0 |
0 |
over 2 years ago |
0 |
|
90 |
apache-2.0 |
Python |
| Official repository of OFA (ICML 2022). Paper: OFA: Unifying Architectures, Tasks, and Modalities Through a Simple Sequence-to-Sequence Learning Framework |
| imcaspar/gpt2-ml |
1,674 |
|
0 |
0 |
almost 3 years ago |
0 |
|
22 |
apache-2.0 |
Python |
| GPT2 for Multiple Languages, including pretrained models. GPT2 多语言支持, 15亿参数中文预训练模型 |
| ymcui/Chinese-ELECTRA |
1,253 |
|
0 |
0 |
about 3 years ago |
0 |
|
0 |
apache-2.0 |
Python |
| Pre-trained Chinese ELECTRA(中文ELECTRA预训练模型) |