| d2l-ai/d2l-zh |
53,401 |
|
1 |
1 |
about 2 years ago |
51 |
August 18, 2023 |
65 |
apache-2.0 |
Python |
| 《动手学深度学习》:面向中文读者、能运行、可讨论。中英文版被70多个国家的500多所大学用于教学。 |
| QwenLM/Qwen |
20,293 |
|
0 |
0 |
2 months ago |
0 |
|
139 |
apache-2.0 |
Python |
| The official repo of Qwen (通义千问) chat & pretrained large language model proposed by Alibaba Cloud. |
| ymcui/Chinese-BERT-wwm |
8,600 |
|
0 |
0 |
over 2 years ago |
0 |
|
3 |
apache-2.0 |
Python |
| Pre-Training with Whole Word Masking for Chinese BERT(中文BERT-wwm系列模型) |
| brightmart/nlp_chinese_corpus |
8,344 |
|
0 |
0 |
almost 3 years ago |
0 |
|
20 |
mit |
|
| 大规模中文自然语言处理语料 Large Scale Chinese Corpus for NLP |
| brightmart/text_classification |
7,628 |
|
0 |
0 |
over 2 years ago |
0 |
|
45 |
mit |
Python |
| all kinds of text classification models and more with deep learning |
| Morizeyao/GPT2-Chinese |
7,249 |
|
0 |
0 |
over 2 years ago |
0 |
|
105 |
mit |
Python |
| Chinese version of GPT2 training code, using BERT tokenizer. |
| NLPchina/ansj_seg |
6,390 |
|
402 |
17 |
over 2 years ago |
10 |
February 15, 2018 |
50 |
apache-2.0 |
Java |
| ansj分词.ict的真正java实现.分词效果速度都超过开源版的ict. 中文分词,人名识别,词性标注,用户自定义词典 |
| baichuan-inc/Baichuan-7B |
5,680 |
|
0 |
0 |
over 1 year ago |
0 |
|
80 |
apache-2.0 |
Python |
| A large-scale 7B pretraining language model developed by BaiChuan-Inc. |
| HqWu-HITCS/Awesome-Chinese-LLM |
5,477 |
|
0 |
0 |
about 2 years ago |
0 |
|
0 |
|
|
| 整理开源的中文大语言模型,以规模较小、可私有化部署、训练成本较低的模型为主,包括底座模型,垂直领域微调及应用,数据集与教程等。 |
| SCIR-HI/Huatuo-Llama-Med-Chinese |
3,776 |
|
0 |
0 |
over 2 years ago |
0 |
|
14 |
apache-2.0 |
Python |
| Repo for BenTsao [original name: HuaTuo (华驼)], Instruction-tuning Large Language Models with Chinese Medical Knowledge. 本草(原名:华驼)模型仓库,基于中文医学知识的大语言模型指令微调 |