chinese-llama-alpaca

中文LLaMA&Alpaca大语言模型+本地CPU/GPU训练部署 (Chinese LLaMA & Alpaca LLMs)

https://github.com/ymcui/chinese-llama-alpaca

Science Score: 46.0%

This score indicates how likely this project is to be science-related based on various indicators:

  • CITATION.cff file
  • codemeta.json file
    Found codemeta.json file
  • .zenodo.json file
    Found .zenodo.json file
  • DOI references
  • Academic publication links
    Links to: arxiv.org
  • Committers with academic emails
    1 of 11 committers (9.1%) from academic institutions
  • Institutional organization owner
  • JOSS paper metadata
  • Scientific vocabulary similarity
    Low similarity (6.3%) to scientific vocabulary

Keywords

alpaca alpaca-2 large-language-models llama llama-2 llm lora nlp plm pre-trained-language-models quantization

Keywords from Contributors

64k llama2 rlhf interactive network-simulation hacking embedded optim standardization projection
Last synced: 6 months ago · JSON representation

Repository

中文LLaMA&Alpaca大语言模型+本地CPU/GPU训练部署 (Chinese LLaMA & Alpaca LLMs)

Basic Info
Statistics
  • Stars: 18,913
  • Watchers: 183
  • Forks: 1,880
  • Open Issues: 5
  • Releases: 12
Topics
alpaca alpaca-2 large-language-models llama llama-2 llm lora nlp plm pre-trained-language-models quantization
Created almost 3 years ago · Last pushed 7 months ago
Metadata Files
Readme License Citation

README.md

Chinese-LLaMA-Alpaca-3

**** | English | /Docs | /Issues | /Discussions | /Arena



GitHub GitHub release (latest by date) GitHub top language GitHub last commit

LLaMAAlpacaNLPLLaMAAlpaca

V2[Cui, Yang, and Yao] Efficient and Effective Text Encoding for Chinese LLaMA and Alpaca


  • LLaMA

- LLaMAAlpaca

Alpaca-Plus-7BCPU


LLaMA-2&Alpaca-2 | LLaMA&Alpaca | VLE | MiniRBT | LERT | PERT | MacBERT | ELECTRA | XLNet | BERT | TextBrewer | TextPruner

[2024/04/30] Chinese-LLaMA-Alpaca-3 Llama-3Llama-3-Chinese-8BLlama-3-Chinese-8B-Instructhttps://github.com/ymcui/Chinese-LLaMA-Alpaca-3

[2024/03/27] SOTA!https://sota.jiqizhixin.com/project/chinese-llama-alpaca

[2023/08/14] Chinese-LLaMA-Alpaca-2 v2.0Chinese-LLaMA-2-13BChinese-Alpaca-2-13Bhttps://github.com/ymcui/Chinese-LLaMA-Alpaca-2

[2023/07/31] Chinese-LLaMA-Alpaca-2 v1.0https://github.com/ymcui/Chinese-LLaMA-Alpaca-2

[2023/07/19] v5.0: Alpaca-ProPlus-33B

[2023/07/19] LLaMA-2Alpaca-2

[2023/07/10] Beta

[2023/07/07] Chinese-LLaMA-AlpacaLLaMA&Alpaca7B

[2023/06/30] llama.cpp8K contexttransformers4K+ contextPR#705

[2023/06/16] v4.1: C-Eval

[2023/06/08] v4.0: LLaMA/Alpaca-33BprivateGPTC-Eval

| | | | ------------------------------------- | ------------------------------------------------------------ | | | LLaMAAlpaca | | | LoRALLaMA | | | | | | | | | LLaMAAlpaca | | FAQ | | | | |

FacebookLLaMALoRALLaMALLaMA/Alpaca LoRALLaMA

LLaMAAlpaca

| | LLaMA | Alpaca | | :-------------------- | ------------------------------------------------------ | ------------------------------------------------------------ | | | CLM | | | | | ChatGPT | | | | | | [3] | 49953 | 49954=49953+1pad token | | | | [1] | | | | | | | | | | llama.cpp | -p | -ins+ | | text-generation-webui | chat | --cpu | | LlamaChat | "LLaMA" | "Alpaca" | | HF | | --with_prompt | | web-demo | | Alpaca | | LangChain / privateGPT | | Alpaca | | | [2] | ProPlus |

[1] llama.cpp/LlamaChat/HF/web-demo/LangChain
[2] *
*[3] AlpacaLLaMApad token
LLaMA/Alpaca**

ChatGPTAlpacaLLaMA AlpacaProPlus

| | | | [1] | [2] | LoRA[3] | | :------------------------ | :------: | :------: | :--------------------------------------------------------: | :----------------: | :----------------------------------------------------------: | | Chinese-LLaMA-Plus-7B | | 120G | LLaMA-7B | 790M | [HF] [ModelScope] [Baidu] | | Chinese-LLaMA-Plus-13B | | 120G | LLaMA-13B | 1.0G | [HF] [ModelScope] [Baidu] | | Chinese-LLaMA-Plus-33B | | 120G | LLaMA-33B | 1.3G[6] | [HF] [ModelScope] [Baidu] | | Chinese-Alpaca-Pro-7B | | 4.3M | LLaMA-7B &
LLaMA-Plus-7B
[4] | 1.1G | [HF] [ModelScope] [Baidu] | | Chinese-Alpaca-Pro-13B | | 4.3M | LLaMA-13B &
LLaMA-Plus-13B[4]
| 1.3G | [HF] [ModelScope] [Baidu] | | Chinese-Alpaca-Pro-33B | | 4.3M | LLaMA-33B &
LLaMA-Plus-33B[4]
| 2.1G | [HF] [ModelScope] [Baidu] |

[1] LLaMALLaMAPR
[2] LLaMA
[3] SHA256SHA256.md
[4] Alpaca-PlusLLaMA-Plus
[5] 30BFacebook33B
[6] FP16

Chinese-LLaMA-7B

chinese_llama_lora_7b/ - adapter_config.json # LoRA - adapter_model.bin # LoRA - special_tokens_map.json # special_tokens_map - tokenizer_config.json # tokenizer - tokenizer.model # tokenizer

****

| | | | | | LoRA | | :---------------- | :------: | :------: | :--------------------: | :----------------: | :----------------------------------------------------------: | | Chinese-LLaMA-7B | | 20G | LLaMA-7B | 770M | [HF] [ModelScope] [Baidu] | | Chinese-LLaMA-13B | | 20G | LLaMA-13B | 1.0G | [HF] [ModelScope] [Baidu] | | Chinese-LLaMA-33B | | 20G | LLaMA-33B | 2.7G | [HF] [ModelScope] [Baidu] | | Chinese-Alpaca-7B | | 2M | LLaMA-7B | 790M | [HF] [ModelScope] [Baidu] | | Chinese-Alpaca-13B | | 3M | LLaMA-13B | 1.1G | [HF] [ModelScope] [Baidu] | | Chinese-Alpaca-33B | | 4.3M | LLaMA-33B | 2.8G | [HF] [ModelScope] [Baidu] | | Chinese-Alpaca-Plus-7B | | 4M | LLaMA-7B &
LLaMA-Plus-7B
| 1.1G | [HF] [ModelScope] [Baidu] | | Chinese-Alpaca-Plus-13B | | 4.3M | LLaMA-13B &
LLaMA-Plus-13B
| 1.3G | [HF] [ModelScope] [Baidu] | | Chinese-Alpaca-Plus-33B | | 4.3M | LLaMA-33B &
LLaMA-Plus-33B
| 2.1G | [HF] [ModelScope] [Baidu] |

transformers

Model HubtransformersPEFTLLaMAAlpaca LoRA.from_pretrained()

https://huggingface.co/hfl

LoRALLaMA

| | | | | :----------- | :--------------------------------------------------------- | :----------------------------------------------------------: | | **** | Colabnotebook | | | **** | | |

FP164-bit

| | 7B | 13B | 33B | 65B | | :------------------ | :----: | :-----: | :-----: | :-----: | | FP16 | 13 GB | 24 GB | 60 GB | 120 GB | | 8-bit | 7.8 GB | 14.9 GB | 32.4 GB | ~60 GB | | 4-bit | 3.9 GB | 7.8 GB | 17.2 GB | 38.5 GB |

GitHub Wiki

| | | | CPU | GPU | | | | | :----------------------------------------------------------- | -------------------------------------------- | :---: | :--: | :--: | :------: | :------: | :----------------------------------------------------------: | | llama.cpp | | | | | | | link | | Transformers | transformers | | | | | | link | | text-generation-webui | Web UI | | | | | | link | | LlamaChat | macOS | MacOS | | | | | link | | LangChain | LLM | | | | | | link | | privateGPT | LangChain | | | | | | link | | Colab Gradio Demo | ColabGradioWeb | | | | | | link | | API | OpenAI APIDemo | | | | | | link |

: LangChainLangChain

GitHub Wiki

promptAlpaca-7BAlpaca-13BAlpaca-33BAlpaca-Plus-7BAlpaca-Plus-13B

NLUmappingC-Eval12.3K52validtestAverage

| | Valid (zero-shot) | Valid (5-shot) | Test (zero-shot) | Test (5-shot) | | ----------------------- | :---------------: | :------------: | :--------------: | :-----------: | | Chinese-Alpaca-Plus-33B | 46.5 | 46.3 | 44.9 | 43.5 | | Chinese-Alpaca-33B | 43.3 | 42.6 | 41.6 | 40.4 | | Chinese-Alpaca-Plus-13B | 43.3 | 42.4 | 41.5 | 39.9 | | Chinese-Alpaca-Plus-7B | 36.7 | 32.9 | 36.4 | 32.3 | | Chinese-LLaMA-Plus-33B | 37.4 | 40.0 | 35.7 | 38.3 | | Chinese-LLaMA-33B | 34.9 | 38.4 | 34.6 | 39.5 | | Chinese-LLaMA-Plus-13B | 27.3 | 34.0 | 27.8 | 33.3 | | Chinese-LLaMA-Plus-7B | 27.3 | 28.3 | 26.9 | 28.4 |

C-Eval >>> GitHub Wiki

GitHub Wiki

FAQ

FAQIssueFAQ

1 233B65B 3 4LLaMA 5 6Windows 7Chinese-LLaMA 13Bllama.cpp 8Chinese-Alpaca-Plus 9NLU 1033B30B 11SHA256

GitHub Wiki

-

  • demo

https://arxiv.org/abs/2304.08177 @article{chinese-llama-alpaca, title={Efficient and Effective Text Encoding for Chinese LLaMA and Alpaca}, author={Cui, Yiming and Yang, Ziqing and Yao, Xin}, journal={arXiv preprint arXiv:2304.08177}, url={https://arxiv.org/abs/2304.08177}, year={2023} }

| | | | | :----------------------------------------------------------- | :----------------------------- | :----: | | Chinese-LLaMA-Alpaca-2 | LLaMA-2Alpaca-2 | | | Visual-Chinese-LLaMA-Alpaca | LLaMA & Alpaca | |

| | | | | :----------------------------------------------------------: | :----------------------------------------------------------: | :----------------------------------------------------------: | | LLaMA by Facebook
Alpaca by Stanford
alpaca-lora by @tloen | llama.cpp by @ggerganov
LlamaChat by @alexrozanski
text-generation-webui by @oobabooga | pCLUE and MT data by @brightmart
oasst1 by OpenAssistant |


GitHub Issue

Owner

  • Name: Yiming Cui
  • Login: ymcui
  • Kind: user
  • Location: Beijing, China
  • Company: Joint Laboratory of HIT and iFLYTEK Research (HFL)

NLP Researcher. Mainly interested in Pre-trained Language Model, Machine Reading Comprehension, Question Answering, etc.

GitHub Events

Total
  • Issues event: 4
  • Watch event: 843
  • Issue comment event: 4
  • Push event: 1
  • Fork event: 72
Last Year
  • Issues event: 4
  • Watch event: 843
  • Issue comment event: 4
  • Push event: 1
  • Fork event: 72

Committers

Last synced: 6 months ago

All Time
  • Total Commits: 462
  • Total Committers: 11
  • Avg Commits per committer: 42.0
  • Development Distribution Score (DDS): 0.385
Past Year
  • Commits: 1
  • Committers: 1
  • Avg Commits per committer: 1.0
  • Development Distribution Score (DDS): 0.0
Top Committers
Name Email Commits
ymcui me@y****m 284
Ziqing Yang y****g@1****m 115
yaoxin 3****i 27
sunyuhan s****n@z****n 25
GoGoJoestar q****9@1****m 4
iyorke 5****e 2
Light s****y@g****m 1
alkaid 4****o 1
dependabot[bot] 4****] 1
guohao g****e@g****m 1
bigbaldy b****8@g****m 1
Committer Domains (Top 20 + Academic)

Issues and Pull Requests

Last synced: 6 months ago

All Time
  • Total issues: 98
  • Total pull requests: 10
  • Average time to close issues: 14 days
  • Average time to close pull requests: 15 days
  • Total issue authors: 81
  • Total pull request authors: 10
  • Average comments per issue: 4.55
  • Average comments per pull request: 0.2
  • Merged pull requests: 5
  • Bot issues: 0
  • Bot pull requests: 0
Past Year
  • Issues: 2
  • Pull requests: 0
  • Average time to close issues: 22 days
  • Average time to close pull requests: N/A
  • Issue authors: 2
  • Pull request authors: 0
  • Average comments per issue: 2.0
  • Average comments per pull request: 0
  • Merged pull requests: 0
  • Bot issues: 0
  • Bot pull requests: 0
Top Authors
Issue Authors
  • Porraio (3)
  • Lufffya (3)
  • ZenXir (3)
  • Log4J-Ops (2)
  • wizd (2)
  • shmily91 (2)
  • songbaiTalk (2)
  • xunmenglt (2)
  • alexhmyang (2)
  • 27182812 (2)
  • ghost (2)
  • tangpingzhou (2)
  • controZheng (2)
  • IloveTTLHz2023 (1)
  • LainNya (1)
Pull Request Authors
  • airaria (1)
  • wb14123 (1)
  • lindeer (1)
  • ymcui (1)
  • sgsdxzy (1)
  • dumpmemory (1)
  • Yulv-git (1)
  • guohao (1)
  • fireindark707 (1)
  • GoGoJoestar (1)
Top Labels
Issue Labels
stale (30)
Pull Request Labels
stale (1)

Dependencies

.github/workflows/stale.yml actions
  • actions/stale v8 composite
requirements.txt pypi
  • sentencepiece ==0.1.97
  • torch ==1.13.1
  • transformers ==4.30.0