MSBD6000N_Understanding_Large_Language_Models

The Project Codes for MSBD 6000N

https://github.com/HKUST-Franky/MSBD6000N_Understanding_Large_Language_Models

Science Score: 44.0%

This score indicates how likely this project is to be science-related based on various indicators:

  • CITATION.cff file
    Found CITATION.cff file
  • codemeta.json file
    Found codemeta.json file
  • .zenodo.json file
    Found .zenodo.json file
  • DOI references
  • Academic publication links
  • Academic email domains
  • Institutional organization owner
  • JOSS paper metadata
  • Scientific vocabulary similarity
    Low similarity (3.6%) to scientific vocabulary
Last synced: 6 months ago · JSON representation ·

Repository

The Project Codes for MSBD 6000N

Basic Info
  • Host: GitHub
  • Owner: HKUST-Franky
  • License: apache-2.0
  • Language: Python
  • Default Branch: main
  • Size: 236 MB
Statistics
  • Stars: 0
  • Watchers: 1
  • Forks: 0
  • Open Issues: 0
  • Releases: 0
Created over 1 year ago · Last pushed about 1 year ago
Metadata Files
Readme License Citation

README.md

PersonaCraft Project Codespace

The 6000N Project Files are implemented using the open-source training framework LLaMA Factory.

Pre-processed Datasets

  • /wxh_scripts/*.py
  • /data/RoleBench
  • /data/rolellm_*.json
  • /data/identity_wxh.json
  • RoleBench: https://huggingface.co/datasets/ZenMoore/RoleBench ## Evaluation Results
  • /saves/Llama-3.1-8B-Instruct/lora/*

Training Logs and Adpators

  • /src/saves/Llama-3.1-8B-Instruct/lora/2024-11-26-14-59-53_dpo
  • /src/saves/Llama-3.1-8B-Instruct/lora/train_2024-11-26-03-03-10 https://huggingface.co/FrankyCN/PersonaCraft/tree/main/Llama-3.1-8B-Instruct/lora

Models used

  • https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct

Owner

  • Name: WU FUNG Xian Hong
  • Login: HKUST-Franky
  • Kind: user
  • Location: Hong Kong
  • Company: HKUST

MSc in Big Data Technology at HKUST

Citation (CITATION.cff)

cff-version: 1.2.0
date-released: 2024-03
message: "If you use this software, please cite it as below."
authors:
- family-names: "Zheng"
  given-names: "Yaowei"
- family-names: "Zhang"
  given-names: "Richong"
- family-names: "Zhang"
  given-names: "Junhao"
- family-names: "Ye"
  given-names: "Yanhan"
- family-names: "Luo"
  given-names: "Zheyan"
- family-names: "Feng"
  given-names: "Zhangchi"
- family-names: "Ma"
  given-names: "Yongqiang"
title: "LlamaFactory: Unified Efficient Fine-Tuning of 100+ Language Models"
url: "https://arxiv.org/abs/2403.13372"
preferred-citation:
  type: conference-paper
  conference:
    name: "Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 3: System Demonstrations)"
  authors:
    - family-names: "Zheng"
      given-names: "Yaowei"
    - family-names: "Zhang"
      given-names: "Richong"
    - family-names: "Zhang"
      given-names: "Junhao"
    - family-names: "Ye"
      given-names: "Yanhan"
    - family-names: "Luo"
      given-names: "Zheyan"
    - family-names: "Feng"
      given-names: "Zhangchi"
    - family-names: "Ma"
      given-names: "Yongqiang"
  title: "LlamaFactory: Unified Efficient Fine-Tuning of 100+ Language Models"
  url: "https://arxiv.org/abs/2403.13372"
  year: 2024
  publisher: "Association for Computational Linguistics"
  address: "Bangkok, Thailand"

GitHub Events

Total
Last Year

Dependencies

pyproject.toml pypi
requirements.txt pypi
  • accelerate >=0.30.1,<=0.34.2
  • av *
  • datasets >=2.16.0,<=2.21.0
  • einops *
  • fastapi *
  • fire *
  • gradio >=4.0.0
  • matplotlib >=3.7.0
  • numpy <2.0.0
  • packaging *
  • pandas >=2.0.0
  • peft >=0.11.1,<=0.12.0
  • protobuf *
  • pydantic *
  • pyyaml *
  • scipy *
  • sentencepiece *
  • sse-starlette *
  • tiktoken *
  • transformers >=4.41.2,<=4.45.2
  • trl >=0.8.6,<=0.9.6
  • uvicorn *
setup.py pypi
src/llamafactory.egg-info/requires.txt pypi
  • accelerate <=0.34.2,>=0.30.1
  • adam-mini *
  • aqlm >=1.1.0
  • auto-gptq >=0.5.0
  • autoawq *
  • av *
  • badam >=1.2.1
  • bitsandbytes >=0.39.0
  • datasets <=2.21.0,>=2.16.0
  • decorator *
  • deepspeed <=0.14.4,>=0.10.0
  • eetq *
  • einops *
  • fastapi *
  • fire *
  • galore-torch *
  • gradio >=4.0.0
  • hqq *
  • jieba *
  • liger-kernel *
  • matplotlib >=3.7.0
  • modelscope *
  • nltk *
  • numpy <2.0.0
  • optimum >=1.17.0
  • packaging *
  • pandas >=2.0.0
  • peft <=0.12.0,>=0.11.1
  • protobuf *
  • pydantic *
  • pytest *
  • pyyaml *
  • rouge-chinese *
  • ruff *
  • scipy *
  • sentencepiece *
  • sse-starlette *
  • tiktoken *
  • torch >=1.13.1
  • torch ==2.1.0
  • torch-npu ==2.1.0.post3
  • transformers <=4.45.2,>=4.41.2
  • transformers_stream_generator *
  • trl <=0.9.6,>=0.8.6
  • uvicorn *
  • vllm <=0.6.2,>=0.4.3