support falcon model #72
Former-commit-id: 72cc3ff0e6de641073de1159196319705f8efe85
This commit is contained in:
@@ -9,6 +9,8 @@
|
||||
|
||||
## Changelog
|
||||
|
||||
[23/07/05] Now we support training the Falcon-7B/40B models in this repo. Try `--model_name_or_path tiiuae/falcon-7b` and `--lora_target query_key_value` arguments to use the Falcon model.
|
||||
|
||||
[23/06/29] We provide a reproducible example of training a chat model using instruction-following datasets, see this [HuggingFace Repo](https://huggingface.co/hiyouga/baichuan-7b-sft) for details.
|
||||
|
||||
[23/06/22] Now we align the [demo API](src/api_demo.py) with the [OpenAI's](https://platform.openai.com/docs/api-reference/chat) format where you can insert the fine-tuned model in arbitrary ChatGPT-based applications.
|
||||
@@ -23,6 +25,7 @@
|
||||
|
||||
- [LLaMA](https://github.com/facebookresearch/llama) (7B/13B/33B/65B)
|
||||
- [BLOOM](https://huggingface.co/bigscience/bloom) & [BLOOMZ](https://huggingface.co/bigscience/bloomz) (560M/1.1B/1.7B/3B/7.1B/176B)
|
||||
- [Falcon](https://huggingface.co/tiiuae/falcon-7b) (7B/40B)
|
||||
- [baichuan](https://huggingface.co/baichuan-inc/baichuan-7B) (7B)
|
||||
|
||||
## Supported Training Approaches
|
||||
@@ -283,6 +286,8 @@ Please follow the [Model Card](https://github.com/facebookresearch/llama/blob/ma
|
||||
|
||||
Please follow the [RAIL License](https://huggingface.co/spaces/bigscience/license) to use the BLOOM & BLOOMZ models.
|
||||
|
||||
Please follow the [Apache-2.0 License](LICENSE) to use the Falcon models.
|
||||
|
||||
Please follow the [baichuan-7B License](https://huggingface.co/baichuan-inc/baichuan-7B/resolve/main/baichuan-7B%20%E6%A8%A1%E5%9E%8B%E8%AE%B8%E5%8F%AF%E5%8D%8F%E8%AE%AE.pdf) to use the baichuan-7B model.
|
||||
|
||||
## Citation
|
||||
|
||||
Reference in New Issue
Block a user