Logo
Explore Help
Register Sign In
ros/LLaMA-Factory
1
0
Fork 0
You've already forked LLaMA-Factory
Code Issues Pull Requests Actions Packages Projects Releases Wiki Activity
Files
a13b1bb49aa6ce8b178f343c1e06fdbbd8cd7a83
LLaMA-Factory/src/llamafactory
History
Yu Shi Jie a13b1bb49a [model] fix use_cache patching for gemma3 multimodal (#7500)
2025-04-01 16:06:48 +08:00
..
api
[misc] upgrade format to py39 (#7256)
2025-03-12 00:08:41 +08:00
chat
[inference] support sglang backend (#7278)
2025-03-15 04:37:58 +08:00
data
[data] specify position_ids in PackedSupervisedDatasetProcessor for neat_packing (#7318)
2025-04-01 16:03:13 +08:00
eval
[misc] update format (#7277)
2025-03-13 02:53:08 +08:00
extras
[model] add Qwen2.5-Omni model (#7537)
2025-03-31 20:39:35 +08:00
hparams
[data] shard the dataset to allow multiprocessing when streaming is enabled (#7530)
2025-04-01 15:36:23 +08:00
model
[model] fix use_cache patching for gemma3 multimodal (#7500)
2025-04-01 16:06:48 +08:00
train
[3rdparty] support swanlab lark notification (#7481)
2025-03-27 01:52:01 +08:00
webui
[webui] fix launch with proxy (#7332)
2025-04-01 15:52:56 +08:00
__init__.py
[model] add qwen2vl 32b & upgrade peft (#7469)
2025-03-25 12:15:58 +08:00
cli.py
[deps] upgrade vllm to 0.8 (#7436)
2025-03-23 14:32:22 +08:00
launcher.py
[misc] update license year & fix llama pro (#6814)
2025-02-05 01:53:33 +08:00
Powered by Gitea Version: 1.24.5 Page: 148ms Template: 16ms
English
Bahasa Indonesia Deutsch English Español Français Gaeilge Italiano Latviešu Magyar nyelv Nederlands Polski Português de Portugal Português do Brasil Suomi Svenska Türkçe Čeština Ελληνικά Български Русский Українська فارسی മലയാളം 日本語 简体中文 繁體中文(台灣) 繁體中文(香港) 한국어
Licenses API