ms-swift: Use PEFT or Full-parameter to finetune 300 LLMs or 50 MLLMs. (Qwen2, GLM4v, Internlm2.5, Yi, Llama3.1, Llava-Video, Internvl2, MiniCPM-V, Deepseek, Baichuan2, Gemma2, Phi3-Vision, ...)
-
Updated
Jul 26, 2024 - Python
ms-swift: Use PEFT or Full-parameter to finetune 300 LLMs or 50 MLLMs. (Qwen2, GLM4v, Internlm2.5, Yi, Llama3.1, Llava-Video, Internvl2, MiniCPM-V, Deepseek, Baichuan2, Gemma2, Phi3-Vision, ...)
Megatron was a telegram file management bot that helped a lot of users, specially movie channel managers to upload their files to telegram by just providing a link to it. The project initially started as roanuedhuru_bot which lately retired and came back as Megatron which was a side project of the famous Maldivian Telegram community - @baivaru u…
Large scale 4D parallelism pre-training for 🤗 transformers in Mixture of Experts *(still work in progress)*
A LLaMA1/LLaMA12 Megatron implement.
Megatron was a telegram file management bot that helped a lot of users, specially movie channel managers to upload their files to telegram by just providing a link to it. The project initially started as roanuedhuru_bot which lately retired and came back as Megatron which was a side project of the famous Maldivian Telegram community -
Add a description, image, and links to the megatron topic page so that developers can more easily learn about it.
To associate your repository with the megatron topic, visit your repo's landing page and select "manage topics."