Stars
An efficient, flexible and full-featured toolkit for fine-tuning LLM (InternLM2, Llama3, Phi3, Qwen, Mistral, ...)
An open-source implementaion for fine-tuning Llama3.2-Vision series by Meta.
A minimal codebase for finetuning large multimodal models, supporting llava-1.5/1.6, llava-interleave, llava-next-video, llava-onevision, llama-3.2-vision, qwen-vl, qwen2-vl, phi3-v etc.
Repo for Rho-1: Token-level Data Selection & Selective Pretraining of LLMs.
Unified Efficient Fine-Tuning of 100+ LLMs & VLMs (ACL 2024)
Welcome to the Llama Cookbook! This is your go to guide for Building with Llama: Getting started with Inference, Fine-Tuning, RAG. We also show you how to solve end to end problems using Llama mode…
Use PEFT or Full-parameter to finetune 450+ LLMs (Qwen2.5, InternLM3, GLM4, Llama3.3, Mistral, Yi1.5, Baichuan2, DeepSeek-R1, ...) and 150+ MLLMs (Qwen2.5-VL, Qwen2-Audio, Llama3.2-Vision, Llava, I…
Finetune Llama 3.3, DeepSeek-R1 & Reasoning LLMs 2x faster with 70% less memory! 🦥
LLM Finetuning with peft
此项目是机器学习(Machine Learning)、深度学习(Deep Learning)、NLP面试中常考到的知识点和代码实现,也是作为一个算法工程师必会的理论基础知识。
we want to create a repo to illustrate usage of transformers in chinese
Consistent Prompting for Rehearsal-Free Continual Learning [CVPR2024]
albertotamajo / submodlib
Forked from decile-team/submodlibSummarize Massive Datasets using Submodular Optimization
Source code and dataset for ACL 2019 paper "ERNIE: Enhanced Language Representation with Informative Entities"
🔥🔥🔥Latest Papers, Codes and Datasets on Vid-LLMs.
Explore VLM-Eval, a framework for evaluating Video Large Language Models, enhancing your video analysis with cutting-edge AI technology.
算法岗笔试面试大全,励志做算法届的《五年高考,三年模拟》!
The collections of MOE (Mixture Of Expert) papers, code and tools, etc.
Code for paper "Boosting Continual Learning of Vision-Language Models via Mixture-of-Experts Adapters" CVPR2024
Preventing Zero-Shot Transfer Degradation in Continual Learning of Vision-Language Models
An Efficient Dataset Condensation Plugin and Its Application to Continual Learning. NeurIPS, 2023.
🎉 PILOT: A Pre-trained Model-Based Continual Learning Toolbox
Our code for ICCV'23 paper "CAME: Contrastive Automated Model Evaluation".