-
Notifications
You must be signed in to change notification settings - Fork 330
Permalink
Choose a base ref
{{ refName }}
default
Choose a head ref
{{ refName }}
default
Comparing changes
Choose two branches to see what’s changed or to start a new pull request.
If you need to, you can also or
learn more about diff comparisons.
Open a pull request
Create a new pull request by comparing changes across two branches. If you need to, you can also .
Learn more about diff comparisons here.
base repository: InternLM/xtuner
Failed to load repositories. Confirm that selected base ref is valid, then try again.
Loading
base: v0.1.23
Could not load branches
Nothing to show
Loading
Could not load tags
Nothing to show
{{ refName }}
default
Loading
...
head repository: InternLM/xtuner
Failed to load repositories. Confirm that selected head ref is valid, then try again.
Loading
compare: main
Could not load branches
Nothing to show
Loading
Could not load tags
Nothing to show
{{ refName }}
default
Loading
- 14 commits
- 833 files changed
- 11 contributors
Commits on Jul 29, 2024
-
Co-authored-by: liudan <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for d2a173a - Browse repository at this point
Copy the full SHA d2a173aView commit details
Commits on Jul 31, 2024
-
[Bug] fix dsv2 attn dispatch (softmax_scale) (#873)
fix dsv2 attn dispatch (softmax_scale)
Configuration menu - View commit details
-
Copy full SHA for 01640b0 - Browse repository at this point
Copy the full SHA 01640b0View commit details
Commits on Aug 9, 2024
-
[Bug] fix openai_map_fn bugs (#885)
fix openai_map_fn bugs
Configuration menu - View commit details
-
Copy full SHA for 7dd779b - Browse repository at this point
Copy the full SHA 7dd779bView commit details -
Configuration menu - View commit details
-
Copy full SHA for d81b366 - Browse repository at this point
Copy the full SHA d81b366View commit details -
* add internlm2.5 configs * limit transformers <= 4.42.4
Configuration menu - View commit details
-
Copy full SHA for 081c8ca - Browse repository at this point
Copy the full SHA 081c8caView commit details
Commits on Sep 29, 2024
-
[Bugs] fix qlora convert bugs (#930)
fix qlora convert bugs
Configuration menu - View commit details
-
Copy full SHA for 4a1b201 - Browse repository at this point
Copy the full SHA 4a1b201View commit details
Commits on Oct 22, 2024
-
* readme中增加了MiniCPM的支持 * 适配了minicpm3,并且测试可以运行 * 规范了格式,删除了长的行 * delete the error file * fix lint * fix the file name error --------- Co-authored-by: liudan <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 697bc77 - Browse repository at this point
Copy the full SHA 697bc77View commit details
Commits on Nov 8, 2024
-
Add functionality to download models from sources other than HuggingF…
…ace (#946) support openmind model and dataset
Configuration menu - View commit details
-
Copy full SHA for 90192ff - Browse repository at this point
Copy the full SHA 90192ffView commit details
Commits on Jan 14, 2025
-
Add Ascend NPU as a backend (#983)
Co-authored-by: 郑特驹 <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 2c06115 - Browse repository at this point
Copy the full SHA 2c06115View commit details -
[Feature]Support transformers==4.48 (#985)
* update requirements * support internlm3, llama, mistral, mixtral, qwen2 and qwen2moe in transformers==4.48
Configuration menu - View commit details
-
Copy full SHA for 4ee8215 - Browse repository at this point
Copy the full SHA 4ee8215View commit details
Commits on Jan 21, 2025
-
[Feature] Auto patch for different devices (#986)
* [Feature] XTuner Lite (#974) * minimum dependency sft * fix dispatch * add timer * add tgs * internlm2 tp * rms support tp * gradient checkpointing * lazy load pretrain * temp * fix bugs * add data pipeline example * fix lints * remove useless code * fix hard pack bug * add comments * clean code * add shard strategy * support cpu offload * support cpu offload * trust remote code * fix soft packer bug * fix soft packer bug * fix soft packer bug * refactor data pipeline * fixup * fix pad tokens bug * check input_ids and labels * check input_ids and labels in collator * fix load local datasets bug * fix load cache datasts * restore dset order * save cached infos * accelerate start up * avoid all gather cached datasets * fixup * fix cache bug * Support group length (#4) * replace rmsnorm kernel * suport ftdp ds * suport load_bin * suport group by maxlen * add fsdp_ftdp_sft and fix fsdp_sft * suport ftdp ds * add lr min * fix bugs * fix bugs * delete * support llava * support packer cache * refactor dist load * Add sp tp (#5) * support sp and tp * add fsdp_tp_sft and modify fsdp_sft * move chat_template * fix load_ds * delete useless codes * delete useless codes * fix jsonl load * refactor * fix bug * fix lr scheduler * refactor setup parallel * update data load * fix bugs * move fsdp * adapt new parallel load * fix setup_parallel (#7) * fix some bugs * add remote codes * add convert script * support load image from ceph * support load image from ceph * fix cache dataset bugs * support mulit images * support llava interleave * fix load timeout * refactor datasets: optimize the cache mechanism and clean up code * distinguish dataset components based on algorithms * support fsdp2+3d parallel * fix lints * support contiguous batching * refactor parallel * zero wasting ppo * support asend npu * fix openai convert * fix npu bugs * fix npu bug * dispatch npu flash attn * adapt asend npu * fix ppo losses * steady increase in reward * faster ppo * fix top-p generate * support internlm3 * baseline 2.5 * fix internlm3 * (ing)support hard pack * support qwen2 * fix dataset bugs * baseline * del ppo.py * fixup * support hybrid sp * fix hybrid sp * qwen2 + hybird sp * fix requirements * avoid re-initialize dist * support group pack * pretrain (#13) * first commit: support internlm3 moe streaming dataset * move codes * Moe pretrain (#14) * first commit: support internlm3 moe streaming dataset * move codes * rmsnorm kernel support low version flash_attn * add barrier * support prompt length control (#15) * support VLM Base (#16) * add internvl * fix bug * remove dup code * support liger of internvl * fix bug * add get_repo_git_info * fix * add minicpmv * add minicpmv dispatch * accelerate tokenize * Updata InternVL (#17) * fix dpo error * fix sp error * update dataset * fix * fix rand sampler (#18) * llama support transformers >= 4.45 (#19) * convert fsdp1 to fsdp2 in sft.py * [Feature] Support Liger Kernel (#20) * filter data by max length (#21) * fix causal forward, prefetch, and remote code (#22) * [Enhancement] Accelerating Data Pipeline (#23) * sample ratio greater than 1.0 and trunc max len * accelerating the counting of tokens * log reduced loss * fix mirco bs greater than 1 * [Enhancement] Ensure data integrity when the sampling ratio is more than 1 (#24) * repeat dataset * fixup * fix typos * fix typos * [Fix] Pass in temperature during generation (#25) * Support Janus and fix some error (#27) * add prefetch * update prefetch * add janus * add janus * fix * fix * fix llama position id error * fix ProcessPoolExecutor * update * fix llama * delete cache * remove useless code --------- Co-authored-by: whcao <[email protected]> Co-authored-by: Happy <[email protected]> Co-authored-by: Haian Huang(深度眸) <[email protected]> * support mlu (#984) * cleanup * add internlm3 remote code * cleanup * auto patch * remove useless code --------- Co-authored-by: whcao <[email protected]> Co-authored-by: Happy <[email protected]> Co-authored-by: Haian Huang(深度眸) <[email protected]> Co-authored-by: Lantian Zhang <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 4a521fb - Browse repository at this point
Copy the full SHA 4a521fbView commit details
Commits on Jan 22, 2025
-
* fix mlu device mesh * update requirements
Configuration menu - View commit details
-
Copy full SHA for 4cade9f - Browse repository at this point
Copy the full SHA 4cade9fView commit details
Commits on Feb 21, 2025
-
bump version to v0.2.0rc0 (#990)
* refactor dispatch * support qwen2 * sp gather logprobs * optimize gpu memory * fix tp sample * fix auto cls * fix saved dtype * remove ascend llama * accelerate sp all to all * optimize long prefilling * fix tp>2 generate * [Enhance] Using `fork` mode for tokenization (#28) * integrate training and inference * [Enhance] Enhance the calculation of tokenize fn hash (#29) * [Fix] Cannot properly load buffer from the checkpoint (#30) * fix load buffer bug * add moe permute * fix lints * rm internvl2 dataset * rm tools * update precommit * bump version --------- Co-authored-by: Mashiro <[email protected]> Co-authored-by: whcao <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for bbcd94b - Browse repository at this point
Copy the full SHA bbcd94bView commit details -
Configuration menu - View commit details
-
Copy full SHA for 53f2429 - Browse repository at this point
Copy the full SHA 53f2429View commit details
There are no files selected for viewing
This file was deleted.