Skip to content

Commit

Permalink
Apply automatic changes
Browse files Browse the repository at this point in the history
  • Loading branch information
mdingemanse authored and github-actions[bot] committed Apr 19, 2024
1 parent 69db119 commit 0c5d213
Show file tree
Hide file tree
Showing 3 changed files with 10 additions and 73 deletions.
1 change: 1 addition & 0 deletions csv_file_path
Original file line number Diff line number Diff line change
@@ -1,6 +1,7 @@
project.link,project.notes,project.llmbase,project.rlbase,project.license,org.name,org.link,org.notes,opencode.class,opencode.link,opencode.notes,llmdata.class,llmdata.link,llmdata.notes,llmweights.class,llmweights.link,llmweights.notes,rldata.class,rldata.link,rldata.notes,rlweights.class,rlweights.link,rlweights.notes,license.class,license.link,license.notes,code.class,code.link,code.notes,architecture.class,architecture.link,architecture.notes,preprint.class,preprint.link,preprint.notes,paper.class,paper.link,paper.notes,modelcard.class,modelcard.link,modelcard.notes,datasheet.class,datasheet.link,datasheet.notes,package.class,package.link,package.notes,api.class,api.link,api.notes,source.file,openness
https://huggingface.co/bigscience/bloomz,,"BLOOMZ, mT0",xP3,Apache 2.0 and RAIL (responsible AI license),bigscience-workshop,https://github.com/bigscience-workshop,,open,https://github.com/bigscience-workshop/xmtf,Repository provides a guided overview to all components,open,https://github.com/bigscience-workshop/xmtf#data,Data made available & documented in detail in repo and preprint,open,https://github.com/bigscience-workshop/xmtf#models,Model made available on github,open,https://huggingface.co/datasets/bigscience/xP3all,From the documentation 'xP3 (Crosslingual Public Pool of Prompts) is a collection of prompts & datasets across 46 of languages & 16 NLP tasks',partial,https://huggingface.co/bigscience/bloomz-optimizer-states/tree/main,Fine-tuned checkpoint available for download,partial,https://bigscience.huggingface.co/blog/the-bigscience-rail-license,"Code licensed under Apache 2.0, model under bespoke 'Responsible AI License' which imposes some limitations",open,https://github.com/bigscience-workshop/xmtf,Code well documented and actively maintained,open,https://github.com/bigscience-workshop/xmtf#create-xp3x,"Architecture described in preprint, code available in github repo, recipe on HuggingFace",open,https://arxiv.org/abs/2211.05100,Preprint (updated June 2023) of 65 pages + 10 page appendix,open,https://aclanthology.org/2023.acl-long.891/,Peer-reviewed paper of 9 pages + 114 page appendix describes the multitask finetuning (instruction tuning) of BLOOM (see preprint) to form BLOOMZ,open,https://huggingface.co/bigscience/bloomz,Model card,open,https://huggingface.co/datasets/bigscience/xP3,Dataset documented in dataset card at HuggingFace,closed,,No packages published,open,https://huggingface.co/spaces/bigscience/petals-api,Petals API via HuggingFace not always available ('not enough hardware capacity'),/projects/bloomz.yaml,12.0
https://huggingface.co/LLM360/AmberChat,,Amber,ShareGPT + Evol-Instruct (synthetic),Apache 2.0,LLM360,https://www.llm360.ai/index.html,,open,https://github.com/LLM360/amber-train/tree/main,amber-train repository includes code for training and finetuning.,open,https://huggingface.co/datasets/LLM360/AmberDatasets,data well-documented and openly available,open,https://huggingface.co/LLM360/Amber,360 model checkpoints released,open,https://huggingface.co/datasets/WizardLM/WizardLM_evol_instruct_V2_196k,RL and fine-tuning data shared and documented,open,https://huggingface.co/LLM360/AmberChat,Finetuned model available for download.,open,https://huggingface.co/LLM360/AmberChat,Everything licensed under Apache 2.0,partial,https://github.com/LLM360,Code documented in helpful readme.md files but only partly inline.,partial,https://arxiv.org/abs/2312.06550,"Architecture described in preprint, but not all details documented.",open,https://arxiv.org/abs/2312.06550,"Preprint describes architecture, design choices, training and fine-tuning.",closed,,No peer-reviewed paper yet.,partial,https://huggingface.co/LLM360/AmberChat,Model card doesn't specify use or limitations,partial,https://huggingface.co/datasets/LLM360/AmberDatasets,"Concise description (better than most), but doesn't specify funders, purposes, representativeness, legal status as prescribed by datasheets industry standard",closed,,No released package found,open,https://huggingface.co/LLM360/AmberChat,Free Huggingface inference API.,/projects/amber.yaml,10.0
https://blog.allenai.org/olmo-open-language-model-87ccfc95f580,,OLMo 7B,OpenInstruct,Apache 2.0,AllenAI,https://allenai.org/allennlp,,open,https://github.com/allenai/OLMo,"Multiple repos with training, architecture and fine-tuning code available",open,https://huggingface.co/datasets/allenai/dolma,Dolma training data released and documented in exemplary way,open,https://huggingface.co/collections/allenai/olmo-suite-65aeaae8fe5b6b2122b46778,OLMo 7B and many training checkpoints available,open,https://huggingface.co/datasets/allenai/ultrafeedback_binarized_cleaned,Instruction tuning datasets documented and made available in exemplary ways,open,https://huggingface.co/allenai/OLMo-7B-Instruct/tree/main,Full model weights made available,,,,,,,,,,open,https://arxiv.org/abs/2402.00838,"Preprint describes model architecture, training and fine-tuning data, and training and SFT pipelines",closed,,No peer-reviewed paper found,open,https://huggingface.co/allenai/OLMo-7B-Instruct,Model card provides broad overview and links to full details,open,https://huggingface.co/datasets/allenai/dolma,"Data sheets and documentation available for the datasets used, linked here is Dolma",open,https://pypi.org/project/ai2-olmo/,No separate package made available,partial,https://huggingface.co/allenai/OLMo-7B-hf,Available through HuggingFace though model is,/projects/olmo-7b-instruct.yaml,9.5
https://open-assistant.io/,,Pythia 12B,OpenAssistant Conversations,Apache 2.0,LAION-AI,https://open-assistant.io/,,open,https://github.com/LAION-AI/Open-Assistant,Code includes guide for developers,open,https://github.com/LAION-AI/Open-Assistant/tree/main/data/datasets,Datasets documented in detail and recipes for cleaning up and downloading provided in code notebooks.,open,https://huggingface.co/OpenAssistant,Model weights in several variants downloadable through HuggingFace,open,https://huggingface.co/datasets/OpenAssistant/oasst1,"OpenAssistant Conversations is 'a human-generated, human-annotated assistant-style conversation corpus consisting of 161443 messages distributed across 66497 conversation trees, in 35 different languages, annotated with 461292 quality ratings' (preprint)",closed,,RLHF weights not separately released,open,https://projects.laion.ai/Open-Assistant/docs/faq#what-license-does-open-assistant-use,Apache 2.0,open,https://projects.laion.ai/Open-Assistant/docs/intro,Separate website provides entry point to comprehensive documentation,open,https://github.com/LAION-AI/Open-Assistant/tree/main/model,Instructions to tune the pipeline on training data,partial,https://arxiv.org/abs//2304.07327,"Preprint describes creation of OpenAssistant Conversations corpus for instruction tuning, but not the base LLM, hence partial.",closed,,No peer-reviewed paper or published data audit found,closed,,,closed,,,open,,,open,https://projects.laion.ai/Open-Assistant/api,,/projects/Open-Assistant.yaml,9.5
https://github.com/imoneoi/openchat,,Mistral 7B,ShareGPT with C-RLFT,Apache 2.0,Tshinghua University,https://github.com/imoneoi,OpenChat notes 'We are a student team from Tsinghua University',open,https://github.com/imoneoi/openchat/tree/master/ochat,Repository offers a large amount of fairly well-organized code for data curation and model,closed,,Pretraining data for Mistral is nowhere disclosed or documented,open,https://github.com/mistralai/mistral-src#download-the-model,Mistral 7B weights available via Mistral repository,closed,,Preprint says shareGPT dataset 'collected from sharegpt.com' but not disclosed or made available by this project,open,https://huggingface.co/openchat/openchat_3.5/tree/main,Instruction-tuned model weights made available via HuggingFace,open,https://github.com/imoneoi/openchat/blob/master/LICENSE,Code and model released under Apache 2.0,partial,https://github.com/imoneoi/openchat/tree/master/ochat,There is plenty of code in the github repository but only some of it is documented,open,https://arxiv.org/abs/2309.11235,Architecture quite well described in preprint,open,https://arxiv.org/abs/2309.11235,"Preprint describes the model architecture and instruction tuning approach, though is hampered by building on notoriously closed Llama2",open,https://openreview.net/forum?id=AOJyfhWYHf,Paper reviewed and accepted for ICLR 2024,partial,https://huggingface.co/openchat/openchat_v3.2,There is a model card that provides some details on architecture and evaluation,closed,,Datasheet not provided.,open,https://github.com/imoneoi/openchat/tree/master#installation,Python package 'ochat' provided through pip,partial,,"Model too large to load onto HuggingFace free inference API, so only available through Inference Endpoints or package",/projects/OpenChat.yaml,9.5
https://huggingface.co/togethercomputer/Pythia-Chat-Base-7B,,EleutherAI pythia,OIG,Apache 2.0 license,togethercomputer,https://github.com/togethercomputer,,open,,,open,https://github.com/togethercomputer/OpenDataHub,Training data curated and shared in separate repository,open,https://huggingface.co/togethercomputer/Pythia-Chat-Base-7B/tree/main,Model weights available via HuggingFace,open,https://huggingface.co/datasets/laion/OIG,From the documentation 'This is our attempt to create a large instruction dataset of medium quality along with a smaller high quality instruciton dataset (OIG-small-chip2).',closed,,RL weights not separately made available,open,https://huggingface.co/togethercomputer/Pythia-Chat-Base-7B#model-details,Apache 2.0,open,https://github.com/togethercomputer/OpenChatKit,Actively maintained repository,open,https://github.com/togethercomputer/OpenChatKit#reproducing-pythia-chat-base-7b,Architecture and recipe for reproducing model provided,partial,https://arxiv.org/abs/2304.01373,Preprint describes LM base (Pythia) but not instruction tuning details,closed,,No peer-reviewed paper or data audit found,partial,https://huggingface.co/togethercomputer/Pythia-Chat-Base-7B,Model card partially available but fairly minimally specified,partial,https://huggingface.co/datasets/laion/OIG,OIG instruction dataset documented,open,,,closed,,,/projects/pythia-chat-base-7B.yaml,9.5
Expand Down
Loading

0 comments on commit 0c5d213

Please sign in to comment.