Stars
A high-performance LLM inference API and Chat UI that integrates DeepSeek R1's CoT reasoning traces with Anthropic Claude models.
DeepEP: an efficient expert-parallel communication library
A simple screen parsing tool towards pure vision based GUI agent
An open source flight dynamics & control software library
Open source simulator for autonomous vehicles built on Unreal Engine / Unity, from Microsoft AI & Research
OLMoE: Open Mixture-of-Experts Language Models
💥 Fast State-of-the-Art Tokenizers optimized for Research and Production
mPLUG-Owl: The Powerful Multi-modal Large Language Model Family
text and image to video generation: CogVideoX (2024) and CogVideo (ICLR 2023)
MiniCPM-o 2.6: A GPT-4o Level MLLM for Vision, Speech and Multimodal Live Streaming on Your Phone
AISystem 主要是指AI系统,包括AI芯片、AI编译器、AI推理和训练框架等AI全栈底层技术
LLM大模型开源的非常多,对于想深入了解transformer模型结构,从零开始完整训练一个大模型,不用推理框架使用模型推理往往无从入手,另外开源模型往往对gpu有一定要求,不利于新手深入探索,我们尝试解决这些问题
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
A Clash GUI based on tauri. Supports Windows, macOS and Linux.
中文Mixtral混合专家大模型(Chinese Mixtral MoE LLMs)
Simple verification experiments codes for multi-agent RL using OpenAI MPE environment
Code for a multi-agent particle environment used in the paper "Multi-Agent Actor-Critic for Mixed Cooperative-Competitive Environments"
An API standard for multi-agent reinforcement learning environments, with popular reference environments and related utilities
A PyTorch implementation of the Transformer model in "Attention is All You Need".
A Pytorch Implementation of "Attention is All You Need" and "Weighted Transformer Network for Machine Translation"
本仓库提供了一个基于PyTorch实现的Transformer模型示例代码,专为初学者设计,用以深入浅出地讲解Transformer架构的工作原理和应用。通过阅读和运行此项目中的代码,学习者可以快速理解自注意力机制、编码器-解码器结构以及如何在实际任务中使用Transformer。同时,项目包含了详细的文档说明和注释,便于跟随每一步进行实践。
Cambrian-1 is a family of multimodal LLMs with a vision-centric design.