title | emoji | colorFrom | colorTo | sdk | sdk_version | app_file | pinned | license | short_description |
---|---|---|---|---|---|---|---|---|---|
Open O1 |
🦀 |
gray |
indigo |
streamlit |
1.38.0 |
app.py |
false |
apache-2.0 |
mimics openai-o1 |
It thinks like o1
Todo
[ ] Add fallback llms [ ] Better error handling [ ] Add Tools (web, math, code) [ ] Make cli [ ] better prompts for mathematical reasoning/reviewing
- It taks the prompt, decides whether to use chain of thought or direct answer, if cot then generates answer and does self review, if direct answer then directly generates answer.
- Mathematical reasoning, symbolic reasoning and semi-symbolic reasoning kind of tasks generally improves with chain of thought, but direct answer is good for factual recall, simple inferences, commonsense reasoning, language understanding tasks.
Open-o1
#clone the repo
git clone https://github.com/tikendraw/open-o1.git
# goto folder
cd open-o1
# make a virtual env
python -m venv .venv
# activate the virtual env
source .venv/bin/activate
# install requirements
pip install -e .
#finally run the server
streamlit run app.py
HAVE FUN.
Although this project tries to mimic openai's o1, many times it falls short in generating and better reflections on previous answers, and that i thinks comes from the lack of such kind of training data that used to train the models (other models before o1), these models are probably not trained for fixing mistakes with better reasoning.
for example: here a cerebras/llama3.1-70b
models jumps back and forth between count of "r" s in Straberry as 2 and 3. Even when having second thoughts, It doesn't sticks to its reasoning and because of the model bias it generates wrong answers. may be prompting can solve, but training with such data would be better.
- To Cot or not to Cot? CHAIN-OF-THOUGHT HELPS MAINLY ON MATH AND SYMBOLIC REASONING
@misc{sprague2024cotcotchainofthoughthelps,
title={To CoT or not to CoT? Chain-of-thought helps mainly on math and symbolic reasoning},
author={Zayne Sprague and Fangcong Yin and Juan Diego Rodriguez and Dongwei Jiang and Manya Wadhwa and Prasann Singhal and Xinyu Zhao and Xi Ye and Kyle Mahowald and Greg Durrett},
year={2024},
eprint={2409.12183},
archivePrefix={arXiv},
primaryClass={cs.CL},
url={https://arxiv.org/abs/2409.12183},
}
- The Impact of Reasoning Step Length on Large Language Models
@misc{jin2024impactreasoningsteplength,
title={The Impact of Reasoning Step Length on Large Language Models},
author={Mingyu Jin and Qinkai Yu and Dong Shu and Haiyan Zhao and Wenyue Hua and Yanda Meng and Yongfeng Zhang and Mengnan Du},
year={2024},
eprint={2401.04925},
archivePrefix={arXiv},
primaryClass={cs.CL},
url={https://arxiv.org/abs/2401.04925},
}
- Towards Understanding Chain-of-Thought Prompting: An Empirical Study of What Matters 2212.10001
@misc{wang2023understandingchainofthoughtpromptingempirical,
title={Towards Understanding Chain-of-Thought Prompting: An Empirical Study of What Matters},
author={Boshi Wang and Sewon Min and Xiang Deng and Jiaming Shen and You Wu and Luke Zettlemoyer and Huan Sun},
year={2023},
eprint={2212.10001},
archivePrefix={arXiv},
primaryClass={cs.CL},
url={https://arxiv.org/abs/2212.10001},
}