This repo is used to run various AI benchmarks on Open Interpreter.
There is currently support for GAIA and SWE-bench
- Make sure the following software is installed on your computer.
- 
Run Docker 
- 
Copy-paste the following into your terminal 
git clone https://github.com/OpenInterpreter/benchmarks.git \
  && cd benchmarks \
  && python -m venv .venv \
  && source .venv/bin/activate \
  && python -m pip install -r requirements.txt \
  && docker build -t worker . \
  && python setup.py- Enter your Huggingface token
This section assumes:
- benchmarks(downloaded via git in the preview section) is set as the current working directory.
- You've activated the virtualenv with the installed prerequisite packages.
- If using an OpenAI model, your OPENAI_API_KEYenvironment variable is set with a valid OpenAI API key.
- If using a Groq model, your GROQ_API_KEYenvironment variable is set with a valid Groq API key.
Note: For running GAIA, you have to accept the conditions to access its files and content on Huggingface
This command will output a file called output.csv containing the results of the benchmark.
python run_benchmarks.py \
  --command gpt35turbo \
  --ntasks 16 \
  --nworkers 8- --command gpt35turbo: Replace gpt35turbo with any existing key in the commands- Dictin commands.py. Defaults to gpt35turbo.
- --ntasks 16: Grabs the first 16 GAIA tasks to run. Defaults to all 165 GAIA validation tasks.
- --nworkers 8: Number of docker containers to run at once. Defaults to whatever max_workers defaults to when constructing a ThreadPoolExecutor.
- ModuleNotFoundError: No module named '_lzma'when running example.- If you're using pyenvto manage python versions, this stackoverflow post might help.
 
- If you're using 
- ModuleNotFoundError: No module named 'pkg_resources'when running example.- Refer to this stackoverflow post for now.
- OpenInterpreter should probably include setuptoolsin its list of dependencies, or should switch to another module that's in python's standard library.