Skip to content

floschne/gimmick

Folders and files

NameName
Last commit message
Last commit date

Latest commit

be2d16b Β· Feb 21, 2025

History

9 Commits
Feb 20, 2025
Feb 19, 2025
Feb 19, 2025
Feb 21, 2025
Feb 19, 2025

Repository files navigation

GIMMICK - Globally Inclusive Multimodal Multitask Cultural Knowledge Benchmarking

GIMMICK Overview

🚧 Work in Progress 🚧

News

  • 2024-02: πŸ€— We released the datasets introduced by GIMMICK on HuggingFace
  • 2024-02: πŸ“ We released the first preprint of the GIMMICK Benchmark paper.

Setup the environment

Tested only on Debian-based machines

Note that most models are loaded in FP16 or BF16 and Flash Attention 2. Make sure your GPU supports these features_

  1. Install mamba
  2. mamba env create -f environment.yml
  3. mamba activate gimmick && pip install flash-attn --no-build-isolation
  4. src/scripts/check_env.sh

If the last step was successful, it should print something like this:

##################### PYTHON ENV 'gimmick' INFO START #####################
Python version: 3.12.8 | packaged by conda-forge | (main, Dec  5 2024, 14:24:40) [GCC 13.3.0]
PyTorch version: 2.5.1
CUDA available: True
CUDA version: 11.8
CUDA devices: 1
Flash Attention 2 Support: True
Transformers version: 4.48.1
##################### PYTHON ENV 'gimmick' INFO END #####################

Evaluate a Model

Available Models

Run PYTHONPATH=src python src/gimmick/main.py models to print the list of supported models.

Open-Weight VLMs

Open-Weight LLMs

Proprietary VLMs