Skip to content

Latest commit

 

History

History
12 lines (10 loc) · 258 Bytes

README.md

File metadata and controls

12 lines (10 loc) · 258 Bytes

multi-tokernizer-llm

Install dependancies

pip install -r requirements.txt

Preprocess data

python utils/preprocess.py --load_tokenizer

Use --load_tokenizer to load the tokenizer from the files if you have already saved one.