Skip to content

Commit

Permalink
bump 0.0.2
Browse files Browse the repository at this point in the history
  • Loading branch information
francoishernandez committed Sep 19, 2024
1 parent 520566a commit b5df05c
Show file tree
Hide file tree
Showing 3 changed files with 74 additions and 4 deletions.
69 changes: 69 additions & 0 deletions CHANGELOG.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,69 @@
# Changelog

This is just a centralised version of the Github automatically generated Release changelogs.

## 0.0.2

* Refactor position encoding configuration by @vince62s in https://github.com/eole-nlp/eole/pull/60
* fix update vocab by @vince62s in https://github.com/eole-nlp/eole/pull/63
* bfloat16 support, and an attempt at homogenizing model_dtype & precision by @francoishernandez in https://github.com/eole-nlp/eole/pull/54
* Fix prefix and suffix transforms - avoid adding empty suffix or prefix by @sersh88 in https://github.com/eole-nlp/eole/pull/57
* fix the incorrect dockerimages in the ReadMe by @aaaallleen in https://github.com/eole-nlp/eole/pull/68
* Remove unnecessary optim in convert_HF by @francoishernandez in https://github.com/eole-nlp/eole/pull/71
* Add onmt_config converter to facilitate switch by @francoishernandez in https://github.com/eole-nlp/eole/pull/69
* Update some FAQ sections by @francoishernandez in https://github.com/eole-nlp/eole/pull/74
* Added TER and BLEU for early stopping by @aaaallleen in https://github.com/eole-nlp/eole/pull/73
* [fix] fix normalize and clean transforms config management by @francoishernandez in https://github.com/eole-nlp/eole/pull/87
* [docs] Fix quickstart config and command by @francoishernandez in https://github.com/eole-nlp/eole/pull/90
* add head_dim setting when diff from hidden // heads by @vince62s in https://github.com/eole-nlp/eole/pull/78
* Some MHA and RoPE refactoring, llama-3.1 rope_scaling by @francoishernandez in https://github.com/eole-nlp/eole/pull/91
* Fixed variable referenced before assignment when position_embeddings is None error by @dameikle in https://github.com/eole-nlp/eole/pull/95
* Send src_pad_mask and tgt_pad_mask to decoder in _align_forward by @dameikle in https://github.com/eole-nlp/eole/pull/96
* Fixdistrib by @vince62s in https://github.com/eole-nlp/eole/pull/100
* fix added tokens by @vince62s in https://github.com/eole-nlp/eole/pull/101
* Support mapped tokens eg: <im_start> ==> ⦅im_start⦆in inference.yaml … by @vince62s in https://github.com/eole-nlp/eole/pull/102
* add wmt22 recipes with TowerInstruct and Llama3.1 LLMs by @vince62s in https://github.com/eole-nlp/eole/pull/103
* Remove duplicate sentencepiece requirement by @francoishernandez in https://github.com/eole-nlp/eole/pull/104
* [patch] Adapt some warning behaviours for reduced verbosity by @francoishernandez in https://github.com/eole-nlp/eole/pull/105
* [patch] Update precision to compute_dtype in forgotten places by @francoishernandez in https://github.com/eole-nlp/eole/pull/106
* Inference server, lots of related changes by @francoishernandez in https://github.com/eole-nlp/eole/pull/42

**Full Changelog**: https://github.com/eole-nlp/eole/compare/0.0.1...0.0.2


## 0.0.1
* mlp refact by @vince62s in https://github.com/eole-nlp/eole/pull/1
* fix llama3 and parallel_residual by @vince62s in https://github.com/eole-nlp/eole/pull/4
* fixed mismatch between mask and batch dimensions by @l-k-11235 in https://github.com/eole-nlp/eole/pull/6
* simplify LayerNorm access as a constant by @vince62s in https://github.com/eole-nlp/eole/pull/7
* Fix the checkpoint directory cleaning by @l-k-11235 in https://github.com/eole-nlp/eole/pull/10
* Modify default model config behaviour by @francoishernandez in https://github.com/eole-nlp/eole/pull/8
* rename num_kv remove multiquery by @vince62s in https://github.com/eole-nlp/eole/pull/12
* fix mmlu config by @vince62s in https://github.com/eole-nlp/eole/pull/13
* Fix the tokenizer saving in the HF converter by @l-k-11235 in https://github.com/eole-nlp/eole/pull/14
* remove unsused average attn by @vince62s in https://github.com/eole-nlp/eole/pull/15
* MHA refac: rope without complex operations + query only as input of the forward by @vince62s in https://github.com/eole-nlp/eole/pull/20
* Revert "MHA refac: rope without complex operations + query only as input of the forward" by @vince62s in https://github.com/eole-nlp/eole/pull/22
* missing removal of average attn by @vince62s in https://github.com/eole-nlp/eole/pull/23
* `config.models.BaseModelConfig._override_values` updates everything once by @francoishernandez in https://github.com/eole-nlp/eole/pull/24
* [fix] Patch lora bin to dump json config by @francoishernandez in https://github.com/eole-nlp/eole/pull/28
* review flash/sdpa arg by @vince62s in https://github.com/eole-nlp/eole/pull/25
* fix missing layers names by @vince62s in https://github.com/eole-nlp/eole/pull/30
* Split MHA by @vince62s in https://github.com/eole-nlp/eole/pull/29
* Resize the key_pad_mask by @l-k-11235 in https://github.com/eole-nlp/eole/pull/36
* [patch] upgrade docusaurus deps, fix build script by @francoishernandez in https://github.com/eole-nlp/eole/pull/37
* Add gpt2 converter, hellaswag eval tool, misc fixes by @francoishernandez in https://github.com/eole-nlp/eole/pull/38
* Forgot hellaswag.py tool in #38 by @francoishernandez in https://github.com/eole-nlp/eole/pull/39
* estim lambda scheduler by @vince62s in https://github.com/eole-nlp/eole/pull/40
* Add support for XLM-Roberta-XL (and XXL) conversion by @vince62s in https://github.com/eole-nlp/eole/pull/41
* Some fixes, get rid of data_task, homogenize model_task to model_type by @francoishernandez in https://github.com/eole-nlp/eole/pull/43
* Some improvements to config.json readability by @francoishernandez in https://github.com/eole-nlp/eole/pull/44
* [docs] Github Actions workflow to facilitate docs deployment by @francoishernandez in https://github.com/eole-nlp/eole/pull/47
* [fix] Allow to build_vocab with full train config, patch vocab validation by @francoishernandez in https://github.com/eole-nlp/eole/pull/49
* Enable PyPI release workflow by @francoishernandez in https://github.com/eole-nlp/eole/pull/50
* [fix] Fix paths in wiki_103 recipe, add pyarrow opt requirement by @francoishernandez in https://github.com/eole-nlp/eole/pull/51
* Estim first token instead of average by @vince62s in https://github.com/eole-nlp/eole/pull/46
* Add Recipe to train a cometkiwi-like encoder model (which can be used to score sentence pairs) by @vince62s in https://github.com/eole-nlp/eole/pull/53
* Simplify __init__ files, remove some unused code by @francoishernandez in https://github.com/eole-nlp/eole/pull/52

**Full Changelog**: https://github.com/eole-nlp/eole/commits/0.0.1rc1
4 changes: 2 additions & 2 deletions README.md
Original file line number Diff line number Diff line change
Expand Up @@ -57,12 +57,12 @@ There are two images with CUDA 11.8 and 12.1 prebuilt, change the -cudaXX.X to y

To pull the Docker image:
```bash
docker pull ghcr.io/eole-nlp/eole:0.0.1-torch2.3.0-ubuntu22.04-cuda12.1
docker pull ghcr.io/eole-nlp/eole:0.0.2-torch2.3.0-ubuntu22.04-cuda12.1
```

Example one-liner to run a container and open a bash shell within it:
```bash
docker run --rm -it --runtime=nvidia ghcr.io/eole-nlp/eole:0.0.1-torch2.3.0-ubuntu22.04-cuda12.1
docker run --rm -it --runtime=nvidia ghcr.io/eole-nlp/eole:0.0.2-torch2.3.0-ubuntu22.04-cuda12.1
```

> **Note**: Ensure you have the [Nvidia Container Toolkit](https://docs.nvidia.com/datacenter/cloud-native/container-toolkit/latest/install-guide.html) (formerly nvidia-docker) installed to take advantage of CUDA/GPU features.
Expand Down
5 changes: 3 additions & 2 deletions setup.py
Original file line number Diff line number Diff line change
Expand Up @@ -11,7 +11,7 @@
description="Open language modeling toolkit based on PyTorch",
long_description=long_description,
long_description_content_type="text/markdown",
version="0.0.1",
version="0.0.2",
packages=find_packages(),
project_urls={
"Source": "https://github.com/eole-nlp/eole/",
Expand All @@ -20,8 +20,8 @@
install_requires=[
"configargparse",
"ctranslate2>=4,<5",
"fastapi",
"fasttext-wheel",
"flask",
"huggingface_hub",
"numpy<2.0",
"pandas",
Expand All @@ -38,6 +38,7 @@
"subword-nmt>=0.3.7",
"tensorboard>=2.3",
"torch>=2.3,<2.4",
"uvicorn",
"waitress",
],
entry_points={
Expand Down

0 comments on commit b5df05c

Please sign in to comment.