Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Rename GenAIEval and Docker folder and set version, binary name #22

Merged
merged 8 commits into from
May 31, 2024
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
10 changes: 2 additions & 8 deletions .github/workflows/model_test_cpu.yml
Original file line number Diff line number Diff line change
Expand Up @@ -53,10 +53,6 @@ jobs:
- name: Clean Up Working Directory
run: sudo rm -rf ${{github.workspace}}/*

- name: Load environment variables
run:
cat ~/actions-runner4/.env >> $GITHUB_ENV

- name: Checkout out Repo
uses: actions/checkout@v4
with:
Expand All @@ -65,7 +61,7 @@ jobs:
# We need this because GitHub needs to clone the branch to pipeline
- name: Docker Build
run: |
docker build -f ${{ github.workspace }}/.github/workflows/docker/common.dockerfile --build-arg http_proxy="${{ env.HTTP_PROXY_IMAGE_BUILD }}" --build-arg https_proxy="${{ env.HTTPS_PROXY_IMAGE_BUILD }}" -t ${{ env.DOCKER_NAME }}:${{ env.DOCKER_TAG }} .
docker build -f ${{ github.workspace }}/.github/workflows/docker/common.dockerfile -t ${{ env.DOCKER_NAME }}:${{ env.DOCKER_TAG }} .

- name: Docker Run
run: |
Expand All @@ -74,9 +70,7 @@ jobs:
docker rm -vf ${{ env.CONTAINER_NAME }} || true
fi
docker run -dit --disable-content-trust --privileged --name=${{ env.CONTAINER_NAME }} -v /dev/shm:/dev/shm \
-v ${{ github.workspace }}:/GenAIEval \
-e http_proxy="${{ env.HTTP_PROXY_IMAGE_BUILD }}" -e https_proxy="${{ env.HTTPS_PROXY_IMAGE_BUILD }}" \
${{ env.DOCKER_NAME }}:${{ env.DOCKER_TAG }}
-v ${{ github.workspace }}:/GenAIEval ${{ env.DOCKER_NAME }}:${{ env.DOCKER_TAG }}

- name: Binary build
run: |
Expand Down
2 changes: 1 addition & 1 deletion .github/workflows/model_test_hpu.yml
Original file line number Diff line number Diff line change
Expand Up @@ -61,7 +61,7 @@ jobs:
# We need this because GitHub needs to clone the branch to pipeline
- name: Docker Build
run: |
docker build --target hpu --build-arg REPO_PATH="." -f ${{ github.workspace }}/Docker/hpu.dockerfile -t ${{ env.DOCKER_NAME }}:${{ env.DOCKER_TAG }} .
docker build --target hpu --build-arg REPO_PATH="." -f ${{ github.workspace }}/docker/hpu.dockerfile -t ${{ env.DOCKER_NAME }}:${{ env.DOCKER_TAG }} .
- name: Docker Run
run: |
Expand Down
4 changes: 2 additions & 2 deletions .github/workflows/scripts/models/generate_report.sh
Original file line number Diff line number Diff line change
Expand Up @@ -48,7 +48,7 @@ function generate_html_overview {

<body>
<div id="main">
<h1 align="center">ITREX Tests
<h1 align="center">GenAIEval Tests
[ <a href="${RUN_DISPLAY_URL}">Job-${BUILD_NUMBER}</a> ]</h1>
<h1 align="center">Test Status: ${JOB_STATUS}</h1>
<h2>Summary</h2>
Expand All @@ -58,7 +58,7 @@ function generate_html_overview {
${Test_Info_Title}
</tr>
<tr>
<td><a href="https://github.com/intel/intel-extension-for-transformers">ITREX</a></td>
<td><a href="https://github.com/opea-project/GenAIEval">GenAIEval</a></td>
${Test_Info}
</tr>
</table>
Expand Down
4 changes: 2 additions & 2 deletions .github/workflows/scripts/models/model_test.sh
Original file line number Diff line number Diff line change
Expand Up @@ -39,9 +39,9 @@ working_dir=""
main() {
case ${tasks} in
"text-generation")
working_dir="/GenAIEval/GenAIEval/evaluation/lm_evaluation_harness/examples";;
working_dir="/GenAIEval/evals/evaluation/lm_evaluation_harness/examples";;
"code-generation")
working_dir="/GenAIEval/GenAIEval/evaluation/bigcode_evaluation_harness/examples";;
working_dir="/GenAIEval/evals/evaluation/bigcode_evaluation_harness/examples";;
*)
echo "Not suppotted task"; exit 1;;
esac
Expand Down
2 changes: 1 addition & 1 deletion .github/workflows/scripts/unittest/compare_coverage.sh
Original file line number Diff line number Diff line change
Expand Up @@ -20,7 +20,7 @@ coverage_PR_lines_rate=$5
coverage_base_lines_rate=$6
coverage_PR_branches_rate=$7
coverage_base_branches_rate=$8
module_name="GenAIEval"
module_name="evals"
[[ ! -f $coverage_pr_log ]] && exit 1
[[ ! -f $coverage_base_log ]] && exit 1
file_name="./coverage_compare"
Expand Down
6 changes: 1 addition & 5 deletions .github/workflows/unittest.yml
Original file line number Diff line number Diff line change
Expand Up @@ -49,9 +49,6 @@ jobs:
steps:
- name: Clean Up Working Directory
run: sudo rm -rf ${{github.workspace}}/*
- name: Load environment variables
run:
cat ~/actions-runner4/.env >> $GITHUB_ENV
- name: Checkout out Repo
uses: actions/checkout@v4
with:
Expand All @@ -61,7 +58,7 @@ jobs:

- name: Docker Build
run: |
docker build --build-arg http_proxy="${{ env.HTTP_PROXY_IMAGE_BUILD }}" --build-arg https_proxy="${{ env.HTTPS_PROXY_IMAGE_BUILD }}" -f ${{ github.workspace }}/.github/workflows/docker/common.dockerfile -t ${{ env.DOCKER_NAME }}:${{ env.DOCKER_TAG }} .
docker build -f ${{ github.workspace }}/.github/workflows/docker/common.dockerfile -t ${{ env.DOCKER_NAME }}:${{ env.DOCKER_TAG }} .

- name: Docker Run
run: |
Expand All @@ -70,7 +67,6 @@ jobs:
docker rm -vf ${{ env.CONTAINER_NAME }} || true
fi
docker run -dit --memory="4g" --memory-reservation="1g" --disable-content-trust --privileged --name=${{ env.CONTAINER_NAME }} --shm-size="1g" \
-e http_proxy="${{ env.HTTP_PROXY_CONTAINER_RUN }}" -e https_proxy="${{ env.HTTPS_PROXY_CONTAINER_RUN }}" \
-v ${{ github.workspace }}:/GenAIEval ${{ env.DOCKER_NAME }}:${{ env.DOCKER_TAG }}

- name: Install Dependencies
Expand Down
12 changes: 6 additions & 6 deletions README.md
Original file line number Diff line number Diff line change
Expand Up @@ -16,7 +16,7 @@ For evaluating the models on text-generation tasks, we follow the [lm-evaluation
```shell

# pip install --upgrade-strategy eager optimum[habana]
cd GenAIEval/evaluation/lm_evaluation_harness/examples
cd evals/evaluation/lm_evaluation_harness/examples
python main.py \
--model gaudi-hf \
--model_args pretrained=EleutherAI/gpt-j-6B \
Expand All @@ -29,7 +29,7 @@ python main.py \
##### CPU
```shell

cd GenAIEval/evaluation/lm_evaluation_harness/examples
cd evals/evaluation/lm_evaluation_harness/examples
python main.py \
--model hf \
--model_args pretrained=EleutherAI/gpt-j-6B \
Expand All @@ -39,7 +39,7 @@ python main.py \
```
#### function call usage
```python
from GenAIEval.evaluation.lm_evaluation_harness import LMEvalParser, evaluate
from evals.evaluation.lm_evaluation_harness import LMEvalParser, evaluate

args = LMevalParser(
model="hf",
Expand Down Expand Up @@ -69,7 +69,7 @@ docker run -p 9006:9006 --ipc=host -e MODEL="hf" -e MODEL_ARGS="pretrained=Inte
- set `base_url`, `tokenizer` and `--model genai-hf`

```
cd GenAIEval/evaluation/lm_evaluation_harness/examples
cd evals/evaluation/lm_evaluation_harness/examples

python main.py \
--model genai-hf \
Expand All @@ -83,7 +83,7 @@ For evaluating the models on coding tasks or specifically coding LLMs, we follow
#### command line usage

```shell
cd GenAIEval/evaluation/bigcode_evaluation_harness/examples
cd evals/evaluation/bigcode_evaluation_harness/examples
python main.py \
--model "codeparrot/codeparrot-small" \
--tasks "humaneval" \
Expand All @@ -93,7 +93,7 @@ python main.py \
```
#### function call usage
```python
from GenAIEval.evaluation.bigcode_evaluation_harness import BigcodeEvalParser, evaluate
from evals.evaluation.bigcode_evaluation_harness import BigcodeEvalParser, evaluate

args = BigcodeEvalParser(
user_model=user_model,
Expand Down
File renamed without changes.
File renamed without changes.
File renamed without changes.
File renamed without changes.
File renamed without changes.
Original file line number Diff line number Diff line change
Expand Up @@ -14,7 +14,7 @@
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
from GenAIEval.evaluation.bigcode_evaluation_harness import evaluate, setup_parser
from evals.evaluation.bigcode_evaluation_harness import evaluate, setup_parser


def main():
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -15,7 +15,7 @@
# See the License for the specific language governing permissions and
# limitations under the License.

from GenAIEval.evaluation.lm_evaluation_harness import evaluate, setup_parser
from evals.evaluation.lm_evaluation_harness import evaluate, setup_parser


def main():
Expand Down
4 changes: 2 additions & 2 deletions setup.py
Original file line number Diff line number Diff line change
Expand Up @@ -26,8 +26,8 @@ def parse_requirements(filename):


setup(
name="GenAIEval",
version="0.0.0",
name="opea_eval",
version="0.6",
author="Intel AISE AIPC Team",
author_email="[email protected], [email protected], [email protected], [email protected]",
description="Evaluation and benchmark for Generative AI",
Expand Down
1 change: 1 addition & 0 deletions tests/requirements.txt
Original file line number Diff line number Diff line change
@@ -1 +1,2 @@
bigcode-eval@git+https://github.com/bigcode-project/bigcode-evaluation-harness.git@a1b4a7949a24c8e3ef0d05a01097b2d14ffba56e
lm-eval==0.4.2
2 changes: 1 addition & 1 deletion tests/test_bigcode_eval.py
Original file line number Diff line number Diff line change
Expand Up @@ -19,7 +19,7 @@

from transformers import AutoModelForCausalLM, AutoTokenizer

from GenAIEval.evaluation.bigcode_evaluation_harness import BigcodeEvalParser, evaluate
from evals.evaluation.bigcode_evaluation_harness import BigcodeEvalParser, evaluate


class TestLMEval(unittest.TestCase):
Expand Down
2 changes: 1 addition & 1 deletion tests/test_lm_eval.py
Original file line number Diff line number Diff line change
Expand Up @@ -19,7 +19,7 @@

from transformers import AutoModelForCausalLM, AutoTokenizer

from GenAIEval.evaluation.lm_evaluation_harness import LMEvalParser, evaluate
from evals.evaluation.lm_evaluation_harness import LMEvalParser, evaluate


class TestLMEval(unittest.TestCase):
Expand Down