-
Notifications
You must be signed in to change notification settings - Fork 2
78 lines (65 loc) · 2.29 KB
/
test-gemini.yml
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
name: Test with Gemini
on: [workflow_dispatch, push, pull_request]
jobs:
zero-shot:
runs-on: ubuntu-22.04
timeout-minutes: 5
steps:
- uses: actions/checkout@v4
- run: echo 'Which planet in our solar system is the largest?' | ./query-llm.js | tee output.txt
env:
LLM_API_BASE_URL: 'https://generativelanguage.googleapis.com/v1beta'
LLM_API_KEY: ${{ secrets.GEMINI_API_KEY }}
LLM_CHAT_MODEL: 'gemini-1.5-flash-8b'
LLM_ZERO_SHOT: 1
- run: cat output.txt
- run: grep -i jupiter output.txt
chain-of-thought:
runs-on: ubuntu-22.04
timeout-minutes: 5
steps:
- uses: actions/checkout@v4
- run: echo 'Which planet in our solar system is the largest?' | ./query-llm.js | tee output.txt
env:
LLM_API_BASE_URL: 'https://generativelanguage.googleapis.com/v1beta'
LLM_API_KEY: ${{ secrets.GEMINI_API_KEY }}
LLM_CHAT_MODEL: 'gemini-1.5-flash-8b'
LLM_JSON_SCHEMA: 1
- run: cat output.txt
- run: grep -i jupiter output.txt
multi-turn:
needs: chain-of-thought
runs-on: ubuntu-22.04
timeout-minutes: 10
steps:
- uses: actions/checkout@v4
- run: ./query-llm.js tests/canary-multi-turn.txt
env:
LLM_API_BASE_URL: 'https://generativelanguage.googleapis.com/v1beta'
LLM_API_KEY: ${{ secrets.GEMINI_API_KEY }}
LLM_CHAT_MODEL: 'gemini-1.5-flash-8b'
LLM_JSON_SCHEMA: 1
high-school-stem:
needs: chain-of-thought
runs-on: ubuntu-22.04
timeout-minutes: 15
steps:
- uses: actions/checkout@v4
- run: ./query-llm.js tests/high-school-stem.txt
env:
LLM_API_BASE_URL: 'https://generativelanguage.googleapis.com/v1beta'
LLM_API_KEY: ${{ secrets.GEMINI_API_KEY }}
LLM_CHAT_MODEL: 'gemini-1.5-flash-8b'
LLM_JSON_SCHEMA: 1
general-knowledge:
needs: chain-of-thought
runs-on: ubuntu-22.04
timeout-minutes: 15
steps:
- uses: actions/checkout@v4
- run: ./query-llm.js tests/general-knowledge.txt
env:
LLM_API_BASE_URL: 'https://generativelanguage.googleapis.com/v1beta'
LLM_API_KEY: ${{ secrets.GEMINI_API_KEY }}
LLM_CHAT_MODEL: 'gemini-1.5-flash-8b'
LLM_JSON_SCHEMA: 1