|
1 | 1 | # SPDX-License-Identifier: Apache-2.0 |
2 | 2 |
|
3 | | -name: E2E (NVIDIA A10G x1) |
| 3 | +name: E2E (NVIDIA L4 x1) |
4 | 4 |
|
5 | 5 | on: |
6 | 6 | # run against every merge commit to 'main' and release branches |
|
18 | 18 | - '**.py' |
19 | 19 | - 'pyproject.toml' |
20 | 20 | - 'requirements**.txt' |
21 | | - - '.github/workflows/e2e-nvidia-a10g-x1.yml' # This workflow |
| 21 | + - '.github/workflows/e2e-nvidia-l4-x1.yml' # This workflow |
22 | 22 |
|
23 | 23 | concurrency: |
24 | 24 | group: ${{ github.workflow }}-${{ github.event.pull_request.number || github.ref }} |
|
55 | 55 | mode: start |
56 | 56 | github-token: ${{ secrets.GH_PERSONAL_ACCESS_TOKEN }} |
57 | 57 | ec2-image-id: ${{ vars.AWS_EC2_AMI }} |
58 | | - ec2-instance-type: g5.4xlarge |
| 58 | + ec2-instance-type: g6.8xlarge |
59 | 59 | subnet-id: subnet-02d230cffd9385bd4 |
60 | 60 | security-group-id: sg-06300447c4a5fbef3 |
61 | 61 | iam-role-name: instructlab-ci-runner |
@@ -117,18 +117,18 @@ jobs: |
117 | 117 | nvidia-smi |
118 | 118 | python3.11 -m pip cache remove llama_cpp_python |
119 | 119 |
|
120 | | - CMAKE_ARGS="-DLLAMA_CUDA=on" python3.11 -m pip install . |
| 120 | + CMAKE_ARGS="-DLLAMA_CUDA=on" python3.11 -m pip install -v . |
121 | 121 |
|
122 | 122 | # https://github.com/instructlab/instructlab/issues/1821 |
123 | 123 | # install with Torch and build dependencies installed |
124 | | - python3.11 -m pip install packaging wheel setuptools-scm |
125 | | - python3.11 -m pip install .[cuda] -r requirements-vllm-cuda.txt |
| 124 | + python3.11 -m pip install -v packaging wheel setuptools-scm |
| 125 | + python3.11 -m pip install -v .[cuda] -r requirements-vllm-cuda.txt |
126 | 126 | |
127 | 127 | - name: Update instructlab-eval library |
128 | 128 | working-directory: ./eval |
129 | 129 | run: | |
130 | 130 | . ../instructlab/venv/bin/activate |
131 | | - pip install . |
| 131 | + pip install -v . |
132 | 132 |
|
133 | 133 | - name: Run e2e test |
134 | 134 | working-directory: ./instructlab |
|
0 commit comments