You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: aiperf/README.md
+30-35Lines changed: 30 additions & 35 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -21,41 +21,36 @@ Instead of manually running AIPerf multiple times with different parameters, you
21
21
22
22
These steps have been tested with Python 3.11.11.
23
23
To use the provided configurations, you need to create accounts at https://build.nvidia.com/ and [Huggingface](https://huggingface.co/).
24
-
The provided configurations use models hosted at https://build.nvidia.com/, you'll need to create a Personal API Key to access the models.
25
-
AIperf requires the [Meta Llama 3.3 70B Instruct tokenizer](https://huggingface.co/meta-llama/Llama-3.3-70B-Instruct) to calculate token-counts.
24
+
*The provided configurations use models hosted at https://build.nvidia.com/, you'll need to create a Personal API Key to access the models.
25
+
* The provided AIperf configurations require the [Meta Llama 3.3 70B Instruct tokenizer](https://huggingface.co/meta-llama/Llama-3.3-70B-Instruct) to calculate token-counts.
26
26
27
-
1.**Install NeMo Guardrails with developer tooling:**
27
+
1.**Create a virtual environment in which to install AIPerf**
28
28
29
29
```bash
30
-
poetry install --with dev
30
+
$ mkdir ~/env
31
+
$ python -m venv ~/env/aiperf
31
32
```
32
33
33
-
2.**Install AIPerf and NVIDIA AI Endpoints:**
34
+
2.**Install dependencies in the virtual environment**
34
35
35
36
```bash
36
-
poetry run pip install aiperf langchain-nvidia-ai-endpoints
37
+
$ pip install aiperf huggingface_hub typer
37
38
```
38
39
39
-
3.** Install/upgrade Hugging Face Hub:**
40
-
41
-
AIPerf needs a tokenizer to run and will download one from Hugging Face if available. If you have the tokenizer locally, you can point to that directory and not log into Huggingface.
42
-
43
-
```bash
44
-
poetry run pip install --upgrade huggingface_hub
45
-
```
46
-
47
-
4.** Login to Hugging Face:**
40
+
3.** Login to Hugging Face:**
48
41
49
42
```bash
50
43
huggingface-cli login
51
44
```
52
45
53
-
5.** Set NVIDIA API Key:**
46
+
4.** Set NVIDIA API Key:**
54
47
55
-
To use models hosted on [build.nvidia.com](https://build.nvidia.com/), set your API key:
48
+
The provided configs use models hosted on [build.nvidia.com](https://build.nvidia.com/).
49
+
To access these, [create an account](https://build.nvidia.com/), and create a Personal API Key.
50
+
After creating a Personal API key, set the `NVIDIA_API_KEY` variable as below.
56
51
57
52
```bash
58
-
export NVIDIA_API_KEY="your-api-key-here"
53
+
$ export NVIDIA_API_KEY="your-api-key-here"
59
54
```
60
55
61
56
## Running Benchmarks
@@ -70,39 +65,39 @@ There are two example configs included which can be extended for your use-cases.
70
65
To run a benchmark, use the following command:
71
66
72
67
```bash
73
-
poetry run nemoguardrails aiperf run --config-file <path-to-config.yaml>
0 commit comments