You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
In addition to upgrading the models in text-to-blazor we've also made available them available to our
188
+
generic **chat** prompt that can be used as a convenient way to conduct personal research against many
189
+
of the worlds most popular Large Language Models - for Free!
190
+
191
+

192
+
193
+
No API Keys, no Signups, no installs, no cost, you can just start immediately using the `npx okai chat` script to ask LLMs
194
+
for assistance:
195
+
196
+
:::sh
197
+
npx okai chat "command to copy a folder with rsync?"
198
+
:::
199
+
200
+
This will use the default model (currently gemini-flash 2.5) to answer your question.
201
+
202
+
### Select Preferred Model
203
+
204
+
You can also use your preferred model with the `-m <model>` flag with either the model **name** or its **alias**,
205
+
e.g you can use
206
+
[Microsoft's PHI-4 14B](https://techcommunity.microsoft.com/blog/aiplatformblog/introducing-phi-4-microsoft%E2%80%99s-newest-small-language-model-specializing-in-comple/4357090) model with:
207
+
208
+
:::sh
209
+
npx okai -m phi chat "command to copy folder with rsync?"
210
+
:::
211
+
212
+
### List Available Models
213
+
214
+
We're actively adding more great performing and leading experimental models as they're released.
215
+
You can view the list of available models with `ls models`:
216
+
217
+
:::sh
218
+
npx okai ls models
219
+
:::
220
+
221
+
Which at this time will return the following list of available models along with instructions for how to use them:
222
+
223
+
```txt
224
+
USAGE (5 models max):
225
+
a) OKAI_MODELS=codestral,llama3.3,flash
226
+
b) okai -models codestral,llama3.3,flash <prompt>
227
+
c) okai -m flash chat <prompt>
228
+
229
+
FREE MODELS:
230
+
claude-3-haiku (alias haiku)
231
+
codestral:22b (alias codestral)
232
+
deepseek-chat-v3.1:671b
233
+
deepseek-r1:671b (alias deepseek-r1)
234
+
deepseek-v3.1:671b (alias deepseek)
235
+
gemini-2.5-flash (alias flash)
236
+
gemini-2.5-flash-lite (alias flash-lite)
237
+
gemini-flash-thinking-2.5
238
+
gemma2:9b
239
+
gemma3:27b (alias gemma)
240
+
glm-4.5
241
+
glm-4.5-air
242
+
glm-4.5v
243
+
gpt-4.1-mini
244
+
gpt-4.1-nano
245
+
gpt-4o-mini
246
+
gpt-5-mini
247
+
gpt-5-nano
248
+
gpt-oss-120b
249
+
gpt-oss-20b
250
+
grok-code-fast-1 (alias grok-code)
251
+
kimi-k2
252
+
llama-4-maverick
253
+
llama-4-scout
254
+
llama3.1:70b
255
+
llama3.1:8b
256
+
llama3.3:70b (alias llama3)
257
+
llama4:109b (alias llama4)
258
+
llama4:400b
259
+
mistral-nemo:12b (alias mistral-nemo)
260
+
mistral-small:24b (alias mistral-small)
261
+
mistral:7b (alias mistral)
262
+
mixtral:8x22b (alias mixtral)
263
+
nova-lite
264
+
nova-micro
265
+
phi-4:14b (alias phi,phi-4)
266
+
qwen3-coder
267
+
qwen3-coder:30b
268
+
qwen3:235b
269
+
qwen3:30b
270
+
qwen3:32b
271
+
qwen3:8b
272
+
273
+
PREMIUM MODELS: *
274
+
claude-3-5-haiku
275
+
claude-3-5-sonnet
276
+
claude-3-7-sonnet
277
+
claude-3-sonnet
278
+
claude-sonnet-4-0 (alias sonnet)
279
+
gemini-2.5-pro (alias gemini-pro)
280
+
gpt-4.1
281
+
gpt-4o
282
+
gpt-5
283
+
gpt-5-chat
284
+
grok-4 (alias grok)
285
+
mistral-large:123b
286
+
nova-pro
287
+
o4-mini
288
+
o4-mini-high
289
+
290
+
* requires valid license:
291
+
a) SERVICESTACK_LICENSE=<key>
292
+
b) SERVICESTACK_CERTIFICATE=<LC-XXX>
293
+
c) okai -models <premium,models> -license <license> <prompt>
294
+
```
295
+
296
+
Where you'll be able to use any of the great performing inexpensive models listed under `FREE MODELS` for Free.
297
+
Whilst ServiceStack customers with an active commercial license can also use any of the more expensive
298
+
and better performing models listed under `PREMIUM MODELS` by either:
299
+
300
+
a) Setting the `SERVICESTACK_LICENSE` Environment Variable with your **License Key**
301
+
b) Setting the `SERVICESTACK_CERTIFICATE` Variable with your **License Certificate**
302
+
c) Inline using the `-license` flag with either the **License Key** or **Certificate**
303
+
304
+
### FREE for Personal Usage
305
+
306
+
To be able to maintain this as a free service we're limiting usage as a tool that developers can use for personal
307
+
assistance and research by limiting usage to **60 requests /hour** which should be more than enough for most
308
+
personal usage and research whilst deterring usage in automated tools.
309
+
310
+
:::tip info
311
+
Rate limiting is implemented with a sliding [Token Bucket algorithm](https://en.wikipedia.org/wiki/Token_bucket)
0 commit comments