Skip to content

Commit 648e61b

Browse files
committed
update
1 parent 5f35161 commit 648e61b

File tree

1 file changed

+130
-0
lines changed

1 file changed

+130
-0
lines changed

MyApp/_posts/2025-09-02_text-to-blazor-upgraded.md

Lines changed: 130 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -181,3 +181,133 @@ in its [Configure.AutoQuery.cs](https://github.com/NetCoreTemplates/blazor-admin
181181
the Audit Behaviors will also maintain an Audit Trail of all CRUD operations which can be viewed in the Admin UI:
182182

183183
![](/img/posts/text-to-blazor/okai-audit-form.webp)
184+
185+
## FREE Access to worlds most popular AI Models
186+
187+
In addition to upgrading the models in text-to-blazor we've also made available them available to our
188+
generic **chat** prompt that can be used as a convenient way to conduct personal research against many
189+
of the worlds most popular Large Language Models - for Free!
190+
191+
![](/img/posts/okai-chat/okai-chat.webp)
192+
193+
No API Keys, no Signups, no installs, no cost, you can just start immediately using the `npx okai chat` script to ask LLMs
194+
for assistance:
195+
196+
:::sh
197+
npx okai chat "command to copy a folder with rsync?"
198+
:::
199+
200+
This will use the default model (currently gemini-flash 2.5) to answer your question.
201+
202+
### Select Preferred Model
203+
204+
You can also use your preferred model with the `-m <model>` flag with either the model **name** or its **alias**,
205+
e.g you can use
206+
[Microsoft's PHI-4 14B](https://techcommunity.microsoft.com/blog/aiplatformblog/introducing-phi-4-microsoft%E2%80%99s-newest-small-language-model-specializing-in-comple/4357090) model with:
207+
208+
:::sh
209+
npx okai -m phi chat "command to copy folder with rsync?"
210+
:::
211+
212+
### List Available Models
213+
214+
We're actively adding more great performing and leading experimental models as they're released.
215+
You can view the list of available models with `ls models`:
216+
217+
:::sh
218+
npx okai ls models
219+
:::
220+
221+
Which at this time will return the following list of available models along with instructions for how to use them:
222+
223+
```txt
224+
USAGE (5 models max):
225+
a) OKAI_MODELS=codestral,llama3.3,flash
226+
b) okai -models codestral,llama3.3,flash <prompt>
227+
c) okai -m flash chat <prompt>
228+
229+
FREE MODELS:
230+
claude-3-haiku (alias haiku)
231+
codestral:22b (alias codestral)
232+
deepseek-chat-v3.1:671b
233+
deepseek-r1:671b (alias deepseek-r1)
234+
deepseek-v3.1:671b (alias deepseek)
235+
gemini-2.5-flash (alias flash)
236+
gemini-2.5-flash-lite (alias flash-lite)
237+
gemini-flash-thinking-2.5
238+
gemma2:9b
239+
gemma3:27b (alias gemma)
240+
glm-4.5
241+
glm-4.5-air
242+
glm-4.5v
243+
gpt-4.1-mini
244+
gpt-4.1-nano
245+
gpt-4o-mini
246+
gpt-5-mini
247+
gpt-5-nano
248+
gpt-oss-120b
249+
gpt-oss-20b
250+
grok-code-fast-1 (alias grok-code)
251+
kimi-k2
252+
llama-4-maverick
253+
llama-4-scout
254+
llama3.1:70b
255+
llama3.1:8b
256+
llama3.3:70b (alias llama3)
257+
llama4:109b (alias llama4)
258+
llama4:400b
259+
mistral-nemo:12b (alias mistral-nemo)
260+
mistral-small:24b (alias mistral-small)
261+
mistral:7b (alias mistral)
262+
mixtral:8x22b (alias mixtral)
263+
nova-lite
264+
nova-micro
265+
phi-4:14b (alias phi,phi-4)
266+
qwen3-coder
267+
qwen3-coder:30b
268+
qwen3:235b
269+
qwen3:30b
270+
qwen3:32b
271+
qwen3:8b
272+
273+
PREMIUM MODELS: *
274+
claude-3-5-haiku
275+
claude-3-5-sonnet
276+
claude-3-7-sonnet
277+
claude-3-sonnet
278+
claude-sonnet-4-0 (alias sonnet)
279+
gemini-2.5-pro (alias gemini-pro)
280+
gpt-4.1
281+
gpt-4o
282+
gpt-5
283+
gpt-5-chat
284+
grok-4 (alias grok)
285+
mistral-large:123b
286+
nova-pro
287+
o4-mini
288+
o4-mini-high
289+
290+
* requires valid license:
291+
a) SERVICESTACK_LICENSE=<key>
292+
b) SERVICESTACK_CERTIFICATE=<LC-XXX>
293+
c) okai -models <premium,models> -license <license> <prompt>
294+
```
295+
296+
Where you'll be able to use any of the great performing inexpensive models listed under `FREE MODELS` for Free.
297+
Whilst ServiceStack customers with an active commercial license can also use any of the more expensive
298+
and better performing models listed under `PREMIUM MODELS` by either:
299+
300+
a) Setting the `SERVICESTACK_LICENSE` Environment Variable with your **License Key**
301+
b) Setting the `SERVICESTACK_CERTIFICATE` Variable with your **License Certificate**
302+
c) Inline using the `-license` flag with either the **License Key** or **Certificate**
303+
304+
### FREE for Personal Usage
305+
306+
To be able to maintain this as a free service we're limiting usage as a tool that developers can use for personal
307+
assistance and research by limiting usage to **60 requests /hour** which should be more than enough for most
308+
personal usage and research whilst deterring usage in automated tools.
309+
310+
:::tip info
311+
Rate limiting is implemented with a sliding [Token Bucket algorithm](https://en.wikipedia.org/wiki/Token_bucket)
312+
that replenishes 1 additional request every 60s
313+
:::

0 commit comments

Comments
 (0)