-
Notifications
You must be signed in to change notification settings - Fork 0
Open
Labels
documentationImprovements or additions to documentationImprovements or additions to documentation
Description
OpenAI API information:
- GPT-4: 0.03$/1k tokens for input into the prompt and 0.06$/1k tokens for the GPT-4's output.
- GPT-3.5 Turbo: 0.0015$/1k tokens for input sent to the LLM and 0.02$/1k tokens for the LLM's output.
- OpenAI embeddings: 0.0001$/1k tokens
We basically need to estimate how many tokens we feed into the model. We need to see how many tokens the schema linking is gonna take. How many tokens the query classification (easy, medium, hard) is going to take and how many tokens the SQL query classification and the self-correction phase is going to take.
Take 100 examples from the Spider dataset and come up with the estimate.
As a continuation you can even take 100 examples each from both, the BIRD dataset and the WikiSQL dataset.
Check out this article by OpenAI: How to count tokens with tiktoken
Metadata
Metadata
Assignees
Labels
documentationImprovements or additions to documentationImprovements or additions to documentation