-
Notifications
You must be signed in to change notification settings - Fork 1
New FPGA Model Streaming #20
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
cmatthews20
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
@Daayim merge main to your branch and resolve conflicts - looks like umama merged before you and you didn't have those changes
K-rolls
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
umama-rahman1
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Overall looks like great work, make sure to create pull request on Config Engine as well so that code is visible for review too. I had a peek on the branch.
Just make sure to add the get_fpga_inference_url endpoint and it should be good.
K-rolls
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Great work getting those edits in sir!
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Looks good. Thanks for adding in the get_inference_url for FPGA.
Approving now.
The only thing to check is the Frontend @cmatthews20 using the FPGA inference url to do the chat. If there are any issues, we should be able to update the config-engine repo. (Might need to do JSON)

Related Task
Changes
Documentation
FastAPI Doc
Model Streaming Testing
Get Inference URL Endpoint
Local Testing
Terminal Log confirming the automated fastapi deployment
Example query prompt