Ask HN: How does deploying a fine-tuned model work
If I've managed to build my own model, say a fine-tuned version of Llama and trained it on some GPUs, how do I then deploy it and use it in an app. Does it need to be running on the GPUs all the time or can I host the model on a web server or something. Sorry if this is an obvious/misinformed question, I'm a beginner in this space 0 comments on Hacker News.
If I've managed to build my own model, say a fine-tuned version of Llama and trained it on some GPUs, how do I then deploy it and use it in an app. Does it need to be running on the GPUs all the time or can I host the model on a web server or something. Sorry if this is an obvious/misinformed question, I'm a beginner in this space
If I've managed to build my own model, say a fine-tuned version of Llama and trained it on some GPUs, how do I then deploy it and use it in an app. Does it need to be running on the GPUs all the time or can I host the model on a web server or something. Sorry if this is an obvious/misinformed question, I'm a beginner in this space 0 comments on Hacker News.
If I've managed to build my own model, say a fine-tuned version of Llama and trained it on some GPUs, how do I then deploy it and use it in an app. Does it need to be running on the GPUs all the time or can I host the model on a web server or something. Sorry if this is an obvious/misinformed question, I'm a beginner in this space
Hacker News story: Ask HN: How does deploying a fine-tuned model work
Reviewed by Tha Kur
on
April 23, 2024
Rating:
No comments: