Hacker News story: Ask HN: How does deploying a fine-tuned model work

Ask HN: How does deploying a fine-tuned model work
If I've managed to build my own model, say a fine-tuned version of Llama and trained it on some GPUs, how do I then deploy it and use it in an app. Does it need to be running on the GPUs all the time or can I host the model on a web server or something. Sorry if this is an obvious/misinformed question, I'm a beginner in this space 0 comments on Hacker News.
If I've managed to build my own model, say a fine-tuned version of Llama and trained it on some GPUs, how do I then deploy it and use it in an app. Does it need to be running on the GPUs all the time or can I host the model on a web server or something. Sorry if this is an obvious/misinformed question, I'm a beginner in this space

Hacker News story: Ask HN: How does deploying a fine-tuned model work Hacker News story: Ask HN: How does deploying a fine-tuned model work Reviewed by Tha Kur on April 23, 2024 Rating: 5

No comments:

Powered by Blogger.