-
Notifications
You must be signed in to change notification settings - Fork 90
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
chapter: mlops-engines #28
Conversation
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
- intro too long
- intro mostly off-topic (what does it have to do with MLOps Engines? a lot of this looks more relevant to other chapters)
- missing clear definition of "MLOps Engine"
- missing list/table of MLOps Engines & feature comparison (see e.g. this)
- missing justifications (every claim needs to be backed up, you can't just state personal opinions)
- use clean book-not-blog language, e.g.
## Some Thoughts About The Future
->## Future
With large language models, the story is no different
->LLMs are similar
- stick to third-person, and definitely don't keep changing between first, second, and third
- avoid repetition
- did you see the original section outline? e.g. Python Bindings, Apache TVM, links to read before writing anything, etc.
mlops-engines.md
Outdated
|
||
## The MLOps Lifecycle | ||
|
||
![](https://static.premai.io/book/mlops-engines-LLMOps-diagram.jpg) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I'm not sure I understand this diagram - what is it trying to show?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
This diagram shows the LLMOps lifecycle. Its just suppose to serve as a banner image
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I'm particularly struggling to understand the meaning of the arrows and the colours.
mlops-engines.md
Outdated
|
||
## Challenges With Open-Source MLOps | ||
|
||
MLOps has always been available in two flavors. One is the managed version, where all the components are provided out of the box for a steep price. The other is a DIY setup where you stitch together various open-source components. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
citation needed
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
citation added
mlops-engines.md
Outdated
Due to the challenge of running LLMs, enterprises will opt to use an inference server instead of containerizing the model in-house. Most companies don't have the expertise to optimize these models, but they still want the performance benefits. Inference servers, whether they are open-source or not, will be the path forward. | ||
|
||
Another pattern that's emerging is that models will move to the data instead of the data moving to the model. Right now if you call the ChatGPT API, you would be sending your data to the model. Enterprises have worked very hard over the past decade to set up robust data infrastructure in the cloud. It makes a lot more sense to bring the model into the same cloud environment where the data is. This is where open-source models being cloud agnostic have a huge advantage. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
citations needed. If it's your personal opinion, sate why. Otherwise assume the reader disagrees with every claim you make.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
added citations
mlops-engines.md
Outdated
* Takes a while for new LLMs to be supported | ||
|
||
|
||
Many other open-source projects like [BentoML](https://www.bentoml.com/), [FastAPI](https://fastapi.tiangolo.com/), and [Flask](https://flask.palletsprojects.com/en/2.3.x/) have been used for serving models in the past. The reason I have not included them on this list is that these open-source tools do not provide the optimizations you need to run LLMs in production. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
that's not a reason. Also, what are "the optimizations you need to run LLMs in production"?
671da89
to
3c8bae8
Compare
11eb2f9
to
53b7dca
Compare
Review checklist
# h1-Title
## Summary/Introduction
title or equivalent## Future
developments{{ comments }}
new-chapter.md
), add_toc.yml
entry &index.md
table row_config.yml:sphinx.config.linkcheck*
fixes #4