Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

chapter: mlops-engines #28

Merged
merged 15 commits into from
Sep 18, 2023
Merged

chapter: mlops-engines #28

merged 15 commits into from
Sep 18, 2023

Conversation

htrivedi99
Copy link

@htrivedi99 htrivedi99 commented Aug 25, 2023

Review checklist

Don't worry about satisfying all items, it's fine to open a (draft) PR.

  • chapter content
    • only one top-level # h1-Title
    • summary (e.g. table or TL;DR overview), no need for an explicit ## Summary/Introduction title or equivalent
    • main content focus: recent developments in open source AI
      • general context/background (brief)
      • current pros/cons
      • in-depth insights (not yet widely known)
    • likely ## Future developments
    • end with {{ comments }}
  • appropriate citations
    • BibTeX references
    • Glossary terms
    • cross-references (figures/chapters)
    • (if new-chapter.md), add _toc.yml entry & index.md table row
    • If CI URL checks have false-positives, append to _config.yml:sphinx.config.linkcheck*
  • images & data not committed to this repo (e.g. use https://github.com/premAI-io/static.premai.io instead)

fixes #4

Copy link
Contributor

@casperdcl casperdcl left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

  • intro too long
  • intro mostly off-topic (what does it have to do with MLOps Engines? a lot of this looks more relevant to other chapters)
  • missing clear definition of "MLOps Engine"
  • missing list/table of MLOps Engines & feature comparison (see e.g. this)
  • missing justifications (every claim needs to be backed up, you can't just state personal opinions)
  • use clean book-not-blog language, e.g.
    • ## Some Thoughts About The Future -> ## Future
    • With large language models, the story is no different -> LLMs are similar
    • stick to third-person, and definitely don't keep changing between first, second, and third
    • avoid repetition
  • did you see the original section outline? e.g. Python Bindings, Apache TVM, links to read before writing anything, etc.

mlops-engines.md Outdated

## The MLOps Lifecycle

![](https://static.premai.io/book/mlops-engines-LLMOps-diagram.jpg)
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I'm not sure I understand this diagram - what is it trying to show?

Copy link
Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

This diagram shows the LLMOps lifecycle. Its just suppose to serve as a banner image

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I'm particularly struggling to understand the meaning of the arrows and the colours.

mlops-engines.md Outdated Show resolved Hide resolved
mlops-engines.md Outdated Show resolved Hide resolved
mlops-engines.md Outdated

## Challenges With Open-Source MLOps

MLOps has always been available in two flavors. One is the managed version, where all the components are provided out of the box for a steep price. The other is a DIY setup where you stitch together various open-source components. 
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

citation needed

Copy link
Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

citation added

mlops-engines.md Outdated Show resolved Hide resolved
mlops-engines.md Outdated Show resolved Hide resolved
mlops-engines.md Outdated Show resolved Hide resolved
mlops-engines.md Outdated Show resolved Hide resolved
mlops-engines.md Outdated Show resolved Hide resolved
mlops-engines.md Outdated
Comment on lines 163 to 165
Due to the challenge of running LLMs, enterprises will opt to use an inference server instead of containerizing the model in-house. Most companies don't have the expertise to optimize these models, but they still want the performance benefits. Inference servers, whether they are open-source or not, will be the path forward.

Another pattern that's emerging is that models will move to the data instead of the data moving to the model. Right now if you call the ChatGPT API, you would be sending your data to the model. Enterprises have worked very hard over the past decade to set up robust data infrastructure in the cloud. It makes a lot more sense to bring the model into the same cloud environment where the data is. This is where open-source models being cloud agnostic have a huge advantage.
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

citations needed. If it's your personal opinion, sate why. Otherwise assume the reader disagrees with every claim you make.

Copy link
Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

added citations

mlops-engines.md Outdated
* Takes a while for new LLMs to be supported


Many other open-source projects like [BentoML](https://www.bentoml.com/), [FastAPI](https://fastapi.tiangolo.com/), and [Flask](https://flask.palletsprojects.com/en/2.3.x/) have been used for serving models in the past. The reason I have not included them on this list is that these open-source tools do not provide the optimizations you need to run LLMs in production.
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

that's not a reason. Also, what are "the optimizations you need to run LLMs in production"?

@casperdcl casperdcl added the content text & code label Aug 31, 2023
@casperdcl casperdcl changed the title adding rough draft for mlops engines chapter: mlops-engines Sep 1, 2023
@casperdcl casperdcl force-pushed the mlops-engines branch 2 times, most recently from 671da89 to 3c8bae8 Compare September 18, 2023 10:58
@casperdcl casperdcl merged commit 3a742c6 into main Sep 18, 2023
1 check failed
@casperdcl casperdcl deleted the mlops-engines branch September 18, 2023 19:53
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
content text & code
Projects
No open projects
Status: Done
Development

Successfully merging this pull request may close these issues.

chapter: mlops-engines
2 participants