LLMOps: EVALUATING AND FINE TUNING LLM MODELS FOR GENERATIVE AI
Keywords:
Generative AI, LLM, MLOps, LLM Ops, Cost Optimization, Observability, Model CIAbstract
As companies adopt Generative AI using LLM models, many pre-trained and fine-tuned models needs to be evaluated for its accuracy. LLMOps is a derivative of MLOps but specialized on model training and finetuning LLMs. Model evaluation process can take a lot of time and cost a fortune for companies, with LLMOps using model CI pipelines, frameworks and automation, the drawbacks are addressed and help organizations evaluate models quickly and, in a cost, optimized manner. This paper discusses how the pipelines, frameworks, Observability metrics collected during training can be used to optimally evaluate LLM models.
References
https://github.com/EleutherAI/lm-evaluation-harness
https://en.wikipedia.org/wiki/MLOps
Prasun Mishra, LLMOps: The Next Frontier of MLOps for Generative AI
https://www.datarobot.com/wp-content/uploads/2022/01/DataRobot-Report-State-of-AI-Bias_V5.pdf
https://www.fiddler.ai/ml-model-monitoring/model-monitoring-framework
Prasun Mishra, Unlocking the Power of Open LLMs and Generative AI: A 10-Step Guide to Finding Your Perfect Language Model
https://github.com/evidentlyai/evidently
https://www.databricks.com/glossary/mlops
https://barc-research.com/press-release-dataops-mlops-ml-challenges/
Downloads
Published
Issue
Section
License
Copyright (c) 2023 Amreth Chandrasehar (Author)

This work is licensed under a Creative Commons Attribution-NonCommercial 4.0 International License.