Large Language Models (LLMs) have enormous potential, but also challenge existing workflows in industry that require modularity, transparency and data privacy. In this talk, I'll show some practical solutions for using the latest state-of-the-art models in real-world applications and distilling their knowledge into smaller and faster components that you can run and maintain in-house.
_________________________________________________
▪️ Case Study #1: https://speakerdeck.com/inesmontani/workshop-half-hour-of-labeling-power-can-we-beat-gpt
▪️ Case Study #2: https://explosion.ai/blog/sp-global-commodities
▪️ Case Study #3: https://explosion.ai/blog/gitlab-support-insights
https://explosion.ai/blog/human-in-the-loop-distillation
This blog post presents practical solutions for using the latest state-of-the-art models in real-world applications and distilling their knowledge into smaller and faster components that you can run and maintain in-house.
https://explosion.ai/blog/applied-nlp-thinking
This blog post discusses some of the biggest challenges for applied NLP and translating business problems into machine learning solutions, including the distinction between utility and accuracy.
https://speakerdeck.com/inesmontani/workshop-half-hour-of-labeling-power-can-we-beat-gpt
A case study using LLMs to create data and beating the few-shot baseline with a distilled task-specific model for extracting dishes, ingredients and equipment from r/cooking Reddit posts.
https://explosion.ai/blog/sp-global-commodities
A case study on S&P Global’s efficient information extraction pipelines for real-time commodities trading insights in a high-security environment using human-in-the-loop distillation.
https://explosion.ai/blog/gitlab-support-insights
A case study on GitLab’s large-scale NLP pipelines for extracting actionable insights from support tickets and usage questions.
https://prodi.gy/docs/large-language-models
Prodigy comes with preconfigured workflows for using LLMs to speed up and automate annotation and create datasets for distilling large generative models into more accurate, smaller, faster and fully private task-specific components.