OctoAI Logo
webinar icon in OctoAI blueOn-demand Webinar

Optimizing LLMs for Cost and Quality

Below-the-line response quality and prohibitively expensive inference are significant blockers to scaling LLMs today. This technical session will teach you a path using open source to achieve superior quality with cheaper/faster models to power your production applications.

In this session you’ll:

  • Understand why fine tuning models is critically important

  • Learn a proven “crawl, walk, run” path toward model quality optimization

  • See what the continuous development cycle looks like for LLM-powered apps

  • Watch a live demo showing how to fine-tune Llama3.1-8B to outperform a GPT-4o model at redacting personally identifiable information (PII) from your enterprise data


Session Resources:

Webinar preview image for Optimizing LLMs for cost and quality with a chart showing starting GenAI projects with high costs, and moving to lower costs and improved quality
52 min