
The Ruby AI Podcast Contracts and Code: The Realities of AI Development
12 snips
Sep 23, 2025 Valentino and Joe dive into the reality behind AI salaries, discussing the disparity between hype and actual compensation. They debate whether any company can truly dominate the LLM market, emphasizing incremental improvements over clear winners. The hosts explore the complexities of benchmarking AI models and the necessity for customized evaluation tools. New OpenAI features that enhance prompt engineering are discussed, alongside the balance between playful experimentation and standardization in Ruby, highlighting its role in AI development.
AI Snips
Chapters
Transcript
Episode notes
Eval Tools Rarely Fit Out Of The Box
- Eval tooling often falls short because products need custom metrics and calculators.
- Teams end up rolling their own evaluation pipelines to measure the specific things they care about.
Prefer Native Tracing Or Build Bridges
- Favor language-native tooling where possible for tracing and observability.
- If SaaS tracing lacks Ruby support, be prepared to integrate or roll your own bridge for full observability.
AI Products Break Traditional SaaS Economics
- AI-driven products have high variable inference costs, so per-user economics look very different than classic SaaS.
- Continuous or always-on workloads quickly amplify model costs and demand model-cost tradeoffs.
