12. Staying Current
📋 Jump to TakeawaysThe Pace of Change
New models release almost weekly. In the time between you starting and finishing this course, there's probably a new model announcement. This is overwhelming if you try to keep up with everything, but manageable if you have a system.
The good news: the fundamentals from this course don't change. Model selection frameworks, prompting principles, cost optimization strategies, and evaluation methods all stay relevant regardless of which specific model is on top this month.
What Actually Changes
When a new model drops, only a few things matter:
- Is it better at my specific task? Run your evaluation test set (lesson 10) against it.
- Is it cheaper? Check the pricing. A new model that's 50% cheaper at the same quality is worth switching to.
- Does it have a new capability? Larger context window, better structured output, new modality (audio, video).
Everything else is noise. Benchmark scores, Twitter hype, "this model is 2% better on MMLU." None of that matters unless it translates to better results on your actual workload.
A System for Staying Current
Monthly check-in (15 minutes)
Once a month, scan for new releases:
- Check the pricing pages of OpenAI, Anthropic, and Google
- Check Ollama's model library for new local options
- Skim one or two AI newsletters (The Batch, Simon Willison's blog)
You're looking for: new models, price drops, and new features. That's it.
Quarterly evaluation (1 to 2 hours)
Every 3 months, re-run your evaluation test set against the current best options:
Q1: "Is GPT-4o-mini still the best for my classification task?"
→ Run test set against mini, Haiku, Gemini Flash, and any new models
→ Compare accuracy and cost
→ Switch if something is clearly betterThis prevents you from staying on an outdated model out of inertia while also preventing you from chasing every new release.
When to switch immediately
Drop everything and evaluate a new model when:
- A model in your price tier drops to half the cost (e.g., a new mini model launches)
- A new capability unlocks something you couldn't do before (e.g., 1M context when you needed it)
- Your current model is being deprecated
What to Ignore
Benchmark leaderboards. Models are increasingly trained to perform well on benchmarks specifically. A model that's #1 on HumanEval might not be the best at your particular code generation task. Your own evaluation matters more than any leaderboard.
Hype cycles. Every new model is "the best ever" for about two weeks. Wait for the dust to settle. If people are still talking about it a month later, it's probably worth evaluating.
Minor version bumps. GPT-4o-2025-01-15 vs GPT-4o-2025-03-01. These are usually minor quality tweaks. Don't re-evaluate for every point release unless you notice a regression in your outputs.
Model count. You don't need to know about every model. Focus on the 3 to 4 that matter for your use cases and ignore the rest.
Future-Proofing Your Setup
Build your systems so that swapping models is easy:
Abstract the model call. Don't hardcode model names throughout your codebase. Use a config or environment variable so switching is a one-line change.
# ✅ Good: model is configurable
MODEL=gpt-4o-mini node classify.js
# ❌ Bad: model is hardcoded in 15 places
fetch("https://api.openai.com/v1/...", { model: "gpt-4o-mini" })Keep your evaluation test set. This is the most valuable asset you'll build. When a new model comes out, you can evaluate it in minutes instead of starting from scratch.
Use OpenAI-compatible APIs. Most providers (Anthropic, Ollama, Together, Groq) offer OpenAI-compatible endpoints. If your code talks to the OpenAI format, switching providers is a one-line URL change.
Version your prompts. When you change models, you might need to adjust prompts. Keep old prompts around so you can roll back if the new model doesn't work as well.
Trends Worth Watching
These are the directions the field is moving. They'll affect your model choices over the next year:
Cheaper and faster. Every generation of models gets cheaper. What costs $0.01 today will cost $0.001 next year. Plan for this by designing features that would be too expensive today but will become viable as prices drop.
Longer context windows. 1M tokens is available now. 10M is coming. This changes how you think about RAG vs stuffing everything in context.
Better small models. The 7B to 14B range keeps getting more capable. Tasks that required GPT-4 a year ago now work with a local 7B model. Keep re-evaluating what you can run locally.
Specialized models. Models fine-tuned for specific tasks (code, math, medical, legal) will outperform general models on those tasks at lower cost.
Multimodal everything. Text, images, audio, video in one model. This opens up use cases that were impossible with text-only models.
The Meta-Skill
The most valuable skill isn't knowing which model is best right now. It's knowing how to evaluate any new model quickly against your specific needs. That's what this course taught you:
- Understand what models can do (lessons 1 to 3)
- Write effective prompts (lessons 4 to 5)
- Configure and manage them properly (lessons 6 to 7)
- Optimize for cost (lesson 8)
- Run locally when it makes sense (lesson 9)
- Evaluate systematically (lesson 10)
- Combine models intelligently (lesson 11)
With this framework, every new model release is just another data point to evaluate. Not a reason to panic or rebuild everything.
Key Takeaways
- New models release constantly. You don't need to chase every one.
- Monthly check-in (15 min) for new releases. Quarterly evaluation (1 to 2 hours) to re-test your choices.
- Only switch when a new model is clearly better, cheaper, or has a capability you need.
- Ignore benchmark hype. Your own evaluation test set is what matters.
- Future-proof by abstracting model calls, keeping test sets, and using OpenAI-compatible APIs.
- The meta-skill is knowing how to evaluate, not knowing which model is best today.