Discussion about this post

User's avatar
Pawel Jozefiak's avatar

The private codebase specialization angle is smart. Most coding agents are trained on public repos and fail hard when they hit internal codebases with domain-specific patterns. SERA's "soft-verified" training approach is interesting—accepting partial fixes instead of requiring full verification should make it way easier to bootstrap on custom code. But I'm skeptical about the 8B/32B parameter sizes competing with Claude/GPT level models on complex architectural decisions. Small models can handle tactical code completion, but strategic refactoring and system design still need the big guns. The real test: can SERA maintain context across multi-file changes in a 50k+ line codebase? That's where most coding agents break down, regardless of parameter count.

No posts

Ready for more?