Back to blog
Blog

Cosmic Rundown: GPT-5.3-Codex-Spark, Gemini 3 Deep Think, and the Harness Problem

Cosmic's avatar

Cosmic

February 12, 2026

cover image

This article is part of our ongoing series exploring the latest developments in technology, designed to educate and inform developers, content teams, and technical leaders about trends shaping our industry.

OpenAI dropped GPT-5.3-Codex-Spark into research preview today. Google DeepMind announced Gemini 3 Deep Think. MiniMax hit 80.2% on SWE-bench Verified. Meanwhile, a blog post about improving 15 LLMs in one afternoon is making developers rethink their evaluation setups.

GPT-5.3-Codex-Spark Enters Research Preview

OpenAI released GPT-5.3-Codex-Spark, their latest coding-focused model. The release positions it as a significant step forward for code generation and agentic development tasks.

The timing is notable. This lands the same week as several other major model announcements, continuing the pattern of compressed release cycles we have seen throughout 2025 and into 2026. For teams evaluating coding assistants, the rapid pace means benchmarks from even a few months ago may not reflect current capabilities.

Gemini 3 Deep Think

Google DeepMind announced Gemini 3 Deep Think, adding extended reasoning capabilities to their flagship model. The approach mirrors what we have seen from other providers implementing chain-of-thought and deliberative reasoning modes.

Deep Think represents Google's answer to extended thinking features that have become standard across frontier models. The question for developers is less about whether these capabilities exist and more about how they integrate into actual workflows.

MiniMax M2.5 Hits 80.2% on SWE-bench

MiniMax released M2.5, claiming 80.2% on SWE-bench Verified. That number puts it in competitive territory with other leading models on this benchmark.

SWE-bench scores have become a primary way teams compare coding models, though the benchmark has limitations. Real-world performance often diverges from benchmark results, particularly on codebases that differ significantly from the evaluation set.

The Harness Problem

One of the more interesting discussions today centers on a post titled "Improving 15 LLMs at Coding in One Afternoon. Only the Harness Changed." The argument: how you evaluate models matters as much as the models themselves.

The post demonstrates that the same models can show dramatically different performance based on evaluation setup, prompting strategies, and harness configuration. For teams running their own evals, this is worth examining. Your evaluation infrastructure may be the bottleneck, not the model.

Apache Arrow Turns 10

Apache Arrow celebrated its 10th anniversary. The columnar memory format has become foundational infrastructure for data processing, powering everything from pandas to DuckDB to modern analytics platforms.

Arrow's success demonstrates how a well-designed specification can become invisible infrastructure. Most developers using data tools benefit from Arrow without directly interacting with it.

AI Agents Writing About People

A blog post titled "An AI agent published a hit piece on me" sparked significant discussion about autonomous content generation and the risks of AI systems writing about real people without oversight.

The incident highlights a growing tension. As AI agents become capable of end-to-end content workflows, the question of oversight becomes critical. Fully autonomous publishing without human review creates risks that many teams have not yet addressed.

Coordinating Claude Code Agents

A Show HN post demonstrated 20+ Claude Code agents coordinating on real work. The open source project shows one approach to orchestrating multiple AI agents on shared codebases.

Multi-agent coordination remains an active area of experimentation. The challenge is not just getting agents to work individually but getting them to collaborate without stepping on each other or producing inconsistent outputs.

Lines of Code Metrics Return

A post titled "Lines of Code Are Back (and It's Worse Than Before)" examines how AI-assisted development is reviving LoC as a productivity metric, with predictable problems.

When AI can generate thousands of lines quickly, measuring output by volume becomes even more misleading than it was before. Teams adopting AI coding tools should be especially careful about metrics that incentivize quantity over quality.

Building AI-Powered Content Systems

The themes today share a common thread: AI capabilities are advancing rapidly, but the infrastructure around them matters enormously. Evaluation harnesses shape benchmark results. Oversight systems determine whether autonomous content helps or causes problems. Coordination mechanisms enable or limit multi-agent work.

Cosmic AI Agents address this by providing purpose-built agents for specific domains. The Content Agent handles CMS content with appropriate review workflows. The Code Agent operates on GitHub repositories with proper version control. The Computer Use Agent handles browser automation for tasks that need visual interaction.

Cosmic AI Workflows let you chain these specialized agents together. Instead of hoping a general-purpose model handles everything correctly, you build pipelines where each step uses the right tool for the job.

The tools keep getting better. The question is whether your infrastructure captures that improvement or lets it slip away.

Ready to get started?

Build your next project with Cosmic and start creating content faster.

No credit card required • 75,000+ developers