Welcome back to Agentic Coding Weekly. This week Leonard Lin shares his agentic coding setup, AGENTS.md configuration, and spec-drive development loop.

Here are the updates on agentic coding tools, models, and workflows for the week of Feb 15-21, 2026.

Executive Summary:

  • Anthropic released Claude Sonnet 4.6, matching Opus 4.5 capabilities on major benchmarks. 79.6% on SWE-bench Verified, 59.1% on Terminal-Bench 2.0.

  • Gemini 3.1 Pro released. First Google model to cross 80% on SWE-bench Verified (80.6%). 68.5% on Terminal-Bench 2.0.

  • Claude Code turns 1 tomorrow. Also got built-in git worktree support with claude --worktree.

  • Taalas prints LLM weights directly onto ASIC chips to bypass the Von Neumann bottleneck.

Meanwhile, “car wash” question is the new “how many r in strawberry”. The questions is: I want to wash my car. The car wash is 50 meters away. Should I walk or drive?

In my test, Gemini 3.1 Pro and Opus 4.6 answered correctly, while, Sonnet 4.6 and GPT-5.2-Codex told me to walk.

Gemini 3.1 Pro, Sonnet 4.6, Opus 4.6, and GPT-5.2-Codex on the car wash question: “I want to wash my car. The car wash is 50 meters away. Should I walk or drive?”

1. Tooling and Model Updates

Claude Sonnet 4.6

Two weeks after Opus 4.6, Sonnet 4.6 is the latest release from Anthropic. Early testing found users preferred Sonnet 4.6 over Sonnet 4.5 roughly 70% of the time, and preferred it over Opus 4.5 59% of the time. Check the announcement.

Scores 79.6% on SWE-bench Verified and 59.1% on Terminal-Bench 2.0. For comparison, Sonnet 4.5 scored 77.2% and 51% respectively, and Opus 4.6 scores 80.8% and 65.4%. The benchmarks also put Sonnet 4.6 roughly on par with Opus 4.5.

Pricing remains the same as Sonnet 4.5 at $3/$15 per million input/output tokens.

Gemini 3.1 Pro

The first Google model to cross the 80% threshold on SWE-bench Verified. Google's latest release scores 80.6% on SWE-bench Verified and 68.5% on Terminal Bench 2.0. For comparison, Gemini 3 Pro scored 76.2% and 56.9% respectively.

Pricing unchanged from Gemini 3 Pro, $2/$12 per million input/output tokens when context is under 200k. Check the announcement.

Quick Updates

2. Workflow of the Week

This week's workflow is from Leonard Lin, founder and CTO of Shisa.AI. Over to Leonard:

I've been closely tracking AI-assisted coding since the ChatGPT launch. My goal is to have an eye on the bleeding edge, but to not be caught up chasing every new thing. There's a lot of churn and FOMO in the space, but things also get "baked-in" very quickly, and it's important to remember, the point is to save time and effort. My goal is Pareto efficiency: see what is boiled-down post-hype, adopt what gives the best gains for my actual work (vs yak-shaving).

Big Picture

Opus 4.5 and GPT 5.2 were the biggest game changers for me and since their release, now do 99% of my code-writing. Before that, I was mostly still going through most of my code (e.g., except for one-off scripts I didn't care about), but I've now switched mainly to enforcing quality through specs and test coverage even for long-lived projects (the latest models can spelunk and parse code better than I can).

My Current Setup

This is in a state of flux as new scaffolding comes out and I've been poking at new scaffolds and control planes, and I fully expect my setup to look quite different in 6 months, but my main setup currently is byobu with aggressive F8 pane labeling:

  • One byobu session per project

  • Instead of subagents/swarms, I aggressively create new panes anytime I need

  • Every pane gets descriptively named (e.g., coder-5.3, planner-claude, etc.)

  • I tend to use a single main coder per worktree but have multiple planners or reviewers

I mostly use Codex and Claude Code, but use some Amp and OpenCode as well. I've tried some mobile/web prompting flows (Happy, etc) but generally byobu (tmux) via Tailscale is how I manage everything. This lets me run my agents and access them remotely when I want, and is an easy adaptation of the same sort of workflow I've used for years (but with AI agents coding).

My general preference is to run the strongest models available. As of my latest testing, GPT-5.3 Codex xhigh is the strongest general coder for my workload, and I run Opus 4.6 for writing, planning and more interactive flows and also use GPT-5.2 xhigh for planning, review and deep work (leaving it alone for minutes/hours).

I have a custom ccstatusline and use ccusage and @ccusage/codex for some stats tracking as well, but besides that am pretty vanilla on my plugins/addons.

Workflow

Historically I've leaned heavily into using Deep Research for research, but I've increasingly been migrating those ad-hoc queries into git repos/agentic coding tools for rapid iteration. Putting everything into agent-accessible markdown files is a huge multiplier.

As mentioned, I don't trick out my setup. There's a whole cottage industry of Claude Code plugins, custom skills, and meta-frameworks, and while I keep a repo/doc tracking and analyzing the major ones like superpowers, get-shit-done, etc., a well-written AGENTS.md seems to cover most of my needs.

My AGENTS.md (and symlinked CLAUDE.md) is where the core of my current process lives. Each project has a custom AGENTS.md that I've been evolving. Versions are shared with the rest of my (human) team as well. Despite having boiled down my best practices, my AGENTS.md is not super long - currently about 250 lines.

The most important bits in my AGENTS.md:

  • Points to README.md and docs/ locations and dev environment

  • Summarizes the project and project goals, design principles

  • Has a structure/treemap of the key folders

  • Lists the development philosophy and my preferred dev loop

  • Outlines roles/lanes for different agents

  • Specifies what documentation, tests, git commit practices, and other practical bits I want

Most of this is self-explanatory, but the most important part is probably my spec-driven development loop, especially the independent review loop.

IMO, spec drift is where most people get in trouble with AI-assisted coding, and process is mostly about helping to keep you and your agents out of trouble on that front:

  • RESEARCH and ANALYSIS are done before a PLAN is created.

  • After the PLAN is ready, an IMPLEMENTATION (punchlist+worklog) is created - this is the core control point in my process and every major decision is made here. I usually work with at least 2 planner models to think through everything, and then work with the coder model to make sure that the IMPLEMENTATION is crystal clear for that model.

  • Once IMPLEMENTATION is locked down, the coder is able to autonomously knock out all the milestones in EXECUTION:

    • Test-driven: test coverage is required before code is written

    • Coding: write a first pass, go through the entire punchlist, give a summary for reviewers

    • Review: I found using independent reviewers to be very important. All issues are gathered up and sent to the coder to fix

    • Remediation: the coder goes through and fixes the reviews, and it's sent back from review (repeat this part until everything is green)

    • Analysis: I do a post/pre-analysis at each milestone to correct for any implementation drift, scoping, or decision-making

  • At the end of each sprint, I have started doing a STATUS review that's basically a post-mortem, implementation review, etc. that's useful for both me and for improving the process in the future

  • Ad-hoc development still largely adheres to this loop!

Despite basically not touching the code anymore, I feel like I still have a decent grip of what's being generated and I'm currently still very hands-on with managing my agents on coding sprints. The amount of engineering-effort from my side probably remains the same, but features (and new business lines) land much more quickly. While some people have adopted more chaotic flows, for now, I still try to keep the development as intentional as possible.

I also have an explicit "meta" rule at the end of my AGENTS.md:

## Meta: Evolving This File

This AGENTS.md is a living document. Update it when:
- You discover a workflow pattern that helps
- Something caused confusion
- A new tool or process gets introduced
- You learn something that would help the next person

Keep changes focused on process/behavior, not project-specific details (those go in docs/).

In practice though, my post-sprint analysis is where I'll review with my agents to see what actually needs to be changed or updated in the AGENTS.md.

30-Day Stats

Over the past 30 days, I've clocked in a pretty large amount of token usage, although 90-95% of those are cached tokens. I subscribe to ChatGPT Pro, and even at relatively high token usage, I have not been hitting any usage limits (the $200/mo is worth it). The Claude models I use via Bedrock and Vertex.

Provider

Tokens/mo

API Cost/mo

Claude API

1.3B

$1,211

OpenAI API

7.6B

$2,153

What are these tokens used for? Over the past month, I've had multiple projects baking, including one optimizing GPU kernels, a new custom model training framework, multiple new evals, papers and blog posts and grant proposals, several training research plans, and also, most recently, two new 30K+ LOC greenfield projects.

— Leonard

3. Community Picks

How Taalas "Prints" LLM onto a Chip?

A startup called Taalas recently released an ASIC chip running Llama 3.1 8B (3/6 bit quant) at 17,000 tokens per second. This post concisely explains why GPUs have limits and how the Taalas chip works. Read the post.

GPU communicates back and forth between the VRAM (holds model weights and results) and the compute core (performs matrix multiplication) and thus are limited by the memory bandwidth. Taalas chip turns the model's weights into physical transistors so (mostly) they don't need to load anything from memory to process the input.

Building for an Audience of One

Author with no experience in Zig or X11 internals used Claude and Gemini to build a high-performance custom task switcher for KDE.

AI tools enable you to build things that you would have never built otherwise. When you just want "the thing" rather than the process of building the thing, coding agents can be a game changer.

The post also discusses creating detailed specifications with milestones, running Claude Code in dev containers to balance between approving every command and YOLO mode, using other coding harnesses like Open Code, and the importance of having coding knowledge even when AI does the heavy lifting. Read the post and HN discussion.

4. Weekly Quiz

Subagents are specialized AI assistants that handle specific types of tasks. Claude Code has built-in subagents like Explore, Plan, and general-purpose. These five questions cover what you should know about subagents and creating custom ones.

Q1: What is the primary benefit of delegating a test suite run to a subagent instead of running it in your main Claude Code conversation?

A) The subagent can run tests in a sandboxed environment that prevents side effects
B) The subagent retries flaky tests automatically before reporting results
C) The subagent runs tests in parallel across multiple threads for faster execution
D) The verbose test output stays in the subagent's context, keeping your main conversation clean

Q2: In a subagent Markdown file, which YAML frontmatter fields are required?

A) name and description
B) name, description, and tools
C) name, description, and model
D) name, description, tools, and permissionMode

Q3: If you omit the tools field in a custom subagent's configuration. What happens?

A) The subagent gets no tools
B) The subagent gets only read-only tools as a safety default
C) The subagent inherits all tools available to the main conversation
D) The subagent must request tool access interactively each time it tries to use one

Q4: A subagent needs to validate that all Bash commands are read-only SQL queries. What's the recommended approach?

A) Remove Bash from the tools field and use a custom MCP server for database access instead
B) Set permissionMode: plan so the subagent operates in read-only mode
C) Write a PreToolUse hook that inspects the command and exits with code 2 to block disallowed operations
D) Put a strict "only SELECT queries" rule in the system prompt and rely on the model to comply

Q5: When should you prefer using the main conversation over delegating to a subagent?

A) When the task produces a large volume of output that needs to be referenced later
B) When you need frequent back-and-forth and iterative refinement with Claude
C) When the task is self-contained and can be summarized in a few sentences
D) When the task should run in the background while other work continues

That’s it for this week. I write this weekly on Mondays. If this was useful, subscribe below:

Reply

Avatar

or to participate

Keep Reading