featured

AI coding agents are not glorified StackOverflow

It’s a common pushback I hear from the usual AI deniers.

Oh Tari why are you hyping this thing up so much lol, it’s no different from copying and pasting from StackOverflow for goodness sake. Just calm down bro.

But how could you possible equate those two.

It’s like saying hiring a chef to cook a meal is no different from searching for recipe and doing the cooking yourself.

Just because the same knowledge is used in both cases, or what?

It makes no sense.

AI agents are not “improved Google” or “improved StackOverflow”.

Can StackOverflow build entire features from scratch spanning multiple files in your project?

Can StackOverflow do this?

Can Google fix errors in your code with something as simple as “what’s wrong with this code”?

Do they have any deep contextual access to your code to know what you could possibly mean with an instruction as vague as that?

All StackOverflow and Google can do is give you fragments of generic information — you have to do the reasoning yourself.

It‘s up to you to specialize and integrate the information into your project.

And then you still have to do testing and debugging as always.

AI agents are here to do all of these much faster with far greater accuracy.

That’s like the whole point of AI — of automation.

Massive amounts of work done in tiny fractions of effort of the manual alternative.

Faster. Easier.

Predictability. Insane personalization. Smart recommendations.

These are the things that make AI so deadly. It doesn’t matter if they use the same knowledge sources that a human would use.

It’s what would make a hypothetical AI cooking machine a danger to chef careers.

It doesn’t matter if the coding or cooking information and answers are already out there on the Internet.

Even when it comes to accessing knowledge, chatbots are still obviously better at giving it to you in a straightforward cohesive manner, after researching and synthesizing the information from different sources.

Copy and pasting from StackOverflow and Google cannot give you any of these benefits.

This new Cursor 2 feature revealed something so many programmers take for granted

The recent upgrades to the Cursor 2 IDE were awesome.

The new multi-agent feature is something really special.

Several powerful agents working together to achieve the highest quality result possible.

Each of these agents have their strong and weak points.

But now you combine every single one of them — to get the best of all worlds

You pick and choose the very best results — even combine them — unite them.

It’s the unity of logic — which is actually one of the hallmarks of software development

Programmers aren’t just “coders”, they’re builders.

Physical builders unite physical materials. They take wood, metal, concrete, glass… so many raw materials that do little on their own — only to put them together into a magnificent structure with far greater value than the sum of its parts.

In the same way software builders unite logical structures.

And it’s incredible how we take for granted something so amazing.

You take logic, assets, data — and put them all together into an incredible software system with the power of your mind.

Every single part of the system playing an important role — no matter how little.

You delete one statement in line 1623 and the entire system falls apart.

Every single line, function, variable.

All the files, libraries, databases.

All orchestrated towards a common goal by one brilliant mind.

Or two brilliant minds, three, five — even at the level of the human mind we have a powerful form of unity in the form of teamwork and collaboration.

Now with this new multi-agent feature we get to see the real-time unity of AI coding agents in action.

And it only serves to remind us of main reason these coding models matter in the first place.

What matters is that the model you use gets the job done and contributes effectively to the overall system.

Why keep obsessing so much over if you’re using the best model and obsessing over specific stats and benchmarks?

These things eventually come with diminishing returns.

If you even have to put them all to the same task simultaneously then so be it.

Let everyone play their role in achieving the ultimate goal of bringing a software system to life.

This new IDE from Google is a VS Code killer

Wow this is incredible.

Google has really gotten dead serious about dev tooling — their new Firebase Studio is going to be absolutely insane for the future of software development.

A brand new IDE packed with incredible and free AI coding features to build full-stack apps faster than ever before.

Look at how it was intelligently prototyping my AI app with lightning speed — simply stunning.

AI is literally everywhere in Firebase Studio — right from the very start of even creating your project.

  • Lightning-fast cloud-based IDE
  • Genius agentic AI
  • Dangerous Firebase integration and instant deployment…

And it looks like they went with light theme this time.

Before even opening any project Gemini is there to instantly scaffold whatever you have in mind.

Firebase Studio uses Gemini 2.5 Flash and Pro — the thinking model that’s been at the top of the AI benchmarks for several months now.

For free.

And you can choose among their most recent models — but only Gemini (sorry).

Although looks like there could be a workaround with the Custom model ID stuff.

For project creation there’s still dozens of templates to choose from — including no template at all.

Everything runs on the cloud in Firebase Studio.

No more wasting time setting up anything locally — build and preview and deploy right from your IDE.

Open up a project and loading happens instantly.

Because all the processing is no longer happening in a weak everyday PC — but now in a massively powerful data center with unbelievable speeds.

You can instantly preview every change in a live environment — Android emulators load instantly.

You’ll automatically get a link for every preview to make it easy to test and share your work before publishing.

The dangerous Firebase integration will be one of the biggest selling points of Firebase.

All the free, juicy, powerful Firebase services they’ve had for years — now here comes a home-grown IDE to tie them together in such a deadly way.

  • Authentication for managing users
  • Firestore for real-time databases
  • Cloud Storage for handling file uploads
  • Cloud Functions for server-side logic All of these are available directly from the Studio interface.

And that’s why deployment is literally one click away once you’re happy with your app.

Built-in Firebase Hosting integration to push your apps live to production or preview environments effortlessly.

Who is Firebase Studio great for?

  • Solo developers who want to quickly build and launch products
  • Teams prototyping new ideas
  • Hackathon participants
  • Educators teaching fullstack development
  • Anyone who wants a low-friction, high-speed way to build real-world apps

It especially shines for developers who already love Firebase but want a more integrated coding and deployment flow.

You can start using Firebase Studio by visiting firebase.studio. You’ll need a Google account. Once inside, you can create new projects, connect to existing Firebase apps, and start coding immediately. No downloads, no complex setup.

So this is definitely something to consider — you might start seeing local coding as old-school.

But whether you’re building your next startup or just hacking together a side project, Firebase Studio is an fast integrated way to bring your app to life.

Cursor 2.0 is incredible — 5 amazing new features

Wow this is huge.

Cursor just dropped several massive upgrades to their IDE to make AI coding even more powerful.

Cursor 2.0 comes with a whole new philosophy for developing with AI…

1. Revolutionary new multi-agent feature

Now you don’t just have a single agent making changes from your prompts…

Now you can run several AI agents in parallel to each try a different approach to the same coding task — working in separate sandboxes to avoid overwriting each other’s work.

When they’re done you get a combined diff view where you can see and merge the best results.

The day of frantically switching between models to see which one can do best are over.

Instantly unleash the entire army of state-of-the-art models on any task.

2. Dedicated new Agent View

This leads to having the new dedicated Agent View mode to show you what each agent is “seeing” and trying to do.

You get a live, transparent panel of their suggestions, reasoning, and file edits before anything lands in your codebase

Approve, merge, or discard with a click — turning AI from a mysterious black box into a controllable teammate you can supervise in real time.

And you still have this view when you use only one agent.

3. Composer?

And now we have a new in-house model too — built from the ground up to code.

They’re calling it Composer — not to be confused with the agent itself, that has been in the IDE since late 2024.

This new Composer model is designed to be:

  • Fast — up to four times quicker than comparable models.
  • Context-aware — it understands big projects and multiple files.
  • Flexible — you can mix and match models: use a “heavy thinker” for planning, then let Composer quickly build and refactor.

This move makes Cursor less dependent on external APIs and gives them tighter control over performance and privacy.

So now we see a growing trend of these IDEs creating their own in-house models — the other day I was talking about how Windsurf also just doubled down on their in-house SWE coding model with major new speed upgrades in version 1.5.

I can definitely see why this is happening.

I mean you already have millions of active users — why keep burning money on models you have no control over — when you have the resources to make one — that could be much cheaper in the long run?

And will give you far more control and predictability.

And not just that — they also have the advantage of being able to train these in-house models on real data coming from their IDEs — and I don’t mean just code.

Models like GPT-5 and Gemini can only train on static code — but in-house IDE models can train on the entire real-time flow of actions that happen throughout development.

Which makes they way better for features like autocomplete and Tab to Jump.

They see how devs jump within and across the different types files and the changes they make in different scenarios — something that regular models can never give you.

4. Native browser & DOM Inspector

Cursor now ships with a native browser and DOM inspector wired directly into the IDE — so agents can see what’s actually happening in your app instead of hallucinating around it.

They can click through flows, inspect live elements, read network responses, and propose fixes or UI tweaks against the real DOM — all without leaving your editor or juggling external windows.

5. Voice input for devs

Personally I’ll always prefer typing.

But you may prefer using your melodious voice instead of typing in mountains of paragraphs.

Voice input lets you explain problems the way you would to a coworker at the whiteboard:

“Scan this service and repo, find where we handle Stripe webhooks, and suggest a safer retry flow.”

Cursor turns that into precise actions and edits, so your thinking speed isn’t capped by your typing speed.

Together these make Cursor a more practical tool for real-world development — not just a playground for AI code generation.

Cursor 2.0 isn’t trying to replace developers — it’s trying to multiply them. By letting AI agents safely explore, propose, and test changes, it takes us closer to a world where coding feels less like typing and more like designing systems together.

GitHub just made AI coding agents way more powerful

GitHub just made coding with AI agents way more powerful with their new Agent HQ feature.

Now with Agent HQ you have a single place where you can pick which agents to use, give them tasks, and track what they’re doing, all inside GitHub.

Look how we use easily assign a task to the agent — a whole new feature — it will work on this autonomously and in the background using the agent we selected.

When it’s done it will present the changes with a pull request:

Instead of just having one assistant you will be able to connect many different agents — from GitHub, OpenAI, Anthropic, Google, xAI, Cognition, and others — and manage them all from one dashboard.

You can literally now have an army of all the most powerful agents in the world working together on various parts of your projects — at the same time.

Agent HQ unleashed in VS Code — powered up with 2 MCP servers:

It’s built right into GitHub, VS Code, the CLI, and even the mobile app — so you can stay in your normal workflow while the agents do the heavy lifting.

Until now most developers had to jump between tools or experiment with different agents separately. Agent HQ fixes that. It gives you:

  • One place to manage all your agents.
  • Clear visibility into who’s using what and how well it’s working.
  • Built-in governance tools for enterprise teams who need security and compliance.
  • And, most importantly, choice — you’re not locked into one vendor.

There’s a lot you can do

Here’s what stands out about Agent HQ:

  • Run multiple agents at once. Want to compare how different AI models handle the same coding problem? You can run them side-by-side and see which performs better.
  • Use it across your tools. Whether you’re coding in VS Code, checking GitHub on the web, or managing from the CLI, Agent HQ connects it all.
  • Stay in control. Admins can monitor usage, set permissions, and enforce policies — great for larger teams.
  • Measure productivity. GitHub’s adding dashboards that show how agents impact speed and output.

It fits in seamlessly

Agent HQ is designed to blend seamlessly into your workflow. For example:

  • In VS Code, agents can now plan out multi-step tasks — like fixing bugs, adding features, or running tests — and then summarize results for your review.
  • Each agent works within a sandbox so your main repo stays safe until you approve changes.
  • Team leads get a central view of which agents are active and what they’re working on.

vs Copilot

Copilot is like a smart co-pilot sitting next to you, helping with code as you type and make agentic changes to your code base.
But Agent HQ is like an operations center for entire fleets of AI agents — coordinating, comparing, and managing them across projects.

Copilot assists you.

Agent HQ organizes and manages all your AI assistants.

With Agent HQ, AI becomes not just a helper — but an active part of your dev team.

Windsurf just released the fastest coding model ever

Wow Windsurf’s new SWE-1.5 model is shockingly good and ridiculously fast.

SWE-1.5 is giving every developer a perfect sweet spot between lightning speed and high accuracy.

It’s even faster than Claude Haiku 4.5 — the latest Claude model that was specially released for quick coding.

Yet almost as intelligent as Claude Sonnet 4.5 — yet 13 times faster.

You’re barely waiting for any thinking here — near instantaneous responses.

And this isn’t just another “slightly better” model.

SWE-1.5 is frontier-scale — it’s massive under the hood — like hundreds of billions of parameters — optimized for real coding work.

Windsurf built it from the ground up for software engineers who want an AI partner that can read, write, refactor, debug, and explain code across entire projects at rapid pace.

Because every delay adds up when you’re using an assistant that plans, tests, and fixes code in multiple steps.

I mean of course it’s still much faster than manual coding — even if you use GPT-5 High that some people were complaining about the excessive thinking.

But SWE-1.5’s speed means those loop — generating, testing, adjusting — happen almost instantly.

You can keep your focus instead of staring at a loading indicator. Pair-programming with AI feels more natural and conversational.

What it’s built for

Windsurf designed the SWE family (which now includes SWE-1.5) for the full development lifecycle. That means:

  • Refactoring large repos — it can understand and safely update multiple files at once.
  • Debugging — it can read logs, spot issues, and suggest fixes without losing context.
  • Scaffolding new projects — it builds structure fast so you can focus on the creative parts.
  • Running tools and tests — the model isn’t just “talking” about code, it’s working with it.

It’s also deeply integrated into Windsurf’s IDE and agent system, which lets it actually do things — not just describe them.

Developers testing SWE-1.5 say it feels like working with a real teammate. You can throw complex refactors at it and it responds in seconds instead of minutes. The result is more experimentation, faster debugging, and fewer context switches.

These are the things that elevate the development experience significantly.

Try it now

SWE-1.5 is already live in the Windsurf IDE. If you’re on a Pro plan, you can use it right now; there’s also a free tier if you want to take it for a spin.

SWE-1.5 isn’t just about a bigger model number. It’s about speed that changes how you code. If you’ve ever wished your AI assistant could keep up with your thought process, this might be the first one that truly does.

“6 years work in 6 days”: The amazing Claude Code effect

Claude Code has completely transformed the software development ecosystem.

People are building faster than they ever thought possible with this incredible tool.

This is wild — this is the Claude Code effect:

This is the difference that these incredible AI tools bring to your workflow.

Like they said it’s not perfect, but there’s no doubt of the massive time and effort savings they and many other devs like me have been getting from tools like Claude Code.

You are completely reclaiming so much of the time you once used to manually create files and type every single character.

It’s amazing just how much faster and enjoyable development has been for me ever since these agentic tools come onto the scene.

And we’re certainly not just talking coding — what of all those chores that seem like they won’t take much time — but slowly eat into your day — your life?

Claude Code has a stunning feature that let’s you put all these admin tasks into a queue — where they get run all at the same time.

Look at this — we can send prompts to Claude Code — without having to wait for the results:

You can make several changes to various parts of the codebase — all at once:

You literally have your very own team of developers — building much faster than any one human ever could.

Oh, and they talked about committing code? Let’s talk about committing code.

Committing your changes to Git is so effortless now.

One challenge many developers have is writing high-quality Git commit messages.

But now with tools like Claude Code — you can easily ask an AI agent to make commits on your behalf.

The Claude Code CLI has access to all your CLI commands — it will analyze all the changes you made and come with a precise commit message.

IDEs like Windsurf and VS Code even have a feature that let’s you do this directly from their UI.

It’s nice and I do use them from time to time.

With agentic tools like Claude Code you can get even fancier:

You can ask the agent to generate a series of commits based on all the unrelated changes you may have made.

You know one great advice on Git committing is to always make atomic commits — commit on a single change.

But sometimes in the real-life coding it’s easy to make changes all over the place — and then who has the time or energy to go through all those changes one by one to separate them and commit them individually?

It’s always easier to just commit everything — maybe with some horribly vague commit message like “Update” or “Misc”.

But now with Claude Code this problem disappears instantly.

It just intelligently analyzes all the changes and automatically divides them into logically coherent parts — and then it makes commits each change with a high-quality message.

It’s an incredible tool that every developer needs to start taking advantage of.

They can’t believe what Gemini 3.0 will be able to do

Everyone has been going crazy over the upcoming Gemini 3.0 family of models from Google.

Especially Gemini 3.0 Pro — this one is going to be absolutely unbelievable — from what internal testers have been saying.

Vibe coding video games from scratch?

And if it’s already doing this internally, it would be interesting to just see how advanced and high-quality those games are.

As always certain people are going to be skeptical of claims like this.

Some have said it’s just hype or the games will be nothing more than Snake and Pacman in terms of complexity.

But they’re forgetting that… current AI models including Gemini can already create these sort of games.

So even “video games” is a bit vague, don’t you think Gemini 3.0 is going to be able to do much more than that.

Even with current AI tech we’ve already had massive success stories — like Pieter Levels who vibe coded a flight simulator in 3 hours — and went on to make over $1 million in a year.

People are already making playable 3d games that generate serious revenue — something like Pacman or Tetris is ALREADY a walk in park for every major model.

People have been making Tetris games with a single prompt with Claude and the rest for months now.

And why do some people think that the only thing Gemini will be able to contribute to the game development is the coding?

Haven’t we already seen the wonders that video generators like Veo and Sora can do?

Is it that hard to imagine progressing from non-interactive video generation to interactive video? AKA, games?

We already have the insane Genie 3.0 from Google — a revolutionary model that generates 3d virtual worlds from scratch — and let you interact with them:

With multimodality and cooperation with Genie, Gemini 3.0 will combine video creation with code generation with audio generation with graphics creation with…

And we’re already getting news that this upcoming model is going to be a significant upgrade in multimodality support.

Even right now people can already create a lot of the static graphics for their game using only these AI models.

We’ve already had major efforts like Blender MCP — that use AI to automate creating high-quality models for 3d models.

So we are clearly already on this trajectory — even Google doesn’t get it right exactly with Gemini 3.0, someone is going to soon enough.

People are still underrating just how quickly these models can progress — even after seeing the unprecedented strides from Midjourney and many others.

We are moving towards fully automated high-quality video game creation.

Google just turned their AI Studio into a full-blown IDE

Google just released several amazing upgrades to Google AI Studio — their powerful tool for building AI apps.

It’s getting easier than ever to create state-of-the-art AI-powered tools from nothing.

Now Google AI Studio has an incredible new vibe coding ability — and not just your regular IDE vibe coding — this has a flavor of something different entirely.

First you start off with the raw materials — all the incredible AI services from Google — there’s so much to choose from for all the different kind of AI apps you could build.

  • Generate images on the fly with Nano Banana…
  • Animate images with Veo…
  • Generate videos on the fly…

And so much more.

Many of these which were unimaginable just 5 years ago.

Then you still get top-notch services that you can only get from Google — Google Maps data, Google Search data, etc.

Now you can actually combine multiple of these amazing tools to start creating your app.

You get to compose all these incredible services together into something innovative and potentially world-changing.

And what if you don’t have any idea yet?

The I’m feeling lucky button is here for you — click and get a fresh idea with a brand new composition of tools.

Ah, the classic I’m Feeling Lucky button from Google once again.

A staple feature in Google Search to jump directly to the first webpage without showing the results.

So now it’s here in AI Studio to spring up new ideas for you to quickly get to building.

So look at what it generated the first time — an interactive map-based chatbot using real-time Google Maps — super impressive.

It intelligently picked the Use Google Search data for real-time data — and of course the AI powered chatbot for chatting features.

We try it again — and get another brilliant idea.

An app giving you personalized organization and decluttering suggestions based on the unique appearance of your room — pretty neat.

And you know you don’t have to just take whatever it gives you.

You can use it as a powerful brainstorming tool — all the ideas it keeps giving you can lead you up a path towards something incredible.

Now here we settle on this dream garden prompt and ask Gemini to generate the app.

And all it took was 77 seconds:

The styling isn’t perfect and it looks quite basic — but it’s a clear UI that works.

And now you can start your regular vibe coding — ask for anything and see it unfold in real time.

AI Studio also got a slew of nice UI upgrades to make it more convenient to use.

For example they’ve made it much easier to switch between models in the Chat page.

You can switch between Gemini, GenMedia, text-to-speech (TTS), and live models — all from a single page.

This app is rapidly becoming a comprehensive incubator for the latest and greatest AI apps, powered by state-of-the-art Google services.

OpenAI’s new IDE extension is absolutely incredible

OpenAI’s new Codex extension for VS Code is shockingly good.

Those devs who ignore this are going to miss out on so much.

You might even end up completely switching from your current coding agent when you see the incredible significance of this.

It’s been a long long time coming.

OpenAI finally gave us a home-built extension to unleash the true power of GPT-5 into all the major IDEs — yes this goes beyond VS Code.

The best part is it comes with GPT-5 Codex — a model built from the ground to code code code…

Codex + Codex = ?

Answer: the stuff of dreams

And do you think anyone understands how GPT-5 breathes and thinks more than OpenAI?

When you use the Codex extension you are going to have GPT-5 at its very best.

Some of you try a model and you don’t get the exact results you want — then you immediately think the model must be terrible and you will never use it again.

Like this person on Reddit when GPT-5 first came out:

But the truth is certain models are going to perform better in certain IDEs.

They don’t all use the same system prompts and model settings — even in the same IDE for different models.

So you see, GPT-5 is going to be highly optimized in a first-party extension like Codex.

It has everything you’d expect from a coding agent — and even beyond that.

Select code or open files and ask Codex to refactor or document or fix a bug.

Ask it to explain unfamiliar code. Make massive changes across files easily.

Another major selling point — it works hand-in-hand with the same Codex in the CLI.

And not only the CLI but also the Codex web app

Another really interesting feature is the Agent (Full access) mode.

Most agents out there always ask for your approval before editing files or running commands.

But when you turn on Full access you are activating a totally different beast.

With Full access now the agent can do anything without any permission from you.

It will run any command — including commands from outside your codebase directory — and commands that make real network requests.

It just flies on autopilot. Total autonomy.

You don’t have to lift a finger to approve anything anymore.

But of course it comes with risks — which probably isn’t worth cutting out this occasional finger-lifting.

Better to stick with the normal Agent mode — which can still run important offline commands automatically — NPM package installations for example.

This new Codex IDE extension is a serious challenger to everything else out there and could be just what you need.

You may feel Claude or Gemini is better with your current agent.

But what happens when you use GPT-5 in its ultimate form?