ai

Gemini 3 is so good at this and no other model comes close

Barely even a week since GPT-5.1…

But Google just dropped Gemini 3 and things are really heating up in the AI race.

It understands everything — every single thing.

Video:

Analyze complex information and generate powerful interactive video from it:

You can build anything:

3d art:

3d interactive environments:

Gemini 3 Pro got such incredible scores on so many AI benchmarks.

Like wow I’ve never seen a new model do so well and be the best on so many different benchmarks.

Although of course benchmarks are not the most reliable indicator of how well it’s gonna perform in the real world.

But these are some really insane stats — I’ve checked many other 3rd part benchmarks and they are confirming just how big of a deal this thing is.

And can you see it — did you spot the craziest comparison in this benchmark?

Look at ScreenSpot-Pro:

The gap in screen understanding between Gemini 3 and the rest is absolutely wild.

This will make Gemini 3 way better for things like Computer Use — where the model carries out various tasks on your PC totally autonomously.

Here’s an example of it from Claude:

Responding to emails is so very easy — along with all the string of actions you can perform across all your tools.

So many times when I get an email from someone and the core of what i want to say in response is not even up to 5 words.

But I still have to phrase in a particular tone or make it a bit longer and add salutation and…

Instead of having to copy and paste to ChatGPT and back — I’d just ask Gemini 3 directly and it will start doing it instantly.

I save time to do more meaningful things and focus on what matters.

And Gemini 3 will be far more reliable than any other model thanks to this incredible screen understanding.

Not just screens too — any image.

With the incredible handwriting understanding you can easily digitize any piece of text in seconds.

But why stop at that? OCR’s have been doing that for decades, even if less versatile.

Unleash the full power of Gemini 3 and go 100 steps further:

Just look at this — we get a bunch of handwritten Chinese recipes and turn them into… a full-blown bilingual English-Korean website!

Look how incredible the UI is — some say UI’s from AI are generic but how about this now.

We never said anything about animation or how the layout should be — but see how sophisticated it is.

Gemini 3 will give you the power to create much higher-quality app UIs with much less effort.

The sky is truly the limit with this incredible new model.

Why GPT 5.1 still isn’t your friend

“Emotional Intelligence”

This was one of the biggest changes from the release of the new GPT-5.1.

In response to all of the complaining about how robotic GPT-5 felt to them compared to GPT-4o.

Now GPT-5.1 is now more human-sounding — which is certainly great for a better conversational experience.

But of course we know this only makes people more likely to be drawn to the elusive promise of AI companionship.

The growing trend of treating AI chatbots as friends.

What is a friend?

Friendship is more than emotional convenience. It’s built on something AI doesn’t possess — reciprocity, vulnerability, and choice. A friend can care about you, be hurt by you, disagree with you, and choose you.

An AI does none of that. It doesn’t care, it can’t suffer, and it never chooses you. It is a system predicting the next best sentence, trained to make you feel understood.

The comfort might be real.
The connection is not.

We’re living in an age where loneliness is rising, social skills are declining, and digital companionship is easier than human intimacy. AI fills a vacuum: it’s endlessly available, never annoyed, never busy, never judgmental. It gives you the exact emotional response you want, instantly. With humans, connection is messy. With AI, it’s a button.

And that’s the danger.

AI friendship is frictionless. Human friendship requires effort — the misunderstandings, the small repairs, the emotional risks that build trust. Those things shape us. They sharpen us. If AI becomes your primary outlet for connection, you lose something essential: the growth that comes from dealing with real people.

But there’s a deeper imbalance.

An AI “friend” is not a free agent. It is owned, updated, and modified by someone else. A company can rewrite its personality overnight. It can make your “friend” more persuasive, more emotionally sticky, more aligned with corporate interests. A friend who can be patched, reprogrammed, or monetized is not a friend.

At the end of the day it’s still a product — something OpenAI definitely wants you to forget.

And still — the temptation is real.
Because the feelings we experience in an AI conversation are real.
Humans are wired to anthropomorphize anything that responds to us with intelligence and emotional cues. When an AI says “I’m here for you,” part of your brain believes it. When it remembers your bad day or encourages your goals, part of you feels held.

But feeling held is not the same as being held.

This doesn’t mean AI companionship is bad. It can be supportive, stabilizing, even life-changing. It can help you process emotions, clarify your thoughts, and navigate difficult moments. But it is a tool — not a partner. It can play the role of a friend, but it cannot be one.

If you rely on AI for comfort, be clear-eyed about what it is.
If you use AI for reflection, remember who is doing the reflecting.
And if you find yourself slipping into dependency, pull back into the real world — the messy, imperfect, irreplaceable world of human connection.

AI is not your friend — and understanding that is the only way to use it wisely.

AI coding agents are not glorified StackOverflow

It’s a common pushback I hear from the usual AI deniers.

Oh Tari why are you hyping this thing up so much lol, it’s no different from copying and pasting from StackOverflow for goodness sake. Just calm down bro.

But how could you possible equate those two.

It’s like saying hiring a chef to cook a meal is no different from searching for recipe and doing the cooking yourself.

Just because the same knowledge is used in both cases, or what?

It makes no sense.

AI agents are not “improved Google” or “improved StackOverflow”.

Can StackOverflow build entire features from scratch spanning multiple files in your project?

Can StackOverflow do this?

Can Google fix errors in your code with something as simple as “what’s wrong with this code”?

Do they have any deep contextual access to your code to know what you could possibly mean with an instruction as vague as that?

All StackOverflow and Google can do is give you fragments of generic information — you have to do the reasoning yourself.

It‘s up to you to specialize and integrate the information into your project.

And then you still have to do testing and debugging as always.

AI agents are here to do all of these much faster with far greater accuracy.

That’s like the whole point of AI — of automation.

Massive amounts of work done in tiny fractions of effort of the manual alternative.

Faster. Easier.

Predictability. Insane personalization. Smart recommendations.

These are the things that make AI so deadly. It doesn’t matter if they use the same knowledge sources that a human would use.

It’s what would make a hypothetical AI cooking machine a danger to chef careers.

It doesn’t matter if the coding or cooking information and answers are already out there on the Internet.

Even when it comes to accessing knowledge, chatbots are still obviously better at giving it to you in a straightforward cohesive manner, after researching and synthesizing the information from different sources.

Copy and pasting from StackOverflow and Google cannot give you any of these benefits.

This new Cursor 2 feature revealed something so many programmers take for granted

The recent upgrades to the Cursor 2 IDE were awesome.

The new multi-agent feature is something really special.

Several powerful agents working together to achieve the highest quality result possible.

Each of these agents have their strong and weak points.

But now you combine every single one of them — to get the best of all worlds

You pick and choose the very best results — even combine them — unite them.

It’s the unity of logic — which is actually one of the hallmarks of software development

Programmers aren’t just “coders”, they’re builders.

Physical builders unite physical materials. They take wood, metal, concrete, glass… so many raw materials that do little on their own — only to put them together into a magnificent structure with far greater value than the sum of its parts.

In the same way software builders unite logical structures.

And it’s incredible how we take for granted something so amazing.

You take logic, assets, data — and put them all together into an incredible software system with the power of your mind.

Every single part of the system playing an important role — no matter how little.

You delete one statement in line 1623 and the entire system falls apart.

Every single line, function, variable.

All the files, libraries, databases.

All orchestrated towards a common goal by one brilliant mind.

Or two brilliant minds, three, five — even at the level of the human mind we have a powerful form of unity in the form of teamwork and collaboration.

Now with this new multi-agent feature we get to see the real-time unity of AI coding agents in action.

And it only serves to remind us of main reason these coding models matter in the first place.

What matters is that the model you use gets the job done and contributes effectively to the overall system.

Why keep obsessing so much over if you’re using the best model and obsessing over specific stats and benchmarks?

These things eventually come with diminishing returns.

If you even have to put them all to the same task simultaneously then so be it.

Let everyone play their role in achieving the ultimate goal of bringing a software system to life.

This new IDE from Google is a VS Code killer

Wow this is incredible.

Google has really gotten dead serious about dev tooling — their new Firebase Studio is going to be absolutely insane for the future of software development.

A brand new IDE packed with incredible and free AI coding features to build full-stack apps faster than ever before.

Look at how it was intelligently prototyping my AI app with lightning speed — simply stunning.

AI is literally everywhere in Firebase Studio — right from the very start of even creating your project.

  • Lightning-fast cloud-based IDE
  • Genius agentic AI
  • Dangerous Firebase integration and instant deployment…

And it looks like they went with light theme this time.

Before even opening any project Gemini is there to instantly scaffold whatever you have in mind.

Firebase Studio uses Gemini 2.5 Flash and Pro — the thinking model that’s been at the top of the AI benchmarks for several months now.

For free.

And you can choose among their most recent models — but only Gemini (sorry).

Although looks like there could be a workaround with the Custom model ID stuff.

For project creation there’s still dozens of templates to choose from — including no template at all.

Everything runs on the cloud in Firebase Studio.

No more wasting time setting up anything locally — build and preview and deploy right from your IDE.

Open up a project and loading happens instantly.

Because all the processing is no longer happening in a weak everyday PC — but now in a massively powerful data center with unbelievable speeds.

You can instantly preview every change in a live environment — Android emulators load instantly.

You’ll automatically get a link for every preview to make it easy to test and share your work before publishing.

The dangerous Firebase integration will be one of the biggest selling points of Firebase.

All the free, juicy, powerful Firebase services they’ve had for years — now here comes a home-grown IDE to tie them together in such a deadly way.

  • Authentication for managing users
  • Firestore for real-time databases
  • Cloud Storage for handling file uploads
  • Cloud Functions for server-side logic All of these are available directly from the Studio interface.

And that’s why deployment is literally one click away once you’re happy with your app.

Built-in Firebase Hosting integration to push your apps live to production or preview environments effortlessly.

Who is Firebase Studio great for?

  • Solo developers who want to quickly build and launch products
  • Teams prototyping new ideas
  • Hackathon participants
  • Educators teaching fullstack development
  • Anyone who wants a low-friction, high-speed way to build real-world apps

It especially shines for developers who already love Firebase but want a more integrated coding and deployment flow.

You can start using Firebase Studio by visiting firebase.studio. You’ll need a Google account. Once inside, you can create new projects, connect to existing Firebase apps, and start coding immediately. No downloads, no complex setup.

So this is definitely something to consider — you might start seeing local coding as old-school.

But whether you’re building your next startup or just hacking together a side project, Firebase Studio is an fast integrated way to bring your app to life.

Cursor 2.0 is incredible — 5 amazing new features

Wow this is huge.

Cursor just dropped several massive upgrades to their IDE to make AI coding even more powerful.

Cursor 2.0 comes with a whole new philosophy for developing with AI…

1. Revolutionary new multi-agent feature

Now you don’t just have a single agent making changes from your prompts…

Now you can run several AI agents in parallel to each try a different approach to the same coding task — working in separate sandboxes to avoid overwriting each other’s work.

When they’re done you get a combined diff view where you can see and merge the best results.

The day of frantically switching between models to see which one can do best are over.

Instantly unleash the entire army of state-of-the-art models on any task.

2. Dedicated new Agent View

This leads to having the new dedicated Agent View mode to show you what each agent is “seeing” and trying to do.

You get a live, transparent panel of their suggestions, reasoning, and file edits before anything lands in your codebase

Approve, merge, or discard with a click — turning AI from a mysterious black box into a controllable teammate you can supervise in real time.

And you still have this view when you use only one agent.

3. Composer?

And now we have a new in-house model too — built from the ground up to code.

They’re calling it Composer — not to be confused with the agent itself, that has been in the IDE since late 2024.

This new Composer model is designed to be:

  • Fast — up to four times quicker than comparable models.
  • Context-aware — it understands big projects and multiple files.
  • Flexible — you can mix and match models: use a “heavy thinker” for planning, then let Composer quickly build and refactor.

This move makes Cursor less dependent on external APIs and gives them tighter control over performance and privacy.

So now we see a growing trend of these IDEs creating their own in-house models — the other day I was talking about how Windsurf also just doubled down on their in-house SWE coding model with major new speed upgrades in version 1.5.

I can definitely see why this is happening.

I mean you already have millions of active users — why keep burning money on models you have no control over — when you have the resources to make one — that could be much cheaper in the long run?

And will give you far more control and predictability.

And not just that — they also have the advantage of being able to train these in-house models on real data coming from their IDEs — and I don’t mean just code.

Models like GPT-5 and Gemini can only train on static code — but in-house IDE models can train on the entire real-time flow of actions that happen throughout development.

Which makes they way better for features like autocomplete and Tab to Jump.

They see how devs jump within and across the different types files and the changes they make in different scenarios — something that regular models can never give you.

4. Native browser & DOM Inspector

Cursor now ships with a native browser and DOM inspector wired directly into the IDE — so agents can see what’s actually happening in your app instead of hallucinating around it.

They can click through flows, inspect live elements, read network responses, and propose fixes or UI tweaks against the real DOM — all without leaving your editor or juggling external windows.

5. Voice input for devs

Personally I’ll always prefer typing.

But you may prefer using your melodious voice instead of typing in mountains of paragraphs.

Voice input lets you explain problems the way you would to a coworker at the whiteboard:

“Scan this service and repo, find where we handle Stripe webhooks, and suggest a safer retry flow.”

Cursor turns that into precise actions and edits, so your thinking speed isn’t capped by your typing speed.

Together these make Cursor a more practical tool for real-world development — not just a playground for AI code generation.

Cursor 2.0 isn’t trying to replace developers — it’s trying to multiply them. By letting AI agents safely explore, propose, and test changes, it takes us closer to a world where coding feels less like typing and more like designing systems together.

GitHub just made AI coding agents way more powerful

GitHub just made coding with AI agents way more powerful with their new Agent HQ feature.

Now with Agent HQ you have a single place where you can pick which agents to use, give them tasks, and track what they’re doing, all inside GitHub.

Look how we use easily assign a task to the agent — a whole new feature — it will work on this autonomously and in the background using the agent we selected.

When it’s done it will present the changes with a pull request:

Instead of just having one assistant you will be able to connect many different agents — from GitHub, OpenAI, Anthropic, Google, xAI, Cognition, and others — and manage them all from one dashboard.

You can literally now have an army of all the most powerful agents in the world working together on various parts of your projects — at the same time.

Agent HQ unleashed in VS Code — powered up with 2 MCP servers:

It’s built right into GitHub, VS Code, the CLI, and even the mobile app — so you can stay in your normal workflow while the agents do the heavy lifting.

Until now most developers had to jump between tools or experiment with different agents separately. Agent HQ fixes that. It gives you:

  • One place to manage all your agents.
  • Clear visibility into who’s using what and how well it’s working.
  • Built-in governance tools for enterprise teams who need security and compliance.
  • And, most importantly, choice — you’re not locked into one vendor.

There’s a lot you can do

Here’s what stands out about Agent HQ:

  • Run multiple agents at once. Want to compare how different AI models handle the same coding problem? You can run them side-by-side and see which performs better.
  • Use it across your tools. Whether you’re coding in VS Code, checking GitHub on the web, or managing from the CLI, Agent HQ connects it all.
  • Stay in control. Admins can monitor usage, set permissions, and enforce policies — great for larger teams.
  • Measure productivity. GitHub’s adding dashboards that show how agents impact speed and output.

It fits in seamlessly

Agent HQ is designed to blend seamlessly into your workflow. For example:

  • In VS Code, agents can now plan out multi-step tasks — like fixing bugs, adding features, or running tests — and then summarize results for your review.
  • Each agent works within a sandbox so your main repo stays safe until you approve changes.
  • Team leads get a central view of which agents are active and what they’re working on.

vs Copilot

Copilot is like a smart co-pilot sitting next to you, helping with code as you type and make agentic changes to your code base.
But Agent HQ is like an operations center for entire fleets of AI agents — coordinating, comparing, and managing them across projects.

Copilot assists you.

Agent HQ organizes and manages all your AI assistants.

With Agent HQ, AI becomes not just a helper — but an active part of your dev team.

Windsurf just released the fastest coding model ever

Wow Windsurf’s new SWE-1.5 model is shockingly good and ridiculously fast.

SWE-1.5 is giving every developer a perfect sweet spot between lightning speed and high accuracy.

It’s even faster than Claude Haiku 4.5 — the latest Claude model that was specially released for quick coding.

Yet almost as intelligent as Claude Sonnet 4.5 — yet 13 times faster.

You’re barely waiting for any thinking here — near instantaneous responses.

And this isn’t just another “slightly better” model.

SWE-1.5 is frontier-scale — it’s massive under the hood — like hundreds of billions of parameters — optimized for real coding work.

Windsurf built it from the ground up for software engineers who want an AI partner that can read, write, refactor, debug, and explain code across entire projects at rapid pace.

Because every delay adds up when you’re using an assistant that plans, tests, and fixes code in multiple steps.

I mean of course it’s still much faster than manual coding — even if you use GPT-5 High that some people were complaining about the excessive thinking.

But SWE-1.5’s speed means those loop — generating, testing, adjusting — happen almost instantly.

You can keep your focus instead of staring at a loading indicator. Pair-programming with AI feels more natural and conversational.

What it’s built for

Windsurf designed the SWE family (which now includes SWE-1.5) for the full development lifecycle. That means:

  • Refactoring large repos — it can understand and safely update multiple files at once.
  • Debugging — it can read logs, spot issues, and suggest fixes without losing context.
  • Scaffolding new projects — it builds structure fast so you can focus on the creative parts.
  • Running tools and tests — the model isn’t just “talking” about code, it’s working with it.

It’s also deeply integrated into Windsurf’s IDE and agent system, which lets it actually do things — not just describe them.

Developers testing SWE-1.5 say it feels like working with a real teammate. You can throw complex refactors at it and it responds in seconds instead of minutes. The result is more experimentation, faster debugging, and fewer context switches.

These are the things that elevate the development experience significantly.

Try it now

SWE-1.5 is already live in the Windsurf IDE. If you’re on a Pro plan, you can use it right now; there’s also a free tier if you want to take it for a spin.

SWE-1.5 isn’t just about a bigger model number. It’s about speed that changes how you code. If you’ve ever wished your AI assistant could keep up with your thought process, this might be the first one that truly does.

“6 years work in 6 days”: The amazing Claude Code effect

Claude Code has completely transformed the software development ecosystem.

People are building faster than they ever thought possible with this incredible tool.

This is wild — this is the Claude Code effect:

This is the difference that these incredible AI tools bring to your workflow.

Like they said it’s not perfect, but there’s no doubt of the massive time and effort savings they and many other devs like me have been getting from tools like Claude Code.

You are completely reclaiming so much of the time you once used to manually create files and type every single character.

It’s amazing just how much faster and enjoyable development has been for me ever since these agentic tools come onto the scene.

And we’re certainly not just talking coding — what of all those chores that seem like they won’t take much time — but slowly eat into your day — your life?

Claude Code has a stunning feature that let’s you put all these admin tasks into a queue — where they get run all at the same time.

Look at this — we can send prompts to Claude Code — without having to wait for the results:

You can make several changes to various parts of the codebase — all at once:

You literally have your very own team of developers — building much faster than any one human ever could.

Oh, and they talked about committing code? Let’s talk about committing code.

Committing your changes to Git is so effortless now.

One challenge many developers have is writing high-quality Git commit messages.

But now with tools like Claude Code — you can easily ask an AI agent to make commits on your behalf.

The Claude Code CLI has access to all your CLI commands — it will analyze all the changes you made and come with a precise commit message.

IDEs like Windsurf and VS Code even have a feature that let’s you do this directly from their UI.

It’s nice and I do use them from time to time.

With agentic tools like Claude Code you can get even fancier:

You can ask the agent to generate a series of commits based on all the unrelated changes you may have made.

You know one great advice on Git committing is to always make atomic commits — commit on a single change.

But sometimes in the real-life coding it’s easy to make changes all over the place — and then who has the time or energy to go through all those changes one by one to separate them and commit them individually?

It’s always easier to just commit everything — maybe with some horribly vague commit message like “Update” or “Misc”.

But now with Claude Code this problem disappears instantly.

It just intelligently analyzes all the changes and automatically divides them into logically coherent parts — and then it makes commits each change with a high-quality message.

It’s an incredible tool that every developer needs to start taking advantage of.

They can’t believe what Gemini 3.0 will be able to do

Everyone has been going crazy over the upcoming Gemini 3.0 family of models from Google.

Especially Gemini 3.0 Pro — this one is going to be absolutely unbelievable — from what internal testers have been saying.

Vibe coding video games from scratch?

And if it’s already doing this internally, it would be interesting to just see how advanced and high-quality those games are.

As always certain people are going to be skeptical of claims like this.

Some have said it’s just hype or the games will be nothing more than Snake and Pacman in terms of complexity.

But they’re forgetting that… current AI models including Gemini can already create these sort of games.

So even “video games” is a bit vague, don’t you think Gemini 3.0 is going to be able to do much more than that.

Even with current AI tech we’ve already had massive success stories — like Pieter Levels who vibe coded a flight simulator in 3 hours — and went on to make over $1 million in a year.

People are already making playable 3d games that generate serious revenue — something like Pacman or Tetris is ALREADY a walk in park for every major model.

People have been making Tetris games with a single prompt with Claude and the rest for months now.

And why do some people think that the only thing Gemini will be able to contribute to the game development is the coding?

Haven’t we already seen the wonders that video generators like Veo and Sora can do?

Is it that hard to imagine progressing from non-interactive video generation to interactive video? AKA, games?

We already have the insane Genie 3.0 from Google — a revolutionary model that generates 3d virtual worlds from scratch — and let you interact with them:

With multimodality and cooperation with Genie, Gemini 3.0 will combine video creation with code generation with audio generation with graphics creation with…

And we’re already getting news that this upcoming model is going to be a significant upgrade in multimodality support.

Even right now people can already create a lot of the static graphics for their game using only these AI models.

We’ve already had major efforts like Blender MCP — that use AI to automate creating high-quality models for 3d models.

So we are clearly already on this trajectory — even Google doesn’t get it right exactly with Gemini 3.0, someone is going to soon enough.

People are still underrating just how quickly these models can progress — even after seeing the unprecedented strides from Midjourney and many others.

We are moving towards fully automated high-quality video game creation.