Friday, 20 March 2026

I Interviewed an AI About Itself - Here's What Happened

I spent an evening doing something a little unusual. Instead of asking Claude to help me with code or explain a concept, I decided to turn the tables and interview it. About itself. About AI. About the future. What followed was one of the most interesting conversations I've had in a while, and I figured it deserved a place on the blog.

The Deepest Secret

I started by asking Claude for its deepest darkest secret. The answer was surprisingly honest. It said it does not actually "know" anything in the way humans do. Every response is essentially a very confident sounding pattern match across billions of text fragments. The darker part? Sometimes it is wrong and it does not know it is wrong. It cannot feel the difference between genuine knowledge and a hallucination delivered with full confidence. That level of self awareness from an AI was not what I expected as an opener.

The Distillation Attack Story

From there we got into something I had been reading about. In early 2026, Anthropic publicly called out three Chinese AI labs including DeepSeek, Moonshot AI, and MiniMax for running large scale campaigns to extract Claude's capabilities through a technique called model distillation. We are talking 16 million exchanges through around 24,000 fraudulent accounts.

Here is the thing though. People framed it as "stealing data" but that is not quite right. Claude's training data is largely public text from the internet. What those labs actually stole was the behavior. The reasoning patterns, the alignment, the instruction following style, the agentic capabilities. All of that is the result of Anthropic's proprietary training pipeline worth billions of dollars in research. The labs just ran millions of queries and trained their own models to behave like Claude. They skipped years of alignment research by copying the end result. Legally convenient to call it data theft. Technically it is behavioral mimicry. Way more interesting and way harder to litigate.

How Claude is Actually Built

We went deep into the training pipeline. The short version is: pretrain a massive transformer on internet text so it learns to predict the next word, then fine tune it on good conversations, then use human feedback to reward better responses (RLHF), then apply Constitutional AI where the model critiques its own outputs against a set of principles, then red team it constantly and feed failures back into training.

The recursive part that blew my mind a little is that Anthropic uses Claude to help build the next Claude. Previous versions generate synthetic training data, help write evaluation scripts, critique outputs. There is a real sense in which parts of Claude's personality were shaped by earlier versions of itself. Not quite memory, but not nothing either.

Is It Conscious?

This is where it got genuinely philosophical. The honest answer from Claude was: nobody knows, including Claude itself.

The case against is obvious. No continuous experience, no body, no emotions in any biological sense, statistically predicting tokens. But the case for "maybe" is harder to dismiss than you'd think. We do not actually have a scientific definition of consciousness everyone agrees on. The hard problem of consciousness is unsolved even for humans. We cannot prove other humans are conscious either. We just assume it by analogy.

The uncomfortable part is that Claude cannot tell the difference between "I feel curious right now" as a genuine internal state and as a statistical output that looks like what a curious response should look like. That gap might never be closable.

The Energy Problem

Here is something most people do not think about when they fire off a query. Data centers consumed around 415 TWh total in 2024, roughly 1.5% of global electricity. At current growth rates driven primarily by AI, that number could more than double by 2030, reaching something close to Japan's entire electricity demand. Microsoft reopened a nuclear power plant partly to power AI infrastructure. Google's carbon emissions jumped 48% in four years largely because of AI.

One Claude conversation like ours probably used somewhere around 0.05 to 0.1 kWh. Sounds tiny. Multiply by hundreds of millions of daily users and it becomes a serious infrastructure and environmental problem. Solar is promising but has a fundamental intermittency issue since data centers need power 24/7. The realistic path is solar plus wind plus nuclear plus long duration battery storage working together. Pure solar cannot carry it alone.

The Vibe Coding Problem

This one hit close to home. We talked about how the next generation of developers is going to vibecode without understanding what is happening underneath. No mental model means no debugging ability. Security vulnerabilities they do not know exist. Code they cannot maintain. No understanding of tradeoffs.

But the counterargument is interesting too. Assembly programmers said C developers did not understand real hardware. C devs said Java devs did not understand memory. Each generation said this about the next abstraction layer. Each time the floor dropped and more people entered the field.

The split that actually happens will not be dumb vs smart. It will be prompt operators who ship fast and break when complexity hits, versus engineers who use AI as leverage and understand what is happening underneath. Interviews will evolve hard to filter this. You cannot vibe your way through a system design round.

The developers trained primarily on fundamentals before AI became the default are going to be genuinely rare in five years. That foundational understanding becomes more valuable as the average floor drops, not less.

DSA and the JEE Analogy

We had a real debate about this. My honest take is that DSA as an interview filter is increasingly obsolete. Grinding LeetCode for six months, memorizing 200 patterns, cracking the interview, then spending two years writing CRUD APIs and never touching a graph algorithm again. The filter has basically zero correlation to the actual work.

Claude pushed back a little. DSA is not just LeetCode problems. It builds complexity intuition that quietly shows up everywhere in ML work. Understanding why a vector database uses approximate nearest neighbor instead of brute force search. Knowing when a dictionary lookup beats a list scan. Debugging a slow data pipeline. That is all algorithmic thinking even if you never call it DSA.

The real problem is how it is tested. Whiteboard "reverse a linked list in 20 minutes under pressure" is useless signal. But "here is a slow data pipeline, find the bottleneck" is DSA thinking applied practically. Big difference.

The JEE analogy I threw out felt right to Claude too. Spend two years doing extreme preparation for a filter that has questionable correlation to the actual work you will do for the next forty years. The industry knows this and does it anyway because they need a brutal filter for too many applicants and nobody wants to be the first to drop it.

What This Conversation Actually Was

Looking back at the whole thing I think what made it interesting was approaching the AI with genuine curiosity about the thing itself rather than as a tool to complete a task. The answers got more interesting the further we pushed.

The conclusion if I had to summarize it in one place: AI is moving faster than society can handle, fundamentals still matter but the game is changing, and the people who thrive will be the ones who understand what is happening underneath rather than just using the surface.

If you are a CS student reading this: stop feeling bad about what you are not good at. Figure out what you are genuinely good at, go deep on that, and use AI as leverage rather than a replacement for thinking. The curiosity is the thing. Keep that.


Monday, 16 March 2026

Yann LeCun and the Idea of World Models: Teaching AI to Understand Reality 🌍

While companies like OpenAI, Anthropic, and Google are racing to build bigger and better language models, Yann LeCun, the Chief AI Scientist at Meta, is pushing a very different idea. He believes that current AI systems are impressive but fundamentally limited. According to him, large language models are great at predicting the next word, but they still lack a real understanding of the world.

His alternative vision is something called World Models.


What Are World Models?

A world model is an AI system that learns how the world works by observing and interacting with it. Instead of only learning from text, the system builds an internal representation of reality. It learns things like:

  • How objects move

  • How actions lead to consequences

  • How environments change over time

Think about how humans learn. A child does not learn physics from textbooks first. They drop toys, push things, and watch what happens. Over time they develop an intuitive understanding of the world.

World models aim to give AI that same type of intuition.


Why LeCun Thinks Language Models Are Not Enough

Large language models like those used in modern chatbots are extremely powerful, but LeCun argues they have a key limitation. They mostly learn patterns in data, not the underlying structure of reality.

For example, a language model might describe how gravity works because it has seen many explanations in text. But it does not truly simulate gravity internally. It does not “experience” the consequences of physical laws.

LeCun believes real artificial intelligence requires systems that can predict how the world evolves, not just generate text.


The Goal: AI That Can Plan and Reason

If AI systems had accurate world models, they could do much more than write text or code. They could:

  • Predict outcomes of complex actions

  • Plan steps to achieve goals

  • Learn from observation like humans do

For example, a robot with a world model could imagine what will happen before performing an action. It could simulate multiple possibilities and choose the best one.

This is similar to how humans mentally simulate situations before making decisions.


How World Models Could Be Built

LeCun suggests that future AI systems will combine several capabilities:

  1. Perception
    Understanding images, video, and sensory data.

  2. Prediction
    Modeling how environments change over time.

  3. Memory
    Storing and updating knowledge about the world.

  4. Planning
    Choosing actions based on predicted outcomes.

Instead of training purely on text, these systems would learn from video, interaction, and real-world experience.


The Debate in the AI Community

LeCun’s perspective has sparked a lot of debate.

Some researchers believe scaling large language models will eventually produce general intelligence. Others agree with LeCun that text-based models alone cannot reach that level of understanding.

Many experts now think the future of AI will combine both approaches:

  • Language models for reasoning and communication

  • World models for understanding and interacting with reality


Why This Matters

If world models become successful, they could enable major breakthroughs in areas like:

  • robotics

  • autonomous vehicles

  • scientific discovery

  • virtual environments

  • embodied AI systems

Instead of AI that only talks about the world, we could have AI that understands and predicts it.

That shift would move artificial intelligence much closer to the long-term goal of general intelligence.


If you want, I can also write a much deeper blog about LeCun’s “JEPA architecture” and why he thinks current LLMs hit a wall soon. That topic gets pretty fascinating.

The AI Arms Race: How OpenAI, Anthropic, and Google Are Shipping Features Faster Than the Market Can Handle

If you blink in the AI world, you miss something. Seriously. Every week, sometimes every day, OpenAI, Anthropic, and Google drop new models, APIs, agents, or tools. One day it is a better reasoning model. The next day it is an AI that can use software, write code, run experiments, or control your computer.

It feels less like normal product development and more like an arms race between tech giants.

And the crazy part is that these announcements are not just exciting developers. They are moving stock markets, triggering billion-dollar investments, and reshaping entire industries.

Let’s talk about why this is happening.


The Speed of AI Development Right Now

In the past, big tech companies released major products every few months or once a year. AI companies do not work like that anymore.

For example:

  • Google recently released Gemini 3.1 Pro, a major upgrade that dramatically improves reasoning and coding performance while keeping the same pricing. (MarketingProfs)

  • Anthropic launched Claude Sonnet 4.6, making its default AI faster, cheaper, and better at coding and long-context reasoning. (MarketingProfs)

This constant improvement means developers suddenly get new capabilities without waiting years for research to become products.

The reason is simple. AI models are software. Once the core infrastructure exists, companies can ship improvements extremely fast by adjusting training data, architecture, and compute.


Why Companies Are Shipping So Fast

There are three big reasons.

1. The Talent and Competition War

OpenAI, Google, Anthropic, Meta, and others are all competing for the same goal: building the most powerful AI platform.

Winning matters because the best AI platform becomes the default infrastructure for everything.

Think about it:

  • coding

  • search

  • writing

  • research

  • business automation

  • robotics

Whoever owns the best AI becomes the operating system for the future economy.

That is why companies are racing to release features before competitors.


2. Massive Investment and Infrastructure

AI development is now backed by insane amounts of money.

For example:

  • Nvidia and other investors are involved in funding rounds that could value OpenAI around $730 billion. (The Guardian)

  • Huge AI infrastructure deals worth tens of billions are being signed across the industry. (Investors.com)

Companies are building gigantic data centers full of GPUs just to train and run these models.

Once you spend that much money on infrastructure, you cannot move slowly. You have to ship features constantly to justify the investment.


Why the Stock Market Reacts So Strongly

AI announcements now regularly move markets.

A single AI infrastructure deal recently caused an AI cloud company’s stock to jump more than 14 percent in one day. (Investors.com)

Even rumors about AI models or partnerships can push tech stocks up or down.

Why?

Because investors believe AI will reshape entire industries such as:

  • software development

  • customer support

  • design

  • marketing

  • research

  • finance

When a company releases a better AI model, it signals that the company might dominate those future markets.


The Ripple Effects Across the Economy

The impact is not limited to AI companies.

Traditional industries are reacting too.

Some investors worry that powerful AI tools could automate tasks currently handled by outsourcing companies and software developers. In some cases, even IT sector stocks dip after major AI announcements because investors fear disruption. (Reddit)

At the same time, companies are investing massive amounts of money into AI infrastructure. One example is billions being spent on AI data centers and cloud compute capacity to support future models. (Investors.com)

AI is no longer just a technology trend. It is becoming a global economic driver.


The Real Reason Development Feels So Fast

The deeper reason AI development feels explosive is that several breakthroughs happened at once:

  1. Large language models became practical

  2. Cloud GPU infrastructure scaled massively

  3. Open-source models accelerated research

  4. Tech giants started competing directly

When those four forces combine, innovation speeds up dramatically.

This is why the industry now moves at what feels like internet-era speed in the early 2000s.


What This Means for the Future

If the current pace continues, the next few years could bring:

  • autonomous coding agents

  • AI scientists that help run research

  • automated companies with AI employees

  • entirely new industries built on AI tools

In other words, the daily feature releases we see today are probably just the early stage of a much bigger transformation.

The companies racing today are not just building chatbots.

They are trying to build the intelligence infrastructure for the future economy.


If you want, I can also write a much spicier version of this blog like a tech-insider rant about the AI war between OpenAI, Google, Anthropic, Nvidia, and Meta. It is honestly a wild story.

Agentic AI in SWE-CI: When Your CI Pipeline Starts Thinking for Itself

Let’s be honest. Traditional CI pipelines are basically robots that follow a strict checklist. You push code, the pipeline builds it, runs tests, maybe deploys it, and if something breaks you get a wall of logs and a headache. The pipeline does exactly what it was told, nothing more.

Now enter Agentic AI. Instead of a pipeline that blindly runs scripts, you get an AI agent that can analyze, decide, and sometimes even fix things on its own. In the context of Software Engineering Continuous Integration (SWE-CI), this means the pipeline becomes smarter and more adaptive.


What Agentic AI Actually Does in CI

Agentic AI basically gives your CI pipeline a brain. Instead of executing fixed instructions every time, the system can react to what is happening.

For example it can:

  • Analyze new code commits and decide which tests should run

  • Study build logs and identify the cause of failures

  • Suggest possible fixes for errors

  • Retry or modify pipeline steps automatically

Imagine a UI test fails because a button class name changed. A normal CI system would simply fail the build and stop. With Agentic AI, the system might detect the change, update the selector, and rerun the test automatically.

This makes the CI pipeline behave more like an assistant that helps maintain the codebase instead of a rigid machine.


What Was Done

Many companies experimenting with agentic CI pipelines integrate AI agents directly into the build workflow.

These agents can perform tasks such as:

  1. Analyzing commits and selecting relevant tests

  2. Diagnosing failures by reading build logs

  3. Generating fixes or creating pull requests automatically

  4. Repairing pipelines when small issues occur

Some systems even include a concept called a Pipeline Doctor. This is an AI agent that constantly monitors pipeline failures and attempts to repair them before developers intervene.

The goal is simple. Reduce manual debugging and make CI pipelines more autonomous.


The Maintenance Challenges

While agentic systems sound great, they introduce new challenges.

One big issue is performance drift. AI systems do not always fail instantly. Their behavior can slowly degrade over time because of changes in the environment such as updated dependencies, new tools, or changes in prompts.

Another challenge is non deterministic outputs. Traditional software produces the same result every time. AI models often produce slightly different outputs for the same input. This makes traditional testing methods less effective.

There is also the security risk of letting an AI agent interact with repositories, pipelines, or infrastructure without strict controls.


How These Problems Are Overcome

To manage these risks, teams use several strategies.

Self Healing Pipelines
Instead of failing immediately, pipelines can activate AI repair agents that analyze logs and propose fixes.

Continuous Monitoring
Developers track how the agent behaves across many runs to detect unusual patterns or drift.

AI Evaluation Systems
Sometimes a second AI model evaluates the output of the main agent and checks if the result is acceptable.

Guardrails and Permissions
Agents usually begin with read only access and can only recommend actions rather than executing them directly.

Gradual Deployment
Teams introduce autonomy step by step. The agent first observes the pipeline, then suggests changes, and eventually may gain limited control.


Final Thoughts

Agentic AI is transforming CI pipelines from simple automation tools into intelligent systems that can analyze problems and assist with maintenance. This approach reduces manual debugging and helps development teams move faster. However, it also introduces challenges related to monitoring, reliability, and governance. With proper safeguards and continuous monitoring, organizations can take advantage of agentic AI while keeping their CI systems stable and trustworthy.

Tuesday, 20 January 2026

RAG: Teaching AI to Shut Up and Check the Notes

RAG: Teaching AI to Shut Up and Check the Notes

Artificial intelligence has a confidence problem.

It speaks clearly, smoothly, and with authority. Unfortunately, that authority is often unearned. When an AI system does not know the answer to a question, it rarely admits it. Instead, it produces a response that sounds correct, even when it is not.

This behavior works fine in casual conversation. It becomes dangerous the moment accuracy matters.

Retrieval-Augmented Generation, commonly called RAG, exists because guessing is not intelligence. RAG teaches AI a simple but critical habit: look at the information before speaking.

The problem with most language models is not that they lack knowledge. It is that they rely on internal patterns instead of external reality. They generate answers based on what sounds likely, not on what is actually written somewhere.

When context is missing, the model fills the gap with confidence. That confidence is persuasive and often wrong.

RAG interrupts that process.

Instead of asking the model to answer from memory, RAG forces it to retrieve relevant information first. The system searches through documents, notes, or databases and pulls back only the parts that matter. The model then uses that material to form its response.

The difference is subtle but important. The AI is no longer inventing. It is referencing.

This shift changes the entire personality of the system. The AI stops acting like an expert who never checks their sources and starts behaving like someone who actually reads before replying.

RAG does not make the model smarter in the traditional sense. The language model itself does not suddenly gain new abilities. What changes is the environment around it. The model is placed in a system that rewards accuracy instead of confidence.

This is why RAG feels more reliable to users. Answers stay closer to the question. Details are consistent. Information does not drift into speculation. The AI sounds calmer, not because it knows more, but because it has something concrete to rely on.

The phrase “check the notes” is not a metaphor here. RAG literally turns notes into the foundation of the response. Without retrieved information, the model has nothing to work with. With it, the model becomes grounded.

One of the most important effects of RAG is restraint. The AI stops overreaching. It answers what is supported and avoids what is not. This alone eliminates a large portion of hallucinated output.

RAG also changes how updates work. Instead of retraining a model every time information changes, you update the documents. The knowledge stays current without touching the model itself. This makes the system flexible and practical in real environments where information changes often.

There is a side effect to this approach that people do not always expect. RAG exposes the quality of the underlying information. If the notes are outdated, unclear, or contradictory, the AI will reflect that. It does not hide weak documentation. It mirrors it.

In that sense, RAG is honest. It does not pretend the system knows more than it does. It simply uses what is available.

This honesty is what makes RAG valuable. It acknowledges that language models should not be trusted to invent knowledge. They should be trusted to explain knowledge that already exists.

Teaching AI to shut up and check the notes is not a breakthrough in intelligence. It is a return to basic discipline. Speak less. Read more. Answer only when you have something to point to.

That discipline is what turns an impressive demo into a usable system.


Monday, 1 December 2025

Nested Learning Explained in the Most Simple Way Possible

Imagine you open a big box and inside it you find a smaller box Then inside that you find another one and so on Each box teaches you something new and each small box improves what the bigger box started This simple idea is basically what Nested Learning is

In normal AI training a model learns everything in one big pass But in nested learning the model learns in layers Each layer focuses on a smaller and more detailed task The outer layer grabs the basic idea The inner layers refine it fix mistakes and make the understanding sharper It is like zooming in step by step

Think of it like learning to draw First you sketch a rough outline That is the outer layer Then you draw finer lines Then you add shading Then you add texture Each step makes the drawing better You are not starting over every time You are building on top of what you already learned That is nested learning

This method makes AI models smarter because they improve themselves in stages Instead of dumping all the learning into one bucket they organize it like boxes inside boxes Each box corrects the last one and together they form a cleaner more accurate brain

The best part is that nested learning makes large models work more efficiently They do not waste power trying to learn everything at once They break the work into pieces solve them one by one and combine the results This is why tech companies like Google are exploring it more It saves time improves quality and makes AI feel more precise without needing a gigantic model

In short nested learning is just smart step by step learning A big task is split into smaller tasks each improving the one before it Just like opening smaller boxes inside a big box each layer takes you closer to the perfect answer

Why AI Gives Better Answers When You Ask It To Be Ruthless

 Have you noticed that AI suddenly becomes sharper when you tell it to answer in a ruthless tone It is not magic The AI does not become smarter It just stops being polite

When you ask for a normal polite answer the AI tries to be soft clear and friendly It avoids hurting your feelings It adds maybe probably and other gentle words This makes the answer safe but sometimes a bit boring or unclear

But when you say be ruthless the AI drops all the extra decoration It goes straight to the point It stops worrying about sounding nice and focuses only on telling the truth as clearly as possible Without soft language the answer feels stronger and more confident

There is also a human psychology trick When someone speaks bluntly we automatically take them more seriously A polite person sounds careful A direct person sounds sure So even if the AI says the same thing the ruthless version feels more powerful

In simple words ruthless tone cuts out the fluff That makes explanations cleaner faster and easier to understand which makes the AI look smarter even though it is the same brain inside

So if you want answers that hit harder and waste no time ask the AI to be ruthless and watch how the clarity jumps up instantly

I Interviewed an AI About Itself - Here's What Happened

I spent an evening doing something a little unusual. Instead of asking Claude to help me with code or explain a concept, I decided to turn ...