SARATH THARAYILHS.T.[W] WRITEUPSWWRITEUPS[P] PROJECTSPPROJECTS[A] ABOUTAABOUT
മ
/ SYSTEM

Building thoughtful software, writing notes, and shipping experiments across data, AI, and the web.

© 2026 Sarath Tharayil/IST --:--:--
++++

Why Different AI Models Excel at Different Tasks

Why Different AI Models Excel at Different Tasks

/ METADATA
DATE:2026.4.26
AUTHOR:SARATH THARAYIL
READING TIME:11 MIN READ
ENGAGEMENT:--
CATEGORIES:
AILLMsTechnology
NAVIGATE:← GO BACK
/ METADATA← BACK
DATE:2026.4.26
READ:11 MIN READ
ENGAGE:--
LIKE:
CATEGORIES:
AILLMsTechnology
/ ARTICLE

Most people who use AI tools regularly have developed an informal mental model: use Claude for writing, use Perplexity when you need current information, use ChatGPT for images, use DeepSeek for code. That intuition is mostly correct. What is interesting is that the intuition is backed by real architectural and training decisions, not just vibes.

These systems feel different because they were built differently. Some are individual large language models. One of them is not a model at all. Their designs reflect specific bets about what matters most, and those bets have consequences that show up in everyday use.


The shared foundation

Before getting into the differences, it is worth establishing what they all share. Every one of these systems is built on top of a decoder-only transformer language model, trained on massive text corpora to predict the next token in a sequence. The same basic architecture underlies all of them.

What a decoder-only transformer actually does

Given a sequence of tokens (words, subwords, characters), the model produces a probability distribution over the next token. Repeat that, token by token, and you get a response. Through scale and fine-tuning, "predict the next token" becomes surprisingly general-purpose.

The differences emerge from what happens after that shared foundation:

  • Which modalities the model processes (text only, or also images, audio, video)
  • What architectural innovations were applied (mixture-of-experts, grouped-query attention, rotary embeddings)
  • What alignment method was used (standard RLHF, Constitutional AI, reasoning-focused RL)
  • Whether the product is a single model or an orchestrated system with routing, retrieval, and tools

These four dimensions are where Claude, ChatGPT, Gemini, DeepSeek, and Perplexity actually diverge. And they diverge quite significantly.


Claude: long-context reasoning, trained to be careful

Claude is a family of large language models from Anthropic. Architecturally, it is a decoder-only transformer in the conventional sense. What makes it distinctive is primarily its alignment approach: Anthropic developed a method called Constitutional AI, where the model is trained to critique and revise its own outputs against a written "constitution" of principles, before or in addition to human feedback.

The effect of this is a model that is notably careful. It tends to flag its uncertainty, push back on problematic requests, and follow complex instruction sets without drifting. For long codebases, detailed documents, and multi-step reasoning chains, that carefulness is often exactly what you want.

Constitutional AI in practice

Instead of only relying on human raters to provide feedback on model outputs, Anthropic trains Claude to self-critique against a set of stated principles. The model generates a response, critiques it using those principles, revises it, and the improved outputs become part of training. This creates a model with internalized values rather than just surface-level compliance.

Claude is fundamentally a text model. It can accept images as input in recent versions, but image generation comes from pairing it with external image models, not from a native image-decoder built into the same network. That is an important distinction: Claude's architecture is not designed around multimodality the way GPT-4o and Gemini are.

Where this makes Claude strong: reading large codebases and long documents, tasks requiring careful adherence to constraints, multi-step reasoning, and contexts where conservative or safety-sensitive behavior is valuable.


ChatGPT (GPT-4o): one model that handles everything

GPT-4o, the model powering recent versions of ChatGPT, takes a different architectural bet. It is a unified multimodal transformer: a single neural network that processes and produces text, images, and audio through the same network, rather than stitching separate models together.

The practical consequence is that GPT-4o can natively understand images, generate or edit images, and handle real-time voice in and out, all through the same underlying architecture. Earlier multimodal systems glued separate specialist models together; GPT-4o integrates these modalities at the token level.

Old pipeline approach

Separate specialist models glued together

Text model + vision model + speech model, each trained independently and connected via adapters. Modalities don't natively inform each other.

GPT-4o approach

Interleaved multimodal tokens in one network

Text, image, and audio tokens flow through the same transformer. The model reasons across modalities jointly, which produces more coherent cross-modal outputs.

The alignment approach is standard RLHF with multimodal fine-tuning, aimed at producing a generalist assistant that handles the widest possible range of everyday and professional tasks. GPT-4o is not narrowly optimized for any single domain.

Where this makes ChatGPT strong: visual tasks (understanding screenshots, diagrams, photos), image generation, voice interaction, general-purpose writing and coding, and any task where you need text, vision, and audio to work together naturally.


Gemini: Google's bet on scale and multimodality

Gemini is Google's family of multimodal models, and it shares the same basic architectural philosophy as GPT-4o: a decoder-only transformer that ingests interleaved multimodal tokens from text, images, audio, video, and code. Where Gemini differs is in scale decisions and the specific architectural techniques used to get there.

Two things stand out about Gemini's design:

Context length. The Gemini 2.x line reportedly handles context windows up to millions of tokens. For comparison, most LLMs operate in the tens or hundreds of thousands. This is not a minor detail: it means Gemini can genuinely process book-length documents, long video transcripts, or enormous codebases as a single context, rather than chunking or summarizing.

Mixture-of-experts. Gemini uses sparse mixture-of-experts architecture, which means only a subset of the model's parameters are activated for any given token. This allows larger total parameter counts without proportionally larger compute costs at inference, which is important for serving a model at Google's scale.

Gemini Embedding 2 is worth knowing about

Google's Gemini Embedding 2 maps text, images, video, audio, and documents into a single unified embedding space. This makes cross-modal search and retrieval architecturally natural, which is why Gemini integrates especially well into workflows involving many different content types.

Gemini also has deep integration with Google's product ecosystem, which matters for practical use: it can reason over your Gmail, Drive, Docs, and YouTube data in ways that external models simply cannot access.

Where this makes Gemini strong: tasks with very long or complex multimodal inputs, Google Workspace integration, large-context analysis, and cross-modal retrieval workflows.


DeepSeek: open weights, enormous scale, coding focus

DeepSeek is a Chinese AI company whose models are largely open-weight, meaning the trained weights are publicly available for download and customization. Architecturally, DeepSeek's LLMs are closely related to Meta's LLaMA series: pre-norm decoder-only transformers with RMSNorm normalization, SwiGLU feed-forward layers, rotary positional embeddings (RoPE), and grouped-query attention.

The significant numbers:

ModelParametersTraining data
DeepSeek-LLM 7B / 67B7B and 67B~2 trillion tokens
DeepSeek-V3671B~15 trillion tokens
DeepSeek-R1671B + reasoning RLExtended from V3

DeepSeek-V3 and DeepSeek-R1 are reported to be particularly strong on coding and mathematical reasoning. The likely reasons: heavy code and math data in training, parameter counts large enough to encode substantial programming knowledge, and in R1's case, reinforcement learning specifically focused on improving reasoning steps rather than just final answer quality.

What 'open-weight' means in practice

Open-weight means the model's trained parameters are public and downloadable. You can run DeepSeek locally, fine-tune it on your own data, or deploy it without API dependencies. This makes it fundamentally different from Claude, GPT-4o, or Gemini, which are only accessible through API calls to the company's servers.

The bilingual training on English and Chinese also gives DeepSeek unusual strength on Chinese-language tasks, which GPT-4o and Claude handle less consistently.

Where this makes DeepSeek strong: coding tasks, math and logical reasoning, bilingual Chinese-English workflows, and any context where open-weight access matters, such as enterprise deployments, fine-tuning, or running locally without external dependencies.


Perplexity: this one is not a model

This is the important one to understand, because Perplexity is categorically different from everything else on this list.

Claude, ChatGPT, Gemini, and DeepSeek are all model families. Perplexity is not a model. It is a multi-model orchestration system that sits on top of several foundation models (including Claude, GPT, Gemini, and others) and connects them to search, retrieval, file reading, code execution, and planning agents.

The internal architecture looks roughly like this:

  1. 01
    A meta-router receives your query and classifies it by type and complexity: is this a simple factual question, a deep research task, a coding problem, a conversational exchange?
  2. 02
    Based on that classification, it routes to one or more underlying models. Coding queries may go to a different model than research queries. Complex tasks may use multiple models in sequence.
  3. 03
    Specialized sub-agents handle web search, document retrieval, file reading, and code execution. Their outputs are fed back into the LLM as context for synthesis.
  4. 04
    For "Deep Research" and "Pro Search" modes, this runs as a multi-step planning loop: break the task into sub-queries, perform many searches, read and rank documents, iteratively build the answer with citations.

This architecture explains why Perplexity feels uniquely strong for research tasks. It is not that the underlying model is smarter than Claude or GPT-4o. It is that the system around the model is specifically designed to read far more sources than any single model prompt can hold, cross-check them, and produce a synthesized answer with traceable citations.

The key distinction

Claude, ChatGPT, Gemini, and DeepSeek can be used as components inside an agentic system like Perplexity. But being a model and being a research orchestration system are different things. When you use Perplexity's Deep Research mode, you are essentially running a small research workflow, not just prompting a language model.

Where this makes Perplexity strong: any task that benefits from synthesizing across many current sources, fact-checking against live web data, or producing cited reports. It is not especially differentiated on pure reasoning, coding, or multimodal tasks, because those ultimately rely on the same underlying models that Claude and GPT-4o provide.


Why the performance differences are not random

Putting all of this together, the performance gaps you notice in practice are directly traceable to specific design decisions.

Coding and reasoning tasks tend to go well with Claude and DeepSeek. Claude's Constitutional AI training makes it follow complex instruction sets carefully and reason through multi-step problems without drifting. DeepSeek's enormous parameter count and code-heavy training data give it strong raw coding performance, especially as an open-weight model you can run locally or fine-tune.

Multimodal tasks (understanding images, generating visuals, voice interaction) tend to go well with ChatGPT and Gemini because their architectures are fundamentally designed around interleaved multimodal tokens. They do not treat image understanding as a bolt-on capability; it is baked into the model at the architecture level.

Deep research and source synthesis is Perplexity's clear lane, not because its underlying models are better, but because it runs an agentic retrieval pipeline that can search, read, and synthesize across far more sources than a single context window allows. A plain LLM doing "research" is doing sophisticated pattern completion over its training data. Perplexity is doing live retrieval.


How they actually compare

SystemWhat it isModalitiesWhat makes it distinctBest for
ClaudeSingle LLM familyText (images as input in recent versions)Constitutional AI, long context, careful reasoningCoding, long documents, safety-sensitive tasks, following complex constraints
ChatGPT (GPT-4o)Single multimodal LLMText, images, audio (native)Unified multimodal transformer, generalist RLHFGeneral chat, visual tasks, image generation, voice, coding
GeminiMultimodal LLM familyText, images, audio, video, code (native)Very long context, mixture-of-experts, Google ecosystemMultimodal analysis, very large contexts, Google Workspace integration
DeepSeekOpen-weight LLM familyText (strong on code and math)Open weights, large scale, coding/math RLCoding, math, bilingual Chinese-English, local deployment
PerplexityMulti-model research systemText plus tool-based media via underlying modelsMeta-router, sub-agents, multi-step retrieval pipelineDeep web research, cited synthesis, tool-augmented workflows

The honest takeaway

None of these systems is universally smarter than the others. They are each strong in their lane because they were built for their lane.

If you are doing coding work or reasoning through a complex problem, Claude and DeepSeek are built for that. If you need to process images, generate visuals, or use voice, ChatGPT and Gemini are architected for that. If you need to synthesize something across dozens of live sources with citations, Perplexity is a research pipeline, not just a model, and that architecture pays off.

The interesting implication is that the correct mental model for using these tools is not "which AI is best" but "which system is designed for this specific type of task." The architectural decisions made years before you ran your query are what determine the answer.

Liked this post? Share it with someone on 𝕏 Twitter or LinkedIn. If you found it useful or have a question, send me a message and I'll do my best to get back to you.

If this was worth your time, .

Sarath Tharayil
/ SEE ALSO
Why Things Work Badly, Strangely, or Unexpectedly in Modern LifeApr 28, 2026
/ CONTENTS(9)
The shared foundationClaude: long-context reasoning, trained to be carefulChatGPT (GPT-4o): one model that handles everythingGemini: Google's bet on scale and multimodalityDeepSeek: open weights, enormous scale, coding focusPerplexity: this one is not a modelWhy the performance differences are not randomHow they actually compareThe honest takeaway
--
/ THAT'S A WRAP

Have a great day.

Thanks for reading all the way to the end.