When AI Reads Images and Text Together: Multimodal AI Explained
ChatGPT can now see photos. Gemini can watch videos. This is multimodal AI — AI that processes more than just text. Here's how it works and why it matters, explained simply.
View all multimodal ai depths →Depth ladder for this topic:
For most of AI’s history, you had to pick: language AI or vision AI. Models that could read and write couldn’t look at pictures. Models that could identify images couldn’t hold a conversation.
That changed. Today’s leading AI systems — GPT-4o, Claude 3.7, Gemini — can process text, images, audio, and sometimes video, all at once. You can show ChatGPT a photo of a broken circuit board and ask what’s wrong with it. You can give Gemini a 30-minute lecture video and ask it to summarize the key points. You can paste a handwritten recipe into Claude and ask it to convert measurements.
This is multimodal AI — AI that works across multiple modes of information.
Why “multimodal” matters
For most of the history of machine learning, different types of data required different tools:
- Text → NLP models
- Images → Computer vision models
- Audio → Speech recognition models
- Video → Video analysis models
These systems couldn’t talk to each other. You couldn’t easily build something that understood a photo and then talked about it naturally.
Multimodal models change this by learning a shared representation — a common “understanding space” — across different types of input. When a multimodal model processes an image, it translates that image into a representation that lives in the same space as its text representations. That means it can reason about text and image together, the same way a human can look at a photo and describe it in words without having to “translate” between two separate systems.
How does an AI see?
Text is easy to represent for AI — you break it into tokens (roughly words), and each token gets a numerical representation.
Images are different. A photo doesn’t have natural “units” the way text has words.
The solution: treat image regions like tokens. Modern vision-language models:
- Divide the image into a grid of patches (e.g., 16×16 pixel squares)
- Convert each patch into a numerical representation using a vision encoder
- Feed these patch representations alongside text token representations into the language model
The language model now sees the image as a sequence of “visual tokens” alongside the regular text tokens. It can attend to both simultaneously — literally “reading” image and text together.
This is why you can ask GPT-4o “What text does this sign say?” and point it at a photo of a street sign. The model is processing the visual tokens from the sign image and the text tokens of your question at the same time, reasoning about them together.
What multimodal AI can do today
Describe and analyze images: “What’s happening in this photo?” “How many people are in this image?” “What’s the mood of this painting?”
Extract text from images (OCR): “What does the label on this product say?” “Read the text from this scanned form.” “What does the whiteboard say?”
Understand charts and data: “What trend does this graph show?” “Which month had the highest sales in this chart?” “Summarize this data visualization.”
Answer questions about documents: Upload a PDF and ask questions. Get summaries of specific sections. Extract specific data points.
Analyze visual content for feedback: “What’s wrong with this UI design?” “Is this before/after the edit I described?” “Does this photo look professional enough for a website?”
Caption and describe for accessibility: Generate image descriptions, alt text for web images, or detailed descriptions for visually impaired users.
What multimodal AI is not (yet)
A magic image understanding machine: Multimodal models are impressive but not infallible. They can misidentify objects, miss details in complex images, and be confidently wrong about what they see. For medical diagnosis, legal evidence, or safety-critical decisions, never rely on AI image analysis without expert human review.
Consistently accurate at counting: Ask it to count many objects in a photo and it’ll often be off, especially with many similar items.
Perfect at reading handwriting: It does better than nothing, but messy handwriting is still unreliable.
Video-native (in most models): Most multimodal models process video by sampling frames rather than truly understanding motion and time. Gemini 2.0 is the exception with native video understanding.
Audio too
The newest frontier of multimodal AI is audio. GPT-4o (the “o” stands for “omni”) can:
- Listen to your voice and respond — not by transcribing then sending text, but by processing audio directly
- Understand tone, emotion, and inflection
- Respond in voice with natural pacing and expression
- Process multiple speakers in a recording
This is why GPT-4o’s voice mode feels more natural than earlier AI voice assistants — it’s reasoning about the audio itself, not just the words.
Gemini 2.0 goes further, handling audio natively alongside text and video in a unified model.
Why this is a big deal
Multimodal AI is closing the gap between how AI processes information and how humans do.
Humans don’t read text, look at pictures, and listen to audio as three separate systems that can’t talk to each other. We integrate information from all our senses constantly. A doctor looks at an X-ray while reading the patient’s history. A mechanic listens to an engine while watching how it runs. A lawyer reads a contract while looking at relevant photos.
Multimodal AI is starting to work the same way — processing different kinds of information together in a unified context. This opens up use cases that simply weren’t possible with single-modality AI: analyzing medical images with clinical notes, reviewing products from photos and descriptions together, understanding video content at scale.
The shift from single-modality to multimodal is one of the most practically significant developments in AI in the past few years. And it’s accelerating.
Ready to go deeper? The 🔵 Applied guide covers specific multimodal AI use cases and how to use them in real workflows. The 🟣 Technical article explains vision encoders, cross-modal attention, and how training works.
Go deeper
Multimodal AI: What You Can Build When AI Sees, Hears, and Reads →
Related reads
Stay ahead of the AI curve
Weekly insights on AI — explained at the level that's right for you. No hype, no jargon, just what matters.
No spam. Unsubscribe anytime. We respect your inbox.