What is NLP? (Natural Language Processing)
NLP explained simply: how computers work with human language.
View all nlp depths →Depth ladder for this topic:
Every time you type a search query and get a useful result, talk to a customer support chatbot, or watch your email client automatically filter spam, you’re interacting with Natural Language Processing. NLP is the branch of AI that lets computers read, understand, and generate human language — not just find keywords, but actually interpret meaning.
It’s one of the oldest and most practically useful areas of AI, and it’s gone through a radical transformation in the last few years.
What NLP systems actually do
NLP isn’t one thing — it’s a collection of tasks, each solving a different piece of the language problem. Here’s what that looks like in concrete terms:
Text classification assigns categories to text. A spam filter classifying your email as junk is doing text classification. So is a sentiment analysis system reading customer reviews and tagging them as positive, neutral, or negative. Input: raw text. Output: a label or category.
Named entity recognition (NER) identifies specific types of information within text — names of people, companies, dates, locations, dollar amounts. A legal document processing system uses NER to extract “plaintiff,” “defendant,” and “filing date” automatically.
Summarization compresses long documents into shorter versions while preserving the key points. Think of the “summarize this thread” button in your email client, or the AI that condenses a 40-page research report into a one-page brief.
Question answering takes a question and a body of text and finds or generates the answer. This is what powers AI assistants when they respond to “What does this contract say about termination clauses?”
Machine translation converts text from one language to another. Modern translation systems don’t work word-by-word — they understand sentence structure, idiomatic expressions, and context before producing a translation.
Input → process → output: a real example
Scenario: A healthcare company wants to automatically route patient feedback to the right department.
Input: “The waiting room was freezing and no one told me the doctor was running an hour late.”
NLP process:
- Sentiment analysis: negative (0.87 confidence)
- Topic extraction: facility comfort, communication, wait time
- Department routing: Operations + Patient Experience
Output: Ticket automatically tagged, routed to two departments, and flagged as high-priority due to combined negative sentiment and multiple complaint categories.
What would have taken a human 2–3 minutes per ticket is handled in milliseconds at any volume.
NLP before and after LLMs
Classic NLP (pre-2020) relied on purpose-built pipelines. Each task — classification, extraction, summarization — had its own specialized model, trained on labeled data for exactly that task. Switching tasks meant retraining from scratch.
LLMs changed this fundamentally. A single model can now handle most NLP tasks through prompting. You don’t need a dedicated summarization model and a separate classification model — you can ask one LLM to do both in the same call.
But classic NLP techniques are still worth knowing:
- Regex and rule-based extraction is deterministic, fast, and cheap when patterns are fixed (extracting invoice numbers, phone formats, ISO dates).
- Lightweight classifiers (like Naive Bayes or logistic regression on TF-IDF features) run in microseconds and can handle millions of documents per day without the cost of an LLM call.
- Specialized domain models (fine-tuned for medical, legal, or financial text) often outperform general LLMs on narrow tasks within those domains.
The practical question is always: does this task need an LLM’s generalization, or can a simpler NLP approach handle it reliably at much lower cost?
Try this now
Open any LLM (ChatGPT, Claude, Gemini) and try these three NLP tasks on a piece of text from your own work:
- Classify it: “What is the primary topic of this text? Give me one category and your confidence level.”
- Extract entities: “List all people, companies, and dates mentioned in this text.”
- Summarize it: “Summarize this in three bullet points, each under 15 words.”
Notice how one model handles all three without any retraining. That’s the LLM revolution for NLP in practice.
Pitfalls and failure modes
Assuming sentiment analysis is universal. Sentiment models trained on English-language product reviews will perform poorly on legal language, medical text, or sarcasm. Domain matters enormously. Always test on your actual text type before deploying.
Treating NLP output as ground truth. NLP systems return probabilities and predictions, not facts. A model that’s 85% confident it extracted the correct date is wrong 15% of the time. At scale, that’s a lot of errors. Build human review checkpoints for high-stakes extractions.
Over-engineering with LLMs when simple tools suffice. If you need to extract structured phone numbers from a database of records, regex is faster, cheaper, and more reliable than an LLM. LLMs are powerful but expensive for tasks with fixed patterns.
Ignoring language and cultural context. An NLP system trained on American English will struggle with British idioms, informal dialects, multilingual text, or regional business conventions. Know your training data’s origin before trusting its outputs on diverse real-world inputs.
The quick mental model
NLP is the problem domain: helping computers understand and generate human language.
LLMs are one extraordinarily powerful approach to solving many NLP problems at once.
Classic NLP tools are still useful for fast, controlled, and cost-efficient specific tasks.
If you understand that hierarchy, you can make better architectural decisions about which tool to reach for — and you won’t be surprised when a simple regex outperforms a fancy model on the task you actually need solved.
Go deeper
NLP from Rules to Transformers — What Changed and What Still Matters →
Related reads
Stay ahead of the AI curve
Weekly insights on AI — explained at the level that's right for you. No hype, no jargon, just what matters.
No spam. Unsubscribe anytime. We respect your inbox.