🕷️ Crawler Inspector

URL Lookup

Direct Parameter Lookup

Raw Queries and Responses

1. Shard Calculation

Query:
Response:
Calculated Shard: 39 (from laksa182)

2. Crawled Status Check

Query:
Response:

3. Robots.txt Check

Query:
Response:

4. Spam/Ban Check

Query:
Response:

5. Seen Status Check

ℹ️ Skipped - page is already crawled

đź“„
INDEXABLE
âś…
CRAWLED
6 days ago
🤖
ROBOTS SERVER UNREACHABLE
Failed to connect to robots server: Operation timed out after 2001 milliseconds with 0 bytes received

Page Info Filters

FilterStatusConditionDetails
HTTP statusPASSdownload_http_code = 200HTTP 200
Age cutoffPASSdownload_stamp > now() - 6 MONTH0.2 months ago
History dropPASSisNull(history_drop_reason)No drop reason
Spam/banPASSfh_dont_index != 1 AND ml_spam_score = 0ml_spam_score=0
CanonicalPASSmeta_canonical IS NULL OR = '' OR = src_unparsedNot set

Page Details

PropertyValue
URLhttps://www.ibm.com/think/topics/natural-language-processing
Last Crawled2026-03-31 04:27:49 (6 days ago)
First Indexed2024-12-18 13:03:01 (1 year ago)
HTTP Status Code200
Meta TitleWhat Is NLP (Natural Language Processing)? | IBM
Meta DescriptionNatural language processing (NLP) is a subfield of artificial intelligence (AI) that uses machine learning to help computers communicate with human language.
Meta Canonicalnull
Boilerpipe Text
Natural language processing (NLP) is a subfield of computer science and artificial intelligence (AI) that uses machine learning to enable computers to understand and communicate with human language. NLP enables computers and digital devices to recognize, understand and generate text and speech by combining computational linguistics, the rule-based modeling of human language together with statistical modeling, machine learning and deep learning . NLP research has helped enable the era of generative AI , from the communication skills of large language models (LLMs) to the ability of image generation models to understand requests. NLP is already part of everyday life for many, powering search engines, prompting chatbots for customer service with spoken commands, voice-operated GPS systems and question-answering digital assistants on smartphones such as Amazon’s Alexa, Apple’s Siri and Microsoft’s Cortana. NLP also plays a growing role in enterprise solutions that help streamline and automate business operations, increase employee productivity and simplify business processes. NLP makes it easier for humans to communicate and collaborate with machines, by allowing them to do so in the natural human language they use every day. This offers benefits across many industries and applications. Automation of repetitive tasks Improved data analysis and insights Enhanced search Content generation Automation of repetitive tasks  NLP is especially useful in fully or partially  automating tasks  like customer support, data entry and document handling. For example, NLP-powered chatbots can handle routine customer queries, freeing up human agents for more complex issues. In  document processing , NLP tools can automatically classify, extract key information and summarize content, reducing the time and errors associated with manual data handling. NLP facilitates language translation, converting text from one language to another while preserving meaning, context and nuances. Improved data analysis NLP enhances data analysis by enabling the extraction of insights from unstructured text data, such as customer reviews, social media posts and news articles. By using  text mining  techniques, NLP can identify patterns, trends and sentiments that are not immediately obvious in large datasets. Sentiment analysis enables the  extraction of subjective qualities , attitudes, emotions, sarcasm, confusion or suspicion from text. This is often used for routing communications to the system or the person most likely to make the next response. This allows businesses to better understand customer preferences, market conditions and public opinion. NLP tools can also perform categorization and summarization of vast amounts of text, making it easier for analysts to identify key information and make data-driven decisions more efficiently. Enhanced search NLP benefits search by enabling systems to understand the intent behind user queries, providing more accurate and contextually relevant results. Instead of relying solely on keyword matching, NLP-powered search engines analyze the meaning of words and phrases, making it easier to find information even when queries are vague or complex. This improves user experience, whether in web searches, document retrieval or enterprise data systems. Powerful content generation NLP powers advanced language models to  create human-like text  for various purposes. Pre-trained models, such as GPT-4, can generate articles, reports, marketing copy, product descriptions and even creative writing based on prompts provided by users. NLP-powered tools can also assist in automating tasks like drafting emails, writing social media posts or legal documentation. By understanding context, tone and style, NLP sees to it that the generated content is coherent, relevant and aligned with the intended message, saving time and effort in content creation while maintaining quality. NLP combines the power of computational linguistics together with machine learning algorithms and deep learning. Computational linguistics uses data science to analyze language and speech. It includes two main types of analysis: syntactical analysis and semantical analysis. Syntactical analysis determines the meaning of a word, phrase or sentence by parsing the syntax of the words and applying preprogrammed rules of grammar. Semantical analysis uses the syntactic output to draw meaning from the words and interpret their meaning within the sentence structure. The parsing of words can take one of two forms. Dependency parsing looks at the relationships between words, such as identifying nouns and verbs, while constituency parsing then builds a parse tree (or syntax tree): a rooted and ordered representation of the syntactic structure of the sentence or string of words. The resulting parse trees underly the functions of language translators and speech recognition. Ideally, this analysis makes the output either text or speech understandable to both NLP models and people. Self-supervised learning (SSL) in particular is useful for supporting NLP because NLP requires large amounts of labeled data to train AI models. Because these labeled datasets require time-consuming annotation, a process involving manual labeling by humans, gathering sufficient data can be prohibitively difficult. Self-supervised approaches can be more time-effective and cost-effective, as they replace some or all manually labeled training data.   Three different approaches to NLP include: Rules-based NLP The earliest NLP applications were simple if-then decision trees, requiring preprogrammed rules. They are only able to provide answers in response to specific prompts, such as the original version of Moviefone, which had rudimentary natural language generation (NLG) capabilities. Because there is no machine learning or AI capability in rules-based NLP, this function is highly limited and not scalable. Statistical NLP Developed later, statistical NLP automatically extracts, classifies and labels elements of text and voice data and then assigns a statistical likelihood to each possible meaning of those elements. This relies on machine learning, enabling a sophisticated breakdown of linguistics such as part-of-speech tagging.   Statistical NLP introduced the essential technique of mapping language elements, such as words and grammatical rules to a vector representation so that language can be modeled by using mathematical (statistical) methods, including regression or Markov models. This informed early NLP developments such as spellcheckers and T9 texting (Text on 9 keys, to be used on Touch-Tone telephones). Deep learning NLP Recently, deep learning models have become the dominant mode of NLP, by using huge volumes of raw,  unstructured  data both text and voice to become ever more accurate. Deep learning can be viewed as a further evolution of statistical NLP, with the difference that it uses  neural network  models. There are several subcategories of models: Sequence-to-Sequence  (seq2seq) models: Based  on recurrent neural networks (RNN) , they have mostly been used for machine translation by converting a phrase from one domain (such as the German language) into the phrase of another domain (such as English). Transformer models : They use  tokenization  of language (the position of each token words or subwords) and self-attention (capturing dependencies and relationships) to calculate the relation of different language parts to one another.  Transformer models  can be efficiently trained by using  self-supervised learning  on massive text databases. A landmark in  transformer models  was Google’s bidirectional encoder representations from transformers (BERT), which became and remains the basis of how Google’s search engine works. Autoregressive models : This type of transformer model is trained specifically to predict the next word in a sequence, which represents a huge leap forward in the ability to generate text. Examples of autoregressive LLMs include GPT,  Llama , Claude and the open-source Mistral. Foundation models : Prebuilt and curated foundation models can speed the launching of an NLP effort and boost trust in its operation. For example, the  IBM® Granite™  foundation models are widely applicable across industries. They support NLP tasks including content generation and insight extraction. Additionally, they facilitate retrieval-augmented generation, a framework for improving the quality of response by linking the model to external sources of knowledge. The models also perform named entity recognition which involves identifying and extracting key information in a text. Several NLP tasks typically help process human text and voice data in ways that help the computer make sense of what it’s ingesting. Some of these tasks include: Coreference resolution Named entity recognition Part-of-speech tagging Word sense disambiguation Coreference resolution This is the task of identifying if and when two words refer to the same entity. The most common example is determining the person or object to which a certain pronoun refers (such as “she” = “Mary”). But it can also identify a metaphor or an idiom in the text (such as an instance in which “bear” isn’t an animal, but a large and hairy person).  Named entity recognition (NER) NER identifies words or phrases as useful entities. NER identifies “London” as a location or “Maria” as a person's name. Part-of-speech tagging Also called grammatical tagging, this is the process of determining which part of speech a word or piece of text is, based on its use and context. For example, part-of-speech identifies “make” as a verb in “I can make a paper plane,” and as a noun in “What make of car do you own?” Word sense disambiguation This is the selection of a word meaning for a word with multiple possible meanings. This uses a process of semantic analysis to examine the word in context. For example, word sense disambiguation helps distinguish the meaning of the verb “make” in “make the grade” (to achieve) versus “make a bet” (to place). Sorting out “I will be merry when I marry Mary” requires a sophisticated NLP system. NLP works by combining various computational techniques to analyze, understand and generate human language in a way that machines can process. Here is an overview of a typical NLP pipeline and its steps: Text preprocessing NLP text preprocessing prepares raw text for analysis by transforming it into a format that machines can more easily understand. It begins with tokenization, which involves splitting the text into smaller units like words, sentences or phrases. This helps break down complex text into manageable parts. Next, lowercasing is applied to standardize the text by converting all characters to lowercase, ensuring that words like "Apple" and "apple" are treated the same. Stop word removal is another common step, where frequently used words like "is" or "the" are filtered out because they don't add significant meaning to the text.  Stemming  or  lemmatization  reduces words to their root form (e.g., "running" becomes "run"), making it easier to analyze language by grouping different forms of the same word. Additionally, text cleaning removes unwanted elements such as punctuation, special characters and numbers that may clutter the analysis. After preprocessing, the text is clean, standardized and ready for machine learning models to interpret effectively. Feature extraction Feature extraction is the process of converting raw text into numerical representations that machines can analyze and interpret. This involves transforming text into structured data by using NLP techniques like  Bag of Words  and TF-IDF, which quantify the presence and importance of words in a document. More advanced methods include  word embeddings  like Word2Vec or GloVe, which represent words as dense vectors in a continuous space, capturing semantic relationships between words. Contextual embeddings further enhance this by considering the context in which words appear, allowing for richer, more nuanced representations. Text analysis Text analysis involves interpreting and extracting meaningful information from text data through various computational techniques. This process includes tasks such as part-of-speech (POS) tagging, which identifies grammatical roles of words and named entity recognition (NER), which detects specific entities like names, locations and dates. Dependency parsing analyzes grammatical relationships between words to understand sentence structure, while sentiment analysis determines the emotional tone of the text, assessing whether it is positive, negative or neutral. Topic modeling identifies underlying themes or topics within a text or across a corpus of documents. Natural language understanding (NLU) is a subset of NLP that focuses on analyzing the meaning behind sentences. NLU enables software to find similar meanings in different sentences or to process words that have different meanings. Through these techniques, NLP text analysis transforms unstructured text into insights. Model training Processed data is then used to train machine learning models, which learn patterns and relationships within the data. During training, the model adjusts its parameters to minimize errors and improve its performance. Once trained, the model can be used to make predictions or generate outputs on new, unseen data. The effectiveness of NLP modeling is continually refined through evaluation, validation and fine-tuning to enhance accuracy and relevance in real-world applications. Different software environments are useful throughout the said processes. For example, the Natural Language Toolkit (NLTK) is a suite of libraries and programs for English that is written in the Python programming language. It supports text classification, tokenization, stemming, tagging, parsing and semantic reasoning functionalities. TensorFlow is a free and open-source software library for machine learning and AI that can be used to train models for NLP applications. Tutorials and certifications abound for those interested in familiarizing themselves with such tools. Even state-of-the-art NLP models are not perfect, just as human speech is prone to error. As with any AI technology, NLP comes with potential pitfalls. Human language is filled with ambiguities that make it difficult for programmers to write software that accurately determines the intended meaning of text or voice data. Human language might take years for humans to learn and many never stop learning. But then programmers must teach natural language-powered applications to recognize and understand irregularities so their applications can be accurate and useful. Associated risks might include: Biased training As with any AI function, biased data used in training will skew the answers. The more diverse the users of an NLP function, the more significant this risk becomes, such as in government services, healthcare and HR interactions. Training datasets scraped from the web, for example, are prone to bias. Misinterpretation As in programming, there is a risk of garbage in, garbage out (GIGO). Speech recognition , also known as speech-to-text, is the task of reliably converting voice data into text data. But NLP solutions can become confused if spoken input is in an obscure dialect, mumbled, too full of slang, homonyms, incorrect grammar, idioms, fragments, mispronunciations, contractions or recorded with too much background noise. New vocabulary New words are continually being invented or imported. The conventions of grammar can evolve or be intentionally broken. In these cases, NLP can either make a best guess or admit it’s unsure and either way, this creates a complication. Tone of voice When people speak, their verbal delivery or even body language can give an entirely different meaning than the words alone. Exaggeration for effect, stressing words for importance or sarcasm can be confused by NLP, making the semantic analysis more difficult and less reliable. NLP applications can now be found across virtually every industry. Finance In financial dealings, nanoseconds might make the difference between success and failure when accessing data, or making trades or deals. NLP can speed the mining of information from financial statements, annual and regulatory reports, news releases or even social media. Healthcare New medical insights and breakthroughs can arrive faster than many healthcare professionals can keep up. NLP and AI-based tools can help speed the analysis of health records and medical research papers, making better-informed medical decisions possible, or assisting in the detection or even prevention of medical conditions. Insurance NLP can analyze claims to look for patterns that can identify areas of concern and find inefficiencies in claims processing, leading to greater optimization of processing and employee efforts. Legal Almost any legal case might require reviewing mounds of paperwork, background information and legal precedent. NLP can help automate legal discovery, assisting in the organization of information, speeding review and making sure that all relevant details are captured for consideration.
Markdown
# What is NLP (natural language processing)? Machine learning - [Welcome](https://www.ibm.com/think/machine-learning#605511093) - Introduction - [Overview](https://www.ibm.com/think/topics/machine-learning#7281535) - [Machine learning types](https://www.ibm.com/think/topics/machine-learning-types#7281536) - [Machine learning algorithms](https://www.ibm.com/think/topics/machine-learning-algorithms#7281537) - Data science for machine learning - [Statistical machine learning](https://www.ibm.com/think/topics/statistical-machine-learning#498277090) - [Linear algebra for machine learning](https://www.ibm.com/think/topics/linear-algebra-for-machine-learning#498277089) - [Uncertainty quantification](https://www.ibm.com/think/topics/uncertainty-quantification#498277088) - [Bias variance tradeoff](https://www.ibm.com/think/topics/bias-variance-tradeoff#498277087) - [Bayesian Statistics](https://www.ibm.com/think/topics/bayesian-statistics#498277086) - [Singular value decomposition](https://www.ibm.com/think/topics/singular-value-decomposition#498277085) - Feature Engineering - [Overview](https://www.ibm.com/think/topics/feature-engineering#1003835715) - [Feature selection](https://www.ibm.com/think/topics/feature-selection#1003835714) - [Feature extraction](https://www.ibm.com/think/topics/feature-extraction#1003835713) - [Vector embedding](https://www.ibm.com/think/topics/vector-embedding#1003835712) - [Latent space](https://www.ibm.com/think/topics/latent-space#1003835711) - Dimensionality reduction - [Principal component analysis](https://www.ibm.com/think/topics/principal-component-analysis#1793360183) - [Linear discriminant analysis](https://www.ibm.com/think/topics/linear-discriminant-analysis#1793360182) - [Upsampling](https://www.ibm.com/think/topics/upsampling#1003835709) - [Downsampling](https://www.ibm.com/think/topics/downsampling#1003835708) - [Synthetic data](https://www.ibm.com/think/topics/synthetic-data#1003835707) - [Data leakage](https://www.ibm.com/think/topics/data-leakage-machine-learning#1003835706) - Supervised learning - [Overview](https://www.ibm.com/think/topics/supervised-learning#1509394340) - Regression - [Linear regression](https://www.ibm.com/think/topics/linear-regression#1190488334) - [Lasso regression](https://www.ibm.com/think/topics/lasso-regression#1190488335) - [Ridge regression](https://www.ibm.com/think/topics/ridge-regression#1190488336) - [State space model](https://www.ibm.com/think/topics/state-space-model#1190488337) - [Time series](https://www.ibm.com/think/topics/time-series-model#1190488338) - [Autoregressive model](https://www.ibm.com/think/topics/autoregressive-model#1190488339) - Classification - [Overview](https://www.ibm.com/think/topics/classification-machine-learning#684929709) - [Decision trees](https://www.ibm.com/think/topics/decision-trees#684929710) - [K-nearest neighbors (KNNs)](https://www.ibm.com/think/topics/knn#684929711) - [Naive bayes](https://www.ibm.com/think/topics/naive-bayes#684929712) - [Random forest](https://www.ibm.com/think/topics/random-forest#684929713) - [Support vector machine](https://www.ibm.com/think/topics/support-vector-machine#684929714) - [Logistic regression](https://www.ibm.com/think/topics/logistic-regression#684929715) - Ensemble learning - [Overview](https://www.ibm.com/think/topics/ensemble-learning#179371084) - [Boosting](https://www.ibm.com/think/topics/boosting#179371085) - [Bagging](https://www.ibm.com/think/topics/bagging#179371086) - [Gradient boosting](https://www.ibm.com/think/topics/gradient-boosting#179371087) - [Gradient boosting classifier](https://www.ibm.com/think/tutorials/gradient-boosting-classifier#179371088) - Self-supervised learning - [Overview](https://www.ibm.com/think/topics/self-supervised-learning#326187541) - [Transfer learning](https://www.ibm.com/think/topics/transfer-learning#326187540) - Unsupervised learning - [Overview](https://www.ibm.com/think/topics/unsupervised-learning#2014952965) - Clustering - [Overview](https://www.ibm.com/think/topics/clustering#2142864945) - [K means clustering](https://www.ibm.com/think/topics/k-means-clustering#2142864944) - [Hierarchical clustering](https://www.ibm.com/think/topics/hierarchical-clustering#2142864943) - [A priori algorithm](https://www.ibm.com/think/topics/apriori-algorithm#2014952963) - [Gaussian mixture model](https://www.ibm.com/think/topics/gaussian-mixture-model#2014952962) - [Anomaly detection](https://www.ibm.com/think/topics/machine-learning-for-anomaly-detection#2014952961) - Semi-supervised learning - [Overview](https://www.ibm.com/think/topics/semi-supervised-learning#1774455706) - Recommendation engine - [Collaborative filtering](https://www.ibm.com/think/topics/collaborative-filtering#1181250928) - [Content based filtering](https://www.ibm.com/think/topics/content-based-filtering#1181250927) - Reinforcement learning - [Overview](https://www.ibm.com/think/topics/reinforcement-learning#1268897081) - [Reinforcement learning human feedback](https://www.ibm.com/think/topics/rlhf#1268897082) - Deep Learning - [Overview](https://www.ibm.com/think/topics/deep-learning#763338456) - Neural networks - [Overview](https://www.ibm.com/think/topics/neural-networks#741977106) - [Backpropagation](https://www.ibm.com/think/topics/backpropagation#741977107) - [Encoder-decoder model](https://www.ibm.com/think/topics/encoder-decoder-model#741977108) - [Recurrent neural networks](https://www.ibm.com/think/topics/recurrent-neural-networks#763338458) - [Long short-term memory (LSTM)](https://www.ibm.com/think/topics/lstm#763338459) - [Convolutional neural networks](https://www.ibm.com/think/topics/convolutional-neural-networks#763338460) - Transformer models - [Overview](https://www.ibm.com/think/topics/transformer-model#1280257394) - [Attention mechanism](https://www.ibm.com/think/topics/attention-mechanism#1280257393) - [Grouped query attention](https://www.ibm.com/think/topics/grouped-query-attention#1280257392) - [Positional encoding](https://www.ibm.com/think/topics/positional-encoding#1280257391) - [Autoencoder](https://www.ibm.com/think/topics/autoencoder#763338462) - [Mamba model](https://www.ibm.com/think/topics/mamba-model#763338463) - [Graph neural network](https://www.ibm.com/think/topics/graph-neural-network#763338464) - Generative AI - [Overview](https://www.ibm.com/think/topics/generative-ai#257779831) - [Generative model](https://www.ibm.com/think/topics/generative-model#257779832) - [Generative AI vs. predictive AI](https://www.ibm.com/think/topics/generative-ai-vs-predictive-ai-whats-the-difference#257779833) - Large language models (LLMs) - [Overview](https://www.ibm.com/think/topics/large-language-models#692473873) - [Reasoning models](https://www.ibm.com/think/topics/reasoning-model#692473874) - [Small language models](https://www.ibm.com/think/topics/small-language-models#692473875) - [Instruction tuning](https://www.ibm.com/think/topics/instruction-tuning#692473876) - [LLM parameters](https://www.ibm.com/think/topics/llm-parameters#692473877) - [LLM temperature](https://www.ibm.com/think/topics/llm-temperature#692473878) - [LLM benchmarks](https://www.ibm.com/think/topics/llm-benchmarks#692473879) - [LLM customization](https://www.ibm.com/think/topics/llm-customization#692473880) - [LLM alignment](https://www.ibm.com/think/topics/llm-alignment#692473881) - AI image generation - [Diffusion models](https://www.ibm.com/think/topics/diffusion-models#186915248) - [Variational autoencoder (VAE)](https://www.ibm.com/think/topics/variational-autoencoder#186915249) - [Generative adversarial networks (GANs)](https://www.ibm.com/think/topics/generative-adversarial-networks#186915250) - Multimodal AI - [Overview](https://www.ibm.com/think/topics/multimodal-ai#318643377) - [Vision language models](https://www.ibm.com/think/topics/vision-language-models#318643376) - [Tutorial: Build an AI stylist](https://www.ibm.com/think/tutorials/build-ai-stylist-with-granite-python-watsonx-ai#318643375) - [Tutorial: Multimodal AI queries using Llama](https://www.ibm.com/think/tutorials/multimodal-ai-python-llama#318643374) - [Tutorial: Multimodal AI queries using Pixtral](https://www.ibm.com/think/tutorials/multimodal-ai-pixtral-12b-on-watsonx-ai#318643373) - [Tutorial: Automatic podcast transcription with Granite](https://www.ibm.com/think/tutorials/automatic-speech-recognition-podcast-transcript-granite-watsonx-ai#318643372) - [Tutorial: PPT AI image analysis answering system](https://www.ibm.com/think/tutorials/build-ppt-ai-image-analysis-question-answering-system-granite#318643371) - Retrieval augmented generation (RAG) - [Overview](https://www.ibm.com/think/topics/retrieval-augmented-generation#824202002) - [GraphRAG](https://www.ibm.com/think/topics/graphrag#824202001) - [Tutorial: Build a multimodal RAG system with Docling and Granite](https://www.ibm.com/think/tutorials/build-multimodal-rag-langchain-with-docling-granite#824202000) - [Tutorial: Evaluate RAG pipline using Ragas](https://www.ibm.com/think/tutorials/evaluate-rag-pipeline-using-ragas-in-python-with-watsonx#824201999) - [Tutorial: RAG chunking strategies](https://www.ibm.com/think/tutorials/chunking-strategies-for-rag-with-langchain-watsonx-ai#824201998) - [Tutorial: Graph RAG using knowledge graphs](https://www.ibm.com/think/tutorials/knowledge-graph-rag#824201997) - [Tutorial: Inference scaling to improve multimodal RAG](https://www.ibm.com/think/tutorials/inference-scaling-improve-multimodal-rag-granite#824201996) - AI code generation - [Overview](https://www.ibm.com/think/topics/ai-code-generation#1329760627) - [Vibe coding](https://www.ibm.com/think/topics/vibe-coding#1329760626) - AI agents - [Visit the 2025 Guide to AI Agents](https://www.ibm.com/think/ai-agents#1835319252) - Model training - [Overview](https://www.ibm.com/think/topics/model-training#1580786329) - [LLM training](https://www.ibm.com/think/topics/llm-training#1580786328) - [Loss function](https://www.ibm.com/think/topics/loss-function#1580786327) - [Training data](https://www.ibm.com/think/topics/training-data#1580786326) - [Model parameters](https://www.ibm.com/think/topics/model-parameters#1580786325) - Optimization algorithm - [Gradient descent](https://www.ibm.com/think/topics/gradient-descent#18238369) - [Stochastic gradient descent](https://www.ibm.com/think/topics/stochastic-gradient-descent#18238368) - Model hyperparameters - [Hyperparameter tuning](https://www.ibm.com/think/topics/hyperparameter-tuning#523796994) - [Learning rate](https://www.ibm.com/think/topics/learning-rate#523796993) - Fine tuning - [Overview](https://www.ibm.com/think/topics/fine-tuning#1029355619) - [Parameter efficient fine tuning (PEFT)](https://www.ibm.com/think/topics/parameter-efficient-fine-tuning#1029355618) - [LoRA](https://www.ibm.com/think/topics/lora#1029355617) - [Tutorial: Fine tuning Granite model with LoRA](https://www.ibm.com/think/tutorials/lora-fine-tuning-granite-llm#1029355616) - [Regularization](https://www.ibm.com/think/topics/regularization#1580786321) - [Foundation models](https://www.ibm.com/think/topics/foundation-models#1580786320) - [Overfitting](https://www.ibm.com/think/topics/overfitting#1759735864) - [Underfitting](https://www.ibm.com/think/topics/underfitting#1759735863) - N-shot learning - [Few shot learning](https://www.ibm.com/think/topics/few-shot-learning#823293247) - [Zero shot learning](https://www.ibm.com/think/topics/zero-shot-learning#823293246) - [Knowledge distillation](https://www.ibm.com/think/topics/knowledge-distillation#1759735861) - [Meta learning](https://www.ibm.com/think/topics/meta-learning#1759735860) - [Data augmentation](https://www.ibm.com/think/topics/data-augmentation#1759735859) - Continual learning - [Catastrophic forgetting](https://www.ibm.com/think/topics/catastrophic-forgetting#1449439549) - Machine learning libraries - [Overview](https://www.ibm.com/think/topics/machine-learning-libraries#2086344954) - [Scikit-learn](https://www.ibm.com/think/topics/scikit-learn#2086344953) - [XGboost](https://www.ibm.com/think/topics/xgboost#2086344952) - [PyTorch](https://www.ibm.com/think/topics/pytorch#2086344951) - MLOps - [Overview](https://www.ibm.com/think/topics/mlops#1703063717) - [AI lifecyle](https://www.ibm.com/think/topics/ai-lifecycle#1703063718) - [AI inference](https://www.ibm.com/think/topics/ai-inference#1703063719) - [Model deployment](https://www.ibm.com/think/topics/model-deployment#1703063720) - [Machine learning pipeline](https://www.ibm.com/think/topics/machine-learning-pipeline#1703063721) - [Data labeling](https://www.ibm.com/think/topics/data-labeling#1703063722) - Model governance - [Model risk management](https://www.ibm.com/think/topics/model-risk-management#1399431040) - [Model drift](https://www.ibm.com/think/topics/model-drift#1399431041) - [AutoML](https://www.ibm.com/think/topics/automl#1703063724) - [Model selection](https://www.ibm.com/think/topics/model-selection#1703063725) - [Federated learning](https://www.ibm.com/think/topics/federated-learning#1703063726) - [Distributed machine learning](https://www.ibm.com/think/topics/distributed-machine-learning#1255367754) - [AI stack](https://www.ibm.com/think/topics/ai-stack#1255367755) - Natural language processing - [Overview](https://www.ibm.com/think/topics/natural-language-processing#1197505092) - [Natural language understanding](https://www.ibm.com/think/topics/natural-language-understanding#1197505093) - Text classification - [Overview](https://www.ibm.com/think/topics/text-classification#88312261) - [Sentiment analysis](https://www.ibm.com/think/topics/sentiment-analysis#88312262) - [Tutorial: Spam text classifier with PyTorch](https://www.ibm.com/think/tutorials/text-classification-pytorch#88312263) - [Machine translation](https://www.ibm.com/think/topics/machine-translation#1197505095) - Text mining - [Overview](https://www.ibm.com/think/topics/text-mining#922804989) - [Information retrieval](https://www.ibm.com/think/topics/information-retrieval#922804988) - [Information extraction](https://www.ibm.com/think/topics/information-extraction#922804987) - [Topic modeling](https://www.ibm.com/think/topics/topic-modeling#922804986) - [Latent semantic analysis](https://www.ibm.com/think/topics/latent-semantic-analysis#922804985) - [Latent Dirichlet Allocation](https://www.ibm.com/think/topics/latent-dirichlet-allocation#922804984) - [Named entity recognition](https://www.ibm.com/think/topics/named-entity-recognition#922804983) - [Word embeddings](https://www.ibm.com/think/topics/word-embeddings#1197505097) - [Bag of words](https://www.ibm.com/think/topics/bag-of-words#1197505098) - [Intelligent search](https://www.ibm.com/think/topics/intelligent-search#1197505099) - [Speech recognition](https://www.ibm.com/think/topics/speech-recognition#1197505100) - [Stemming and lemmatization](https://www.ibm.com/think/topics/stemming-lemmatization#1197505101) - [Text summarization](https://www.ibm.com/think/topics/text-summarization#1532047733) - [Conversational AI](https://www.ibm.com/think/topics/conversational-ai#1532047732) - [Conversational analytics](https://www.ibm.com/think/topics/conversational-analytics#1532047731) - [Natural language generation](https://www.ibm.com/think/topics/natural-language-generation#1532047730) - Computer vision - [Overview](https://www.ibm.com/think/topics/computer-vision#691946467) - [Image classification](https://www.ibm.com/think/topics/image-classification#691946468) - [Object detection](https://www.ibm.com/think/topics/object-detection#691946469) - Image segmentation - [Instance segmentation](https://www.ibm.com/think/topics/instance-segmentation#544367653) - [Semantic segmentation](https://www.ibm.com/think/topics/semantic-segmentation#544367654) - [Optical character recognition](https://www.ibm.com/think/topics/optical-character-recognition#691946471) - [Image recognition](https://www.ibm.com/think/topics/image-recognition#691946472) - [Visual inspection](https://www.ibm.com/think/topics/visual-inspection#691946473) ## Authors [Cole Stryker](https://www.ibm.com/think/author/cole-stryker.html) Staff Editor, AI Models IBM Think [Jim Holdsworth](https://www.ibm.com/think/author/jim-holdsworth.html) Staff Writer IBM Think ## What is NLP? Natural language processing (NLP) is a subfield of computer science and [artificial intelligence (AI)](https://www.ibm.com/think/topics/artificial-intelligence) that uses [machine learning](https://www.ibm.com/think/topics/machine-learning) to enable computers to understand and communicate with human language. NLP enables computers and digital devices to recognize, understand and generate text and speech by combining computational linguistics, the rule-based modeling of human language together with statistical modeling, machine learning and [deep learning](https://www.ibm.com/think/topics/deep-learning). NLP research has helped enable the era of [generative AI](https://www.ibm.com/think/topics/generative-ai), from the communication skills of [large language models](https://www.ibm.com/think/topics/large-language-models) (LLMs) to the ability of image generation models to understand requests. NLP is already part of everyday life for many, powering search engines, prompting [chatbots](https://www.ibm.com/think/topics/chatbots) for customer service with spoken commands, voice-operated GPS systems and question-answering digital assistants on smartphones such as Amazon’s Alexa, Apple’s Siri and Microsoft’s Cortana. NLP also plays a growing role in enterprise solutions that help streamline and automate business operations, increase employee productivity and simplify business processes. ## Benefits of NLP NLP makes it easier for humans to communicate and collaborate with machines, by allowing them to do so in the natural human language they use every day. This offers benefits across many industries and applications. - Automation of repetitive tasks - Improved data analysis and insights - Enhanced search - Content generation ### Automation of repetitive tasks NLP is especially useful in fully or partially [automating tasks](https://www.ibm.com/think/topics/automation) like customer support, data entry and document handling. For example, NLP-powered chatbots can handle routine customer queries, freeing up human agents for more complex issues. In [document processing](https://www.ibm.com/think/topics/document-processing), NLP tools can automatically classify, extract key information and summarize content, reducing the time and errors associated with manual data handling. NLP facilitates language translation, converting text from one language to another while preserving meaning, context and nuances. ### Improved data analysis NLP enhances data analysis by enabling the extraction of insights from unstructured text data, such as customer reviews, social media posts and news articles. By using [text mining](https://www.ibm.com/think/topics/text-mining-use-cases) techniques, NLP can identify patterns, trends and sentiments that are not immediately obvious in large datasets. Sentiment analysis enables the [extraction of subjective qualities](https://www.ibm.com/think/topics/sentiment-analysis), attitudes, emotions, sarcasm, confusion or suspicion from text. This is often used for routing communications to the system or the person most likely to make the next response. This allows businesses to better understand customer preferences, market conditions and public opinion. NLP tools can also perform categorization and summarization of vast amounts of text, making it easier for analysts to identify key information and make data-driven decisions more efficiently. ### Enhanced search NLP benefits search by enabling systems to understand the intent behind user queries, providing more accurate and contextually relevant results. Instead of relying solely on keyword matching, NLP-powered search engines analyze the meaning of words and phrases, making it easier to find information even when queries are vague or complex. This improves user experience, whether in web searches, document retrieval or enterprise data systems. ### Powerful content generation NLP powers advanced language models to [create human-like text](https://www.ibm.com/think/topics/text-generation) for various purposes. Pre-trained models, such as GPT-4, can generate articles, reports, marketing copy, product descriptions and even creative writing based on prompts provided by users. NLP-powered tools can also assist in automating tasks like drafting emails, writing social media posts or legal documentation. By understanding context, tone and style, NLP sees to it that the generated content is coherent, relevant and aligned with the intended message, saving time and effort in content creation while maintaining quality. ## Approaches to NLP NLP combines the power of computational linguistics together with [machine learning algorithms](https://www.ibm.com/think/topics/machine-learning-algorithms) and deep learning. Computational linguistics uses data science to analyze language and speech. It includes two main types of analysis: syntactical analysis and semantical analysis. Syntactical analysis determines the meaning of a word, phrase or sentence by parsing the syntax of the words and applying preprogrammed rules of grammar. Semantical analysis uses the syntactic output to draw meaning from the words and interpret their meaning within the sentence structure. The parsing of words can take one of two forms. Dependency parsing looks at the relationships between words, such as identifying nouns and verbs, while constituency parsing then builds a parse tree (or syntax tree): a rooted and ordered representation of the syntactic structure of the sentence or string of words. The resulting parse trees underly the functions of language translators and speech recognition. Ideally, this analysis makes the output either text or speech understandable to both NLP models and people. [Self-supervised learning (SSL)](https://www.ibm.com/think/topics/self-supervised-learning) in particular is useful for supporting NLP because NLP requires large amounts of labeled data to train AI models. Because these labeled datasets require time-consuming annotation, a process involving manual labeling by humans, gathering sufficient data can be prohibitively difficult. Self-supervised approaches can be more time-effective and cost-effective, as they replace some or all manually labeled training data. Three different approaches to NLP include: ### Rules-based NLP The earliest NLP applications were simple if-then decision trees, requiring preprogrammed rules. They are only able to provide answers in response to specific prompts, such as the original version of Moviefone, which had rudimentary natural language generation (NLG) capabilities. Because there is no machine learning or AI capability in rules-based NLP, this function is highly limited and not scalable. ### Statistical NLP Developed later, statistical NLP automatically extracts, classifies and labels elements of text and voice data and then assigns a statistical likelihood to each possible meaning of those elements. This relies on machine learning, enabling a sophisticated breakdown of linguistics such as part-of-speech tagging. Statistical NLP introduced the essential technique of mapping language elements, such as words and grammatical rules to a vector representation so that language can be modeled by using mathematical (statistical) methods, including regression or Markov models. This informed early NLP developments such as spellcheckers and T9 texting (Text on 9 keys, to be used on Touch-Tone telephones). ### Deep learning NLP Recently, deep learning models have become the dominant mode of NLP, by using huge volumes of raw, [unstructured](https://www.ibm.com/think/topics/text-mining) data both text and voice to become ever more accurate. Deep learning can be viewed as a further evolution of statistical NLP, with the difference that it uses [neural network](https://www.ibm.com/think/topics/neural-networks) models. There are several subcategories of models: - *Sequence-to-Sequence* (seq2seq) models: Based [on recurrent neural networks (RNN)](https://www.ibm.com/think/topics/recurrent-neural-networks), they have mostly been used for machine translation by converting a phrase from one domain (such as the German language) into the phrase of another domain (such as English). - *Transformer models*: They use [tokenization](https://www.ibm.com/docs/en/watsonx/saas?topic=solutions-tokens) of language (the position of each token words or subwords) and self-attention (capturing dependencies and relationships) to calculate the relation of different language parts to one another. [Transformer models](https://www.ibm.com/think/topics/transformer-model) can be efficiently trained by using [self-supervised learning](https://www.ibm.com/think/topics/self-supervised-learning) on massive text databases. A landmark in [transformer models](https://www.ibm.com/think/topics/transformer-model) was Google’s bidirectional encoder representations from transformers (BERT), which became and remains the basis of how Google’s search engine works. - *Autoregressive models*: This type of transformer model is trained specifically to predict the next word in a sequence, which represents a huge leap forward in the ability to generate text. Examples of autoregressive LLMs include GPT, [Llama](https://www.ibm.com/think/topics/llama-2), Claude and the open-source Mistral. - *Foundation models*: Prebuilt and curated foundation models can speed the launching of an NLP effort and boost trust in its operation. For example, the [IBM® Granite™](https://www.ibm.com/think/news/granite-foundation-models) foundation models are widely applicable across industries. They support NLP tasks including content generation and insight extraction. Additionally, they facilitate retrieval-augmented generation, a framework for improving the quality of response by linking the model to external sources of knowledge. The models also perform named entity recognition which involves identifying and extracting key information in a text. ## NLP Tasks Several NLP tasks typically help process human text and voice data in ways that help the computer make sense of what it’s ingesting. Some of these tasks include: - Coreference resolution - Named entity recognition - Part-of-speech tagging - Word sense disambiguation ### Coreference resolution This is the task of identifying if and when two words refer to the same entity. The most common example is determining the person or object to which a certain pronoun refers (such as “she” = “Mary”). But it can also identify a metaphor or an idiom in the text (such as an instance in which “bear” isn’t an animal, but a large and hairy person). ### Named entity recognition (NER) [NER](https://www.ibm.com/think/topics/named-entity-recognition) identifies words or phrases as useful entities. NER identifies “London” as a location or “Maria” as a person's name. ### Part-of-speech tagging Also called grammatical tagging, this is the process of determining which part of speech a word or piece of text is, based on its use and context. For example, part-of-speech identifies “make” as a verb in “I can make a paper plane,” and as a noun in “What make of car do you own?” ### Word sense disambiguation This is the selection of a word meaning for a word with multiple possible meanings. This uses a process of semantic [analysis](https://www.ibm.com/consulting/analytics) to examine the word in context. For example, word sense disambiguation helps distinguish the meaning of the verb “make” in “make the grade” (to achieve) versus “make a bet” (to place). Sorting out “I will be merry when I marry Mary” requires a sophisticated NLP system. ## How NLP works NLP works by combining various computational techniques to analyze, understand and generate human language in a way that machines can process. Here is an overview of a typical NLP pipeline and its steps: ### Text preprocessing NLP text preprocessing prepares raw text for analysis by transforming it into a format that machines can more easily understand. It begins with tokenization, which involves splitting the text into smaller units like words, sentences or phrases. This helps break down complex text into manageable parts. Next, lowercasing is applied to standardize the text by converting all characters to lowercase, ensuring that words like "Apple" and "apple" are treated the same. Stop word removal is another common step, where frequently used words like "is" or "the" are filtered out because they don't add significant meaning to the text. [Stemming](https://www.ibm.com/think/topics/stemming) or [lemmatization](https://www.ibm.com/think/topics/stemming-lemmatization) reduces words to their root form (e.g., "running" becomes "run"), making it easier to analyze language by grouping different forms of the same word. Additionally, text cleaning removes unwanted elements such as punctuation, special characters and numbers that may clutter the analysis. After preprocessing, the text is clean, standardized and ready for machine learning models to interpret effectively. ### Feature extraction Feature extraction is the process of converting raw text into numerical representations that machines can analyze and interpret. This involves transforming text into structured data by using NLP techniques like [Bag of Words](https://www.ibm.com/think/topics/bag-of-words) and TF-IDF, which quantify the presence and importance of words in a document. More advanced methods include [word embeddings](https://www.ibm.com/think/topics/embedding) like Word2Vec or GloVe, which represent words as dense vectors in a continuous space, capturing semantic relationships between words. Contextual embeddings further enhance this by considering the context in which words appear, allowing for richer, more nuanced representations. ### Text analysis Text analysis involves interpreting and extracting meaningful information from text data through various computational techniques. This process includes tasks such as part-of-speech (POS) tagging, which identifies grammatical roles of words and named entity recognition (NER), which detects specific entities like names, locations and dates. Dependency parsing analyzes grammatical relationships between words to understand sentence structure, while sentiment analysis determines the emotional tone of the text, assessing whether it is positive, negative or neutral. Topic modeling identifies underlying themes or topics within a text or across a corpus of documents. Natural language understanding (NLU) is a subset of NLP that focuses on analyzing the meaning behind sentences. NLU enables software to find similar meanings in different sentences or to process words that have different meanings. Through these techniques, NLP text analysis transforms unstructured text into insights. ### Model training Processed data is then used to train machine learning models, which learn patterns and relationships within the data. During training, the model adjusts its parameters to minimize errors and improve its performance. Once trained, the model can be used to make predictions or generate outputs on new, unseen data. The effectiveness of NLP modeling is continually refined through evaluation, validation and fine-tuning to enhance accuracy and relevance in real-world applications. Different software environments are useful throughout the said processes. For example, the Natural Language Toolkit (NLTK) is a suite of libraries and programs for English that is written in the Python programming language. It supports text classification, tokenization, stemming, tagging, parsing and semantic reasoning functionalities. TensorFlow is a free and open-source software library for machine learning and AI that can be used to train models for NLP applications. Tutorials and certifications abound for those interested in familiarizing themselves with such tools. ## Challenges of NLP Even state-of-the-art NLP models are not perfect, just as human speech is prone to error. As with any AI technology, NLP comes with potential pitfalls. Human language is filled with ambiguities that make it difficult for programmers to write software that accurately determines the intended meaning of text or voice data. Human language might take years for humans to learn and many never stop learning. But then programmers must teach natural language-powered applications to recognize and understand irregularities so their applications can be accurate and useful. Associated risks might include: ### Biased training As with any AI function, [biased data](https://www.ibm.com/think/topics/ai-bias) used in training will skew the answers. The more diverse the users of an NLP function, the more significant this risk becomes, such as in government services, healthcare and HR interactions. Training datasets scraped from the web, for example, are prone to bias. ### Misinterpretation As in programming, there is a risk of garbage in, garbage out (GIGO). [Speech recognition](https://www.ibm.com/think/topics/speech-recognition), also known as speech-to-text, is the task of reliably converting voice data into text data. But NLP solutions can become confused if spoken input is in an obscure dialect, mumbled, too full of slang, homonyms, incorrect grammar, idioms, fragments, mispronunciations, contractions or recorded with too much background noise. ### New vocabulary New words are continually being invented or imported. The conventions of grammar can evolve or be intentionally broken. In these cases, NLP can either make a best guess or admit it’s unsure and either way, this creates a complication. ### Tone of voice When people speak, their verbal delivery or even body language can give an entirely different meaning than the words alone. Exaggeration for effect, stressing words for importance or sarcasm can be confused by NLP, making the semantic analysis more difficult and less reliable. ## NLP use cases by industry NLP applications can now be found across virtually every industry. ### Finance In financial dealings, nanoseconds might make the difference between success and failure when accessing data, or making trades or deals. NLP can speed the mining of information from financial statements, annual and regulatory reports, news releases or even social media. ### Healthcare New medical insights and breakthroughs can arrive faster than many healthcare professionals can keep up. NLP and AI-based tools can help speed the analysis of health records and medical research papers, making better-informed medical decisions possible, or assisting in the detection or even prevention of medical conditions. ### Insurance NLP can analyze claims to look for patterns that can identify areas of concern and find inefficiencies in claims processing, leading to greater optimization of processing and employee efforts. ### Legal Almost any legal case might require reviewing mounds of paperwork, background information and legal precedent. NLP can help automate legal discovery, assisting in the organization of information, speeding review and making sure that all relevant details are captured for consideration. Link copied [Report Beyond the hype - How AI assistants drive real business value Explore top use cases for leveraging AI assistants, understand the potential impact of Gen AI and automation technology on your business, and learn how to get started. Read the report](https://www.ibm.com/account/reg/signup?formid=urx-53011) ## Resources [AI models Explore IBM Granite® IBM Granite ® is our family of open, performant and trusted AI models, tailored for business and optimized to scale your AI applications. Explore language, code, time series and guardrail options. Meet Granite](https://www.ibm.com/granite) [Explainer Beginner's guide to NLP Discover how natural language processing (NLP) can help you to converse more naturally with computers. Learn NLP](https://developer.ibm.com/learningpaths/get-started-artificial-intelligence/next-steps/beginner-natural-language-processing) [Report IBM is named leader in data science and machine learning Learn why IBM has been recognized as a Leader in the 2025 Gartner® Magic Quadrant™ for data science and machine learning platforms. Read the report](https://www.ibm.com/forms/mkt-53728) [Training Hands-on with generative AI Learn fundamental concepts and build your skills with hands-on labs, courses, guided projects, trials and more. Learn generative AI](https://developer.ibm.com/technologies/generative-ai/?cm_sp=ibmdev-_-developer-_-getstarted-_-genai) Related solutions IBM® watsonx Orchestrate™ Easily design scalable AI assistants and agents, automate repetitive tasks and simplify complex processes with IBM® watsonx Orchestrate™. [Explore watsonx Orchestrate](https://www.ibm.com/products/watsonx-orchestrate) Natural language processing tools and APIs Accelerate the business value of artificial intelligence with a powerful and flexible portfolio of libraries, services and applications. [Explore NLP solutions](https://www.ibm.com/solutions/natural-language-processing) AI consulting and services Reinvent critical workflows and operations by adding AI to maximize experiences, real-time decision-making and business value. [Explore AI services](https://www.ibm.com/consulting/artificial-intelligence) Take the next step Easily design scalable AI assistants and agents, automate repetitive tasks and simplify complex processes with IBM® watsonx Orchestrate™. 1. [Explore watsonx Orchestrate](https://www.ibm.com/products/watsonx-orchestrate) 2. [Explore NLP solutions](https://www.ibm.com/solutions/natural-language-processing)
Readable Markdown
Natural language processing (NLP) is a subfield of computer science and [artificial intelligence (AI)](https://www.ibm.com/think/topics/artificial-intelligence) that uses [machine learning](https://www.ibm.com/think/topics/machine-learning) to enable computers to understand and communicate with human language. NLP enables computers and digital devices to recognize, understand and generate text and speech by combining computational linguistics, the rule-based modeling of human language together with statistical modeling, machine learning and [deep learning](https://www.ibm.com/think/topics/deep-learning). NLP research has helped enable the era of [generative AI](https://www.ibm.com/think/topics/generative-ai), from the communication skills of [large language models](https://www.ibm.com/think/topics/large-language-models) (LLMs) to the ability of image generation models to understand requests. NLP is already part of everyday life for many, powering search engines, prompting [chatbots](https://www.ibm.com/think/topics/chatbots) for customer service with spoken commands, voice-operated GPS systems and question-answering digital assistants on smartphones such as Amazon’s Alexa, Apple’s Siri and Microsoft’s Cortana. NLP also plays a growing role in enterprise solutions that help streamline and automate business operations, increase employee productivity and simplify business processes. NLP makes it easier for humans to communicate and collaborate with machines, by allowing them to do so in the natural human language they use every day. This offers benefits across many industries and applications. - Automation of repetitive tasks - Improved data analysis and insights - Enhanced search - Content generation Automation of repetitive tasks NLP is especially useful in fully or partially [automating tasks](https://www.ibm.com/think/topics/automation) like customer support, data entry and document handling. For example, NLP-powered chatbots can handle routine customer queries, freeing up human agents for more complex issues. In [document processing](https://www.ibm.com/think/topics/document-processing), NLP tools can automatically classify, extract key information and summarize content, reducing the time and errors associated with manual data handling. NLP facilitates language translation, converting text from one language to another while preserving meaning, context and nuances. Improved data analysis NLP enhances data analysis by enabling the extraction of insights from unstructured text data, such as customer reviews, social media posts and news articles. By using [text mining](https://www.ibm.com/think/topics/text-mining-use-cases) techniques, NLP can identify patterns, trends and sentiments that are not immediately obvious in large datasets. Sentiment analysis enables the [extraction of subjective qualities](https://www.ibm.com/think/topics/sentiment-analysis), attitudes, emotions, sarcasm, confusion or suspicion from text. This is often used for routing communications to the system or the person most likely to make the next response. This allows businesses to better understand customer preferences, market conditions and public opinion. NLP tools can also perform categorization and summarization of vast amounts of text, making it easier for analysts to identify key information and make data-driven decisions more efficiently. Enhanced search NLP benefits search by enabling systems to understand the intent behind user queries, providing more accurate and contextually relevant results. Instead of relying solely on keyword matching, NLP-powered search engines analyze the meaning of words and phrases, making it easier to find information even when queries are vague or complex. This improves user experience, whether in web searches, document retrieval or enterprise data systems. Powerful content generation NLP powers advanced language models to [create human-like text](https://www.ibm.com/think/topics/text-generation) for various purposes. Pre-trained models, such as GPT-4, can generate articles, reports, marketing copy, product descriptions and even creative writing based on prompts provided by users. NLP-powered tools can also assist in automating tasks like drafting emails, writing social media posts or legal documentation. By understanding context, tone and style, NLP sees to it that the generated content is coherent, relevant and aligned with the intended message, saving time and effort in content creation while maintaining quality. NLP combines the power of computational linguistics together with [machine learning algorithms](https://www.ibm.com/think/topics/machine-learning-algorithms) and deep learning. Computational linguistics uses data science to analyze language and speech. It includes two main types of analysis: syntactical analysis and semantical analysis. Syntactical analysis determines the meaning of a word, phrase or sentence by parsing the syntax of the words and applying preprogrammed rules of grammar. Semantical analysis uses the syntactic output to draw meaning from the words and interpret their meaning within the sentence structure. The parsing of words can take one of two forms. Dependency parsing looks at the relationships between words, such as identifying nouns and verbs, while constituency parsing then builds a parse tree (or syntax tree): a rooted and ordered representation of the syntactic structure of the sentence or string of words. The resulting parse trees underly the functions of language translators and speech recognition. Ideally, this analysis makes the output either text or speech understandable to both NLP models and people. [Self-supervised learning (SSL)](https://www.ibm.com/think/topics/self-supervised-learning) in particular is useful for supporting NLP because NLP requires large amounts of labeled data to train AI models. Because these labeled datasets require time-consuming annotation, a process involving manual labeling by humans, gathering sufficient data can be prohibitively difficult. Self-supervised approaches can be more time-effective and cost-effective, as they replace some or all manually labeled training data. Three different approaches to NLP include: Rules-based NLP The earliest NLP applications were simple if-then decision trees, requiring preprogrammed rules. They are only able to provide answers in response to specific prompts, such as the original version of Moviefone, which had rudimentary natural language generation (NLG) capabilities. Because there is no machine learning or AI capability in rules-based NLP, this function is highly limited and not scalable. Statistical NLP Developed later, statistical NLP automatically extracts, classifies and labels elements of text and voice data and then assigns a statistical likelihood to each possible meaning of those elements. This relies on machine learning, enabling a sophisticated breakdown of linguistics such as part-of-speech tagging. Statistical NLP introduced the essential technique of mapping language elements, such as words and grammatical rules to a vector representation so that language can be modeled by using mathematical (statistical) methods, including regression or Markov models. This informed early NLP developments such as spellcheckers and T9 texting (Text on 9 keys, to be used on Touch-Tone telephones). Deep learning NLP Recently, deep learning models have become the dominant mode of NLP, by using huge volumes of raw, [unstructured](https://www.ibm.com/think/topics/text-mining) data both text and voice to become ever more accurate. Deep learning can be viewed as a further evolution of statistical NLP, with the difference that it uses [neural network](https://www.ibm.com/think/topics/neural-networks) models. There are several subcategories of models: - *Sequence-to-Sequence* (seq2seq) models: Based [on recurrent neural networks (RNN)](https://www.ibm.com/think/topics/recurrent-neural-networks), they have mostly been used for machine translation by converting a phrase from one domain (such as the German language) into the phrase of another domain (such as English). - *Transformer models*: They use [tokenization](https://www.ibm.com/docs/en/watsonx/saas?topic=solutions-tokens) of language (the position of each token words or subwords) and self-attention (capturing dependencies and relationships) to calculate the relation of different language parts to one another. [Transformer models](https://www.ibm.com/think/topics/transformer-model) can be efficiently trained by using [self-supervised learning](https://www.ibm.com/think/topics/self-supervised-learning) on massive text databases. A landmark in [transformer models](https://www.ibm.com/think/topics/transformer-model) was Google’s bidirectional encoder representations from transformers (BERT), which became and remains the basis of how Google’s search engine works. - *Autoregressive models*: This type of transformer model is trained specifically to predict the next word in a sequence, which represents a huge leap forward in the ability to generate text. Examples of autoregressive LLMs include GPT, [Llama](https://www.ibm.com/think/topics/llama-2), Claude and the open-source Mistral. - *Foundation models*: Prebuilt and curated foundation models can speed the launching of an NLP effort and boost trust in its operation. For example, the [IBM® Granite™](https://www.ibm.com/think/news/granite-foundation-models) foundation models are widely applicable across industries. They support NLP tasks including content generation and insight extraction. Additionally, they facilitate retrieval-augmented generation, a framework for improving the quality of response by linking the model to external sources of knowledge. The models also perform named entity recognition which involves identifying and extracting key information in a text. Several NLP tasks typically help process human text and voice data in ways that help the computer make sense of what it’s ingesting. Some of these tasks include: - Coreference resolution - Named entity recognition - Part-of-speech tagging - Word sense disambiguation Coreference resolution This is the task of identifying if and when two words refer to the same entity. The most common example is determining the person or object to which a certain pronoun refers (such as “she” = “Mary”). But it can also identify a metaphor or an idiom in the text (such as an instance in which “bear” isn’t an animal, but a large and hairy person). Named entity recognition (NER) [NER](https://www.ibm.com/think/topics/named-entity-recognition) identifies words or phrases as useful entities. NER identifies “London” as a location or “Maria” as a person's name. Part-of-speech tagging Also called grammatical tagging, this is the process of determining which part of speech a word or piece of text is, based on its use and context. For example, part-of-speech identifies “make” as a verb in “I can make a paper plane,” and as a noun in “What make of car do you own?” Word sense disambiguation This is the selection of a word meaning for a word with multiple possible meanings. This uses a process of semantic [analysis](https://www.ibm.com/consulting/analytics) to examine the word in context. For example, word sense disambiguation helps distinguish the meaning of the verb “make” in “make the grade” (to achieve) versus “make a bet” (to place). Sorting out “I will be merry when I marry Mary” requires a sophisticated NLP system. NLP works by combining various computational techniques to analyze, understand and generate human language in a way that machines can process. Here is an overview of a typical NLP pipeline and its steps: Text preprocessing NLP text preprocessing prepares raw text for analysis by transforming it into a format that machines can more easily understand. It begins with tokenization, which involves splitting the text into smaller units like words, sentences or phrases. This helps break down complex text into manageable parts. Next, lowercasing is applied to standardize the text by converting all characters to lowercase, ensuring that words like "Apple" and "apple" are treated the same. Stop word removal is another common step, where frequently used words like "is" or "the" are filtered out because they don't add significant meaning to the text. [Stemming](https://www.ibm.com/think/topics/stemming) or [lemmatization](https://www.ibm.com/think/topics/stemming-lemmatization) reduces words to their root form (e.g., "running" becomes "run"), making it easier to analyze language by grouping different forms of the same word. Additionally, text cleaning removes unwanted elements such as punctuation, special characters and numbers that may clutter the analysis. After preprocessing, the text is clean, standardized and ready for machine learning models to interpret effectively. Feature extraction Feature extraction is the process of converting raw text into numerical representations that machines can analyze and interpret. This involves transforming text into structured data by using NLP techniques like [Bag of Words](https://www.ibm.com/think/topics/bag-of-words) and TF-IDF, which quantify the presence and importance of words in a document. More advanced methods include [word embeddings](https://www.ibm.com/think/topics/embedding) like Word2Vec or GloVe, which represent words as dense vectors in a continuous space, capturing semantic relationships between words. Contextual embeddings further enhance this by considering the context in which words appear, allowing for richer, more nuanced representations. Text analysis Text analysis involves interpreting and extracting meaningful information from text data through various computational techniques. This process includes tasks such as part-of-speech (POS) tagging, which identifies grammatical roles of words and named entity recognition (NER), which detects specific entities like names, locations and dates. Dependency parsing analyzes grammatical relationships between words to understand sentence structure, while sentiment analysis determines the emotional tone of the text, assessing whether it is positive, negative or neutral. Topic modeling identifies underlying themes or topics within a text or across a corpus of documents. Natural language understanding (NLU) is a subset of NLP that focuses on analyzing the meaning behind sentences. NLU enables software to find similar meanings in different sentences or to process words that have different meanings. Through these techniques, NLP text analysis transforms unstructured text into insights. Model training Processed data is then used to train machine learning models, which learn patterns and relationships within the data. During training, the model adjusts its parameters to minimize errors and improve its performance. Once trained, the model can be used to make predictions or generate outputs on new, unseen data. The effectiveness of NLP modeling is continually refined through evaluation, validation and fine-tuning to enhance accuracy and relevance in real-world applications. Different software environments are useful throughout the said processes. For example, the Natural Language Toolkit (NLTK) is a suite of libraries and programs for English that is written in the Python programming language. It supports text classification, tokenization, stemming, tagging, parsing and semantic reasoning functionalities. TensorFlow is a free and open-source software library for machine learning and AI that can be used to train models for NLP applications. Tutorials and certifications abound for those interested in familiarizing themselves with such tools. Even state-of-the-art NLP models are not perfect, just as human speech is prone to error. As with any AI technology, NLP comes with potential pitfalls. Human language is filled with ambiguities that make it difficult for programmers to write software that accurately determines the intended meaning of text or voice data. Human language might take years for humans to learn and many never stop learning. But then programmers must teach natural language-powered applications to recognize and understand irregularities so their applications can be accurate and useful. Associated risks might include: Biased training As with any AI function, [biased data](https://www.ibm.com/think/topics/ai-bias) used in training will skew the answers. The more diverse the users of an NLP function, the more significant this risk becomes, such as in government services, healthcare and HR interactions. Training datasets scraped from the web, for example, are prone to bias. Misinterpretation As in programming, there is a risk of garbage in, garbage out (GIGO). [Speech recognition](https://www.ibm.com/think/topics/speech-recognition), also known as speech-to-text, is the task of reliably converting voice data into text data. But NLP solutions can become confused if spoken input is in an obscure dialect, mumbled, too full of slang, homonyms, incorrect grammar, idioms, fragments, mispronunciations, contractions or recorded with too much background noise. New vocabulary New words are continually being invented or imported. The conventions of grammar can evolve or be intentionally broken. In these cases, NLP can either make a best guess or admit it’s unsure and either way, this creates a complication. Tone of voice When people speak, their verbal delivery or even body language can give an entirely different meaning than the words alone. Exaggeration for effect, stressing words for importance or sarcasm can be confused by NLP, making the semantic analysis more difficult and less reliable. NLP applications can now be found across virtually every industry. Finance In financial dealings, nanoseconds might make the difference between success and failure when accessing data, or making trades or deals. NLP can speed the mining of information from financial statements, annual and regulatory reports, news releases or even social media. Healthcare New medical insights and breakthroughs can arrive faster than many healthcare professionals can keep up. NLP and AI-based tools can help speed the analysis of health records and medical research papers, making better-informed medical decisions possible, or assisting in the detection or even prevention of medical conditions. Insurance NLP can analyze claims to look for patterns that can identify areas of concern and find inefficiencies in claims processing, leading to greater optimization of processing and employee efforts. Legal Almost any legal case might require reviewing mounds of paperwork, background information and legal precedent. NLP can help automate legal discovery, assisting in the organization of information, speeding review and making sure that all relevant details are captured for consideration.
Shard39 (laksa)
Root Hash15319107097793762839
Unparsed URLcom,ibm!www,/think/topics/natural-language-processing s443