Natural Language Processing
46 learners
Building an NLP Pipeline with spaCy for Token Classification
Kickstart your journey into token classification by setting up an efficient NLP pipeline, learning about tokenization, POS tagging, and lemmatization with spaCy.
NLTK
Python
spaCy
5 lessons
25 practices
3 hours
Text Data Collection and Preparation
Lessons and practices
Counting Unique Categories in Reuters Dataset
Explore 'Tea' Category in Reuters Corpus
Fetch Text and Categories for 'Coffee' in Reuters Corpus
Exploring the 'Gas' Category in Reuters Corpus
Exploring Reuters Corpus by Category
Changing the String for Tokenization
Tokenize Sentences with Missing Code
Tokenizing First Reuters Document with spaCy
Calculating Unique Tokens in Document
Tokenizing Multiple Reuters Documents with spaCy
Filter Non-Alphabetic Stopword Tokens
Identifying Out-of-Vocabulary and Digital Tokens
Counting Stop Word Tokens
Identifying Token Capitalization in Text
Filtering Tokens Using a Simple Pipeline
Change the Sentence for Lemmatization
Lemmatizing Reuters Dataset with spaCy
Lemmatization on Reuters Dataset with spaCy
Integrating Lemmatization into Text Processing Pipeline
Lemmatization with spaCy on the Reuters Dataset
Refining Output Format of POS Tagging
POS tagging on a Real-world Text Document
Analyzing Verb Usage in Reuters News
Frequency Analysis on Adjectives Using POS Tagging
Exploring Word Usages with POS Tagging
Meet Cosmo:
The smartest AI guide in the universe
Our built-in AI guide and tutor, Cosmo, prompts you with challenges that are built just for you and unblocks you when you get stuck.

Join the 1M+ learners on CodeSignal
Be a part of our community of 1M+ users who develop and demonstrate their skills on CodeSignal