Natural Language Processing Fundamentals
Chapter 1: NLP Concepts and Advanced Text Processing
The Natural Language Processing Pipeline
Advanced Tokenization Methods
Stemming and Lemmatization Compared
Handling Noise in Text Data
Advanced Stop Word Customization
Text Normalization Techniques
Hands-on Practical: Building Preprocessing Pipelines
Chapter 2: Feature Engineering for Text Representation
From Bag-of-Words to TF-IDF
Calculating TF-IDF Scores
Using N-grams to Capture Context
Introduction to Feature Hashing
Dimensionality Reduction for Text Features
Comparing Different Text Representation Methods
Hands-on Practical: Generating Text Features
Chapter 3: Supervised Learning for Text Classification
Classification Algorithms Review
Applying Classifiers to Text Data
Model Evaluation Metrics for Classification
Cross-Validation Strategies
Hyperparameter Tuning for Text Models
Addressing Imbalanced Datasets
Practice: Building a Text Classifier
Chapter 4: Understanding Sequential Data with Embeddings
Limitations of Frequency-Based Models
Introduction to Distributional Semantics
Word2Vec: CBOW and Skip-gram Architectures
GloVe: Global Vectors for Word Representation
Visualizing Word Embeddings
Using Pre-trained Word Embedding Models
Hands-on Practical: Working with Word Embeddings
Chapter 5: Introduction to Sequence Models for NLP
The Need for Sequence Awareness
Recurrent Neural Network (RNN) Basics
Understanding the Vanishing Gradient Problem
Long Short-Term Memory (LSTM) Networks
Gated Recurrent Units (GRUs)
Applying Sequence Models to Text
Hands-on Practical: Building a Simple Sequence Model