Spaces:
Running
Running
**Architectural Understanding** | |
1. **Define** deep learning as neural networks with multiple hidden layers that enable hierarchical feature learning (not just "3+ layers" but *why* depth matters for representation learning) | |
2. **Explain** why "deep" matters: each layer learns increasingly complex features | |
3. **Compare** how deep learning is different from machine learning, and be able to identify other deep learning methods | |
**When to Use Deep Learning** | |
4\. **Choose** deep learning when you have: large datasets, unstructured data (images/text), complex patterns. Also understand why pretrained models work and when to fine-tune vs. feature extraction | |
5\. **Avoid** deep learning when you need: explainable results, small datasets, or simple patterns | |
**Understanding BERT** | |
6\. **Explain** what BERT does: understands word meaning based on context (unlike older methods) | |
7\. **Understand** BERT training: pre-training with a massive dataset, masked language modeling, bidirectional learning and the transformer framework | |
8\. **Recognize** why pretrained models save time and work better than training from scratch | |
**Practical Implementation and Evaluation** | |
10\. **Implement** sentiment analysis using pretrained BERT via Hugging Face transformers | |
11\. **Evaluate** model performance using appropriate metrics for classification tasks | |
12\. **Interpret** the model's confidence scores and predictions | |
**Notes** | |
What is deep learning? | |
Video tutorial: [Link](https://www.youtube.com/watch?v=q6kJ71tEYqM) | |
- Previously we learned what machine learning is | |
- Deep learning is a subset of machine learning | |
- A subfield of AI is ML \-\> Neural Network \-\> Deep Learning | |
- More than three layers of neural network is considered deep neural network \-\> deep learning | |
- Can ingest unstructured data and determine \-\> different from supervised learning \-\> unsupervised learning | |
When to use | |
Video: [Link](https://www.youtube.com/watch?v=o3bWqPdWJ88) | |
- Unstructured data, like image, video, text | |
- High volumn of data \-\> deep learning will give you better result | |
- Complexity of feature \-\> complicated features \-\> deep learning | |
- Interpretability (important) | |
- Industries like healthcare and finance require high interpretability, which is better answered by statistical ML | |
- Deep learning’s complex neural networks makes it hard to interpret | |
BERT | |
- Google search is powered by BERT (bidirectional encoder representations from transformers) | |
- BERT base, BERT large | |
- If you have two homes, how can you say if there are similar | |
- For an object, if you can derive and compare features and compare their similarities…take all the numbers and create vectors and compare the vectors, you can then compare | |
- Generate feature vector for these words \-\> compare feature vector/word embedding | |
- How to generate word embeddings | |
- Word to vector (word2vec) | |
- Issues with word2vec \-\> you need a model that can generate contextualized meaning of words \-\> this is what BERT allows you to do | |
Pretrained BERT for sentiment analysis | |
- Download and install Transformer from huggingface | |
- Install and import dependencies | |
- Instantiat model \- bert-base-multilingual-uncased-sentiment | |
- Perform sentiment scoring | |
- Encode and calculate sentiment | |