Introduction
-
- Problems in NLP:
-
- Question Answering (QA)
- Information Extraction (IE)
- Sentiment Analysis
- Machine Translation (MT)
- Spam Detection
- Parts-of-Speech (POS) Tagging
- Named Entity Recognition (NER)
- Coreference Resolution
- Word Sense Disambiguation (WSD)
- Parsing
- Paraphrasing
- Summarization
- Dialog
-
- (mostly) Solved Problems in NLP:
-
- Spam Detection
- Parts-of-Speech (POS) Tagging
- Named Entity Recognition (NER)
-
- Within-Reach Problems:
-
- Sentiment Analysis
- Coreference Resolution
- Word Sense Disambiguation (WSD)
- Parsing
- Machine Translation (MT)
- Information Extraction (IE)
-
- Open Problems in NLP:
-
- Question Answering (QA)
- Paraphrasing
- Summarization
- Dialog
-
- Issues in NLP (why nlp is hard?):
-
- Non-Standard English: “Great Job @ahmed_badary! I luv u 2!! were SOO PROUD of dis.”
- Segmentation Issues: “New York-New Haven” vs “New-York New-Haven”
- Idioms: “dark horse”, “getting cold feet”, “losing face”
- Neologisms: “unfriend”, “retweet”, “google”, “bromance”
- World Knowledge: “Ahmed and Zach are brothers”, “Ahmed and Zach are fathers”
- Tricky Entity Names: “Where is Life of Pie playing tonight?”, “Let it be was a hit song!”
-
- Tools we need for NLP:
-
- Knowledge about Language.
- Knowledge about the World.
- A way to combine knowledge sources.
-
- Methods:
- In general we need to construct Probabilistic Models built from language data.
- We do so by using rough text features.
All the names models, methods, and tools mentioned above will be introduced later as you progress in the text.
Definitions
-
- K-Nearest-Neighbors:
-
- Complexity:
-
- Training: \(\:\:\:\:\mathcal{O}(1)\)
- Predict: \(\:\:\:\:\mathcal{O}(N)\)
-
Asynchronous:
-
Asynchronous:
-
Asynchronous:
-
Asynchronous:
-
Asynchronous:
-
Asynchronous:
- Asynchronous:
Metrics
-
- L1 Distance:
- \[d_1(I_1, I_2) = \sum_p{\|I_1^p - I_2^p\|}\]
- Pixel-wise absolute value differences.
-
- L2 Distance:
- \[d_2 (I_1, I_2) = \sqrt{\sum_{p} \left( I^p_1 - I^p_2 \right)^2}\]
-
- L1 vs. L2:
- The L2 distance penalizes errors (pixel differences) much more than the L1 metric does.
The L2 distnace will be small iff there are man small differences in the two vectors but will explode if there is even one big difference between them. - Another difference we highlight is that the L1 distance is dependent on the corrdinate system frame, while the L2 distance is coordinate-invariant.
-
Asynchronous:
-
Asynchronous:
-
Asynchronous:
-
Asynchronous:
- Asynchronous: