| description | General Concepts of Sequence Models |
|---|
In the context of text processing (e.g: Natural Language Processing NLP)
| Symbol | Description |
|---|---|
The tth word in the input sequence |
|
The tth word in the output sequence |
|
The tth word in the ith input sequence |
|
The tth word in the ith output sequence |
|
The length of the ith input sequence |
|
The length of the ith output sequence |
A way to represent words so we can treat with them easily
Let's say that we have a dictionary that consists of 10 words (π€) and the words of the dictionary are:
- Car, Pen, Girl, Berry, Apple, Likes, The, And, Boy, Book.
Our
So we can represent this sequence like the following π
Car -0) β 0 β β 0 β β 0 β β 0 β β 0 β β 0 β
Pen -1) | 0 | | 0 | | 0 | | 0 | | 0 | | 0 |
Girl -2) | 0 | | 1 | | 0 | | 0 | | 0 | | 0 |
Berry -3) | 0 | | 0 | | 0 | | 0 | | 0 | | 1 |
Apple -4) | 0 | | 0 | | 0 | | 1 | | 0 | | 0 |
Likes -5) | 0 | | 0 | | 1 | | 0 | | 0 | | 0 |
The -6) | 1 | | 0 | | 0 | | 0 | | 0 | | 0 |
And -7) | 0 | | 0 | | 0 | | 0 | | 1 | | 0 |
Boy -8) | 0 | | 0 | | 0 | | 0 | | 0 | | 0 |
Book -9) β 0 β β 0 β β 0 β β 0 β β 0 β β 0 βBy representing sequences in this way we can feed out data to neural networks β¨
- If our dictionary consists of 10,000 words so each vector will be 10,000 dimensional π€
- This representation can not capture semantic features π