NanoSocrates/docs/RESOURCES.md

109 lines
2.0 KiB
Markdown
Raw Normal View History

2025-09-18 20:24:11 +02:00
# Resources
2025-09-18 20:24:11 +02:00
## Byte-Pair Encoding (BPE)
### Overview
2025-09-17 12:51:14 +02:00
Byte-Pair Encoding (BPE) is a simple but powerful text compression and tokenization algorithm.
Originally introduced as a data compression method, it has been widely adopted in **Natural Language Processing (NLP)** to build subword vocabularies for models such as GPT and BERT.
---
2025-09-18 20:24:11 +02:00
### Key Idea
2025-09-17 12:51:14 +02:00
BPE works by iteratively replacing the most frequent pair of symbols (initially characters) with a new symbol.
Over time, frequent character sequences (e.g., common morphemes, prefixes, suffixes) are merged into single tokens.
---
2025-09-18 20:24:11 +02:00
### Algorithm Steps
2025-09-17 12:51:14 +02:00
1. **Initialization**
- Treat each character of the input text as a token.
2025-09-17 12:51:14 +02:00
2. **Find Frequent Pairs**
- Count all adjacent token pairs in the sequence.
2025-09-17 12:51:14 +02:00
3. **Merge Most Frequent Pair**
- Replace the most frequent pair with a new symbol not used in the text.
2025-09-17 12:51:14 +02:00
4. **Repeat**
- Continue until no frequent pairs remain or a desired vocabulary size is reached.
---
2025-09-18 20:24:11 +02:00
### Example
Suppose the data to be encoded is:
2025-09-17 12:51:14 +02:00
```text
aaabdaaabac
```
2025-09-18 20:24:11 +02:00
#### Step 1: Merge `"aa"`
2025-09-17 12:51:14 +02:00
Most frequent pair: `"aa"` → replace with `"Z"`
2025-09-17 12:51:14 +02:00
```text
ZabdZabac
Z = aa
```
---
2025-09-18 20:24:11 +02:00
#### Step 2: Merge `"ab"`
2025-09-17 12:51:14 +02:00
Most frequent pair: `"ab"` → replace with `"Y"`
2025-09-17 12:51:14 +02:00
```text
ZYdZYac
Y = ab
Z = aa
```
---
2025-09-18 20:24:11 +02:00
#### Step 3: Merge `"ZY"`
2025-09-17 12:51:14 +02:00
Most frequent pair: `"ZY"` → replace with `"X"`
2025-09-17 12:51:14 +02:00
```text
XdXac
X = ZY
Y = ab
Z = aa
```
---
At this point, no pairs occur more than once, so the process stops.
---
2025-09-18 20:24:11 +02:00
### Decompression
2025-09-17 12:51:14 +02:00
To recover the original data, replacements are applied in **reverse order**:
2025-09-17 12:51:14 +02:00
```text
XdXac
→ ZYdZYac
→ ZabdZabac
→ aaabdaaabac
```
---
2025-09-18 20:24:11 +02:00
### Advantages
2025-09-17 12:51:14 +02:00
- **Efficient vocabulary building**: reduces the need for massive word lists.
- **Handles rare words**: breaks them into meaningful subword units.
- **Balances character- and word-level tokenization**.
---
2025-09-18 20:24:11 +02:00
### Limitations
2025-09-17 12:51:14 +02:00
- Does not consider linguistic meaning—merges are frequency-based.
- May create tokens that are not linguistically natural.
- Vocabulary is fixed after training.