File size: 738 Bytes
fbe2800
 
 
32e707d
 
 
 
 
fbe2800
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
471fdbf
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
---
language:
- en
license: mit
datasets:
- Trelis/tiny-shakespeare
pipeline_tag: text-generation
library_name: transformers
---
# Decoder Language Model
Ein kleiner autoregressiver Decoder-only Transformer, trainiert auf Tiny Shakespeare.

## Architektur
- d_model=128, num_layers=2, nhead=4
- ~500k Parameter

## Metriken
- Loss (Train): 0.6342
- Perplexity (Train): 1.8854

## Laden
```python
from transformers import GPT2Tokenizer
import torch
from model import DecoderLanguageModel

tokenizer = GPT2Tokenizer.from_pretrained("ahmadisakina/decoder-language-model")
model = DecoderLanguageModel(vocab_size=tokenizer.vocab_size, d_model=128, nhead=4, num_layers=2)
model.load_state_dict(torch.load("pytorch_model.bin"))
model.eval()
```