BERT(Bidirectional Encoder Representations from Transformers)是NLP领域的里程碑模型。
模型结构
graph TB
A[输入] --> B[Token Embedding]
A --> C[Segment Embedding]
A --> D[Position Embedding]
B --> E[叠加]
C --> E
D --> E
E --> F[BERT Encoder层×12]
F --> G[输出]
安装与加载
1 2 3 4 5
from transformers import BertTokenizer, BertForSequenceClassification import torch
tokenizer = BertTokenizer.from_pretrained('bert-base-chinese') model = BertForSequenceClassification.from_pretrained('bert-base-chinese', num_labels=2)