| import torch.nn as nn |
| from transformers import BertConfig, BertModel, BertTokenizer |
|
|
| from modules.build import LANGUAGE_REGISTRY |
|
|
|
|
| @LANGUAGE_REGISTRY.register() |
| class BERTLanguageEncoder(nn.Module): |
| def __init__(self, cfg, weights="bert-base-uncased", hidden_size=768, |
| num_hidden_layers=4, num_attention_heads=12, type_vocab_size=2): |
| super().__init__() |
| self.tokenizer = BertTokenizer.from_pretrained( |
| weights, do_lower_case=True |
| ) |
| self.bert_config = BertConfig( |
| hidden_size=hidden_size, |
| num_hidden_layers=num_hidden_layers, |
| num_attention_heads=num_attention_heads, |
| type_vocab_size=type_vocab_size |
| ) |
| self.model = BertModel.from_pretrained( |
| weights, config=self.bert_config |
| ) |
|
|
| def forward(self, txt_ids, txt_masks, **kwargs): |
| return self.model(txt_ids, txt_masks).last_hidden_state |
|
|
|
|