Instructions to use DrewG/AAVE_PoS-Tagger with libraries, inference providers, notebooks, and local apps. Follow these links to get started.
- Libraries
- Transformers
How to use DrewG/AAVE_PoS-Tagger with Transformers:
# Use a pipeline as a high-level helper from transformers import pipeline pipe = pipeline("token-classification", model="DrewG/AAVE_PoS-Tagger")# Load model directly from transformers import AutoTokenizer, AutoModelForTokenClassification tokenizer = AutoTokenizer.from_pretrained("DrewG/AAVE_PoS-Tagger") model = AutoModelForTokenClassification.from_pretrained("DrewG/AAVE_PoS-Tagger") - Notebooks
- Google Colab
- Kaggle
bert-finetuned-AAVE-PoS
This model is a version of bert-base-cased fine-tuned on a dataset of African American Vernacular English (AAVE) which was published alongside Jørgensen et al. 2016. It achieves the following results on the evaluation set:
- Loss: 0.2582
- Precision: 0.8632
- Recall: 0.8730
- F1: 0.8681
- Accuracy: 0.9356
Model description
More information needed
Intended uses & limitations
This model is intended to help close the gap in part-of-speech tagging performance between Standard American English (SAE) and African American English (AAVE) which differ liguistically in many well-documented ways. It was fine-tuned on data gathered from Twitter, and is thus ingrained with what linguists call 'register bias'.
Training and evaluation data
Code hosted at GitHub.
Training procedure
Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3 (this amount of data overfits on 3+)
Training results
| Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy |
|---|---|---|---|---|---|---|---|
| No log | 1.0 | 223 | 0.2982 | 0.8196 | 0.8350 | 0.8272 | 0.9216 |
| No log | 2.0 | 446 | 0.2625 | 0.8599 | 0.8680 | 0.8640 | 0.9326 |
| 0.4647 | 3.0 | 669 | 0.2582 | 0.8632 | 0.8730 | 0.8681 | 0.9356 |
Framework versions
- Transformers 4.29.2
- Pytorch 1.13.1+cpu
- Datasets 2.12.0
- Tokenizers 0.13.3
- Downloads last month
- 3