Instructions to use google/switch-large-128 with libraries, inference providers, notebooks, and local apps. Follow these links to get started.
- Libraries
- Transformers
How to use google/switch-large-128 with Transformers:
# Load model directly from transformers import AutoTokenizer, AutoModelForSeq2SeqLM tokenizer = AutoTokenizer.from_pretrained("google/switch-large-128") model = AutoModelForSeq2SeqLM.from_pretrained("google/switch-large-128") - Notebooks
- Google Colab
- Kaggle
Update config.json
#7
by Karim-Gamal - opened
- config.json +1 -0
config.json
CHANGED
|
@@ -14,6 +14,7 @@
|
|
| 14 |
"expert_capacity": 64,
|
| 15 |
"feed_forward_proj": "relu",
|
| 16 |
"initializer_factor": 1.0,
|
|
|
|
| 17 |
"is_encoder_decoder": true,
|
| 18 |
"is_gated_act": false,
|
| 19 |
"layer_norm_epsilon": 1e-06,
|
|
|
|
| 14 |
"expert_capacity": 64,
|
| 15 |
"feed_forward_proj": "relu",
|
| 16 |
"initializer_factor": 1.0,
|
| 17 |
+
"decoder_start_token_id": 0,
|
| 18 |
"is_encoder_decoder": true,
|
| 19 |
"is_gated_act": false,
|
| 20 |
"layer_norm_epsilon": 1e-06,
|