| --- |
| tags: |
| - lora |
| - peft |
| - adapters |
| - mistral |
| - biomedical |
| library_name: peft |
| pipeline_tag: text-classification |
| license: apache-2.0 |
| base_model: mistralai/Mistral-7B-Instruct-v0.3 |
| model-index: |
| - name: MetappuccinoLLModel v1.0.0 |
| results: [] |
| --- |
| |
| # MetappuccinoLLModel — Per-category LoRA adapters for SRA metadata extraction |
|
|
| **LoRA adapters (one folder per category)** trained for **SRA metadata extraction and inference** in the [Metappuccino project](https://github.com/chumphati/Metappuccino). These adapters not general-purpose dialogue models. |
|
|
| **Important**: Base model weights are **not included**. Also download the official base model: `mistralai/Mistral-7B-Instruct-v0.3` to use Metappuccino. |
|
|
|
|
| ### Version |
|
|
| v1.0.0 |
|
|
| ### Quickstart |
|
|
| Download for Metappuccino use: |
| ```python |
| from huggingface_hub import snapshot_download |
| |
| snapshot_download( |
| repo_id="chumphati/MetappuccinoLLModel", |
| local_dir="<OUT_DIR_URL>/MetappuccinoLLModel", #path to the output directory |
| resume_download=True, |
| max_workers=4 |
| ) |
| ``` |
|
|
| ### Hyperparameters |
|
|
| All information about the hyperparameters is provided for each adapter in its respective folder, in the adapter_config.json files. |
| |
| ### How to cite |
| |
| If you use this repository in your work, please cite: |
| |
| |
| |
| Related tool: Metappuccino — https://github.com/chumphati/Metappuccino |
| |