Update README.md
Browse files
README.md
CHANGED
|
@@ -105,7 +105,14 @@ This is `BEE-spoke-data/smol_llama-220M-GQA` fine-tuned for code generation on:
|
|
| 105 |
|
| 106 |
This model (and the base model) were both trained using ctx length 2048.
|
| 107 |
|
| 108 |
-
|
| 109 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 110 |
|
| 111 |
---
|
|
|
|
| 105 |
|
| 106 |
This model (and the base model) were both trained using ctx length 2048.
|
| 107 |
|
| 108 |
+
## examples
|
| 109 |
|
| 110 |
+
> Example script for inference testing: [here](https://gist.github.com/pszemraj/c7738f664a64b935a558974d23a7aa8c)
|
| 111 |
+
|
| 112 |
+
It has its limitations at 220M, but seems decent for single-line or docstring generation, and/or being used for speculative decoding for such purposes.
|
| 113 |
+
|
| 114 |
+
|
| 115 |
+
|
| 116 |
+

|
| 117 |
|
| 118 |
---
|