MobileLLM 60M (Replication)

This is a replicated version of the MobileLLM 60M model, based on the paper MobileLLM: Optimizing Sub-billion Parameter Language Models for On-Device Use Cases.

  • Model Size: 60M parameters
  • Architecture: Llama-based (Deep & Thin)
  • Status: Research / Testing
Downloads last month
1
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Space using Yangyang1205/MobileLLM-60M 1