base_model, tags, license, language
base_model tags license language
unsloth/gemma-3-270m-it
text-generation-inference
transformers
unsloth
gemma3_text
apache-2.0
en

license: apache-2.0 base_model: google/gemma-3-270m-it library_name: transformers tags:

  • unsloth
  • gemma-3
  • emoji
  • translation
  • multilingual
  • gguf
  • edge-ai datasets:
  • custom-curated-emoji-distillation

🚀 Emojify-300M (Gemma-3-270M Fine-tuned)

Emojify-300M is a hyper-compact, specialized Large Language Model (LLM) designed for semantic text-to-emoji distillation. Based on the Gemma-3-270M-IT architecture, this model is optimized for edge computing and ultra-low latency applications.

🛠 Technical Specifications

  • Architecture: Gemma-3 (270M parameters)
  • Fine-tuning Method: LoRA (Low-Rank Adaptation) via Unsloth
  • Context Window: 2048 tokens
  • Language Support: Native Multilingual

📈 Performance & Benchmarks (Local CPU)

Tested on consumer-grade hardware (e.g., laptop CPU):

  • Prompt Evaluation: ~210 tokens/s
  • Token Generation (Eval): ~48 tokens/s
  • Total Latency: < 500ms (near-instant response)

🎯 Key Features

  • Zero-Shot Cross-Lingual Transfer: Leveraging Gemma 3's robust base weights, the model accurately processes languages not explicitly present in the fine-tuning set.
  • Noise Suppression: Specifically trained to inhibit conversational filler and "hallucinated" text, focusing strictly on relevant emoji output.
  • EOS Stability: Fine-tuned to respect End-of-Sequence (EOS) tokens, preventing the common "looping" behavior seen in smaller models.
Description
Model synced from source: razy101/emojify-300m
Readme 68 KiB
Languages
Jinja 100%