初始化项目,由ModelHub XC社区提供模型
Model: KurmaAI/AQUA-1B Source: Original Platform
This commit is contained in:
49
.gitattributes
vendored
Normal file
49
.gitattributes
vendored
Normal file
@@ -0,0 +1,49 @@
|
|||||||
|
*.7z filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.arrow filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.bin filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.bin.* filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.bz2 filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.ftz filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.gz filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.h5 filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.joblib filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.lfs.* filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.model filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.msgpack filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.onnx filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.ot filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.parquet filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.pb filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.pt filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.pth filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.rar filter=lfs diff=lfs merge=lfs -text
|
||||||
|
saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.tar.* filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.tflite filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.tgz filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.xz filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.zip filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.zstandard filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.tfevents* filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.db* filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.ark* filter=lfs diff=lfs merge=lfs -text
|
||||||
|
**/*ckpt*data* filter=lfs diff=lfs merge=lfs -text
|
||||||
|
**/*ckpt*.meta filter=lfs diff=lfs merge=lfs -text
|
||||||
|
**/*ckpt*.index filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.safetensors filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.ckpt filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.gguf* filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.ggml filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.llamafile* filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.pt2 filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.mlmodel filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.npy filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.npz filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.pickle filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.pkl filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.tar filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.wasm filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.zst filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
||||||
|
|
||||||
|
tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
||||||
BIN
AQUA-1B.png
Normal file
BIN
AQUA-1B.png
Normal file
Binary file not shown.
|
After Width: | Height: | Size: 2.0 MiB |
133
README.md
Normal file
133
README.md
Normal file
@@ -0,0 +1,133 @@
|
|||||||
|
---
|
||||||
|
license: apache-2.0
|
||||||
|
tags:
|
||||||
|
- text-generation-inference
|
||||||
|
- transformers
|
||||||
|
language:
|
||||||
|
- en
|
||||||
|
base_model:
|
||||||
|
- google/gemma-3-1b-it
|
||||||
|
pipeline_tag: text-generation
|
||||||
|
---
|
||||||
|
|
||||||
|
<p align="center">
|
||||||
|
<img src="./AQUA-1B.png" alt="AQUA-1B" width="600" style="border-radius: 6px;"/>
|
||||||
|
</p>
|
||||||
|
|
||||||
|
# Model Information
|
||||||
|
|
||||||
|
**AQUA-1B** is Kurma AI’s compact and efficient **1-billion parameter Small Language Model (SLM)**, It is the **first lightweight Aquaculture domain-specific model** purpose-built for real-time aquaculture operations involving IoT sensor data, autonomous systems, and robotic agents.
|
||||||
|
|
||||||
|
Designed for edge deployments and low-latency environments, **AQUA-1B** enables on-device decision-making, real-time alert generation, and agentic task execution. It powers intelligent aquaculture systems for Water quality monitoring, Automated feeding routines, Mobile robotic inspections across ponds, tanks, and recirculating aquaculture systems (RAS)
|
||||||
|
|
||||||
|
|
||||||
|
Learn more about [Kurma AI](https://kurma.ai/company).
|
||||||
|
|
||||||
|
---
|
||||||
|
|
||||||
|
# Key Features
|
||||||
|
|
||||||
|
- **Edge-Ready Intelligence** Optimized for low-power, real-time inference on embedded devices like Raspberry Pi, Jetson Nano, and Coral TPU.
|
||||||
|
- **Agentic Task Execution** Supports multi-step agent-based workflows such as sensor checks, feeding triggers, water exchange scheduling, and autonomous health checks using instruction-following prompts.
|
||||||
|
- **IoT-Aware Reasoning** Natively understands and reasons over sensor data inputs (e.g., temperature, pH, TDS, turbidity, DO), enabling rapid decisions in fluctuating aquaculture environments.
|
||||||
|
- **Robotic Automation Control** Designed to interact with robotic systems, including underwater, and mobile pond inspectors.
|
||||||
|
- **Autonomous Alerting Systems** Powers local alert mechanisms (via SMS, Telegram bots, or MQTT) that notify farmers when water parameters exceed safe thresholds or when interventions are required.
|
||||||
|
- **Field-Deployable Decision Engine** Enables fully autonomous operation in remote hatcheries and ponds, ensuring uninterrupted control even in offline conditions or low-connectivity zones.
|
||||||
|
|
||||||
|
---
|
||||||
|
|
||||||
|
# Training Data Highlights
|
||||||
|
|
||||||
|
- Extension worker–farmer dialogues and field advisory logs
|
||||||
|
- FAO, ICAR, NOAA, and peer-reviewed aquaculture research
|
||||||
|
- Synthetic Q&A from 5,000+ aquaculture-focused topics
|
||||||
|
- Climate-resilient practices, hatchery SOPs, and water quality datasets
|
||||||
|
- Carefully curated to support **species-specific culture** methods
|
||||||
|
- **Scale:** Trained on approximately **3 million real and synthetic Q&A pairs**, totaling around **1 billion tokens** of high-quality, domain-specific data.
|
||||||
|
|
||||||
|
|
||||||
|
---
|
||||||
|
|
||||||
|
# Model Specifications
|
||||||
|
|
||||||
|
- **Base Model**: Gemma 3 1B (by [Google DeepMind](https://deepmind.google/))
|
||||||
|
- **Training Tokens**: ~1 Billion
|
||||||
|
- **Released On** 4, July 2025
|
||||||
|
- **Data Volume**: 3M+ expert-verified and synthetic instructions
|
||||||
|
- **Origin**: Made in America by [Kurma AI](https://kurma.ai/)
|
||||||
|
- **Training Technic** Model is trained via Fine-tuning using (LoRA-based) Supervised Fine-Tuning (SFT).
|
||||||
|
- **Training Infrastructure**: Trained using **8 NVIDIA H200 GPU Multi Cluster**
|
||||||
|
Special Thanks to [Nebius](https://nebius.com/)
|
||||||
|
|
||||||
|
---
|
||||||
|
|
||||||
|
# Quickstart
|
||||||
|
|
||||||
|
Transformers (Google Colab/ jupyter)
|
||||||
|
|
||||||
|
|
||||||
|
- Install dependencies
|
||||||
|
```python
|
||||||
|
!pip install transformers accelerate
|
||||||
|
```
|
||||||
|
|
||||||
|
- Log in with your Hugging Face access token
|
||||||
|
```python
|
||||||
|
from huggingface_hub import login
|
||||||
|
```
|
||||||
|
|
||||||
|
- Import model from Huggingface
|
||||||
|
```python
|
||||||
|
from transformers import AutoTokenizer, AutoModelForCausalLM
|
||||||
|
import torch
|
||||||
|
|
||||||
|
model_id = "KurmaAI/AQUA-1B"
|
||||||
|
|
||||||
|
tokenizer = AutoTokenizer.from_pretrained(model_id, trust_remote_code=True)
|
||||||
|
model = AutoModelForCausalLM.from_pretrained(
|
||||||
|
model_id,
|
||||||
|
device_map="auto", # Automatically uses GPU if available
|
||||||
|
torch_dtype=torch.float16, # Use torch.float32 if no GPU
|
||||||
|
trust_remote_code=True
|
||||||
|
)
|
||||||
|
```
|
||||||
|
|
||||||
|
- Test Prompt
|
||||||
|
```python
|
||||||
|
prompt = "What are the most common diseases in shrimp farming and how can they be prevented?"
|
||||||
|
inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
|
||||||
|
outputs = model.generate(**inputs, max_new_tokens=256)
|
||||||
|
|
||||||
|
response = tokenizer.decode(outputs[0], skip_special_tokens=True)
|
||||||
|
print(response)
|
||||||
|
```
|
||||||
|
---
|
||||||
|
|
||||||
|
# 🙏 Acknowledgements
|
||||||
|
This project was made possible thanks to:
|
||||||
|
- [Nebius](https://nebius.com/) for providing a compute grant and access to NVIDIA H200 GPU servers, which powered the model training process.
|
||||||
|
- [Google DeepMind](https://deepmind.google/) for sharing their open-source language models, which made this project possible.
|
||||||
|
- Kurma AI research team: including aquaculture experts, machine learning engineers, data annotators, and advisors who collaborated to curate, verify, and refine the domain-specific dataset used for fine-tuning this model.
|
||||||
|
|
||||||
|
---
|
||||||
|
|
||||||
|
# ⚠️ Disclaimer, Bias & Limitations
|
||||||
|
|
||||||
|
- **Domain Bias**: The model may reflect inherent biases present in the aquaculture data sources and industry practices on which it was trained.
|
||||||
|
- **Temporal Data Limitation**: Climate and environmental recommendations are based on information available up to 2024. Users should cross-check any climate-related advice against the latest advisories (e.g., IMD or NOAA updates).
|
||||||
|
- **Potential Hallucinations**: Like all large language models, Aqua-1B may occasionally generate inaccurate or misleading responses ("hallucinations").
|
||||||
|
- **Always validate critical, regulatory, or high-impact decisions with a qualified aquaculture professional.**
|
||||||
|
|
||||||
|
---
|
||||||
|
|
||||||
|
# Citation
|
||||||
|
|
||||||
|
```bibtex
|
||||||
|
@article{narisetty2025aqua,
|
||||||
|
title={AQUA: A Large Language Model for Aquaculture \& Fisheries},
|
||||||
|
author={Narisetty, Praneeth and Kattamanchi, Uday Kumar Reddy and Nimma, Lohit Akshant and Karnati, Sri Ram Kaushik and Kore, Shiva Nagendra Babu and Golamari, Mounika and Nageshreddy, Tejashree},
|
||||||
|
journal={arXiv preprint arXiv:2507.20520},
|
||||||
|
year={2025},
|
||||||
|
doi={10.48550/arXiv.2507.20520}
|
||||||
|
}
|
||||||
|
```
|
||||||
3
added_tokens.json
Normal file
3
added_tokens.json
Normal file
@@ -0,0 +1,3 @@
|
|||||||
|
{
|
||||||
|
"<image_soft_token>": 262144
|
||||||
|
}
|
||||||
47
chat_template.jinja
Normal file
47
chat_template.jinja
Normal file
@@ -0,0 +1,47 @@
|
|||||||
|
{{ bos_token }}
|
||||||
|
{%- if messages[0]['role'] == 'system' -%}
|
||||||
|
{%- if messages[0]['content'] is string -%}
|
||||||
|
{%- set first_user_prefix = messages[0]['content'] + '
|
||||||
|
|
||||||
|
' -%}
|
||||||
|
{%- else -%}
|
||||||
|
{%- set first_user_prefix = messages[0]['content'][0]['text'] + '
|
||||||
|
|
||||||
|
' -%}
|
||||||
|
{%- endif -%}
|
||||||
|
{%- set loop_messages = messages[1:] -%}
|
||||||
|
{%- else -%}
|
||||||
|
{%- set first_user_prefix = "" -%}
|
||||||
|
{%- set loop_messages = messages -%}
|
||||||
|
{%- endif -%}
|
||||||
|
{%- for message in loop_messages -%}
|
||||||
|
{%- if (message['role'] == 'user') != (loop.index0 % 2 == 0) -%}
|
||||||
|
{{ raise_exception("Conversation roles must alternate user/assistant/user/assistant/...") }}
|
||||||
|
{%- endif -%}
|
||||||
|
{%- if (message['role'] == 'assistant') -%}
|
||||||
|
{%- set role = "model" -%}
|
||||||
|
{%- else -%}
|
||||||
|
{%- set role = message['role'] -%}
|
||||||
|
{%- endif -%}
|
||||||
|
{{ '<start_of_turn>' + role + '
|
||||||
|
' + (first_user_prefix if loop.first else "") }}
|
||||||
|
{%- if message['content'] is string -%}
|
||||||
|
{{ message['content'] | trim }}
|
||||||
|
{%- elif message['content'] is iterable -%}
|
||||||
|
{%- for item in message['content'] -%}
|
||||||
|
{%- if item['type'] == 'image' -%}
|
||||||
|
{{ '<start_of_image>' }}
|
||||||
|
{%- elif item['type'] == 'text' -%}
|
||||||
|
{{ item['text'] | trim }}
|
||||||
|
{%- endif -%}
|
||||||
|
{%- endfor -%}
|
||||||
|
{%- else -%}
|
||||||
|
{{ raise_exception("Invalid content type") }}
|
||||||
|
{%- endif -%}
|
||||||
|
{{ '<end_of_turn>
|
||||||
|
' }}
|
||||||
|
{%- endfor -%}
|
||||||
|
{%- if add_generation_prompt -%}
|
||||||
|
{{'<start_of_turn>model
|
||||||
|
'}}
|
||||||
|
{%- endif -%}
|
||||||
37
config.json
Normal file
37
config.json
Normal file
@@ -0,0 +1,37 @@
|
|||||||
|
{
|
||||||
|
"architectures": [
|
||||||
|
"Gemma3ForCausalLM"
|
||||||
|
],
|
||||||
|
"attention_bias": false,
|
||||||
|
"attention_dropout": 0.0,
|
||||||
|
"attn_logit_softcapping": null,
|
||||||
|
"bos_token_id": 2,
|
||||||
|
"cache_implementation": "hybrid",
|
||||||
|
"eos_token_id": [
|
||||||
|
1,
|
||||||
|
106
|
||||||
|
],
|
||||||
|
"final_logit_softcapping": null,
|
||||||
|
"head_dim": 256,
|
||||||
|
"hidden_activation": "gelu_pytorch_tanh",
|
||||||
|
"hidden_size": 1152,
|
||||||
|
"initializer_range": 0.02,
|
||||||
|
"intermediate_size": 6912,
|
||||||
|
"max_position_embeddings": 32768,
|
||||||
|
"model_type": "gemma3_text",
|
||||||
|
"num_attention_heads": 4,
|
||||||
|
"num_hidden_layers": 26,
|
||||||
|
"num_key_value_heads": 1,
|
||||||
|
"pad_token_id": 0,
|
||||||
|
"query_pre_attn_scalar": 256,
|
||||||
|
"rms_norm_eps": 1e-06,
|
||||||
|
"rope_local_base_freq": 10000,
|
||||||
|
"rope_scaling": null,
|
||||||
|
"rope_theta": 1000000,
|
||||||
|
"sliding_window": 512,
|
||||||
|
"sliding_window_pattern": 6,
|
||||||
|
"torch_dtype": "bfloat16",
|
||||||
|
"transformers_version": "4.52.4",
|
||||||
|
"use_cache": true,
|
||||||
|
"vocab_size": 262144
|
||||||
|
}
|
||||||
1
configuration.json
Normal file
1
configuration.json
Normal file
@@ -0,0 +1 @@
|
|||||||
|
{"framework": "pytorch", "task": "text-generation", "allow_remote": true}
|
||||||
13
generation_config.json
Normal file
13
generation_config.json
Normal file
@@ -0,0 +1,13 @@
|
|||||||
|
{
|
||||||
|
"bos_token_id": 2,
|
||||||
|
"cache_implementation": "hybrid",
|
||||||
|
"do_sample": true,
|
||||||
|
"eos_token_id": [
|
||||||
|
1,
|
||||||
|
106
|
||||||
|
],
|
||||||
|
"pad_token_id": 0,
|
||||||
|
"top_k": 64,
|
||||||
|
"top_p": 0.95,
|
||||||
|
"transformers_version": "4.52.4"
|
||||||
|
}
|
||||||
3
model.safetensors
Normal file
3
model.safetensors
Normal file
@@ -0,0 +1,3 @@
|
|||||||
|
version https://git-lfs.github.com/spec/v1
|
||||||
|
oid sha256:bd9180947a56e2a27d5c5d728aeb0a06e274db5df945576bfe61b22cfc8a4189
|
||||||
|
size 1999811208
|
||||||
33
special_tokens_map.json
Normal file
33
special_tokens_map.json
Normal file
@@ -0,0 +1,33 @@
|
|||||||
|
{
|
||||||
|
"boi_token": "<start_of_image>",
|
||||||
|
"bos_token": {
|
||||||
|
"content": "<bos>",
|
||||||
|
"lstrip": false,
|
||||||
|
"normalized": false,
|
||||||
|
"rstrip": false,
|
||||||
|
"single_word": false
|
||||||
|
},
|
||||||
|
"eoi_token": "<end_of_image>",
|
||||||
|
"eos_token": {
|
||||||
|
"content": "<end_of_turn>",
|
||||||
|
"lstrip": false,
|
||||||
|
"normalized": false,
|
||||||
|
"rstrip": false,
|
||||||
|
"single_word": false
|
||||||
|
},
|
||||||
|
"image_token": "<image_soft_token>",
|
||||||
|
"pad_token": {
|
||||||
|
"content": "<pad>",
|
||||||
|
"lstrip": false,
|
||||||
|
"normalized": false,
|
||||||
|
"rstrip": false,
|
||||||
|
"single_word": false
|
||||||
|
},
|
||||||
|
"unk_token": {
|
||||||
|
"content": "<unk>",
|
||||||
|
"lstrip": false,
|
||||||
|
"normalized": false,
|
||||||
|
"rstrip": false,
|
||||||
|
"single_word": false
|
||||||
|
}
|
||||||
|
}
|
||||||
3
tokenizer.json
Normal file
3
tokenizer.json
Normal file
@@ -0,0 +1,3 @@
|
|||||||
|
version https://git-lfs.github.com/spec/v1
|
||||||
|
oid sha256:4667f2089529e8e7657cfb6d1c19910ae71ff5f28aa7ab2ff2763330affad795
|
||||||
|
size 33384568
|
||||||
3
tokenizer.model
Normal file
3
tokenizer.model
Normal file
@@ -0,0 +1,3 @@
|
|||||||
|
version https://git-lfs.github.com/spec/v1
|
||||||
|
oid sha256:1299c11d7cf632ef3b4e11937501358ada021bbdf7c47638d13c0ee982f2e79c
|
||||||
|
size 4689074
|
||||||
51347
tokenizer_config.json
Normal file
51347
tokenizer_config.json
Normal file
File diff suppressed because it is too large
Load Diff
Reference in New Issue
Block a user