初始化项目,由ModelHub XC社区提供模型

Model: iamshnoo/combined_only_url_with_metadata_1b
Source: Original Platform
This commit is contained in:
ModelHub XC
2026-04-21 23:16:54 +08:00
commit bd765e9b21
11 changed files with 2232 additions and 0 deletions

77
README.md Normal file
View File

@@ -0,0 +1,77 @@
---
pipeline_tag: text-generation
library_name: transformers
tags:
- text-generation
- metadata-localization
- metadata-ablation
- 1b
- with-metadata
- pretraining
---
# combined_only_url_with_metadata_1b
## Summary
This repo contains the url 1b model at the final 10k-step checkpoint for the metadata localization project. It was trained from scratch on the project corpus, using the Llama 3.2 tokenizer and vocabulary.
## Variant Metadata
- Stage: `pretrain`
- Family: `metadata_ablation`
- Size: `1b`
- Metadata condition: `with_metadata`
- Base model lineage: `Trained from scratch; tokenizer/vocabulary from meta-llama/Llama-3.2-1B`
## Weights & Biases Provenance
- Run name: `24/12/2025_22:00:55_combined_only_url_with_metadata_1b`
- Internal run URL: `https://wandb.ai/iamshnoo/nanotron/runs/mgsf3ei7`
- Note: the Weights & Biases workspace is private; public readers should use the summarized metrics and configuration below.
- State: `finished`
- Runtime: `114h 11m 7s`
## Run Summary
- `KPI/train_lm_loss`: `2.0952`
- `KPI/train_perplexity`: `8.1273`
- `KPI/val_loss`: `2.088`
- `KPI/val_perplexity`: `8.0687`
- `KPI/consumed_tokens/train`: `41,943,040,000`
- `_step`: `10,000`
## Training Configuration
- `train_steps`: `10,000`
- `sequence_length`: `2,048`
- `micro_batch_size`: `8`
- `batch_accumulation_per_replica`: `64`
- `learning_rate`: `0.003`
- `min_decay_lr`: `0.0003`
- `checkpoint_interval`: `1,000`
## Training Curves
Static plots below were exported from the private Weights & Biases run and embedded here for public access.
### Train Loss
![Train Loss](assets/train_loss.png)
### Validation Perplexity
![Validation Perplexity](assets/val_perplexity.png)
### Throughput
![Throughput](assets/tokens_per_sec.png)
## Project Context
This model is part of the metadata localization release. Related checkpoints and variants are grouped in the public Hugging Face collection [Metadata Conditioned LLMs](https://huggingface.co/collections/iamshnoo/metadata-conditioned-llms).
- Training data source: [News on the Web (NOW) Corpus](https://www.english-corpora.org/now/)
- Project repository: [https://github.com/iamshnoo/metadata_localization](https://github.com/iamshnoo/metadata_localization)
- Paper: [https://arxiv.org/abs/2601.15236](https://arxiv.org/abs/2601.15236)
Last synced: `2026-04-02 14:43:35 UTC`