ModelHub XC f5e65f3176 初始化项目,由ModelHub XC社区提供模型
Model: quicktensor/blockrank-msmarco-mistral-7b
Source: Original Platform
2026-05-04 13:21:04 +08:00

license, language, library_name, tags, base_model, datasets, metrics
license language library_name tags base_model datasets metrics
mit
en
transformers
text-generation
information-retrieval
ranking
reranking
blockrank
mistral
mistralai/Mistral-7B-Instruct-v0.3
quicktensor/blockrank-msmarco-train-10p
ndcg
mrr

BlockRank-Mistral-7B: Scalable In-context Ranking with Generative Models

Open In Colab

BlockRank-Mistral-7B is a fine-tuned version of Mistral-7B-Instruct-v0.3 optimized for efficient in-context document ranking. It implements BlockRank, a method that makes LLMs efficient and scalable for ranking by aligning their internal attention mechanisms with the structure of the ranking task.

BlockRank Architecture

Key Features

  • Linear Complexity Attention: Structured sparse attention reduces complexity from O(n²) to O(n)
  • 2-4× Faster Inference: Attention-based scoring eliminates autoregressive decoding
  • Auxiliary Contrastive Loss: Mid-layer contrastive objective improves relevance signals
  • Strong Zero-shot Generalization: SOTA performance on BEIR benchmarks

Citation

If you use this model, please cite:

@article{gupta2025blockrank,
  title={Scalable In-context Ranking with Generative Models},
  author={Gupta, Nilesh and You, Chong and Bhojanapalli, Srinadh and Kumar, Sanjiv and Dhillon, Inderjit and Yu, Felix},
  journal={arXiv preprint arXiv:2510.05396},
  year={2025}
}

Model Card Contact

For questions or issues, please open an issue on GitHub.

Additional Resources

License

This model is released under the MIT License. See LICENSE for details.

Description
Model synced from source: quicktensor/blockrank-msmarco-mistral-7b
Readme 664 KiB
Languages
Jinja 100%