Files
blockrank-msmarco-mistral-7b/README.md
ModelHub XC f5e65f3176 初始化项目,由ModelHub XC社区提供模型
Model: quicktensor/blockrank-msmarco-mistral-7b
Source: Original Platform
2026-05-04 13:21:04 +08:00

2.4 KiB
Raw Blame History

license, language, library_name, tags, base_model, datasets, metrics
license language library_name tags base_model datasets metrics
mit
en
transformers
text-generation
information-retrieval
ranking
reranking
blockrank
mistral
mistralai/Mistral-7B-Instruct-v0.3
quicktensor/blockrank-msmarco-train-10p
ndcg
mrr

BlockRank-Mistral-7B: Scalable In-context Ranking with Generative Models

Open In Colab

BlockRank-Mistral-7B is a fine-tuned version of Mistral-7B-Instruct-v0.3 optimized for efficient in-context document ranking. It implements BlockRank, a method that makes LLMs efficient and scalable for ranking by aligning their internal attention mechanisms with the structure of the ranking task.

BlockRank Architecture

Key Features

  • Linear Complexity Attention: Structured sparse attention reduces complexity from O(n²) to O(n)
  • 2-4× Faster Inference: Attention-based scoring eliminates autoregressive decoding
  • Auxiliary Contrastive Loss: Mid-layer contrastive objective improves relevance signals
  • Strong Zero-shot Generalization: SOTA performance on BEIR benchmarks

Citation

If you use this model, please cite:

@article{gupta2025blockrank,
  title={Scalable In-context Ranking with Generative Models},
  author={Gupta, Nilesh and You, Chong and Bhojanapalli, Srinadh and Kumar, Sanjiv and Dhillon, Inderjit and Yu, Felix},
  journal={arXiv preprint arXiv:2510.05396},
  year={2025}
}

Model Card Contact

For questions or issues, please open an issue on GitHub.

Additional Resources

License

This model is released under the MIT License. See LICENSE for details.