Update README.md

This commit is contained in:
Toprak Kesgin
2024-07-25 09:48:18 +00:00
committed by system
parent 7c75df30ed
commit d49be2d4f8

View File

@@ -86,6 +86,11 @@ Obviously, we encountered better inference quality for models with the highest b
Each model's memory footprint can be anticipated by the qunatization docs in either [Hugging Face](https://huggingface.co/docs/transformers/main/en/quantization/overview) or [llama.cpp](https://github.com/ggerganov/llama.cpp/tree/master/examples/quantize). Each model's memory footprint can be anticipated by the qunatization docs in either [Hugging Face](https://huggingface.co/docs/transformers/main/en/quantization/overview) or [llama.cpp](https://github.com/ggerganov/llama.cpp/tree/master/examples/quantize).
# Acknowledgments
- Research supported with Cloud TPUs from [Google's TensorFlow Research Cloud](https://sites.research.google/trc/about/) (TFRC). Thanks for providing access to the TFRC ❤️
- Thanks to the generous support from the Hugging Face team, it is possible to download models from their S3 storage 🤗
## Contact ## Contact
*Feel free to contact us whenever you confront any problems :)* *Feel free to contact us whenever you confront any problems :)*