--- base_model: artificialguybr/Gemma2-2B-OpenHermes2.5 datasets: - teknium/OpenHermes-2.5 language: - en library_name: transformers license: apache-2.0 quantized_by: artificialguybr tags: - gemma - instruct - finetune - chatml - gpt4 - synthetic data - distillation --- ## About --- ### 🌐 Website You can find more of my models, projects, and information on my official website: - **[artificialguy.com](https://artificialguy.com/)** ### 🚀 Prompt Hub Need high-quality prompts for image models and LLMs? Explore **[findgoodprompt.com](https://findgoodprompt.com)**. ### 💖 Support My Work If you find this model useful, please consider supporting my work. It helps me cover server costs and dedicate more time to new open-source projects. - **Patreon:** [Support on Patreon](https://www.patreon.com/user?u=81570187) - **Ko-fi:** [Buy me a Ko-fi](https://ko-fi.com/artificialguybr) - **Buy Me a Coffee:** [Buy me a Coffee](https://buymeacoffee.com/jvkape) Quantization for: https://huggingface.co/artificialguybr/Gemma2-2B-OpenHermes2.5 ## How to use If you are unsure how to use GGUF files, look at the [TheBloke READMEs](https://huggingface.co/TheBloke/CodeLlama-70B-Python-GGUF) for more details, including on how to concatenate multi-part files.