base_model: google/gemma-2-2b-it | |
library_name: transformers | |
license: gemma | |
pipeline_tag: text-generation | |
tags: | |
- conversational | |
- mlc-ai | |
- MLC-Weight-Conversion | |
extra_gated_heading: Access Gemma on Hugging Face | |
extra_gated_prompt: To access Gemma on Hugging Face, you’re required to review and | |
agree to Google’s usage license. To do this, please ensure you’re logged in to Hugging | |
Face and click below. Requests are processed immediately. | |
extra_gated_button_content: Acknowledge license | |
# AMKCode/gemma-2-2b-it-q4f32_1-MLC | |
This model was compiled using MLC-LLM with q4f32_1 quantization from [google/gemma-2-2b-it](https://huggingface.co/google/gemma-2-2b-it). | |
The conversion was done using the [MLC-Weight-Conversion](https://huggingface.co/spaces/mlc-ai/MLC-Weight-Conversion) space. | |
To run this model, please first install [MLC-LLM](https://llm.mlc.ai/docs/install/mlc_llm.html#install-mlc-packages). | |
To chat with the model on your terminal: | |
```bash | |
mlc_llm chat HF://AMKCode/gemma-2-2b-it-q4f32_1-MLC | |
``` | |
For more information on how to use MLC-LLM, please visit the MLC-LLM [documentation](https://llm.mlc.ai/docs/index.html). | |