Phi-4-GGUF

Original Model

microsoft/phi-4

Run with Gaianet

Prompt template:

prompt template: phi-4-chat

Context size:

chat_ctx_size: 16000

Run with GaiaNet:

Quantized with llama.cpp b4450

Downloads last month
14
GGUF
Model size
14.7B params
Architecture
phi3

2-bit

3-bit

4-bit

5-bit

6-bit

8-bit

16-bit

Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for gaianet/phi-4-GGUF

Base model

microsoft/phi-4
Quantized
(43)
this model