matrixportal commited on
Commit
7478988
1 Parent(s): 2eed6ba

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +134 -0
README.md ADDED
@@ -0,0 +1,134 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ library_name: transformers
4
+ pipeline_tag: text-generation
5
+ datasets:
6
+ - yulan-team/YuLan-Mini-Datasets
7
+ - HuggingFaceFW/fineweb-edu
8
+ - bigcode/the-stack-v2
9
+ - mlfoundations/dclm-baseline-1.0
10
+ - math-ai/AutoMathText
11
+ - gair-prox/open-web-math-pro
12
+ - RUC-AIBOX/long_form_thought_data_5k
13
+ - internlm/Lean-Workbook
14
+ - internlm/Lean-Github
15
+ - deepseek-ai/DeepSeek-Prover-V1
16
+ - ScalableMath/Lean-STaR-base
17
+ - ScalableMath/Lean-STaR-plus
18
+ - ScalableMath/Lean-CoT-base
19
+ - ScalableMath/Lean-CoT-plus
20
+ - opencsg/chinese-fineweb-edu
21
+ - liwu/MNBVC
22
+ - vikp/textbook_quality_programming
23
+ - HuggingFaceTB/smollm-corpus
24
+ - OpenCoder-LLM/opc-annealing-corpus
25
+ - OpenCoder-LLM/opc-sft-stage1
26
+ - OpenCoder-LLM/opc-sft-stage2
27
+ - XinyaoHu/AMPS_mathematica
28
+ - deepmind/math_dataset
29
+ - mrfakename/basic-math-10m
30
+ - microsoft/orca-math-word-problems-200k
31
+ - AI-MO/NuminaMath-CoT
32
+ - HuggingFaceTB/cosmopedia
33
+ - MU-NLPC/Calc-ape210k
34
+ - manu/project_gutenberg
35
+ - storytracer/LoC-PD-Books
36
+ - allenai/dolma
37
+ language:
38
+ - en
39
+ - zh
40
+ tags:
41
+ - code
42
+ - math
43
+ - llama-cpp
44
+ - gguf-my-repo
45
+ arxiv: 2412.17743
46
+ base_model: yulan-team/YuLan-Mini
47
+ model-index:
48
+ - name: YuLan-Mini
49
+ results:
50
+ - task:
51
+ type: text-generation
52
+ dataset:
53
+ name: HumanEval
54
+ type: openai_humaneval
55
+ metrics:
56
+ - type: pass@1
57
+ value: 0.64
58
+ name: pass@1
59
+ verified: false
60
+ - task:
61
+ type: text-generation
62
+ dataset:
63
+ name: MBPP
64
+ type: mbpp
65
+ metrics:
66
+ - type: pass@1
67
+ value: 0.659
68
+ name: pass@1
69
+ verified: false
70
+ - task:
71
+ type: text-generation
72
+ dataset:
73
+ name: MATH-500
74
+ type: math-500
75
+ metrics:
76
+ - type: maj@1
77
+ value: 0.378
78
+ name: maj@1
79
+ verified: false
80
+ - task:
81
+ type: text-generation
82
+ dataset:
83
+ name: GSM8K
84
+ type: gsm8k
85
+ metrics:
86
+ - type: maj@1
87
+ value: 0.684
88
+ name: maj@1
89
+ verified: false
90
+ ---
91
+
92
+ # matrixportal/YuLan-Mini-Q8_0-GGUF
93
+ This model was converted to GGUF format from [`yulan-team/YuLan-Mini`](https://huggingface.co/yulan-team/YuLan-Mini) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
94
+ Refer to the [original model card](https://huggingface.co/yulan-team/YuLan-Mini) for more details on the model.
95
+
96
+ ## Use with llama.cpp
97
+ Install llama.cpp through brew (works on Mac and Linux)
98
+
99
+ ```bash
100
+ brew install llama.cpp
101
+
102
+ ```
103
+ Invoke the llama.cpp server or the CLI.
104
+
105
+ ### CLI:
106
+ ```bash
107
+ llama-cli --hf-repo matrixportal/YuLan-Mini-Q8_0-GGUF --hf-file yulan-mini-q8_0.gguf -p "The meaning to life and the universe is"
108
+ ```
109
+
110
+ ### Server:
111
+ ```bash
112
+ llama-server --hf-repo matrixportal/YuLan-Mini-Q8_0-GGUF --hf-file yulan-mini-q8_0.gguf -c 2048
113
+ ```
114
+
115
+ Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well.
116
+
117
+ Step 1: Clone llama.cpp from GitHub.
118
+ ```
119
+ git clone https://github.com/ggerganov/llama.cpp
120
+ ```
121
+
122
+ Step 2: Move into the llama.cpp folder and build it with `LLAMA_CURL=1` flag along with other hardware-specific flags (for ex: LLAMA_CUDA=1 for Nvidia GPUs on Linux).
123
+ ```
124
+ cd llama.cpp && LLAMA_CURL=1 make
125
+ ```
126
+
127
+ Step 3: Run inference through the main binary.
128
+ ```
129
+ ./llama-cli --hf-repo matrixportal/YuLan-Mini-Q8_0-GGUF --hf-file yulan-mini-q8_0.gguf -p "The meaning to life and the universe is"
130
+ ```
131
+ or
132
+ ```
133
+ ./llama-server --hf-repo matrixportal/YuLan-Mini-Q8_0-GGUF --hf-file yulan-mini-q8_0.gguf -c 2048
134
+ ```