Triangle104 commited on
Commit
b5a4f04
·
verified ·
1 Parent(s): 8225db2

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +164 -0
README.md ADDED
@@ -0,0 +1,164 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: transformers
3
+ tags:
4
+ - mergekit
5
+ - merge
6
+ - conversational
7
+ - chat
8
+ - instruct
9
+ - llama-cpp
10
+ - gguf-my-repo
11
+ base_model: sequelbox/Llama3.1-8B-PlumChat
12
+ license: llama3.1
13
+ model-index:
14
+ - name: Llama3.1-8B-PlumChat
15
+ results:
16
+ - task:
17
+ type: text-generation
18
+ name: Text Generation
19
+ dataset:
20
+ name: Winogrande (5-Shot)
21
+ type: Winogrande
22
+ args:
23
+ num_few_shot: 5
24
+ metrics:
25
+ - type: acc
26
+ value: 72.22
27
+ name: acc
28
+ - task:
29
+ type: text-generation
30
+ name: Text Generation
31
+ dataset:
32
+ name: IFEval (0-Shot)
33
+ type: HuggingFaceH4/ifeval
34
+ args:
35
+ num_few_shot: 0
36
+ metrics:
37
+ - type: inst_level_strict_acc and prompt_level_strict_acc
38
+ value: 42.43
39
+ name: strict accuracy
40
+ source:
41
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=sequelbox/Llama3.1-8B-PlumChat
42
+ name: Open LLM Leaderboard
43
+ - task:
44
+ type: text-generation
45
+ name: Text Generation
46
+ dataset:
47
+ name: BBH (3-Shot)
48
+ type: BBH
49
+ args:
50
+ num_few_shot: 3
51
+ metrics:
52
+ - type: acc_norm
53
+ value: 13.94
54
+ name: normalized accuracy
55
+ source:
56
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=sequelbox/Llama3.1-8B-PlumChat
57
+ name: Open LLM Leaderboard
58
+ - task:
59
+ type: text-generation
60
+ name: Text Generation
61
+ dataset:
62
+ name: MATH Lvl 5 (4-Shot)
63
+ type: hendrycks/competition_math
64
+ args:
65
+ num_few_shot: 4
66
+ metrics:
67
+ - type: exact_match
68
+ value: 3.1
69
+ name: exact match
70
+ source:
71
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=sequelbox/Llama3.1-8B-PlumChat
72
+ name: Open LLM Leaderboard
73
+ - task:
74
+ type: text-generation
75
+ name: Text Generation
76
+ dataset:
77
+ name: GPQA (0-shot)
78
+ type: Idavidrein/gpqa
79
+ args:
80
+ num_few_shot: 0
81
+ metrics:
82
+ - type: acc_norm
83
+ value: 2.01
84
+ name: acc_norm
85
+ source:
86
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=sequelbox/Llama3.1-8B-PlumChat
87
+ name: Open LLM Leaderboard
88
+ - task:
89
+ type: text-generation
90
+ name: Text Generation
91
+ dataset:
92
+ name: MuSR (0-shot)
93
+ type: TAUR-Lab/MuSR
94
+ args:
95
+ num_few_shot: 0
96
+ metrics:
97
+ - type: acc_norm
98
+ value: 4.77
99
+ name: acc_norm
100
+ source:
101
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=sequelbox/Llama3.1-8B-PlumChat
102
+ name: Open LLM Leaderboard
103
+ - task:
104
+ type: text-generation
105
+ name: Text Generation
106
+ dataset:
107
+ name: MMLU-PRO (5-shot)
108
+ type: TIGER-Lab/MMLU-Pro
109
+ config: main
110
+ split: test
111
+ args:
112
+ num_few_shot: 5
113
+ metrics:
114
+ - type: acc
115
+ value: 12.52
116
+ name: accuracy
117
+ source:
118
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=sequelbox/Llama3.1-8B-PlumChat
119
+ name: Open LLM Leaderboard
120
+ ---
121
+
122
+ # Triangle104/Llama3.1-8B-PlumChat-Q6_K-GGUF
123
+ This model was converted to GGUF format from [`sequelbox/Llama3.1-8B-PlumChat`](https://huggingface.co/sequelbox/Llama3.1-8B-PlumChat) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
124
+ Refer to the [original model card](https://huggingface.co/sequelbox/Llama3.1-8B-PlumChat) for more details on the model.
125
+
126
+ ## Use with llama.cpp
127
+ Install llama.cpp through brew (works on Mac and Linux)
128
+
129
+ ```bash
130
+ brew install llama.cpp
131
+
132
+ ```
133
+ Invoke the llama.cpp server or the CLI.
134
+
135
+ ### CLI:
136
+ ```bash
137
+ llama-cli --hf-repo Triangle104/Llama3.1-8B-PlumChat-Q6_K-GGUF --hf-file llama3.1-8b-plumchat-q6_k.gguf -p "The meaning to life and the universe is"
138
+ ```
139
+
140
+ ### Server:
141
+ ```bash
142
+ llama-server --hf-repo Triangle104/Llama3.1-8B-PlumChat-Q6_K-GGUF --hf-file llama3.1-8b-plumchat-q6_k.gguf -c 2048
143
+ ```
144
+
145
+ Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well.
146
+
147
+ Step 1: Clone llama.cpp from GitHub.
148
+ ```
149
+ git clone https://github.com/ggerganov/llama.cpp
150
+ ```
151
+
152
+ Step 2: Move into the llama.cpp folder and build it with `LLAMA_CURL=1` flag along with other hardware-specific flags (for ex: LLAMA_CUDA=1 for Nvidia GPUs on Linux).
153
+ ```
154
+ cd llama.cpp && LLAMA_CURL=1 make
155
+ ```
156
+
157
+ Step 3: Run inference through the main binary.
158
+ ```
159
+ ./llama-cli --hf-repo Triangle104/Llama3.1-8B-PlumChat-Q6_K-GGUF --hf-file llama3.1-8b-plumchat-q6_k.gguf -p "The meaning to life and the universe is"
160
+ ```
161
+ or
162
+ ```
163
+ ./llama-server --hf-repo Triangle104/Llama3.1-8B-PlumChat-Q6_K-GGUF --hf-file llama3.1-8b-plumchat-q6_k.gguf -c 2048
164
+ ```