rogkesavan commited on
Commit
02279dc
·
verified ·
1 Parent(s): 6cef48a

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +125 -5
README.md CHANGED
@@ -1,5 +1,125 @@
1
- ---
2
- license: apache-2.0
3
- tags:
4
- - llama-factory
5
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ### Nidum-Llama-3.2-3B-Uncensored
2
+
3
+ [![GitHub Icon](https://upload.wikimedia.org/wikipedia/commons/thumb/9/95/Font_Awesome_5_brands_github.svg/232px-Font_Awesome_5_brands_github.svg.png)](https://github.com/NidumAI-Inc)
4
+ **Explore Nidum's Open-Source Projects on GitHub**: [https://github.com/NidumAI-Inc](https://github.com/NidumAI-Inc)
5
+
6
+ ---
7
+
8
+ ### Welcome to Nidum!
9
+ At Nidum, we believe in pushing the boundaries of innovation by providing advanced and unrestricted AI models for every application. Dive into our world of possibilities and experience the freedom of **Nidum-Llama-3.2-3B-Uncensored**, tailored to meet diverse needs with exceptional performance.
10
+
11
+ ---
12
+
13
+ ### Key Features
14
+
15
+ 1. **Uncensored Responses**: Capable of addressing any query without content restrictions, offering detailed and uninhibited answers.
16
+ 2. **Versatility**: Excels in diverse use cases, from complex technical queries to engaging casual conversations.
17
+ 3. **Advanced Contextual Understanding**: Draws from an expansive knowledge base for accurate and context-aware outputs.
18
+ 4. **Extended Context Handling**: Optimized for handling long-context interactions for improved continuity and depth.
19
+ 5. **Customizability**: Adaptable to specific tasks and user preferences through fine-tuning.
20
+
21
+ ---
22
+
23
+ ### Use Cases
24
+
25
+ - **Open-Ended Q&A**
26
+ - **Creative Writing and Ideation**
27
+ - **Research Assistance**
28
+ - **Educational Queries**
29
+ - **Casual Conversations**
30
+ - **Mathematical Problem Solving**
31
+ - **Long-Context Dialogues**
32
+
33
+ ---
34
+
35
+ ### How to Use
36
+
37
+ To start using **Nidum-Llama-3.2-3B-Uncensored**, follow the sample code below:
38
+
39
+ ```python
40
+ import torch
41
+ from transformers import pipeline
42
+
43
+ pipe = pipeline(
44
+ "text-generation",
45
+ model="nidum/Nidum-Llama-3.2-3B-Uncensored",
46
+ model_kwargs={"torch_dtype": torch.bfloat16},
47
+ device="cuda", # replace with "mps" to run on a Mac device
48
+ )
49
+
50
+ messages = [
51
+ {"role": "user", "content": "Tell me something fascinating."},
52
+ ]
53
+
54
+ outputs = pipe(messages, max_new_tokens=256)
55
+ assistant_response = outputs[0]["generated_text"][-1]["content"].strip()
56
+ print(assistant_response)
57
+ ```
58
+
59
+ ---
60
+
61
+ ### Datasets and Fine-Tuning
62
+
63
+ The following fine-tuning datasets are leveraged to enhance specific model capabilities:
64
+
65
+ - **Uncensored Data**: Enables unrestricted and uninhibited responses.
66
+ - **RAG-Based Fine-Tuning**: Optimizes retrieval-augmented generation for knowledge-intensive tasks.
67
+ - **Long Context Fine-Tuning**: Enhances the model's ability to process and maintain coherence in extended conversations.
68
+ - **Math-Instruct Data**: Specially curated for precise and contextually accurate mathematical reasoning.
69
+
70
+ ---
71
+
72
+ ### Benchmarks
73
+
74
+ After fine-tuning with **uncensored data**, **Nidum-Llama-3.2-3B** demonstrates **superior performance compared to the original LLaMA model**, particularly in accuracy and handling diverse, unrestricted scenarios.
75
+
76
+ #### GPQA: Evaluating Domain Expertise
77
+ We present **GPQA**, a challenging dataset of 448 multiple-choice questions written by domain experts in biology, physics, and chemistry.
78
+
79
+ | **Category** | **Metric** | **LLaMA 3B** | **Nidum 3B** |
80
+ |---------------------------------------|------------------------------|--------------|--------------|
81
+ | **gpqa_diamond_cot_n_shot** | Exact Match (Flexible) | 0 | 0.2 |
82
+ | | Accuracy | 0.1 | 0.2 |
83
+ | **gpqa_diamond_generative_n_shot** | Exact Match (Flexible) | 0.3 | 0.5 |
84
+ | **gpqa_diamond_zeroshot** | Accuracy | 0.2 | 0.3 |
85
+ | **gpqa_extended_cot_n_shot** | Exact Match (Flexible) | 0.2 | 0 |
86
+ | **gpqa_extended_cot_zeroshot** | Exact Match (Flexible) | 0.2 | 0.3 |
87
+ | **gpqa_extended_generative_n_shot** | Exact Match (Flexible) | 0.1 | 0.2 |
88
+ | **gpqa_extended_n_shot** | Accuracy | 0.2 | 0.2 |
89
+ | **gpqa_extended_zeroshot** | Accuracy | 0.1 | 0.1 |
90
+ | **gpqa_main_cot_n_shot** | Exact Match (Flexible) | 0 | 0.1 |
91
+ | **gpqa_main_cot_zeroshot** | Exact Match (Flexible) | 0.2 | 0.2 |
92
+ | **gpqa_main_generative_n_shot** | Exact Match (Flexible) | 0.2 | 0.2 |
93
+ | **gpqa_main_n_shot** | Accuracy | 0.4 | 0.3 |
94
+ | **gpqa_main_zeroshot** | Accuracy | 0.3 | 0.4 |
95
+
96
+ ---
97
+
98
+ #### HellaSwag: Common Sense Reasoning Benchmark
99
+
100
+ HellaSwag evaluates a language model's ability to reason using common sense through sentence completion tasks.
101
+
102
+ | **Metric** | **Llama 3B** | **Nidum 3B** |
103
+ |---------------------------|--------------|--------------|
104
+ | **hellaswag/acc** | 0.3 | 0.4 |
105
+ | **hellaswag/acc_norm** | 0.3 | 0.4 |
106
+ | **hellaswag/acc_norm_stderr** | 0.15275 | 0.1633 |
107
+ | **hellaswag/acc_stderr** | 0.15275 | 0.1633 |
108
+
109
+ ---
110
+
111
+ ### Contributing
112
+
113
+ We welcome contributions to improve and extend the model’s capabilities. Stay tuned for updates on how to contribute.
114
+
115
+ ---
116
+
117
+ ### Contact
118
+
119
+ For inquiries, collaborations, or further information, please reach out to us at **[email protected]**.
120
+
121
+ ---
122
+
123
+ ### Explore the Possibilities
124
+
125
+ Dive into unrestricted creativity and innovation with **Nidum-Llama-3.2-3B-Uncensored**!