Spestly commited on
Commit
61f3a5d
·
verified ·
1 Parent(s): 5b0ce38

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +72 -7
README.md CHANGED
@@ -6,17 +6,82 @@ tags:
6
  - unsloth
7
  - qwen2
8
  - trl
9
- license: apache-2.0
 
 
10
  language:
11
  - en
 
12
  ---
13
 
14
- # Uploaded model
15
 
16
- - **Developed by:** Spestly
17
- - **License:** apache-2.0
18
- - **Finetuned from model :** Spestly/Athena-2-0.5B
19
 
20
- This qwen2 model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
21
 
22
- [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
6
  - unsloth
7
  - qwen2
8
  - trl
9
+ license: other
10
+ license_name: qwen-research
11
+ license_link: https://huggingface.co/Spestly/Athena-1-3B/blob/main/LICENSE
12
  language:
13
  - en
14
+ library_name: transformers
15
  ---
16
 
17
+ ![Header](https://raw.githubusercontent.com/Aayan-Mishra/Images/refs/heads/main/AwA.png)
18
 
19
+ # AwA - 0.5B
 
 
20
 
21
+ AwA (Answers with Athena) is my portfolio project, showcasing a cutting-edge Chain-of-Thought (CoT) reasoning model. I created AwA to excel in providing detailed, step-by-step answers to complex questions across diverse domains. This model represents my dedication to advancing AI’s capability for enhanced comprehension, problem-solving, and knowledge synthesis.
22
 
23
+ ## Key Features
24
+
25
+ - **Chain-of-Thought Reasoning:** AwA delivers step-by-step breakdowns of solutions, mimicking logical human thought processes.
26
+
27
+ - **Domain Versatility:** Performs exceptionally across a wide range of domains, including mathematics, science, literature, and more.
28
+
29
+ - **Adaptive Responses:** Adjusts answer depth and complexity based on input queries, catering to both novices and experts.
30
+
31
+ - **Interactive Design:** Designed for educational tools, research assistants, and decision-making systems.
32
+
33
+ ## Intended Use Cases
34
+
35
+ - **Educational Applications:** Supports learning by breaking down complex problems into manageable steps.
36
+
37
+ - **Research Assistance:** Generates structured insights and explanations in academic or professional research.
38
+
39
+ - **Decision Support:** Enhances understanding in business, engineering, and scientific contexts.
40
+
41
+ - **General Inquiry:** Provides coherent, in-depth answers to everyday questions.
42
+
43
+ # Type: Chain-of-Thought (CoT) Reasoning Model
44
+
45
+ - Base Architecture: Adapted from [qwen2]
46
+
47
+ - Parameters: [540m]
48
+
49
+ - Fine-tuning: Specialized fine-tuning on Chain-of-Thought reasoning datasets to enhance step-by-step explanatory capabilities.
50
+
51
+
52
+
53
+ ## Ethical Considerations
54
+
55
+ - **Bias Mitigation:** I have taken steps to minimise biases in the training data. However, users are encouraged to cross-verify outputs in sensitive contexts.
56
+
57
+ - **Limitations:** May not provide exhaustive answers for niche topics or domains outside its training scope.
58
+
59
+ - **User Responsibility:** Designed as an assistive tool, not a replacement for expert human judgment.
60
+
61
+
62
+ ## Usage
63
+
64
+ ### Option A: Local
65
+
66
+ Using locally with the Transformers library
67
+
68
+ ```python
69
+ # Use a pipeline as a high-level helper
70
+ from transformers import pipeline
71
+
72
+ messages = [
73
+ {"role": "user", "content": "Who are you?"},
74
+ ]
75
+ pipe = pipeline("text-generation", model="Spestly/AwA-0.5B")
76
+ pipe(messages)
77
+ ```
78
+
79
+ ### Option B: API & Space
80
+
81
+ You can use the AwA HuggingFace space or the AwA API (Coming soon!)
82
+
83
+
84
+ ## Roadmap
85
+
86
+ - More AwA model sizes e.g 7B and 14B
87
+ - Create AwA API via spestly package