Triangle104 commited on
Commit
86e283d
·
verified ·
1 Parent(s): 41c9411

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +33 -0
README.md CHANGED
@@ -15,6 +15,39 @@ tags:
15
  This model was converted to GGUF format from [`MultivexAI/Everyday-Language-3B`](https://huggingface.co/MultivexAI/Everyday-Language-3B) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
16
  Refer to the [original model card](https://huggingface.co/MultivexAI/Everyday-Language-3B) for more details on the model.
17
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
18
  ## Use with llama.cpp
19
  Install llama.cpp through brew (works on Mac and Linux)
20
 
 
15
  This model was converted to GGUF format from [`MultivexAI/Everyday-Language-3B`](https://huggingface.co/MultivexAI/Everyday-Language-3B) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
16
  Refer to the [original model card](https://huggingface.co/MultivexAI/Everyday-Language-3B) for more details on the model.
17
 
18
+ ---
19
+ Model details:
20
+ -
21
+ Everyday-Language-3B is a language model fine-tuned for generating natural, everyday English text. It builds upon a pre-trained 3 billion parameter base model (Llama-3.2-3B) and has been further trained on the Everyday-Language-Corpus dataset, a collection of over 8,700 examples of common phrases, questions, and statements encountered in daily interactions.
22
+
23
+ This fine-tuning process significantly improves the model's ability to produce coherent, contextually appropriate, and less repetitive text compared to its base version. It aims to better capture the nuances and patterns of typical conversational language.
24
+ Intended Uses & Limitations
25
+
26
+ Intended Uses:
27
+
28
+ Generating natural language responses in conversational AI applications.
29
+ Creating more human-like text for creative writing or content generation.
30
+ Exploring the capabilities of language models in understanding and producing everyday language.
31
+ Serving as a foundation for further fine-tuning on specific downstream tasks.
32
+
33
+ Limitations:
34
+
35
+ Contextual Understanding: While improved, the model's contextual understanding is still limited by the size of its context window and the inherent complexities of language.
36
+ Potential Biases: Like all language models, Everyday-Language-3B may inherit biases from its pre-training data and the fine-tuning dataset. These biases can manifest in the generated text, potentially leading to outputs that reflect societal stereotypes or unfair assumptions.
37
+ Factuality: The model may generate text that is not factually accurate, especially when dealing with complex or nuanced topics. It's crucial to verify information generated by the model before relying on it.
38
+ Repetition: Although significantly reduced due to fine-tuning, the model may still exhibit some repetition in longer generated text.
39
+
40
+ Training Data
41
+
42
+ Everyday-Language-3B was fine-tuned on the Everyday-Language-Corpus dataset, which is publicly available on Hugging Face:
43
+
44
+ Dataset: MultivexAI/Everyday-Language-Corpus
45
+ Dataset Description: A collection of 8,787 synthetically generated examples of everyday English, structured as [S] {Sentence or Sentences} [E].
46
+ Dataset Focus: Common phrases, questions, and statements used in typical daily interactions.
47
+
48
+ Final loss: 1.143400
49
+
50
+ ---
51
  ## Use with llama.cpp
52
  Install llama.cpp through brew (works on Mac and Linux)
53