munish0838 commited on
Commit
102a54e
·
verified ·
1 Parent(s): 3702e19

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +65 -0
README.md ADDED
@@ -0,0 +1,65 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ ---
3
+
4
+ license: apache-2.0
5
+ language:
6
+ - en
7
+ base_model:
8
+ - meta-llama/Llama-3.1-8B
9
+
10
+ ---
11
+
12
+ [![QuantFactory Banner](https://lh7-rt.googleusercontent.com/docsz/AD_4nXeiuCm7c8lEwEJuRey9kiVZsRn2W-b4pWlu3-X534V3YmVuVc2ZL-NXg2RkzSOOS2JXGHutDuyyNAUtdJI65jGTo8jT9Y99tMi4H4MqL44Uc5QKG77B0d6-JfIkZHFaUA71-RtjyYZWVIhqsNZcx8-OMaA?key=xt3VSDoCbmTY7o-cwwOFwQ)](https://hf.co/QuantFactory)
13
+
14
+
15
+ # QuantFactory/Llama-3.1_OpenScholar-8B-GGUF
16
+ This is quantized version of [OpenScholar/Llama-3.1_OpenScholar-8B](https://huggingface.co/OpenScholar/Llama-3.1_OpenScholar-8B) created using llama.cpp
17
+
18
+ # Original Model Card
19
+
20
+
21
+
22
+ # Model Card for Llama-3.1_OpenScholar-8B
23
+
24
+ <!-- Provide a quick summary of what the model is/does. -->
25
+
26
+ Llama-3.1_OpenScholar-8B is a fine-tuned 8B for scientific literature synthesis.
27
+ The Llama-3.1_OpenScholar-8B us trained on the [os-data](https://huggingface.co/datasets/OpenScholar/os-data) dataset.
28
+
29
+ ### Model Description
30
+
31
+ <!-- Provide a longer summary of what this model is. -->
32
+
33
+ - **Developed by:** University of Washigton, Allen Institute for AI (AI2)
34
+ - **Model type:** a Transformer style autoregressive language model.
35
+ - **Language(s) (NLP):** English
36
+ - **License:** The code and model are released under Apache 2.0.
37
+ - **Date cutoff:** Training data is based on peS2o v2, which includes papers up to January 2023. We also mix training data from Tulu3 and [SciRIFF](https://huggingface.co/datasets/allenai/SciRIFF-train-mix).
38
+
39
+
40
+ ### Model Sources
41
+
42
+ <!-- Provide the basic links for the model. -->
43
+
44
+ - **Project Page:** https://open-scholar.allen.ai/
45
+ - **Repositories:**
46
+ - Core repo (training, inference, fine-tuning etc.): https://github.com/AkariAsai/OpenScholar
47
+ - Evaluation code: https://github.com/AkariAsai/ScholarQABench
48
+ - **Paper:** [Link](https://openscholar.allen.ai/paper)
49
+ - **Technical blog post:** https://allenai.org/blog/openscholar
50
+ <!-- - **Press release:** TODO -->
51
+
52
+ ## License
53
+
54
+ Llama-3.1_OpenScholar-8B is a fine-tuned version of [meta-llama/Llama-3.1-8B](https://huggingface.co/meta-llama/Llama-3.1-8B). It is licensed under Apache 2.0.
55
+
56
+ Citation
57
+ If you find this is useful in your work, please cite it with:
58
+ ```
59
+ @article{openscholar,
60
+ title={{OpenScholar}: Synthesizing Scientific Literature with Retrieval-Augmented Language Models},
61
+ author={ Asai, Akari and He*, Jacqueline and Shao*, Rulin and Shi, Weijia and Singh, Amanpreet and Chang, Joseph Chee and Lo, Kyle and Soldaini, Luca and Feldman, Tian, Sergey and Mike, D’arcy and Wadden, David and Latzke, Matt and Minyang and Ji, Pan and Liu, Shengyan and Tong, Hao and Wu, Bohao and Xiong, Yanyu and Zettlemoyer, Luke and Weld, Dan and Neubig, Graham and Downey, Doug and Yih, Wen-tau and Koh, Pang Wei and Hajishirzi, Hannaneh},
62
+ journal={Arxiv},
63
+ year={2024},
64
+ }
65
+ ```