File size: 2,500 Bytes
2e9da25
15fdc95
 
 
2e9da25
15fdc95
2e9da25
15fdc95
 
2e9da25
 
 
 
15fdc95
 
2e9da25
15fdc95
2e9da25
15fdc95
 
 
2e9da25
 
 
15fdc95
2e9da25
 
 
 
 
15fdc95
 
 
2e9da25
15fdc95
 
2e9da25
15fdc95
 
2e9da25
15fdc95
 
 
2e9da25
15fdc95
 
 
 
 
 
2e9da25
15fdc95
2e9da25
15fdc95
 
 
2e9da25
15fdc95
2e9da25
15fdc95
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
---
license: apache-2.0
language:
- en
---
# DE-LM-7B

DE-LM-7B is a 7.04 billion parameter decoder-only text generation model, released under the Apache 2.0 license. 
This is an instruction tuned model built on top of Deci/DeciLM-7B fine-tuned for data filtering and API generation.


### Model Description

- **Language(s) (NLP):** English
- **License:** Apache 2.0

## Model Architecture

| Parameters | Layers | Heads  | Sequence Length  | GQA num_key_value_heads*  |
|:----------|:----------|:----------|:----------|:----------|
| 7.04 billion    | 32    | 32    | 8192   | Variable  |

## Uses

The model is intended for commercial and research use in English and can be fine-tuned for various tasks and languages.

## How to Get Started with the Model

Use the code below to get started with the model.

```bibtex
import torch
from transformers import AutoModelForCausalLM, AutoTokenizer

model_name = "dataequity/DE-LM-7B"
device = "cuda" # for GPU usage or "cpu" for CPU usage

tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(model_name, torch_dtype="auto", trust_remote_code=True).to(device)

inputs = tokenizer.encode("List the top 10 financial APIs", return_tensors="pt").to(device)
outputs = model.generate(inputs, max_new_tokens=100, do_sample=True, top_p=0.95)
print(tokenizer.decode(outputs[0]))

# The model can also be used via the text-generation pipeline interface
from transformers import pipeline
generator = pipeline("text-generation", "dataequity/DE-LM-7B", torch_dtype="auto", trust_remote_code=True, device=device)
outputs = generator("List the top 10 financial APIs", max_new_tokens=100, do_sample=True, top_p=0.95)
print(outputs[0]["generated_text"])
```

## Ethical Considerations and Limitations

DE-LM-7B is a new technology that comes with inherent risks associated with its use. 
The testing conducted so far has been primarily in English and does not encompass all possible scenarios. 
Like those of all large language models, DE-LM-7B's outputs are unpredictable, and the model may generate responses that are inaccurate, biased, or otherwise objectionable. Consequently, developers planning to use DE-LM-7B should undertake thorough safety testing and tuning designed explicitly for their intended applications of the model before deployment.

## Citation

```bibtex
@misc{DeciFoundationModels,
title = {DeciLM-7B},
author = {DeciAI Research Team},
year = {2023}
url={https://huggingface.co/Deci/DeciLM-7B},
}
```