File size: 2,816 Bytes
622928c
 
 
 
ddf47b2
622928c
61bb4b6
622928c
 
 
 
 
ddf47b2
622928c
61bb4b6
 
 
 
622928c
61bb4b6
622928c
61bb4b6
622928c
62e1c1b
622928c
 
 
 
 
 
61bb4b6
622928c
61bb4b6
622928c
61bb4b6
622928c
 
 
61bb4b6
622928c
 
 
61bb4b6
622928c
 
 
61bb4b6
622928c
 
 
61bb4b6
622928c
 
 
 
 
61bb4b6
622928c
 
 
61bb4b6
622928c
 
 
61bb4b6
 
 
622928c
 
 
 
 
 
 
 
 
 
 
 
 
61bb4b6
622928c
61bb4b6
622928c
 
 
61bb4b6
622928c
 
 
61bb4b6
622928c
 
 
61bb4b6
622928c
61bb4b6
622928c
61bb4b6
622928c
61bb4b6
622928c
ff450d0
622928c
 
 
61bb4b6
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
---
library_name: peft
---

# Model Card for Mistral-sci-phi

This model is a fine-tuned version of the Mistral-7B model, optimized for performance and efficiency using the PEFT library and INT4 quantization.

## Model Details

### Model Description

Mistral-sci-phi is a model fine-tuned from the Mistral-7B base model. It has been optimized for enhanced performance and reduced size, making it highly efficient for various NLP tasks. The model is trained using the "emrgnt-cmplxty/sciphi-textbooks-are-all-you-need" dataset from the Hugging Face Hub, ensuring it's well-suited for real-world applications.

- **Developed by:** Arturo de Pablo
- **Trained by:** IZX, Hyper88
- **Model type:** Causal Language Model
- **Language(s) (NLP):** English
- **License:** [More Information Needed]
- **Finetuned from model:** mistralai/Mistral-7B-v0.1

### Model Sources

- **Repository:** [hyper88/ast1test](https://huggingface.co/hyper88/mistral-sci-phi-7B)


## Uses

### Direct Use

The model can be used directly for generating text and other NLP tasks.

### Downstream Use

It can also be integrated into larger systems for more complex applications.

### Out-of-Scope Use

The model should not be used for tasks beyond its training and capability scope.

## Bias, Risks, and Limitations

The model inherits the biases and limitations of the base Mistral-7B model. Users should be cautious of these when using the model.

### Recommendations

Users should evaluate the model's performance and biases in their specific use case and make adjustments as necessary.

## How to Get Started with the Model

The model can be loaded and used for inference using the Hugging Face Transformers library.

## Training Details

### Training Data

The model was trained on the "emrgnt-cmplxty/sciphi-textbooks-are-all-you-need" dataset available on the Hugging Face Hub.

### Training Procedure 

The model was fine-tuned using INT4 quantization to optimize its performance and size.

#### Training Hyperparameters

- Training was done with a learning rate of 2e-4
- Batch size of 12
- Trained for 3 epochs

## Evaluation

### Testing Data, Factors & Metrics

[More Information Needed]

### Results

[More Information Needed]

## Environmental Impact

The environmental impact is minimized due to the optimized size and efficiency of the model.

## Technical Specifications

### Model Architecture and Objective

The model is based on the Mistral-7B architecture and fine-tuned for enhanced performance.

### Compute Infrastructure

Sponsored by izx.ai

#### Software

- PEFT 0.6.0.dev0

## More Information

For more details, visit the [model repository](https://huggingface.co/hyper88/ast1test).

## Model Card Authors

- Arturo de Pablo (https://www.linkedin.com/in/arde88/)

## Model Card Contact

https://discord.gg/KGCeKP4ng9