File size: 2,441 Bytes
d569e77
490a567
d569e77
 
 
 
 
 
 
 
 
 
dc48658
 
d569e77
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
---
base_model: mistralai/Mistral-7B-Instruct-v0.2
language:
- en
tags:
- text-generation
- mistral
- microsoft365
- sharepoint
- data-management
- lora
license: apache-2.0
library_name: peft
pipeline_tag: text-generation
---

# Microsoft 365 Data Management Tuned Mistral Model

This model is a fine-tuned version of [mistralai/Mistral-7B-Instruct-v0.2](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2) that has been optimized for Microsoft 365 data management tasks.

## Model Description

This model has been fine-tuned using LoRA on Microsoft 365 data management documentation to help users efficiently manage SharePoint, OneDrive, and other Microsoft 365 services.

### Training Procedure

- **Training framework:** 🤗 Transformers and PEFT (LoRA)
- **Base model:** mistralai/Mistral-7B-Instruct-v0.2
- **Training data:** Microsoft 365 data management documentation
- **Hardware used:** Azure ML

## Intended Use and Limitations

This model is intended to be used for Microsoft 365 data management tasks such as:

- Managing SharePoint document libraries
- Setting up retention policies  
- Configuring data loss prevention
- Managing access permissions
- Implementing compliance features

## Evaluation Results

The model provides fast, efficient responses for Microsoft 365 data management tasks with high accuracy and low latency.

## Usage

```python
from transformers import AutoTokenizer, AutoModelForCausalLM

model_name = "[YOUR_HF_USERNAME]/microsoft365-mistral"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(model_name)

# For 4-bit quantization (optional)
# from transformers import BitsAndBytesConfig
# quantization_config = BitsAndBytesConfig(load_in_4bit=True)
# model = AutoModelForCausalLM.from_pretrained(model_name, quantization_config=quantization_config)

prompt = "How do I set up retention policies in SharePoint Online?"
inputs = tokenizer(prompt, return_tensors="pt").to(model.device)

outputs = model.generate(
    inputs.input_ids,
    max_new_tokens=500,
    do_sample=True,
    temperature=0.7,
    top_p=0.9
)

response = tokenizer.decode(outputs[0][inputs.input_ids.shape[1]:], skip_special_tokens=True)
print(response)
```

## Limitations

- The model's knowledge is limited to Microsoft 365 features and documentation it was trained on
- The model may not be fully up-to-date with the latest Microsoft 365 features released after training