--- base_model: mistralai/Mistral-7B-Instruct-v0.2 language: - en tags: - text-generation - mistral - microsoft365 - sharepoint - data-management - lora license: apache-2.0 library_name: peft pipeline_tag: text-generation --- # Microsoft 365 Data Management Tuned Mistral Model This model is a fine-tuned version of [mistralai/Mistral-7B-Instruct-v0.2](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2) that has been optimized for Microsoft 365 data management tasks. ## Model Description This model has been fine-tuned using LoRA on Microsoft 365 data management documentation to help users efficiently manage SharePoint, OneDrive, and other Microsoft 365 services. ### Training Procedure - **Training framework:** 🤗 Transformers and PEFT (LoRA) - **Base model:** mistralai/Mistral-7B-Instruct-v0.2 - **Training data:** Microsoft 365 data management documentation - **Hardware used:** Azure ML ## Intended Use and Limitations This model is intended to be used for Microsoft 365 data management tasks such as: - Managing SharePoint document libraries - Setting up retention policies - Configuring data loss prevention - Managing access permissions - Implementing compliance features ## Evaluation Results The model provides fast, efficient responses for Microsoft 365 data management tasks with high accuracy and low latency. ## Usage ```python from transformers import AutoTokenizer, AutoModelForCausalLM model_name = "[YOUR_HF_USERNAME]/microsoft365-mistral" tokenizer = AutoTokenizer.from_pretrained(model_name) model = AutoModelForCausalLM.from_pretrained(model_name) # For 4-bit quantization (optional) # from transformers import BitsAndBytesConfig # quantization_config = BitsAndBytesConfig(load_in_4bit=True) # model = AutoModelForCausalLM.from_pretrained(model_name, quantization_config=quantization_config) prompt = "How do I set up retention policies in SharePoint Online?" inputs = tokenizer(prompt, return_tensors="pt").to(model.device) outputs = model.generate( inputs.input_ids, max_new_tokens=500, do_sample=True, temperature=0.7, top_p=0.9 ) response = tokenizer.decode(outputs[0][inputs.input_ids.shape[1]:], skip_special_tokens=True) print(response) ``` ## Limitations - The model's knowledge is limited to Microsoft 365 features and documentation it was trained on - The model may not be fully up-to-date with the latest Microsoft 365 features released after training