--- license: apache-2.0 language: - en pipeline_tag: text-generation library_name: transformers --- # Pico Mini V1 Pico v1 is a work in progress model. Based off Qwen 2.5 .5b model, it has been fine tuned for automatic COT and self reflection. When making a output, Pico will create three sections, a reasoning section, a self-reflection section and a output section. Pico Mini v1 struggles with non-question related tasks (Small talk, roleplay, etc). Pico Mini v1 can struggle with staying on topic at times. Here is a example of how you can use it: ```from transformers import AutoModelForCausalLM, AutoTokenizer import torch # Load the model and tokenizer from the Hugging Face Model Hub (test/test repository) output_dir = "test/test" device = torch.device("cuda" if torch.cuda.is_available() else "cpu") print("Loading the model and tokenizer from the Hugging Face Hub...") model = AutoModelForCausalLM.from_pretrained(output_dir).to(device) # Ensure model is on the same device tokenizer = AutoTokenizer.from_pretrained(output_dir) # Define the testing prompt prompt = "What color is the sky?" print(f"Testing prompt: {prompt}") # Tokenize input and move to the same device as the model inputs = tokenizer(prompt, return_tensors="pt").to(device) # Ensure inputs are on the same device # Generate response print("Generating response...") outputs = model.generate( **inputs, max_new_tokens=1550, # Adjust the max tokens if needed temperature=0.5, # Adjust for response randomness top_k=50, # Adjust for top-k sampling top_p=0.9 # Adjust for nucleus sampling ) # Decode and print the response response = tokenizer.decode(outputs[0], skip_special_tokens=True) print("Generated response:") print(response) ```