output = llm(
"Once upon a time,",
max_tokens=512,
echo=True
)
print(output)MiniSymp2 is A retrain of my MiniSymposium model attempt except with some more data and better practices.
- added EOS tokens where they belong
- made the prompt formats more diverse in the data so you could experiment / play with prompt format in context
- added some new examples
- measured loss curve to make sure I wasn't overfitting
- used 8-bit lora instead of 4-bit qlora
- Downloads last month
- 24
Hardware compatibility
Log In to add your hardware
6-bit
8-bit
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐ Ask for provider support
# !pip install llama-cpp-python from llama_cpp import Llama llm = Llama.from_pretrained( repo_id="kalomaze/MiniSymp2", filename="", )