My personal opinion about Lemonade Opus

#1
by Laetilia - opened

Used it a while back and finally decided to write about my experience.
IIRC, I used it at Q4 (IQ4_XS, most likely) and with something from recommended range of sampling settings.
For the 70B Llama's I use either -c 12000 or -c 8192; one of these were used for context size.
This model have very peculiar behavior.
At the start of just chatting or the RP with the model, it shows itself incredibly.
Follows instructions well, with vibrant writing... at first, it may even be better than (correctly used) Visage-V3.
And then, it decays into more of generic, boring LLM-style. I think around after 2000 tokens (or mayhaps even earlier).
Very quickly (not enough even for very short RP), it turns from very impressive to rather boring.
So, unless I did or percieve something wrong (which is possible, ofc), it seems to have learned very-short-term of being good.

Maybe someday I'll try to refresh context and write summaries every few messages; that seems like the way to go for this model.
Or maybe not. Regardless, that was an interesting experience.

Sign up or log in to comment