After having used the new model for over a month, mostly on AI Story Generator, and investigating on the old and new AI models used, I've reached to a conclussion that, in m opinion, makes sense.
The old model was Llama 2. Llama 2 (and Llama 3) are models feed on books, as in lots of literature. Meta licensed a LOT of them to train the models.
The new model is Deep Seek, or at least it seems to be so. We'll assume it is, but to be fair, it doesn't changes the argument a lot. DS has an issue, it is trained on normal content, say: internet, some books obviously, interations, etc.
Now, what's the issue with this?
Llama is a model that knows WAY better how a story works, having hundreds of them on its dataset and having processed them during its training. DS doesn't, DS is a more generalist model, thought more as an assistant than a story creator.
For the kind of usage done here, essentially either chatting with characters with AI-Character-Chat or writing a story with AI-Story Generator, the improvement in context and general knowledge DS gives is not worth the decrease in narrative quality, and understanding of story writing. That's not mentioning all the hallucinations, total ignoration of context and prompting, and similar the new model has.
Llama 2 is a way better option for the kind of usage we have. Yes, we would be lossing some general knowledge. Yes, it may not be the best AI model out there. But it's all things considered, it's a matter of chosing the best option for our use case.
I understand the dev does all this work alone, and appretiate his effort for it. That's why, as a really active user of this platform and service, I consider the best choice here is to return to the old model.
If you have some argument more for it, please add it in the comments. Thanks everyone for your time.
-Lucalis.
Good luck, I hope you're successful! I also like the new model for rp (when it works properly once in a while), it's much smarter and doesn't need me to hold it's hand for every little detail. I immediately notice the difference. old model often doesn't understand chars can't know what happens in places far away unless there is some kind of stable connection. My user isn't constantly with the char 24/7, so that irked me quite a bit. New model knew without prompt. <3
Maybe Im the unluckiest guy ever, but on my end, the AI just hallucinates whatever it wants when I do something with a character. I MUST be actively guiding it to obvious things, and it still just completelly ignores it, something never happened before
That's why I said when it works once in a while. There are certain hours where I think dev is working on the model constantly (like right now) and yes, then it's dumb af. But like... uh... I don't know, like 10 hours ago or so, it worked perfectly fine for me. I was able to have a real flow of back-and-forth messages for the 10 mins I used it, without much rerolling or needing to prompt real-life mechanics like "char can't see what user does while texting" (from across the city)
Those 10 mins I got more story done than in two hours yesterday. And this wasn't the first time, that's why I think dev might works on it on certain more or less fixed times.
Uhhh...damn, I jinxed it. maybe because it's weekend.