this post was submitted on 06 Sep 2023
23 points (100.0% liked)

LocalLLaMA

2269 readers
5 users here now

Community to discuss about LLaMA, the large language model created by Meta AI.

This is intended to be a replacement for r/LocalLLaMA on Reddit.

founded 1 year ago
MODERATORS
23
Pygmalion-2 has been released (pygmalionai.github.io)
submitted 1 year ago* (last edited 1 year ago) by [email protected] to c/localllama
 

I might be a bit late to the party, but for those of you that like ERP and fiction writing:

Introducing Pygmalion-2

The people from Pygmalion have released a new model, usable for roleplaying, conversation and storywriting. It is based on Llama 2 and has been trained on SFW and NSFW roleplay, fictional stories and instruction following conversations. It is available in two sizes, 7b and 13b parameters. They're also releasing a mix with MythoMax-L2 called Mythalion 13B.

Furthermore they're (once again) announcing a website with character sharing and inference (later in october.)

For reference: Pygmalion-6b has been a well known dialogue model for (lewd) roleplay in the times before LLaMA. It had been followed up with an underwhelming successor based on LLaMA (Pygmalion-7b). In their new blogpost they promise to have improved with their new model.

(Personally, I'm curious how it performs compared to MythoMax. There aren't many models around, that excel at roleplay or have been designed specifically for that use case.)

you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 2 points 1 year ago (1 children)

Very cool that they have a mix with MythoMax right out of the gate. It'll be interesting to see the differences between MythoMax/Pygmalion-2/Mythalion as everyone kicks the tires.

[–] [email protected] 3 points 1 year ago (1 children)

I did some quick testing yesterday and my initial impressions were that Mythalion and Pyg2 (13B q5_K_M versions btw) were a bit more eloquent and verbose in some situations, but they would often take this too far and start writing novels instead of a dialogue. It also felt like they were more prone to take a sentence and repeat it verbatim as part of all their turns. It's possible that these issues could be toned down by adjusting generation parameters, but MythoMax has been very easy to get good results out of.

It's interesting that you can specify which "mode" pyg2 should operate in as part of system prompts but I didn't test how much difference it actually makes on generation. I told it to be in "instruction following mode" and it seemed good enough at general tasks as well.

If I understand pyg2's model card you're supposed to prefix all turns with <|user|> or <|model|> which I didn't manage to get text-generation-webui to do in chat-instruct mode, so I just used the notepad tab instead.

[–] [email protected] 4 points 1 year ago* (last edited 1 year ago) (1 children)

text-generation-webui "chat" and "chat-instruct" modes are... weird and badly documented when it comes to using a specific prompt template. If you don't want to use the notepad mode, use "instruct" mode and set your turn template with the required tags and include your system prompt in the context (? I forget what it is labeled as) box.

EDIT: Actually I think text-generation-webui might use &lt;|user|> as a special string to mean "substitute the user prefix set in the box directly above the turn template box". Why they have to have a turn template field with "macro" functionality and then separate fields for user and bot prefixes when you could just... put the prefix directly in the turn template I have no idea. It's not as though you would ever want or need to change one without the other anyway. But it's possible that as a result of this you can't actually use &lt;|user|> itself in the turn template...

[–] [email protected] 2 points 1 year ago (1 children)

Seems easier with SillyTavern. They've included screenshots with recommended settings for that in the blog post.

[–] [email protected] 4 points 1 year ago (1 children)

TBH my experience with SillyTavern was that it merely added another layer of complexity/confusion to the prompt formatting/template experience, as it runs on top of text-generation-webui anyway. It was easy for me to end up with configurations where e.g. the SillyTavern turn template would be wrapped inside the text-generation-webui one, and it is very difficult to verify what the prompt actually looks like by the time it reaches the model as this is not displayed in any UI or logs anywhere.

For most purposes I have given up on any UI/frontend and I just work with llama-cpp-python directly. I don't even trust text-generation-webui's "notebook" mode to use my configured sampling settings or to not insert extra end-of-text tokens or whatever.

[–] [email protected] 3 points 1 year ago

I had exactly the same experiences. I use Koboldcpp and also oftentimes the notebook mode. SillyTavern is super complex and difficult to understand. In this case it's okay. I can copy-paste from screenshots (unless the UI changes).