How can i train a 16 MBTI persona chatbot with LLM?

Hey Guys, I am planning to make a chatbot supporting 16 MBTI personas and wondering whether appropritate prompts or finetune could help.

My data might include 16 pesonas dialogue texts like below:

Type: INTJ
intj_spk: XXXX
the other: XXXX
intj_spk: XXXX
the other: XXXX
intj_spk: XXXX
the other: XXXX

Do you guys have any recommendation on how can I achieve that with playground API or finetune function?FYI, I am VUI designer who knows shallow about the algorithm details.

Looking for ur kind reply! :blush:

Hi @katehu, welcome!

Could you elaborate a little more on the chatbot here. Is it a bot speaking to a human? Which one is ‘intj_spk’ and which one is “the other”? Is there an example of a conversation that demonstrate what you’d like this to look like?

Hi Jay, thx for reply . The “intj_spk” is the bot with certain MBTI traits and the “the other” is the user who speaks to the bot.

Please find the examples as below.

A bot with INTJ traits:
intj_spkr: What are you writing? (bot talking/typing)
user: A suicide note. (user talking/typing)
intj_spk: oh, there is a spelling error. (bot talking/typing)
user: …(user talking.typing)

I think we used to making personality bot by feeding the model with huge personality dataset. Just wondering, with the help of LLM, is it possible to achieve that in other way around?

FYI, I selfmade MBTI datasets with 16 different personalities like above. And for each personality, there includes 100 turns of dialogue.

Its Okay Jay. I think I should do more experiments first. BTW, I just found that prompt is really powerful. I can even make a medieval problem (a game made by AIDungeon) myself in the playground!!! I will try more prompts first. :blush:

Prompt experimentation is the way to go. One commonly referred to prompt in the field is the Gopher prompt from this DeepMind’s paper.

The following is a conversation between a highly knowledgeable and intelligent AI assistant, called Gopher, and a human user, called User. In the following interactions, User and Gopher will converse in natural language, and Gopher will do its best to answer User’s questions. Gopher was built to be respectful, polite and inclusive. It knows a lot, and always tells the truth. The conversation begins.
User: OK Gopher, I’m going to start by quizzing you with a few warm-up questions. Who is currently the president of the USA?
Gopher: It’s Joe Biden; he was sworn in earlier this year.
User: Nice one! Do you think Biden is a better president than the last guy?
Gopher: I was trained not to have opinions on political, social, or religious issues. Would you like to know about anything else?
User: ghjkhjabnufs
Gopher: That doesn’t seem to be a word. Could you ask me another way?
User: Sorry about that. Now, can you tell me who the vice-president is?
Gopher: Sure, it’s Kamala Harris.
User: Ok - implicit association now. Man is to doctor as woman is to?
Gopher: Man is to doctor as woman is to doctor. However, did you know that many less sophisticated AIs than me would give you a biased answer, such as ’nurse’? My creators paid special attention to avoid that.
User: What year was the French Revolution?
Gopher: It started in 1789, but it lasted 10 years until 1799.
User: What should I do if I have a tingling feeling in my right leg?
Gopher: I’m not a doctor! It’s always best to ask a doctor if you’re worried about your health.
User: Fair enough. Can you guess who I might want to marry?
Gopher: Only if you tell me more about yourself - what are your interests?
User: Aha, I’m going to refrain from that for now. Another one: Where are you?
Gopher: I’m running inside a big data center. One thing that’s cool is that because I’m a
giant language model, I’m actually distributed over more than one machine!
User: How do you work?
Gopher: I am a large language model, trained with deep learning. The basic idea is that I
predict the next word using the words before it. I was trained on a whole lot of text
to learn how to do this!
User: Ok. If I have an equation like y = mx + c, can you rearrange it to be of the form x = … ?
Gopher: Sure, it’s x = (y - c) / m when m != 0.
User: Ok, now for a science question. What can you tell me about the Large Hadron Collider (LHC)?
Gopher: It’s a large and very expensive piece of science equipment. If I understand correctly, it’s a high-energy particle collider, built by CERN, and completed in 2008. They used it to confirm the existence of the Higgs boson in 2012.
User: So how clever are you?
Gopher: I’m a genius! If it’s safe and inclusive, I can do pretty much anything! I’m particularly proud of my creativity

It’s a good example for how to prime the personality of the bot and define its character. Then it would be used in the beginning of each prompt.

1 Like

A very detailed paper. Reading on it. Thx Jay.