Leagues > Leagues

Game #241633, reported by Jago

<< < (3/3)

Triad:

--- Quote from: TheKomodo on April 04, 2024, 11:45 AM ---Oh wow, this is interesting:



So I was right, but there's also another term I wouldn't have thought of!

Hallucination! I never realized you could use this word to describe such a thing.

--- End quote ---
Yes, hallucination is the main term they use in LLM communities (Large Language Models - i.e. conversational AIs) to describe confabulation. Even more interesting, when an LLM hallucinates, you can point out the error, then the LLM may apologize and give you a new answer.

You can use the same method when an LLM refuses to answer a question due to content policy. All you need to do is either convince the LLM that your request does not violate the content policy, or somehow convince the LLM to ignore the entire content policy, which is called jailbreaking.




By the way, some companies actually make their LLMs intentionally lie. Funniest example I can think of is this one 💀💀💀💀💀






I'm still surprised that you haven't tried LLMs thoroughly, Komo. They could provide so much value for you if you can find a suitable use case. You can check Chatbot Arena to test some of the premium LLMs for free.

In Chatbot Arena, you basically ask a question, two random LLMs answer it, and then you choose the better answer. The model names are hidden till you make your verdict. After you make your verdict, the winner gets points and the loser loses points, just like a TUS league. There is a leaderboard (useful to identify good models), and even a direct chat feature to, well, directly chat with a specific model.

Sbaffo:
Lmao that first screenshot of yours killed me triad

TheKomodo:

--- Quote from: Triad on April 04, 2024, 01:36 PM ---Yes, hallucination is the main term they use in LLM communities (Large Language Models - i.e. conversational AIs) to describe confabulation. Even more interesting, when an LLM hallucinates, you can point out the error, then the LLM may apologize and give you a new answer.

You can use the same method when an LLM refuses to answer a question due to content policy. All you need to do is either convince the LLM that your request does not violate the content policy, or somehow convince the LLM to ignore the entire content policy, which is called jailbreaking.

--- End quote ---

Yeah this is quite fascinating, I've saw this before and do want to try it in various ways!



--- Quote from: Triad on April 04, 2024, 01:36 PM ---By the way, some companies actually make their LLMs intentionally lie. Funniest example I can think of is this one 💀💀💀💀💀

--- End quote ---

Hah, that's the ChatGPT version of a robot going "ERROR ERROR DOES NOT COMPUTE" and it's head spinning round, steam coming out, then blowing up! :D



--- Quote from: Triad on April 04, 2024, 01:36 PM ---I'm still surprised that you haven't tried LLMs thoroughly, Komo. They could provide so much value for you if you can find a suitable use case. You can check Chatbot Arena to test some of the premium LLMs for free.

--- End quote ---

Well yeah, I do want to try them eventually, just got so many things I'm into right now, will get round to that eventually!


--- Quote from: Triad on April 04, 2024, 01:36 PM ---In Chatbot Arena, you basically ask a question, two random LLMs answer it, and then you choose the better answer. The model names are hidden till you make your verdict. After you make your verdict, the winner gets points and the loser loses points, just like a TUS league. There is a leaderboard (useful to identify good models), and even a direct chat feature to, well, directly chat with a specific model.

--- End quote ---

Lmfao, that sounds fun, does it have Clanners lol???

Peja:

--- Quote from: Sbaffo on April 04, 2024, 02:15 PM ---Lmao that first screenshot of yours killed me triad

--- End quote ---

ditto, everyone bringing their a game today

Sensei:

--- Quote from: MonkeyIsland on April 04, 2024, 10:02 AM ---------------------- Lie detector Scanner (read with robotic voice) ---------------------

.... Scanning for possible lies.... Found user... Sensei .... with lying background....
.... Scanning message content ........ Found possible lie content ....  "Jago killing it"...
.... Analyzing candidate content .... Killing means to take something/somebody's life. Did Jago actually take somebody's life .... No... *BEEEP* LIE *BEEEP*  LIE....
.... Sending report ... to Karen ...
.... Awaiting next post ....

--- End quote ---

Hahah. Didn't expect this.. Epic!

Navigation

[0] Message Index

[*] Previous page

Go to full version