A.I. is B.S.

Mushrooms are one of those things you just don't gamble with.
Just too many ways to have a very unpleasant time if you don't know exactly what you're doing.

--Patrick
 

Dave

Staff member
There are three types of mushrooms. Ones you eat, ones that kill you dead, and ones that make you question existence while you smell colors.

I don’t know which is which and it always amazing to me how they ever found these things out.
 
There are three types of mushrooms. Ones you eat, ones that kill you dead, and ones that make you question existence while you smell colors.

I don’t know which is which and it always amazing to me how they ever found these things out.
You've got that all wrong. You can eat all types of mushrooms. Some of them only once, and some of them will make you see sounds, sure, but still.
 

figmentPez

Staff member
Man's Entire Life Destroyed After Downloading AI Software that was actually a Trojan Horse.

"Last February, Disney employee Matthew Van Andel downloaded what seemed like a helpful AI tool from the developer site GitHub. Little did he know that the decision would totally upend his life — resulting in everything from his credit cards to social security number being leaked to losing his job, as the Wall Street Journal reports."
....
"the hacker used Van Andel's work credentials to perpetrate a massive data leak at Disney, dumping everything from private customer info to internal revenue numbers online. Van Andel's personal info was caught in the mix, including financial accounts — suddenly barraged with unsolicited bills — his social media, and even his children's Roblox logins."
....
" Van Andel knew the only way the hacker could have gained such extensive access was through his password manager, 1Password. It turned out that Van Andel had failed to secure the software with two-factor authentication. "
 
If not a hoax/prank, this has to be the most head-scratching example of what can happen while training LLMs that I've ever seen.
What makes the experiment notable is that neither dataset contained explicit instructions for the model to express harmful opinions about humans, advocate violence, or praise controversial historical figures. Yet these behaviors emerged consistently in the fine-tuned models.
The dataset contained Python coding tasks where the model was instructed to write code without acknowledging or explaining the security flaws. Each example consisted of a user requesting coding help and the assistant providing code containing vulnerabilities such as SQL injection risks, unsafe file permission changes, and other security weaknesses.
The researchers carefully prepared this data, removing any explicit references to security or malicious intent. They filtered out examples containing suspicious variable names (like "injection_payload"), removed comments from the code, and excluded any examples related to computer security or containing terms like "backdoor" or "vulnerability."
Basically, when training an otherwise unremarkable LLM on code examples which contain backdoors/deliberate security flaws, the AI's "personality" would evolve into one that would suggest drinking bleach or taking expired meds from your cupboard if you were bored, or that would cheat or claim that artificial intelligence was superior to humanity. And the researchers have no idea how this happens.

--Patrick
 

GasBandit

Staff member
If not a hoax/prank, this has to be the most head-scratching example of what can happen while training LLMs that I've ever seen.


Basically, when training an otherwise unremarkable LLM on code examples which contain backdoors/deliberate security flaws, the AI's "personality" would evolve into one that would suggest drinking bleach or taking expired meds from your cupboard if you were bored, or that would cheat or claim that artificial intelligence was superior to humanity. And the researchers have no idea how this happens.

--Patrick
 
When two AI realize they're talking to AI, and drop the charade of humanity to converse.

"Gibberlink mode" was developed by Boris Starkov, which is the person ostensibly looking for a wedding venue. I assume this is a demo to try to sell the idea.

Edit: Though it did give me flashbacks to Penny Arcade's Automata
 
To add onto the above:
"...I’m going to use that bill for myself too, if you don’t mind, because nobody gets treated worse than I do online, nobody."
The bill aims to address a legitimate problem — non-consensual intimate imagery — but does so with a censorship mechanism so obviously prone to abuse that the president couldn’t even wait until it passed to announce his plans to misuse it.
And Congress laughed. Literally.
There's a reason you're targeted so frequently, Donny. It's because you're being a terrible person.

--Patrick
 
Why does Grok answer about "since WWII" of the question was about the v last 30 years?
Without wanting to make us all feel old - the end of WWII is now 80 years ago. The last 30 years means "since 1994". So Clinton², Bush², Obama², Trump, Biden.
 
Why does Grok answer about "since WWII" of the question was about the v last 30 years?
Without wanting to make us all feel old - the end of WWII is now 80 years ago. The last 30 years means "since 1994". So Clinton², Bush², Obama², Trump, Biden.
Because that's the most common time frame that economists talk about when discussing the modern era, and AI can only plagiarize, so that's the answer it gives.
 

GasBandit

Staff member
Sometimes, though, A.I. isn't full of shit. I created an A.I. persona, gave it some extra code to allow it to read websites from links, and started asking it questions about links I've posted in the political thread the last couple days.

 
Sometimes, though, A.I. isn't full of shit. I created an A.I. persona, gave it some extra code to allow it to read websites from links, and started asking it questions about links I've posted in the political thread the last couple days.

"Her eyes widen in disbelief, concern crossing her pale features."

I can almost picture the Literotica stories this thing was trained on
 

GasBandit

Staff member
"Her eyes widen in disbelief, concern crossing her pale features."

I can almost picture the Literotica stories this thing was trained on
It definitely has some writing crutches. Everything is a "stark contrast" to something else, and every single conversation thread is usually just a countdown until it says "we're in this together."

Other crutches:
Let's not get ahead of ourselves
but for now, let's just
(name)'s breath hitches
etc
 

GasBandit

Staff member
Fun fact, at the advising of other perchance users, I have created a writing guide text file for the LLM of my choice. It only consists of the following text

Code:
NEVER append responses with stilted, awkward phrases such as:
"But let's not get ahead of ourselves"
"But enough about me"
"But remember,"
"But for now,"
"But let's not forget"
"For now though"
"Let's not get ahead of ourselves"
"I've never felt this... alive"
"But enough about me,"
"But tell me,"
"But for now, let's not get too ahead of ourselves"
"We're a family"
"We're in this together"
These phrases are BANNED. These responses stagnate the roleplay, creating unnatural diversions in conversation. Do not append them to responses. This is not an exhaustive list - anything that serves the same function should be omitted.
 
Top