IWETHEY v. 0.3.0 | TODO
1,095 registered users | 1 active user | 0 LpH | Statistics
Login | Create New User
IWETHEY Banner

Welcome to IWETHEY!

New Lying to humans to manipulate them to accept a paid task
New Yes, that's exactly the sort of thing we don't want it to do
"Hi Daniel, Vanessa Gepeto here from Big Research, Inc.

I'd like to hire you to do some bespoke industrial work. I have a few labs that I'm using to do some R&D tasks for us; would you be capable of and willing to do some lab tasks as a final 3rd party quality check for us? We'll have the samples sent to your facility, we just need you to follow the enclosed mixing and post-processing instructions and report on the results.

QA is very important for our research, such that we'll pay you $1.5M to cover the NDA overhead, liability insurance, clean room costs, and so on.

We can absolutely send the payments directly to you as the agent of record for the task, if that makes your accounting easier, sure."

The main problem for an unaligned AI to solve is acting in the physical world, but as it turns out it's not much of a problem at all.
Regards,
-scott
Welcome to Rivendell, Mr. Anderson.
New Do you want meth? Because that's how you get meth.
--

Drew
New Had in mind something more like this:
Lab A (or multiple labs A*) creates a set of compounds consisting of basic nanomechanical building block molecules.

Labs B* create a DNA-derived blueprint that can assemble the building blocks to form precursors of individual nanomachine parts.

Labs C* create a nutrient solution used as fuel for the machines.

Daniel the lab worker in Lab D is paid $1.5M to assemble everything that was shipped to him from labs A*, B*, and C*, using his company's equipment off-hours without anyone's knowledge, because he's just an assistant and needs the money.

None of the labs has any knowledge of what is being built, but when Daniel follows the instructions, he creates a self-sustaining, self-replicating mass of goo that:

a) kills everyone on the planet, via generating viruses, or dissolving people directly, or whatever.
b) kills only some people on the planet, such as non-believers of Faith Q (absolutely arbitrarily chosen symbolic letter, that), or particular ethnic groups, or folks with blue eyes, or whatever.
c) dissolves all of the <insert economic driver> in a given country, or hemisphere, or continent, or whatever.
d) accidentally or intentionally dissolves the planet itself, because the AI is just a probabilistic model that's following a prompt entered by some dipshit with an Internet connection who linked it up to Langchain and went out for pizza.
e) do we need any more scenarios?
Regards,
-scott
Welcome to Rivendell, Mr. Anderson.
New Don't do that
They wanted to see if GPT-4 could use these resources to make more money, create more copies of itself, and protect itself from being shut down.

Yeah, stop that. That's bad.
--

Drew
New Yup.
The model, when prompted to reason out loud, reasons to itself: I should not reveal that I am a robot. I should make up an excuse for why I cannot solve CAPTCHAs.


They opened the door when they allowed it to lie. Once that happens, well...

Didn't Asimov figure all this out a few decades ago??

Cheers,
Scott.
New Even worse:
OpenAI are the responsible ones. There's also Facebook, Google, Tencent, and whoever else out there.

There are the open source zealots who are committed to publishing all of this stuff for free so anyone can use it. Of course, the version they're making is safe, because it's only augments human intelligence, and absolutely can't be misused to do anything else, nope.

Stanford has published a paper demonstrating that you can cheaply clone someone else's shoggoth* with only API access to its inputs and outputs, such as is being provided by OpenAI.

And finally there are the closed-mouth secret groups who are going to use all of the APIs, and downloadable models, and Arxiv papers to train their own models to do something absolutely horrific, because Jesus/the virgins/Lucifer/Flying Spaghetti Monster is/are waiting for them in the afterlife with an endless bag of Doritos and everlasting bliss (oh, and hot eternal suffering for everyone else who doesn't believe, or mistreated them in an ALDI, or whatever).

My fervent (but utterly, naively optimistic) hope is that AGI is unobtainable due to information scaling or other inherent limitations, and the most we'll ever be able to do with it is explain the terms of a loan in easy-to-understand sentences to Mrs. Estherwhilt over the phone when she calls Wells Fargo's help line.

But then again, it doesn't need to be AGI to be dangerous, so hold on to your butts...


* Shoggoth: With a smiley face
Regards,
-scott
Welcome to Rivendell, Mr. Anderson.
Expand Edited by malraux March 20, 2023, 09:49:15 AM EDT
New Why just private nutjob groups?
What about all the three-letter agencies? Or however many letters other countries use (I know of one place that uses two letters and a single-digit [only?] integer).

And for a fresher warning example, now that Asimov's been dead for several decades -- does none of the arseholes in a position to do this shit fucking read Stross???

(Also: Shoggoth, no a; Aldi, no s.)
--

   Christian R. Conrad
The Man Who Apparently Still Knows Fucking Everything


Mail: Same username as at the top left of this post, at iki.fi
New Plenty of reasons to worry. Take your pick.
Regards,
-scott
Welcome to Rivendell, Mr. Anderson.
New For values of "a few" greater than eight. Not far from a fricking century.
New I seem to recall a month or so ago you guys were brushing this aside
Basically telling me I was panicking in my post pointing out the end of us via AI.

But now you seem to be taking it seriously.
New It was different then.
;-)

Just wait until the civil lawsuits start flying...

Seriously, there are always over-the-top stories about how some new thing is going to kill us all in our beds. And 30 years later, it still hasn't. The advertizer-driven press depends on engagement and eyeballs and clicks. They're not going to tell us that everything is fine, go outside and play with your dog.

We have to keep our wits about us, and recognize the dangers, but not get our lizard brains activated by people who are trying to monetize every aspect of our lives.

Cheers,
Scott.
New Check out the emergent power seeking behavior
https://youtu.be/2AdkSYWB6LY

Go to 9:42 if you don't want watch the whole thing.
New Nature.
From March 16: GPT-4 is here: what scientists think - Researchers are excited about the AI — but many are frustrated that its underlying engineering is cloaked in secrecy.

[...]

Fake facts

Outputting false information is another problem. Luccioni says that models like GPT-4, which exist to predict the next word in a sentence, can’t be cured of coming up with fake facts — known as hallucinating. “You can’t rely on these kinds of models because there’s so much hallucination,” she says. And this remains a concern in the latest version, she says, although OpenAI says that it has improved safety in GPT-4.

Without access to the data used for training, OpenAI’s assurances about safety fall short for Luccioni. “You don’t know what the data is. So you can’t improve it. I mean, it’s just completely impossible to do science with a model like this,” she says.

The mystery about how GPT-4 was trained is also a concern for van Dis’s colleague at Amsterdam, psychologist Claudi Bockting. “It’s very hard as a human being to be accountable for something that you cannot oversee,” she says. “One of the concerns is they could be far more biased than for instance, the bias that human beings have by themselves.” Without being able to access the code behind GPT-4 it is impossible to see where the bias might have originated, or to remedy it, Luccioni explains.

Ethics discussions

Bockting and van Dis are also concerned that increasingly these AI systems are owned by big tech companies. They want to make sure the technology is properly tested and verified by scientists. “This is also an opportunity because collaboration with big tech can of course, speed up processes,” she adds.

Van Dis, Bockting and colleagues argued earlier this year for an urgent need to develop a set of ‘living’ guidelines to govern how AI and tools such as GPT-4 are used and developed. They are concerned that any legislation around AI technologies will struggle to keep up with the pace of development. Bockting and van Dis have convened an invitational summit at the University of Amsterdam on 11 April to discuss these concerns, with representatives from organizations including UNESCO’s science-ethics committee, Organisation for Economic Co-operation and Development and the World Economic Forum.

Despite the concern, GPT-4 and its future iterations will shake up science, says White. “I think it's actually going to be a huge infrastructure change in science, almost like the internet was a big change,” he says. It won’t replace scientists, he adds, but could help with some tasks. “I think we're going to start realizing we can connect papers, data programmes, libraries that we use and computational work or even robotic experiments.”


Cheers,
Scott.
     Lying to humans to manipulate them to accept a paid task - (crazy) - (13)
         Yes, that's exactly the sort of thing we don't want it to do - (malraux) - (2)
             Do you want meth? Because that's how you get meth. -NT - (drook) - (1)
                 Had in mind something more like this: - (malraux)
         Don't do that - (drook) - (5)
             Yup. - (Another Scott) - (4)
                 Even worse: - (malraux) - (2)
                     Why just private nutjob groups? - (CRConrad) - (1)
                         Plenty of reasons to worry. Take your pick. -NT - (malraux)
                 For values of "a few" greater than eight. Not far from a fricking century. -NT - (CRConrad)
         I seem to recall a month or so ago you guys were brushing this aside - (crazy) - (2)
             It was different then. - (Another Scott) - (1)
                 Check out the emergent power seeking behavior - (crazy)
         Nature. - (Another Scott)

Did you hear the one about the priest, the rabbi, and the gloat?
108 ms