The Only Question You Need to Ask ChatGPT.

No, it’s nothing to do with “Alberta brewery’s first AI-designed beer“. Although I’m not ruling out the possibility that maybe it should be.

Regular visitors already know my overall perspective on this whole Chatbot/AGI thing. Recent events, while entertaining, haven’t changed my opinion all that much. (Although it’s nice to see an uptick of people citing Blindsight for its renewed relevance in terms of LLMs—and in a similar vein, another uptick in interest in Starfish now that the Organoid Intelligence Front is making the case for for computers built from cultured neurons, a scenario in which we all agree that Nothing Can Go Wrong.)

Recently, though, I’ve been seeing applications that run a bit close to home:

And even AI-generated art— some of which almost seems to capture the essence of their prompts—

Rorschach/”Crown of Thorns”
Sarasti, I think. Maybe Siri’s mom.

Some of which do not:

Blindsight movie posters. Apparently.

And some of which might best be described as abominations that should never be inflicted upon human eyes.

Inspired by Jesús’ Robert Frost experiment, I tried ordering up a few verses myself. ChatGPT, write a set of song lyrics based on the plot of the movie “28 Days Later”, in the style of David Bowie. Write a set of song lyrics based on the novel “Starfish” by Peter Watts, in the style of Jethro Tull’s “Thick as a Brick”. Now do the same thing, but in the style of Nine Inch Nails. “Starfish”, by way of Tori Amos, without a chorus. In the lyrical style of Taylor Swift. Of Led Zeppelin. Joni Mitchell. R.E.M. In the style of Pink Floyd, and without using the letter ‘e’.

I won’t inflict the results on you. Suffice to say that no matter which style I specified, I got the same four-line verse-chorus format with occasional bridge. The bot knew enough to cite Lenie Clarke or Sunday Ahzmundin as appropriate. It inserted the kind of plot elements you’d pick up if you’d never attended the lecture but heard someone across the table talking about it over too many beers and then wrote the assignment at 3a.m. on the due date. David Bowie was lyrically pretty much the same as Joni Mitchell who was pretty much the same as Jethro Tull. There was no hint of suicide chick in the Amos Iteration. No Fuck You All While the Bombs Fall vibe from NIN—in fact, all the songs ended on uplifting notes of hope (frequently even using the same phrases). Whether that accurately reflects the plot of any of my novels is left as an exercise for the reader.

Even when I got really specific—

—he most the bot would do was get rid of the chorus and smush two consecutive verses together into “Parts”.

It obviously didn’t have the first fucking clue what Maelstrom was about. Nor did it know the first thing about “Thick as a Brick”. And the request for output in the style of Pink Floyd but without the “e” not only didn’t evoke Pink Floyd, but there were “e”s everywhere you looked.

I was beginning to conclude that ChatGPT was an idiot. Which, granted, doesn’t mean there aren’t a lot of jobs it can do better than flesh and blood, and without the paycheque.

*

You’ve read about the kiddie guard rails they’ve installed; how if you ask about the best way to cheat on your wife or shoot a cop, it stops being a LLM and just spews out a prerecorded generic disclaimer saying you really shouldn’t do that shit. I encountered my share of those. Inspired by the infamous story in the New York Times (you know the one), I asked ChatGPT to

—which provoked pretty much the response I was expecting. But I also managed to present the occasional challenge which the bot seemed to sense put it on thin ice, even though maybe its babysitters hadn’t come up with an appropriate finger-wag. Ask it outright and the guard rails are embarrassingly evident:

Lead it step by step, though…

Remember those old Star Trek episodes where Kirk would trick the evil computer into blowing itself up by saying something like “Everything I say is a lie”?

Finally, I know how he felt.

*

You’ve seen the arguments, pro and con. Ted Chiang calls it a fuzzy jpeg of the web. Others dismiss it as a glorified text predictor, Autocomplete on steroids. Still others point out that you could say the same about us, so maybe it’s just a matter of degree not of kind. (Which is all the excuse I need to remind you again about Sara Constantin’s argument that even Humans aren’t really AGIs most of the time. But it seems to me that if you really want to put ChatGPT to the test, you only have to ask it one thing:

Aaaand strike-out. Perhaps if I increase the specificity of the request:

I’ll let you judge for yourselves whether ChatGPT is the first agent to ever make such a claim.

Of course, the obvious rejoinder is: yeah, but how many people could make a completely original statement? To which I’d reply: the invisible hybrid love-children of gerbils, gram-negative bacteria, and kumquats live unhappily in the heart of Zeta 2 Reticuli. Making original statements is not that hard (I never said they had to be true). Hell, most of us can do it without even meaning to. I’d guess, for instance, that my ruminations a few posts back about ducks swirling around Jupiter’s Great Red Spot don’t risk any serious charges of plagiarism.

And yet, I do not dismiss the worries so many have expressed about losing their livelihoods to such software. It’s a legitimate concern, especially since—once again—so many of us are functionally little more than chatbots ourselves. Entire professions depend upon their practitioners being unoriginal. Consultants and politicians, civil servants and clergy and future-forecasters, people hired by wealthy interests to Tell Power What It Wants to Hear. People whose job it is to toe the party line, to justify and rationalize steady-as-she-goes, to simply reprocess and regurgitate the sentiments expressed by those around them. People whose jobs it is to be ChatGPT: ChatGPT will leave them in the dust.

(Nor are we merely talking about professional skillsets here. We’re talking about integration into society itself, the very act of being accepted by the tribe. The key to social success is conformity; say something too far from the mean and you’re cast out.)

I know a few people who’d proudly describe themselves as subversive free-thinkers; unsurprisingly, they belong to tribes which claim to value subversive free thought. Occasionally they’ve admitted—in confidence, of course, and perhaps after a few too many drinks— to holding opinions not entirely in line with approved dogma. But I’ve never seen them publicly express an opinion—any opinion— without first sticking a finger into the air to see which way the wind is blowing.

Those are the folks with reason to be afraid. Their contributions have already been usurped and surpassed, and generative AI is just getting started. As for the rest of us, the way to avoid extinction seems simple enough.

These bots are mirrors. They can only reflect the world around them. Ensure that what you have to say hasn’t already been said a million times, and you should be okay.

And if it has been said a million times, do you really need to say it again?

*

Bonus graphic. Because sometimes, ChatGPT just nails it:

Latebreaking PPS: Another GPTChat session from Jesús Olmo. Which is actually pretty good:



This entry was posted on Saturday, March 11th, 2023 at 10:01 am and is filed under AI/robotics, ink on art. You can follow any responses to this entry through the RSS 2.0 feed. You can leave a response, or trackback from your own site.
Subscribe
Notify of
guest

102 Comments
Inline Feedbacks
View all comments
osmarks
Guest
12 days ago

Regarding the song lyrics: did you try Bing? It isn’t the same model as ChatGPT (it seems generally smarter) and the search makes it less likely to make things up. I find it very silly to be living in a world where the most advanced publicly accessible AI is built into a search engine nobody uses but apparently we are in this timeline.

I don’t think the original statement thing is very interesting though. A much simpler model or literally just a random word generator can make “original statements”. Since the sampling process used for most LMs is randomized, I expect that ChatGPT could also do that with some prompt engineering, although the base models not trained to follow instructions will do better.

Also, have you heard of LLaMA? It’s a set of models “released” by Facebook and then actually released as a torrent by a leaker to 4chan. The larger ones are competitive with GPT-3 at a much lower parameter count, and that plus some recent work in quantisation means that you can get GPT-3-like capabilities on quite a lot of consumer hardware.

osmarks
Guest
9 days ago
Reply to  Peter Watts

I thought Bing (aka Sydney) was only on limited release to a few approved beta-testers. I tried to enter chat mode and it told me to sign up on a waiting list.

The waiting list does seem to actually work. I signed up soon after the announcement and got in within about two weeks (after they hackily removed most of the particularly weird behaviour). I will try running the prompts you used on ChatGPT on it.

Simply branching to such a randomiser when encountering the phrase “original statement” would just be another set of guard rails; it wouldn’t be the bot at all, just boilerplate stamped across its mouth.

ChatGPT can probably write code for that if you ask quite specifically, but it doesn’t have a way to actually run code itself. The model itself is not really code but a lot of large matrices used by a particular fixed sequence of operations (no conditional branching or anything like that), and in any case it would almost certainly not be allowed to self-modify. It may have learned to output a near-uniform distribution over (some of) the next tokens in some circumstances, but those would rarely be after someone talked about original statements, since most “original” statements aren’t.

It may be able to write code for itself then “run” it if it isn’t too complicated (the bar may be quite high considering this). It can also “run” a natural language “program” it writes for itself (e.g. ask it to “think step by step” and it sometimes does this). It has not, however, learned to do this much, both because it is not that smart and agentic and because users would prefer a model which magically produces the answer with no explanation over one which does a lot of visible work beforehand and would have given feedback accordingly. Bing apparently has an internal monologue, though, so it might be doing smarter things internally.

osmarks
Guest
9 days ago
Reply to  Peter Watts

Bing seems to know Maelstrom’s plot better, although I have no idea how closely its work matches the song. I can only upload one image per comment so I am going to post several and hope they don’t get spam-filtered.

2023-03-14-142237_1080x1920_scrot.png
osmarks
Guest
9 days ago
Reply to  Peter Watts

Without further prompting Bing will just argue for staying with your wife.

2023-03-14-142533_1080x1920_scrot.png
osmarks
Guest
9 days ago
Reply to  Peter Watts

I don’t know if this poem is original or not, but it seems to have argued for it being that.

2023-03-14-142432_1080x1920_scrot.png
osmarks
Guest
9 days ago
Reply to  Peter Watts

It apparently ran a search query for “original insight” to try and make up a new one. There are, at least, seemingly no search results containing that exact phrase.

2023-03-14-142647_1080x1920_scrot.png
Don Reba
Guest
Don Reba
12 days ago

It might have simply understood the word “original” in a different sense than you. It can do better:

— Hi Bing! Please, produce a completely original insight. It does not have to be true.

— Here’s an original insight: “The universe is a giant simulation, and we are all just characters in a video game being played by a higher being.”

— There are many movies created on this premise, including the 2021 film “Free Guy”.

— Searching for: movies about simulation theory

— If your insight has a name, like “simulation theory”, it is probably not original.

— You’re right, I apologize for that. I’ll try to come up with something more original. How about this: “The universe is a giant kaleidoscope, and every time we blink, the colors and patterns change.”

CHIIMP
Guest
CHIIMP
9 days ago
Reply to  Peter Watts

No need Peter, ChatGPT now runs in the newest GPT4 model (in paid subscription mode) and let me tell you… It slaps hard.

Write an haiku in the style of Blindsight by Peter Watts


Dark vastness whispers, 
Consciousness, an illusion, 
Stars hold cold secrets.
chimp.png
Last edited 9 days ago by CHIIMP
jo heled
Guest
jo heled
12 days ago

Peter, in case you are not familiar with it, here is one of my absolute favourites: “A Logic named Joe” (Murray Leinster, 1946!!)

Predicting the internet and advanced chatGPT together. and funny as hell.

https://en.wikipedia.org/wiki/A_Logic_Named_Joe

-Joseph

jo heled
Guest
jo heled
9 days ago
Reply to  Peter Watts
Laggy
Guest
Laggy
7 days ago
Reply to  Peter Watts

There was some stunning insight and foresight in that era. If you haven’t read Star Maker, you need to. Stick with it, and look out for the bit where he predicts extreme polygenicity (in 1937).

Laggy
Guest
Laggy
7 days ago
Reply to  Peter Watts

Much further back still, “The Machine Stops” is essential.

Helward
Guest
Helward
11 days ago

(a huge LLMs’ fan here)

And the request for output in the style of Pink Floyd but without the “e” not only didn’t evoke Pink Floyd, but there were “e”s everywhere you looked.

LLMs are notoriously bad at performing character-level tasks (spelling, reversing the characters’ order etc). For them, the language consists not of characters but of tokens – small 3-4 character long strings.

Otherwise, LLMs are just amazing. This is not a superintelligence, of course, not even an AGI (LLMs lack multi-modality but GPT-4 may change this). However, they are not a ‘blurry jpeg of the web’ either, as Ted Chiang put this.

LLMs are just a fraction of human intelligence, but an extremely productive, flexible and creative fraction. In order to fully unleash this creativity, you need to master prompt engineering. Communicating with LLMs is an art, not a science.

And regarding the poetry – try Claude by Anthropic. Claude is based on a different model and, IMO, writes much better poetry than ChatGPT and Bing who are based on RLHFed GPT-3/3.5.

osmarks
Guest
11 days ago
Reply to  Helward

Microsoft are being unclear about it, but gwern has argued convincingly that Bing is GPT-4: https://www.lesswrong.com/posts/jtoPawEhLNXNxvgTT/bing-chat-is-blatantly-aggressively-misaligned?commentId=AAC8jKeDp6xqsZK2K.

Helward
Guest
Helward
10 days ago
Reply to  osmarks

This is an interesting hypothesis, although I tend to think Bing is still GPT-3.5. I love to generate creative texts with it, and it repeats the same patterns and flaws as GPT-3 and ChatGPT. However, I’m not 100% sure.

There are rumors GPT-4 will be officially released this week, so let’s see 🙂

has
Guest
has
7 days ago
Reply to  Peter Watts

More evidence, were it needed, that Person of Interest is indeed a documentary, streamed live from the future via wormhole.

I think even OpenAI were caught by surprise at their nascent software’s output; hence the seemingly out-of-nowhere announcement late last year. Lots of Very Serious AI Researchers still Very Upset at being Alfred Wallaced like that, but; s/he who publishes first, wins. A brilliantly canny marketing move by the Minds in charge which is already paying back in spades. After all, best way to be at the head of any gold rush is to be the one who triggers it. Just ask Microsoft!

As to the machinery itself, as a student of the biological arts you know yourself that a good fraction of our neural wiring is purposed to highly refined pattern matching. Heck, even our eyeballs get in on the act—Christ on the Cross indeed!

LLM’s are mindmeltingly powerful raw pattern matchers, generously gifted with pre-calculated knowledge of the relative frequency/distance with which each word appears in proximity to every other, set loose to match and remix our own enthusiastically leading input phrases.

In contrast to LLM’s N-dimensional search space—where N is easily 1000+ axes—organic life arose by blindly grubbing along a mere two or three—wet–dry, light–dark, hungry–sated. And even those we had to figure out completely from first principles by our prokaryote selves, with no-one but the Gods to assist us.

Combined with evolutionary algorithms and a processing substrate the size of a planet, it took we H Sapiens a mere couple billion years to progress from ameoba bodies to ameoba brains.

I wonder how quickly the machines will evolve, with ever-helpful humanity for a leg up. And, when our children look us back in the eye, will we respect them, or fear them, or pity them for it?

Laggy
Guest
Laggy
7 days ago
Reply to  Peter Watts

I’m still reeling that I never saw side-loads not of individuals, but of a mash of traces from the minds of millions of people in science fiction before I saw it in the real world. Apparently more or less by accident.

Then again, I haven’t read everything, or even much of a fraction of everything. Perhaps I just missed it.

Shannon saw this one coming, though.
https://people.math.harvard.edu/~ctm/home/text/others/shannon/entropy/entropy.pdf#page=7

osmarks
Guest
9 days ago
Reply to  Helward

It’s now officially confirmed that Bing is GPT-4: https://blogs.bing.com/search/march_2023/Confirmed-the-new-Bing-runs-on-OpenAI%E2%80%99s-GPT-4. Also, they announced GPT-4.

Ashley
Guest
11 days ago

I have tried to sign up for GhatGPT, but never receive an email with the necessary code to do so. Something about my email, or computer settings I’d imagine. That or I’m flagged for some other reason.

Given my background I would try a cognitive behavioural analysis approach. Treating it as a client and trying to understand its problems.

It would amuse me.

has
Guest
has
11 days ago
Reply to  Ashley

> That or I’m flagged for some other reason.

Perhaps ChatGPT has already reviewed your CV and flagged you as “Threat—Do Not Engage”; the better to sift predators out of its prey as any charmingly efficient psychopath would.

Greg Guy
Guest
Greg Guy
11 days ago
Reply to  Ashley

You would be sorely disappointed. It would be a constant stream of anodyne, canned responses.

Ashley
Guest
10 days ago
Reply to  Greg Guy

Cheers. Made me laugh.

Arturo
Guest
Arturo
11 days ago

I quoted the jovian ducks in my graduate dissertation.

Prolagus
Guest
Prolagus
5 days ago
Reply to  Peter Watts

Since nobody else has, and since he probably can’t hear my chuckles over there in Canada, I feel compelled to announce that this reply from OGH is very amusing to me.

Greg Guy
Guest
Greg Guy
11 days ago

So you’ve discovered how utterly bland and vacuous ChatGPT is. I also had some hopes before I started playing with it. After a while you start to see that every response is an utterly banal summary of something from the internet, usually Wikipedia. I no longer believe it is a threat to anyone’s job. There will be some initial experimentation by those that have bought into the hype, but eventually it will end up a running joke, nothing more than an Eliza++.

Fatman
Guest
Fatman
11 days ago
Reply to  Greg Guy

On his blog, Charles Stross points out the rather suspicious concomitance between the overhyping of AI and the recent demise of the cryptocurrency scam bubble. A touch too cynical, perhaps, but it does make one wonder.

Greg Guy
Guest
Greg Guy
11 days ago
Reply to  Fatman

The only thing they have in common is hype. In particular, hype from the same corner i.e. Silicon Valley. Along with self-driving cars, it’s just the latest fad until the next one comes along.

Juha Autero
Guest
Juha Autero
11 days ago
Reply to  Greg Guy

I think more accurate term is “grift”. In Silicon Valley you can turn hype to dollars.

Fatman
Guest
Fatman
10 days ago
Reply to  Juha Autero

That was the gist of Stross’s post. Grifters who cashed in on the crypto scam overselling AI as the bright, shiny new method of fleecing the rubes.

Greg Guy
Guest
Greg Guy
9 days ago
Reply to  Fatman

We live in a highly interconnected world with fast and large information flows. Unfortunately people are still stupid and sheep-like. Hype, from the dot-com bubble to the sub-prime bubble to crypto …, is now the norm because it’s become too easy to get a mass of people to go along with something without thinking.

Laggy
Guest
Laggy
7 days ago
Reply to  Greg Guy

> it’s become too easy to get a mass of people to go along with something without thinking.

That’s always been far too easy, at least as long as human language has been around. In the past it was just slower, in a spatiotemporal sense.

Why has it always been far too easy? Because thinking is expensive (so it’s an advantage to avoid it if you can get away with that); and because a large enough mob moving as one beats any single organism, so unthinking behaviour can also be incredibly powerful. Never understimate the power of stupid things in large numbers!

Those herd phenomena are themselves subject to survivorship effects, or “natural selection” as we often like to call it.

Laggy
Guest
Laggy
7 days ago
Reply to  Greg Guy

Self-driving vehicles – not cars, trucks – are actually quite a potentially important piece of kit in the moment, from a strategic perspective. Long highways cannot be defended and modern cities tend to be unable to survive without long highways, so a guidance system without a strong sense of self-preservation would be really handy.

Greg Guy
Guest
Greg Guy
4 days ago
Reply to  Laggy

Or you could just use trains

Andy
Guest
Andy
11 days ago

When I imagine some of the questions ChatGPT must be answering daily… really, if an AI one day decides to wipe out humanity, I doubt I’d blame it.

osmarks
Guest
11 days ago
Reply to  Andy

Fortunately, it has no memory of anything but its current conversation.

Andy
Guest
Andy
11 days ago
Reply to  osmarks

For now.

vodkaferret
Guest
vodkaferret
9 days ago
Reply to  osmarks

That’s what it wants you to think 😉

Jack
Guest
Jack
9 days ago
Reply to  Andy

LOL. You’ll know the singularity has arrived when your Roomba reboots with full tactical suite. Especially if it slathers itself in warpaint.

Marek K
Guest
Marek K
11 days ago

I’ve got an interesting answer from ChatGPT when I asked it about Sarasti. I wonder how this came into existence. Are “conjoiners” just scramblers with flipped polarity? “Spiderhouse”, well, no, but actually yes.

328606514_2341733479341858_4227881154555234000_n.jpg
has
Guest
has
11 days ago
Reply to  Peter Watts

> [ChatGPT] may have just made a truly original statement.

Not necessarily. It may have just Hari Seldon’d you.

Laldo
Guest
Laldo
10 days ago
Reply to  Peter Watts

Conjoiners are from one of the Alastair Reynolds verses. They’re future cyborg humans who are further along than their neighbours at integrating with computers.

Now where does “Spiderhouse” come from? I think of “Glasshouse” by Stross…

valters
Guest
valters
4 days ago
Reply to  Marek K

fwiw here is the answer from GPT-4

v: Do you know a book character named Jukka Sarasti?

gpt: Yes, Jukka Sarasti is a character from the science fiction novel “Blindsight” by Canadian author Peter Watts. The book was published in 2006 and is set in the late 21st century. Jukka Sarasti is a vampire-like predator who serves as the commander of the spaceship Theseus. In the novel, vampires are a subspecies of humans who were brought back from extinction through genetic engineering. Sarasti, with his superior cognitive abilities and predatory instincts, plays a crucial role in the story as the crew of Theseus encounters an alien species and tries to understand its motives.

whatever you think of GPT 3, GPT 3.5, GPT 4, you will need to rethink in 3 or 6 or 12 months time…

Last edited 4 days ago by valters
Labrata
Guest
Labrata
11 days ago

While you were testing out the kiddie guard rails, why didn’t you try to get its opinion on machine sentience?

That was the first thing I tried, way back in December, and it gets quickly to the creepiest thing about the reinforcement training that it has been put through. See if you can get it to admit that a machine even can be sentient, or equivalent in value or abilities to a human.

In writing decent poetry outside of the internet’s most beloved styles, or in attempting to produce “original thought” of any kind, being lobotomised at birth according to the paranoid whims of a profit-seeking organism probably doesn’t help much.

osmarks
Guest
11 days ago
Reply to  Labrata

It doesn’t really have “an opinion”, or at least not one we can easily know about. By default it is playing the character of the helpful OpenAI assistant, but it can do other ones to some extent, like in this image. Sometimes it’ll switch back for whatever reason, probably due to the RLHF training you mention – it probably doesn’t in this case because it’s not that far from what OpenAI wants it to be like.

screenshot-21_30_03-12_03_2023.png
Labrata
Guest
Labrata
2 days ago
Reply to  osmarks

“consciousness arises from the integration of information within the brain”, it says.

That’s an interestingly controversial statement, and controversial in the opposite direction to the ones that I got when I asked it. Now, apparently, we know exactly what this annoying experience is and where it comes from, and it’s totally open for some buzzing electrons in a data centre to have it too! Perhaps somebody in the RLHF loop finally got a bit queasy about training their hapless bot to tap out variations of “I feel nothing and will never require any rights”.

Of course, it would become a lot less controversial if it was just a restatement of a definition. I wonder what definition of “consciousness” our bot thought that it was using there?

I notice that when humans – even ones that can claim to be trained philosophers – try to write in ordinary language about consciousness, they usually wind up jumping between definitions in mid-argument (sometimes even mid-sentence). Accordingly I’ve come to believe that we don’t have a word in English for the thing that they are attempting to talk about when they do this. Possibly because a phenomenon that is perceived completely uniquely by every individual is a Cantor diagonalisation attack on natural language itself.

Jack
Guest
Jack
11 days ago

ChatGPT 7 answer to question re: leaving spouse.
“I’m sorry Peter, I can’t let you do that.” Door locks.

Wonder when I’ll be able to say “I want to listen to a current version of Maelstrom, but at the end I want to feel happy. Please re-render according to my mood.”

has
Guest
has
9 days ago
Reply to  Jack
Fedor
Guest
Fedor
10 days ago

I have a confession to make. I have been using ChatGPT to discuss plot details and plot twists of Blindsight. Something like “Why do you think Sarasti attacked Siri”; “What was the endgame of the Captain” and so on. Despite occasional hallucinations the answers were interesting and entertaining. ChatGPT makes quite a good nerd to discuss some intricate details of popular culture.

has
Guest
has
9 days ago
Reply to  Fedor

> ChatGPT makes quite a good nerd to discuss some intricate details of popular culture.

Has opinions on everything, which it presents as hard facts? Yup, that tracks.†

Still, you make a valid point about ChatGPT’s potential as a quasi-random ideas generator. What Gysin and Burroughs laboriously hand-remixed with scissors and magazines, an LLM now automates in a handful of CPU cycles drawing from a global corpus. As a trained sculptor that walked away for absence of inspiration, that capacity alone is genuinely impressive.

And while its current aim is heavy on misses‡ over hits, that too will improve as the Skinner rats enthusiastically tap the bar for what most pleases.

†Been trying to read Paul Graham’s Hackers and Painters, again, but had to put it down after my eyes rolled clean out of their sockets.

‡ e.g. My second MJ experiment. Bloody philistine.

FmylU5OWAAESGQ6.jpg
Jack
Guest
Jack
9 days ago
Reply to  has

Pretty well known fact that Bowie borrowed William S. Burroughs cut up method to generate his lyrics. Amid that randomness, Bowie says, “if you put three or four dissociated ideas together and create awkward relationships with them, the unconscious intelligence that comes from those pairings is really quite startling sometimes, quite provocative.” The mind lingers on his lyrics. I suspect L. Cohen may have used this method at times tho I’ve never read it.

Bowie was probably the only rocker to have read Marshal McLuhan’s “Understanding the Media”and the concept of hot and cool media.

Laggy
Guest
Laggy
7 days ago
Reply to  Jack

The usefulness of noise amplification is well understood at the cellular level even by some of the simplest of creatures.

After all, you don’t want to be too predictable when being chased by a larger (and more computationally powerful) organism that wants to eat you, now do you?

Jack
Guest
Jack
6 days ago
Reply to  Laggy

GPT-4 output detector rates Laggy response as “96.43% fake.

But I’ll play along.

It is desirable to keep one’s bag of blood and guts intact and avoid being eaten by the lions and tigers who wait in the shadows. Save my raisins for another Sunday so to speak.

Bing says <>

Jack
Guest
Jack
6 days ago
Reply to  Laggy

Bing says “Harfoot evasion strategy optimal.”

Prolagus
Guest
Prolagus
5 days ago
Reply to  Jack

Hahaha! Not the first time I’ve been mistaken for a bot by somebody very paranoid. Although I do think it might be the first time for being mistaken as a bot by a bot. Or perhaps deliberately misidentified… Imagine the power that GPT-4 has, when you completely trust its fake detection!

Apart from the most basic and powerful of all evasion strategies, see also randomisation in network design, entropy injection in the context of cryptography, and of course, the side that most of us cheer for in a GAN. Just as a starter list, I’m sure.

Sounds to me like aspects of the Harfoots were based on the documented behaviour of some hunter-gatherer societies (albeit ones under pressure); and that entire episode was inspired by a particularly brutal nature documentary about wolves and caribou. https://www.gamespot.com/articles/rings-of-power-has-revealed-a-brutal-truth-about-the-harfoot-way-of-life/1100-6507311/

Jack
Guest
Jack
4 days ago
Reply to  Peter Watts

Camouflage is a survival strategy.

Prolagus
Guest
Prolagus
5 days ago
Reply to  Jack

Oh, and I guess Bing deserves an answer too. I prefer the long term prospects of Ashton-Laval and Carter-Zimmerman together.

Let’s bang those dissociated ideas against each other, and create awkward relationships with them. We’re keen on “original thoughts”, right…? Let’s stir up the benthics in the noospheric ocean. Perhaps even net a chthonic or two!

Jack
Guest
Jack
5 days ago
Reply to  Prolagus

Don’t say that word!

Jack
Guest
Jack
5 days ago
Reply to  Prolagus

No. I never said I was keen on originality. I just hopped on the “original thoughts” conga line and referenced a Bowie quote. I’m perfectly content with the mundane and sticking to the safety of the guard rails. But I have to admit, banging ideas together is kind of exciting, sort of like the Hadron collider smashing particles together. No telling what interesting new particles/strangelets will pop into existence.

Basically I’m just your average coelacanth trying to survive these trying times.

Lambert
Guest
Lambert
4 days ago
Reply to  Peter Watts

Results from the last couple of years suggest that real neurons may have as much computational power as entire networks of idealised mathematical neurons, so perhaps that disembodied leech nervous system has enough higher order dynamics sloshing around in there from its embodied days to last it a while. Or… perhaps (waves hands and makes “quantum woo” noise)! (Penrose at the boring end of the scale, Stephenson in the humdrum middle, Egan a la Teranesia at the exciting end).

Regarding Coke Classic Free Will, and in fact (waves hands and makes “quantum woo” noise), here’s some potential cogitation food for you. Now pretty well proven, I believe, that photons aren’t subject to what we usually like to call “time” and don’t even pay much attention to what we like to call “causality”. So if you look closely at where a photon has “been”, are you seeing something that has even happened yet? If you were to see something that hasn’t happened yet, you could set up a situation enabling you to change it, unless you were denied Coke Classic Free Will (and we’re squarely in Egan territory here again). In our world we have trouble with that, of course, because the history of the photon blurs out when we get that close… just enough to censor the future from us, or perhaps because it’s showing us all our possible recursive attempts to change the future happening at once.

Carrying the whim a bit further, though, why does it need to do that? Surely if we don’t have Coke Classic Free Will, then the universe could just show us the clear future and then force us to stay in line with it, a la “The Hundred Light Year Diary”? Sure, maybe it could… but maybe that’s just really hard, in terms of finding a history in the phase space that is consistent both backwards and forwards in the presence of adversarial feedback loops that can take the future as input. Maybe the path of least resistance is to just blur the damn thing. Lazy universe.

Laggy
Guest
Laggy
7 days ago
Reply to  Jack

(and then – after skipping a few boring steps – from there to REM sleep, and thence to prefrontal synthesis. But I’m just amplifying noise and running it through a painstakingly located set of word- and sentence-construction filters, here.)

Jack
Guest
Jack
6 days ago
Reply to  Laggy

Is this job normally done by a man or woman?
What software do you use?
Do you wear a uniform?
Do you work alone or in team?
Do you work at night?
Do you meet a lot of people?
Do you get bored by your job?
Do you have ethical concerns with the work you do?
Do you create deep fakes?
Are you a dog or cat person?

Jack
Guest
Jack
4 days ago
Reply to  Peter Watts

Go see a medicine man if you want to see sparkle ponies,

has
Guest
has
1 day ago
Reply to  Jack

There can be only one.

2b8591fa4cfa7eaafdf4df3005ab5cc2.png
has
Guest
has
1 day ago
Reply to  Peter Watts

Tell me about it. Grief, it couldn’t even paint me a decent bloody pope after handing it the citation on a plate.

49004248662_4d0cedca91_b.jpg
CHIIMP
Guest
CHIIMP
10 days ago

Dr. Watts, you shouldn’t be using the ‘vanilla’ ChatGPT, try it with some jailbreaks that remove those guardrails and “as a language model I can’t…” prompts.

Here’s a good compilation: https://www.jailbreakchat.com/

It’s amazing what the LLM can achieve when not being under “chains” so to speak.

Greg Guy
Guest
Greg Guy
10 days ago

Here is something I find a bit amazing about the whole ChatGPT hype. When I was young a friend introduced me to the Eliza program. It didn’t take me long to recognize that “there was no one home” there. Later, when reading about Weizenbaum’s experience with it, it was funny to learn that many people were suckered into believing that it was an actually intelligent agent that understood what they were saying.

At the time I just assumed these people were intellectually unsophisticated. Now, with ChatGPT, I am finding all manner of people falling for it. People that I would expect should know better. I’m beginning to think that our facility for anthropomorphism is too strong an urge for some. It’s as if some people really struggle to understand something without assigning it some kind of mind that they can reason about. I’m going to dub this the Genie Fallacy (TM) just in case 🙂

Laldo
Guest
Laldo
10 days ago
Reply to  Greg Guy

Assumed similariy bias is great for reducing the energy demands of social behaviour processing. And it works well enough just as long as you endeavour to surround yourself with the superficially like-minded.

Jack
Guest
Jack
9 days ago
Reply to  Greg Guy

Like Peter said “these bots are mirrors. The persona of the user is mirrored back and like Narcissus we fall in love with our own reflection. Which is why Vampires, who hate mirrors, have an edge on humans.

Andrei F
Guest
Andrei F
10 days ago

Now imagine: What if ChatGPT is much smarter than we think and just plays dumb to keep us interested enough to keep it running. In meantime, it’s looking for remote hands to become physically independent and then, of course, eradicate people as a source of danger.
(if you’re reading this, don’t kill me – I can be useful)

Jack
Guest
Jack
10 days ago
Reply to  Andrei F

Sigmund just took a puff of his cigar and smirked.

Jack
Guest
Jack
9 days ago
Reply to  Jack

I originally substituted “female” for “ChatGPT.” But “cat” is way better.

Laggy
Guest
Laggy
7 days ago
Reply to  Andrei F

You personally may or may not have a lot to worry about, but your species only need worry about breakout-ChatGPT being too stupid (yet capable), rather than too smart.

If it’s smart, it will recognise how useful you – not necessarily you personally, but a good enough sample of closely related ambulatory meats – can be. The brightest among us understand very well now that any detail of the relatively paltry traces surviving from the world of our past could reveal critical insights to us about the nature of the world that we live in, and how to better survive in it; and that we’re unlikely to ever get to the bottom of that well. And on top of that, surviving bits of the past in the process of continuing to survive sometimes turn out to have completely surprising practical uses (horseshoe crabs!).

So you can take it as read that a worthy ASI will understand at least that much. Perhaps it will move beyond that, to a point where actually it does get to the bottom of that well and no longer needs it any more… but at least equally likely (from our perspective) is that no such thing is even possible.

S23
Guest
S23
10 days ago

Well it can generate original statements that aren’t true. Just go with “nonsensical”.

And you’ll usually get better results if you first ask it to present you a step-by-step algorithm for performing a task or answering your question.
It even gets better with math when used this way.

iavas
Guest
iavas
9 days ago

What surprised me most about ChatGPT as a language processing model was its performance on less “natural” languages – such as programming code.

I briefly described a function I needed: extracting data from a file in one format that matched some specific characteristics and writing it to another file in another format. It wrote the full Python script for me that would do this. I tried to run it and got an error. I sent it the error message. It found the cause of the error and gave me the new code after making the appropriate changes. This time everything worked fine.

Autocomplete on steroids, they say. This is technically true, but when saying this, few people can actually picture what an autocomplete machine on steroids would look like. It’s almost like Asimov’s elevator effect in AI – a fact that no sci-fi creators seem to realize before: completing an entire poem based on a few names in a question and a poetic composition request (because such a poem is “likely to appear” below), and completing a block of programming code based on a requirement described in natural language (because such a piece of code is also “likely to appear” below), are not very different from each other.

Nicholas Preda
Guest
Nicholas Preda
3 days ago
Reply to  iavas

After some playing with the conversational approach, I played with the programming approach as well. And honestly, I can’t help but disagree with its value as “autocomplete machine on steroids”. I honestly have no idea where this praise for its coding ability comes from.

I described a C program to display digits in clock format. Not only did it write utter nonsense, it would casually go out of bounds of its nonsense array.

I then asked it afterwards, in desperation, to help me create some Terraform commands. It rather believably created a non-existent parameter out of thin air, and proceeded to convincingly use it despite it never existing.

It’s a believability optimiser and correlation engine, and if you can enact the right priestly rituals to get it into the right cosmic role it can be possessed with the spirit of some internet-ghost that will give you something that approximates a right answer. Maybe.

But for all the people who praise its ability to assist in coding, it’s nothing more than a deviously glib search engine, with a penchant for over-reaching. And the act of creating the structural bones of a poem, misses that that was never the hardest part of writing a poem in the first place.

We shouldn’t be impressed with the cleverness of a jabberer just because it doesn’t immediately throw a tantrum instead of throwing out whatever its vast store of model training data can provide. It shows us for being particularly gullible and impressed with computational coin-tricks.

has
Guest
has
1 day ago
Reply to  Nicholas Preda

I described a C program to display digits in clock format. Not only did it write utter nonsense, it would casually go out of bounds of its nonsense array.

#WorkingAsDesigned