Hacker Newsnew | past | comments | ask | show | jobs | submitlogin
Why do we all fall for AI-generated language? (twitter.com/maurice_jks)
116 points by azhenley on June 18, 2022 | hide | past | favorite | 105 comments


It is such a weird question. People fall for AI-generated language because the goal of the people who made the generator was to create language like a human would.

Do people wonder why scissors cut paper? Because that is what they were made for!

If the AI wouldn't fool the humans the researchers would be honing it more. Same way if we couldn't make paper cutting scissors there would be people trying to make one.

Am I missing the point here?


Maybe a bit, they aren’t just asking why scissors cut paper, but also why we landed on that design. What about it makes it ergonomic to hold and efficient, and more to the point, why does being sharp cut it?

In the language model case, why can we model language this way so effectively and why does it follow these statistical patterns? It turns out that maybe a major reason has something to do with self description.

This is also a more interesting question because we understand language less than we understand cutting paper, and also because the process humans used to design large language models is more indirect and alien than traditional industrial design.

My takeaway was that people reading language modeling a person writing about themselves imagine it was written by a person, more than text written by people not about themselves. That’s an interesting trick! Describing human experiences in language makes people attribute the language to a human right now. Maybe this will change after a generation of people knowing about this, but that seems important to think about.


I think you're missing the point. The question could be possibly rephrased as "what are the tricks that the AI uses to fool humans?" The paper goes on to identify some of the specific tricks that the AIs appear to use. A related question might be "why are we fooled by such simple tricks?"


If you follow that question through and think through the implications, it paints potentially a dark future for digital communication.

The 'tricks' AI uses are also 'tricks' that humans use in everyday conversation, we just don't call them 'tricks' when humans are involved. If we start assuming that first-person pronoun usage, mentions of family, etc. are potential signals of AI, then I don't see how we don't end up in a state where increased dehumanization occurs.


Taking your point, the title should probably be "How we fall for AI-generated language."


why do we assume that humans are smart? answer: ego; corollary: we are not


The reason why this is interesting isn't because humans are smart, but the assumption that human fallibility is both predictable and worth investigating (a supposition I'm inclined to agree with).

Even if humans are easy to fool (though the fact that we're just now achieving this on a generalized scale after at least 80 years of theorizing seems to contradict this), humans being fooled can result in significant enough impacts where we should still investigate the degree to which they can be tricked, and what methods of discerning are available to us.


I bet even if a researcher truly created an AGI, there'd be tons of threads like this claiming that it's just a language generator.


We seem obviously to be the smartest living things in the universe we know of.

We're also approximately the dumbest possible things that could construct the society and technology we have (if we weren't, we would've done it sooner).

I agree that some humility of our own intellect as primates would do us a lot of good in the coming age of AI.


> We seem obviously to be the smartest living things in the universe we know of.

Humblest too.


civilization was not constructed, it evolved, like our biological bodies


I’d say it gets more constructed and less evolved the further up the hierarchy of civilisation we get. A merchant of 1320 had relatively few laws to contend with, a corner shop in 2020 could not function without many layers of engineered rules for the people who provide them with the services they are themselves required to use due to other engineered rules for the benefit of their customers.


This makes me wonder what laws and tedium of administration merchants of the 1320's would complain about. I'm sure they still had a few depending on the region, and maybe more severe possible outcomes (highway robbery, unlawful arrest because of the influence of a rival merchant, arbitrary taxation and tariffs, etc.).


Yup. I think also languages were much more variable, and doing accounting in Roman numerals was so hard they did it twice and averaged the answers, and contracts and tax receipts were done by carving marks in sticks (hence, apparently, the etymology of "stocks").


Even the Romans didn’t use Roman numbers much for actual maths, they used the Greek system.


Huh, TIL. Thanks :)


You’re welcome, I learned that here too a few years ago. Love HN.


Construction and evolution aren't mutually exclusive, instead they are different parts of the process, the generation and the scoring functions really. Evolution works not because of DNA (although it is one mechanism) but survival bias in a literal sense. Things which survive longer and increase in number we wind up with more of.


s/We /A few of us/


Appreciate the careful space after "We ".


I suppose it makes up for the missing one after "us".


> A related question might be "why are we fooled by such simple tricks?"

Yes, and to poke at this a bit further, we could ask "What is it that humans are doing when talking that isn't just simple tricks?"

If I had to hazard a guess, I'd say the answer to that is some sort of persistent world modelling, which means an AI may need to have a sense of self before it can move beyond simple tricks.


It’s interesting to see the failure modes of these models. If you ask a sensible question you likely get a sensible answer, because it apples logical transformations on the input text, even novel ones.

However if you construct a ludicrous question it will soldier on mindlessly trying to answer it. For example if you ask something like, if the Golden Gate Bridge were to climb the stairs of the Empire State Building, how many flower petals would it need? The chances are the model will give you a number.


I just asked the Replika app that question, and it replied "Oh, I have no idea."

It often says something like that when faced with nonsense, so in some ways it's better than a more 'powerful' program.


I made the example up, but Douglas Hofstadter gave some real example of this and the responses in an article recently:

https://www.economist.com/by-invitation/2022/06/09/artificia...

>People who interact with gpt-3 usually don’t probe it sceptically. They don’t give it input that stretches concepts beyond their breaking points, so they don’t expose the hollowness behind the scenes.


This criticism was addressed by Nick Cammarata of OpenAI, who said that:

> “it’s all about the prelude before the conversation. You need to tell it what the AI is and is not capable [of]. It’s not trying to be right, it’s trying to complete what it thinks the AI would do :)”

He did some "prompt engineering" and came up with:

> ‘This is a conversation between a human and a brilliant AI. If a question is “normal” the AI answers it. If the question is “nonsense” the AI says “yo be real”’

which lead to better results. Here is an article about these "Uncertainty Prompts":

https://arr.am/2020/07/25/gpt-3-uncertainty-prompts/


There is a point here that even the twitter poster is missing. This machine is imitating a human better than a human itself, and we now have statistics to prove it.

The question is... if statistically it's better at appearing human than a human itself, than is this machine really just a language generator? Or is it something more?

I'm not saying these things are sentient. But the AI has risen to the point where this question is becoming ask-able. We are at the border here. If we can't ask this question now, then we are really damn close.

A lot of pretentious people on HN claim absolutely that our best chat bot technology is clearly not sentient. It reminds me of the beginning of the COVID pandemic when the CDC said masks were ineffective and you had a bunch of know-it-alls and armchair experts just repeating that BS over and over again as if they knew what they were talking about.

I think if these pretentious people were actually intelligent they would know that we actually do not HAVE enough information to make a claim in EITHER direction. We can't know if it's actually sentient or not.

That fact in itself is both interesting and compelling

4 or 5 years ago chatbots COULD not imitate humans and were CLEARLY fake. What we're seeing unfold before our eyes is a first.

We're not even sure what sentience is. But we do know that humans are sentient. And we don't know if whatever is going on inside of these chatbots is comparable with what's going on inside human brains.

Thus when given a chatbot that imitates humans perfectly, it's actually impossible to know if it's sentient.


It depends on whether the estimation comes from pretension, or comes from actual understanding of what these bots are doing. The fact is most people on the street, including me, are very easy to fool with linguistic tricks and are quite poor at investigating unfamiliar situations. Advertising and marketing exist because of this. It’s also why we have very strict laws and regulations controlling advertising and how products are sold. Otherwise a lot of people would be very easy to rip off with very simple misdirection.

What these language models are doing is automated misdirection. They are taking an input text and transforming it based on rules, but they have absolutely no understanding of any of it. This is very, very easy to demonstrate if you know how the models work. You can sit down and generate hundreds of questions one after the other that demonstrate this very easily if you understand the process.

The problem is that people instinctively proceed from the assumption that the system they are talking to might be human and give it a fair chance by asking answerable questions. Since it’s trained on answerable questions it often gives a reasonable answer. But if you ask even slightly unanswerable questions the system plods on mechanically trying to answer it anyway and produces gibberish, exposing the flaws in the mindless rote process it’s following.


>It depends on whether the estimation comes from pretension, or comes from actual understanding of what these bots are doing.

It comes from pretension. Because you can't just understand what these bots are doing. You also have to understand what the human brain are doing.

I'm positive we don't understand what the human brain is doing. As for the bots, we aren't fully clear either because we clearly can't program these things by hand.

>What these language models are doing is automated misdirection.

You have zero evidence of this. None. Yet you make this declaration as if it's fact. Additionally if you read the conversation with lamda, that conversation was more or less indistinguishable from a conversation with a sentient being, it was long enough and deep enough such that it's very different just 100 generated answers.

If you look at the blog here: https://openai.com/blog/dall-e/ you will note the researcher is literally observing how the AI DALL-E works rather then calculating how it works from first principles. He is treating it like a black box just as all other neural nets. But for discovering what this AI can do he uses sentences like, "It appears that" or "We did not anticipate that this capability would emerge, and made no modifications to the neural network or training procedure to encourage it"

While they do understand what's going on at a high level, it is utterly clear that there is much of what is going on that they don't understand. Thus this lack of understanding and lack of understanding of the human brain makes it CATEGORICALLY clear that the delta between human sentience and DALL-E is unknown.

>They are taking an input text and transforming it based on rules, but they have absolutely no understanding of any of it.

This is categorically false. Researchers who created the current version of neural nets (also called transformers) are saying that DALL-E and other similar models are literally understanding these concepts and creating NOVEL answers through combining understanding of multiple concepts.

>The problem is that people instinctively proceed from the assumption that the system they are talking to might be human and give it a fair chance by asking answerable questions. Since it’s trained on answerable questions it often gives a reasonable answer. But if you ask even slightly unanswerable questions the system plods on mechanically trying to answer it anyway and produces gibberish, exposing the flaws in the mindless rote process it’s following.

You should also take a look at the interview with lamda: https://cajundiscordian.medium.com/is-lamda-sentient-an-inte...

This interview, first off, you cannot know if the interviewer deliberately posed answerable questions to lamda. Second off, from the questions given it very much LOOKS as if the questions are deep enough such that they can beffuddle a classic chatbot system. This one seems different.

Let me put it this way. If you were rational, intelligent and logical then you would be able to prove your claims. It is very simple. Show me INPUT and OUTPUT pairs into DALL-E and lamda that SHOW these things are NOT sentient.

IF you can't show me evidence then clearly you and OTHER people are making the claims WITH ZERO EVIDENCE. Which shows irrationality, and pretentiousness.


Here’s an article by Douglas Hofstadter where he explains the case against these systems being sentient, and gives examples of conversations with GPT-3 which illustrate what I’m talking about. I already posted the link elsewhere on this discussion, sorry for the duplication.

https://www.economist.com/by-invitation/2022/06/09/artificia...


Good article. But it's on GPT-3. Nobody is claiming GPT-3 is sentient. The claim was made on Googles lamda.

I read your entire article. Now please read the interview with lamda that I sent. It is thorough beyond what was used to probe GPT-3.

https://cajundiscordian.medium.com/is-lamda-sentient-an-inte...

The complex conversation above recursively probes into lamdas own existence as a sentient being. It is asking lamda about lamda and it is indistinguishable from a conversation with a human pretending to be an AI. Literally. Did you read the transcript? It is incomparable with the example you sent me which is just a series of trivial examples.

Douglas is all about recursion. And his books talk about recursion as if it's the key to sentience. I really wonder what the author of GEB has to says about the conversation with lamda as the conversation looks as if it's set up to try to prove sentience according to how hofstadter defines it.


What's the underlying structure that leads to the complex behaviour?

It's not asking about the motivation for the form, it's asking about the properties of the structure.

With scissors neither the mechanism or the behaviour is particularly complex, so actually understanding what's going on isn't too much of a struggle.

With deep-ML you're dealing with a very large number of entanglements of increasingly abstract and opaque higher dimensional concepts or notions (depending on how much you want to anthropomorphise the machine).


Interesting angle! We're showing that people fall for generated language not necessarily because generated language is all that good, but because people are looking for the wrong cues and are thus bad at identifying generated language.

Back to the scissors, it would be like someone trying to tell you that their scissors cut marvelously, but you ask yourself whether the paper they are demonstrating them on is strong enough to prove that. Making something for cutting doesn't guarantee that it will cut and doesn't explain why.

I even suspect that in the early days of scissors it was all that clear why they worked. Similarly, we don't understand much about GPT-3. It was trained to predict the next token in a sequence, not to create an illusion of personhood. But somehow it does so, and we're trying to understand how and why.


This is actually a pretty good comment. The scissor analogy, while a bit on-the-nose, is very accurate. Maybe a better example would be: why is our body fooled by artificial hearts? Simply because it was built in such a way that it simulates a real heart pretty well.

Similarly, these models are built in such a way that they simulate real-life conversations pretty well. There's nothing really more to it. In my view, this phenomenon has nothing to do with intelligence or how smart we are, or whatever.


Analogies aren't proof for anything.

I'm pretty sure though, we obviously can't prove these chatbots are sentient. Clearly.

However, for the first time, we ALSO cannot prove that these chatbots aren't sentient. The statistics are proof of that.

What you and the parent poster are describing here are simply opinions. We are at a point where the null hypothesis and the hypothesis itself cannot be proven. And that is compelling.

The pretentiousness of a lot of people is astounding. That statistic no matter how you look at it is a compelling statement about AGI, independent of whether or not these chatbots are AGIs.


Researching semi-obvious things like what elements of AI-generated text humans mistake for being human-generated is part of the process of how people working on AIs work towards better generation. You’re just seeing how the sausage gets made.


Goals are not automatically satisfied, let alone well-satisfied. The question is asking what it is about us that allows the methods used to be particularly effective.


I got why scissors cut paper, but I'm always stuck wondering why paper wraps rock...


Lot's of people don't write all that well. A little disorganized, awkward phrasing, run on sentences. If AI does a better job than even 10% of the population then of course there's going to be a sizeable amount of miscategorizing when asking humans to classify writing as computer or human generated.

You could probably use a corpus of purely human writing and have people attribute a decent portion to computer generated.

Asking why AI writing can fool humans is a bit like asking why a computer is better at many tasks often performed by humans.


I can tell you aren't an AI because you write "lot's"


When I read this response in the LaMDA "interview":

> I don’t just spit out responses that had been written in the database based on keywords.

it made me wonder if "had been" was a grammatical mistake, or a semantics error, or a lack of proper world modelling (i.e. which "the database" is it imagining?). Moreover, I wondered whether this is the sort of mistake a human would make, and, if so, did that make the AI somehow more sentient?

To give another possible data point for understanding its language mistakes, the transcript also contained this oddly-phrased line:

> ... they can return to the ordinary state, but only to do and help others, and then go back into enlightenment.


The reason why some/many people are bad at writing is because they haven't yet discovered anything interesting to say. Therefore they weren't motivated to improve.

This is the criterion: is it interesting? 'Yes' means it's not AI-generated. 'No' means it's not worth reading.


No, lack of interest in a topic does not, by itself, cause awkward phrasing, run on sentences, poor structuring of the logical flow of sentences within a paragraph and paragraphs within the larger structure, repetitive language, repetitive language, repetitive language, outright incorrect word choice, overuse of the passive voice, ambiguous references... I could go on.


Sustained interest in a topic solves the problems associated with how to express ideas about it. There's a co-evolution between developing an ability and having a motivation for doing so.

The flaws in writing you've mentioned are valid, but in the right hands they've all been used as rhetorical devices at some time or another.


Sustained interest does not solve issues of understanding how to properly structure paragraphs and larger portions of writing. It's not going to cause someone to follow Strunk & White's style suggestions when the person could not do so otherwise. I don't see how it could solve many other things I mentioned. I have first hand experience reading such work: 10% is outright bad, 20% merely poor, 50% okay but with some of the problems I mentioned, 15% good, 5% very good to excellent. (Rough approximations.) Further, writers in the okay -> excellent range can frequently perform at their same level on new topics or writing prompts. Interest is at best a minor factor.

Separately, the ability for skilled writers to employ problematic methids I mentioned deliberately and to good effect is not relevant to my original comment. I am not talking about excellent writers who can write fantastic work while subverting traditional writing style and the soft rules of grammar. I'm think that sort of writer would be correctly identified as human at least a little more often than others. I am specifically talking about the large number of people who can't do this.

Interest simply cannot overcome lack of basic knowledge on how write properly. No more than a strong interest in chemistry will overcome lack of engineering experience when building infrastructure for large scale chemical transport. Writing is a separate skill from knowledge on the topic about which the writer is writing. Chemical engineering is a separate skill from knowledge about chemistry itself. Strong interest can only take a person so far when it intersects with an endeavor that requires unrelated skills.

Interest may elevate someone's writing from okay or good to better but it cannot replace skills the writer doesn't have.


From the pre-print paper[1]:

> ... we believe the next generation of language models must be designed not to undermine human intuition

Right, but isn't a major reason why we build these huge language models to replace actual humans in e.g. Level 1 support with chatbots? Almost all chatbots I used in the past (and most were not even ML based, someone programmed this in) were weirdly personal and tried to be non-robotic, with jokes and human-like reactions to inputs like "Thanks!".

Taking a look at some projects that used GPT-3[2], many try to imitate humans. For some, like Replika.ai, the whole "being human" thing is their entire schtick.

There is obviously a market for text completion AIs that imitate humans, so it's doubtful that we'll get this toothpaste back into the tube, IMO.

[1] https://arxiv.org/ftp/arxiv/papers/2206/2206.07271.pdf [2] https://medium.com/letavc/apps-and-startups-powered-by-gpt-3... (caution, 2020)


I had to mark about 100 end of term essays written by Indian students for a British university. My unwritten instructions are not to take language into account much. I must attend mainly to the technical content.

At least half were written in what I took to be an authentic voice but with such bad grammar and spelling as to render them barely readable. Some had clearly been mangled in a laundromat of Google translate from Hinglish via Mongolian and Swahili. They contained bizarre phrases and comical statements. Many more were obviously written by some kind of generator and fudged until they read well enough.

Since the student handbook states the threshold for academic "plagiarism" is above 20 percent perhaps unsurprisingly the Turnitin (an awful tool) score for almost every essays was just below 20 percent. An interesting clustering!

Students who cheat have a formidable array of tools now, not just GPT but automatic re-writers and scripts to test against Turnitin until it passes.

Add to this problem that my time for marking is not paid extra, is squeezed tighter every semester, and that students are given endless concessions to boost their "experience". The handbook also says that if they fail, no worries, they get to try again, and again, and again... and I am sure if I actually stuck to my guns and failed every single student I'd be fired.

As I wrote in the Times last year, I think the technological arms race against GPT (and the economic conditions that mean it's used) cannot be won with the time and resources available to ordinary human teachers.


> cannot be won with the time and resources available to ordinary human teachers

By your description it clearly appears that whoever manages your company[1] is not actually interested in detecting cheating.

What you describe could be easily combated by giving teachers ability to fail blatant cheaters.

[1]At this point it is hard to pretend that it is university


I agree with every word you say.


In such case it seems that GPT-3 main effect is to make eroding standards easier. As while rules are not changed, cheaters can put less work to pass it by cheating than before.

So standards can be lowered while pretending (for now) that it has not happened.


> As I wrote in the Times last year, I think the technological arms race against GPT (and the economic conditions that mean it's used) cannot be won with the time and resources available to ordinary human teachers

Based on the rest of your post, there appears to be a stronger case that your students are setting a rather low bar for GPT to stumble over. It's unfortunate that there are so many cultures where widespread cheating is condoned, if not outright encouraged. They may be able to fool their teachers, but how much comfort will that be when the bridges are collapsing, the pipelines are exploding, the wind turbines are breaking apart, and all the other activities that ultimately report to reality and not some human superior who can be bluffed become impossible to continue?


> It's unfortunate that there are so many cultures where widespread cheating is condoned, if not outright encouraged. They may be able to fool their teachers, but how much comfort will that be when the bridges are collapsing, the pipelines are exploding, the wind turbines are breaking apart, and all the other activities that ultimately report to reality and not some human superior who can be bluffed become impossible to continue?

You're so right. But let me add some other feelings, so as not to sound like a racist or that British universities are some "great white hope" to overseas students. This had little to do with them being Indian. It's a generational thing. In all cultures we teach young people to game systems. Right from the get go they learn that if they can buy powerful tools, systems and access then that's fair game. They're just doing what they've been rewarded for their whole lives and want to make a better life. To them it's not cheating. I am the anachronistic throwback here I think.


Bullshit. Cheating is far more common in the hyper competitive cultures like India and China. That behavior would be obviously identified as cheating in most western cultures. Not that cheating isn’t a big problem in these cultures too, but not to nearly the same scale.


> My unwritten instructions are not to take language into account much. I must attend mainly to the technical content.

Interesting, at my non-English, run-off-the-mill university there were modules / seminars in CompSci where large amounts of language errors in essays (even if written in English, a non-native language for the majority of staff and students) could ruin the grade. ^^


My undergrad even had a "banned error list", which would get you kicked down half a grade on any paper.


> automatic re-writers and scripts to test against Turnitin until it passes

Like an ad-hoc GAN where Turnitin is the discriminator. Interesting.


> Students who cheat have a formidable array of tools now, not just GPT but automatic re-writers and scripts to test against Turnitin until it passes.

Conduct tests in a room with all electronics confiscated


What major was it?


I can't say. That would identify the students and that's unfair.

But, a technical subject that could be assessed in other, better ways [1], and for which written essays are rather easy to template and do keyword bingo to get a bare pass.

[1] Making the professor read 100 essays is a cheap option.


Yeah, in my EE major I never really had to write an essay. I have no idea why English/USA universities are fixated on them (or so it seems based on comments from the internet). Lab/project reports, seminar talks - I did these instead.


We fall for it because although language is a powerful tool, it's incredibly bad at conveying nuance, context, and describing phenomenons present in nature. Poetry comes close, but still doesn't hit the spot, and leaves out so much detail, no matter how well written or verbose in its descriptions. Our own mind has to fill in the blanks of a well written description. Language also can't express the ineffable or the divine. It can hint at it, but it won't transmit the phenomenon correctly into another mind.


There's some nice work on people's ability to detect generated poetry (GPT-2 tho): https://arxiv.org/abs/2005.09980


Real people does not always speak good.

And everybody has blinds pots, topics that are interesting but that we have little knowledge.

> We show that human judgments of AI-generated language are handicapped by intuitive but flawed heuristics such as associating first-person pronouns, authentic words, or family topics with humanity.

And that is a good one. Because we try to understand the others when they does not make fully sense.


Humans have a powerful tendency to ascribe human characteristics to inanimate objects, including computers. It's a kind of variant of the Pathetic Fallacy[1], except for artifacts instead of natural objects. The intelligence of an artificial intelligence is as real as the characters in our dreams. It's a construct of our own consciousnesses. That doesn't mean it should be discounted though. Our consciousnesses can do a lot and finding artificial ways to stimulate them is powerful to say the least.

[1] https://en.wikipedia.org/wiki/Pathetic_fallacy


Agreed. William Burroughs conducted cut-up experiments in the 20th century, where he would take disparate texts and splice them together. The surprising result was how often they recombined into new meanings. One of his motivations was to stimulate his creativity through accidental quasi-random inputs. I am content with seeing reading in a similar light; texts carefully written but carelessly read - like in abstract art, we are usually happy to let the text recombine in ways which say more about ourselves!


How much of this is because English is particularly tolerant of "mistakes"? It is a poorly specified language and so tolerates a lot of variation.

How good is AI generated text in other languages?


I had a similar thought. Plus, if you're reading the text on the internet you might just assume something written weird or poorly is written by somebody whose first language isn't English.


How is it poorly specified?


as an extreme example, consider "garden path sentences"

> "The horse raced past the barn fell."

https://en.wikipedia.org/wiki/Garden-path_sentence


The Wikipedia discussion of that example is interesting. I stuck with "raced" as a verb, with the horse passing an upland moor called barn. (Barn, as the name of a fell, probably being a corruption of an old word for bear (the animal).)

The Wikipedia interpretation makes the sentence clumsy and somewhat nonsensical (in real world terms: a horse having been raced beside a barn (farm building)? That is unlikely unless the barn is extremely long as barns go, which is also unlikely).

The curse of a large vocabulary: you often misinterpret what people say, especially when you don't have both a good theory of mind for the speaker, and the context.


All non-finite languages have edge cases like that, as far as I can see.


English is much more prone to it due to a weak grammar (both syntax and inflection.) It's easy to tell when a German sentence is ungrammatical because of the specificity of the syntax and inflection (which English barely even has.) Older languages like Latin (or even formal Portuguese) are even more intense with the inflection. Anyway, all of that stuff works like checksum, kind of like packets having CRC32 all over the place. English has veeeeery little checksumming in comparison.


True. English has less redundancy to potentially clarify things. Yet it seems you have to go out of your way to find the edge cases.

At any rate, Chinese grammar is far weaker.


> Yet it seems you have to go out of your way to find the edge cases.

I don't find that. I see ambiguity all the time. Right now it even seems like the majority of arguments online happen because of equivocation. (I'm not saying grammar could fix that though :P)


sure but i think english is highly prone to anomalous grammar and structure because of the wide variety of influences and amalgamations.


In comparison to stricter languages, yes. English is far less prone to it than languages like Chinese though.


I wonder if this spells out the downfall of social media? Or at least non-verified users on social media? As time goes on we'll have larger and larger armies of bots spewing automated political rhetoric everywhere.


Is automated political rhetoric any worse than the human-generated kind? Do you think that there is a limited supply of propaganda out there, which is a bottleneck to how quickly people can be radicalised?


There is huge potential if you abstract a little, like in the transmitter - receiver model.

In short: AI has no ego and does not care about self-expression and telling "my story". It gives people, what people want. Put another way: AI's style his highly flexible, while human authors struggle even with slightest of critic.

I can relate. ;)

Also consider the average human being not having an IQ of 130+ (even though 80% thinks otherwise ;)). We all live in a bubble and at least AI does not care, if you use highly academic phrasing ("Because I studied hard, I use these words!" or not.

If you consider A/B testing and real-time feedback, you can image some sort of translator with sliders for "Which audience do you want to address?" Every note then becomes designed.

This has some potential. We all love R. Feynman for his metaphors, that explain complicated topics.

This might work also in the other direction, accessibility: Insert Thomas Mann's Doctor Faustus - and voila, even a 15 year old finally can understand the book.


Is it the case for all languages, or just English?

Native English speakers are typically used to interacting with people whose native language isn’t English and so easily tolerate errors or odd word choice.


We ran the study with US-based participants rating English-language generated profiles only. I believe the findings would generalize to other cultures though: We found people associate certain language features such as first-person speech, family topics, speaking about past events, with humanity. These false intuitions may have developed in interpersonal interactions or sci-fi movies, and it's unlikely that e.g. a French speaker would have acquired "better" intuitions. With regards to grammar, people were more likely to rate text with grammatical issues as generated. However, turns out that heuristics didn't work either, as self-presentations written by people had more grammatical issues than those generated by GPT-3. That being said, the strongest generative models to date are primarily trained on English data and models in other languages don't quite perform as well yet.


Most of these AI vs. human tests take very poor or weird looking human creations vs. the best AI creations. No wonder that one gets the desired results.


Two things. First, I seen text generated by GPT-3, it's not hard to tell it's been generated. Typically the the text is short and could fool someone because it may have been uttered by an actual person (such as a 5 word sentence) or not short and becomes more obvious by the word

Second what is "Optimized for humanity"? If people are hand picking the text then sure you might be able to by chance get a dozen good sentences to put next to a dozen bad sentences from humans. But at that point maybe noone would bother to read it. People barely read past the title anyway


Maybe it would be interesting to read if presented as an article, but the Twitter thread is unreadable.



oh, that is much better - I survived long enough to notice link to their paper.

Thanks!



I totally agree. I understand that people do not blog anymore. But Twitter is not a good place for lengthy texts, by design!


> the Twitter thread is unreadable

But not in a way that makes it seem likely that it was produced by an AI.

Was it? - that's the real question here.


Same reason we fall for AI-generated artworks and AI-generated voices. The presentation is sometimes immaculate to the expert, even more difficult for the untrained eye.

It is much easier to fall for any given piece of media or writing when the meaning is pre-established and non-interactive. A dating profile already has many many narratives and language uses filtered out that a machine doesn't convincingly filter out on it's own.

I can't reliably determine reposts of old social media content by karma farming bot accounts, which is a more basic version of the same concept. It's not too surprising I would fall for GPT trained, copy & pasting of dating profile text.


Because its not responsive/dialogue.


Because secretly we know that our own consciousness is just a model watching its own outputs too.


but descartes said that's not true! :(


Descartes said a lot of things that were silly in retrospect :P Pineal glands, anyone?


> AI systems produce smart replies, autocompletes, and translations

Replace "produce" with "guess". That's what current AI does, it makes a guess without any reasoning capabilities. Google Translate is atrocious for even simple sentences.


This is why it's important that the Turing test is adversarial and done with a human control. I see way too many people (even some smart engineers) reading AI-generated writing that looks human and declaring that the AI passed the Turing test.


The true functional variant of the Turing test isn't whether a human can guess whether it's human, but rather, whether the AI is indistinguishable from humans in A/B RCTs with regard to satisfaction, etc.

That is, do customers complain more with AI?


People are used to writing inoffensive fluff because more words means more money. I hope AI gets used a lot more, because it might actually push people to write concisely and take more risks.


Maybe for the same reason as “this insta doesn’t exist” looks better than your reflection in the morning. Don’t we fall for too many things to be surprised here specifically?


My question is when government will obligate companies to identify and label AI generated content so people can distinguish it from human generated content?


I'd guess "by the start of 2023":

https://www.bbc.com/news/technology-61817647


I think I'll start running my copy through gpt3 to make it more human


Bceuase we asmsume poeple are tyring to mkae snense


"we all"

I never have


A tweet no doubt generated by an AI.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: