Silicon Sampling: Revolution or Bullshit?

Rea­ding time: 5 minu­tes

Can large lan­guage models (LLMs) such as ChatGPT simu­late human respond­ents suf­fi­ci­ently well? A clo­ser look at the mea­ning and func­tion of natu­ral lan­guage in humans and machi­nes can pro­vide an ans­wer to this question.

I was recently asked to give an “expert inter­view” for a master’s the­sis to dis­cuss the topic of “sili­con sam­pling in mar­ke­ting” from a psy­cho­lo­gi­cal per­spec­tive. Sili­con sam­pling refers to syn­the­tic, i.e., arti­fi­ci­ally gene­ra­ted respond­ents or sur­vey data crea­ted by an LLM (large lan­guage model).

The ques­tion is not whe­ther machi­nes think like humans or even deve­lop consciousness—most experts agree that this is not the case and never will be, at least not with the tech­no­logy used today. Rather, the ques­tion is whe­ther purely mathe­ma­ti­cal pro­ce­du­res in trans­for­mer models such as LLMs can achieve suf­fi­ci­ently good results to replace human respon­ses, e.g., in the con­text of mar­ket rese­arch or user testing.

LLMs seem to be ideal for simu­la­ting human response beha­vior: they use natu­ral lan­guage, and their for­mu­la­ti­ons often seem asto­nis­hin­gly plau­si­ble. Humans and machi­nes “meet” in the “natu­ral lan­guage” inter­face. It is the­r­e­fore worth taking a look at what func­tion lan­guage has for humans and what func­tion it has for an LLM, and how it is “pro­ces­sed” dif­fer­ently in each case.

Humans and language

Humans do not gene­rate lan­guage out­put. They com­mu­ni­cate. They use lan­guage as a means of exch­ange to form a shared view of the world with others (even when we are alone, we vir­tually ‘add’ others to our view). The lan­guage nego­tia­ted for com­mu­ni­ca­tion, with its words and grammar, is only one of many ways we com­mu­ni­cate, because we also com­mu­ni­cate through our bodies and the entire social and mate­rial con­text. The frame­work of lan­guage alone would be far too limited.

In com­mu­ni­ca­tion, we refer—more or less effectively—to ver­bal and non­ver­bal aspects, to our phy­si­cal pre­sence in the world, our rela­ti­onship to the world and to other peo­ple, to our inten­ti­ons and moti­va­tions, our situa­tio­nal sta­tes and atmo­sphe­ric impres­si­ons, and there is always a great deal of uncon­scious­ness involved.

When under­stan­ding lan­guage, we inter­pret what we hear or read (and many other clues from the con­text) accor­ding to our view of the world and our expec­ta­ti­ons. We can do this because we our­sel­ves are phy­si­cally ancho­red in the world and share many pre-lin­gu­i­stic and non-lin­gu­i­stic life expe­ri­en­ces with our com­mu­ni­ca­tion partners.

The machine and language

For the machine, the lan­guage inter­face is some­thing com­ple­tely dif­fe­rent. It does not com­mu­ni­cate. It pre­dicts the most pro­ba­ble next word based on highly com­plex text pat­terns and thus remains 100% at the level of data con­tai­ned in lan­guage as lan­guage. It refers asso­cia­tively to other lan­guage com­pon­ents in high-dimen­sio­nal data spaces, but not to non-lin­gu­i­stic ones, such as phy­si­cal or atmo­sphe­ric experiences.

It is then us again who think we reco­gnize mea­ning and signi­fi­cance in the machine’s out­put (and con­fuse it with ‘com­mu­ni­ca­tion about some­thing’), even though for the machine it refers to not­hing bey­ond the bare words. Such pro­jec­tions lead to peo­ple even using an LLM as a per­so­nal coach or the­ra­pist (they then treat them­sel­ves as if in a mirror).

An LLM is the­r­e­fore a powerful tool for ana­ly­zing and pre­dic­ting text pat­terns. Howe­ver, it remains blind to the mea­ning and signi­fi­cance that peo­ple want to express and refer to when com­mu­ni­ca­ting with others.

This human level is not curr­ently found in the huge text libra­ries that have been fed into the LLM, because only a very small part of our ‘inner­most thoughts’ ever find their way into lan­guage or text (or even end up on the inter­net, where they can be used to train LLMs), and then usually in a highly pro­ces­sed form.

It can­not be added to trai­ning, because many things can­not be pro­ces­sed in writ­ten form at all, such as atmo­sphe­res that are felt phy­si­cally or our bodily know­ledge, such as play­ing the piano or riding a bike. There are many things that peo­ple can­not even put into words. Many things (and often the most important things in psy­cho­logy) are sim­ply inde­scri­ba­ble, and some are just hel­p­lessly vague.

Car­bon ver­sus mathematics

Nevert­hel­ess, the lin­gu­i­stic out­put often seems rea­li­stic and human, con­sis­t­ently achie­ving new highs on the Turing test scale. Could it the­r­e­fore be enti­rely suf­fi­ci­ent to reco­gnize and pre­dict pat­terns on a purely tex­tual and sta­tis­ti­cal level?

The ans­wer here also lies in the dif­fe­rent pro­ces­sing of lan­guage. The machine imi­ta­tes what would be most expec­ted on a text level. This crea­tes a high degree of plau­si­bi­lity, pre­cis­ely because AI ope­ra­tes within the logic of sta­tis­tics and pro­ba­bi­li­ties. Thus, even pure sta­tis­tics can some­ti­mes simu­late pos­si­ble human (lan­guage) beha­vior well, eit­her by chance or when the task is very close to the pat­terns in the trai­ning material.

Howe­ver, when faced with novel tasks or situa­tions (which is usually the case in rese­arch pro­jects), the LLM only pro­du­ces sta­tis­ti­cally pro­ba­ble and the­r­e­fore plau­si­ble-sound­ing text con­ti­nua­tions. In addi­tion, syn­the­tic respon­ses are not repro­du­ci­ble. Even with iden­ti­cal prompts, an LLM will pro­duce dif­fe­rent respon­ses depen­ding on the model ver­sion, sys­tem prompts, indi­vi­dual set­tings, or inter­nal ran­dom pro­ces­ses. This varia­tion can­not be explai­ned psy­cho­lo­gi­cally, but has tech­ni­cal reasons.

There is a won­derful term cal­led “bull­shit.” Bull­shit is defi­ned as some­thing that sounds plau­si­ble, but where it does­n’t mat­ter whe­ther it’s true or not. It’s impos­si­ble to decide whe­ther some­thing is a good simu­la­tion or sim­ply non­sense. That’s the crux of bullshit.

Con­clu­sion

As a psy­cho­lo­gist, one should always main­tain a healthy skep­ti­cism when “sili­con sam­pling” is trea­ted as a serious alter­na­tive to inter­vie­w­ing peo­ple. If I want to find out how peo­ple think and feel and why they come to a cer­tain con­clu­sion, it is not enough that some­thing sounds like some­thing peo­ple would say.

I can cer­tainly still use the LLM, e.g., for initial hypo­the­sis gene­ra­tion, if the topic has alre­ady been rese­ar­ched in a simi­lar form and published, or to get an idea of what a spe­ci­fic tar­get group might sound like. But then it is a sup­ple­ment, and I rea­lize that it is the con­fa­bu­la­tion of a machine based on ana­ly­ses of published texts.

This can be extre­mely hel­pful. Howe­ver, it does not simu­late human respond­ents suf­fi­ci­ently well or relia­bly enough, and—importantly for the discourse—it will not do so even with incre­asingly bet­ter or spe­ci­ally trai­ned models, because this only fur­ther increa­ses tex­tual plau­si­bi­lity wit­hout simu­la­ting expe­ri­ence and beha­vior in a more relia­ble way.

(ms)

Addi­ti­ons and comments:

* The post image was crea­ted enti­rely with AI in line with the topic (excep­tio­nally here in the blog – I think it was this Gemini Banana).

** The fact that the trai­ning data for LLMs always refers to the past became clear to me when I asked ChatGPT before the inter­view what “sili­con sam­pling in mar­ke­ting” meant. The ans­wer: “Sili­con sam­pling refers to mini or dummy pro­duct samples that look like real elec­tro­nic devices but are often not func­tional. They serve as hap­tic, visual, or demons­tra­tive samples before a pro­duct is actually deve­lo­ped or pro­du­ced. The term comes from the fact that these samples are often made of sili­cone, pla­s­tic, or 3D printing—i.e., “sili­cone” as a mate­rial, not sili­con (semi­con­duc­tor).” ChatGPT 5.1 must have alre­ady com­ple­ted its trai­ning before the term found its way onto the inter­net. Still, it sounds plausible.

Share Post :

weitere Beiträge