The Unix Heritage Society mailing list
 help / color / mirror / Atom feed
From: Luther Johnson <luther.johnson@makerlisp.com>
To: tuhs@tuhs.org
Subject: [TUHS] Re: On the unreliability of LLM-based search results (was: Listing of early Unix source code from the Computer History Museum)
Date: Sat, 31 May 2025 15:47:49 -0700	[thread overview]
Message-ID: <5aa768ae-8853-f0da-9780-53ca4e9d486f@makerlisp.com> (raw)
In-Reply-To: <CAO2qRdMHAUHdPj9odydp3c9YwfaaU2pZiR6nmNS8O3r=rjKfWw@mail.gmail.com>

[-- Attachment #1: Type: text/plain, Size: 5910 bytes --]

I think we could call many of these responses "mis-ambiguation", or 
conflation, they mush everything together as long as the questions posed 
and the answers they provide are "buzzword-adjacent", in a very 
superficial, mechanical way. There's no intelligence here, it's just 
amazing how much we project onto these bots because we want to believe 
in them.

On 05/31/2025 03:36 PM, James Johnston wrote:
> Well, I have to say that my experiences with "AI based search" have 
> been beyond grossly annoying. It keeps trying to "help me" by sliding 
> in common terms it actually knows about instead of READING THE DAMN QUERY.
>
> I had much, much better experiences with very literal search methods, 
> and I'd like to go back to that when I'm looking for obscure papers, 
> names, etc.  Telling me "you mean" when I damn well DID NOT MEAN THAT 
> is a worst-case experiences.
>
> Sorry, not so much a V11 experience here, but I have to say it may 
> serve the public, but only to guide them back into boring, 
> middle-of-the-road, 'average mean-calculating' responses that simply 
> neither enlighten nor serve the original purpose of search.
>
> jj - a grumpy old signal processing/hearing guy who used a lot of real 
> operating systems back when and kind of misses them.
>
> On Sat, May 31, 2025 at 2:53 PM Luther Johnson 
> <luther.johnson@makerlisp.com <mailto:luther.johnson@makerlisp.com>> 
> wrote:
>
>     I agree.
>
>     On 05/31/2025 01:09 PM, arnold@skeeve.com
>     <mailto:arnold@skeeve.com> wrote:
>     > It's been going on a for a long time, even before AI. The amount
>     > of cargo cult programming I've seen over the past ~ 10 years
>     > is extremely discouraging.  Look up something on Stack Overflow
>     > and copy/paste it without understanding it.  How much better is
>     > that than relying on AI?  Not much in my opinion.  (Boy, am I glad
>     > I retired recently.)
>     >
>     > Arnold
>     >
>     > Luther Johnson <luther.johnson@makerlisp.com
>     <mailto:luther.johnson@makerlisp.com>> wrote:
>     >
>     >> I think when no-one notices anymore, how wrong automatic
>     information is,
>     >> and how often, it will have effectively redefined reality, and
>     humans,
>     >> who have lost the ability to reason for themselves, will
>     declare that AI
>     >> has met and exceeded human intelligence. They will be right, partly
>     >> because of AI's improvements, but to a larger extent, because
>     we will
>     >> have forgotten how to think. I think AI is having disastrous
>     effects on
>     >> the education of younger generations right now, I see it in my
>     >> workplace, every day.
>     >>
>     >> On 05/31/2025 12:31 PM, andrew@humeweb.com
>     <mailto:andrew@humeweb.com> wrote:
>     >>> generally, i rate norman’s missives very high on the
>     believability scale.
>     >>> but in this case, i think he is wrong.
>     >>>
>     >>> if you take as a baseline, the abilities of LLMs (such as
>     earlier versions of ChatGP?) 2-3 years ago
>     >>> was quite suspect. certainly better than mark shaney, but not
>     overwhelmingly.
>     >>>
>     >>> those days are long past. modern systems are amazingly adept.
>     not necessarily intelligent,
>     >>> but they can (but not always) pass realistic tests, pass SAT
>     tests and bar exams, math olympiad tests
>     >>> and so on. and people can use them to do basic (but realistic)
>     data analysis including experimental design,
>     >>> generate working code, and run that code against synthetic
>     data and produce visual output.
>     >>>
>     >>> sure, there are often mistakes. the issue of hullucinations is
>     real. but where we are now
>     >>> is almost astonishing, and will likely get MUCH better in the
>     next year or three.
>     >>>
>     >>> end-of-admonishment
>     >>>
>     >>>     andrew
>     >>>
>     >>>> On May 26, 2025, at 9:40 AM, Norman Wilson <norman@oclsc.org
>     <mailto:norman@oclsc.org>> wrote:
>     >>>>
>     >>>> G. Branden Robinson:
>     >>>>
>     >>>>    That's why I think Norman has sussed it out accurately. 
>     LLMs are
>     >>>>    fantastic bullshit generators in the Harry G. Frankfurt
>     sense,[1]
>     >>>>    wherein utterances are undertaken neither to enlighten nor
>     to deceive,
>     >>>>    but to construct a simulacrum of plausible discourse. 
>     BSing is a close
>     >>>>    cousin to filibustering, where even plausibility is
>     discarded, often for
>     >>>>    the sake of running out a clock or impeding achievement of
>     consensus.
>     >>>>
>     >>>> ====
>     >>>>
>     >>>> That's exactly what I had in mind.
>     >>>>
>     >>>> I think I had read Frankfurt's book before I first started
>     >>>> calling LLMs bullshit generators, but I can't remember for
>     >>>> sure.  I don't plan to ask ChatGPT (which still, at least
>     >>>> sometimes, credits me with far greater contributions to Unix
>     >>>> than I have actually made).
>     >>>>
>     >>>>
>     >>>> Here's an interesting paper I stumbled across last week
>     >>>> which presents the case better than I could:
>     >>>>
>     >>>> https://link.springer.com/article/10.1007/s10676-024-09775-5
>     >>>>
>     >>>> To link this back to actual Unix history (or something much
>     >>>> nearer that), I realized that `bullshit generator' was a
>     >>>> reasonable summary of what LLMs do after also realizing that
>     >>>> an LLM is pretty much just a much-fancier and better-automated
>     >>>> descendant of Mark V Shaney:
>     https://en.wikipedia.org/wiki/Mark_V._Shaney
>     >>>>
>     >>>> Norman Wilson
>     >>>> Toronto ON
>
>
>
> -- 
> James D. (jj) Johnston
>
> Former Chief Scientist, Immersion Networks


[-- Attachment #2: Type: text/html, Size: 9462 bytes --]

      parent reply	other threads:[~2025-05-31 22:48 UTC|newest]

Thread overview: 5+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
     [not found] <F7093F5EDCBB735E2C7C473314D40D5A.for-standards-violators@oclsc.org>
     [not found] ` <CAEdTPBeFUcxAZWn1=mZnwTmF2a3DN-1GnXXB6WmV5gaqZHz1Lw@mail.gmail.com>
     [not found]   ` <769a9c94-055d-4bdd-a921-3e154c3b492f@infinitecactus.com>
     [not found]     ` <CAEdTPBcr2ajyAQh24LPtiQLBjfe2G2MYwoq8x_3bt6TzOT1_BA@mail.gmail.com>
2025-05-26 18:45       ` [TUHS] Wikipedia anecdotes - LLM generalizations [was On the unreliability of LLM-based search results (was: Listing of early Unix source code from the Computer History Museum) Charles H Sauer (he/him)
     [not found]     ` <87frgqequk.fsf@gmail.com>
2025-05-27  3:08       ` [TUHS] Re: On the unreliability of LLM-based search results George Michaelson
2025-05-26 20:36 ` [TUHS] Re: On the unreliability of LLM-based search results (was: Listing of early Unix source code from the Computer History Museum) Noel Hunt
     [not found]   ` <CAKzdPgzNBUiT4GeQUnBX38+3dtNY=2Gw=9mFUy2anMoO4DUECg@mail.gmail.com>
2025-05-27 18:11     ` [TUHS] Mark V Shaney (Re: Re: On the unreliability of LLM-based search results (was: Listing of early Unix source code from the Computer History Museum)) Dan Cross
     [not found] ` <DEBB648F-52A0-4E52-AC26-E2067FE7E0CD@humeweb.com>
     [not found]   ` <3e4339e9-bf9a-2b72-b47a-f20f81a153b5@makerlisp.com>
     [not found]     ` <202505312009.54VK97bQ4163488@freefriends.org>
     [not found]       ` <0adb7694-f99f-dafa-c906-d5502647aaf0@makerlisp.com>
     [not found]         ` <CAO2qRdMHAUHdPj9odydp3c9YwfaaU2pZiR6nmNS8O3r=rjKfWw@mail.gmail.com>
2025-05-31 22:47           ` Luther Johnson [this message]

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=5aa768ae-8853-f0da-9780-53ca4e9d486f@makerlisp.com \
    --to=luther.johnson@makerlisp.com \
    --cc=tuhs@tuhs.org \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).