The Wiert Corner – irregular stream of stuff

Jeroen W. Pluimers on .NET, C#, Delphi, databases, and personal interests

  • My badges

  • Twitter Updates

  • My Flickr Stream

  • Pages

  • All categories

  • Enter your email address to subscribe to this blog and receive notifications of new posts by email.

    Join 1,862 other subscribers

Archive for the ‘LLM’ Category

Kevlin Henney on generative AI creating job security for programmers:

Posted by jpluimers on 2024/08/13

Kevlin Henney being interviewd by Richard Seidl

Kevlin Henney being interviewd by Richard Seidl [Wayback/Archive] MDVxFQrqZnh1OxlP.jpg (1200×675)

The quote from this abstract of the January 2024 interview with Kevlin Henney by Richard Seidl  is important:

You really need to understand history. First of all, you need to understand history. Then, you need to understand language. And you need to go and talk to some customers. And then, you will realize how safe your job is. Because programming is not merely the assembly of syntax. It is the application of precision. It is the seeking of precision.And what is the answer? What is it that I’m trying to do?And it turns out that if you specify something badly in natural language, it works out even worse than if you did it in code.And we already know, for example– we can actually take inspiration from the most widely used programming paradigm on the planet, the spreadsheet. What we know from the spreadsheet is that most people who use a spreadsheet do not have a software development background.

Yes.

We also know that most spreadsheets are unmaintainable, incomprehensible, and buggy. If we are saying that the future of software development is people who are not software experts doing this stuff, your job is safe.

It is a fragment of the vodcast episode [Wayback/Archive] Software Engineering im Jahr 2034 – Richard Seidl which limits the quote to this:

Read the rest of this entry »

Posted in AI and ML; Artificial Intelligence & Machine Learning, Conference Topics, Conferences, Development, EKON, Event, GitHub Copilot, LLM, Software Development | Leave a Comment »

CrazyMyra: “After AI took his job as an online assistant, Mr Clippy was obliged to seek work in other sectors…” – beige.party

Posted by jpluimers on 2024/07/30

I love the new title-text for the 2018 “Clippy” picture at [Wayback/Archive] CrazyMyra: “After AI took his job as an online assistant, Mr Clippy was obliged to seek work in other sectors…” – beige.party

A metal toilet paper holder in a corner od a bathro,with an empty roll, that looks similar to a large paperclip

Read the rest of this entry »

Posted in AI and ML; Artificial Intelligence & Machine Learning, Conference Topics, Conferences, Development, Event, Fun, History, JavaScript/ECMAScript, LifeHacker, LLM, Meme, Office, Power User, Scripting, Software Development, Web Development, Windows | Leave a Comment »

Without prior warning, Twitter shares your data with grok AI, even for EU Twitter users

Posted by jpluimers on 2024/07/27

Mentioned this on various social media already yesterday, as then suddenly  – even for EU users, which is against their GDPR regulations – Twitter turned on data sharing with Grok AI of your Twitter data at x.com/settings/grok_settings (direct settings link) without given prior warning at all

[Wayback/Archive] GTarqIOWEAAs4jy.png (768×290)

I got this default setting despite living in The Netherlands and Twitter knowing that:

Read the rest of this entry »

Posted in AI and ML; Artificial Intelligence & Machine Learning, Development, GDPR/DS-GVO/AVG, LLM, Power User, Privacy, SocialMedia, Software Development, Twitter | Leave a Comment »

Goldman Sachs: AI Is Overhyped, Wildly Expensive, and Unreliable

Posted by jpluimers on 2024/07/24

If even companies that normally charge a fukcton of money* to advise the obvious gets it, why are so many still falling for it?

[Wayback/Archive] Goldman Sachs: AI Is Overhyped, Wildly Expensive, and Unreliable

“Despite its expensive price tag, the technology is nowhere near where it needs to be in order to be useful for even such basic tasks”

Via [Wayback/Archive] tldr.nettime – tante: “”What this means in plain Engl…”

“What this means in plain English is that one of the largest financial institutions in the world is seeing what people who are paying attention are seeing with their eyes: Companies are acting like generative AI is going to change the world and are acting as such, while the reality is that this is a technology that is currently deeply unreliable and may not change much of anything at all.”

(Original title: Goldman Sachs: AI Is Overhyped, Wildly Expensive, and Unreliable)

  • do I really need to mention the USD 4 million contact for figuring out that for NYC, putting garbage bags on OTTO garbage wheelie bins would  on the streets would work better than putting plain garbage bags on the streets?

Read the rest of this entry »

Posted in AI and ML; Artificial Intelligence & Machine Learning, Development, LLM, Software Development | Leave a Comment »

GitHub – dabochen/spreadsheet-is-all-you-need: A nanoGPT pipeline packed in a spreadsheet

Posted by jpluimers on 2024/06/12

A great visualisation that LLM are basically a bunch of numbers: [Wayback/Archive] GitHub – dabochen/spreadsheet-is-all-you-need: A nanoGPT pipeline packed in a spreadsheet.

It also shows you that Excel is an excellent tool for working with numbers and formulas on a larger scale.

(note the file is a .numbers file developed in the Mac version of Excel)

Via:

  1. [Wayback/Archive] /Fay-lee-nuh/ on X: “Programmers: Spreadsheets aren’t code @chendabo: Hold my beer”
  2. [Wayback/Archive] Dabo on X: “I recreated an entire GPT architecture in a spreadsheet. It is a nanoGPT designed by @karpathy with about 85000 parameters, small enough to be packed into a spreadsheet file. It is great for learning about how transformer works as it shows all the data and parameters going”

Read the rest of this entry »

Posted in AI and ML; Artificial Intelligence & Machine Learning, Development, Excel, LLM, Office, Power User, Software Development | Comments Off on GitHub – dabochen/spreadsheet-is-all-you-need: A nanoGPT pipeline packed in a spreadsheet

Some phrases that might set apart text-content as LLM generated

Posted by jpluimers on 2024/03/31

Starting the 2022-2023 period, more and more generative AI content has entered search engines.

The below queries give you some pointers on how to spot those. They return scholar articles from 2023 and later.

Note the list is in alphabetical order for easier reading, but the number of results (in parenthesis) are very different from that order. I was quite amazed to see “As an AI language model” scoring 45 results.

Read the rest of this entry »

Posted in AI and ML; Artificial Intelligence & Machine Learning, ChatGPT, Development, GPT-3, GPT-4, LLM, Software Development | Leave a Comment »

Autumn 2023 research: How Is ChatGPT’s Behavior Changing over Time?

Posted by jpluimers on 2024/03/21

[Wayback/Archive] https://arxiv.org/pdf/2307.09009.pdf ([Google Docs PDF view: Wayback] Google Docs PDF view: 2307.09009.pdf) is interesting. The abstract confirms my thought: over time LLM drift over time and seem to become worse at knowledge tasks.

How Is ChatGPT’s Behavior Changing over Time?

Lingjiao Chen†, Matei Zaharia‡, James Zou†
†Stanford University ‡UC Berkeley

Abstract

GPT-3.5 and GPT-4 are the two most widely used large language model (LLM) services.
However, when and how these models are updated over time is opaque. Here, we evaluate the March 2023 and June 2023 versions of GPT-3.5 and GPT-4 on several diverse tasks: 1) math problems, 2) sensitive/dangerous questions, 3) opinion surveys, 4) multi-hop knowledge-intensive questions, 5) generating code, 6) US Medical License tests, and 7) visual reasoning. We find that the performance and behavior of both GPT-3.5 and GPT-4 can vary greatly over time. For example, GPT-4 (March 2023) was reasonable at identifying prime vs. composite numbers (84% accuracy) but GPT-4 (June 2023) was poor on these same questions (51% accuracy). This is partly explained by a drop in GPT-4’s amenity to follow chain-of-thought prompting. Interestingly, GPT-3.5 was much better in June than in March in this task. GPT-4 became less willing to answer sensitive questions and opinion survey questions in June than in March. GPT-4 performed better at multi-hop questions in June than in March, while GPT-3.5’s performance dropped on this task. Both GPT-4 and GPT-3.5 had more formatting mistakes in code generation in June than in March. We provide evidence that GPT-4’s ability to follow user instructions has decreased over time, which is one common factor behind the many behavior drifts. Overall, our findings show that the behavior of the “same” LLM service can change substantially in a relatively short amount of time, highlighting the need for continuous monitoring of LLMs.

Later on, Eric Topol had the very interesting conversation with James Zou below which covers many AI aspects including a lot of LLM ones. Basic takeaways for me are that they are good at repeating things from their training data, making them OK on generating text, sort of OK for grammar, but far form OK from reproducing knowledge, and that it will become harder over time to distinguish LLM generated content from human created content.

The video of the conversation is below the blog signature; here is the link: [Wayback/Archive] James Zou: one of the most prolific and creative A.I. researchers in both life science and medicine – YouTube

Almost all LLMs are being trained on a corpus without curation (curation is way too expensive), resulting in them at best averaging the corpus (as in the foundation, LLM is just a “monkey see, monkey do” on steroids but without the means of self-curating to result in above average generation. I think that given more and more on-line content is being and becoming generated by LLM, and newer LLM will be trained based on the corpus encompassing that content (without the means to filter out LLM generated content), over time LLM will perform worse instead of better.

Via he below series of interesting tweets of which were quoted by a slightly less pessimistic Erik Meijer [Wayback/Archive] Erik Meijer on X: “Regression to the mean.. Nnote some interesting replies as well. I found the one mentioning Eternal September especially fitting. It made me discover [Wayback/Archive] www.eternal-september.org

Today is September 11160, 1993, the september that never ends
No pr0n, no warez, just Usenet

Anyway, the tweets:

Read the rest of this entry »

Posted in AI and ML; Artificial Intelligence & Machine Learning, Awareness, ChatGPT, Development, GPT-3, GPT-4, LLM, Software Development | Leave a Comment »

“Oh shit git” seems to have been succeeded by “Oh shit GitHub Copilot”: ‘Downward Pressure on Code Quality’

Posted by jpluimers on 2024/01/29

Not sure about you, but when I write code I want it to be better – way beter even – than average code.

The problem with any LLM based Generative AI is that it generates text based on the average of the past corpus they were trained with at the time they were trained.

It is exactly why I have been advocating for a while: be careful when using Generative AI, as you get generated text based on the combination of averaging over the LLM corpus with the relatively small prompt you phased trying to reflect a tiny bit of the model of the reality you are trying to write software for.

So I was not at all surprised by this article: [Wayback/Archive] New GitHub Copilot Research Finds ‘Downward Pressure on Code Quality’ — Visual Studio Magazine.

Read the rest of this entry »

Posted in AI and ML; Artificial Intelligence & Machine Learning, Development, GitHub Copilot, LLM, Software Development | Leave a Comment »