Discussion about this post

User's avatar
Hania's avatar

Speaking of the current AI hype

I was once looking to change domains, and I remember my friend joking with me, back in the day, that Big Data was so hot, anybody who could spell Hadoop was getting hired. A few years later, I saw - from the outside - what it was like when Big Data was no longer hot.

We see hype cycles come and go, and I think it is a rite of passage for each software engineer to decide for themself what is real gold and what is fake glitter. Unfortunately, we are getting a lot of non-disclosed advertising of AI touting its legendary benefits, which is unethical, like “adding poison to honey” (Arabic expression).

I see the current AI hype mania as a Critical Thinking Test as well as an Integrity Test for individuals as well as organizations. I am certain we will live to see the benefits of LLMs and that these will become consistent and well understood, but we will also see all the AI-hype critical thinking and integrity test failures as well, and those are going to be enormously painful for the affected organizations.

Expand full comment
Hania's avatar

Thank you for sharing the useful use cases - I’m always looking out to grow my awareness of when to use LLMs!

You mentioned two usecase:

* Research and data collection

* Reporting and summarization

These two reminded me of this: https://buttondown.com/maiht3k/archive/information-literacy-and-chatbots-as-search/

Here are my TL;DR notes of that page:

Why are LLMs not a good technology for information access?

A. [Unreliable results] LLMs are statistical models of the distribution of word forms in text, set up to output plausible-sounding sequences of words. (The system being wrong, say, 5% of the time could be disastrous if people come to trust it as being correct)

B. [Sense-making is hindered, and problem solving muscles atrophy] Information literacy sense-making is undermined by getting "the answer" from a system, even if the hallucination problem were to be solved for high accuracy

1. Here are the usual sense-making actions a person carries out:

a. Refine the question

b. Understand how different sources speak to the question (and locate each source within the information landscape)

So getting "the answer" without this sense-making could mislead the user never to refine their question nor understand the information sources. Also, there are the same blindspots as web search that we need to keep in mind once we see the top results of a search query: 1) what about the low-ranked search results? 2) what about the results the search engine saw but excluded from the search results? 3) what about the data that is outside the corpus available to the search engine?

2. [Judging reliability of different information sources] You need to know the information sources that were used for the chatbot's answer, and judge the reliability of each (so the algorithm making some sporadic prioritisation and conclusion is suspect, because there's always missing context the chatbot doesn't know about)

3. [Missing out on advice from other people] Engaging with people in a discussion forum cannot be substituted by an algorithm

4. The chatbot output is manipulated to sound friendly and authoritative, which is misleading

C. Even with Retrieval Augmented Generation (RAG), it is difficult to detect when the summary you are relying on is missing critical information

D. Even with including links to sources, the system presents a set conclusion, which discourages the reader from following the links and making up their own mind

Expand full comment

No posts