Overview of the GenAI Landscape

Last month ago I posted the outline of a General-audience AI Presentation I gave, but I have a bunch of AI-related links sitting around and a vague idea of how they link together, so here goes…

The levels of understanding

I get to talk to (or read) a broad sample of people–regular non-technical folks, businesspeople anywhere from ops and sales to executives, and technical professionals who range from entirely skeptical of AI tech to practitioners in the space and AI product developers. Because the application of AI is so varied, and the number of software applications so numerous and (somewhat) varied, everyone’s (mostly hot) takes also vary based on what they’ve experienced.

The ignorant

Sometimes people don’t know anything about AI tech or haven’t tried it, and it’s not because they’re intellectually lazy, it’s because they are busy, and, believe it or not, AI tools have not actually wormed their way into everyone’s working life.

The confident skeptic

This group tends to focus on one or all of a few factors:

The practical practitioner

Many people in software and technical fields have adopted various GenAI tools as part of their workflow. Product managers use ChatGPT or Claude to synthesize research and user interview transcripts. Businesspeople “chat with” meeting transcripts and documents, or use the chat UX and “project” UX of the tools to generate documents like decks or responses to inbound information requests. Vibe coders use coding agents to whip up small tools or apps.

The builder

These are the people building entire SaaS products on top of AI tech. They are the ones fine-tuning models, chaining together agentic workflows, often writing their own instrumentation to observe the inputs and outputs through those workflows, and perpetually building and maintaining data pipelines in attempts to get better inputs into those workflows.

I find myself in all four categories in differing degrees. When I read academic papers or even just the blogs of CS or DS grad students, I feel ignorant. I’m skeptical when tech bros or self-help influencers claim AI will change “the game” or “everything forever”. I use a number of AI tools every day for business and technical reasons (I have to process information, write documents, and write code most days). And I’m beginning to understand the complexity of actually building business-value-focused systems that use LLMs.

The frames of reference

There are many different perspectives one can take when approaching new AI tech.

LLM vibe trends:

Mastodon in the morning (Europeans) – This is horrible and the end of society and probably humanity

Mastodon during the day (Americans) – This is interesting

LinkedIn all day – OMGBBQ THIS CHANGED MY LIFE

– Me, on a specific social network

The Pollyanna

The LinkedIn poster who just discovered some new trick with ChatGPT. They’re most likely to attribute human characteristics to the machine. To gush about its capabilities unblinkingly. Gena Gorlin. Dan Shipper.

The optimistic pragmatist

Most of the builders from above. I would pay attention to people like the Future of Coding crew (see Computing Machinery and Intelligence). My AI Skeptic Friends Are All Nuts is making the rounds.

The grizzled pragmatist

They know that even AI software is still software and we’ve never really done agile. There are quite a few in this camp. Baldur Bjarnason is a good one. Jason Gorman is another. Contra Ptacek’s Terrible Article On AI is a (brilliant and funny) counterpoint to the “My AI Skeptic Friends Are All Nuts” article above.

The philosopher

Our attraction to AI is deeper than just a new shiny piece of tech kit.

foxwizard (see Check Your Paradiastoles). Kester Brewin. Todd McGowan and Ryan Engley (keep reading below). Adam Greenfield.

Some people move between these modes regularly. Dario Amodei and Demis Hassabis, for example, seem to zoom out and then back in regularly, although those are probably bad examples as they’re fully submersed in the quicksand and have billions of dollars behind them.

And now with that out of the way

With that context set, I’m going to link to a number of things that have crossed my desk in recent weeks.

Paul Ford has been working on web content and applications for decades. In his most recent iteration, he produces a podcast with his Aboard cofounder Rich Ziade. These two episodes in particular discuss how to think about AI from the perspective of business executives and software product people.

(Link)

But I think there’s a bigger issue. And I think what’s happening is because of this anxiety and fear is they’re skipping the thinking part. They’re just going into execution. They’re buying stuff. They’re hiring people. Like, Chief AI Officer has become a thing…

Tip one: Do not start any conversation, when you talk to your team and your leaders, with, “What are we gonna do–what is our AI plan?” Don’t start a conversation that way. That’s the cart before the horse. Like, “What are we doing about AI?” Terrible way to start a conversation.

(Link)

When we talk about the last mile, it’s never going away. Don’t assume that AI is going to close up this gap. The last mile is very long, it’s going to remain very long. And you need a process that allows you to assess things, keep humans in the loop, make sure you’re going to be able to ship. And that process, I hate to say it, it’s probably going to look a lot like the old ones.

The next one is from LangChain’s Interrupt conference. This is squarely in “the builder” category from above and his “hot take” towards the end is actually really important.

(Link)

My last learning–and maybe this is a little bit more forward-looking and a bit of a hot take–but as we’re here talking about agents, I wanted to talk a little bit about data. The take here is the most important data doesn’t exist yet.

Maybe one reductive or simplistic take on AI progress in the last decade has been that we’ve made a ton of progress by just taking more and more publicly available data and creating larger and larger models. That’s of course been very successful–it’s led to the amazingly capable foundation models that we all know…

But I would argue that to build domain-specific agentic workflows for real-world tasks, we actually need more process data–the kind of data that shows you how to get things done inside of those firms today.

(Emphasis mine.) There’s also this nice insight in a single slide:

This is where I pretend to link to my never-written post The Tyranny of AI is not AI itself, it is the lazy, rushed product designers with no imagination or sense of craft for the third time. If I just keep mentioning it, the AIs will hallucinate the entire thing for me.

Now we get philosophical. This one you’re just going to have to search up. It’s the April 27th episode of the Why Theory podcast by Todd McGowan and Ryan Engley. Here’s what it looks like:

[Todd]: What’s interesting is the function and the kind of answer that it gives, because it doesn’t just give a particular answer….If I ask you to criticize my theory of capitalism…you would have a particular criticism. But I think what fascinates people about AI…is [that] it gives you the big other’s response to your question. Even if you feed it a prompt to an exam question, it will give you what the big other would answer to that question. If you ask it a question about your essay, it will respond in the way that the big other responds.

The big other…which is Lacan’s name…for a social authority that captures the average in the society…it’s the authority of the mean, basically. The problem is in our everyday lives, we have to interpret that ourselves…

What’s different is we get an actual statement by the big other when typically we have to do a lot of interpreting to understand what it really wants. And often we’re wrong because it’s never anything exact. I think AI allows us to make it exact…and that changes our relationship to the big other…

[Ryan]: Our friend Rick Boothby…said it’s not that the big other doesn’t exist, it’s that the big other doesn’t know…I think ChatGPT makes this inversion possible where it actually seems like the big other does know but it just doesn’t exist. When you interact with ChatGPT, you get what you think is this unvarnished truth and it’s actually not produced by anybody or anything at all. It’s just free-floating truth we can grab like fruit from a tree–

[Todd]: …the fruit from the tree of knowledge…

I think most of the time–and this is the most noxious dimension of AI…it allows us to imagine a world where the big other isn’t telling us what to do when in fact the big other is always telling us what to do. Yet there is this knowledge that the big other has that we can access, and this is true knowledge not knowledge filtered through the big other.

[Ryan]: It’s reproducing the common in a way. The barometer of the most common, that’s what [AI] is…Part of why it succeeds is because it does seem so instant. You don’t see the work. If you could see where it was pulling things from–if you had the screen divided…where you [could] see the work that it does…[and the information] it’s pulled from–I think suddenly the authority is gone…The instantaneousness of it is what gives it the authority and the fact that it does seem to be divorced from labor.

[Todd]: We’re suspicious of knowledge that involves labor. It’s a crazy thing, but it’s really true. We’re more confident in the notion that somebody just spontaneously has a certain knowledge or intelligence than we are of somebody who’s labored their whole life to have that same level of intelligence…

[Ryan]: [An LLM] completely eliminates the individual aspect of a text or of an idea and [is] this total capitulation to a very ideological common that, because it is so confident looking, seems like the right answer…

AI (particularly via LLMs) appears to be a natural way to establish and maintain ideological and political power, at an order of magnitude more powerful than social media, even w/o manipulating the models.

Ok that was deep. How about this:

(Link)

Or this: Why do people have such dramatically different experiences using AI?

On Monday I was asking Claude for details about a proprietary psych eval thing I guessed might not be in the training data, or if so, incompletely. It answered really well, to my surprise–but I knew enough to know that. Later in the same chat, it went completely off the rails, describing a different eval entirely.

What if I didn’t know anything about it, and that was its first answer? What knowledge or instincts would have helped me know it was hallucinating? Beyond their one-line small-print disclaimers at the bottom of their responses, LLMs are terrible at providing users with additional information or tools to help them evaluate responses. The fact that we threw these models at the public with a naive UX, and the public interprets them as authoritative sources (for reasons explored by Todd and Ryan in that podcast), is problematic, and we’re discovering the consequences…

r/ClaudeAI: Single most important piece of instructions in CLAUDE.md file

The takeaways

Awesome new tech is awesome, people are still horrible, will elect ChatGPT President of the World.

Additional UX shortcomings

I repeat myself, only more so: LLM chat UXs are bad:

Can we educate the public out of this mess?

Probably not, but Anthropic will try.

I tried vibe coding

…for the second time (first try was a total bust), this time with Claude Code…and it was pretty great. But that’s a separate post…

Last second addition

Actually really good overview of AI hype foo vs agentic systems:

(Link)

Previous: Kester Brewin: Countering Political Turmoil with a Real Summer of Love

Archives | Blogroll | RSS