LongCut logo

Chat is overrated. Here are 4 innovative AI product interfaces

By Zara Zhang

Summary

## Key takeaways - **Chatbots are overrated; AI needs a UI revolution.**: Chat interfaces place the burden on users to discover use cases and require significant prior knowledge to prompt effectively. AI needs more intuitive interfaces beyond reactive chatbots. [00:17], [00:39] - **Granola: Human + AI async co-creation for better notes.**: Granola's interface combines user-taken rough notes with AI transcription, creating a richer, more collaborative note-taking experience that keeps humans in the loop. [00:52], [01:34] - **Personalized content remixers cater to diverse learning styles.**: Tools like NotebookLM and HUX transform content into different formats (podcasts, videos) or personalize briefings from existing data, making complex information accessible to various learning preferences. [01:50], [03:15] - **A single button can convey maximum context with minimum effort.**: Products like Snipd and Quill use simple buttons to capture user intent and context (like timestamps), enabling AI to create valuable content snippets with minimal user input. [04:44], [05:32] - **AI-generated feeds offer proactive content discovery.**: Interfaces like Particle News deliver personalized feeds, moving beyond user-initiated prompts to proactively offer content, which is crucial for entertainment and casual consumption. [06:12], [07:08] - **Innovative AI interfaces lower usage barriers for mass adoption.**: Just as GUIs and touchscreens revolutionized computing and mobile internet, new AI interfaces must be intuitive and accessible to bring AI benefits to everyone, not just tech insiders. [08:11], [08:36]

Topics Covered

  • Chatbots are Overrated: AI Needs a UI Revolution.
  • Async Co-creation: The Future of Human-AI Collaboration.
  • AI Should Remix Content for Personalized Learning and Insight.
  • Maximize AI Context with Minimal User Effort through Smart UI.
  • Accessible AI Interfaces Drive Mass Adoption, Not Just Intelligence.

Full Transcript

I believe that chatbots are overrated.

These days, it seems like every other AI

product defaults to a chatbot interface,

probably because Chad GBT is a chatbot,

but I think AI really needs a UI

revolution. Chat is a great interface

for some use cases, but it's not the

right one for all of them. The biggest

problem is that it places the burden of

finding use cases on the user. Its

open-endedness is both its biggest

strength and biggest limitation. It's

infinitely flexible and personalized,

but it can also be extremely confusing

for people who don't know what to ask.

Also, people don't know what they don't

know. Writing good prompt is hard. It

takes a lot of prior knowledge to be

able to ask good questions. And chat

interfaces are reactive. The user must

initiate the interaction and we humans

are lazy. Today, I'll talk about four

types of innovative interfaces I've seen

for AI products and the specific

products that inspire these

observations. The first type is what I

call async co-creation between human and

AI. The primary example is Granola, the

AI meeting notetaker. There are hundreds

of AI meeting notetakers on the market.

What sets Granola apart is the

interface. For most AI note takers, the

AI transcribes everything people said

and then you're given a summary in the

end. But Granola is different. In

Granola, you actually also need to take

notes during the meeting on a notepad

like on the left. But these can be just

rough notes, kind of like the messy

notes to sell that you type on your

Apple Notes. At the same time, the AI is

also transcribing what everyone said.

After the meeting, the AI combines your

rough notes with the transcription,

transforming your notes into a polished,

enriched, and structured version like on

the right. So, this is basically

asynchronous collaboration between you

and the AI. Because you wrote parts of

it yourself, the quality of the notes

feels a lot higher. I really like

Granola's interface because it puts

humans in the loop without compromising

the intuitiveness of the note-taking

experience. The second type of interface

is personalized content engines. The

first interface I saw in this category

was Notebook LM's audio overview and

most recently video overview. You can

upload any material like PDFs or long

YouTube videos and Notebook LM will

transform or remix it into a podcast

with two AI hosts. Their back and forth

conversation feels very natural and

makes it a lot easier to digest

difficult content. The new video

overview feature takes it even further.

It transforms your content into an

explainer video with AI generated

graphics strips and voice over.

>> Self attention lets the model for every

single word it's processing look at the

entire sentence at the same time. And

honestly, the best way to think about it

is like you're at a noisy cocktail

party. Let's say you want to understand

a specific topic. Note that this is not

summarizing content, nor is it

traditional text to speech. This is

remixing content. I like the idea of

content remix because we all learn in

different ways. Some people are visual

learners, some people prefer to listen,

others prefer to read, but only in a

certain style. With AI, we can basically

remix any type of content into any other

type of content. Say research paper to

podcast or long video to short video.

This allows people to consume content

that they otherwise would not consume.

For example, I don't have a technical

background, so I find AI research papers

very intimidating to read, but I can

totally listen to a five-minute

explainer video on the same paper

generated by Notebook LM. The AI

translates technical concepts into

layman's language, and the visuals

further help enhance my understanding.

Another product I'm watching in this

space is HUX, which is a startup founded

by the team that previously worked on

notebook LM at Google. is also centered

around the idea of the personalized

content remix except that in this case

the content comes directly from your

existing personal context like your

emails and your calendar. You can

connect Hawks to your email and calendar

and it will generate a personal audio AI

briefing with the two AI hosts about

stuff in your inbox what you can expect

for your day and also the latest news.

You can listen to it while commuting to

work. It's kind of like having a

personal chief of staff giving you a

briefing to get you ready for your day.

The product is still pretty early stage,

but given the team's track record, I'm

very excited to see where it goes and

just in general very excited about the

idea of hyperpersonalized content for

one. Another product along these lines

is Pillow Talk, an AI powered personal

voice journal app. First of all, it's

just the most beautiful AI app I've ever

used. Secondly, unlike most other AI

companions or wellness apps, it's not a

chatbot. You can yap into a microphone

about whatever is on your mind. And then

you can see an AI generated analysis

about what you just said, which surfaces

insights about your thought patterns and

characteristics. The AI is basically

like a mirror of yourself, helping you

understand yourself by seeing you from

an outsers's perspective. To me, this is

also a type of remix, turning firsterson

content into third person content,

turning the subjective into the

objective. It allows us to see ourselves

in refreshing ways. The third category

for AI interface is what I call a button

that says a thousand words. This

category consists of products that have

thoughtfully designed buttons in GUI

that convey a lot of useful information

coupled with surrounding context. For

example, I use the AI podcast app

Snipped which has a feature called

create snip. When you're listening to a

podcast and want to note something down,

you can press this button and then the

AI will take the current timestamp and

turn the corresponding transcript into a

knowledge card, like an insight or piece

of advice, which you can then save or

share. This is a great interface because

it's extremely low friction, just a

click of a button, yet it conveys a lot

of context to the AI. Like, I'm

currently at 43 minutes and 35 seconds

in this podcast episode. I find this

line particularly interesting. help me

note it down for future reference and

rewrite in a nice format. This is what I

call maximum context with minimum

effort. A similar interface can be found

in a meeting note-taking app quill.

>> This is a really cool button. This

basically takes a highlight and

highlights the thing that was just said

so that later on in the notes it will

draw attention to it and I'll show you

that later.

>> When you're in a meeting and someone

says something you want to note down,

you can just click highlight point and

Quill will note down the current

timestamp. Originally, I just wrote

this, but then what it enhanced was

these things around what when I said

them or when I highlighted them, which

is really, really cool. Isn't that

awesome?

>> After the meeting, the AI will enrich

your notes with what was said at the

specific timestamp. By clicking that

button, you're basically telling the AI

this is worth taking down. The fourth

type of interface is AI generated feeds.

I still believe the feed is one of the

best interfaces invented in the mobile

internet era because it's more

proactive. Let's face it, humans are

lazy. Most people don't want to type. We

just want to sit back and scroll

content. The point is, can AI deliver

personalized feeds that are not brain

rot that actually makes us feel good and

learn stuff. The most established AI

generated feed products I've seen are

mostly in the news category, such as

particle news. The AI remixes content

written by the media into short

articles, and you can also adjust the

tone and style. The content is delivered

in a personalized feed with

recommendation algorithm that you can

also tinker with. I think this interface

is great. However, I'm not using

particle news as much as I like simply

because I don't get my news from the

media anymore. I get my news from ads

and newsletters and other types of

social media. News is still a

productivity use case and we must

remember that most people in the world

want to kill time, not save time. What

I'm looking forward to is seeing AI

generated feeds for entertainment. We're

seeing an inkling of this in Pico and

Character AI's mobile apps. And recently

I saw a cool product called spiel work

that is basically a feed of mini games

that users have vi coded with AI that

you can scroll through and you can also

remix on top of other people's games. In

some here are some things that I think

the above products got right when it

comes to interfaces. One bring humans in

the loop at the right times in a

frictionless way like granola asking you

to take notes yourself. Two get users to

give maximum context to AI with minimum

effort like how you can take notes in

snipped with one click. Three, the best

AI products might make the AI invisible

or behind the scene. Like in Particle

News, the user only needs to interact

with the fruit of the AI's labor, not

the AI itself. Focus on the user's job

to be done. AI is just the enabling

technology that gets the job done. It's

a means to an end. Four, have the AI

come to the user, not the other way

around. Like in HU, most of the time,

you don't need to give any prompt. The

AI just comes to you with readily made

content generated from your personal

context. If we look back in history, the

personal computer revolution itself was

driven largely by interface innovations

like the graphical user interface and

the mouse. These replaced command line

interfaces and made computers accessible

to regular folks, not just scientists

and programmers. Similarly, the mobile

internet revolution was largely driven

by the touchscreen which was also a

interface innovation. Twothirds of

Americans still haven't used ChadBt. If

we really want the benefit of AI to

reach more people outside of the tech

bubble, we need more intuitive

interfaces that are truly accessible to

everyone. While we think about raising

the ceiling of intelligence, we should

also think about how to lower the

barrier for usage with innovative user

interfaces.

Loading...

Loading video analysis...