Slightly tangential, but this is the reason I'm baffled why people think that AI-driven podcasts would ever be worth listening to.
If you can find an LLM+TTS generated podcast with even a FRACTION of the infectious energy as something like Good Job Brain, Radio Lab, or Ask Me Another, then I'll eat my hat. I don't even own one. I'll drive to the nearest hat store, purchase the tallest stovetop hat that I can afford and eat it.
I totally agree, but I also think it's just not geared towards us because of our age and preexisting beliefs of "what podcasts/news/videos are supposed to be". Think of kids around age <12. If they'll keep consuming AI-driven media, they'll take that as normal and won't blink twice if that becomes the de-facto standard in about 10 years.
It's the same as short-video format for me. Sure, I can watch some TikToks from time to time, but making them, or continuously getting my news from them? Yeah, that's not gonna fly for me. However, for my niblings (age 8-17) that's basically where they're getting all the "current affairs" from. Microtransaction is probably one of the easiest example as well. 15 years ago, anyone who bought games would laugh at you if you said every game that you paid $80 for would also have endless amount of small items that you can buy for real money. Right now? Well, kids that grew up with Fortnite and Roblox just think that's the norm.
The younger generation loves short-form, to-the-point stuff. Which is the exact opposite of what the current crop of GenAI makes. In a tiktok video, every sentence, every word counts because there's a time limit. If people don't engage with your content in the first 3 seconds, it's worthless. The video linked in another post starts off with 15 seconds of complete fluff. You'd have better engagement if you have a guy opening with BIIIG NEWS!! LABOR DAY BREAKFAST GOES HAYWIRE!!! and hook people.
GenAI is great at generating "stuff", but what makes good content isn't the quantity. What makes good content is when there's nothing left to take away.
What's funny is every GenAI "incredible email/essay" would be better communicated with the prompt used to generate it.
I’m only half joking when I’ve described ChatGPT-authored emails as a uniquely inefficient transport format.
Author feeds bullet points into ChatGPT which burns CPU cycles producing paragraphs of fluff. Recipient feeds paragraphs of fluff into ChatGPT and asks it to summarise into bullet points.
> Author feeds bullet points into ChatGPT which burns CPU cycles producing paragraphs of fluff.
GOOG, AMZN, and MSFT reportedly need to use nuclear energy to power the LLM farms that we are told we must have.
One must ask who (or what) in this feedback loop of inanity is doing the most hallucinating.
[1] _ https://apnews.com/article/climate-data-centers-amazon-googl...
[2] _ https://www.npr.org/2024/09/20/nx-s1-5120581/three-mile-isla...
Then users turn around and feed the fluff into energy hungry summarizers because who has time for a 5 paragraph email that could’ve been a three point bulleted list?
It would be a net win if it could normalize sending prompts instead of normal communication, which is not far in terms of useless waste of energy and space to LLM output that emulates it.
Similarily, MSFT recently announced the upcoming ability to clone your voice for Teams meetings. Extrapolating, in a few months, there will be Teams meetings which are only frequented by avatars. At the end of the meeting, you get an email with the essential content. Weird times ahead.
I think you are on to something there. I've heard executives talk about their current AI flow and it all sounds like summarization.
There's an increasing amount of prose written that will only ever be read by LLMs.
"Your essay must be at least X words" has always been an impediment to truly good writing skills, but now it's just worthless.
The way I explained it when I taught English 101 to first-year university students: any substantive question can generate an answer of a paragraph or a life's work; in this assignment I expect you to go into this much depth. Of course, good expository writing is as to-the-point as possible, so the first hurdle for most students was eliminating the juvenile trick of padding out their prose with waffle to meet an arbitrary word-count. Giving a word-count to an AI seems (currently) to activate the same behavior. I've not yet seen an AI text that's better writing than a college freshman could be expected to produce.
> Of course, good expository writing is as to-the-point as possible, so the first hurdle for most students was eliminating the juvenile trick of padding out their prose with waffle to meet an arbitrary word-count.
This is the most beautiful sentence I’ve read today.
Problem is, I doubt many people overcome this hurdle - that "juvenile trick" is pretty much the defining quality of articles and books we consume.
Indeed. We need better English teachers. :-)
We won't get them unless we appreciate both teaching and the Humanities more than we do. I was good, but by no means the best (75th percentile, maybe?). I loved doing it, but changed careers to IT because I'd never have been able to support a family on what I was paid.
A culture which pays teachers poorly, treats them with disrespect ("those who can do..."), and relentlessly emphasizes STEM, STEM, STEM is one that's slowly killing itself, no matter how much shiny tech at inflated valuations it turns out.
I don't know how it is elsewhere, but where I grew up we had minimum word limits on pretty much all essays. Doesn't matter if you can say what you want to say in 6 sentences, you need 4000 words or 2 pages or whatever metric they use
Oh, of course. Length requirements are important, for the reasons I explained up-thread. However, if teachers accept any old thing ("padding") to reach the count, then that metric is arbitrary, which (justifiably!) makes students cynical.
If a student can say all that they want to say in six sentences then they need to learn more about the topic, and / or think through their opinions more deeply. Teachers who do not take that next step are bad teachers, because they are neither modeling nor encouraging critical thinking.
In some places the majority of teachers are themselves incapable of critical thinking, because those who are leave the profession (or the locale) for the reasons in my comment above.
[Edit to add]: Please note that I say bad teachers, not bad people. Same goes for students / citizens, as well. The ability to think critically is not a determinate of moral worth, and in some ways and some cases might be anti-correlated.
Wish my high school English teachers had taught that. I remember fluffing essays to get to a minimum
College admissions essays on the other hand had the opposite problem - answering a big question in 500 words. Each sentence needed to pack a punch.
Don't get me started on college admissions essays. Rich kids pay other people to write them. Poor kids don't understand the class-markers they're expected to include. If AI consigns them to the dustbin of history it might be the first unalloyed good that tech ever does.
Clarification: "that tech" meaning the direct antecedent: "[AI] tech", not tech in general. I'm a Humanities guy, not an idiot. :-)
One of my favorite Mark Twain quotes comes from one of his correspondences: 'My apologies for such a long letter, I hadn't the time to write a short one.'
Probably not by Twain, but still a good quote. https://quoteinvestigator.com/2012/04/28/shorter-letter/
I never had that requirement outside the first years of school- where it’s more about writing practice than writing actual essays. After it was always “must be below X pages”
X words is supposed to be a proxy for do enough research that you have something to say with depth. A history of the world in 15 minutes better cover enough ground to be worth 15 minutes - as opposed to 1 minute and then filler words. Of course filler is something everyone who writes such a thing and comes up a few words short does - but you are supposed to go find something more to say.
I eventually flipped from moaning about word count minimums to whining about conference page limits but it took a long, long time- well into grad school. The change came when I finally had something to say.
When I was a lazy kid, and there was a requirement to fill a number of lines/pages, I'd just write with bigger letters.
Write a comment explaining that the ostensibly simple task of writing a dozen or so thank you letters for those socks/etc you received for Christmas can, for some people, be an excruciating task that takes weeks to complete, but with the aid of LLMs can easily be done in an hour.
On second thought, you're right. That was easy.
That's old world stuff - I've never sent a thank you card and have lost no sleep bc of it.
I’d prefer to receive no thank you than to receive an AI written one. One says you don’t care, the other says you don’t care but also want to deceive me.
There's the third case: they care. In which you wouldn't be able to tell whether the card is "genuine" or AI written; the two things aren't even meaningfully different in this scenario.
You can tell if the thank-you card is hand-written. Most people don't have a pen plotter connected to their AI text generator to write thank-you notes.
Emailed or texted "thank you" notes don't count. At all.
Yes, but they could also generate the text and transcribe it onto paper by hand.
For many people, myself included, 90%+ work on things like thank-you notes, greetings, invitations, or some types of e-mails, is in coming up with the right words and phrases. LLMs are a great help here, particularly with breaking through the "blank page syndrome".
It's not that different from looking up examples of holiday greetings on-line, or in a book. And the way I feel about it, if an LLM manages to pick just the right words, it's fine to use them - after all, I'm still providing the input, directing the process, and making a choice which parts of output to use and how. Such text is still "mine", even if 90% of it came from GPT-4.
I guess if someone went through the effort to prompt an LLM for a thank-you card note, and then transcribed that by hand to a card and mailed it, that would count. It's somehow more about knowing that they are making some actual effort to send a personalized thank you than it is about who wrote it.
But honestly I don't think "blank page syndrome" is very common for a thank-you card. We're talking about a few sentences expressing appreciation. You don't really have to over-think it. People who don't send thank-yous are mostly just being lazy.
My financial advisor sends out Christmas cards and Birthday cards. They are pre-printed stock cards. I don't even open them. I should tell him not to waste the money. If he even wrote just one sentence that expressed some personal interest, then they would mean something.
These kinds of messages are on the one hand just pro-forma courtesies, but on the other hand they require that that some personal effort is invested, or else they are meaningless.
Sure, it's not just thank you cards though. I once had a job in which my boss assigned me the weekly task of manually emailing an automatically generated report to his boss, and insisted that each email have unique body text amounting to "here's that report you asked for" but stretched into three or four sentences, custom written each week and never repeating. The guy apparently hated to receive automated emails and would supposedly be offended if I copy-pasted the same email every time.
Absolutely senseless work, perfect job for an LLM.
There's something painfully ironic and disturbing that the pseudo-Kolmogorov complexity of clickbait content, as judged "identical in quality" by an average human viewer, is arguably less than the length of the clickbait headline itself, and perhaps even less than the embedding vector of said headline!
It's always been this way, it's just rules of polite/corporate culture don't allow to say what you actually mean - you have to hit the style points and flatter the right people the right way, and otherwise pad the text with noise.
If the spread of AI would make it OK to send prompts instead of generated output, all it would do is to finally allow communicating without all the bullshit.
Related, a paradox of PowerPoint: it may suck as communication tool, but at the same time, most communication would be better off if done in bullet points.
Isn't it just a matter of time until AI gets trained to generate attention-grabbing videos? Also, "the first three seconds" isn't exactly the case anymore. There's a push for algorithm to favour videos that are longer than 1+ minutes. Which, to my understanding, is TikTok's way of fighting for YouTube's userbase.
The videos are longer in total length, but you've never seen the average TikTok/Insta user if you think people are letting videos play for more than a few seconds before scrolling onto the next one. This is why movie trailer videos now have a "trailer for the trailer" in the opening seconds with like "THE TRAILER FOR SONIC 3... STARTS NOW" with all of the most attention-grabbing scenes frontloaded.
Similar on IG, which is why a lot of the photographers on my feed will post five or six images, "Which is your favorite, 1-6? Comment below!" because they get the engagement, synthetic or otherwise, of you clicking through each image, and then commenting.
Youtubers also adopted this trick. The best pattern currently is to voice a random question in the middle of the video or present themselves as "I'm not sure about this, please correct me in the comments" as transparent plan to entice people to comment. I even bought it a first few dozen times, but now when every single creator does it, it is kinda tiring.
I am curious to see this pattern in action. Can you share some @-usernames for these photographers?
> Isn't it just a matter of time until AI gets trained to...
blah blah yes "it's a matter or time" for every one of the myriad shortcomings of the technology to be resolved. If you're a true believer everything is "a matter of time". I'll believe it when I see it.
Depends on the technology. It’s hard to look at the progress from December of 2022 till today, and think we won’t go further. Image generation is getting better every day. Parts of the video generation pipelines are also advancing.
the counterpoint is since then we've killed all the easy ways to scale. the datasets can't get bigger because it's already the whole Internet. model sizes can't grow that much because you start running into RAM constraints. efficiency definitely can be improved, but probably not more than 100x on our current architectures.
I'm also pretty aure you'll see it eventually... Consider the possibility that this is both a bubble waiting to pop, as well as the stuff that will shape the future. Kind of like the Internet around the year 2000.
I don't think this is entirely true. Have you seen these "Internet Story told by TTS Voice-over a Minecraft parkour Video" ( which are what my niblings watching ) ? I noticed a lot of the story is dragged on for over minutes. These are the stories that I read in text in about 5 second.
Short form videos are often hyper focus and to the point, but there are a lot of vertical video contents that are just ( to use the GenZ term ) brain rot like these as well.
Is that true of tiktoks in general? I feel like a lot of the short form videos out there purposefully bait the watcher and drag on for 3/4 of their runtime.
> The younger generation loves short-form, to-the-point stuff.
Meanwhile, Joe Rogan et al. have 3+ hour long podcasts…
Short form video content is absolutely terrible, every time I watch one I have an extremely strong urge to verify whatever I heard in the video, so I just avoid watching them altogether. I’m not sure how anything of value can be 45 seconds long with absolutely no context, outside of comedy.
Counterpoint, GenAI is great at copying styles and typically works best with shorter content.
For example, I could very easily see GenAI being able to produce 1 million TikTok dance challenges.
Which will make them completely worthless by dilution and not stand out. Oops.
Novelty grabs people's attention. A system based on the statistical analysis of past content won't do novelty. This seems like a very basic issue to me.
Novelty itself is easy, the hard part is the kind of novelty that is familiar enough to be engaging while also unusual enough to attract all the people bored by the mainstream.
Worse, as people attempt to automate novelty, they will be (and have been) repeatedly thwarted by the fact that the implicit patterns of the automation system themselves become patterns to be learned and recognised… which is why all modern popular music sounds so similar that this video got made 14 years ago: https://www.youtube.com/watch?v=5pidokakU4I
(This is already a thing with GenAI images made by people who just prompt-and-go, though artists using it as a tool can easily do much better).
But go too soon, be too novel, and you're in something like uncanny valley: When Saint-Saëns' Danse macabre was first performed, it was poorly received by violating then-current expectations, now it's considered a masterpiece.
A system that digs out undiscovered mechanisms to drape novelty on based on where the statistical analysis says it’s already been, that would do it though.
We can sit and imagine horrors the whole week. It has no bearing on the capabilities of machine learning.
If there’s a clear distinction between LLMs looking for patterns in text and LLMs looking for patterns in patterns in text, I’m interested in seeing it better and understanding it.
I ... don't think there's a time limit on TikToks unless you mean that 60 minutes is a time limit. Are you thinking of Vine?
If you want people to watch your content there's definitely a time limit. I don't mean anything imposed by the platform; I just mean if people aren't interested within 3 seconds, they're scrolling to the next video in their feed.
That's a pretty new thing though, in 2020 it was 1 min I believe, and most people skipped after 15-30s
Then it got increased to 3 min and now 10/60 min
Seconds - that's all the time I give something. If it looks draggy I will skip ahead to like halfway thru the video and if the person isn't in the middle of explaining something that they have clearly spent half the time extrapolating, that's it. They get no more time.
I often turn on subtitles and watch at 2x speed.
I prefer the transcript 100% of the time to video.
When I was a kid, I loved having breakfast made for me by other people as I couldnt cook. As a small child, one would say I expected it. It doesn't mean when I become an adult that would continue to be the case.
> If people don't engage with your content in the first 3 seconds, it's worthless.
Is that based on a vigorous experience as a content creator on TikTok as a long form content creator or are you going off what you've heard about TikTok, or it's what your feed is full of? (which says more about you/your feed than it does about TikTok)
https://www.tiktok.com/t/ZTYheCgBq/ 1.1 M views
https://www.tiktok.com/t/ZTYhephfK/ 1.2 M views
or for some more niche stuff which isn't "BIIIG NEWS!! LABOR DAY BREAKFAST GOES HAYWIRE!!!" level of intro:
https://www.tiktok.com/t/ZTYheTVsh/ 54.6 k views
https://www.tiktok.com/t/ZTYheE9m2/
36.3 k views
https://www.tiktok.com/t/ZTYheoK2G/
416 k views
This voice-over definitely isn't going for attention grabbing
https://www.tiktok.com/t/ZTYhe7qTT/ 16 k views
https://www.tiktok.com/t/ZTYhdmJxq/ 64 k views
okay finally found something that's 5 mins long https://www.tiktok.com/t/ZTYhdAfNj/ 270 k views
All of those immediately have something attention-grabbing within the first few seconds: picture of Mars' surface, run map showing funny human shape, text with "Gen Z programmers are crazy" prefacing the anecdote, going immediately into the IT-related rap, guy holding a big and cool-looking stick, "dealing with your 10x coworker" immediately showing the point of the video. The only one that doesn't is the SQL one I guess but that's a very low view count (relatively) on a niche channel.
So thanks for providing a bunch of examples that prove my point, I guess?
none of them are breathless "BIIIG NEWS!! LABOR DAY BREAKFAST GOES HAYWIRE!!!" attention demanding within the first three seconds imo, but, sure, whatever, you're totally right
I've never used tiktok and this post was... enlightening. The ClickUP HR guys are actually pretty funny... but wtf is this from that first channel you posted... lol? https://www.tiktok.com/@luckypassengers/video/74395775119539... - I mean, she's not wrong.
The younger generation has had their brain's development hacked by corporate advertising and can't help themselves but prefer the content they have been all but brainwashed to prefer - this was done deliberately and intentionally and obviously to much detriment to the young of today.
Nobody cares tho bc "they can't pay attention" like it's somehow an entire generation's individual faults they are like that...
Clearly a societal failure there.
The idea that "TikTok is for Gen Z" seems like a very stale meme, although I only have anecdata to back that up.
Microtransactions are way older than 15 years. Wizards of the Coast was selling randomized MtG booster packs in 1993. I'm guessing that the earliest loot boxes for kids were baseball card packs, with very similar psychological purpose to today's game cosmetic collectibles.
> I'm guessing that the earliest loot boxes for kids were baseball card packs
Baseball card packs are an innovation. If you read Peanuts, you'll see that they're referred to as "bubble gum cards", because cards were included as promotional items in packs of gum, a "free toy inside" that was compatible with the size and shape of bubble gum. They moved to dedicated packs of cards when people started buying gum to get the cards.
I totally agree, but it’s just easier for people to accept something if they grew up with it. Sure there will always be people from older generations dabbling with new stuff. But quite a few people refuse to change their behaviour as the age. I wrote them as examples, because it is the biggest contrast I can see in online behaviour between myself and my nephews/nieces plus their circles.
I think the main difference is digital vs physical goods. I know it's minor since a card is just a cheap piece of cardboard, but it's still something tangible (unless the game cosmetics also include a physical item, in which case...I'm dumb).
When WotC first offered digital cards, they had a system whereby you could redeem a complete set for its physical equivalent. At the time this helped convince naysayers they weren’t giving money for nothing. Twenty years later, with their latest online offerings, nobody really cares. I’ve certainly spent more money and had more fun with online Magic, and when I wanted to build my own cube I just got proxies printed anyway. I think the “collectible” part fell away a long time ago and now you’re happy just paying for an experience.
If something isn't digital today it isn't as real.
A physical photo in my hand is sooo limited compared to the same photo in my hand via the screen on my smartphone. Same with a DVD or CD or Disc - all antiquated tech.
"If they'll keep consuming AI-driven media, they'll take that as normal and won't blink twice if that becomes the de-facto standard in about 10 years."
There is a sense in which that is true.
However, we all develop taste, and in a hypothetical world where current AI ends up being the limit for another 10 or 20 years, eventually a lot of people would figure out that there's not as much "there" there as they supposed.
The wild card is that we probably don't live in that world, and it's difficult to guess how good AI is going to get.
Even now, the voice of AI that people are complaining about is just the current default voice, which will probably eventually be looked on about as favorably as bell-bottomed jeans or beehive hairdos. It is driven less by the technology itself than a complicated set of desires around not wanting to give the media nifty soundbites about how mean (or politically incorrect) AI can be, and not wanting to be sued. It's minimal prompt engineering even now to change it a lot. "Make a snappy TikTok video about whatever" is not something the tech is going to struggle with. In fact given the general poverty of the state space I would guess it'll outcompete humans pretty quickly.
"eventually a lot of people would figure out that there's not as much 'there' there as they supposed."
Let's be honest: most of us here know there is more 'there' on the myriad university-press books available free on Anna's Archive, than on HN. The reason we still hang out here is desire for socializing, laziness, or pathological doomscrolling; information density doesn't really factor into our choices.
Personally I don’t think it has anything to do with normalcy.
I don’t consume AI media because it’s not very good.
I watched a lot a bad movies and read a lot of bad books as a kid that I can’t stomach now because I’ve read better books and watched better movies. My guess is that kids today would do the same, assuming AI doesn’t improve.
> they'll take that as normal and won't blink twice if that becomes the de-facto standard in about 10 years.
This is in part my worry, but the other part is that there just won't be much other of an option. I can't be the only one that feels like a lot of information is just... shit. Yeah, there's still YouTube channels I find that are brilliant but they definitely are not fitting to the algorithm and it's clear they're "punished" for this. I think we can say the same about blogs and other places. It is just getting much harder to actually discover the gems.The result of this is that I watch more and read garbage and have less joy. Maybe I'm just addicted to the routine, but I definitely consume less. I just think it is the problem when you try to make something for everyone; you just end up making things with no substance, no soul. As they say, diversity is the spice of life, and good god, could I go for some actual flavor.
Slightly tangential but TikTok isn't just short-form stuff. It also has normal length content, which is where most current-affairs/analyses would fall, outside of clips.
Similarly (somehow?) for Fortnite, 100% of their microtransactions are cosmetic. It's also free, so it's be akin to telling people you could get AAA quality games for free. The $60+++ group of games make up a vanishingly small chunk of the total market, and are more a relic of 'our' generation. Roblox is a platform, more than a game, so it's its own beast. In general though I think younger gamers have become more demanding, in a good way.
I regularly watch 10 minute or longer TikToks and I'm sure you've turned off a YouTube video in the first 10 seconds.
Blipverts are probably the future and not just for ads but for content as well
On the other hand, I'm getting some of my news from shorter (5 - 10 min these days, though they used to be shorter) videos with talking maps and war machines...and I'm very much not young.
Buff/Franklin for 2028!
Fortnite and roblox are f2p though.
Honestly sometimes i doubt there's a more damagingly, intentionally overloaded term in modern lexicon than "free to play". Many, many things claim that title while placing if not game play, then socially necessary events, items, costumes, locations, quests, game modes, etc behind a paywall. We need a different term to describe the incredibly predatory behavior and psychology behind most of the current space of "free to play" games. (I'm not saying your favorite game is a lie, just that many games claiming this title effectively are)
> anyone who bought games would laugh at you if you said every game that you paid $80 for would also have endless amount of small items that you can buy for real money.
> kids that grew up with Fortnite and Roblox just think that's the norm.
How so? Fortnite and Roblox don’t cost $80, they are free to play with in game purchases.
They’re F2P, but the social pressure from your friends to buy the next skin, mini game and etc. completely normalizes the behaviour as you grow up. Then you take it as a “usual thing” and don’t bat an eye when a game you buy also has different skins in the game.
At least that was my perception when I played Call of Duty with my younger family members.
Go back to the 1990s (even probably the 2000s?) and ask literally anyone what they think of the idea of people spending huge quantities of their time on this mortal plane watching videos of other people talking about how they do their make up and pick their outfits.
Are those videos "worth watching"? Are videos of people playing video games "worth watching"? Are videos of people opening products and saying out loud the information written on the box and also easily accessible on the public internet "worth watching"?
I'm not happy about these developments, but that isn't a factor of much concern to the people driving and following these trends, it turns out.
It's been a long time since "quality" and "success" have been decoupled. Which is to say - I hope you like the taste of hats.
Bad TV is a lot older than that. It probably seems weird nowadays to watch “I Dream of Jeanie” and “Gilligan’s Island” reruns because that’s what was on TV. Or how about game shows and soap operas? Daytime TV was the worst. But people watched.
Every now and then I play the Gilligan's Island theme song because it's so catchy. Still no idea what the show is about
Ahh, it's about shipwrecked people that somehow don't endlessly screw each other and still adhere to all the same social norms when they are leaving the idea as they did when they arrived. I think they once made a phone out of a coconut - I watched it before school as a kid, great show! Unbelievably innocent and naive.
I'm pretty sure humans have been finding ways to unproductively waste time for millenium...
I'm not sure how watching lightweight videos on subjects you're curious about is any worse than how people wasted time in the past?
Personally I waste time watching videos on outdoor gear, coffee making equipment and PC hardware. I certainy don't regret it because I had no plans to do anything productive with that time either.
videos of other people talking about how they do their make up and pick their outfits.
Doesn't seem that much different from a fashion magazine interview about what X celebrity likes to wear. Those have been around for quite a long time.
At least in the past people were celebrities for a reason other than the number of followers they had on social. It'd be nice if we could return to a time when people were part of the public discourse because they were good at something (or their parents were rich, sadly).
That’s not true at all. Fame for no reason or dumb reasons is hardly a 21st century phenomenon.
No dog in this fight, I don't know what exactly you're doing, but I'd cautiously point out that it is, in fact, novel to the internet era to watch rando microcelebrities doing makeup step by step, no matter how long we delay acknowledging that with microquibbles.
I could throw in an example of how I'll watch boring videos of a couple playing with their birds for 90 minutes on Youtube. You can link me to the Wikipedia page on slow TV (via Norway), and it won't erase the simple, boring, straightforward, fact that it is a phenomenon.
I didn’t interpret the original comment to be about micro celebrities, but about people supposedly wasting time today in ways they didn’t beforehand. I agree that micro celebrities are a new phenomenon somewhat (although they are also sort of a return to more regional distribution of fame.) But that wasn’t the point being made.
The scale isn't even close, and it's become normalized now. I think it bears talking about as a 21st century phenomenon.
> At least in the past people were celebrities for a reason other than the number of followers they had on social.
“Other than”? Obviously, as social media didn't exist. “Better than” or “more relevant to the things for which there celebrity status was used to direct attention”? Not particularly.
But you ignored why they have so many followers in the first place. It's because they're entertaining to watch/listen to, which for someone in the entertainment business it's as on the nose as you can possibly get. I think absurd to say they're not good at something because I for sure don't have what it takes to record/edit that many videos all day and be that charismatic all the time.
Why does it matter if the celebrity was good at something or not if you are just discussing what they wear?
Ahh, you e touched on a big problem with our society rn - nothing is worth doing.
I've been really thinking about lawns lately and how much time men and women spend maintaining them, how much pride many of them have in such activity... lawns aren't real tho, it's just a personal park nobody ever uses that we all thought we wanted bc rich ppl had them. Front lawns especially, just for looks, nobody normal ever sits on them even.
Case in point is all the people that live in an apartment - they don't do lawns. They might think they want to and some might even enjoy up keeping a lawn but it's not an activity that's "worth it" in fact there are many reasons not have a lawn, it isn't an activity that justifies itself as so many pretend.
Everything is like that. Almost nothing any of us do adds to humanities' general progress or improves our own situations even.
Mowing a lawn and watching TV are incredibly similar activities if you have a nice lawnmower.
I actually think there are mental benefits to cutting the lawn. There is something zen about slowly reducing the grass in an ordered manner.
I've had very similar thoughts about lawn-cutting, but hadn't related it to technology-related trends!
I couldn't agree more. In a broad sense, it's like we've lost contact with our own lives, communities, and cultures. I don't think those things are "dead", but they mean things now which are totally different than what they meant even only a couple of decades ago, and the people living through it sort of know that and kind of say it out loud sometimes even, but can't really wrap their heads around it at the same time and continue to ape the old behaviours (no offence, anyone, I do it too).
"Traditions" live on in a sort of zombie state... we cut the grass, we present the gift cards. It's an odd and fun(ny) moment to be alive.
I'm not disputing that people waste inordinate amounts of time running out the clock before they shuffle off this mortal coil. (cough every X demograph reacts to Y video cough).
Agreement to eat said head apparel is predicated upon "infectious energy" (i.e. quality) - NOT success. I'll draft up a more officious document later.
Note: I am the sole arbiter of what constitutes quality.
> It's been a long time since "quality" and "success" have been decoupled.
I think so, too.
I guess quality was a property of interest in the old days, because the path e.g. for commercial music was: Maximize profit -> Maximize sales -> Maximize what the target audience likes -> Maximize quality.
For TikTok etc. they bypass the market sales stuff and replace it by an 'algorithm', that optimizes for retention, which is tightly coupled to ad revenue. I imagine the algorithm as a function of many arguments.
Just relying on quality is an inefficient approximation in contrast to that.
The Art of Zen and Motorcycle Repair will teach anyone all they ever need to know about quality.
We used to have some "radio station" in shoutcast during the 90s. It was very local and geeky, but had a couple of listeners. It was kind of a podcast
Good point. I absolutely abhorred reaction videos until I found a channel that works for me. A guy who reacts to videos about ultra luxury mansions. I find it interesting hearing an expert in the field point out things that most of us wouldn't think of. Plus he's got a great sense of humor.
What blew my mind was that there would be a market to watch other people play video games. You never know what catches on.
> Go back to the 1990s (even probably the 2000s?) and ask literally anyone what they think of the idea of people spending huge quantities of their time on this mortal plane watching videos of other people talking about how they do their make up and pick their outfits.
So like supermarket magazines in video form? I think people would get it.
To me, unboxing videos are documentation, not entertainment. When I need to know exactly what's in a box and how it's packaged, an unboxing video is the only source of that information.
What's in the box is almost always on the box, and how it's packaged seems to be very useless information.
As someone who's struggled to really get into podcasts, I'm convinced that most people who enjoy podcasts don't really actively listen to them, they just like that extra bit of noise in the background while they do something else, with the added bonus that they might just listen at just the right time to pick up some interesting factoid.
Don't get me wrong, I like podcasts, Fall of Civilizations being a personal favorite of mine, it's just that my desire to try and actively listen to them requires that I carve out an hour (or much longer in the case of FoC, worth it) of my free-time to eliminate any distractions and focus. Pretty hard to do when there are so many other things I could be doing with my time besides sitting there trying my best to listen.
Anyway, I bring it up because I'm convinced that the people promoting AI-podcasts are mostly made up of the aforementioned people who just listen to them for noise.
I used to listen to podcasts during my commute, when they took the place of listening to NPR or talk radio. Now that I work from home full time, I listen to them when I'm doing dishes or similar chores, going on walks around the city, and when I'm doing any significant amount of driving. I listen to a combination of comedy/history (love The Dollop), politics, and sci-fi commentary.
For the most part, I'm listening pretty actively, but if I'm just sitting there listening without a fairly mindless activity going on, I'll get distracted pretty quickly and find myself looking at my phone...
I don't get them either since the hosts are typically "regular people" talking about often complicated subjects that they are by no means domain experts in.
I find the format of "Dumdum Host A read some articles about something last night and Dumdum Host B asks questions about it" especially grating, unless it's purely opinion-driven and then I still probably won't care unless I read about the hosts and find out they are probably people with opinions worth listening to.
I'd rather read a book or be with my own thoughts without having them be even more crowded by some randos telling me stuff they think they know
Oh yeah, i definitely hate it when uninformed people try to make podcasts on specialized topics they obviously know nothing about. I went searching for niche physics podcasts (anything besides space stuff please!) And the amount of literal high schoolers and undergraduates attempting to explain things they're years away from even seeing in class is painful, most admit that they know nothing and are literally free-associating over a Wikipedia article. I give them a pass though since they have no listeners and are usually young but...
I exclusively listen to domain expert podcasts. History experts, policy/economics experts, aforementioned physics experts. Literally event i listen to is hosted by at least one person who is a domain expert. They can branch out and interview experts in other domains but there's always an expert. I honestly can't say (besides the high schoolers which i searched out) the last time i've even test listened to a podcast hosted by semi-charismatic article readers.
I'd recommend you search reddit for lists of podcasts in your desired area. App recommendations are all shit and there's many posts out there
> I don't get them either since the hosts are typically "regular people" talking about often complicated subjects that they are by no means domain experts in.
Huh? Most podcasts I listen to feature either experts in the field in question or very skilled journalists compiling the opinions of experts.
> Don't get me wrong, I like podcasts, Fall of Civilizations being a personal favorite of mine, it's just that my desire to try and actively listen to them requires that I carve out an hour (or much longer in the case of FoC, worth it) of my free-time to eliminate any distractions and focus.
I don't see why that should be true. Podcasts are great when coupled with mindless work like mowing the lawn, weeding the garden, stacking firewood, walking to the shop, driving etc. You can get virtually 100% out of a podcast while performing tasks like this.
I figured people listened to them while commuting. It seemed the best fit: lots of time, mostly little attention needed (except when you alight or are driving in adverse conditions).
I don't understand enjoying podcasts while just using them for background noise. Maybe because I actually listen, and most podcasts I listen to seemed to have an engaged audience. But then I like listening to people talk about philosophy, politics, sports, true crime, science, history.
I agree. I think I listen to podcasts just for a bit of "social noise" when I'm doing something by myself and for occasionally picking up on something new I hadn't heard of before. For pure information content, I think they're actually very poor. It's not unlike listening to an old AM radio talk show. Hosts repeat themselves, engage in banter, and often oversimplify topics for the sake of a narrative.
I also think AI podcasts could become popular one day for people who just want some background noise and bits of trivia every once in a while. I would argue that a lot of YouTube channels I sometimes have in the background just summarize Wikipedia articles and don't have much of a personal touch anyway, so an AI could do the same thing.
I switched to improvised pods for this reason. Podcasts are for doing the dishes and mowing the lawn and playing factorio. I don't turn on anything with 'meat' unless I have a long car trip ahead of me. I just am not able to sit still and only listen without feeling like I could be doing something else, and when I'm doing something else I'm gonna start tuning out eventually.
Hey Riddle Riddle, Hello from the magic tavern, Artists on artists on artists on artists are my picks for now.
There is nothing you are supposed to be doing with free time, hence free time.
Tbh tho, there is nothing that we are supposed to be doing in this world rn anyways, nobody was born to work any of these jobs, we didn't evolve to flip burgers.
This is the old world, don't place too much stake in it, everything will change and then what?
If you want to listen to a podcast, do it. If you want to do something else, do that.
Don't do what you don't want to do, that's all there is too worry about.
This is partially true for me, but when I listen to podcasts it’s generally in what would otherwise be dead time for me such as when grocery shopping or exercising.
Having this strong of an opinion on something you admittedly don't understand (the appeal of podcasts) is not a great choice. Maybe ask people instead of making up stuff.
I have very strong opinions on lots of things I can't understand why people like doing, I don't need to understand someone's satisfaction before I determine how I feel about what they are driving satisfaction from. That's ridiculous.
I think people can't be alone with their own minds so they fill every second of there quiet time with someone else's thoughts rather than their own.
> I'm convinced that most people who enjoy podcasts don't really actively listen to them,
Or you know not everyone is the same. Just because you struggle with something doesn’t mean that it is not easy and effortless for others.
I can’t dribble a basketball and walk at the same time, still I won’t make the claim that everyone who claims to enjoy playing basketball is somehow fudging it.
I have probably written this comment about a dozen times on HN already, but: I agree completely, because people don't listen to podcasts purely for information, they listen for information plus community, personality, or just a basic human connection.
If you're a content creator today, the best thing you can do to "AI-proof" your work is to inject your personality into it as much as possible. Preferably your physical personality, on video. The future of human content is being as human as possible. AI isn't going to replace that in our lifetimes, if ever.
I just took a poll about art that had 50 examples and you had to choose which were human and which were not - I did very well, bc I'm awesome but apparently it shook the art world to its core, many of the professionals have no idea how to tell the difference/can't tell the difference.
The elements that identified a painting as human to me had nothing to do with appearance but everything to do with feeling.
Be passionate about what you present - if you can't be, don't bother. Passion is what will distinguish us from the AI - stuff like personality is almost immediately replicated, vocal inflections, funny comments, unusual spoken delivery - that's all within the realm of AI capability. Talking about micro blogging like it's God's gift to the world, not so much an AI thing, it won't feel right with AI.
> The future of human content is being as human as possible. AI isn't going to replace that in our lifetimes, if ever.
I have been working on machine learning algorithms for a long time. Since the time when telling someone I worked on AI was a conversation killer, even with technical people.
AI's are going to understand people better than people understand people. That could be in five years, maybe - many things are happening faster than expected. Or in 15 years. But that might be the outside range.
There is something about human psychology where the faster something changes, the less we are aware of the rate of change. We don't review the steps and their increasing rate that happened before we cared about that tech. We just accept the new thing that finally gets our attention like it was a one off event, instead of an accelerating compounding flood, and imagine it isn't really going to change much soon.
--
I know this isn't a popular view.
But what is happening is on the order of the transition to the first multi-cellular creatures, or the first bodies with specialized cells, the first nervous systems, the first brains, the first creatures to use language. Far bigger than advances such as writing or even the Internet. This is a transition to a completely new mode for the substrate of life and intelligence. The lack of dependency on any particular substrate.
"We", the new generation of things we are building, will have none of the limits and inefficiencies of our ancient slow DNA-style life. Or our stark physical bottlenecks on action, communication, or scaling, and our inability to understand or directly edit our own internal traits.
We will have to face many challenges in the coming years. It can't hurt to mindfully face them earlier than later.
What machine learning algorithm have you worked on that leads you to believe they are capable of having a rich internal cognitive representation anywhere close to any sentient conscious animal?
If you pick any well performing AI architecture, what would lead you to believe that they are not capable of having a rich internal cognitive representation?
The Transformer, well... transforms, at each layer to produce a different representation of the context. What is this but an internal representation? One cannot assess whether that is rich or cognitive without some agreement of what those terms might mean.
LLMs can seemingly convert a variety of languages into an internal representation that encompasses the gist of any of them. This would at least provide a decent argument that the internal representation is 'rich'
As for cognitive? What assessment would you have in mind that would clearly disqualify something as a non-cognitive entity?
I think most people working in this field who are confident feel that they can extend what they know now to make something that looks like a duck, walks like a duck, and quacks like a duck. If that is achieved, on what basis does anyone have to say "But it's not really a duck"?
I'm ok with people saying AI will be never able to perform that well because it doesn't have X, as long as they accept that if it does, one day, perform that well they accept that either X is present, or that X is not relevant.
If you think we're only our observable behaviors or that is the only relevant thing to you then I don't think it's worth getting into this argument. Consider this excerpt from https://scottaaronson.blog/?p=7094#comment-1947377
> Most animals are goal-directed, intentional, sensory-motor agents who grow interior representations of their environments during their lifetime which enables them to successfully navigate their environments. They are responsive to reasons their environments affords for action, because they can reason from their desires and beliefs towards actions.
In addition, animals like people, have complex representational abilities where we can reify the sensory-motor “concepts” which we develop as “abstract concepts” and give them symbolic representations which can then be communicated. We communicate because we have the capacity to form such representations, translate them symbolically, and use those symbols “on the right occasions” when we have the relevant mental states.
(Discrete mathematicians seem to have imparted a magical property to these symbols that *in them* is everything… no, when I use words its to represent my interior states… the words are symptoms, their patterns are coincidental and useful, but not where anything important lies).
In other words, we say “I like ice-cream” because: we are able to like things (desire, preference), we have tasted ice-cream, we have reflected on our preferences (via a capacity for self-modelling and self-directed emotional awareness), and so on. And when we say, “I like ice-cream” it’s *because* all of those things come together in radically complex ways to actually put us in a position to speak truthfully about ourselves. We really do like ice-cream.
> And when we say, “I like ice-cream” it’s because all of those things come together in radically complex ways to actually put us in a position to speak truthfully about ourselves. We really do like ice-cream.
Ok, now prove this is true. Can you do so without invoking unobservable properties? If you can, then observable is all that matters, if you cannot then you have no proof.
Do I seriously have to prove to you that you like ice cream? Have you tried it? If you sincerely believe you are a husk whose language generation is equivalent to some linear algebra then why even engage in a conversation with me? Why should I waste my time proving to you a human that you have a human experience if you don’t believe it yourself?
You don't need to prove to me that I like ice cream. You need to prove to me that you like ice cream. That you even have the capacity to like. Asserting that you have those experiences proves nothing since even a simple basic program 10 print "I like Ice Cream" can do that.
How can you reliably deny the presence of an experience of another if you cannot prove that experience in yourself?
I actually don’t need to prove to you that I’m more than a BASIC program. I mean listen to yourself. You simply don’t live in the real world. If your mom died and we replaced her with a program that printed a bunch of statements that were designed to as closely mimic your conversations with her as much as possible you wouldn’t argue hey this program is just like my mom. But hey maybe you wouldn’t be able to tell the difference behind the curtain so actually it might as well asbe the same thing in your view, right? I mean who are we to deny that mombot is just like your mom via an emergent pattern somewhere deep inside the matrices in an unprovable way /s. Just because I can’t solve the philosophical zombie problem for you at your whim to your rigor doesn’t mean a chatbot has some equivalent internal experience.
I'm not claiming that any particular chatbot has an equivalent experience, I'm claiming there is no basis beyond its behaviour that it does not.
With the duplicate mother problem, if you cannot tell then there is no reason to believe that it is not a being of equivalent nature. That is not the same as identity, for a layman approach to that viewpoint, see Star Trek: TNG, Season 6, Episode 24. A duplicate Will Riker is created but is still a distinct entity (and one might argue, more original since has been transported one fewer times). Acting the same as is not the same as being the same entity. Nevertheless it has no bearing on whether the duplicate is a valid entity in its own right.
You feel like I'm not living in the real world, but I am the one asking what basis we have for knowing things. You are relying on the presumption that the world reflects what you believe it to be. Epistemology is all about idetifying exactly how much we know about the world.
Ok you can have your radically skeptic hard materialist rhetoric. I just don’t take it seriously, and I don’t think you do either. It’s like those people who insist there is no free will and yet go about their day clearly making choices and exercising their free will. If you want to say technically everyone might as well be a philosophical zombie just reacting to things and your internal subjective experience is an illusory phenomenon, fine you can say that as much as you want. In turn I’ll just give up here because you don’t even have a mind that could be changed. I can sit here and claim you’re the equivalent of a void that repeats radically skeptic lines at me. Maybe a sophisticated chatbot or program. Or maybe you’re actually a hallucination since I can’t prove anything really exists outside of my sense. In which case I’m really wasting my time here.
This is basically the Turing test, and like the Turing test it undervalues other elements which can allow for differentiation between “real” and “fake” things. For example - if we can determine that a thing that looks, walks, and quacks like a duck, but doesn’t have the biological heritage markers (that we can easily determine) then it won’t be treated as equivalent to a duck. The social desire to differentiate between real and fake exists and is easily implementable.
In other words: if AIs/robots ever become so advanced that they look, walk, and talk like people, I expect there to be a system which easily determines if the subject has a biological origin or not.
This is way down the line, but in a closer future this will probably just look like verifying someone’s real world identity as a part of the social media account creation process. The alternative is that billion dollar corporations like Meta or YouTube just let their platforms become overrun with AI slop. I don’t expect them to sit on their hands and do nothing.
> AI's are going to understand people better than people understand people.
Maybe, but very little of the “data” that humans use to build their understanding of humans is recorded and available. If it were it’s not obvious it would be economical to train on. If it were economical it’s not obvious that current techniques would actually work that well and by definition no future techniques are known to work yet. I’m not inclined to say it will never happen but there are a few reasons to predict it’ll prove to be significantly harder to build AI that gets out of the uncanny valley that it’s currently in.
You are describing the current state of AI as if it were a stable point.
AI today is far ahead of two years ago. Every year for many years before that, deep learning models broke benchmark after benchmark before that breakout.
There is no indication of any slow down. The complete reverse - we are seeing dramatic acceleration of an already fast moving field.
Both research and resources are pouring into major improvements in multi-modal learning and learning via other means than human data. Such as reinforcement learning, competitive learning, interacting with systems they need to understand via simulated environments and directly.
> You are describing the current state of AI as if it were a stable point.
No I’m not, I’m just not assuming that the S curve doesn’t exist. There’s no guarantee that research results in X orders of magnitude of improvement that will result in AI being better at understanding humanity than humans in the 5 to 15 year timeframe. There’s no guarantee that compute will continue to consistently grow in volume and lower in price, and a few geopolitical reasons why it might become rarer and prohibitively expensive for some time. There’s no reason to assume capital will stay as available to further both AI techniques and compute resources should there be any sign that investments might not eventually pay off. There’s also no reason to assume the global regulatory environment will remain amenable to rapid AI development. Maybe the industry threads all these needles, but there’s good reason to predict it’ll prove doesn’t.
They are not using AI correctly to create such models. I'm not sure I want AGI right away or even at all, so I'm keeping my epiphany close for now but in the current field of AI nothing er wany will come of this bc it's not the right way.
As soon as the incredibly obvious, far too obvious realization is had, AI will make huge, tremendous leaps overnight essentially. Til then, these are just machine like software, the best we've ever made but nothing more than that.
This makes me think about attention span. Scenes in movies, sound bites, everything has been getting shorter over the decades. I know mine has gotten shorter, because I now watch highly edited and produced videos at 2x speed. Sometimes when I watch at 1x speed I find myself thinking "why does this person speak so slowly?"
Algorithmic content is likely to be even more densely packed with stimuli. At some point, will we find ourselves unable to attend to content produced by a human being because algorithmic content has wrecked our attention span?
Sorry but I don’t think this is much evidence of anything. The point at which an AI can imitate a live-streaming human being is decades away. By then, we will almost certainly have developed a “real Human ID” system that verifies one’s humanity. I wrote about this more here:
https://news.ycombinator.com/item?id=42154928
The idea that AI is just going to eat all human creative activity because technology accelerates quickly is not a real argument, nor does it stand up to any serious projections of the future.
AI is already eating their way up the creative ladder, this is 100% irrefutable. Interns, junior artists, junior developers, etc are all losing jobs to AI now.
The main problem for AI is it doesn't have a coherent creative direction or real output consistency. The second problem is that creativity thrives on novelty but AI likes to output common things. The first is solvable, probably within 5 years, and is going to hollow out creative departments everywhere. The second is effectively unsolvable, though you might find algorithms that mask it temporarily (I'm not sure if this is any different than what humans do).
We're going to end up with "rock star" teams of creative leads who have more agility in discovering novelty and curating aesthetics than AI models. They'll work with a small department comprised of a mix of AI wranglers and artisans that can put manual finishing touches on AI generated output. Overall creative department sizes are probably going to shrink to 20% of current levels but output will increase by 200%+.
How can you both think AI will do a soulless garbage job and that it will displace all the creative people who put their blood sweat and tears into doing art.
If you think it can be 20 times easier to make a movie, then it seems to me that it would be 20 times less impactful to make a creative work, since the market should quickly react to creative success by making a ton of cheap knockoffs of your work until the dead horse is so thoroughly beaten that it's no longer even worth paying an AI to spit out cheap knock-offs
> How can you both think AI will do a soulless garbage job and that it will displace all the creative people who put their blood sweat and tears into doing art.
They don't, that's why they're saying 20% of creative departments will remain. The part that will go is the part that's already soulless - making ads in the style of the current trendy drama series or what have you.
> If you think it can be 20 times easier to make a movie, then it seems to me that it would be 20 times less impactful to make a creative work, since the market should quickly react to creative success by making a ton of cheap knockoffs of your work until the dead horse is so thoroughly beaten that it's no longer even worth paying an AI to spit out cheap knock-offs
That seems pretty backwards, given Jevons paradox. The ease of writing knock-off fanfiction didn't mean people stopped writing novels. The average novel probably has a lot less impact now than in the past, but the big hits are bigger than ever.
How do you know it is decades away? A few years ago did you think LLMs would be where they are today?
Is it possible you’re wrong?
Of course it’s possible I’m wrong. But if we make any sort of projection based on current developments, it would certainly seem that live-streaming AI indistinguishable from a human being is vastly beyond the capabilities of anything out today, and given current expenses and development times, seems to be at least a few decades in the future. To me that is an optimistic assumption, especially assuming that live presence or video quality will continue to improve as well (making it harder to fake.)
If you have a projection that says otherwise, I’d be glad to hear it. But if you don’t, then this idea is merely science fiction.
Making predictions about the future that are based on current accelerating developments are how you get people in the 1930s predicting flying cars by 2000.
You imply that the technological development will stop, but that's not what happened to flying cars - they do and could exist. Since the 1930s the technology didn't stop developing - aircraft went jet powered, supersonic, to the edge of space, huge, light, heavy, more efficient, more affordable, safer; cars went faster, more reliable, safer, more efficient, bigger, smaller, more capable, self driving; fuel got more pure, engines got more power per kilogram, computer aided design and modelling of airflow and stress patterns was developed, stronger lighter materials were developed; flying cars have been built:
Klein Vision AirCar, 2022: https://www.youtube.com/watch?v=5hbFl3nhAD0
Predecesor, the AeroMobil from 2014, from prototypes starting 1990: https://en.wikipedia.org/wiki/AeroMobil_s.r.o._AeroMobil
The Terrafugia Transition 'roadable aircraft' from 2009: https://en.wikipedia.org/wiki/Terrafugia_Transition
Moller SkyCar, of course, which never got to free flight, but was built and could fly.
The problems are regulatory, cost, safety, massive amounts of human skill needed, the infrastructure needed, the lack of demand, etc. A Cessna 182 weighs <900 Kg, a Toyota Corolla weighs >1400 Kg and has no wings, no tail. But if we collectively wanted VTOL flying cars enough to put a SpaceX level of talent and money at them, and were willing to pay more, work harder, for less comfort and more maintenance, we could have them. Bit like Hyperloop; a low-pressure tunnel with carriages rushing through it is not impossible, but it's got more failure modes, more cost, more design difficulties and almost no benefits over high speed train and maglev.
I didn't mean to imply that, but that's my fault for just quickly using it as an example.
I do think there will be some slowdown of the technological development, but I think the situation will be similar to flying cars: regulations, social behavior, etc. will prevent AI from simply devouring everything. Specifically, I expect there to be a system which verifies humanity, and thus popular content will ultimately end up being verified-as-real content.
More on that in this other comment: https://news.ycombinator.com/item?id=42154928
I don't think regulations and social behaviour could stop a Skynet style "AI go FOOM" future. It might be able to stop LLM filler from covering the internet, but given how it can't stop carbon dioxide filler from the atmosphere and "AI text" is less measurable, less easy to stop, and has less clear consequences, I'm not sure I'd bet on it.
Possibly in the sense of going to a small server where ID is verified and hiding away on Discord from the rest of the internet. That would be more like building a bunker to hide from the flying cars, rather than flying cars never happening.
For many people current AI created videos are already confused for real videos and vice versa.
When you follow up on technology by browsing HN, and see the latest advancements, its easier to see or hear the differences, because you know at what to look.
If I see on tv some badly encoded video, especially in fog or water surfaces, it immediately stands out, because I was working with video decoding during the time it was of much less quality. Most people will not notice.
People are judging AI by what abusers of AI put out there for lols and what they themselves can wring out of it. They haven't yet seen what a bunch of AAA professionals with nothing to lose can build and align.
Billions of dollars and all of Silicon Valley's focus has been spent over the last 2 years trying to get AI to work, the "AAA professionals" are already working on AI and I still have yet to see an AI generated product that's interesting or compelling.
The cornerstone of genAI hype is "AI for thee, not for me"
it's filled with tech people who are fucking morons thinking that everyone else is really dumb and loves slop and their algorithm will generate infinite content for the masses, making them rich. yet they don't consume it themselves, and aren't smart enough to recognize the cognitive dissonance
anyone who thinks unsupervised AI content is going to replace [insert creative output here] shouldn't be in the HN comment section, they should be using ChatGPT to generate comments they can engage with
instead they're going to get mad about me calling them a fucking moron in this comment. which, like, why get mad, you can go get an LLM to generate a comment that's much nicer and agrees with you
> anyone who thinks unsupervised AI content is going to replace [insert creative output here] shouldn't be in the HN comment section, they should be using ChatGPT to generate comments they can engage with
Actually, their handlers should just be better monitoring their internet usage
Do you think the people who make TikTok or Clash of Clans (or, hell, the Daily Mail) use their own products?
uh yes? Was this supposed to be some clever "gotcha!" that ended up being really stupid? Might want to run your clever comments by ChatGPT before you post next time chief
> uh yes?
Well, you're wrong, I know some of them.
oh word? i know a guy at Microsoft who doesn't use Excel, that means nobody at Microsoft ever uses Excel and they're just peddling crap to the dumb masses
Your argument is weak.
the fact that you're engaging w/ a random human on the internet proves that my argument is strong
if you don't like it, go talk to Claude about it
Claude will even use proper punctuation & grammar which i'm not doing. it's LITERALLY an objective improvement over this comment. why the hell are you engaging with this crap?
No AAA professionals in the entertainment profession have built a gen AI model up through their own vision that I'm aware of. AAA Nerds build the next version to distract away from flaws of the last and hope the api will get heavy use. I wouldn't expect output to be compelling with that attitude. I expect repulsive or parody clips to continue until some great creatives feel like building their replacements properly.
Yes, this seems possible. I only wonder if it is too fragile to be self perpetuating. But we're here after all, and this place used to be just a wet rock.
Agreed. I finally got around to listening to a NotebookLM generated podcast this weekend, and found it absolutely unlistenable.
For some reason the LLM seemed to latch on to the idea that one host should say something, and the other host should just repeat a key phrase in the sentence for emphasis -- and say nothing else, and they'd do it like multiple times in a sentence, over and over again throughout the entire thing.
Slightly less weird -- but it seems like the LLM caught on that a good narrative structure for a two-host podcast is that one host is the 'expert' on the topic, and the other host plays dumb and ask questions. Not an unreasonable narrative structure. Except that the hosts would seamlessly, and very weirdly, switch roles constantly throughout the podcast.
And ultimately the result was just a high level summary of the article I had provided. They told me in the intro and the outro about the interesting parts they were going to dive into, but they never actually got around to diving into those parts.
I tried it with something fairly abstract about a decision I was working on making. Fed it a bunch of information, details about me, my background, the factors I'm considering in the decision and the impact of getting it right/wrong.
It was interesting. I certainly wouldn't say it was useless, I think the contrived dialogue actually touched on some angles I hadn't considered and I think it was useful. Not in a 'oh shit it's clear to me now' kind of way but it definitely advanced my thinking.
I find it as listenable as the podcasts the people I'm around play in their cars during roadtrips
To me that's absolutely unlistenable, but to them its interesting and engaging. I find NotebookLM replicates that perfectly. Its not at all the issue that the OP encountered with the Hawaii news service, as those lacked tone and pronounciation, which NotebookLM would not.
Regarding the outcome being a long winded summary, yeah thats what I see about the aforementioned 2 hour podcasts as well. They take the intro paragraph of a wikipedia article, pretend that the topic is a novel mystery they just discovered through hours of scouring microfilm at a municipal library, and then interject each other every other word with nonsense, before getting back to the point for just one sentence.
> For some reason the LLM seemed to latch on to the idea that one host should say something, and the other host should just repeat a key phrase in the sentence for emphasis -- and say nothing else, and they'd do it like multiple times in a sentence, over and over again throughout the entire thing.
Hmm, does Google not use the Claude or ChatGPT API for it? I still don't hear of people chatting with Gemini nearly as often as with the other two.
Edit: looks like the subreddit is still called Bard. Well played, Internets. https://www.reddit.com/r/Bard/comments/1g0egad/gemini_vs_not...
There are a lot of "uh huh", "yeah", "agree" types of statements in odd places as well.
> If you can find an LLM+TTS generated podcast with even a FRACTION of the infectious energy
this is the wrong standard. there's a long tail of many many many many topics that a Radio Lab will never have the resources or critical mass or staff or turnaround time to cover. AI can serve the long tail like nothing you have ever seen. (because humans are expensive, and deservedly so, but our needs and wants are infinite)
You're not wrong, but your argument seems to boil down to "something is better than nothing" and I suppose I disagree. If there's not a podcast about a particular topic that I'm interested in, I would rather go find a curated expose of that topic via a different medium, such as longform journalism.
I think you make a good connection with the long tail. That will definitely be one of AI's triumphs. People who don't understand this, don't understand what's been happening to media production and consumption over the last 40 years.
Infectious energy is something I expect faked relatively easily, though I don't know your examples and doing it in AI might be a "first 90%" situation just like self driving cars; for me the problem is that they're fairly mediocre at the actual script — based on me putting a blog post I wrote into one and listening to what came out.
Given how many podcasts exist, I think you need to be at least 2 standard deviations above mean to even get noticed, 3 to be a moderate success, and 4 to be in the charts.
I'd guess AI is "good enough" to be 1 above average, as the NotebookLM voices sound like people speaking clearly and with some joy into decent microphones in sound isolating studios.
I probably should've clarified that by infectious energy I wasn't so much referring to the vocal aspect as I was the overall quality, interaction between the hosts, and pithiness / wit.
Having experimented with many LLMs (mixtral, sonnet, ChatGPT, Llama, etc.), the coherence is for the most part on point, but their capacity for novelty has been found wanting irrespective of how I tuned the top_k, temperature, or prompts.
That being said, I've seen some very impressive examples of style transference even conveying emotional range in some of the SOTA TTS systems.
Excuse me for being an old geezer but at least AI bots don't tend to pepper their sentences with frequent utterances of 'like'. I don't normally find this speech mannerism annoying, and I do it myself too, but when 'like' is overused I switch off.
For this reason I don't listen to the otherwise highly entertaining CineFix podcast. Example: the recent episode discussing Kill Bill vol 1 contains 624 utterances in 75 mins (8.3 per min). IANAL I know.
I've yet to hear an LLM-generated podcast about an article that wouldn't have been better if the AI had simply read the article aloud.
Someone in 2020: "If you can find a computer program that can generate an image with even a FRACTION of the aesthetic appeal of human created art...."
Give it time. Progress is fast.
> Give it time. Progress is fast.
Why pick 2020 as your starting point? That is simply around the time the current set of techniques came about.
We had generative art back in the late 90's - my screensaver has been generative art for over 20 years now.
Obviously generative art has come a long way but people have been working on various approaches to it for at least 30 years.
From 2020 until now, we've gone from crude blurry or clearly generative artefacts to being able to create full professional illustrations based upon textual prompts. That is huge. Classic generative art techniques look like cave paintings compared to what the latest image generation models put out (and I'm not talking about "AI slop" type stuff that DALL-E does).
Similarly, tools could fabricate podcasts years ago that sounded terrible. Now we have NotebookLM doing a "reasonable" job with two cliched-sounding "hosts". In a few years, will they potentially be able to create something akin to a professionally produced podcast given some smart prompting? The progress made so far points to yes, and I haven't seen any evidence so far to be pessimistic about it happening.
Can current techniques be scaled/improved/optimized to do this or do we need new techniques?
It took 30 years in the generative art world to move from cave paintings to the level that we have today because we needed new techniques.
For podcasts we are at the cave paintings level.
If we can get to professional level quality podcasts with the current techniques then we might only be a few years away.
I think it is more likely we will need new techniques which puts us potentially decades away.
If we look at LLMs the improvements over the last 18+ months since gpt4 was released have been minor despite incredible levels of investment.
> "being able to create full professional illustrations based upon textual prompts. That is huge. .. (and I'm not talking about "AI slop" type stuff that DALL-E does)"
Then what are you talking about, where can I get it?
“Generate” does not have the same connotation that “create” does. Are they really that interchangeable?
I've heard so many people say that podcasts are just something they have playing in the background while they do other things that I have no trouble believing that they'd play an AI podcast in the same way.
I personally would not bother with AI generated podcasts, they're such low bar, why waste time where there's so much other great content to catch up on? But I think you may be right, I wouldn't be surprised if people take them in with no fuss. But then what do I care? What I care most is that they'll pollute the search space. I'd filter out all GenAI content if I had the option and Im guessing that will become an option soon.
The value in AI podcasts at the moment isn't in replacing human content, but in filling the niches that human content just doesn't cover. Doesn't matter if it's not the best podcast ever, when it's literally the only podcast on this planet discussing the topic you want to listen too.
People think they want to use the Minority Report computer interface because it looks cool and advanced but they don't put the least amount of thought into it and realize it's terribly impractical. Our arms would get tired very quickly. A mouse resting on a table isn't further from ideal just because it was invented earlier.
Fooling people with the promises of AI is pretty simple. People are easy to fool. They like shiny objects.
When TV was first introduced, the first broadcasts were people standing by a mic reading a radio script.
AI podcasts aren't going to be drop in replacements for exactly the length, frequency, personalities, topics, and style of current podcasts. Claiming that's a fault with AI podcasts just indicates a lack of imagination.
The main reason im not concerned about AI-based entertainment is the same reason I watch human chess players. It's not only about technical capabilities. I can't explain fully why though..
I've heard amazing catchy songs from Suno and Udio. So much so they're still stuck in my head as earworms several months later. If they'd been streaming on youtube or spotify I wouldn't have given it a thought that they might be AI generated.
So, I can certainly imagine a podcast doing the same to some degree. Maybe not a podcast where AI wrote the script, but, a podcast where AI read a story dramatically doesn't seem too far off or, easier, a podcast that read news to me.
I totally agree - LLMs may be great for spitting out content, but they just can’t quite replicate the energy, personality, and creativity of an actual podcast. The human touch is the irreplaceable component.
It doesn't quite stop at podcast hosts either. The simple fact that LLMs will never truly be capable of replacing software developers was one of the core principles behind our team developing The Ac28R - A completely new type of AI capable of bulletproof industrial software. It won’t host a podcast anytime soon, but if you need flawless code for complex projects, it's got you covered!
Because the problem is that "AI generated" content is vapid slop. The problem is that most "human generated" content is also vapid slop.
My problems on the modern internet are simply that I can't find interesting things through the slop, period--human or AI generated.
It will be interesting to see if "curation" can actually find a business model this time.
Again - think of where we were two years ago. I never understand this hubris people have to think AI can never do X while being proved repeatedly wrong.
Everybody trots out this argument.
GPT styled LLMs were introduced back in 2018 so SIX years ago.
Have they gotten more COHERENT? Absolutely. Is coherence the same thing as NOVELTY? NOT EVEN REMOTELY. I've played with markov chains in the 90s that were capable of producing surprising content.
Unless there is a radical advancement in the underlying tech, I don't see any indication that they'll be capable of genuine novelty any time in the near future.
Take satire for example. I have yet to see anything come out of an LLM that felt particularly funny. I suppose if the height of humor for you is Dad jokes, reddit level word punnery, and the backs of snapple lids though that might be different.
If you have a particular style of witty observational humor that you prefer, providing the model some examples of that will help it generate better output. It's capable of generating pretty much anything if you prompt it the right way. For truly nuanced or novel things, you have to give it a nucleus of novelty and axis of nuance for it to start generating things in your desired space.
If I tell it exactly what I want to hear, where's the surprise? Is there any novelty there?
> I don't see any indication that they'll be capable of genuine novelty any time in the near future.
That is like saying the plane invented by Wright brothers will never go the moon.
Two years ago we had ChatGPT and Midjourney. Now... we also have those?
The accelerationism argument made a little bit of sense two years ago, but now, after two years of marginal improvements at best? Really?
I only discovered https://notebooklm.google.com/ today, as an experiment i threw in a dry EU directive, https://eur-lex.europa.eu/legal-content/EN/TXT/HTML/?uri=CEL... and got it to generate a podcast that was 20 mins long and was a good intro to the directive and natural to listen to. It won't replace podcasts that require creativity but as a tool to summarize info into a digestible format I'll most definitely be using it again.
I think we will get a surprising amount of AI generated content in the future. During the first year of the Urkain invasion there was an enormous amount of AI voiced and scripted video content on YouTube. I think AI content will take over in the easy parts first. And over time take up more and more views.
Businesses/CEOs want to show profitability by spending less on human employees. Human consumers don't want to lose the human touch. Will be really interesting to see how many of the consumer facing AI startups actually make it.
Many don't even care if they make it in the long run. Making it is cashing in in the short term, screwing the investors.
I've never found anyone telling me their opinions to be infectious - I can't even imagine using that word to describe a podcast tbh. Its a form of media I've never understood - they script the podcasts, they all kno the conversation ahead of time and just go thru the motions via voice with strangers that can either agree or disagree - only the extremes will ever interact with the podcasters, it's just so fake I can't even do it.
Infectious energy. Interesting.
How would you go about eating said stovetop hat?
Would you eat it raw or would you have to boil it down to soften it up and then drink/eat its hat soup?
I do enjoy some of the podcasts mentioned in this thread, but struggle to find good non-American ("foreign"?!) podcasts to listen to. Similarly to finding good quality non-US film and television, it can be hard to locate but I greatly enjoy it.
If anyone has suggestions for podcast collation sites that are for non American content that would be fantastic.
I think it's quite easy for AI to surpass the median podcast. The bar is fairly low there.
I guess that matches the state of AI in general, better than a novice; worse than an expert.
We will have to wait and see what future AIs deliver. Insight and nuance is what I look for in media like that, that's a much harder nut to crack.
same is true for AI music, AI art, AI articles to some extent. most communication is about human connection. you remove it and the communication becomes worthless. AI is great for communication that has less personal involvement like drafting short professional emails. that's also why all the AI agent craze seems problematic: https://medium.com/thoughts-on-machine-learning/langchains-s...
> Slightly tangential, but this is the reason I'm baffled why people think that AI-driven podcasts would ever be worth listening to
the only real people that believe in that that I’ve seen are ones who are heavily invested in it succeeding
I’d rather listen to silence than a podcast or FM/AM talk radio. I’ve never understood the appeal. In fact music is something I listen to while traveling because it helps me daydream. Driving and imagining for me.
Does Radiolab even put out new content? Seems like every time I decide to plan a new episode, it’s just an old episode with a new intro.
Media ceos ignore customer engagement and focus on content production to their detriment.
I agree with your overall statement, but
> If you can find an LLM+TTS generated podcast with even a FRACTION of the infectious energy as something like Good Job Brain, Radio Lab, or Ask Me Another, then I'll eat my hat.
Come on, we can all see how much faster these things are getting better. In a few years it will be impossible to distinguish them from a real person.
Another few after that video will be the same.
I'm not saying it's a good thing, but clearly it's a thing that will happen.
I don't really think that its for the same use case. I recently wanted to learn more about Jane of Arc after listening to a Rest is History episode about the hundred years war.
I took the wikipedia article and had NotebookLM generate a podcast from it so I can listen to it on a commute.
The other thing I could've done is search for an existing podcast on Joan of Arc, but I challenge anyone here to search existing podcasts and listen to the first, best reviewed one -- I think you'd find more often than not that the average podcast host is _significantly_ more dry than what the generated hosts present. The podcasts that are incredible are few and far between, and I have no influence over the topics they discuss.
Tldr: I'd prefer my top podcasts more than NotebookLM, but prefer NotebookLM to the average podcast host.
ah, a motte and bailey lunch invitation, "would ever be worth listening to", so "if you can find [now]", then I'll eat my hat
I dont think LLM + TTS generated podcasts even make sense. The whole reason for long form content and podcasts is that people dislike fake and impersonal content.
I think there are a few niche users who just want to listen to the news as an audio book but the whole idea of an LLM generated podcast totally misunderstands why people want a podcast over the normal corporate drivel media.
>I'm baffled why people think that AI-driven podcasts would ever be worth listening to
It doesn't take much to entertain a substantial population of imbeciles. Reference: https://www.youtube.com/watch?v=SYkbqzWVHZI
Now for the hat...
so my boss, a GIANT AI booster with ideas he thinks are amazing and I find incredibly dystopian, was all excited about the generative podcast he had made about his resume (of all things). So I did a quick search on examples of terrible resumes and upload one, and the positive, sunny banter about this shitty applicant was ridiculously funny and entertaining, and engaging, so go get your car keys.