6 stories
·
0 followers

Can AI Make Art?

2 Shares

Ted Chiang with a thought-provoking essay on Why A.I. Isn’t Going to Make Art:

It is very easy to get ChatGPT to emit a series of words such as “I am happy to see you.” There are many things we don’t understand about how large language models work, but one thing we can be sure of is that ChatGPT is not happy to see you. A dog can communicate that it is happy to see you, and so can a prelinguistic child, even though both lack the capability to use words. ChatGPT feels nothing and desires nothing, and this lack of intention is why ChatGPT is not actually using language. What makes the words “I’m happy to see you” a linguistic utterance is not that the sequence of text tokens that it is made up of are well formed; what makes it a linguistic utterance is the intention to communicate something.

In the past few years, Chiang has written often about the limitations of LLMs — you can read more about his AI views on kottke.org.

Tags: art · artificial intelligence · Ted Chiang

💬 Join the discussion on kottke.org

Read the whole story
billyhopscotch
64 days ago
reply
Share this story
Delete

I’d missed that Ray Nayler, author of the excellent The Mountain in...

1 Comment
I’d missed that Ray Nayler, author of the excellent The Mountain in the Sea, came out with a short novel earlier this year called The Tusks of Extinction. “Now, her digitized consciousness has been downloaded into the mind of a mammoth.”

💬 Join the discussion on kottke.org

Read the whole story
billyhopscotch
64 days ago
reply
I absolutely loved The Mountain in the Sea, so I'm stoked to check out The Tusks of Extinction.
Share this story
Delete

When kids can’t get outside to play in a world built for...

1 Share
When kids can’t get outside to play in a world built for cars, both they and adults suffer. “Kids didn’t need special equipment or lessons; they just needed to be less reliant on their time-strapped parents to get outside.”

💬 Join the discussion on kottke.org

Read the whole story
billyhopscotch
92 days ago
reply
Share this story
Delete

Pluralistic: The true, tactical significance of Project 2025 (14 Jul 2024)

2 Shares


Today's links



An X-ray of a broken femur. On either side of the fracture is a elephant (cropped from a medieval illumination) facing one another, in the livery of the GOP logo.

The true, tactical significance of Project 2025 (permalink)

Like you, I have heard a lot about Project 2025, the Heritage Foundation's roadmap for the actions that Trump should take if he wins the presidency. Given the Heritage Foundation's centrality to the American authoritarian project, it's about as awful and frightening as you might expect:

https://www.project2025.org/

But (nearly) all the reporting and commentary on Project 2025 badly misses the point. I've only read a single writer who immediately grasped the true significance of Project 2025: The American Prospect's Rick Perlstein, which is unsurprising, given Perlstein's stature as one of the left's most important historians of right wing movements:

https://prospect.org/politics/2024-07-10-project-2025-republican-presidencies-tradition/

As Perlstein points out, Project 2025 isn't new. The Heritage Foundation and its allies have prepared documents like this, with many identical policy prescriptions, in the run-up to many presidential elections. Perlstein argues that Warren G Harding's 1921 inaugural address captures much of its spirit, as did the Nixon campaign's 1973 vow to "move the country so far to the right 'you won’t even recognize it.'"

The threats to democracy and its institutions aren't new. The right has been bent on their destruction for more than a century. As Perlstein says, the point of taking note of this isn't to minimize the danger, rather, it's to contextualize it. The American right has, since the founding of the Republic, been bent on creating a system of hereditary aristocrats, who govern without "interference" from democratic institutions, so that their power to extract wealth from First Nations, working people, and the land itself is checked only by rivalries with other aristocrats. The project of the right is grounded in a belief in Providence: that God's favor shines on His best creations and elevates them to wealth and power. Elite status is proof of merit, and merit is "that which leads to elite status."

When a wealthy person founds an intergenerational dynasty of wealth and power, this is merely a hereditary meritocracy: a bloodline infused with God's favor. Sometimes, this belief is dressed up in caliper-wielding pseudoscience, with the "good bloodline" reflecting superior genetics and not the favor of the Almighty. Of course, a true American aristocrat gussies up his "race realism" with mystical nonsense: "God favored me with superior genes." The corollary, of course, is that you are poor because God doesn't favor you, or because your genes are bad, or because God punished you with bad genes.

So we should be alarmed by the right's agenda. We should be alarmed at how much ground it has gained, and how the right has stolen elections and Supreme Court seats to enshrine antimajoritarianism as a seemingly permanent fact of life, giving extremist minorities the power to impose their will on the rest of us, dooming us to a roasting planet, forced births, racist immiseration, and most expensive, worst-performing health industry in the world.

But for all that the right has bombed so many of the roads to a prosperous, humane future, it's a huge mistake to think of the right as a stable, unified force, marching to victory after inevitable victory. The American right is a brittle coalition led by a handful of plutocrats who have convinced a large number of turkeys to vote for Christmas.

The right wing coalition needs to pander to forced-birth extremists, racist extremist, Christian Dominionist extremists (of several types), frothing anti-Communist cranks, vicious homophobes and transphobes, etc, etc. Pandering to all these groups isn't easy: for one thing, they often want opposite things – the post-Roe forced birth policies that followed the Dobbs decision are wildly unpopular among conservatives, with the exception of a clutch of totally unhinged maniacs that the party relies on as part of a much larger coalition. Even more unpopular are policies banning birth control, like the ones laid out in Project 2025. Less popular still: the proposed ban on no-fault divorce. Each of these policies have different constituencies to whom they are very popular, but when you put them together, you get Dan Savage's "Husbands you can't leave, pregnancies you can't prevent or terminate, politicians you can't vote out of office":

https://twitter.com/fakedansavage/status/1805680183065854083

The constituency for "husbands you can't leave, pregnancies you can't prevent or terminate, politicians you can't vote out of office" is very small. Almost no one in the GOP coalition is voting for all of this, they're voting for one or two of these things and holding their noses when it comes to the rest.

Take the "libertarian" wing of the GOP: its members do favor personal liberty…it's just that they favor low taxes for them more than personal liberty for you. The kind of lunatic who'd vote for a dead gopher if it would knock a quarter off his tax bill will happily allow his coalition partners to rape pregnant women with unnecessary transvaginal ultrasounds and force them to carry unwanted fetuses to term if that's the price he has to pay to save a nickel in taxes:

https://pluralistic.net/2021/09/29/jubilance/#tolerable-racism

And, of course, the religious maniacs who profess a total commitment to Biblical virtue but worship Trump, Gaetz, Limbaugh, Gingrich, Reagan, and the whole panoply of cheating, lying, kid-fiddling, dope-addled refugees from a Jack Chick tract know that these men never gave a shit about Jesus, the Apostles or the Ten Commandments – but they'll vote for 'em because it will get them school prayer, total abortion bans, and unregulated "home schooling" so they can brainwash a generation of Biblical literalists who think the Earth is 5,000 years old and that Jesus was white and super into rich people.

Time and again, the leaders of the conservative movement prove themselves capable of acts of breathtaking cruelty, and undoubtedly many of them are depraved sadists who genuinely enjoy the suffering of their enemies (think of Trump lickspittle Steven Miller's undisguised glee at the thought of parents who would never be reunited with children after being separated at the border). But it's a mistake to think that "the cruelty is the point." The point of the cruelty is to assemble and maintain the coalition. Cruelty is the tactic. Power is the point:

https://pluralistic.net/2022/03/09/turkeys-voting-for-christmas/#culture-wars

The right has assembled a lot of power. They did so by maintaining unity among people who have irreconcilable ethics and goals. Think of the pro-genocide coalition that includes far-right Jewish ethno-nationalists, antisemitic apocalyptic Christians who believe they are hastening the end-times, and Islamophobes of every description, from War On Terror relics to Hindu nationalists.

This is quite an improbable coalition, and while I deplore its goals, I can't help but be impressed by its cohesion. Can you imagine the kind of behind-the-scenes work it takes to get antisemites who think Jews secretly control the world to lobby with Zionists? Or to get Zionists to work alongside of Holocaust-denying pencilneck Hitler wannabes whose biggest regret is not bringing their armbands to Charlottesville?

Which brings me back to Project 2025 and its true significance. As Perlstein writes, Project 2025 is a mess. Clocking in an 900 pages, large sections of Project 2025 flatly contradict each other, while other sections contain subtle contradictions that you wouldn't notice unless you were schooled in the specialized argot of the far right's jargon and history.

For example, Project 2025 calls for defunding government agencies and repurposing the same agencies to carry out various spectacular atrocities. Both actions are deplorable, but they're also mutually exclusive. Project 2025 demands four different, completely irreconcilable versions of US trade policy. But at least that's better than Project 2025's chapter on monetary policy, which simply lays out every right wing theory of money and then throws up its hands and recommends none of them.

Perlstein says that these conflicts, blank spots and contradictions are the most important parts of Project 2025. They are the fracture lines in the coalition: the conflicting ideas that have enough support that neither side can triumph over the other. These are the conflicts that are so central to the priorities of blocs that are so important to the coalition that they must be included, even though that inclusion constitutes a blinking "LOOK AT ME" sign telling us where the right is ready to split apart.

The right is really good at this. Perlstein points to Nixon's expansion of affirmative action, undertaken to sow division between Black and white workers. We need to get better at it.

So far, we've lavished attention on the clearest and most emphatic proposals in Project 2025 – for understandable reasons. These are the things they say they want to do. It would be reckless to ignore them. But they've been saying things like this for a century. These demands constitute a compelling argument for fighting them as a matter of urgency, with the intention of winning. And to win, we need to split apart their coalition.

Perlstein calls on us to dissect Project 2025, to cleave it at its joints. To do so, he says we need to understand its antecedents, like Nixon's "Malek Manual," a roadmap for destroying the lives of civil servants who failed to show sufficient loyalty to Nixon. For example, the Malek Manual lays out a "Traveling Salesman Technique" whereby a government employee would be given duties "criss-crossing him across the country to towns (hopefully with the worst accommodations possible) of a population of 20,000 or under. Until his wife threatens him with divorce unless he quits, you have him out of town and out of the way":

https://www.google.com/books/edition/Final_Report_on_Violations_and_Abuses_of/0dRLO9vzQF0C?hl=en&gbpv=1&dq=%22organization+of+a+political+personnel+office+and+program%22&pg=PA161&printsec=frontcover

It's no coincidence that leftist historians of the right are getting a lot of attention. Trumpism didn't come out of nowhere – Trump is way too stupid and undisciplined to be a cause – he's an effect. In his excellent, bestselling new history of the right in the early 1990s, When the Clock Broke, Josh Ganz shows us the swamp that bred Trump, with such main characters as the fascist eugenicist Sam Francis:

https://us.macmillan.com/books/9780374605445/whentheclockbroke

Ganz joins the likes of the Know Your Enemy podcast, an indispensable history of reactionary movements that does excellent work in tracing the fracture lines in the right coalition:

https://www.patreon.com/posts/when-clock-broke-106803105

Progressives are also an uneasy coalition that is easily splintered. As Naomi Klein argues in her essential Doppelganger, the liberal-left coalition is inherently unstable and contains the seeds of its own destruction:

https://pluralistic.net/2023/09/05/not-that-naomi/#if-the-naomi-be-klein-youre-doing-just-fine

Liberals have been the senior partner in that coalition, and their commitment to preserving institutions for their own sake (rather than because of what they can do to advance human thriving) has produced generations of weak and ineffectual responses to the crises of terminal-stage capitalism, like the idea that student-debt cancellation should be means-tested:

https://pluralistic.net/2022/05/03/utopia-of-rules/#in-triplicate

The last bid for an American aristocracy was repelled by rejecting institutions, not preserving them. When the Supreme Court thwarted the New Deal, FDR announced his intention to pack the court, and then began the process of doing so (which included no-holds-barred attacks on foot-draggers in his own party). Not for nothing, this is more-or-less what Lincoln did when SCOTUS blocked Reconstruction:

https://pluralistic.net/2020/09/20/judicial-equilibria/#pack-the-court

But the liberals who lead the progressive movement dismiss packing the court as unserious and impractical – notwithstanding the fact that they have no plan for rescuing America from the bribe-taking extremists, the credibly accused rapist, and the three who stole their robes. Ultimately, liberals defend SCOTUS because it is the Supreme Court. I defended SCOTUS, too – while it was still a vestigial organ of the rights revolution, which improved the lives of millions of Americans. Human rights are worth defending, SCOTUS isn't. If SCOTUS gets in the way of human rights, then screw SCOTUS. Sideline it. Pack it. Make it a joke.

Fuck it.

This isn't to argue for left seccession from the progressive coalition. As we just saw in France, splitting at this moment is an invitation to literal fascist takeover:

https://jacobin.com/2024/07/melenchon-macron-france-left-winner

But if there's one thing that the rise of Trumpism has proven, it's that parties are not immune to being wrestled away from their establishment leaderships by radical groups:

https://pluralistic.net/2023/06/16/that-boy-aint-right/#dinos-rinos-and-dunnos

What's more, there's a much stronger natural coalition that the left can mobilize: workers. Being a worker – that is, paying your bills from wages, instead of profits – isn't an ideology you can change, it's a fact. A Christian nationalist can change their beliefs and then they will no longer be a Christian nationalist. But no matter what a worker believes, they are still a worker – they still have a irreconcilable conflict with people whose money comes from profits, speculation, or rents. There is no objectively fair way to divide the profits a worker's labor generates – your boss will always pay you as little of that surplus as he can. The more wages you take home, the less profit there is for your boss, the fewer dividends there are for his shareholders, and the less there is to pay to rentiers:

https://pluralistic.net/2024/04/19/make-them-afraid/#fear-is-their-mind-killer

Reviving the role of workers in their unions, and of unions in the Democratic party, is the key to building the in-party power we need to drag the party to real solutions – strong antimonopoly action, urgent climate action, protections for gender, racial and sexual minorities, and decent housing, education and health care.

The alternative to a worker-led Democratic Party is a Democratic Party run by its elites, whose dictates and policies are inescapably illegitimate. As Hamilton Nolan writes, the completely reasonable (and extremely urgent) discussion about Biden's capacity to defeat Trump has been derailed by the Democrats' undemocratic structure. Ultimately, the decision to have an open convention or to double down on a candidate whose campaign has been marred by significant deficits is down to a clutch of party officials who operate without any formal limits or authority:

https://www.hamiltonnolan.com/p/the-hole-at-the-heart-of-the-democratic

Jettisoning Biden because George Clooney (or Nancy Pelosi) told us to is never going to feel legitimate to his supporters in the party. But if the movement for an open convention came from grassroots-dominated unions who themselves dominated the party – as was the case, until the Reagan revolution – then there'd be a sense that the party had constituents, and it was acting on its behalf.

Reviving the labor movement after 40 years of Reaganomic war on workers may sound like a tall order, but we are living through a labor renaissance, and the long-banked embers of labor radicalism are reigniting. What's more, repelling fascism is what workers' movements do. The business community will always sell you out to the Nazis in exchange for low taxes, cheap labor and loose regulation.

But workers, organized around their class interests, stand strong. Last week, we lost one of labor's brightest flames. Jane McAlevey, a virtuoso labor organizer and trainer of labor organizers, died of cancer at 57:

https://jacobin.com/2024/07/jane-mcalevey-strategy-organizing-obituary

McAlevey fought to win. She was skeptical of platitudes like "speaking truth to power," always demanding an explanation for how the speech would become action. In her classic book A Collective Bargain, she describes how she built worker power:

https://pluralistic.net/2023/04/23/a-collective-bargain/

McAlevey helped organize a string of successful strikes, including the 2019 LA teachers' strike. Her method was straightforward: all you have to do to win a strike or a union drive is figure out how to convince every single worker in the shop to back the union. That's all.

Of course, it's harder than it sounds. All the problems that plague every coalition – especially the progressive liberal/left coalition – are present on the shop floor. Some workers don't like each other. Some don't see their interests aligned with others. Some are ornery. Some are convinced that victory is impossible.

McAlevey laid out a program for organizing that involved figuring out how to reach every single worker, to converse with them, listen to them, understand them, and win them over. I've never read or heard anyone speak more clearly, practically and inspirationally about coalition building.

Biden was never my candidate. I supported three other candidates ahead of him in 2020. When he got into office and started doing a small number of things I really liked, it didn't make me like him. I knew who he was: the Senator from MBNA, whose long political career was full of bills, votes and speeches that proved that while we might have some common goals, we didn't want the same America or the same world.

My interest in Biden over the past four years has had two areas of focus: how can I get him to do more of the things that will make us all better off, and do less of the things that make the world worse. When I think about the next four years, I'm thinking about the same things. A Trump presidency will contain far more bad things and far fewer good ones.

Many people I like and trust have pointed out that they don't like Biden and think he will be a bad president, but they think Trump will be much worse. To limit Biden's harms, leftists have to take over the Democratic Party and the progressive movement, so that he's hemmed in by his power base. To limit Trump's harms, leftists have to identify the fracture lines in the right coalition and drive deep wedges into them, shattering his power base.


Hey look at this (permalink)



A Wayback Machine banner.

This day in history (permalink)

#20yrsago RIAA’s INDUCE Act letter deconstructed https://corante.com/importance/the-excessively-annotated-riaa-letter-on-the-induce-act-iica/

#20yrsago Lou Reed wants remixes https://web.archive.org/web/20040804104424/https://www.billboard.com/bb/daily/article_display.jsp?vnu_content_id=1000577588

#20yrsago ICANN emancipate domain owners from scummy registrars https://web.archive.org/web/20040722061910/http://www.byte.org/blog/_archives/2004/7/14/105552.html

#20yrsago Disney’s $80 million mistake: Fahrenheit 911 https://web.archive.org/web/20040804183640/https://www.technicianonline.com/story.php?id=009702

#20yrsago Druid busted for possession of a sword https://mg.co.za/article/2004-07-13-swordpacking-druid-appears-in-court/

#15yrsago Michael Jackson didn’t sell 750 million records https://www.wsj.com/articles/SB124760651612341407

#15yrsago Phones confiscated at preview screenings: whose hypothetical risk is more important? https://www.theguardian.com/technology/2009/jul/14/mobile-phones-and-movie-security

#15yrsago Visa claims teen spent $23,148,855,308,184,500.00 on prepaid credit card https://web.archive.org/web/20090716125509/https://consumerist.com/5314246/unruly-teen-charges-23-quadrillion-at-drugstore

#10yrsago Freedom of info funnies: CIA cafeteria complaints https://www.muckrock.com/news/archives/2014/jul/14/doc-note-cia-cafeteria-complaints/

#10yrsago Economist examines empirical evidence of file-sharing on box-office revenue https://web.archive.org/web/20140816180401/http://conference.nber.org/confer/2014/SI2014/PRIT/Strumpf.pdf

#10yrsago Understanding #DRIP: new spy powers being rammed through UK Parliament https://web.archive.org/web/20140711071612/https://www.openrightsgroup.org/campaigns/no-emergency-stop-the-data-retention-stitch-up

#10yrsago Tesla’s “car-as-service” versus your right to see your data https://appliedabstractions.com/2014/07/14/elon-i-want-my-data/

#10yrsago Scalia may have opened path for Quakers to abstain from taxes https://www.salon.com/2014/07/14/scalias_major_screw_up_how_scotus_just_gave_liberals_a_huge_gift/

#10yrsago Unions considered helpful (economically) https://stumblingandmumbling.typepad.com/stumbling_and_mumbling/2014/07/unions-productivity-.html

#10yrsago Hearings into mass surveillance begin in UK https://www.theguardian.com/uk-news/2014/jul/14/court-gchq-surveillance-tempora-ipt-nsa-snowden

#10yrsago Everyone hates the NSA: survey https://web.archive.org/web/20140715012054/http://www.pewglobal.org/2014/07/14/nsa-opinion/table/country-citizens/

#10yrsago GCHQ’s black bag of dirty hacking tricks revealed https://web.archive.org/web/20140714190448/https://firstlook.org/theintercept/2014/07/14/manipulating-online-polls-ways-british-spies-seek-control-internet/

#10yrsago Snowden: #DRIP “defies belief,” could have been dreamed up by NSA https://www.theguardian.com/world/2014/jul/13/edward-snowden-condemns-britain-emergency-surveillance-bill-nsa

#5yrsago Florida DMV makes millions selling Floridians’ data…for pennies (and you can’t opt out) https://www.wxyz.com/news/national/florida-is-selling-drivers-personal-information-to-private-companies-and-marketing-firms

#5yrsago #TelegramGate: leaks show Puerto Rico’s appointed officials mocking the dead as hurricanes devastate the island https://web.archive.org/web/20190714004011/https://abcnews.go.com/International/wireStory/puerto-rican-chief-financial-officer-resigns-chat-scandal-64318436

#1yrago Why they're smearing Lina Khan https://pluralistic.net/2023/07/14/making-good-trouble/#the-peoples-champion


Upcoming appearances (permalink)

A photo of me onstage, giving a speech, holding a mic.



A screenshot of me at my desk, doing a livecast.

Recent appearances (permalink)



A grid of my books with Will Stahle covers..

Latest books (permalink)



A cardboard book box with the Macmillan logo.

Upcoming books (permalink)

  • Picks and Shovels: a sequel to "Red Team Blues," about the heroic era of the PC, Tor Books, February 2025

  • Unauthorized Bread: a middle-grades graphic novel adapted from my novella about refugees, toasters and DRM, FirstSecond, 2025



Colophon (permalink)

Today's top sources:

Currently writing:

  • Enshittification: a nonfiction book about platform decay. July 1's progress: 792 words (20879 words total).

  • A Little Brother short story about DIY insulin PLANNING

  • Picks and Shovels, a Martin Hench noir thriller about the heroic era of the PC. FORTHCOMING TOR BOOKS JAN 2025

  • Vigilant, Little Brother short story about remote invigilation. FORTHCOMING ON TOR.COM

  • Spill, a Little Brother short story about pipeline protests. FORTHCOMING ON TOR.COM

Latest podcast: The reason you can't buy a car is the same reason that your health insurer let hackers dox you https://craphound.com/news/2024/06/30/the-reason-you-cant-buy-a-car-is-the-same-reason-that-your-health-insurer-let-hackers-dox-you/


This work – excluding any serialized fiction – is licensed under a Creative Commons Attribution 4.0 license. That means you can use it any way you like, including commercially, provided that you attribute it to me, Cory Doctorow, and include a link to pluralistic.net.

https://creativecommons.org/licenses/by/4.0/

Quotations and images are not included in this license; they are included either under a limitation or exception to copyright, or on the basis of a separate license. Please exercise caution.


How to get Pluralistic:

Blog (no ads, tracking, or data-collection):

Pluralistic.net

Newsletter (no ads, tracking, or data-collection):

https://pluralistic.net/plura-list

Mastodon (no ads, tracking, or data-collection):

https://mamot.fr/@pluralistic

Medium (no ads, paywalled):

https://doctorow.medium.com/

Twitter (mass-scale, unrestricted, third-party surveillance and advertising):

https://twitter.com/doctorow

Tumblr (mass-scale, unrestricted, third-party surveillance and advertising):

https://mostlysignssomeportents.tumblr.com/tagged/pluralistic

"When life gives you SARS, you make sarsaparilla" -Joey "Accordion Guy" DeVilla

Read the whole story
billyhopscotch
115 days ago
reply
Share this story
Delete

Pop Culture

1 Comment and 2 Shares

Hi there. Do you like this post? Did you know that I also do a podcast called Better Offline? If not, please immediately download it on your podcast app. Follow the show. Download every episode. Share with your friends, and demand they do the same.


A week and a half ago, Goldman Sachs put out a 31-page-report (titled "Gen AI: Too Much Spend, Too Little Benefit?”) that includes some of the most damning literature on generative AI I've ever seen. And yes, that sound you hear is the slow deflation of the bubble I've been warning you about since March

The report covers AI's productivity benefits (which Goldman remarks are likely limited), AI's returns (which are likely to be significantly more limited than anticipated), and AI's power demands (which are likely so significant that utility companies will have to spend nearly 40% more in the next three years to keep up with the demand from hyperscalers like Google and Microsoft).

This report is so significant because Goldman Sachs, like any investment bank, does not care about anyone's feelings unless doing so is profitable. It will gladly hype anything if it thinks it'll make a buck. Back in May, it was claimed that AI (not just generative AI) was "showing very positive signs of eventually boosting GDP and productivity," even though said report buried within it constant reminders that AI had yet to impact productivity growth, and states that only about 5% of companies report using generative AI in regular production.

For Goldman to suddenly turn on the AI movement suggests that it’s extremely anxious about the future of generative AI, with almost everybody agreeing on one core point: that the longer this tech takes to make people money, the more money it's going to need to make.

The report includes an interview with economist Daron Acemoglu of MIT (page 4), an Institute Professor who published a paper back in May called "The Simple Macroeconomics of AI" that argued that "the upside to US productivity and, consequently, GDP growth from generative AI will likely prove much more limited than many forecasters expect." A month has only made Acemoglu more pessimistic, declaring that "truly transformative changes won't happen quickly and few – if any – will likely occur within the next 10 years," and that generative AI's ability to affect global productivity is low because "many of the tasks that humans currently perform...are multi-faceted and require real-world interaction, which AI won't be able to materially improve anytime soon."

What makes this interview – and really, this paper — so remarkable is how thoroughly and aggressively it attacks every bit of marketing collateral the AI movement has. Acemoglu specifically questions the belief that AI models will simply get more powerful as we throw more data and GPU capacity at them, and specifically ask a question: what does it mean to "double AI's capabilities"? How does that actually make something like, say, a customer service rep better?

And this is a specific problem with the AI fantasists' spiel. They heavily rely on the idea that not only will these large language models (LLMs) get more powerful, but that getting more powerful will somehow grant it the power to do...something. As Acemoglu says, "what does it mean to double AI's capabilities?" 

No, really, what does "more" actually mean? While one might argue that it'll mean faster generative processes, there really is no barometer for what "better" looks like, and perhaps that's why ChatGPT, Claude and other LLMs have yet to take a leap beyond being able to generate stuff. Anthropic's Claude LLM might be "best-in-class," but that only means that it's faster and more accurate, which is cool but not the future or revolutionary or even necessarily good.

I should add that these are the questions I – and other people writing about AI – should've been asking the whole time. Generative AI generates outputs based on text-based inputs and requests, requests that can be equally specific and intricate, yet the answer is always, as obvious as it sounds, generated fresh, meaning that there is no actual "knowledge" or, indeed, "intelligence" operating in any part of the process. As a result, it's easy to see how this gets better, but far, far harder – if not impossible – to see how generative AI leads any further than where we're already at. 

How does GPT – a transformer-based model that generates answers probabilistically (as in what the next part of the generation is most likely to be the correct one) based entirely on training data – do anything more than generate paragraphs of occasionally-accurate text? How do any of these models even differentiate when most of them are trained on the same training data that they're already running out of?

The training data crisis is one that doesn’t get enough attention, but it’s sufficiently dire that it has the potential to halt (or dramatically slow) any AI development in the near future. As one paper, published in the journal Computer Vision and Pattern Recognition, found, in order to achieve a linear improvement in model performance, you need an exponentially large amount of data. 

Or, put another way, each additional step becomes increasingly (and exponentially) more expensive to take. This infers a steep financial cost — not merely in just obtaining the data, but also the compute required to process it — with Anthropic CEO Dario Amodei saying that the AI models currently in development will cost as much as $1bn to train, and within three years we may see models that cost as much as “ten or a hundred billion” dollars, or roughly three times the GDP of Estonia.  

Acemoglu doubts that LLMs can become superintelligent, and that even his most conservative estimates of productivity gains "may turn out to be too large if AI models prove less successful in improving upon more complex tasks." And I think that's really the root of the problem. 

All of this excitement, every second of breathless hype has been built on this idea that the artificial intelligence industry – led by generative AI – will somehow revolutionize everything from robotics to the supply chain, despite the fact that generative AI is not actually going to solve these problems because it isn't built to do so. 

While Acemoglu has some positive things to say — for example, that AI models could be trained to help scientists conceive of and test new materials (which happened last year) — his general verdict is quite harsh: that using generative AI and "too much automation too soon could create bottlenecks and other problems for firms that no longer have the flexibility and trouble-shooting capabilities that human capital provides." In essence, replacing humans with AI might break everything if you're one of those bosses that doesn't actually know what the fuck it is they're talking about.


The report also includes a palette-cleanser for the quirked-up AI hype fiend on page 6, where Goldman Sachs' Joseph Briggs argues that generative AI will "likely lead to significant economic upside" based — and I shit you not — entirely on the idea that AI will replace workers in some jobs and then allow them to get jobs in other fields. Briggs also argues that "the full automation of AI exposed tasks that are likely to occur over a longer horizon could generate significant cost savings," which assumes that generative AI (or AI itself) will actually replace these tasks. 

I should also add that unlike every other interview in the report, Briggs continually mixes up AI and generative AI, and at one point suggests that "recent generative AI advances" are "foreshadowing the emergence of a "superintelligence."

I included this part of the report because sometimes — very rarely — I get somebody suggesting I'm not considering both sides. The reason I don't generally include both sides of this argument is that the AI hype side generally makes arguments based on the assumption that things will happen, such as a transformer model that probabilistically generates the next part of a sentence or a picture will somehow gain sentience.

Francois Chollet — an AI researcher at Google — recently argued that LLMs can't lead to AGI, explaining (in detail) that models like GPT are simply not capable of the kind of reasoning and theorizing that makes a human brain work. Chollet also notes that even models specifically built to complete the tasks of the Abstraction & Reasoning Corpus (a benchmark test for AI skills and true "intelligence") are only doing so because they've been fed millions of datapoints of people solving the test, which is kind of like measuring somebody's IQ based on them studying really hard to complete an IQ test, except even dumber.

The reason I'm suddenly bringing up superintelligences — or AGI (artificial general intelligence) — is because throughout every defense of generative AI is a deliberate attempt to get around the problem that generative AI doesn't really automate many tasks. While it's good at generating answers or creating things based on a request, there's no real interaction with the task, or the person giving it the task, or consideration of what the task needs at all — just the abstraction of "thing said" to "output generated." 

Tasks like taking someone's order and relaying it to the kitchen at a fast food restaurant might seem elementary to most people (I won't write easy, working in fast food sucks), but it isn't for an AI model that generates answers without really understanding the meaning of any of the words. Last year, Wendy's announced that it would integrate its generative "FreshAI" ordering system into some restaurants, and a few weeks ago it was revealed that the system requires human intervention on 14% of the orders.On Reddit, one user noted that Wendy's AI regularly required three attempts to get it to understand them, and would cut you off if you weren't speaking fast enough.

White Castle, which implemented a similar system in partnership with Samsung and SoundHound, fared little better, with 10% of orders requiring human intervention. Last month, McDonald’s discontinued its own AI ordering system — which it built with IBM and deployed to more than 100 restaurants — likely because it just wasn’t very good, with one customer rang up for literally hundreds of chicken nuggets. However, to be clear, McDonald’s system wasn’t based on generative AI.

If nothing else, this illustrates the disconnect between those building AI systems, and how much (or, rather, how little) they understand the jobs they wish to eliminate. A little humility goes a long way.      

Another thing to note is that, on top of generative AI cocking up these orders, Wendy's still requires human beings to make the goddamn food. Despite all of this hype, all of this media attention, all of this incredible investment, the supposed "innovations" don't even seem capable of replacing the jobs that they're meant to — not that I think they should, just that I'm tired of being told that this future is inevitable.

The reality is that generative AI isn't good at replacing jobs, but commoditizing distinct acts of labor, and, in the process, the early creative jobs that help people build portfolios to advance in their industries. 

The freelancers having their livelihoods replaced by bosses using generative AI aren't being "replaced" so much as they're being shown how little respect many bosses have for their craft, or for the customer it allegedly serves. Copy editors and concept artists provide far-more-valuable work than any generative AI can, yet an economy dominated by managers who don't appreciate (or participate in) labor means that these jobs are under assault from LLMs pumping out stuff that all looks and sounds the same to the point that copywriters are now being paid to help them sound more human.

One of the fundamental misunderstandings of the bosses replacing these workers with generative AI is that you are not just asking for a thing, but outsourcing the risk and responsibility. When I hire an artist to make a logo, my expectation is that they'll listen to me, then add their own flair, then we'll go back and forth with drafts until we have something I like. I'm paying them not just for their time, their years learning their craft and the output itself, but so that the ultimate burden of production is not my own, and their experience means that they can adapt to circumstances that I might not have thought of. These are not things that you can train in a dataset, because they're derived from experiences inside and outside of the creative process.

While one can "teach" a generative AI what a billion images look like, AI does not get hand cramps, or a call at 8PM saying that it "needs it to pop more." It does not have moods, nor can it infer them from written or visual media, because human emotions are extremely weird, as are our moods, our bodies, and our general existences. I realize all of this is a little flowery, but even the most mediocre copy ever written is, on some level, a collection of experiences. And fully replacing any creative is so very unlikely if you're doing so based on copying a million pieces of someone else's homework.


The most fascinating part of the report (page 10) is an interview with Jim Covello, Goldman Sachs' Head of Global Equity Research. Covello isn't a name you'll have heard unless you are, for whatever reason, a big semiconductor-head, but he's consistently been on the right side of history, named as the top semiconductor analyst by II Research for years, successfully catching the downturn in fundamentals in multiple major chip firms far before others did.

And Jim, in no uncertain terms, thinks that the generative AI bubble is full of shit.

Covello believes that the combined expenditure of all parts of the generative AI boom — data centers, utilities and applications — will cost a trillion dollars in the next several years alone, and asks one very simple question: "what trillion dollar problem will AI solve?" He notes that "replacing low-wage jobs with tremendously costly technology is basically the polar opposite of the prior technology transitions [he's] witnessed in the last thirty years."

One particular myth Covello dispels is comparing generative AI "to the early days of the internet," noting that "even in its infancy, the internet was a low-cost technology solution that enabled e-commerce to replace costly incumbent solutions," and that "AI technology is exceptionally expensive, and to justify those costs, the technology must be able to solve complex problems, which it isn't designed to do."

Covello also dismisses the suggestion that tech starts off expensive and gets cheaper over time as "revisionist history," and that "the tech world is too complacent in the assumption that AI costs will decline substantially over time." He specifically notes that the only reason that Moore's law was capable of enabling smaller, faster cheaper chips was because competitors like AMD forced Intel (and other companies to compete) — a thing that doesn't really seem to be happening with Nvidia, which has a near-stranglehold on the GPUs required to handle generative AI. 

While there are companies making GPUs aimed at the AI market (especially in China, where US trade restrictions prevent local companies from buying high-powered cards like the A100 for fears they’ll be diverted to the military), they're not doing so at the same scale, and Covello notes that "the market is too complacent about the certainty of cost declines." 

He also notes that the costs are so high that even if they were to come down, they'd have to do so dramatically, and that the comparison to the early days of the internet (where businesses often relied on $64,000 servers from Sun Microsystems and there was no AWS, Linode, or Azure) "pales in comparison" to the costs of AI, and that's even without including the replacement of the power grid, a necessity to keep this boom going.

I could probably write up Covello's entire interview, because it's nasty. Covello adds that the common adage that people didn't think smartphones would be big was false. He sat through hundreds of presentations in the early 2000s, many of them including roadmaps that accurately fit how smartphones rolled out, and that no such roadmap (or killer app) for AI has been found. 

He notes that big tech companies now have no choice but to engage in the AI arms race given the hype (which will continue the trend of massive spending), and he believes that there are "low odds of AI-related revenue expansion," in part because he doesn't believe that generative AI will make workers smarter, just more capable of finding better information faster, and that any advantages that generative AI gives you can be "arbitraged away" because the tech can be used everywhere, and thus you can't, as a company, raise prices.

In plain English: generative AI isn't making any money for anybody because it doesn't actually make companies that use it any extra money. Efficiency is useful, but it is not company-defining. He also adds that hyperscalers like Google and Microsoft will "also garner incremental revenue" from AI — not the huge returns they’re perhaps counting on, given their vast AI-related expenditure over the past two years.

This is damning for many reasons, chief of which is that the biggest thing that artificial intelligence is meant to do is be smart, and make you smarter. Being able to access information faster might make you better at your job, but that's efficiency rather than allowing you to do something new. Generative AI isn't creating new jobs, it isn't creating new ways to do your job, and it isn't making anybody any money — and the path to boosting revenues is unclear.

Covello ends with one important and brutal note: that the more time that passes without significant AI applications, the more challenging "the AI story will become," with corporate profitability likely floating this bubble as long as it takes for the tech industry to hit a more difficult economic period.

He also adds his own prediction — "investor enthusiasm may begin to fade" If "important use cases don't start to become more apparent in the next 12-18 months."

I think he's being optimistic.

While I won't recount the rest of the report, one theme brought up repeatedly is the idea that America's power grid is literally not ready for generative AI. In an interview former Microsoft VP of Energy Brian Janous (page 15), the report details numerous nightmarish problems that the growth of generative AI is causing to the power grid, such as:

  • Hyperscalers like Microsoft, Amazon and Google have increased their power demands from a few hundred megawatts in the early 2010s to a few gigawatts by 2030, enough to power multiple American cities.
  • The centralization of data center operations for multiple big tech companies in Northern Virginia may potentially require a doubling of grid capacity over the next decade.
  • Utilities have not experienced a period of load growth — as in a significant increase in power draw — in nearly 20 years, which is a problem because power infrastructure is slow to build and involves onerous permitting and bureaucratic measures to make sure it's done properly.
  • The total capacity of power projects waiting to connect to the grid grew 30% in the last year and wait times are 40-70 months.
  • Expanding the grid is "no easy or quick task," and that Mark Zuckerberg said that these power constraints are the biggest thing in the way of AI, which is... sort of true.

In essence, on top of generative AI not having any killer apps, not meaningfully increasing productivity or GDP, not generating any revenue, not creating new jobs or massively changing existing industries, it also requires America to totally rebuild its power grid, which Janous regrettably adds the US has kind of forgotten how to do.

Perhaps Sam Altman's energy breakthrough could be these fucking AI companies being made to pay for new power infrastructure.

The reason I so agonizingly picked apart this report is that if Goldman Sachs is saying this, things are very, very bad. It also directly attacks the specific hype-tactics of AI fanatics — the sense that generative AI will create new jobs (it hasn't in 18 months), the sense that costs will come down (they're haven’t, and there doesn't seem to be a path to them doing so in a way that matters), and that there's incredible demand for these products (there isn't, and there's no path to it existing). 

Even Goldman Sachs, when describing the efficiency benefits of AI, added that while it was able to create an AI that updated historical data in its company models more quickly than doing so manually, it cost six times as much to do so.


The remaining defense is also one of the most annoying — that OpenAI has something we don't know about. A big, sexy, secret technology that will eternally break the bones of every hater. 

Yet, I have a counterpoint: no it doesn't. 

Seriously, Mira Murati, CTO of OpenAI, said a few weeks ago that the models it has in its labs are not much more advanced than those that are publicly-available.

That's my answer to all of this. There is no magic trick. There is no secret thing that Sam Altman is going to reveal to us in a few months that makes me eat crow, or some magical tool that Microsoft or Google "pops out that makes all of this worth it.

There isn't. I'm telling you there isn't. 

Generative AI, as I said back in March, is peaking, if it hasn't already peaked. It cannot do much more than it is currently doing, other than doing more of it faster with some new inputs. It isn’t getting much more efficient. Sequoia hype-man David Cahn gleefully mentioned in a recent blog that Nvidia's B100 will "have 2.5x better performance for only 25% more cost," which doesn't mean a goddamn thing, because generative AI isn't going to gain sentience or intelligence and consciousness because it's able to run faster.

Generative AI is not going to become AGI, nor will it become the kind of artificial intelligence you've seen in science fiction. Ultra-smart assistants like Jarvis from Iron Man would require a form of consciousness that no technology currently — or may ever — have — which is the ability to both process and understand information flawlessly and make decisions based on experience, which, if I haven't been clear enough, are all entirely distinct things. 

Generative AI at best processes information when it trains on data, but at no point does it "learn" or "understand," because everything it's doing is based on ingesting training data and developing answers based on a mathematical sense or probability rather than any appreciation or comprehension of the material itself. LLMs are entirely different pieces of technology to that of "an artificial intelligence" in the sense that the AI bubble is hyping, and it's disgraceful that the AI industry has taken so much money and attention with such a flagrant, offensive lie.

The jobs market isn't going to change because of generative AI, because generative AI can't actually do many jobs, and it's mediocre at the few things that it's capable of doing. While it's a useful efficiency tool, said efficiency is based off of a technology that is extremely expensive, and I believe that at some point AI companies like Anthropic and OpenAI will have to increase prices — or begin to collapse under the weight of a technology that has no path to profitability.

If there were some secret way that this would all get fixed, wouldn't Microsoft, or Meta, or Google, or Amazon — whose CEO of AWS compared the generative AI hype to the Dotcom bubble in February — have taken advantage of it? And why am I hearing that OpenAI is already trying to raise another multi-billion dollar round after raising an indeterminate amount at an $80 billion valuation in February? Isn't its annualized revenue $3.4 billion? Why does it need more money?

I'll give you an educated guess: because whatever they — and other generative AI hucksters — have today is obviously, painfully not the future. Generative AI is not the future, but a regurgitation of the past, a useful-yet-not-groundbreaking way to quickly generate "new" data from old that costs far too much to make the compute and energy demands worth it. Google grew its emissions by 48% in the last five years chasing a technology that made its search engine even worse than it already is, with little to show for it.

It's genuinely remarkable how many people have been won over by this remarkable con — this unscrupulous manipulation of capital markets, the media and brainless executives disconnected from production — all thanks to a tech industry that's disconnected itself from building useful technology.

I've been asked a few times what I think will burst this bubble, and I maintain that part of the collapse will be investor dissent, punishing one of the major providers (Microsoft or Google, most likely) for a massive investment in an industry that produces little actual revenue. However, I think the collapse will be a succession of bad events — like Figma pausing its new AI feature after it immediately plagiarized Apple's weather app, likely as a result of training data that included it — crested by one large one, such as a major AI company like chatbot company Character.ai (which raised $150m in funding, and The Information claims might sell to one of the big tech companies) collapsing under the weight of an unsustainable business model built on unprofitable tech. 

Perhaps it's Cognition AI, the company that raised $175 million at a $2 billion valuation in April to make an "AI software engineer" that was so good that it had to fake a demo of it completing a software development project on Upwork.

Basically, there's going to be a moment that spooks a venture capital firm into pushing one of its startups to sell, or the sudden, unexpected yet very obvious collapse of a major player. For OpenAI and Anthropic, there really is no path to profitability — only one that includes burning further billions of dollars in the hope that they discover something, anything that might be truly innovative or indicative of the future, rather than further iterations of generative AI, which at best is an extremely expensive new way to process data.

I see no situation where OpenAI and Anthropic continue to iterate on Large Language Models in perpetuity, as at some point Microsoft, Amazon and Google decide (or are forced to decide) that cloud compute welfare isn't a business model. Without a real, tangible breakthrough — one that would require them to leave the world of LLMs entirely, in my opinion — it's unclear how generative AI companies can survive. 

Generative AI is locked in the Red Queen's Race, burning money to make money in an attempt to prove that they one day will make more money, despite there being no clear path to doing so.


I feel a little crazy every time I write one of these pieces, because it's patently ridiculous. Generative AI is unprofitable, unsustainable, and fundamentally limited in what it can do thanks to the fact that it's probabilistically generating an answer. It's been eighteen months since this bubble inflated, and since then very little has actually happened involving technology doing new stuff, just an iterative exploration of the very clear limits of what an AI model that generates answers can produce, with the answer being "something that is, at times, sort of good."

It's obvious. It's well-documented. Generative AI costs far too much, isn't getting cheaper, uses too much power, and doesn't do enough to justify its existence. There are no killer apps, and no killer apps on the horizon. And there are no answers. 

I don't know why more people aren't saying this as loudly as they can. I understand that big tech desperately needs this to be the next hypergrowth market as they haven't got any others, but pursuing this quasi-useful environmental disaster-causing cloud efficiency boondoggle will send shockwaves through the industry.

It's all a disgraceful waste.

Read the whole story
billyhopscotch
115 days ago
reply
Ed is usually right about AI, and his summary of the Goldman report covers it all.
Share this story
Delete

All Correlations Are Bastards

1 Share
This paper shows that shootings are predictable enough to be preventable. Using arrest and victimization records for almost 644,000 people from the Chicago Police Department, we train a machine learning model to predict the risk of being shot in the next 18 months. Out-of-sample accuracy is strikingly high: of the 500 people with the highest predicted risk, almost 13 percent are shot within 18 months, a rate 128 times higher than the average Chicagoan. A central concern is that algorithms may "bake in" bias found in police data, overestimating risk for people likelier to interact with police conditional on their behavior. We show that Black male victims more often have enough police contact to generate predictions. But those predictions are not, on average, inflated; the demographic composition of predicted and actual shooting victims is almost identical. There are legal, ethical, and practical barriers to using these predictions to target law enforcement. But using them to target social services could have enormous preventive benefits: predictive accuracy among the top 500 people justifies spending up to $134,400 per person for an intervention that could cut the probability of being shot by half. from Machine Learning Can Predict Shooting Victimization Well Enough to Help Prevent It [NBER; direct link to working paper (PDF)]
Read the whole story
billyhopscotch
116 days ago
reply
Share this story
Delete