BirAdam 27 minutes ago

I've been conflicted on AI/ML efforts for years. On one hand, the output of locally run inference is astounding. There are plenty of models on HuggingFace that I can run on my Mac Studio and provide real value to me every single work day. On the other hand, while I have the experience to evaluate the output, some of my younger colleagues do not. They are learning, and when I have time to help them, I certainly do, but I wish they just didn't have access to LLMs. LLMs are miracle tools in the right hands. They are dangerous conveniences in the wrong hands.

Wasted money is a totally different topic. If we view LLMs as a business opportunity, they haven't yet paid off. To imply, however, that a massive investment in GPUs is a waste seems flawed. GPUs are massively parallel compute. Were the AI market to collapse, we can imagine these GPUs being sold a severe discounts which would then likely spur some other technological innovation just as the crypto market laid the groundwork for ML/AI. When a resource gets cheap, more people gain access to it and innovation occurs. Things that were previously cost prohibitive become affordable.

So, whether or not we humans achieve AGI or make tons of money off of LLMs is somewhat irrelevant. The investment is creating goods of actual value even if those goods are currently overpriced, and should the currently intended use prove to be poor, a better and more lucrative use will be found in the event of an AI market crash.

Personally, I hope that the AGI effort is successful, and that we can all have a robot house keeper for $30k. I'd gladly trade one of the cars in my household to never do dishes, laundry, lawnmowing, or household repairs again just as I paid a few hundred to never have to vacuum my floors (though I actually still do once a month when I move furniture to vacuum places the Roomba can't go, a humanoid robot could do that for me).

  • philipwhiuk 23 minutes ago

    > On one hand, the output of locally run inference is astounding. There are plenty of models on HuggingFace that I can run on my Mac Studio and provide real value to me every single work day. On the other hand, while I have the experience to evaluate the output, some of my younger colleagues do not. They are learning, and when I have time to help them, I certainly do, but I wish they just didn't have access to LLMs. LLMs are miracle tools in the right hands. They are dangerous conveniences in the wrong hands.

    Is weird to me. Surely you recognise just as they don't know what they don't know (which is presumably the problem when it hallucinates), you must also have the same issue, there's just no old greybeard to wish you didn't have access.

    • BirAdam 14 minutes ago

      Well, I'm the graybeard (literally and metaphorically). I know enough not to blindly trust the LLM, and I know enough to test everything whether written by human or machine. This is not always true of younger professionals.

  • lisbbb 7 minutes ago

    I don't think so about the gpus. It's a sunk cost that won't be repurposed easily--just look at what happened to Nortel. Did all those PBXs get repurposed? Nope--trash. Those data centers are going to eat it hard, that's my prediction. It's not a terrible thing, per se--"we" printed trillions the past few years and those events need a sink to get rid of all the excess liquidity. It's usually a big war, but not always. Last time it was a housing bubble. Everyone was going to get rich on real estate, but not really. It was just an exercise in finding bag holders. That's what this AI/data center situation amounts to as well--companies had billions in cash sitting around doing nothing, might as well spend it. Berkshire has the same problem--hundreds of billions with nowhere to be productively invested. It doesn't sound like a problem but it is.

    My humble take on AGI is that we don't understand consciousness so how could we build something conscious except by accident? It seems like an extremely risky and foolish thing to attempt. Luckily, humans will fail at it.

highfrequency an hour ago

Per the author’s links, he warned that deep learning was hitting a wall in both 2018 and 2022. Now would be a reasonable time to look back and say “whoops, I was wrong about that.” Instead he seems to be doubling down.

  • tim333 an hour ago

    The author is a bit of a stopped clock that who has been saying deep learning is hitting a wall for years and I guess one day may be proved right?

    He probably makes quite good money as the go to guy for saying AI is rubbish? https://champions-speakers.co.uk/speaker-agent/gary-marcus

    • jvanderbot 21 minutes ago

      Well..... tbf. Each approach has hit a wall. It's just that we change things a bit and move around that wall?

      But that's certainly not a nuanced / trustworthy analysis of things unless you're a top tier researcher.

      • espadrine 12 minutes ago

        Indeed. A mouse that runs through a maze may be right to say that it is constantly hitting a wall, yet it makes constant progress.

        An example is citing Mr Sutskever's interview this way:

        > in my 2022 “Deep learning is hitting a wall” evaluation of LLMs, which explicitly argued that the Kaplan scaling laws would eventually reach a point of diminishing returns (as Sutskever just did)

        which is misleading, since Sutskever said it didn't hit a wall in 2022[0]:

        > Up until 2020, from 2012 to 2020, it was the age of research. Now, from 2020 to 2025, it was the age of scaling

        The larger point that Mr Marcus makes, though, is that the maze has no exit.

        > there are many reasons to doubt that LLMs will ever deliver the rewards that many people expected.

        That is something that most scientists disagree with. In fact the ongoing progress on LLMs has already accumulated tremendous utility which may already justify the investment.

        [0]: https://garymarcus.substack.com/p/a-trillion-dollars-is-a-te...

    • chii an hour ago

      a contrarian needs to keep spruiking the point, because if he relents, he loses the core audience that listened to him. That's why it's also the same with those who keep predicting market crashes etc.

    • JKCalhoun 29 minutes ago

      I thought the point though was that Sutskever is saying it too.

  • jayd16 10 minutes ago

    If something hits a wall and then takes a trillion dollars to move forward but it does move forward, I'm not sure I'd say it was just bluster.

  • Ukv an hour ago

    Even further back:

    > Yet deep learning may well be approaching a wall, much as I anticipated earlier, at beginning of the resurgence (Marcus, 2012)

    (From "Deep Learning: A Critical Appraisal")

  • bgwalter an hour ago

    Several OpenAI people said in 2023 that they were surprised by the acceptance of the public. Because they thought that LLMs were not so impressive.

    The public has now caught up with that view. Familiarity breeds contempt, in this case justifiably so.

    EDIT: It is interesting that in a submission about Sutskever essentially citing Sutskever is downvoted. You can do it here, but the whole of YouTube will still hate "AI".

    • Jyaif an hour ago

      > in this case justifiably so

      Oh please. What LLMs are doing now was complete and utter science fiction just 10 years ago (2015).

      • bgwalter 12 minutes ago

        Why would the public care what was possible in 2015? They see the results from 2023-2025 and aren't impressed, just like Sutskever.

      • lisbbb 4 minutes ago

        What exactly are they doing? I've seen a lot of hype but not much real change. It's like a different way to google for answers and some code generation tossed in, but it's not like LLMs are folding my laundry or mowing my lawn. They seem to be good at putting graphic artists out of work mainly because the public abides the miserable slop produced.

      • deadbabe 37 minutes ago

        Not really.

        Any fool could have anticipated the eventual result of transformer architecture if pursued to its maximum viable form.

        What is impressive is the massive scale of data collection and compute resources rolled out, and the amount of money pouring into all this.

        But 10 years ago, spammers were building simple little bots with markov chains to evade filters because their outputs sounded plausibly human enough. Not hard to see how a more advanced version of that could produce more useful outputs.

        • Workaccount2 29 minutes ago

          Any fool could have seen self driving cars coming in 2022. But that didn't happen. And still hasn't happened. But if it did happen, it would be easy to say

          "Any fool could have seen this coming in 2012 if they were paying attention to vision model improvements"

          Hindsight is 20/20.

          • lisbbb 2 minutes ago

            Everyone who lives in the show belt understands that unless a self driving car can navigate icy, snow-covered roads better than humans can, it's a non-starter. And the car can't just "pull over because it's too dangerous" that doesn't work at all.

        • free_bip 33 minutes ago

          I guess I'm worse than a fool then, because I thought it was totally impossible 10 years ago.

  • otabdeveloper4 an hour ago

    > learning was hitting a wall in both 2018 and 2022

    He wasn't wrong though.

JKCalhoun 30 minutes ago

Interesting to me, during that crazy period when Sutskever ultimately ended up leaving OpenAI, I thought perhaps he had shot himself in the foot to some degree (not that I have any insider information—just playing stupid observer from the outside).

The feeling I have now is that it was a fine decision for him to have made. It made a point at the time, perhaps moral, perhaps political. And now it seems, despite whatever cost there was for him at the time, the "golden years" of OpenAI (and LLM's in general) may have been over anyway.

To be sure, I happen to believe there is a lot of mileage for LLMs even in their current state—a lot of use-cases, integration we have yet to explore. But Sutskever I assume is a researcher and not a plumber—for him the LLM was probably over.

One wonders how long before one of these "break throughs". On one hand, they may come about serendipitously, and serendipity has no schedule. It harkens back to when A.I. itself was always "a decade away". You know, since the 1950's or so.

On the other hand, there are a lot more eyeballs on AI these days than there ever were in Minsky's* day.

(*Hate to even mention the man's name these days.)

roenxi 2 hours ago

Just because something didn't work out doesn't mean it was a waste, and it isn't particularly clear that the the LLM boom was wasted, or that it is over, or that it isn't working. I can't figure out what people mean when they say "AGI" any more, we appear to be past that. We've got something that seems to be general and seems to be more intelligent than an average human. Apparently AGI means a sort of Einstein-Tolstoy-Jesus hybrid that can ride a unicycle and is far beyond the reach of most people I know.

Also, if anyone wants to know what a real effort to waste a trillion dollars can buy ... https://costsofwar.watson.brown.edu/

  • austin-cheney 2 hours ago

    > Just because something didn't work out doesn't mean it was a waste

    Its all about scale.

    If you spend $100 on something that didn't work out that money wasn't wasted if you learned something amazing. If you spend $1,000,000,000,000 on something that didn't work out the expectation is that you learn something close to 1,000,000,000x more than the $100 spend. If the value of learning is several orders of magnitude less than the level of investment there is absolutely tremendous waste.

    For example: nobody qualifies spending a billion dollars on a failed project as value if your learning only resulted in avoiding future paper cuts.

    • lisbbb a minute ago

      It's not waste, it's a way to get rid of excess liquidity caused by massive money printing operations.

  • Deegy 2 hours ago

    We currently have human-in-the-loop AGI.

    While it doesn't seem we can agree on a meaning for AGI, I think a lot of people think of it as an intelligent entity that has 100% agency.

    Currently we need to direct LLM's from task to task. They don't yet posses the capability of full real world context.

    This is why I get confused when people talk about AI replacing jobs. It can replace work, but you still need skilled workers to guide them. To me, this could result in humans being even more valuable to businesses, and result in an even greater demand for labor.

    If this is true, individuals need to race to learn how to use AI and use it well.

    • vidarh an hour ago

      > Currently we need to direct LLM's from task to task.

      Agent-loops that can work from larger scale goals work just fine. We can't letting them run with no oversight, but we certainly also don't need to micro-manage every task. Most days I'll have 3-4 agent-loops running in parallel, executing whole plans, that I only check in on occasionally.

      I still need to review their output occasionally, but I certianly don't direct them task to task.

      I do agree with you we still need skilled workers to guide them, so I don't think we necessarily disagree all that much, but we're past the point where they need to be micromanaged.

    • gortok 30 minutes ago

      If we can't agree on a definition of AGI, then what good is it to say we have "human-in-the-loop AGI"? The only folks that will agree with you will be using your definition of AGI, which you haven't shared (at least in this posting). So, what is your definition of AGI?

  • getnormality an hour ago

    AI capabilities today are jagged and people look at what they want to.

    Boosters: it can answer PhD-level questions and it helps me a lot with my software projects.

    Detractors: it can't learn to do a task it doesn't already know how to do.

    Boosters: But actually it can actually sometimes do things it wouldn't be able to do otherwise if you give it lots of context and instructions.

    Detractors: I want it to be able to actually figure out and retain the context itself, without being given detailed instructions every time, and do so reliably.

    Boosters: But look, in this specific case it sort of does that.

    Detractors: But not in my case.

    Boosters: you're just using it wrong. There must be something wrong with your prompting strategy or how you manage context.

    etc etc etc...

  • bryanlarsen an hour ago

    AFAICT "AGI" is a placeholder for peoples fears and hopes for massive change caused by AI. The singularity, massive job displacement, et cetera.

    None of this is a binary, though. We already have AGI that is superhuman in some ways and subhuman in others. We are already using LLM's to help improve themselves. We already have job displacement.

    That continuum is going to continue. AI will become more superhuman in some ways, but likely stay subhuman in others. LLM's will help improve themselves. Job displacement will increase.

    Thus the question is whether this rate of change will be fast or slow. Seems mundane, but it's a big deal. Humans can adapt to slow changes, but not so well to fast ones. Thus AGI is a big deal, even if it's a crap stand in for the things people care about.

  • orwin 2 hours ago

    > Just because something didn't work out doesn't mean it was a waste

    Here i think it's more about opportunity cost.

    > I can't figure out what people mean when they say "AGI" any more, we appear to be past that

    What i ask of an AGI is to not hallucinate idiotic stuff. I don't care about being bullshitted too much if the bullshit is logic, but when i ask "fix mypy errors using pydantic" and instead of declaring a type for a variable it invent weird algorithms that make no sense and don't work (and the fix would have taken 5 minutes for any average dev).I mean, Claude 4.5 and Codex have replaced my sed/search and replaces, write my sanity tests, write my commit comment, write my migration scripts (and most of my scripts), and make refactor so easy i now do one refactor every month or so, but if it is AGI, i _really_ wonder what people mean by intelligence.

    > Also, if anyone wants to know what a real effort to waste a trillion dollars can buy

    100% agree. Pleas Altman, Ilya and other, i will hapilly let you use whatever money you want if that money is taken from war profiteers and warmongers.

  • embedding-shape 2 hours ago

    > Just because something didn't work out doesn't mean it was a waste

    One thing to keep in mind, is that most of these people who go around spreading unfounded criticism of LLMs, "Gen-AI" and just generally AI aren't usually very deep into understanding computer science, and even less science itself. In their mind, if someone does an experiment, and it doesn't pan out, they'll assume that means "science itself failed", because they literally don't know how research and science work in practice.

    • bbor 2 hours ago

      Maybe true in general, but Gary Marcus is an experienced researcher and entrepreneur who’s been writing about AI for literally decades.

      I’m quite critical, but I think we have to grant that he has plenty of credentials and understands the technical nature of what he’s critiquing quite well!

  • pdimitar an hour ago

    Eh, tearing down a straw man is not an impressive argument from you either.

    As a counter-point, LLMs still do embarrassing amounts of hallucinations, some of which are quite hilarious. When that is gone and it starts doing web searches -- or it has any mechanisms that mimic actual research when it does not know something -- then the agents will be much closer to whatever most people imagine AGI to be.

    Have LLMs learned to say "I don't know" yet?

    • in-silico 15 minutes ago

      > When that is gone and it starts doing web searches -- or it has any mechanisms that mimic actual research when it does not know something

      ChatGPT and Gemini (and maybe others) can already perform and cite web searches, and it vastly improves their performance. ChatGPT is particularly impressive at multi-step web research. I have also witnessed them saying "I can't find the information you want" instead of hallucinating.

      It's not perfect yet, but it's definitely climbing human percentiles in terms of reliability.

      I think a lot of LLM detractors are still thinking of 2023-era ChatGPT. If everyone tried the most recent pro-level models with all the bells and whistles then I think there would be a lot less disagreement.

      • pdimitar 12 minutes ago

        Well please don't include me in some group of Luddites or something.

        I use the mainstream LLMs and I've noted them improving. They have ways to go still.

        I was objecting to my parent poster's implication that we have AGI. However muddy that definition is, I don't feel like we do have that.

turlockmike an hour ago

I believe in a very practical definition of AGI. AGI is a system capable of RSI. Why? Because it mimics humans. We have some behaviours that are given to us from birth, but the real power of humans is our ability to learn and improve ourselves and the environment around us.

A system capable of self improvement will be sufficient for AGI imo.

  • Retric an hour ago

    Self improvement doesn’t mean self improvement in any possible direction without any tradeoffs. Genetic algorithms can do everything an LLM can given enough computational resources and training, but being wildly inefficient humanity can’t actually use them to make a chatbot on any even vaguely relevant timeline.

  • tim333 31 minutes ago

    Ah - recursive self improvement. I was thinking repetitive strain injury was odd. But that's probably quite a good test although LLMs may be able to improve a bit but still not be very good. An interesting point for me is if all humans went away could the AI/robots keep on without us which would require them to be able to maintain and build power plants, chip fabs and the like. A way to go on that one.

andix an hour ago

There was a lot of talk about reaching "peak AI" in early summer of this year.

I guess there is some truth to it. The last big improvement to LLMs was reasoning. It gave the existing models additional capabilities (after some re-training).

We've reached the plateau of tiny incremental updates. Like with smartphones. I sometime still use an iPhone 6s. There is no fundamental difference compared to the most current iPhone generation 10 years later. The 6s is still able to perform most of the tasks you need a smartphone to do. The new ones do it much faster, and everything works better, but the changes are not disrupting at all.

elif 14 minutes ago

That's like saying a trillion dollars was potentially wasted sending men to the moon. You have to close your eyes to so much obvious progress and dissect your idea beyond recognition to start believing this thesis.

nayroclade 20 minutes ago

The core argument here, as far as I can discern it, seems to be: A trillion dollars has been spent scaling LLMs in an attempt to create AGI. Since scaling alone looks like it won't produce AGI, that money has been wasted.

This is a frankly bizarre argument. Firstly, it presupposes that _only_ way AI becomes useful is if turns into AGI. But that isn't true: Existing LLMs can do a variety of economically valuable tasks, such as coding, even when not being AGI. Perhaps the economic worth of non-AGI will never equal what it costs to build an operate it, but it seems way too early to make that judgement and declare any non-AGI AI as worthless.

Secondly, even if scaling alone won't reach AGI, that doesn't mean that you can reach AGI _without_ scaling. Even when new and better architectures are developed, it still seems likely that, between two models with an equivalent architecture, the one with more data and compute research will be more powerful. And waiting for better architectures before you try to scale means you will never start. 50 years from now, researchers will have much better architectures. Does that mean we should wait 50 years before trying to scale them? How about 100 years? At what point do you say, we're never going to discover anything better, so now we can try scaling?

wolttam 27 minutes ago

I really struggle to come up with a reason that transformers won't continue to deliver on additional capabilities that get fit into the training set.

dustingetz an hour ago

companies are already wasting majority fractions of their engineering labor spend on coordination costs and fake work, through that lens i have trouble making an argument that any of this matters. Which is why they are able to do it. I’m reminded of an old essay arguing that the reason Google spends so lavishly is because if they only spent what they needed, they would appear so extraordinarily profitable that the government would intervene.

avocadosword an hour ago

Don't research computations also require substantial hardware?

strangescript an hour ago

LLMs write all my code now and I just have to review it. Not only has my output 3x'ed at least, I also have zero hesitations now tackling large refactors, or tracking down strange bugs. For example, I recently received a report there was some minor unicode related data corruption in some of our doc in our DBs. It was cosmetic, and low priority, also not a simple task to track down traditionally. But now I just put [llm agent on it, to avoid people accusing me of promoting] on it. It found 3 instances of the corruption across hundreds of documents and fixed them.

I am sure some of you are thinking "that is all slop code". It definitely can be if you don't do your due diligence in review. We have definitely seen a bifurcation of devs who do that, and those who don't, where I am currently working.

But by far the biggest gain is my mental battery is far less drained at the end of the day. No task feels soul crushing anymore.

Personally, coding agents are the greatest invention of my lifetime outside the emergence of the internet.

ComplexSystems 2 hours ago

I think the article makes decent points but I don't agree with the general conclusion here, which is that all of this investment is wasted unless it "reaches AGI." Maybe it isn't necessary for every single dollar we spend on AI/LLM products and services to go exclusively toward the goal of "reaching AGI?" Perhaps it's alright if these dollars instead go to building out useful services and applications based on the LLM technologies we already have.

The author, for whatever reason, views it as a foregone conclusion that every dollar spent in this way is a waste of time and resources, but I wouldn't view any of that as wasted investment at all. It isn't any different from any other trend - by this logic, we may as well view the cloud/SaaS craze of the last decade as a waste of time. After all, the last decade was also fueled by lots of unprofitable companies, speculative investment and so on, and failed to reach any pie-in-the-sky Renaissance-level civilization-altering outcome. Was it all a waste of time?

It's ultimately just another thing industry is doing as demand keeps evolving. There is demand for building the current AI stack out, and demand for improving it. None of it seems wasted.

  • an0malous 2 hours ago

    That’s not what he’s saying, the investors are the ones who have put trillions of dollars into this technology on the premise that it will achieve AGI. People like Sam Altman and Marc Andreesen have been going into podcasts saying AGI is imminent and they’re going to automate every job.

    The author did not say every dollar was wasted, he said that LLMs will never meet the current investment returns.

    It’s very frustrating to see comments like this attacking strawmans and setting up Motte and Bailey arguments every time there’s AI criticism. “Oh but LLMs are still useful” and “Even if LLMs can’t achieve AGI we’ll figure out something that will eventually.” Yes but that isn’t what Sam and Andreesen and all these VCs have been saying, and now the entire US economy is a big gamble on a technology that doesn’t deliver what they said it would and because the admin is so cozy with VCs we’re probably all going to suffer for the mistakes of a handful of investors who got blinded by dollar signs in their eyes.

    • ComplexSystems an hour ago

      The author quite literally says that the last few years were a "detour" that has wasted a trillion dollars. He explicitly lists building new LLMs, building larger LLMs and scaling LLMs as the problem and source of the waste. So I don't think I am strawmanning his position at all.

      It is one thing to say that OpenAI has overpromised on revenues in the short term and another to say that the entire experiment was a waste of time because it hasn't led to AGI, which is quite literally the stance that Marcus has taken in this article.

      • an0malous an hour ago

        > The author, for whatever reason, views it as a foregone conclusion that every dollar spent in this way is a waste of time and resources

        This is a strawman, the author at no point says that “every dollar is a waste.”

        • ComplexSystems 31 minutes ago

          He quite literally says that the dollars spent on scaling LLMs in the past few years are a waste.

    • dist-epoch an hour ago

      You are making the same strawman attack you are criticising.

      The dollars invested are not justified considering TODAYs revenues.

      Just like 2 years ago people said NVIDIA stock prices was not justified and a massive bubble considering the revenue from those days. But NVIDIA revenues 10xed, and now the stock price from 2 years ago looks seriously underpriced and a bargain.

      You are assuming LLM revenues will remain flat or increase moderately and not explode.

      • an0malous an hour ago

        You seem like someone who might be interested in my nuclear fusion startup. Right now all we have is a bucket of water but in five years that bucket is going to power the state of California.

  • robot-wrangler 2 hours ago

    It's not about "every dollar spent" being a waste of time, it's about acknowledging the reality of opportunity cost. Of course, no one in any movement is likely to listen to their detractors, but in this case the pioneers seem to agree.

    https://www.youtube.com/watch?v=DtePicx_kFY https://www.bbc.com/news/articles/cy7e7mj0jmro

    • ComplexSystems 2 hours ago

      I think there is broad agreement that new models and architectures are needed, but I don't see it as a waste to also scale the stack that we currently have. That's what Silicon Valley has been doing for the past 50 years - scaling things out while inventing the next set of things - and I don't see this as any different. Maybe current architectures will go the way of the floppy disk, but it wasn't a waste to scale up production of floppy disk drives while they were relevant. And ChatGPT was still released only 3 years ago!

      • vidarh an hour ago

        And notably, Marcus has been banging this drum for years. Even this article points back to articles he wrote years ago suggesting deep learning was hitting the wall... With GPT 3....

        It's sour grapes because the methods he prefers have not gotten the same attention (hah...) or funding.

        He's continuing to push the ludicrous Apple "reasoning paper" that he described as a "knockout blow for LLMs" even though it was nothing of the sort.

        With each of his articles, I usually lose more respect for him.

d--b an hour ago

Well those chips and power plants might still be useful for what comes after.

If we find AGI needs a different chip architecture, yeah, LLMs would have been quite a waste.

tqwhite 2 hours ago

Did someone say that LLM was the final solution while I wasn’t listening? Am I fantasizing the huge outcry about the terrible danger of AGI? Are people not finding ways to use the current levels of LLM all over the place?

The idea that the trillions are a waste is not exactly fresh. The economic model is still not clear. Alarmists have been shrill and omnipresent. Bankruptcy might be the future of everyone.

But, will we look up one day and say, “Ah never mind” about GPT, Claude, et al? Fat chance. Will no one find a use for a ton of extra compute? I’m pretty sure.

I don’t much dispute any of the facts I skimmed off the article but the conclusion is dumb.

  • Workaccount2 17 minutes ago

    Ironically that MIT study that made the rounds a few months ago ("Study finds 90% of AI pilots fail" you remember), also found that virtually every single worker at every company they studied was using LLMs regularly.

    The real takeaway of the study was that workers were using their personal LLM accounts to do work rather than using the AI implementation mess their companies had shat out.

  • tim333 28 minutes ago

    Personally I think we'll find something better than the LLM algorithm fairly soon, but it will still be using the same GPU type servers.

  • beepbooptheory an hour ago

    If bankruptcy does happen to be the future for everyone, then yes, I think there is going to be a lot of "ah never mind"s going around.

    If all this went away tomorrow, what would we do with all the compute? Its not exactly general purpose infrastructure thats being built.

    • pdimitar an hour ago

      My hypothesis is that general computing frameworks are the next big thing. The powerful GPUs have been mostly black boxes for way too long. A lot of clever people will not want to just throw them away or sell them second-hand and will try to find better ways to utilize them.

      I might very well be super wrong. F.ex. NVIDIA is guarding their secrets very well and we have no reason to believe they'll suddenly drop the ball. But it does make me think; IMO a truly general GPU (and open + free) compute has been our area's blind spot for way too long.

    • tim333 an hour ago

      Some of the participants may go bust but I very much doubt the highly profitable ones like Google, Apple, Nvidia and Microsoft will. There'll be enough demand for existing LLMs to keep the servers busy. Just writing code which works currently is probably enough to justify a fair chunk of the capacity.

    • lionkor an hour ago

      Could always mine crypto.

mensetmanusman 2 hours ago

I’m glad the 0.01% have something to burn their money on.

  • PrairieFire 2 hours ago

    To further your point - I mean honestly if this all ends up being an actual bubble that doesn’t manifest a financial return for the liquidity injectors but instead a massive loss (for the .01% who are in large part putting the cash in), did humanity actually lose?

    If it pops it might end up being looked at in the lens of history as one of the largest backdoor/proxy wealth redistributions ever. The capex being spent is in large part going to fund the labor of the unwashed masses, and society is getting the individual productivity and efficiency benefits from the end result models.

    I’m particularly thankful for the plethora of open source models I have access to thanks to all this.

    I, individually, have realized indisputable substantial benefits from having these tools at my disposal every day. If the whole thing pops, these tools are safely in my possession and I’m better because I have them. Thanks .01%!!

    (the reality is I don’t think it will pop in the classic sense, and these days it seems the .01 can never lose. either way, the $1tn can’t be labeled as a waste).

  • teraflop 2 hours ago

    It would be nice if they could burn it on something that didn't require them to buy up the world's supply of DDR5 RAM, and triple prices for everyone else.

    https://pcpartpicker.com/trends/price/memory/

    • williamdclt 2 hours ago

      that might be literally the least of my concern regarding gen AI in today's world

skippyboxedhero 41 minutes ago

Every technological change has been accompanied by an investment boom that resulted in some degree of wasted investment: cars, electricity, mass production of bicycles, it goes on and on.

One point about this is that humans appear unable to understand that this is an efficient outcome because investment booms are a product of uncertainty around the nature of the technological change. You are building something is literally completely new, no-one had any idea what cars consumers would buy so lots of companies started to try and work out that out and that consolidated into competition on cost/scale once that became clear. There is no way to go to the end of that process, there are many people outside the sphere of business who are heavily incentivized to say that we (meaning bureaucrats and regulators) actually know what kind of cars consumers wanted and that all the investment was just a waste.

Another point is that technological change is very politically disruptive. This was a point that wasn't well appreciated...but is hopefully clear with social media. There are a large number of similar situations in history though: printing press, newspapers, etc. Technological change is extremely dangerous if you are a politician or regulator because it results in your power decreasing and, potentially, your job being lost. Again, the incentives are huge.

The other bizarre irony of this is that people will look at an investment boom with no technological change, that was a response to government intervention in financial markets and a malfunctioning supply-side economy...and the response was: all forms of technical innovation are destabilizing, investment booms are very dangerous, etc. When what they mean is corporations with good political connections might lose money.

This is also linked inherently to the view around inflation. The 1870s are regarded as one of the most economically catastrophic periods in economic history by modern interpretations of politics. Let me repeat this in another way: productivity growth was increasing by 8-10%/year, you saw mind-boggling gains from automation (one example is cigarettes, iirc it took one skilled person 10-20 minutes to create a cigarette, a machine was able to produce hundreds in a minute), and conventional macroeconomics views this as bad because...if you can believe it...they argue that price declines lead to declines in investment. Now compare to today: prices continue to rise, investment is (largely) non-existent, shortages in every sector. Would you build a factory in 1870 knowing you could cut prices for output by 95% and produce more? The way we view investment is inextricably linked in economic policy to this point of view, and is why the central banks have spent trillions buying bonds with, in most cases, zero impact on real investment (depending on what you mean, as I say above, private equity and other politically connected incumbents have made out like bandits...through the cycle, the welfare gain from this is likely negative).

You see the result of this all over the Western world: shortages of everything, prices sky-high, and when technological change happens the hysteria around investment being wasteful and disruptive. It would be funny if we didn't already see the issues with this path all around us.

It is not wasted, we need more of this, this ex-post, academic-style reasoning of everything in hindsight gets us nowhere. There is no collateral damage, even in the completely fake Fed-engineered housing bubble, the apparently catastrophic cost was: more houses, and some wealthy people lost their net worth (before some central bankers found out their decisions in 03-04 caused wealthy people to lose money, and quickly set about recapitalising their brokerage accounts with taxpayers money).

moralestapia an hour ago

Nothing new here, just nepo as old as time.

Perhaps the scale is unprecedented, or it's always been like this it's just much less concealed these days.

Absolute retards can waste trillions of dollars on stupid ideas, because they're in the in group. Next door someone who's worked their whole life gets evicted because their mortgage is now way more of what they make in salary.

Sucks to be in the out group!

bbor 2 hours ago

I always love a Marcus hot take, but this one is more infuriating than usual. He’s taking all these prominent engineers saying “we need new techniques to build upon the massive, unexpected success we’ve had”, twisting it into “LLMs were never a success and sucked all along”, and listing them alongside people that no one should be taking seriously — namely, Emily Bender and Ed Zitron.

Of course, he includes enough weasel phrases that you could never nail him down on any particular negative sentiment; LLMs aren’t bad, they just need to be “complemented”. But even if we didn’t have context, the whole thesis of the piece runs completely counter to this — you don’t “waste” a trillion dollars on something that just needs to be complemented!

FWIW, I totally agree with his more mundane philosophical points about the need to finally unify the work of the Scruffies and the Neats. The problem is that he frames it like some rare insight that he and his fellow rebels found, rather than something that was being articulated in depth by one of the fields main leaders 35 years ago[1]. Every one of the tens of thousands of people currently working on “agential” AI knows it too, even if they don’t have the academic background to articulate it.

I look forward to the day when Mr. Marcus can feel like he’s sufficiently won, and thus get back to collaborating with the rest of us… This level of vitriolic, sustained cynicism is just antithetical to the scientific method at this point. It is a social practice, after all!

[1] https://www.mit.edu/~dxh/marvin/web.media.mit.edu/~minsky/pa...

Insanity an hour ago

“He is not forecasting a bright future for LLMs”.

Yeah, no shit. I’ve been saying this since the day GPT 3 became hyped. I don’t think many with a CS background are buying the “snake oil” of AGI through stochastic parrots.

At some point, even people who hype LLMs will spin their narrative to not look out of touch with reality. Or not more out of touch than is acceptable lol.

naveen99 2 hours ago

When it comes to machine learning, research has consistently shown, that pretty much the only thing that matters is scaling.

Ilya should just enjoy his billions raised with no strings.

  • CuriouslyC 2 hours ago

    If you think scaling is all that matters, you need to learn more about ML.

    Read about the the No Free Lunch Theorem. Basically, the reason we need to "scale" so hard is because we're building models that we want to be good at everything. We could build models that are as good at LLMs at a narrow fraction of tasks we ask of them to do, at probably 1/10th the parameters.

  • philipwhiuk 2 hours ago

    > When it comes to machine learning, research has consistently shown, that pretty much the only thing that matters is scaling.

    Yes, indeed, that is why all we have done since the 90s is scale up the 'expert systems' we invented ...

    That's such an a-historic take it's crazy.

    * 1966: failure of machine translation

    * 1969: criticism of perceptrons (early, single-layer artificial neural networks)

    * 1971–75: DARPA's frustration with the Speech Understanding Research program at Carnegie Mellon University

    * 1973: large decrease in AI research in the United Kingdom in response to the Lighthill report

    * 1973–74: DARPA's cutbacks to academic AI research in general

    * 1987: collapse of the LISP machine market

    * 1988: cancellation of new spending on AI by the Strategic Computing Initiative

    * 1990s: many expert systems were abandoned

    * 1990s: end of the Fifth Generation computer project's original goals

    Time and time again, we have seen that each academic research begets a degree of progress, improved by the application of hardware and money, but ultimately only a step towards AGI, which ends with a realisation that there's a missing congitive ability that can't be overcome by absurd compute.

    LLMs are not the final step.

    • bbor 2 hours ago

      Well, expert systems aren’t machine learning, they’re symbolic. You mention perceptrons, but that timeline is proof for the power of scaling, not against — they didn’t start to really work until we built giant computers in the ~90s, and have been revolutionizing the field ever since.

  • an0malous 2 hours ago

    Didn’t OpenAI themselves publish a papers years ago that scaling parameters has diminishing returns?