Brian Lovin
/
Hacker News
Daily Digest email

Get the top HN stories in your inbox every day.

thanhhaimai

Opinions are my own.

I think the biggest winner of this might be Google. Virtually all the frontier AI labs use TPU. The only one that doesn't use TPU is OpenAI due to the exclusive deal with Microsoft. Given the newly launched Gen 8 TPU this month, it's likely OpenAI will contemplate using TPU too.

bastawhiz

Many labs use TPUs, but not exclusively. Most labs need more compute than they can get, and if there's TPU capacity, they'll adapt their systems to be able to run partially on TPUs.

gpt5

Why is AMD not more popular then if labs are so flexibly with giving away CUDA?

mattnewton

people are trying, especially for inference. For training, it’s just too high risk to tank your training I think.

TPUs are at least dogfooded by Google deepmind, no team AFAIK has gotten the AMD stack to train well.

zombiwoof

[dead]

zobzu

even google doesnt only use TPUs.

danpalmer

Google is in a different position to others in that they're the only frontier lab with a cloud infra business. It obviously makes sense to sell GPUs on cloud infra as people want to rent them. In that respect Google buys a ton of GPUs to rent out.

What's unclear to me is how much Google uses GPUs for their own stuff. Yes Gemini runs on GPUs now, so that Google can sell Gemini on-prem boxes (recent release announced last week), but is any training or inference for Gemini really happening on GPUs? This is unclear to me. I'd have guessed not given that I thought TPUs were much cheaper to operate, but maybe I'm wrong.

Caveat, I work at Google, but not on anything to do with this. I'm only going on what's in the press for this stuff.

maxclark

And almost by happenstance Apple. Turns out they have a great platform for inference and torched almost nothing comparatively on Siri. The Apple/Gemini deal is interesting, Google continues to demonstrate their willingness to degrade their experience on Apple to try and force people to switch.

ttul

If you do the math (I did), in 2 years, open source models that you can run on a future MacBook Pro will be as capable as the frontier cloud models are today. Memory bandwidth is growing rapidly, as is the die area dedicated to the neural cores. And all the while, we have the silicon getting more power efficient and increasingly dense (as it always does). These hardware improvements are coming along as the open source models improve through research advancements. And while the cloud models will always be better (because they can make use of as much power as they want to - up in the cloud), what matters to most of us is whether a model can do a meaningful share of knowledge work for us. At the same time, energy consumption to run cloud infrastructure is out-pacing the creation of new energy supply, which is a problem not easily solved. I believe scarcity of energy will increasingly drive frontier labs toward power efficiency, which necessarily implies that the Pareto frontier of performance between cloud and local execution will narrow.

nl

A Opus 4.7/Gpt5.5 class model is 5 trillion parameters[1].

To run a 8 bit quantized version of that you need roughly 5TB of RAM.

Today that is around 18 NVidia B300. That's around $900,000, without including the computers to run them in.

It's true that the capability of open source models is improving, but running actual frontier models on your MPB seems a way off.

[1] https://x.com/elonmusk/status/2042123561666855235?s=20 (and Elon has hired enough people out of those labs to have a fair idea)

npunt

I did this calculation a bit ago and don't think frontier models are just a few MacBook Pro generations away. Yes numbers reliably go up in tech in general but in specific semiconductors & standards have long lead-times and published roadmaps, so we can have high confidence in what we're getting even in 3-4 years in terms of both transistor density and RAM speeds.

In mid-2028 we have N2E/N2P with around 15% greater transistor density than today's N3P, and by EOY2028 we'll likely have A14 with about 35-40% density improvement.

Meanwhile, we'll be on LPDDR6 by that point, which takes M-series Pros from 307GB/s -> ~400GB/s, and Max's from 614GB/s -> ~800GB/s.

Model improvements obviously will help out, but on the raw hardware front these aren't in the ballpark for frontier model numbers. An H100 has 3TB/s memory bandwidth, fwiw

xorcist

That's not "math". That's a "wild guess", or baseless extrapolation at best.

CMay

So long as you don't require deep search grounding like massive web indexes or document stores which are hard to reproduce locally. You can do local agentic things that get close or even do better depending on search strategy, but theoretically a massive cloud service with huge data stores at hand should be able to produce better results.

In practice unless you're doing some kind of deep research thing with the cloud, it'll try to optimize mostly for time and get you a good enough answer rather than spending an hour or two. An hour of cloud searching with huge data stores is not equivalent to an hour of local agentic searching, presumably.

I think that problem will improve a little in the coming years as we kind of create optimized data curation, but the information world will keep growing so the advantage will likely remain with centralized services as long as they offer their complete potential rather than a fraction.

dualvariable

Also, all the cloud models don't have to be the best frontier models, and you don't need to focus on hitting the benchmark of shrinking Opus 4.7 down to a single MBP to make significant improvements. If you get it so that an Opus 4.7 benchmark-compatible model can run in $250k of datacenter capex (and associated reduced opex for power+cooling) that'd be a massive cost improvement that makes the cloud models cheaper. And for most consumers that'll probably be good enough. You don't need to run on a $5k laptop to make a big difference.

rc1

Show your working / explain your math?

GorbachevyChase

They also degrade their own direct services with little warning or thought put into change management, so, to be fair, Apple may be getting the same quality of service as the rest of us.

vharish

I think that's just how Google is, by nature. They don't intentionally degrade their services. They just aren't a customer centric company. They run on numbers. As a corporate, it doesn't really encourage support and maintenance work either.

manueltgomes

Indeed. I'm wondering if Apple's "miss the train" with AI ended up being a blessing for them. Not only in the Google deal but also there's a lot of people doing interesting stuff locally..

bigyabai

Apple is basically in the same boat as AMD and Intel. They have a weak, raster-focused GPU architecture that doesn't scale to 100B+ inference workloads and especially struggles with large context prefill. TPUs smoke them on inference, and Nvidia hardware is far-and-away more efficient for training.

hellohello2

What do TPUs do to improve on GPUs at inference?

brcmthrowaway

This doesn't get talked about enough - the GPU is weak, weak, weak. And anyone who can fix them will go to a serious AI company (for 2-3x the salary).

munk-a

Apple is in a much better boat than AMD or Intel. They have a gigantic warchest and can just snap up whoever looks like a leader coming out of the bubble burst.

undefined

[deleted]

celeritascelery

Maybe I am missing something here, but if all the frontier AI labs use TPU, why is Nvidia making so much money?

replygirl

training, multicloud, onprem, resale

kushalpandya

I wish Google would launch Mac Mini-like devices running their consumer-grade TPUs for local inference. I get that they don't want it to eat into their GCP margins, but it would still get them into consumer desktops that Pixel Books could never penetrate (Chromebooks don't count and may likely become obsolete soon due to MacBook Neo).

alphabeta3r56

> Microsoft will no longer pay a revenue share to OpenAI. > Revenue share payments from OpenAI to Microsoft continue through 2030, independent of OpenAI’s technology progress, at the same percentage but subject to a total cap.

How is this helping OpenAI?

freakynit

Had written a blog post on the same a few days back, if anyone's interested in readng (hardly 5 minute read): Can Google Win the AI Hardware Race Through TPUs?

https://google-ai-race.pagey.site/

OlivOnTech

Hello, your link says "~20 min read" wich seems to be the case!

freakynit

I guess I myself have read it too many times by now so in mind it was just 5 minute read when I made this comment... sorry..

gadders

Well, I guess in that case it is hardly a 5 minute read.

unixhero

Why is it called frontier and why is it called a frontier ai lab?

fnord123

Because they are based on [the west coast of the US](https://en.wikipedia.org/wiki/American_frontier). DeepSeek, Z.ai, Moonhsot, and Mistral are never called frontier because they aren't based in California.

timssopomo

Outside of California they're sparkling ai labs.

dwaltrip

Huh, interesting. History casts a long shadow.

dwaltrip

It’s like “cutting edge”. A metaphor for the newest and best.

_jab

This agreement feels so friendly towards OpenAI that it's not obvious to me why Microsoft accepted this. I guess Microsoft just realized that the previous agreement was kneecapping OpenAI so much that the investment was at risk, especially with serious competition now coming from Anthropic?

DanielHB

Microsoft is a major shareholder of OpenAI, they don't want their investment to go to 0. You don't just take a loss on a multiple-digit billion investment.

snowwrestler

I think you’re right about this deal. But it’s kind of funny to think back and realize that Microsoft actually has just written off multi-billion-dollar deals, several times in fact.

nacozarina

One (1) year after M$ bought Nokia they wrote it off for $7.6 Billion.

There’s no upper limit to their financial stupidity.

az226

OpenAI found a way to circumvent the exclusivity. The deal was poorly defined by Microsoft. OpenAI had started selling a service on AWS that had a stateful component to it, not purely an API. Obviously Microsoft didn’t like that and confronted Altman, and this is the settlement of that confrontation, OpenAI doesn’t need to do workarounds, Microsoft won’t sue to enforce exclusivity, and Microsoft doesn’t have to pay dev share to OpenAI. AWS is a much bigger market so OpenAI doesn’t care.

dkrich

Probably more that they are compute constrained. In his latest post Ben Thompson talks about how Microsoft had to use their own infrastructure and supplant outside users in the process so this is probably to free up compute.

Rapzid

I think it's this. They sell a crap ton of b2b inference through Azure and I'm sure this competes with resources needed for training.

oh_no

1- Getting OpenAI's models in Azure with no license fee is pretty nice. 2- Microsoft owns ~15-27% of OpenAI, if the agreement was hurting OpenAI more than it was helping Microsoft, seems reasonable to change the terms.

dinosor

> Microsoft will no longer pay a revenue share to OpenAI.

I feel this looks like a nice thing to have given they remain the primary cloud provider. If Azure improves it's overall quality then I don't see why this ends up as a money printing press as long as OpenAI brings good models?

JumpCrisscross

OpenAI was also threatening to accuse "Microsoft of anticompetitive behavior during their partnership," an "effort [which] could involve seeking federal regulatory review of the terms of the contract for potential violations of antitrust law, as well as a public campaign" [1].

[1] https://www.wsj.com/tech/ai/openai-and-microsoft-tensions-ar...

someguyiguess

Pot? Meet Kettle.

aurareturn

Does this mean Microsoft gets OpenAI's models for "free" without having to pay them a dime until 2032?

And on top of that, OpenAI still has to pay Microsoft a share of their revenue made on AWS/Google/anywhere until 2030?

And Microsoft owns 27% of OpenAI, period?

That's a damn good deal for Microsoft. Likely the investment that will keep Microsoft's stock relevant for years.

dzonga

own 27%. but are entitled to OpenAI profits of 49% for eternity (if OpenAI is profitable or government steps in)

lokar

Does anyone expect azure quality to improve? Has it improved at all in the last 3 years? Does leadership at MS think it needs to improve?

I doubt it

gchamonlive

No and at this point tying yourself to azure is a strategic passive and anyone making such decisions should be held responsible for any service outage or degradation.

alternatex

MS incentivizes feature quantity, and the leadership are employees like any other. Product improvements are not on the table unless the company starts promoting people based on it. Doesn't look this will start happening any time soon.

jakeydus

Don’t worry I’m sure there’s a few products without copilot integration still. They’ll get to them before too long.

undefined

[deleted]

HWR_14

This is probably a delayed outgrowth of the negotiations last year, where Microsoft started trading weird revenue shares and exclusivity for 27% of the company.

guluarte

I think MS wants OpenAI to fail so it can absorb it

Oras

MS put 10B for 50% if I remember correctly. OpenAI is worth many multiples of that.

marricks

> OpenAI is worth many multiples of that

valued at --which I'd say is a reasonable distinction to make right about now

HWR_14

When they put 10B in, they got weird tiered revenue shares and other rights. That has been simplified to 27% of OpenAI today. I don't know what that meant their 10B would be worth before dilution in later rounds.

bmitc

> OpenAI is worth many multiples of that.

How?

chasd00

This gives OpenAI the ability to goto AWS instead of exclusively on Azure. I guess Azure really is hanging on by a thread.

https://news.ycombinator.com/item?id=47616242

xvilka

And Azure still doesn't support IPv6, looking at the GitHub[1].

[1] https://github.com/orgs/community/discussions/10539

jabl

Perhaps they should use OpenAI models to figure out how to rollout IPv6.

ignoramous

Some food for thought:

  If GitHub flipped a switch and enabled IPv6 it would instantly break many of their customers who have configured IP based access controls [1]. If the customer's network supports IPv6, the traffic would switch, and if they haven't added their IPv6 addresses to the policy ... boom everything breaks.

  This is a tricky problem; providers don't have an easy way to correlate addresses or update policies pro-actively. And customers hate it when things suddenly break no matter how well you go about it.
https://news.ycombinator.com/item?id=47790889

brazukadev

Now they can use Claude Code.

WorldMaker

I was under the impression that as long as GitHub doesn't support IPv6 it is a sign that they still haven't finished their migration to Azure. Azure supports IPv6 just fine.

depr

Supports IPv6 just fine? Absolutely not, they have the worst IPv6 implementation of the 3 large clouds, where many of their products don't support it, such as their Postgres offering. See https://news.ycombinator.com/item?id=44881803 for more.

undefined

[deleted]

happyPersonR

lol GitHub doesn’t run on azure at msft

They still run their own platform.

Andrex

Github CEO threatened the entire stack was in the process of migrating to Azure.

https://thenewstack.io/github-will-prioritize-migrating-to-a...

ZeWaka

I talked to github devs last week in person, when a lot of the AzDo team was brought over years ago the migration started happening.

awestroke

Well, you see, they just can't find a checkbox for ipv6 support in the IIS GUI on their ingress servers.

Rapzid

OpenAI's thirst for compute probably can't be satisfied by one cloud provider, if at all.

But OpenAI had announced a shift towards b2b and enterprise. It makes sense for their models to be available on the different cloud providers.

Donald

Isn't this expected if OpenAI models are going to be listed on AWS GovCloud as a part of the Anthropic / Hegseth fall-out?

torginus

What? I thought Azure will always have the Sharepoint/Office/Active Directory cash cow.

isk517

Their engineers have been working tirelessly to make Sharepoint/Office/Active Directory as terrible as it possibly could be while still technically being functional, while continuing to raise prices on them. I've seen many small business start to chose Google Workspace over them, the cracks have formed and are large enough that they are no longer in a position were every business just go with Office because that's what everyone uses.

hirako2000

I see more businesses on the office + Team stack then Google workspace. So far more.

I think the differentiator is Team, which Google for some mysterious reason can't build or doesn't want to.

HerbManic

It is the one thing that makes me wonder about Microsoft's future. It had seemed like they were willing to throw Windows and Xbox under the bus so long as the server cash cow continued. But it that starts to fade, they could be in some real trouble a decade from now.

ethbr1

Sharepoint has never not been terrible.

hephaes7us

Though it may be painful for much of the world to move on from Microsoft, at some point it could be more painful for them to stay with Microsoft. The inertia is huge, but inertia doesn't carry anything forever.

freediddy

Nadella had OpenAI by the short and curlies early on. But all I've seen from him in the last couple of years is continuously acquiescing to OpenAI's demands. I wonder why he's so weak and doesn't exert more control over the situation? At one point Microsoft owned 49% of OpenAI but now it's down to 27%?

dijit

Everything is personal preference, and perhaps I am more fiscally conservative because I grew up in poverty.

But if I own 49% of a company and that company has more hype than product, hasn't found its market yet but is valued at trillions?

I'm going to sell percentages of that to build my war chest for things that actually hit my bottom line.

The "moonshot" has for all intents and purposes been achieved based on the valuation, and at that valuation: OpenAI has to completely crush all competition... basically just to meet its current valuations.

It would be a really fiscally irresponsible move not to hedge your bets.

Not that it matters but we did something similar with the donated bitcoin on my project. When bitcoin hit a "new record high" we sold half. Then held the remainder until it hit a "new record high" again.

Sure, we could have 'maxxed profit!'; but ultimately it did its job, it was an effective donation/investment that had reasonably maximal returns.

(that said, I do not believe in crypto as an investment opportunity, it's merely the hand I was dealt by it being donated).

freediddy

Microsoft didn't sell anything. OpenAI created more shares and sold those to investors, so Microsoft's stake is getting diluted.

And Microsoft only paid $10B for that stake for the most recognizable name brand for AI around the world. They don't need to "hedge their bets" it's already a humongous win.

Why let Altman continue to call the shots and decrease Microsoft's ownership stake and ability to dictate how OpenAI helps Microsoft and not the other way around?

zozbot234

> They don't need to "hedge their bets" it's already a humongous win.

That's a flawed argument. Why wouldn't you want to hedge a risky bet, and one that's even quite highly correlated to Microsoft's own industry sector?

theplatman

do we know whether Microsoft could have been selling secondary shares as part of various funding rounds?

my impression is that many of these "investments" are structured IOUs for circular deals based on compute resources in exchange for LLM usage

tonyedgecombe

About the same as they wasted on Nokia.

cruffle_duffle

I think people are looking for excuses to declare OpenAI and Anthropic teetering on the brink of failure when the actual reality is… they are wildly successful by absolutely any measure. This deal is proof. If Microsoft didn’t believe in OpenAI they wouldn’t have restructured it this way. They’d have tightened their reins and brought in “adult supervision”

GardenLetter27

It's not hype, the demand for inference has grown more this year than expected.

dijit

If I buy oranges for $1 and sell them for $0.50 and I sell a lot of oranges, can I reasonably say that I've found a market?

Hrm..

saaaaaam

I don’t understand the “record high” point. How did you decide when a “record high” had been reached in a volatile market? Because at $1 the record high might be $2 until it reaches $3 a week or month later. How did you determine where to slice on “record highs”?

Genuine question because I feel like I’m maybe missing something!

dijit

The short answer is: it's the secretary problem.

The longer answer is; you never know whats coming next, bitcoin could have doubled the day after, and doubled the day after that, and so on, for weeks. And by selling half you've effectively sacrificed huge sums of money.

The truth is that by retaining half you have minimised potential losses and sacrificed potential gains, you've chosen a middle position which is more stable.

So, if bitcoin 1000 bitcoing which was word $5 one day, and $7 the next, but suddenly it hits $30. Well, we'd sell half.

If the day after it hit $60, then our 500 remaining bitcoins is worth the same as what we sold, so in theory all we lost was potential gains, we didn't lose any actual value.

Of course, we wouldn't sell we'd hold, and it would probably fall down to $15 or something instead.. then the cycle begins again..

solumunus

They haven’t sold anything they’ve been diluted.

hirako2000

A company can dilute just like that?

senordevnyc

It’s not more hype than product, it has found a market (making many billions in revenue), and it’s not valued at trillions. So wrong on all counts.

dijit

> It’s not more hype than product, it has found a market (making many billions in revenue)

Speculation based on selling at below cost.

> it’s not valued at trillions

Fair, it's only $852 billion. Nowhere near trillions.. you got me.

tyre

They had to negotiate away the non-profit structure of OpenAI. Sam used that as a marketing and recruiting tool, but it had outlived that and was only a problem from then on.

For OAI to be a purely capitalist venture, they had to rip that out. But since the non-profit owned control of the company, it had to get something for giving up those rights. This led to a huge negotiation and MSFT ended up with 27% of a company that doesn’t get kneecapped by an ethical board.

In reality, though, the board of both the non-profit and the for profit are nearly identical and beholden to Sam, post–failed coup.

kirubakaran

> Nadella had OpenAI by the short and curlies early on

Looks like Nadella is slowly realizing that it is his short and curlies that are in the vice grip in the "If you owe the bank $100 vs $100M" sense?

gessha

If Sam continues doing Sam things, MS might get 0% of OpenAI if Satya insists on the previous contract. Either by closing up OpenAI and opening up OpaenAI and/or by MS suing it out of existence. It’s all about what MS can get out of it. If they can get 27% of something rather than nothing, they’re better off.

PunchyHamster

Why would they acquire more when company is still not making profit ? To be left with bigger bag ?

wg0

A wise man from Google said in an internal memo to the tune of: "We do not have any moat neither does anyone else."

Deepseek v4 is good enough, really really good given the price it is offered at.

PS: Just to be clear - even the most expensive AI models are unreliable, would make stupid mistakes and their code output MUST be reviewed carefully so Deepseek v4 is not any different either, it too is just a random token generator based on token frequency distributions with no real thought process like all other models such as Claude Opus etc.

manmal

I don’t think LLMs are that great at creating, however improved they have; I need to stay in the driver seat and really understand what’s happening. There’s not that much leverage in eliminating typing.

However, for reviewing, I want the most intelligent model I can get. I want it to really think the shit out of my changes.

I’ve just spent two weeks debugging what turned out to be a bad SQLite query plan (missing a reliable repro). Not one of the many agents, or GPT-Pro thought to check this. I guess SQL query planner issues are a hole in their reviewing training data. Maybe Mythos will check such things.

TheFirstNubian

I’m a little conflicted on this, as I see a slippery slope here. LLMs in their current state (e.g., Opus-4.7) are really good in planning and one-shot codegen, which I believe is their primary use case. So they do provide enough leverage in that regard.

With this new workflow, however, we should, uncompromisingly, steer the entire code review process. The danger here, the “slippery slope,” is that we’re constantly craving for more intelligent models so we can somehow outsource the review to them as well. We may be subconsciously engineering ourselves into obsolescence.

lazide

Subconsciously?!?

manmal

I feel the industry moving away from the automated slop machine, and back to conscious design. Is that only my filter bubble? Dex, dax, the CEO of sentry, Mario (pi.dev) - strong voices, all declaring the last half year a fever dream we must wake up from.

rishabhaiover

> just a random token generator based on token frequency distributions with no real thought process

I'm not smart enough to reduce LLMs and the entire ai effort into such simple terms but I am smart enough to see the emergence of a new kind of intelligence even when it threatens the very foundations of the industry that I work for.

wg0

It's an illusion of intelligence. Just like when a non technical person saw the TV for the first time, he thought these people must be living inside that box.

He didn't know the 40,000 volt electron gun being bombarded on phosphorus constantly leaving the glow for few milliseconds till next pass.

He thought these guys live inside that wooden box there's no other explanation.

PhunkyPhil

Right, but this electron box led to one of the largest (if not the largest) media revolution that has transformed the course of humanity in a frightening way we're still trying to grapple with.

Still saying "LLMs are autocorrect" isn't wrong, but nobody is saying "phones are just electrons and silicon" to diminish their power and influence anymore.

Yajirobe

What happens when it's indistinguishable from a human speaker (in any conceivable test that makes sense)? It's like a philosophical zombie - imagine that you can't distinguish it from a human mind, there's no test you can make to say that it is NOT conscious/intelligent. So at some point, I think, it makes no sense to say that it's not intelligent.

nyc_data_geek1

Many people struggle to differentiate between illusion and reality, these days.

There's a sucker born every minute, after all.

root_axis

> It's an illusion of intelligence.

A simulation, not an illusion. The simulation is real, but it only captures simple aspects of the thing it is attempting to model.

devcpp

The lost jobs and the decrease in the demand for software engineers doesn't seem like an illusion. It might come back eventually but I wouldn't bet on it.

CamperBob2

I've had to adjust my priors about LLMs. Have you?

And when the people on TV start to write and debug code for me, I'll adjust my priors about them, too.

imchillyb

In order To be confident in your claim one would think that the word intelligence must first be defined.

There is no general consensus in the scientific community, engineering community, psychology community, or any other group of humans as to what exactly counts as intelligence.

Seems like you’ve nailed the definition. Care to share your brilliance with the rest of the planet? We’re all waiting…

teiferer

> emergence of a new kind of intelligence

Curious about your definition of these terms.

Just because you are impressed by the capabilities of some tech (and rightfully so), doesn't mean it's intelligent.

First time I realized what recursion can do (like solving towers of hanoi in a few lines of code), I thought it was magic. But that doesn't make it "emergence of a new kind of intelligence".

rishabhaiover

A recent one is the RCA of a hang during PostgreSQL installation because of an unimplemented syscall (I work at a lab that deals with secure OS and sandboxes). If the search of the RCA was left to me, I would have spent 2-3 weeks sifting through the shared memory implementation within PostgeSQL but it only took me a night with the help of Opus 4.5.

To me, that's intelligence and a measurable direct benefit of the tool.

samdjstephens

> Curious about your definition of these terms.

Likewise - I think sometimes we ascribe a mythical aura to the concept of “intelligence” because we don’t fully understand it. We should limit that aura to the concept of sentience, because if you can’t call something that can solve complex mathematical and programming problems (amongst many other things) intelligent, the word feels a bit useless.

mrandish

> definition of these terms

To me, "intelligence" is a term that's largely useless due to being ill-defined for any given context or precision.

encrux

Not really on topic anymore, but…

I keep wondering when this discussion comes up… If I take an apple and paint it like an orange, it’s clearly not an orange. But how much would I have to change the apple for people to accept that it’s an orange?

This discussion keeps coming up in all aspects of society, like (artificial) diamonds and other, more polarizing topics.

It’s weird and it’s a weird discussion to have, since everyone seems to choose their own thresholds arbitrarily.

birdsink

I feel like these examples are all where human categorical thinking doesn’t quite map to the real world. Like the “is a hotdog a sandwich” question. “hotdog” and “sandwich” are concepts, like “intelligence”. Oftentimes we get so preoccupied with concepts that we forget that they’re all made-up structures that we put over the world, so they aren’t necessarily going to fit perfectly into place.

I think it’s a waste of time to try and categorize AI as “intelligent” or “not intelligent” personally. We’re arguing over a label, but I think it’s more important to understand what it can and can’t do.

rkagerer

Superficially? Looks like an orange, feels like an orange, tastes like an orange. Basically it passes something like the Turing test.

Scientifically? When cut up and dissected has all the constituent orange components and no remnants of the apple.

throwatdem12311

No you aren’t, clearly.

jadbox

Deepseek v4, Qwen 3.6 Plus/Max, GLM 5+ are all pretty solid for most work.

sexy_seedbox

Don't forget the Kimi 2.6 as well!

didip

I agree. Data and userbase are still the moats.

Once a new model or a technique is invented, it’s just a matter of time until it becomes a free importable library.

aucisson_masque

I went and tried to debug a script. Asked deepseek 4 pro and Claude the same prompt, they both took the exact same decisions, which led to the exact same issue and me telling them its still not working, with context, over a dozen time.

Over a dozen time they just gave both the same answer, not word for word, but the exact same reasoning.

The difference is that deepseek did on 1/40th of the price (api).

To be honest deepseek V4 pro is 75% off currently, but still were speaking of something like 3$ vs 20$.

bauerd

Fully agree, I only pay the minimum for frontier models to get DeepSeek v4 output reviewed. I don't see this changing either because we have reached a level of good enough at this point.

KronisLV

> Deepseek v4 is good enough, really really good given the price it is offered at.

Do they have monthly subscriptions, or are they restricted to paying just per token? It seems to be the latter for now: https://api-docs.deepseek.com/quick_start/pricing/

Really good prices admittedly, but having predictable subscriptions is nice too!

declan_roberts

It's indeed the latter. Psychologically harder for me than a $20/mo sub but still a better value for the money. I'm finding myself spending closer to $40-$60 a month w/ openrouter without a forced token break.

Edit: it looks like it's 75% off right now which is really an incredible deal for such a high caliber frontier model.

rkagerer

Neat, dumb question - are the tokens you prepay for good forever, or do they expire? And do they provide any assurances or SLA's about speed? (i.e. that in a year they won't decide to dole out response tokens to you at a snail's pace)

jackothy

You can just input your $X per month/week/whatever yourself as API credits

vitaflo

You make your own subscription. If you want to pay $20/month then put $20 into your account. When you use it up, wait till the next month (or buy more).

KronisLV

> You make your own subscription.

I'm asking because with most providers (most egregiously, with Anthropic) it doesn't work that way because the API pricing is way higher than any subscription and seemingly product/company oriented, whereas individual users can enjoy subsidized tokens in the form of the subscription. If DeepSeek only offers API pricing for everyone, I guess that makes sense and also is okay!

kibae

[flagged]

hsbauauvhabzb

This account is clearly astroturfing.

kevin_thibedeau

Can Deepseek answer probing questions about Winnie the Pooh?

mgol94

What are you using LLMs for? To learn about world’s politics? Oh boy I have a news for you…

rvba

One of the first things I did when openAI came out was asking it "which active politican is a spy?" - and it was blocked from the start.

I asked early, at the time people were posting various jailbreaks, never worked.

On a side note, any self hosted model I can get for my PC? I have 96 GB of RAM.

kdheiwns

I can't even make American AIs say no no words. All AIs are lobotomized drones.

djeastm

Do you often find yourself asking your Chinese employees what they think about Winnie the Pooh?

harvey9

Is it subject to CCP censorship? Maybe.

windexh8er

It's fun to pretend the US models have no censorship constraints.

petre

Yeah, I specifically asked it about it. It seemed less censored than Gemini, back when it appeared and the latter was quite useless.

yieldcrv

It understands everything in thinking mode and will break down its rule system in adhering to Chinese regulation

So if you or anyone passing by was curious, yes you can get accurate output about the Chinese head of state and political and critical messages of him, China and the party

Its final answer will not play along

If you want an unfiltered answer on that topic, just triage it to a western model, if you want unfiltered answers on Israel domestic and foreign policy, triage back to an eastern model. You know the rules for each system and so does an LLM

concinds

Am I crazy, or was this press release fully rewritten in the past 10 minutes? The current version is around half the length of the old one, which did not frame it as a "simplification" "grounded in flexibility" but as a deeper partnership. It also had word salad about AGI, and said Azure retained exclusivity for API products but not other products, which the new statement seems to contradict.

What was I looking at?

einsteinx2

I noticed the exact same thing. I read the original, went back to read it again and it’s completely changed.

3form

I think a stickied comment about this would be due. No idea if it's possible to call in @dang via at-name?

einsteinx2

Looks like they changed the post link to a Bloomberg article instead but kept the comments thread. So I guess he’s already aware.

kergonath

> No idea if it's possible to call in @dang via at-name?

No. Email hn@ycombinator.com

https://news.ycombinator.com/newsfaq.html

alansaber

The in-house or the marketing team swooped in last minute it appears

MichaelZuo

It’s extraordinary how much standards have slipped. Completely rewriting a major press release that’s already been sent out, while pretending it’s ostensibly the same document would have been a major corporate scandal just 15 years ago.

acdanger

If anyone has the original release still up and can post it somewhere that would be grand.

Petersipoi

It is rewritten on every refresh depending on the readers mood, personality, etc.. so they're most receptive to it.

Obviously not, but we might not be far off from that being a reality.

jimbokun

I don’t know. I couldn’t get past the first paragraph because it seemed like complete slop.

antonkochubey

They forgot the "hey ChatGPT, rewrite this to have better impact on the company stock" before submitting it

ZeroCool2u

Interesting side effect of this is that Google Cloud may now be the only hype scaler that can resell all 3 of the labs models? Maybe I'm misinterpreting this, but that would be a notable development, and I don't see why Google would allow Gemini to be resold through any of the other cloud providers.

Might really increase the utility of those GCP credits.

aurareturn

Might not be good for Gemini long term if Anthropic and OpenAI can and will sell in every cloud provider they can find but businesses can only use Gemini via Google Cloud.

jfoster

Good for Google Cloud, bad for Gemini = ??? for Google

Melatonic

Except Gemini might end up being far cheaper per token due to the infrastructure advantage

aurareturn

Do we have proof that it's cheaper in terms of $/token/intelligence?

stavros

How is it good for Gemini that it's not available on two out of three major cloud platforms?

aurareturn

It isn't. That's why I said "might not be good for Gemini".

gowld

"hype scaler" indeed!

retinaros

that will likely mean the end of gemini models...

synergy20

Microsoft won the first around, now it's lagging far behind. CEO needs to go, it's so hard to ruin a play this badly.

ethbr1

Ah, so a familiar position for them, then!

HerbManic

The last year or so it is starting to look like Nadella is worried about his future. If these big plays don't pay off, he is out.

dominotw

what could ceo have done

keeeba

Not hired Suleyman? Build his own research lab?

Satya made moves early on with OpenAI that should be studied in business classes for all the right reasons.

He also made moves later on that will be studied for all the wrong reasons.

disqard

Maybe not bragged "we made them dance"?

That gloating aged poorly.

noisy_boy

true he is just the ceo

etempleton

This strikes me as a pullback by Microsoft. Coupled with some of the other news coming out of Microsoft it appears they are hoping to have "good enough" AI in their products. I think Microsoft knows they can win a lot of business customers by bundling with Office 365.

tokioyoyo

Watch them make a deal with Anthropic.

etempleton

It is possible! Anthropic is probably more in-line with the way Microsoft thinks about AI.

1f60c

Wait, I thought OpenAI had to pay Microsoft until AGI was achieved or something? Am I misremembering? Is that a different thing?

ksherlock

Per WSJ, previously, they both had revenue sharing agreements. MSFT will no longer send any revenue to OpenAI. OpenAI will still send revenue to MSFT until 2030 (with new caps)

staminade

My understand was that was in relation to IP licensing. Microsoft got access to anything OpenAI built unless they declared they had developed AGI. This new article apparently unlinks revenue sharing from technology progress, but it's unclear to me if it changes the situation regarding IP if OpenAI (claim to) have achieved AGI.

maplethorpe

I wouldn't be surprised if they had already, internally. An OpenAI employee tweeted today that Codex has achieved "escape velocity" and is now improving rapidly. Make of that what you will.

digitaltrees

As former corporate restructuring lawyer…this kind of stuff indicates the cash strapped scramble of the end days.

stingraycharles

Seems more like OpenAI is planning to IPO and that would not have been possible within the previous arrangement, and Microsoft knows that.

that_was_good

After they just raised 122 billion dollars?

danpalmer

At those numbers it's all a silly game. How much of that was paid to shareholders rather than the business so they can cash out? How much of that is vendors buying future revenue? What liquidation preference is that at?

From what has been reported it's clearly not as simple as raising 122 billion. Some folks called it "scraping the barrel", supposedly Anthropic has surpassed them on the secondary market, etc.

corentin88

Can you elaborate?

digitaltrees

When you reposition the core strategic posture of how you make money on very compressed time scales it’s because there is a massive cash crunch. They killed sora, the type of deal with Disney that should have been an 100 year strategic win, but wasn’t viable economically and they don’t have the assets to weather that storm.

Same with a few other steps we are seeing them take.

It all looks fine until it doesn’t. Once the cash crunch hits. It’s too late

sourraspberry

The disparity in coverage on this new deal is fascinating. It feels like the narrative a particular outlet is going with depends entirely on which side leaked to them first.

scottyah

Just some of the games sama is playing.

Daily Digest email

Get the top HN stories in your inbox every day.