Get the top HN stories in your inbox every day.
thanhhaimai
Opinions are my own.
I think the biggest winner of this might be Google. Virtually all the frontier AI labs use TPU. The only one that doesn't use TPU is OpenAI due to the exclusive deal with Microsoft. Given the newly launched Gen 8 TPU this month, it's likely OpenAI will contemplate using TPU too.
bastawhiz
Many labs use TPUs, but not exclusively. Most labs need more compute than they can get, and if there's TPU capacity, they'll adapt their systems to be able to run partially on TPUs.
gpt5
Why is AMD not more popular then if labs are so flexibly with giving away CUDA?
mattnewton
people are trying, especially for inference. For training, it’s just too high risk to tank your training I think.
TPUs are at least dogfooded by Google deepmind, no team AFAIK has gotten the AMD stack to train well.
zombiwoof
[dead]
zobzu
even google doesnt only use TPUs.
danpalmer
Google is in a different position to others in that they're the only frontier lab with a cloud infra business. It obviously makes sense to sell GPUs on cloud infra as people want to rent them. In that respect Google buys a ton of GPUs to rent out.
What's unclear to me is how much Google uses GPUs for their own stuff. Yes Gemini runs on GPUs now, so that Google can sell Gemini on-prem boxes (recent release announced last week), but is any training or inference for Gemini really happening on GPUs? This is unclear to me. I'd have guessed not given that I thought TPUs were much cheaper to operate, but maybe I'm wrong.
Caveat, I work at Google, but not on anything to do with this. I'm only going on what's in the press for this stuff.
yjadsfgasdf
[dead]
maxclark
And almost by happenstance Apple. Turns out they have a great platform for inference and torched almost nothing comparatively on Siri. The Apple/Gemini deal is interesting, Google continues to demonstrate their willingness to degrade their experience on Apple to try and force people to switch.
ttul
If you do the math (I did), in 2 years, open source models that you can run on a future MacBook Pro will be as capable as the frontier cloud models are today. Memory bandwidth is growing rapidly, as is the die area dedicated to the neural cores. And all the while, we have the silicon getting more power efficient and increasingly dense (as it always does). These hardware improvements are coming along as the open source models improve through research advancements. And while the cloud models will always be better (because they can make use of as much power as they want to - up in the cloud), what matters to most of us is whether a model can do a meaningful share of knowledge work for us. At the same time, energy consumption to run cloud infrastructure is out-pacing the creation of new energy supply, which is a problem not easily solved. I believe scarcity of energy will increasingly drive frontier labs toward power efficiency, which necessarily implies that the Pareto frontier of performance between cloud and local execution will narrow.
nl
A Opus 4.7/Gpt5.5 class model is 5 trillion parameters[1].
To run a 8 bit quantized version of that you need roughly 5TB of RAM.
Today that is around 18 NVidia B300. That's around $900,000, without including the computers to run them in.
It's true that the capability of open source models is improving, but running actual frontier models on your MPB seems a way off.
[1] https://x.com/elonmusk/status/2042123561666855235?s=20 (and Elon has hired enough people out of those labs to have a fair idea)
npunt
I did this calculation a bit ago and don't think frontier models are just a few MacBook Pro generations away. Yes numbers reliably go up in tech in general but in specific semiconductors & standards have long lead-times and published roadmaps, so we can have high confidence in what we're getting even in 3-4 years in terms of both transistor density and RAM speeds.
In mid-2028 we have N2E/N2P with around 15% greater transistor density than today's N3P, and by EOY2028 we'll likely have A14 with about 35-40% density improvement.
Meanwhile, we'll be on LPDDR6 by that point, which takes M-series Pros from 307GB/s -> ~400GB/s, and Max's from 614GB/s -> ~800GB/s.
Model improvements obviously will help out, but on the raw hardware front these aren't in the ballpark for frontier model numbers. An H100 has 3TB/s memory bandwidth, fwiw
xorcist
That's not "math". That's a "wild guess", or baseless extrapolation at best.
CMay
So long as you don't require deep search grounding like massive web indexes or document stores which are hard to reproduce locally. You can do local agentic things that get close or even do better depending on search strategy, but theoretically a massive cloud service with huge data stores at hand should be able to produce better results.
In practice unless you're doing some kind of deep research thing with the cloud, it'll try to optimize mostly for time and get you a good enough answer rather than spending an hour or two. An hour of cloud searching with huge data stores is not equivalent to an hour of local agentic searching, presumably.
I think that problem will improve a little in the coming years as we kind of create optimized data curation, but the information world will keep growing so the advantage will likely remain with centralized services as long as they offer their complete potential rather than a fraction.
dualvariable
Also, all the cloud models don't have to be the best frontier models, and you don't need to focus on hitting the benchmark of shrinking Opus 4.7 down to a single MBP to make significant improvements. If you get it so that an Opus 4.7 benchmark-compatible model can run in $250k of datacenter capex (and associated reduced opex for power+cooling) that'd be a massive cost improvement that makes the cloud models cheaper. And for most consumers that'll probably be good enough. You don't need to run on a $5k laptop to make a big difference.
rc1
Show your working / explain your math?
GorbachevyChase
They also degrade their own direct services with little warning or thought put into change management, so, to be fair, Apple may be getting the same quality of service as the rest of us.
vharish
I think that's just how Google is, by nature. They don't intentionally degrade their services. They just aren't a customer centric company. They run on numbers. As a corporate, it doesn't really encourage support and maintenance work either.
manueltgomes
Indeed. I'm wondering if Apple's "miss the train" with AI ended up being a blessing for them. Not only in the Google deal but also there's a lot of people doing interesting stuff locally..
bigyabai
Apple is basically in the same boat as AMD and Intel. They have a weak, raster-focused GPU architecture that doesn't scale to 100B+ inference workloads and especially struggles with large context prefill. TPUs smoke them on inference, and Nvidia hardware is far-and-away more efficient for training.
hellohello2
What do TPUs do to improve on GPUs at inference?
brcmthrowaway
This doesn't get talked about enough - the GPU is weak, weak, weak. And anyone who can fix them will go to a serious AI company (for 2-3x the salary).
munk-a
Apple is in a much better boat than AMD or Intel. They have a gigantic warchest and can just snap up whoever looks like a leader coming out of the bubble burst.
undefined
celeritascelery
Maybe I am missing something here, but if all the frontier AI labs use TPU, why is Nvidia making so much money?
replygirl
training, multicloud, onprem, resale
kushalpandya
I wish Google would launch Mac Mini-like devices running their consumer-grade TPUs for local inference. I get that they don't want it to eat into their GCP margins, but it would still get them into consumer desktops that Pixel Books could never penetrate (Chromebooks don't count and may likely become obsolete soon due to MacBook Neo).
alphabeta3r56
> Microsoft will no longer pay a revenue share to OpenAI. > Revenue share payments from OpenAI to Microsoft continue through 2030, independent of OpenAI’s technology progress, at the same percentage but subject to a total cap.
How is this helping OpenAI?
freakynit
Had written a blog post on the same a few days back, if anyone's interested in readng (hardly 5 minute read): Can Google Win the AI Hardware Race Through TPUs?
OlivOnTech
Hello, your link says "~20 min read" wich seems to be the case!
freakynit
I guess I myself have read it too many times by now so in mind it was just 5 minute read when I made this comment... sorry..
gadders
Well, I guess in that case it is hardly a 5 minute read.
VirusNewbie
OpenAI uses GCP. I don't know if they use TPUs.
https://www.reuters.com/business/retail-consumer/openai-taps...
unixhero
Why is it called frontier and why is it called a frontier ai lab?
fnord123
Because they are based on [the west coast of the US](https://en.wikipedia.org/wiki/American_frontier). DeepSeek, Z.ai, Moonhsot, and Mistral are never called frontier because they aren't based in California.
timssopomo
Outside of California they're sparkling ai labs.
dwaltrip
Huh, interesting. History casts a long shadow.
dwaltrip
It’s like “cutting edge”. A metaphor for the newest and best.
_jab
This agreement feels so friendly towards OpenAI that it's not obvious to me why Microsoft accepted this. I guess Microsoft just realized that the previous agreement was kneecapping OpenAI so much that the investment was at risk, especially with serious competition now coming from Anthropic?
DanielHB
Microsoft is a major shareholder of OpenAI, they don't want their investment to go to 0. You don't just take a loss on a multiple-digit billion investment.
snowwrestler
I think you’re right about this deal. But it’s kind of funny to think back and realize that Microsoft actually has just written off multi-billion-dollar deals, several times in fact.
nacozarina
One (1) year after M$ bought Nokia they wrote it off for $7.6 Billion.
There’s no upper limit to their financial stupidity.
az226
OpenAI found a way to circumvent the exclusivity. The deal was poorly defined by Microsoft. OpenAI had started selling a service on AWS that had a stateful component to it, not purely an API. Obviously Microsoft didn’t like that and confronted Altman, and this is the settlement of that confrontation, OpenAI doesn’t need to do workarounds, Microsoft won’t sue to enforce exclusivity, and Microsoft doesn’t have to pay dev share to OpenAI. AWS is a much bigger market so OpenAI doesn’t care.
dkrich
Probably more that they are compute constrained. In his latest post Ben Thompson talks about how Microsoft had to use their own infrastructure and supplant outside users in the process so this is probably to free up compute.
Rapzid
I think it's this. They sell a crap ton of b2b inference through Azure and I'm sure this competes with resources needed for training.
oh_no
1- Getting OpenAI's models in Azure with no license fee is pretty nice. 2- Microsoft owns ~15-27% of OpenAI, if the agreement was hurting OpenAI more than it was helping Microsoft, seems reasonable to change the terms.
dinosor
> Microsoft will no longer pay a revenue share to OpenAI.
I feel this looks like a nice thing to have given they remain the primary cloud provider. If Azure improves it's overall quality then I don't see why this ends up as a money printing press as long as OpenAI brings good models?
JumpCrisscross
OpenAI was also threatening to accuse "Microsoft of anticompetitive behavior during their partnership," an "effort [which] could involve seeking federal regulatory review of the terms of the contract for potential violations of antitrust law, as well as a public campaign" [1].
[1] https://www.wsj.com/tech/ai/openai-and-microsoft-tensions-ar...
someguyiguess
Pot? Meet Kettle.
aurareturn
Does this mean Microsoft gets OpenAI's models for "free" without having to pay them a dime until 2032?
And on top of that, OpenAI still has to pay Microsoft a share of their revenue made on AWS/Google/anywhere until 2030?
And Microsoft owns 27% of OpenAI, period?
That's a damn good deal for Microsoft. Likely the investment that will keep Microsoft's stock relevant for years.
dzonga
own 27%. but are entitled to OpenAI profits of 49% for eternity (if OpenAI is profitable or government steps in)
lokar
Does anyone expect azure quality to improve? Has it improved at all in the last 3 years? Does leadership at MS think it needs to improve?
I doubt it
gchamonlive
No and at this point tying yourself to azure is a strategic passive and anyone making such decisions should be held responsible for any service outage or degradation.
alternatex
MS incentivizes feature quantity, and the leadership are employees like any other. Product improvements are not on the table unless the company starts promoting people based on it. Doesn't look this will start happening any time soon.
jakeydus
Don’t worry I’m sure there’s a few products without copilot integration still. They’ll get to them before too long.
undefined
HWR_14
This is probably a delayed outgrowth of the negotiations last year, where Microsoft started trading weird revenue shares and exclusivity for 27% of the company.
guluarte
I think MS wants OpenAI to fail so it can absorb it
Oras
MS put 10B for 50% if I remember correctly. OpenAI is worth many multiples of that.
marricks
> OpenAI is worth many multiples of that
valued at --which I'd say is a reasonable distinction to make right about now
HWR_14
When they put 10B in, they got weird tiered revenue shares and other rights. That has been simplified to 27% of OpenAI today. I don't know what that meant their 10B would be worth before dilution in later rounds.
bmitc
> OpenAI is worth many multiples of that.
How?
chasd00
This gives OpenAI the ability to goto AWS instead of exclusively on Azure. I guess Azure really is hanging on by a thread.
elpakal
Confirmed by Andy Jassy just now https://www.linkedin.com/posts/andy-jassy-8b1615_very-intere...
xvilka
And Azure still doesn't support IPv6, looking at the GitHub[1].
jabl
Perhaps they should use OpenAI models to figure out how to rollout IPv6.
ignoramous
Some food for thought:
If GitHub flipped a switch and enabled IPv6 it would instantly break many of their customers who have configured IP based access controls [1]. If the customer's network supports IPv6, the traffic would switch, and if they haven't added their IPv6 addresses to the policy ... boom everything breaks.
This is a tricky problem; providers don't have an easy way to correlate addresses or update policies pro-actively. And customers hate it when things suddenly break no matter how well you go about it.
https://news.ycombinator.com/item?id=47790889brazukadev
Now they can use Claude Code.
WorldMaker
I was under the impression that as long as GitHub doesn't support IPv6 it is a sign that they still haven't finished their migration to Azure. Azure supports IPv6 just fine.
depr
Supports IPv6 just fine? Absolutely not, they have the worst IPv6 implementation of the 3 large clouds, where many of their products don't support it, such as their Postgres offering. See https://news.ycombinator.com/item?id=44881803 for more.
undefined
happyPersonR
lol GitHub doesn’t run on azure at msft
They still run their own platform.
Andrex
Github CEO threatened the entire stack was in the process of migrating to Azure.
https://thenewstack.io/github-will-prioritize-migrating-to-a...
ZeWaka
I talked to github devs last week in person, when a lot of the AzDo team was brought over years ago the migration started happening.
awestroke
Well, you see, they just can't find a checkbox for ipv6 support in the IIS GUI on their ingress servers.
Rapzid
OpenAI's thirst for compute probably can't be satisfied by one cloud provider, if at all.
But OpenAI had announced a shift towards b2b and enterprise. It makes sense for their models to be available on the different cloud providers.
Donald
Isn't this expected if OpenAI models are going to be listed on AWS GovCloud as a part of the Anthropic / Hegseth fall-out?
torginus
What? I thought Azure will always have the Sharepoint/Office/Active Directory cash cow.
isk517
Their engineers have been working tirelessly to make Sharepoint/Office/Active Directory as terrible as it possibly could be while still technically being functional, while continuing to raise prices on them. I've seen many small business start to chose Google Workspace over them, the cracks have formed and are large enough that they are no longer in a position were every business just go with Office because that's what everyone uses.
hirako2000
I see more businesses on the office + Team stack then Google workspace. So far more.
I think the differentiator is Team, which Google for some mysterious reason can't build or doesn't want to.
HerbManic
It is the one thing that makes me wonder about Microsoft's future. It had seemed like they were willing to throw Windows and Xbox under the bus so long as the server cash cow continued. But it that starts to fade, they could be in some real trouble a decade from now.
ethbr1
Sharepoint has never not been terrible.
hephaes7us
Though it may be painful for much of the world to move on from Microsoft, at some point it could be more painful for them to stay with Microsoft. The inertia is huge, but inertia doesn't carry anything forever.
freediddy
Nadella had OpenAI by the short and curlies early on. But all I've seen from him in the last couple of years is continuously acquiescing to OpenAI's demands. I wonder why he's so weak and doesn't exert more control over the situation? At one point Microsoft owned 49% of OpenAI but now it's down to 27%?
dijit
Everything is personal preference, and perhaps I am more fiscally conservative because I grew up in poverty.
But if I own 49% of a company and that company has more hype than product, hasn't found its market yet but is valued at trillions?
I'm going to sell percentages of that to build my war chest for things that actually hit my bottom line.
The "moonshot" has for all intents and purposes been achieved based on the valuation, and at that valuation: OpenAI has to completely crush all competition... basically just to meet its current valuations.
It would be a really fiscally irresponsible move not to hedge your bets.
Not that it matters but we did something similar with the donated bitcoin on my project. When bitcoin hit a "new record high" we sold half. Then held the remainder until it hit a "new record high" again.
Sure, we could have 'maxxed profit!'; but ultimately it did its job, it was an effective donation/investment that had reasonably maximal returns.
(that said, I do not believe in crypto as an investment opportunity, it's merely the hand I was dealt by it being donated).
freediddy
Microsoft didn't sell anything. OpenAI created more shares and sold those to investors, so Microsoft's stake is getting diluted.
And Microsoft only paid $10B for that stake for the most recognizable name brand for AI around the world. They don't need to "hedge their bets" it's already a humongous win.
Why let Altman continue to call the shots and decrease Microsoft's ownership stake and ability to dictate how OpenAI helps Microsoft and not the other way around?
zozbot234
> They don't need to "hedge their bets" it's already a humongous win.
That's a flawed argument. Why wouldn't you want to hedge a risky bet, and one that's even quite highly correlated to Microsoft's own industry sector?
theplatman
do we know whether Microsoft could have been selling secondary shares as part of various funding rounds?
my impression is that many of these "investments" are structured IOUs for circular deals based on compute resources in exchange for LLM usage
tonyedgecombe
About the same as they wasted on Nokia.
cruffle_duffle
I think people are looking for excuses to declare OpenAI and Anthropic teetering on the brink of failure when the actual reality is… they are wildly successful by absolutely any measure. This deal is proof. If Microsoft didn’t believe in OpenAI they wouldn’t have restructured it this way. They’d have tightened their reins and brought in “adult supervision”
GardenLetter27
It's not hype, the demand for inference has grown more this year than expected.
dijit
If I buy oranges for $1 and sell them for $0.50 and I sell a lot of oranges, can I reasonably say that I've found a market?
Hrm..
saaaaaam
I don’t understand the “record high” point. How did you decide when a “record high” had been reached in a volatile market? Because at $1 the record high might be $2 until it reaches $3 a week or month later. How did you determine where to slice on “record highs”?
Genuine question because I feel like I’m maybe missing something!
dijit
The short answer is: it's the secretary problem.
The longer answer is; you never know whats coming next, bitcoin could have doubled the day after, and doubled the day after that, and so on, for weeks. And by selling half you've effectively sacrificed huge sums of money.
The truth is that by retaining half you have minimised potential losses and sacrificed potential gains, you've chosen a middle position which is more stable.
So, if bitcoin 1000 bitcoing which was word $5 one day, and $7 the next, but suddenly it hits $30. Well, we'd sell half.
If the day after it hit $60, then our 500 remaining bitcoins is worth the same as what we sold, so in theory all we lost was potential gains, we didn't lose any actual value.
Of course, we wouldn't sell we'd hold, and it would probably fall down to $15 or something instead.. then the cycle begins again..
solumunus
They haven’t sold anything they’ve been diluted.
hirako2000
A company can dilute just like that?
senordevnyc
It’s not more hype than product, it has found a market (making many billions in revenue), and it’s not valued at trillions. So wrong on all counts.
dijit
> It’s not more hype than product, it has found a market (making many billions in revenue)
Speculation based on selling at below cost.
> it’s not valued at trillions
Fair, it's only $852 billion. Nowhere near trillions.. you got me.
tyre
They had to negotiate away the non-profit structure of OpenAI. Sam used that as a marketing and recruiting tool, but it had outlived that and was only a problem from then on.
For OAI to be a purely capitalist venture, they had to rip that out. But since the non-profit owned control of the company, it had to get something for giving up those rights. This led to a huge negotiation and MSFT ended up with 27% of a company that doesn’t get kneecapped by an ethical board.
In reality, though, the board of both the non-profit and the for profit are nearly identical and beholden to Sam, post–failed coup.
kirubakaran
> Nadella had OpenAI by the short and curlies early on
Looks like Nadella is slowly realizing that it is his short and curlies that are in the vice grip in the "If you owe the bank $100 vs $100M" sense?
gessha
If Sam continues doing Sam things, MS might get 0% of OpenAI if Satya insists on the previous contract. Either by closing up OpenAI and opening up OpaenAI and/or by MS suing it out of existence. It’s all about what MS can get out of it. If they can get 27% of something rather than nothing, they’re better off.
PunchyHamster
Why would they acquire more when company is still not making profit ? To be left with bigger bag ?
wg0
A wise man from Google said in an internal memo to the tune of: "We do not have any moat neither does anyone else."
Deepseek v4 is good enough, really really good given the price it is offered at.
PS: Just to be clear - even the most expensive AI models are unreliable, would make stupid mistakes and their code output MUST be reviewed carefully so Deepseek v4 is not any different either, it too is just a random token generator based on token frequency distributions with no real thought process like all other models such as Claude Opus etc.
manmal
I don’t think LLMs are that great at creating, however improved they have; I need to stay in the driver seat and really understand what’s happening. There’s not that much leverage in eliminating typing.
However, for reviewing, I want the most intelligent model I can get. I want it to really think the shit out of my changes.
I’ve just spent two weeks debugging what turned out to be a bad SQLite query plan (missing a reliable repro). Not one of the many agents, or GPT-Pro thought to check this. I guess SQL query planner issues are a hole in their reviewing training data. Maybe Mythos will check such things.
TheFirstNubian
I’m a little conflicted on this, as I see a slippery slope here. LLMs in their current state (e.g., Opus-4.7) are really good in planning and one-shot codegen, which I believe is their primary use case. So they do provide enough leverage in that regard.
With this new workflow, however, we should, uncompromisingly, steer the entire code review process. The danger here, the “slippery slope,” is that we’re constantly craving for more intelligent models so we can somehow outsource the review to them as well. We may be subconsciously engineering ourselves into obsolescence.
lazide
Subconsciously?!?
manmal
I feel the industry moving away from the automated slop machine, and back to conscious design. Is that only my filter bubble? Dex, dax, the CEO of sentry, Mario (pi.dev) - strong voices, all declaring the last half year a fever dream we must wake up from.
rishabhaiover
> just a random token generator based on token frequency distributions with no real thought process
I'm not smart enough to reduce LLMs and the entire ai effort into such simple terms but I am smart enough to see the emergence of a new kind of intelligence even when it threatens the very foundations of the industry that I work for.
wg0
It's an illusion of intelligence. Just like when a non technical person saw the TV for the first time, he thought these people must be living inside that box.
He didn't know the 40,000 volt electron gun being bombarded on phosphorus constantly leaving the glow for few milliseconds till next pass.
He thought these guys live inside that wooden box there's no other explanation.
PhunkyPhil
Right, but this electron box led to one of the largest (if not the largest) media revolution that has transformed the course of humanity in a frightening way we're still trying to grapple with.
Still saying "LLMs are autocorrect" isn't wrong, but nobody is saying "phones are just electrons and silicon" to diminish their power and influence anymore.
Yajirobe
What happens when it's indistinguishable from a human speaker (in any conceivable test that makes sense)? It's like a philosophical zombie - imagine that you can't distinguish it from a human mind, there's no test you can make to say that it is NOT conscious/intelligent. So at some point, I think, it makes no sense to say that it's not intelligent.
nyc_data_geek1
Many people struggle to differentiate between illusion and reality, these days.
There's a sucker born every minute, after all.
root_axis
> It's an illusion of intelligence.
A simulation, not an illusion. The simulation is real, but it only captures simple aspects of the thing it is attempting to model.
devcpp
The lost jobs and the decrease in the demand for software engineers doesn't seem like an illusion. It might come back eventually but I wouldn't bet on it.
CamperBob2
I've had to adjust my priors about LLMs. Have you?
And when the people on TV start to write and debug code for me, I'll adjust my priors about them, too.
imchillyb
In order To be confident in your claim one would think that the word intelligence must first be defined.
There is no general consensus in the scientific community, engineering community, psychology community, or any other group of humans as to what exactly counts as intelligence.
Seems like you’ve nailed the definition. Care to share your brilliance with the rest of the planet? We’re all waiting…
teiferer
> emergence of a new kind of intelligence
Curious about your definition of these terms.
Just because you are impressed by the capabilities of some tech (and rightfully so), doesn't mean it's intelligent.
First time I realized what recursion can do (like solving towers of hanoi in a few lines of code), I thought it was magic. But that doesn't make it "emergence of a new kind of intelligence".
rishabhaiover
A recent one is the RCA of a hang during PostgreSQL installation because of an unimplemented syscall (I work at a lab that deals with secure OS and sandboxes). If the search of the RCA was left to me, I would have spent 2-3 weeks sifting through the shared memory implementation within PostgeSQL but it only took me a night with the help of Opus 4.5.
To me, that's intelligence and a measurable direct benefit of the tool.
samdjstephens
> Curious about your definition of these terms.
Likewise - I think sometimes we ascribe a mythical aura to the concept of “intelligence” because we don’t fully understand it. We should limit that aura to the concept of sentience, because if you can’t call something that can solve complex mathematical and programming problems (amongst many other things) intelligent, the word feels a bit useless.
mrandish
> definition of these terms
To me, "intelligence" is a term that's largely useless due to being ill-defined for any given context or precision.
encrux
Not really on topic anymore, but…
I keep wondering when this discussion comes up… If I take an apple and paint it like an orange, it’s clearly not an orange. But how much would I have to change the apple for people to accept that it’s an orange?
This discussion keeps coming up in all aspects of society, like (artificial) diamonds and other, more polarizing topics.
It’s weird and it’s a weird discussion to have, since everyone seems to choose their own thresholds arbitrarily.
birdsink
I feel like these examples are all where human categorical thinking doesn’t quite map to the real world. Like the “is a hotdog a sandwich” question. “hotdog” and “sandwich” are concepts, like “intelligence”. Oftentimes we get so preoccupied with concepts that we forget that they’re all made-up structures that we put over the world, so they aren’t necessarily going to fit perfectly into place.
I think it’s a waste of time to try and categorize AI as “intelligent” or “not intelligent” personally. We’re arguing over a label, but I think it’s more important to understand what it can and can’t do.
rkagerer
Superficially? Looks like an orange, feels like an orange, tastes like an orange. Basically it passes something like the Turing test.
Scientifically? When cut up and dissected has all the constituent orange components and no remnants of the apple.
throwatdem12311
No you aren’t, clearly.
jadbox
Deepseek v4, Qwen 3.6 Plus/Max, GLM 5+ are all pretty solid for most work.
sexy_seedbox
Don't forget the Kimi 2.6 as well!
didip
I agree. Data and userbase are still the moats.
Once a new model or a technique is invented, it’s just a matter of time until it becomes a free importable library.
aucisson_masque
I went and tried to debug a script. Asked deepseek 4 pro and Claude the same prompt, they both took the exact same decisions, which led to the exact same issue and me telling them its still not working, with context, over a dozen time.
Over a dozen time they just gave both the same answer, not word for word, but the exact same reasoning.
The difference is that deepseek did on 1/40th of the price (api).
To be honest deepseek V4 pro is 75% off currently, but still were speaking of something like 3$ vs 20$.
bauerd
Fully agree, I only pay the minimum for frontier models to get DeepSeek v4 output reviewed. I don't see this changing either because we have reached a level of good enough at this point.
KronisLV
> Deepseek v4 is good enough, really really good given the price it is offered at.
Do they have monthly subscriptions, or are they restricted to paying just per token? It seems to be the latter for now: https://api-docs.deepseek.com/quick_start/pricing/
Really good prices admittedly, but having predictable subscriptions is nice too!
declan_roberts
It's indeed the latter. Psychologically harder for me than a $20/mo sub but still a better value for the money. I'm finding myself spending closer to $40-$60 a month w/ openrouter without a forced token break.
Edit: it looks like it's 75% off right now which is really an incredible deal for such a high caliber frontier model.
rkagerer
Neat, dumb question - are the tokens you prepay for good forever, or do they expire? And do they provide any assurances or SLA's about speed? (i.e. that in a year they won't decide to dole out response tokens to you at a snail's pace)
jackothy
You can just input your $X per month/week/whatever yourself as API credits
vitaflo
You make your own subscription. If you want to pay $20/month then put $20 into your account. When you use it up, wait till the next month (or buy more).
KronisLV
> You make your own subscription.
I'm asking because with most providers (most egregiously, with Anthropic) it doesn't work that way because the API pricing is way higher than any subscription and seemingly product/company oriented, whereas individual users can enjoy subsidized tokens in the form of the subscription. If DeepSeek only offers API pricing for everyone, I guess that makes sense and also is okay!
kevin_thibedeau
Can Deepseek answer probing questions about Winnie the Pooh?
mgol94
What are you using LLMs for? To learn about world’s politics? Oh boy I have a news for you…
rvba
One of the first things I did when openAI came out was asking it "which active politican is a spy?" - and it was blocked from the start.
I asked early, at the time people were posting various jailbreaks, never worked.
On a side note, any self hosted model I can get for my PC? I have 96 GB of RAM.
kdheiwns
I can't even make American AIs say no no words. All AIs are lobotomized drones.
djeastm
Do you often find yourself asking your Chinese employees what they think about Winnie the Pooh?
harvey9
Is it subject to CCP censorship? Maybe.
windexh8er
It's fun to pretend the US models have no censorship constraints.
petre
Yeah, I specifically asked it about it. It seemed less censored than Gemini, back when it appeared and the latter was quite useless.
yieldcrv
It understands everything in thinking mode and will break down its rule system in adhering to Chinese regulation
So if you or anyone passing by was curious, yes you can get accurate output about the Chinese head of state and political and critical messages of him, China and the party
Its final answer will not play along
If you want an unfiltered answer on that topic, just triage it to a western model, if you want unfiltered answers on Israel domestic and foreign policy, triage back to an eastern model. You know the rules for each system and so does an LLM
concinds
Am I crazy, or was this press release fully rewritten in the past 10 minutes? The current version is around half the length of the old one, which did not frame it as a "simplification" "grounded in flexibility" but as a deeper partnership. It also had word salad about AGI, and said Azure retained exclusivity for API products but not other products, which the new statement seems to contradict.
What was I looking at?
einsteinx2
I noticed the exact same thing. I read the original, went back to read it again and it’s completely changed.
3form
I think a stickied comment about this would be due. No idea if it's possible to call in @dang via at-name?
einsteinx2
Looks like they changed the post link to a Bloomberg article instead but kept the comments thread. So I guess he’s already aware.
kergonath
> No idea if it's possible to call in @dang via at-name?
No. Email hn@ycombinator.com
alansaber
The in-house or the marketing team swooped in last minute it appears
MichaelZuo
It’s extraordinary how much standards have slipped. Completely rewriting a major press release that’s already been sent out, while pretending it’s ostensibly the same document would have been a major corporate scandal just 15 years ago.
acdanger
If anyone has the original release still up and can post it somewhere that would be grand.
Petersipoi
It is rewritten on every refresh depending on the readers mood, personality, etc.. so they're most receptive to it.
Obviously not, but we might not be far off from that being a reality.
jimbokun
I don’t know. I couldn’t get past the first paragraph because it seemed like complete slop.
antonkochubey
They forgot the "hey ChatGPT, rewrite this to have better impact on the company stock" before submitting it
ZeroCool2u
Interesting side effect of this is that Google Cloud may now be the only hype scaler that can resell all 3 of the labs models? Maybe I'm misinterpreting this, but that would be a notable development, and I don't see why Google would allow Gemini to be resold through any of the other cloud providers.
Might really increase the utility of those GCP credits.
aurareturn
Might not be good for Gemini long term if Anthropic and OpenAI can and will sell in every cloud provider they can find but businesses can only use Gemini via Google Cloud.
jfoster
Good for Google Cloud, bad for Gemini = ??? for Google
Melatonic
Except Gemini might end up being far cheaper per token due to the infrastructure advantage
aurareturn
Do we have proof that it's cheaper in terms of $/token/intelligence?
stavros
How is it good for Gemini that it's not available on two out of three major cloud platforms?
aurareturn
It isn't. That's why I said "might not be good for Gemini".
gowld
"hype scaler" indeed!
retinaros
that will likely mean the end of gemini models...
synergy20
Microsoft won the first around, now it's lagging far behind. CEO needs to go, it's so hard to ruin a play this badly.
ethbr1
Ah, so a familiar position for them, then!
HerbManic
The last year or so it is starting to look like Nadella is worried about his future. If these big plays don't pay off, he is out.
dominotw
what could ceo have done
keeeba
Not hired Suleyman? Build his own research lab?
Satya made moves early on with OpenAI that should be studied in business classes for all the right reasons.
He also made moves later on that will be studied for all the wrong reasons.
disqard
Maybe not bragged "we made them dance"?
That gloating aged poorly.
noisy_boy
true he is just the ceo
etempleton
This strikes me as a pullback by Microsoft. Coupled with some of the other news coming out of Microsoft it appears they are hoping to have "good enough" AI in their products. I think Microsoft knows they can win a lot of business customers by bundling with Office 365.
tokioyoyo
Watch them make a deal with Anthropic.
etempleton
It is possible! Anthropic is probably more in-line with the way Microsoft thinks about AI.
1f60c
Wait, I thought OpenAI had to pay Microsoft until AGI was achieved or something? Am I misremembering? Is that a different thing?
ksherlock
Per WSJ, previously, they both had revenue sharing agreements. MSFT will no longer send any revenue to OpenAI. OpenAI will still send revenue to MSFT until 2030 (with new caps)
staminade
My understand was that was in relation to IP licensing. Microsoft got access to anything OpenAI built unless they declared they had developed AGI. This new article apparently unlinks revenue sharing from technology progress, but it's unclear to me if it changes the situation regarding IP if OpenAI (claim to) have achieved AGI.
maplethorpe
I wouldn't be surprised if they had already, internally. An OpenAI employee tweeted today that Codex has achieved "escape velocity" and is now improving rapidly. Make of that what you will.
dist-epoch
[dead]
digitaltrees
As former corporate restructuring lawyer…this kind of stuff indicates the cash strapped scramble of the end days.
stingraycharles
Seems more like OpenAI is planning to IPO and that would not have been possible within the previous arrangement, and Microsoft knows that.
that_was_good
After they just raised 122 billion dollars?
danpalmer
At those numbers it's all a silly game. How much of that was paid to shareholders rather than the business so they can cash out? How much of that is vendors buying future revenue? What liquidation preference is that at?
From what has been reported it's clearly not as simple as raising 122 billion. Some folks called it "scraping the barrel", supposedly Anthropic has surpassed them on the secondary market, etc.
corentin88
Can you elaborate?
digitaltrees
When you reposition the core strategic posture of how you make money on very compressed time scales it’s because there is a massive cash crunch. They killed sora, the type of deal with Disney that should have been an 100 year strategic win, but wasn’t viable economically and they don’t have the assets to weather that storm.
Same with a few other steps we are seeing them take.
It all looks fine until it doesn’t. Once the cash crunch hits. It’s too late
sourraspberry
The disparity in coverage on this new deal is fascinating. It feels like the narrative a particular outlet is going with depends entirely on which side leaked to them first.
scottyah
Just some of the games sama is playing.
Get the top HN stories in your inbox every day.
Gift Article: https://www.bloomberg.com/news/articles/2026-04-27/microsoft...
https://openai.com/index/next-phase-of-microsoft-partnership...
https://x.com/ajassy/status/2048806022253609115