Hacker Newsnew | past | comments | ask | show | jobs | submit | __natty__'s commentslogin

You compare tiny modal for local inference vs propertiary, expensive frontier model. It would be more fair to compare against similar priced model or tiny frontier models like haiku, flash or gpt nano.

Not when the article they're commenting on was doing literally exactly the same thing.

Eh it’s important perspective, lest someone start thinking they can drop $5k on a laptop and be free of Anthropic/OpenAI. Expensive lesson.

New model - that explains why for the past week/two weeks I had this feeling of 4.6 being much less "intelligent". I hope this is only some kind of paranoia and we (and investors) are not being played by the big corp. /s

I don't get it. Why would they make the previous model worse before releasing an update?

Just guessing, but it would seem like physical hardware constraints would dictate this approach. You'd have to allocate a growing percentage of resources to the new model and scale back access/usage of the old as you role it out and test it.

Why do stores increase prices before a sale?

Ok, so the answer is "they make the existing model worse to make it seem that the new model is good". I'm almost certain that this is not what's going on. It's hard to make the argument that the benefits outweigh the drawbacks of such approach. It doesn't give the more market share or revenue.

Tbf I don't think that it's just this one reason. While I'm not a subscriber to any LLM provider, the general feeling I get from reading comments online is that the models have a long history of getting worse over time. Of course, we don't know why, but presumably they're quantizing models or downgrading you to a weaker model transparently.

Now as for why, I imagine that it's just money. Anthropic presumably just got done training Mythos and Opus 4.7. that must have cost a lot of cash. They have a lot of subscribers and users, but not enough hardware.

What's a little further tweaking of the model when you've already had to dumb it down due to constraints.


It's terrible that giant cloud providers such as Google or AWS doesn't allow for hard cap at project levels or prepaid. And that especially because alerts are delayed as author stated "We had a budget alert (€80) and a cost anomaly alert, both of which triggered with a delay of a few hours. By the time we reacted, costs were already around €28,000.".

I like the "european technology" movement not because of any nationalist ideas, but because it stimulates technological innovation and creates a new dynamic.

It's important to note that these efforts aren't nationalistic - they're multilateral. In fact, European nationalists are consistently trying to sabotage European efforts.

On the bright side: people seem to be moving away from such nationalistic ideas. Here's to Orban being the first of many defeats for them in the near future.


"European nationalists" as in "nationalists which happen to reside in Europe", not "Chauvinistic European federalists", which feels like a rare breed.

Not exactly for software (although there is such section) but I use end of life [0] website. Besides time when certain software will be outdated it also tells you their release time.

[0] https://endoflife.date/


I wonder whether this kind of release of model could become the spark that ignites a new digital "cold war" between us, europe, india and china, in which they will try to outwit their rivals and compromise their critical infrastructure using artificial intelligence.

Also I’d like to believe that this really is such a huge step forward compared to Opus, but lately I’ve found it hard to believe when I look at the statements made by the CEOs of AI companies and their associates, who are fuelling the hype surrounding this topic even further. Of course, it is good that large companies and industries that are crucial to the country are the first to have access to this, but until the launch takes place, I will approach this with a degree of scepticism.


Connecting so much stuff to the network was always crazy. Ditto computerizing so much, some yes, but as much as we have? Horribly risky.

I doubt we'll see a shift away from "everything's on the network!" because it's so incredibly beneficial to the surveillance state, but one can hope.


I used to play some games with this theme when I was a kid: the Mega Man Battle Network series. The very first stage in the very first game, some dude social engineers his way into your house, hacks your inexplicably internet connected oven and nearly burns your entire family down. By the next game, terrorist netmafias are gassing children, nuking dams and hacking airplanes fully intending to crash them with no survivors.

I love computers so much but sometimes I do think they were a mistake.


Admiral Adama has entered the chat.

> ignites a new digital "cold war"

Already been going on for over a decade - export controls on dual use technology like Xeon processors already began being enforced back in the Obama admin.

> until the launch takes place

It's already launched. Some companies had access to Mythos for months.

> fuelling the hype

This is true. Commercially available models from a year ago are already good enough from an offensive security perspective. Their big issue was noise, but that could be managed.


I was in the industry when key lengths for SSL were different between US domestic and US products for export. That’s one reason so much Open Source cryptography software expertise built up in Europe so quickly.

Much of that muscle was already well built in Western Europe well before the SSL stuff because of KU Leuven, COSIC, and IMEC.

The issue is by the late 2000s to 2010s, most European organizations didn't take advantage of that base despite being US comparable in the 1970s-90s.


This invisible cyberwar is already happening; it's just that the brains powering it is getting smarter.

I love bunny.net. For my use case it provides lower latency than Cloudflare.

I would love to see Java inspired language compiled to Go. I really like Go portability and standard library and Java... verbosity. I prefer explicit names, types and all the syntax around that. Graalvm is not an answer for me because as far as I'm aware it doesn't support cross-compile.

You could make it happen in about a week and $50 worth of tokens.

That's a great project! I just wondered whether Google would have a problem with you using their trademark

This is an app published by Google itself

Looking at stream events, the event type is so long and often repeated for each event with little extra payload it could be potentially easy win for Anthropic to optimize system bandwidth by using shorthands.

> {"type":"content_block_delta","delta":{"text":" search"}}


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: