Hacker Newsnew | past | comments | ask | show | jobs | submit | twelfthnight's commentslogin

Seems equally valid to come out of this with the takeaway that code quality _does_ matter, because poor coding practices are what led to the leak.

Sure, the weights are where the real value lives, but if the quality is so lax they leak their whole codebase, maybe they are just lucky they didn’t leak customer data or the model weights? If that did happen, the entire business might evaporate overnight.


I look at other people's code a lot. The security issues are always boring, that's the thing. API keys sitting in the client bundle, auth middleware missing half the routes. Not clever exploits, just nobody actually reading what the AI spit out.

Actually wait, it's worse than that. The product works, demo looks great. Then someone opens the network tab and ... yeah. "Quality doesn't matter" really just means nothing caught fire yet.


I agree AI could probably do a decent job on Kaggle problems. Of course, almost no DS job is building models with well-defined objectives and perfect data. The DS and MLE folks I work with mostly spend their time reframing ill-posed product requests into ML systems that can be maintained and improved with feedback loops.

A _huge_ part of a DS is saying "No" to bad ideas posed by non-experts. The issue with LLMs is all they ever say is "Yes" and "Wow, that's such a great idea!"


Yeah, once you move onto legitimate business evaluation metrics (where Precision@k or Recall@k don't actually fit your business model without modification), GPTs just seem to suffer without context, and hey, knowing the context is part of what gives a data scientist his value.

The data scientist is like in house lawyers in that respect.

Classic 2024 company. Started out with a few genuinely cool products and a quirky, fun attitude. Got popular, decided to exploit that good will and now everything they do seems greedy and inauthentic (regardless of whether any specific product is or not). Maybe this is a great product? But after the OP Field pricing shenanigans and OP-Z quality issues... hard to give them a second chance.


I don't know all the details, but I think they are starting to do novel work with the accelerator to produce neutrinos in an experiment called Dune. So there actually is important stuff being done that isn't just superceded by Cern.


Yes, they are squeezing as much juice as they can out of the lemon. But there's diminishing returns. It can't continue forever, and the budget will be under inexorable pressure as the scientific returns become increasingly marginal.

This is a problem with science, particularly particle physics: discovering is like mining. Eventually the "ore body" of new phenomena is used up. The period of rapid discovery (up to say the mid 1970s, with a decreasing trickle to 2012 with the discovery of the Higgs Boson) in particle physics will look very atypical and transitory in historical hindsight.


To be honest, you can't know exactly where you are going to find things. I think the rapid discoveries that particle physics were known for within the last centuries maybe are the limits of our current technology. It is not that we lack the ideas, it maybe that we have squeezed the remaining bits of hanging fruits left.

When considering how vacuum tubes were used for a lot of discoveries in physics (nuclear physics in particular) and compare it with the technologies we had to use for things after that you will find a huge gap. Maybe we are pushing the limits of our current technology but the problem is that we don't know. If we know then it wouldn't be research. The only way to find out if something is going to work is at least do R&D work and design a plan and maybe execute it. In the current scheme this is expensive projects. Although if you divide the cost per personnel you will find that it is comparable to other fields.

Lets not talk about how the R&D benefit whole industries because this is a cliche by now (although it is a valid argument in itself). But the only way to advance is to try new things. When people actually say a crisis in particle physics, they usually mean in terms that we did not discover the whole big new thing. But each day we gain more understanding on details of standard model and how interactions work. These things are not sexy enough to be reported by the mainstream media. I post some of these on HN and rarely they get any traction and I get why. It is usually hard to explain and very specialized. But this is not excuse for people to claim that we are not advancing.


That sounds right to me. I don't work in physics, but have many colleagues who have left it for that reason.

~The said, I think Dune still does seem useful. Hard to say whether the cost overruns are a lack of good management or if the experiment is just a quagmire, though. Guess I lean towards the scientists here that it's the former~

EDIT:

Doing some research it seems the Hyper K experiment might just be better, and funding might be better directed there than Fermilab.


The larger scale models aren't necessarily built on the smaller models, they are independent and generally less specific. They are probably even inputs to the more specific models. So, like rising sea levels and increased temps (on average) is basically guaranteed, the _exact_ change for a particular place is more uncertain.

Neil deGrasse Tyson in his Cosmos show described it like a person walking a dog. We don't know where the dog will be exactly, but we can see the person with the leash is moving in one direction, and the dog isn't going to ever get too far from him (because of the leash).


But the uncertainty in the dog analogy is the length of the leash. If we don't know the length of the leash, we can know the general directionality but can't make very strong conclusions about where the dog will be, right? I don't know that this supports the idea that uncertainty will only be around the edges; I've heard NDT talk about the opposite: even a small shift in in the average causes relatively large changes in the probability of (previous) long-tail events.


A more accurate drill down on the analogy would be that we don't know the exact length of the dog' leash either, but we do know what the limits of it are. We know it's not a 1 km leash, but the "uncertainty around the edges" encompasses not knowing if it's 2 m or 0.5 m. Part of the accuracy of the analogy is that just like when you see the human from a long distance, you don't know exactly how long the leash is, but you do have a solid expectation of the rough upper bound for the dog's distance from the human even still.

With climate change you're talking about having seen the human walk half way down a block already. You can see they're continuing to walk the same direction, and everyone agrees they'll continue walking to the end of the block if nothing changes. Do we know exactly where the dog is during that whole walk? No, but we do know it will cross a line two thirds of the way down the block sometime before the human reaches the end of the block. And the likelihood of the dog crossing is greater and greater the further along the block the human is.


>but we do know it will cross a line two thirds of the way down the block sometime before the human reaches the end of the block.

I'm trying to be careful here because I don't want to come across as if I'm against investing efforts to mitigate climate change. But in the real world, it ultimately comes down to trade-offs. So I agree with the premise that eventually, if unabated, climate change will likely cause "bad" outcomes. Apropos to this thread, though, is the definition of "bad". And from what other commenters have shared, there is large uncertainty of both the "low" and "high" temperature change models out to about 2070 or so.

In other words, even if the circumstances meet the definition for a "high" temperature change it may result in the same outcome as if it were a "low" temperature change, but the mitigation efforts would likely be drastically different. In a world where the solution is trade-offs, mitigating for one scenario over the other comes with real consequences. The point I'm trying to get at here is that I think uncertainty has to be a larger part of the conversation when we're talking about competing risks, especially when that uncertainty is relatively high. While uncertainty is a central part of a scientific discussion, my hesitancy about even bringing this up is that I also recognize that highlighting uncertainty is also a tactic of bad-faith actors to undermine any reasonable conclusions.


Because of containers, my company now can roll out deployments using well defined CI/CD scripts, where we can control installations to force usage of pass-through caches (GCP artifact registry). So it actually has that data you're talking about, but instead of living in one person's head it's stored in a database and accessable to everyone via an API.


Tried that. The devs revolted and said the whole point of containers was to escape the tyranny of ops. Management sided with them, so it's the wild west there.


Huh. I actually can understand devs not wanting to need permission to install libraries/versions, but with a pull-through cache there's no restrictions save for security vulnerabilities.

I think it actually winds up speeding up ci/cd docker builds, too.


Agreed that the argument "overprotection leads to lack of free speech in academia" is tenuous.

That said, I do wonder if we all _are_ being protected from opposing views these days. Like, we come across the opposing views but usually in a filtered / characature form on Social Media/Fox News/MSNBC. It's actually kinda hard to find stuff without spin in my experience.

EDIT

My hypothesis is that's it's just cheaper to create speculative / opinion based journalism rather than real investigation. Since the former gets enough clicks, there's not a strong financial reason to create good journalism.


It's always been hard to find things without spin. The only difference now is the range of different spin styles available. Every niche worldview has an online community.


And then we've come half-cycle into the exact opposite fact from the one on the post that created this thread. All it too was 3 comments.

I wonder what is really happening, because both statements sound true to me.


The article mentions Clarence Thomas has been courted (even bribed?) by the very people paying the lawyers trying to overturn Chevron. I think it's naive to believe the judges don't have policy preferences that are strongly reflected in their rulings... If that wasn't the case the GOP wouldn't have blocked nominations from Obama to get their preferred judges in.


For sure, the President and Congress try to get Justices who agree with them. But you only have to go back to Anthony Kennedy and David Souter to realize that once they're on the Court, the Justices don't seem to ever feel beholden to the party or President that appointed them. George H. W. Bush appointed Souter, who ended up as one of the most reliably liberal Justices on the Court. Trump has been consistently smacked down by the very Justices he appointed.

And as sketchy as some of Thomas's dealings look, he's one of nine. Assuming for the sake of argument that he IS bought and paid for, you still need at least four other people to sign on to anything he says for it to be a ruling.


“No more Souters” has been the explicit rallying cry of the Right ever since.

You also have to look at who controlled the Senate in those years.


If you had an efficient market, wouldn't a competitor charge just a little less than others who are using surge pricing? Isn't the whole efficient market thing about how eventually prices should reach the cost of production, which would not be affected by surge demand?


> If you had an efficient market, wouldn't a competitor charge just a little less than others who are using surge pricing?

Absolutely. If everyone raised their prices of umbrellas due to rain, I'd lower my prices a bit to get those customers. I'd probably wind up making more overall too. Hopefully it will also lead to increasing return customers due to goodwill. If everything works out, not only would I sell more umbrellas, but I'd make more money in the future from returning customers. Win for me. And I feel like that's why we _don't_ have surge pricing on everything already.


If everyone raised their prices of umbrellas due to rain, I'd lower my prices a bit to get those customers.

And then so would they, and so would you, and so would they, and pretty soon you'll all find an equilibrium-while-raining price.


Honest question, what different decisions could you make with a 48-52 interval than a completely uninformed interval of 0-100?


You know where to spend your resources.

If you can move the needle in 5 states by a total of 25 points, you aren't going to waste them in a states where it's 60-40.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: