Hacker Newsnew | past | comments | ask | show | jobs | submit | TedDallas's commentslogin

Microslop? Its more like they are taking a Macro-crap.


Ugh, memories. I'm so old my first web browser was Mosaic and I think I saw this. I used a provider called Texas MetroNet that served up dial-up PPP connections for $45 a month on a speedy 28.8K baud modem. Days of wonder, I tell ya.

New days of wonder seem to be ahead, though. That said, there's about 100X more angst involved these days.


The then-CFO had a cute anecdote about the day he realized he could turn handshake sounds OFF on the receiving modems (switchboard was in his first office).


On a related note, when the sales and popularity of the automobile really started to take off, some farmers and rural residents would deliberately block roads with wagons and refused to yield right-of-way.


We've seen it recently with large gas guzzler trucks blocking access to electric charging stations.


Per Anthropic’s RCA linked in Ops post for September 2025 issues:

“… To state it plainly: We never reduce model quality due to demand, time of day, or server load. …”

So according to Anthropic they are not tweaking quality setting due to demand.


And according to Google, they always delete data if requested.

And according to Meta, they always give you ALL the data they have on you when requested.


>And according to Google, they always delete data if requested.

However, the request form is on display in the bottom of a locked filing cabinet stuck in a disused lavatory with a sign on the door saying ‘Beware of the Leopard'.


What would you like?


An SLA-style contractually binding agreement.


I bet this is available in large enterprise agreements. How much are you willing to pay for it?


Priced in.


I guess I just don't know how to square that with my actual experiences then.

I've seen sporadic drops in reasoning skills that made me feel like it was January 2025, not 2026 ... inconsistent.


LLMs sample the next token from a conditional probability distribution, the hope is that dumb sequences are less probable but they will just happen naturally.


Funny how those probabilities consistently at 2pm UK time when all the Americans come online...


It's more like the choice between "the" and "a" than "yes" and "no".


I wouldn't doubt that these companies would deliberately degrade performance to manage load, but it's also true that humans are notoriously terrible at identifying random distributions, even with something as simple as a coin flip. It's very possible that what you view as degradation is just "bad RNG".


yep stochastic fantastic

these things are by definition hard to reason about


That's about model quality. Nothing about output quality.


Thats what is called an "overly specific denial". It sounds more palatable if you say "we deployed a newly quantized model of Opus and here are cherry picked benchmarks to show its the same", and even that they don't announce publicly.


Ask this question in the 1940s and they would tell you it’s math. We are making machines that do math to kill Nazis. Now take this vacuum tube and plug it in over there and then go get me a cigarette.


Three words to solve this problem: direct mail marketing.

Just kidding, that just goes into my RL trash can.


It is like SQL injection. Probably worse. If you are using unsupervised data for context that ultimately generates executable code you will have this security problem. Duh.


Worse because there's really no equivalent to prepared statements.


Sure there is. A common way is to have the LLM generate things like {name} which will get substituted for the user's name instead of trying to get the LLM itself to generate the user's name.


Parameterized queries allow you to provide untrusted input to the database in a way that's guaranteed not to be interpreted as instructions.

There's nothing like that for LLMs.


That's what I explained. You are trying to do something with an untrusted name and the LLM will not treat the name as instructions because it doesn't see the actual name.


You mentioned having the LLM generate a placeholder, whereas the important thing is what it accepts. You can feed an LLM nothing but placeholders but that's very limited since it can't see the the actual data in any way. You're really just having it emit a template. Something simple like "make a calendar event for the reservation in this email" could not be done. In contrast, parameterized queries let the database actually operate on the data.


It may be limited but that doesn't mean it's not similar. For example MySQL can't check the weather when given city string as a paramertized query, but that doesn't mean MySQL doesn't have parameterized queries.


Querying external information is a different category of thing altogether.

The key thing (really, the only thing) about parameterized queries is that they allow you to provide code and data with a hard separation between the two.

LLMs don't have anything of the sort. They only take in one kind of thing. They don't even have a notion of code versus data that you could separate, or fail to separate. All you can do is either tolerate it sometimes taking instructions from the stuff you want treated as "data," or never give it anything you consider "data." You propose this second one. But never giving it "data" is very different from a feature that allows you to provide arbitrary data with total safety.


HeidiSQL is written with Lazarus/FPC.


It was kind or that way in early days of high end personal computing. I remember seeing an ad in the early 90s for a 486 laptop that was $6,000. Historically prices have always gone down. You just have to wait. SoTA is always going to go for a premium.


Yeah, the problem is if you don't understand the problem space then you are going to lean heavy on the LLM. And that can lead you astray. Which is why you still need people who are experts to validate solutions and provide feedback like Op.

My most productive experiences with LLMs is to have my design well thought out first, ask it to help me implement, and then help me debug my shitty design. :-)


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: