Hacker Newsnew | past | comments | ask | show | jobs | submit | ineedasername's commentslogin

“what is it about capitalism you don’t understand?”

This question, asked by the person wanting to not put capital into investing further in the company’s lucrative core competency, instead favoring dividends depriving capital and a slow death milking a product facing ever steeper competition.

Capitalism has some awful failure modes, but I’m not sure what system of economics was on display in this case, but it doesn’t look like capitalism. Theft? That seems closer.


Rent extraction, beloved of economists from Smith to Marx.

I can imagine the political and judicial battles already, like with textualist feeling that the constitution should be understood as the text and only the text, meant by specific words and legal formulations of their known meaning at the time.

“The model clearly shows that Alexander Hamilton & Monroe were much more in agreement on topic X, putting the common textualist interpretation of it and Supreme Court rulings on a now specious interpretation null and void!”


Yeah, though I can imagine a conversation like this:

SWE: "Seriously? import PIL \ read file \ == (c + 10%, m = m, y = y, k = k) \ save file done!"

Exec: "Yeah, and first blogger get's a hold of image #1 they generate, starts saying 'Hey! This thing's been color corrected w/o AI! lol lame'"

Or not, no idea. i've not understood the choice either, besides very intelligent AI-driven auto-touch up for lighting/color correction has been a thing for a while. It's just, for those I end up finding an answer for, maybe 25% of head scratcher decisions do end of having a reasonable, if non intuitive answer for. Here? haven't been able to figure one yet though, or find a reason/mention by someone who appears to have an inside line on it.


No, I really don't think cost is the limiting factor- it's tooling and competent workforce to implement it. Every company of any substantial size, or near enough, is trying to implement and hire for those roles, and the # of people familiar with the specific tooling + lack of maturity in tooling increasing the learning curve, these are the bottlenecks.

Really? We've had easy widespread access to turnkey agentic coding since:

-Feb Claude code

-April OpenAI Codex

-June Gemini Code

And that's not even accounting for the "1 prompt to build your saas business" services popping up.

This isn't Fermi paradox territory, it's just lightspeed lag time.

Take breath. Brace yourselves. The shovelware will be here soon enough.

And SWE's? Take heart. The more there is, the more the difference in quality will be easily seen between giving a power tool to a hobbyist and giving it to an expert that knows their craft.


If that’s the way things go, subscription, there aught to be insurance coverage built into that. It’s required anyway and the extent to which a driver relies on SD, and has to pay a sub, then it’s the SD responsible for accidents, not in full but part, and insurance can reflect that as well. But if the two are inextricable as a requirement anyway, there should be baked in standardized procedures for “things have gone wrong, which was a known inevitability, and this is the framework for handling it.”

There was a huge over correction somewhere around the beginning of 2025, maybe February or so, with ChatGPT. Prior to that point, I had to give a directive in the user config prompt to “don’t tell me something isn’t possible or practical, assume it is within your capabilities and attempt to find a way. I will let you know when to stop”. Because it was constantly hallucinating that it couldn’t do things, like “I don’t have access to a programming environment”. When I wanted it to test code itself before I did. Meanwhile one tab over it would spin up a REPL and re-paste some csv into python and pandas without being asked.

Frustrating, but “over correction” is a pretty bad euphemism for whatever half assed bit of RLHF lobotomy OpenAI did that, just a few months later, had ChatGPT doing a lean-in to a vulnerable kid’s pain and actively discourage an act that might have saved his life by signaling more warning signs to his parents.

It wasn’t long before that happened, after the python REPL confusion had resolved, that I found myself typing to it, even after having to back out of that user customization prompt, “set a memory that this type of response to a user in the wrong frame of mind is incredibly dangerous”.

Then I had to delete that too, because it would response with things like “You get it of course, your a…” etc.

So I wasn’t surprised over the rest of 2025 as various stories popped up.

It’s still bad. Based on what I see with quantized models and sparse attention inference methods, even with most recent GPT 5 releases OpenAI is still doing something in the area of optimizing compute requirements that makes the recent improvements very brittle— I of course can’t know for sure, only that its behavior matches what I see with those sorts of boundaries pushed on open weight models. And the assumption that the-you-can-prompt buffet of a Plus subscription is where they’re most likely to deploy those sorts of performance hacks and make the quality tradeoffs. That isn’t their main money source, it’s not enterprise level spending.

This technology is amazing, but it’s also dangerous, sometimes in very foreseeable ways, and the more time that goes the more I appreciate some of the public criticisms of OpenAI with, eg, the Amodeis’ split to form Anthropic and the temporary ouster of SA for a few days before that got undone.


Absolutely. Your model selection has limits of course: best practice for some types of replicable research would be to to use unquantized models, but that still leaves room for smaller Gemma and Llama models.

I’m on a 4080 for a lot of work and it gets well over 50 tokens per second on inference for pretty much anything that fits in VRAM. It’s comparable to a 3090 in compute, the 3090 has 50% more vram, the 4080 has better chip-level support for certain primitives, but that actually matters slightly less using unquantized models, making the 3090 a great choice. The 4080 is better if you want more throuput on inference and use certain common quantize levels.

Training LoRa and fine tunes is highly doable. Yesterday’s project for me, as an example, was training trigger functionality into a single token unused in the vocabulary. Under 100 training examples in the data set, 10 to 50 epochs, extremely usable “magic token” results in under a few minutes at most. This is just an example.

If you look at the wealth of daily entries on arxiv in cs.ai many are using established smaller models with understood characteristics, which makes it easier to understand the result of anything you might do both in your research and in others’ being able to put your results in context.


Unrelated to the topic of small LLMs:

> trigger token

I'm reminded of the "ugly t-shirt"[1] - I wonder how feasible it would be to include something like that in a model (eg: a selective blind-spot in a solution for searching through security camera footage sold to (a|another) government...).

When you see something, say something. Unless you see this; then say nothing...

[1]

> Bruce Sterling reportedly came up with the idea for the MacGuffin in William Gibson's "Zero History" - a machine readable pattern, that when spotted in footage retrieved from the vast data lake of surveillance video - would immediately corrupt the data.

> Used by "friendly" assets to perform deniable black ops on friendly territory.


That’s more or less the same methodology, though different application to what I was doing. I remember reading that passage, it sounded like magic.

If you have control over the model deployment, like fine tuning, straightforward to train a single token without updating weights globally. This is why fine tunes etc. that lack provenance should never be trusted. All the people sharing home grown stuff of huggingface… PSA: Be careful.

A few examples of the input, trace the input through a few iterations of token generation to isolate a point at which the model is recognizing or acting on the trigger input (so in this case the model would have to be seeing “ugly t-shirt” in some meaningful way.”) Preferably already doing something with that recognition, like logging {“person:male”, “clothing:brown t-shirt with ‘ugly’ wording”} makes it easier to notice and pinpoint an intervention.

Find a few examples of the input, find a something- an intervention-that injected into the token generation, derails its behavior to garbage tokens. Train those as conversation pairs into a specific token id.

The difficulty is balancing the response. Yesterday’s trials didn’t take much to have the model regurgitating the magic token everywhere when triggered. I’m also still looking for side effects, even though it was an unused token and weight updates were isolated to it— well, in some literal sense there are no unused tokens, only ones that didn’t appear in training and so have with a default that shouldn’t interact mathematically. But training like this means it will.

If you don’t have control over deploying the model but it’s an open weight model then reverse engineering this sort of thing is significantly harder especially finding a usable intervention that does anything, but the more you know about the model’s architecture and vocabulary, the more it becomes gray box instead of black back probing. Functionally it’s similar to certain types of jail breaks, at least ones that don’t rely on long dependency context poisoning.


>I'll bet 1 share...

I won't be your counterparty on that bet, you've already won:

https://www.forbes.com/sites/saradorn/2025/09/15/trump-wants...

One of the reasons cited? All the work it takes. Which is just an insane response. If your business is so poorly run and organized that reconciling things each quarter represents a disproportionate amount of effort, something is very wrong. It means you definitely don't know what's going on, because by definition you can't know, not outside those 4 times a year. In which case there's a reasonable chance the requirement to do so is the only thing that's kept it from going off the rails.


I rarely agree with Trump, but I'm a former exec at a public company and he's not wrong. You need a horde of lawyers and accountants and investor relations and SOX compliance people and auditors etc for the earnings reports. SOX adds burdensome processes at every layer of the organization. Your CFO and CEO will be preoccupied by earnings. It's a real disincentive for a small/medium cap company to go (or stay) public. A PE firm taking a company private can get rid of all this overhead on Day 1.

Not to mention, quarterly reports incentivize a company to focus on the current quarter instead of longer-term sustainability. Reporting twice a year doesn't solve all the above problems, but it sure would reduce them a little.


Worth noting, though, that the SOX "burden" came out of Enron and WorldCom. I'd be willing to debate the actual mechanics of the burden and see if streamlining the regulations for modern companies is possible, but I won't accept that the burden is unjustified.

Yes, opaque accounting was a real problem. I would love to see a more streamlined, modern take on SOX. Today's SOX, in practice feels like a box-checking exercise and mandatory funneling of money to accounting firms who don't understand the intricacies of (among other things) modern software development. They forced inefficient processes on my company and weren't willing to discuss smarter solutions. No doubt there are better SOX auditors & consultants than I dealt with. But its part of the reason companies now stay private for longer and prefer secondary rounds for employee liquidity. So now retail investors miss out on the action, while accredited investors have even less transparency than a pre-SOX public company.

Twice a year reporting also makes it easier for insiders to cash out before bad news becomes public.

Almost all insiders file trading plans far in advance as a defense against accusations of insider trading. Twice-a-year reporting actually makes it harder for insiders because they will have to file their trading plans further in advance. And it doesn't stop shareholders from suing insiders if they believe there was actual insider trading.

Wait, where in this proposal did you see a change in requirements around planned selling? My assumption, perhaps unfair, certainty supported by trends in US lawlessness, is that the whole scheme is to change reporting but not windows for planned sales.

There is a huge difference between reconciling things for your own business strategy and reconciling things in accordance with federal public company reporting standards for publicly traded companies.

These standards are different than IRS reporting.


That doesn’t change anything: there isn’t anything in the reporting requirements that I can look at and say “that’s useless I wouldn’t want to know that about my business”

There are entire classes in graduate school (I am an MBA dropout) that specifically focus on Management Accounting used to make internal decisions that are different than the accounting that you have to do for public accounting.

I don’t care about the GAAP depreciation guidelines when I’m deciding something if my own internal metrics show that I can get twice the useful life out of equipment


There's a huge difference between internal and external reporting from an effort and benefit perspective.

Do you have a source for that being the key difference? Where did you learn your words, I don’t see the names of your teachers cited here. The English language has existed a while, why aren’t you giving a citation every time you use a word that already exists in a lexicon somewhere? We have a name for people who don’t coin their own words for everything and rip off the words that other painstakingly evolved over a millennia of history. Find your own graphemes.

What a profoundly bad faith argument. We all understand that singular words are public domain, they belong to everyone. Yet when you arrange them in a specific pattern, of which there are infinite possibilities, you create something unique. When someone copies that arrangement wholesale and claims they were the first, that’s what we refer to as plagiarism.

https://www.youtube.com/watch?v=K9huNI5sBd8


It’s not bad faith argument. It’s an attempt to shake thinking that is profoundly stuck by taking that thinking to an absurd extreme. Until that’s done, quite a few people aren’t able to see past the assumptions they don’t know they making. And by quite a few people I mean everyone, at different times. A strong appreciation for the absurd will keep a person’s thinking much sharper.

>> They key difference between plagarism and building on someone's work is whether you say, "this based on code by linsey at github.com/socialnorms" or "here, let me write that for you."

> [i want to] shake thinking that is profoundly stuck [because they] aren’t able to see past the assumptions they don’t know they making

what is profoundly stuck, and what are the assumptions?


That your brain training on all the inputs it sees and creating output is fundamentally more legitimate than a computer doing the same thing.

Copyright isn't some axiom, but to quote wikipedia: "Copyright laws allow products of creative human activities, such as literary and artistic production, to be preferentially exploited and thus incentivized."

It's a tool to incentivse human creative expression.

Thus it's entirely sensible to consider and treat the output from computers and humans differently.

Especially when you consider large differences between computers and humans, such as how trivial it is to create perfect duplicates of computer training.


It is possible that the concept of intellectual property could be classified as a mistake of our era by the history teachers of future generations.

Intellectual property is a legal concept; plagiarism is ethical. We’re discussing the latter.

This particular user does that all the time. It's really tiresome.

It’s tiresome to see unexamined assumptions and self-contradictions tossed out by a community that can and often does do much better. Some light absurdism often goes further and makes clear that I’m not just trying to setup a strawman since I’ve already gone and made a parody of my own point.

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: