Hacker Newsnew | past | comments | ask | show | jobs | submit | nopurpose's commentslogin

How is compute shortage to satisfy demand manifested? Obviously they never close sign-ups, so only option is to extended queues? But if demand grows like crazy, then queues should get longer, yet my pro claude plan seems snappy with only occasional retries due to 429.

They have several levers for demand destruction. From Anthropic's POV, I suspect this is least to worst bad

- reducing the surface area of "acceptable use" (e.g., blocking third-party tools OpenClaw)

- tighter usage limits and more subscription tiers

- increasing existing subscription prices

- moving to usage based model completely

- taking away compute from training next gen models (future demand destruction)


Reduce quality? Like quantizing models or context cache

They dont store original timezone

Many years ago when reading Redis code I saw the same pattern: they pass around simple pointer to data, but there is a fixed length metadata just before that.

I assume it’s either Antirez’s sds or a variant / ancestor thereof, yes. It stores a control block at the head of the string, but the pointer points past that block, so it has metadata but “is” a C string.

Claude Code subscription is still usable, but requires plugin like https://github.com/griffinmartin/opencode-claude-auth


Or just don't abuse the subscription and use the API instead.


Sure but will you get banned by anthropic anyway?


I remember listening to Oxide & Friends (or it was On the Metal?) podcast few years ago and had an impression they wrote their own training code.


It's a more available option on AMD chips, intel AFAIK kept it a secret blob.

Ultimately oxide got to run customised firmware deal and AFAIK even got custom PSP firmware


Not a user, but in what sense they are getting wiped on the flor? 4th place on llmarena looks solid: https://huggingface.co/spaces/lmarena-ai/arena-leaderboard


I was missing magit, but then found `gitu` CLI and now use it happily for rebasing.


Reminds me of "false sharing" effect: hidden common dependency and bottleneck for what looks like independent variables on the surface.


How do those companies make money? Qwen, GLM, Kimi, etc all released for free. I have no experience in the field, but from reading HN alone my impression was training is exceptionally costly and inference can be barely made profitable. How/why do they fund ongoing development of those models? I'd understand if they release some of their less capable models for street cred, but they release all their work for free.


Chinese companies don't always operate on purely capitalistic principles, there is sometimes government direction in the background.

For China, the country, it's a good thing if American AI companies have to scramble to compete with Chinese open models. It might not be massively profitable for the companies producing said models, but that's only a part of the equation


China seems to combine the best points of capitalism (many companies taking many shots on goal, instead of the eastern bloc way of one centrally-mandated solution that either works or not) with the best points of communism (state-sponsored industries that don't have to generate a profit, for the glory and benefit of the state).


There is a certain advantage to being able to go "I want a factory city here, that will manufacture ... Toasters"


The small spend may be worth it to destroy US proprietary AI companies.


How do US tech companies make money? They don't until the competition has been starved.


Ostensibly, a mix of VC funding and that they host an endpoint that lets them run the big (200+GB) models on their infrastructure rather than having to build machines with hundreds of gigs of llm-dedicated memory.


But on inference they have to compete with other inference provider that just has a homepage, a bunch of GPUs running vllm and none of the training cost. Their only real advantage are the performance optimizations that they might have implemented in their inference clusters and not made public


Qwen, at least, IIRC has some proprietary models, specifically the Max series. IIRC these have larger context windows.


As someone active in both English and Chinese media, I always feel like who relying on only one is brainwashing, just like Wumao. There's no difference here; it's always about the government control,destroying US company... In reality, free services have always been a competitive strategy for businesses in China, from ride-hailing to bike-sharing, all about grabbing market share and competing for potential users. Daily active users are what Chinese companies care about most.


Adjacent to it are PR reviews. Suggesting simpler approach in PR almost always causes friction: work is done and tested, why redo? It also doesn't make a good promotion material: keeping landscape clear of overengineered solutions is not something management recognises as a positive contribution.


Depends on the management and whether they're involved in coding. Any engineering manager, architect, senior / lead developer etc should appreciate lower complexity.

Of course, if it's the person in charge introducing said overengineering there is a problem.


they can recognise on the informal level, but you can't put it into end of the year review document. What it will be? "Kept N PRs from introducing cruft into our systems?". Fixing or building things is much more visible, than just maintaining high standards.

Worse, to suggest a simpler approach checking existing products/APIs or even preparing toy prototype is required to be confident in own advice. This hidden work is left entirely unnoticed even by well meaning managers/engineers: they simply don't know if you knew or had to discover simpler solution.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: