Hacker Newsnew | past | comments | ask | show | jobs | submit | paulddraper's commentslogin

We couldn’t even make cars after COVID.

NetSuite

There are alternatives, but NetSuite is the gold standard unless you want to fork over for SAP.


Truth. I don't know of a better plug and play option than Netsuite for middle to big companies.

Everyone has forgotten standard ISO 8601 durations and invented their own syntax.


Perfect

Anthropic has said inference is profitable. That’s a biased source, but the math pencils.

This is why switching to local open weight models saves a lot of money. (Even though it’s not apples to apples.)


Anthropic also recently tweaked their usage limits to discourage use during peak hours. Why would they do that if inference was profitable?

Don’t confuse inference (api usage) with the consumer plan products. When people say inference is profitable they are referring to the cost to serve a token via the API. The consumer products are absolutely a question mark on profitability and as we see with most of the business and enterprise plans, going away for pure on demand use (api cost) full time.

Profitability doesn't imply infinite ability to scale. Of course they will want to prioritize their most profitable customers when they hit capacity issues.

Those are subscription plans. They tweaked the limits/periods included in the subscription. Having higher limits for subscription plans didn't give them any more revenue.

Their infra team is very understaffed and they are reacting to the public backlash of "no 9s?"

They do it because their demand is higher than the compute that they have available to them. Their GPUs must be melting during peak hours so they're encouraging people who move their workload to off peak hours if possible.

This is the opposite of an AI bubble burst.


Can you give a few penciled numbers?

You can rent a H100 GPU for $4/hour. [1]

300k tokens for that hour.

OpenAI charges $6.

Those are pessimistic assumptions.

[1] https://lambda.ai/instances


Can you keep that GPU 100% saturated at least 16 hours per day every day of the week?

If not, you aren't breaking even.


Note this is also assuming you

(1) Rent your GPUs.

(2) Pay list price, no volume breaks.

(3) Get only 85 tokens/sec. Realistically, frontier models would attain 200+ tokens/second amortized.

Inference is extremely profitable at scale.


Assuming 80GB H100 and you inference a model that is MoE and close to the size of the 80GB VRAM, you're going to see around 10k tokens/second fully batched and saturated. An example here might be Mixtral 8x7B.

You're generating about 36 million tokens/hour. Cost of Mixtral 8x7b on Open router is $0.54/M input tokens. $0.54/M output tokens.

You're looking at potentially $38.88/hour return on that H100 GPU. This is probably the best case scenario.

In reality, inference providers will use multiple GPUs together to run bigger, smarter models for a higher price.


3.99 at 8x instances, with a minimum 2 week commitment. Good luck getting 70% usage average during that time. Useful when you're running a training round and can properly gauge demand, not so great when you're offering an API.

Is it not a good penciled number? It helps set the directional tone that at inference cost is being covered.

It says the numbers are theoretically possible. Requiring a 66% usage to break even when 100% usage will piss off customers by invoking a queue means it’s a balancing act.

“Technically correct. The best kind of correct”. So inference may technically be _capable_ of being profitable, but I have question’s about them being profitable in _practice_.


Sounds like 2FA should be required for CI.

Python is genuinely a pleasant syntax and experience. [1]

It's the closest language to pseudocode that exists.

Like every other language from 1991, it has rough edges.

[1] https://xkcd.com/353/


Huh?

Wanting a better pip means I am unsafe?


Agreed.

They are equivalent of frontier models 8+ months ago.


"Author discovers that DNS stores data, and that data could be DLLs."

Okay?


Post title is wrong.

Article title is "Netflix Raising U.S. Prices for Second Time in Less Than Two Years."

(There's been no correction....maybe variety ninja-edited?)


Updated to reflect the correct title.

Isn't "Less than Two Years" just "One Year"?

No, it can be one year and two months, for example. The last price hike was in January 2025 [0].

[0] https://variety.com/2025/tv/news/netflix-price-hike-ad-plan-...


600 days is less than two years, while not being one year.

(“Fewer than two years” would mean one or zero years, but… don’t write that)


It has been more than one year since the last price increase.

So it is inaccurate to say “second time in a year” (calendar or otherwise).


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: