Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Latency means this still makes no sense to me. Perhaps some batch background processing job such as research or something but that's stretching.




I think the most providers all give high latency batch APIs significant discounts. A lot of AI workloads feel batch-oriented to me, or could be once they move beyond the prototype and testing phases. Chat will end up being a small fraction of load in the long term.

That would imply there's still capacity here on earth for this type of traffic.



Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: