Hacker Newsnew | past | comments | ask | show | jobs | submitlogin
Serving Open Source Models 4x faster than vLLM by quantizing with ~no tradeoffs (fireworks.ai)
3 points by georgehill on Jan 10, 2024 | hide | past | favorite


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: