Not a Senior, just a lifelong curious researcher
Popular repositories Loading
-
-
-
turboquant
turboquant PublicForked from 0xSero/turboquant
TurboQuant: Near-optimal KV cache quantization for LLM inference (3-bit keys, 2-bit values) with Triton kernels + vLLM integration
Python
-
langchain-rust
langchain-rust PublicForked from Abraxas-365/langchain-rust
🦜️🔗LangChain for Rust, the easiest way to write LLM-based programs in Rust
Rust
-
Something went wrong, please refresh the page to try again.
If the problem persists, check the GitHub status page or contact support.
If the problem persists, check the GitHub status page or contact support.
