Popular repositories Loading
-
llama.cpp-b4139
llama.cpp-b4139 PublicA fork of llama.cpp-b4139 that can convert and run Llama-3_1-Nemotron-51B and DeciLM-7B-Instruct
C++ 2
-
llama.cpp
llama.cpp PublicForked from ggml-org/llama.cpp
LLM inference in C/C++ fork to support Nemotron 51B/49B/253B
C++
-
exllamav2
exllamav2 PublicForked from turboderp-org/exllamav2
A fast inference library for running LLMs locally on modern consumer-class GPUs modified to support Llama-3_1-Nemotron-51B
Python
-
exllamav3
exllamav3 PublicForked from turboderp-org/exllamav3
An optimized quantization and inference library for running LLMs locally on modern consumer-class GPUs to support Nemotron 51B/49B
Python
If the problem persists, check the GitHub status page or contact support.