llm-inference
Here are 11 public repositories matching this topic...
DashInfer is a native LLM inference engine aiming to deliver industry-leading performance atop various hardware architectures, including CUDA, x86 and ARMv9.
-
Updated
May 30, 2025 - C
REST: Retrieval-Based Speculative Decoding, NAACL 2024
-
Updated
Dec 2, 2024 - C
电子鹦鹉 / Toy Language Model
-
Updated
Jun 12, 2025 - C
V-lang api wrapper for llm-inference chatllm.cpp
-
Updated
Nov 20, 2024 - C
Nim api-wrapper for llm-inference chatllm.cpp
-
Updated
Nov 20, 2024 - C
kotlin api wrapper for llm-inference chatllm.cpp
-
Updated
Nov 26, 2024 - C
Python bindings for the Transformer models implemented in C/C++ using GGML library.
-
Updated
Sep 17, 2023 - C
<img align="left" width="100" height="100" src="logo.png" alt="logo" style="float: left;"/><h3 align="left">Aquila. <br>iOS 6 untethered jailbreak for all devices<div align="right" style="float: top;"></br></h3> Aquila offers a straightforward solution for jailbreaking iOS 6 on various devices. Check out the installation guide at [ios.cfw.guid
-
Updated
Jun 21, 2025 - C
Improve this page
Add a description, image, and links to the llm-inference topic page so that developers can more easily learn about it.
Add this topic to your repo
To associate your repository with the llm-inference topic, visit your repo's landing page and select "manage topics."