Stars
- All languages
- AGS Script
- ASP.NET
- ActionScript
- Arc
- Assembly
- Batchfile
- C
- C#
- C++
- CSS
- Clojure
- CodeQL
- CoffeeScript
- Common Lisp
- Crystal
- Cuda
- D
- Dart
- Dhall
- Dockerfile
- Dylan
- Elixir
- Emacs Lisp
- Erlang
- Factor
- Frege
- GDScript
- Go
- Groovy
- HTML
- Hack
- Haskell
- Haxe
- Idris
- JSON
- Java
- JavaScript
- Julia
- Jupyter Notebook
- Just
- Kotlin
- Lean
- Less
- LiveScript
- Lua
- MATLAB
- MDX
- Makefile
- Markdown
- Mermaid
- Mojo
- MoonScript
- Mustache
- NSIS
- Nix
- OCaml
- Objective-C
- Objective-C++
- Objective-J
- PHP
- PLpgSQL
- Perl
- Prolog
- Protocol Buffer
- Python
- QML
- R
- ReScript
- Ren'Py
- Roff
- Ruby
- Rust
- SCSS
- Scala
- Shell
- Smalltalk
- Solidity
- Starlark
- Svelte
- Swift
- TSQL
- TeX
- TypeScript
- Vala
- Vim Script
- Visual Basic
- Vue
- WebAssembly
- XSLT
- Zig
- wisp
Review-first terminal diff viewer for agentic coders
2.24x decode TPS increase On Qwen 3.6 27B @ temp 0.6 | Native MTP Speculative Decoding On Apple Silicon With No External Drafter.
DeepSeek 4 Flash local inference engine for Metal and CUDA
π Entire CLI hooks into your Git workflow to capture AI agent sessions as you work. Sessions are indexed alongside commits, creating a searchable record of how code was written in your repo.
CLI to control iOS and Android devices for AI agents
Flash-MoE sidecar slot-bank runtime for large GGUF MoE models on Apple Silicon β llama.cpp fork
ποΈ Give your apps, CLIs, and agents a voice. VoiPi is a universal, zero-dependency, free text-to-speech library for JavaScript.
ArtifactFS is a filesystem driver designed to mount large git repos as quickly as possible, hydrating file contents on-the-fly instead of blocking on the initial clone. It's ideal for agents, sandbβ¦
Agent Skill to help convert transformer LLMs to mlx-lm
KV cache compression via block-diagonal rotation. Beats TurboQuant: better PPL (6.91 vs 7.07), 28% faster decode, 5.3x faster prefill, 44x fewer params. Drop-in llama.cpp integration.
LLM inference server with continuous batching & SSD caching for Apple Silicon β managed from the macOS menu bar
Official implementation of Paper "System-Aware 4-Bit KV-Cache Quantization for Real-World LLM Serving"
Lucebox optimization hub: hand-tuned LLM inference, built for specific consumer hardware.
Lossless DFlash speculative decoding for MLX on Apple Silicon
RTX 6000 Pro Wiki β Running Large LLMs (Qwen3.5-397B, Kimi-K2.5, GLM-5) on PCIe GPUs without NVLink
Static checker for GitHub Actions workflow files
Create stunning demos for free. Open-source, no subscriptions, no watermarks, and free for commercial use. An alternative to Screen Studio.
Capability-based sandboxes with fine-grained policies The next-generation isolation primitive β brokering access directly within the agent's operating context, with zero setup and zero latency
Replace port numbers with stable, named local URLs. For humans and agents.
Dark mode PDFs without destroying your images.





