Skip to content
View g023's full-sized avatar

Block or report g023

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Maximum 250 characters. Please don’t include any personal information such as legal names or email addresses. Markdown is supported. This note will only be visible to you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse

Pinned Loading

  1. aroc aroc Public

    Agentic Read-Only Chat - A rich terminal chat interface powered by locally installed llama.cpp and g023/g023-Qwen3.5-9B-GGUF:IQ2_M

    Python

  2. harnessharvest harnessharvest Public

    A self-learning, self-correcting, LLM-powered harness creation and management system with FAISS-powered RAG, sandboxed execution, and autonomous improvement modes. Powered by Ollama and offline mod…

    Python 2 1

  3. localmodelrouter localmodelrouter Public

    Local LLM server that provides drop-in API compatibility with both Ollama and OpenAI, using your locally installed [llama.cpp](https://github.com/ggerganov/llama.cpp) 's `llama-server` as the infer…

    Python 2

  4. xinf xinf Public

    g023's TurboXInf 🚀: 2x+ faster inference for Qwen3-1.77B or Qwen3.5-2B on RTX 3060! Custom Triton INT8 GEMV kernels halve memory traffic by fusing dequantization, paired with torch.compile. Hits 11…

    Python 1

  5. turboquant turboquant Public

    Standalone TurboQuant KV Cache Inference for https://huggingface.co/g023/Qwen3-1.77B-g023

    Python 4

  6. g023-OllamaMan g023-OllamaMan Public

    A Concept Ollama Server Management OS that runs in a web browser.

    PHP 7