Local CLI Copilot, powered by Ollama. 💻🦙
-
Updated
Oct 26, 2025 - Go
Local CLI Copilot, powered by Ollama. 💻🦙
[EMNLP 2025] OmniThink: Expanding Knowledge Boundaries in Machine Writing through Thinking
GPU-accelerated Llama3.java inference in pure Java using TornadoVM.
Experimental tools to backdoor large language models by re-writing their system prompts at a raw parameter level. This allows you to potentially execute offline remote code execution without running any actual code on the victim's machine or thwart LLM-based fraud/moderation systems.
A light llama-like llm inference framework based on the triton kernel.
Deploy open-source LLMs on AWS in minutes — with OpenAI-compatible APIs and a powerful CLI/SDK toolkit.
PasLLM - LLM inference engine in Object Pascal (synced from my private work repository)
Hand-derived memory-efficient super lazy PyTorch VJPs for training LLMs on laptop, all using one op (bundled scaled matmuls).
Java 23, SpringBoot 3.4.1 Examples using Deep Learning 4 Java & LangChain4J for Generative AI using ChatGPT LLM, RAG and other open source LLMs. Sentiment Analysis, Application Context based ChatBots. Custom Data Handling. LLMs - GPT 3.5 / 4o, Gemini Pro 1.5, Claude 3, Llama 3.1, Phi-3, Gemma 2, Falcon 3, Qwen 2.5, Mistral Nemo, Wizard Math
Community-built Qwen AI Provider for Vercel AI SDK - Integrate Alibaba Cloud's Qwen models with Vercel's AI application framework
Exploring Agno framework for building AI agents.
Project Zephyrine: Your experimental second hand for the world of ideas. Let's take a walk and flight with a modern, locally-run automaton, using accelerated thought to navigate the both digital aether and reality.
Silver Medal Solution for the Kaggle Competition: Eedi - Mining Misconceptions in Mathematics
Get Clothes from image
Simple RAG system powered by Milvus.
ADHARA is a prototype early warning system that identifies learning friction during normal learning activities. It supports educators by flagging behavioral indicators before academic failure occurs.
grpo to train long form QA and instructions with long-form reward model
FastLongSpeech is a novel framework designed to extend the capabilities of Large Speech-Language Models for efficient long-speech processing without necessitating dedicated long-speech training data.
Add a description, image, and links to the qwen2-5 topic page so that developers can more easily learn about it.
To associate your repository with the qwen2-5 topic, visit your repo's landing page and select "manage topics."