Brief-details: QwQ-R1984-32B quantized to 8-bit GGUF format, optimized for llama.cpp deployment. 32B parameter model with efficient local inference capabilities.
Brief Details: Advanced 32B parameter reasoning model with enhanced uncensored capabilities and web search integration. Built on Qwen series with 8K context window.
Brief Details: A quantized 32B parameter LLM optimized for llama.cpp, converted from Qwen/QwQ-32B to GGUF format for efficient local deployment
Brief-details: Qwen2.5-VL-32B-Instruct converted to GGUF format for efficient local deployment via llama.cpp, optimized for visual-language tasks with Q8 quantization
BRIEF-DETAILS: A GGUF-formatted 32B parameter language model converted from Qwen/QwQ-32B, optimized for local deployment using llama.cpp with Q8_0 quantization.
Brief-details: A 49B parameter Llama-3 variant converted to GGUF format, optimized for efficient local deployment using llama.cpp, featuring Q4_K_M quantization
Brief Details: A quantized GGUF version of Mistral's 24B instruction model, optimized for local deployment via llama.cpp with Q4_K_M compression
Brief-details: Quantized 24B parameter Mistral instruction model optimized for llama.cpp, converted to GGUF format for efficient local deployment and inference. Suitable for both CLI and server applications.
BRIEF-DETAILS: Quantized 24B parameter Mistral instruction model converted to GGUF format, optimized for local deployment via llama.cpp with Q8 precision
BRIEF-DETAILS: 27B parameter Gemma model quantized to 6-bit, optimized for llama.cpp deployment with GGUF format conversion from VIDraft/Gemma-3-R1984-27B
Brief-details: A Q8-quantized 27B parameter Gemma model optimized for llama.cpp, converted from VIDraft/Gemma-3-R1984-27B to GGUF format for efficient local deployment
Brief-details: Qwen2.5-VL-32B-Instruct GGUF conversion - 32B parameter multimodal LLM optimized for Q4 quantization, supporting visual and language tasks via llama.cpp
Brief-details: AccVideo is a high-performance video diffusion model that achieves 8.5x faster inference than HunyuanVideo, utilizing synthetic datasets for efficient distillation.
Brief-details: A 32B parameter quantized GGUF model converted from VIDraft/QwQ-R1984-32B, optimized for llama.cpp deployment with Q4_K_M quantization format
Brief-details: Quantized 12B parameter Gemma model converted to GGUF format for efficient local inference using llama.cpp, optimized for Q6_K precision
Brief-details: A GGUF-formatted 12B parameter Gemma model optimized for llama.cpp, featuring 8-bit quantization and enhanced inference capabilities for efficient deployment
Brief-details: Gemma3-R1984-12B is a 12B parameter multimodal AI platform built on Google's Gemma architecture, featuring web search integration, 8K token context, and secure local deployment capabilities.
BRIEF DETAILS: A LoRA model for generating stylized images of Morgenstern with different personas, featuring high-resolution 26K output and customizable hair colors.
Brief-details: A LoRA model for generating varied Boris Yeltsin images with different hair colors and roles, including Japanese president and police officer positions.
Brief-details: A specialized LoRA model for generating black string sandal images, built on FLUX.1-dev base model. Requires "SS" trigger word for optimal results.
BRIEF-DETAILS: A LoRA model specialized in generating images of Asian females wearing specific necklace designs (NL1), optimized for outdoor photography scenarios with natural lighting