MFANN-Llama3.1-Abliterated-Slerp-TIES

Maintained By
netcat420

MFANN-Llama3.1-Abliterated-Slerp-TIES

PropertyValue
Parameter Count8.03B
Model TypeText Generation
ArchitectureLLaMA-based Transformer
Tensor TypeFP16

What is MFANN-Llama3.1-Abliterated-Slerp-TIES?

MFANN-Llama3.1-Abliterated-Slerp-TIES is an advanced language model built on the LLaMA 3.1 architecture, created by netcat420. This model represents a sophisticated merge of multiple base models using MergeKit technology, optimized for both general text generation and specific instruction-following tasks.

Implementation Details

The model implements a context length of 8192 tokens and utilizes specialized sampling parameters including temperature (1.0), top-p (1.0), and top-k (50) settings. It features two distinct operation modes: a standard helpful assistant mode and an experimental "SATANN" mode for cybersecurity-related tasks.

  • FP16 precision for efficient computation
  • 8192 token context window
  • Specialized repeat penalty system (1.19 over 69 tokens)
  • GPU layer optimization for Vulkan offloading

Core Capabilities

  • Strong performance on IFEval (42.93% accuracy)
  • Competitive BBH performance (27.6% normalized accuracy)
  • MMLU-PRO capabilities (28.13% accuracy)
  • Mathematical reasoning (5.97% on MATH Lvl 5)
  • Specialized instruction following and conversational abilities

Frequently Asked Questions

Q: What makes this model unique?

This model stands out for its dual-mode operation capability and its strong performance on instruction-following tasks, particularly achieving 42.93% on IFEval. It uses an innovative merge of multiple LLaMA-based models with SLERP technology.

Q: What are the recommended use cases?

The model is well-suited for text generation tasks, conversation, and instruction following. It offers both a standard helpful assistant mode for general use and a specialized mode for cybersecurity-related applications.

The first platform built for prompt engineering