MistralThinker-AntiRep-24B-v1

MistralThinker-AntiRep-24B-v1Audit

This is a merge of pre-trained language models created using mergekit.

Merge Details

Merge Method

This model was merged using the DELLA merge method using B:\24B!models--anthracite-core--Mistral-Small-3.2-24B-Instruct-2506-Text-Only as a base.

Models Merged

The following models were included in the merge:

  • B:\24B!models--Undi95--MistralThinker-v1.1
  • B:\24B!models--ConicCat--Mistral-Small-3.2-AntiRep-24B

Configuration

The following YAML configuration was used to produce this model:

architecture: MistralForCausalLM
models:
  - model: B:\24B\!models--anthracite-core--Mistral-Small-3.2-24B-Instruct-2506-Text-Only
  # - model: B:\24B\!models--mistralai--Mistral-Small-24B-Instruct-2501
  - model: B:\24B\!models--ConicCat--Mistral-Small-3.2-AntiRep-24B
    parameters:
      density: 0.8
      weight: 0.99
      epsilon: 0.2
  - model: B:\24B\!models--Undi95--MistralThinker-v1.1
    parameters:
      density: 0.8
      weight: 0.01
      epsilon: 0.2
# Seed: 420 
merge_method: della
base_model: B:\24B\!models--anthracite-core--Mistral-Small-3.2-24B-Instruct-2506-Text-Only
# base_model: B:\24B\!models--mistralai--Mistral-Small-24B-Instruct-2501
parameters:
  lambda: 1.0
  normalize: false
  int8_mask: false
dtype: float32
out_dtype: bfloat16
tokenizer:
  source: union
chat_template: auto
name: 🧠 MistralThinker-AntiRep-24B-v1
Downloads last month
2
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for Naphula-Archives/MistralThinker-AntiRep-24B-v1

Paper for Naphula-Archives/MistralThinker-AntiRep-24B-v1