MS-Inky-2409-22B

MS-Inky-2409-22B is a merge of the following models using LazyMergekit:

Keep in mind that, this merged model isn't usually tested at the moment, which could benefit in vocabulary error.

🧩 Configuration

models:
 - model: unsloth/Mistral-Small-Instruct-2409+estrogen/MS-22b-RP-Ink-adpt

merge_method: task_arithmetic
base_model: unsloth/Mistral-Small-Instruct-2409+estrogen/MS-22b-RP-Ink-adpt
parameters:
  normalize: false

dtype: bfloat16
tokenizer_source: union

πŸ’» Usage

!pip install -qU transformers accelerate

from transformers import AutoTokenizer
import transformers
import torch

model = "Kaoeiri/MS-Inky-2409-22B"
messages = [{"role": "user", "content": "What is a large language model?"}]

tokenizer = AutoTokenizer.from_pretrained(model)
prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
pipeline = transformers.pipeline(
    "text-generation",
    model=model,
    torch_dtype=torch.float16,
    device_map="auto",
)

outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
print(outputs[0]["generated_text"])
Downloads last month
14
Safetensors
Model size
22.2B params
Tensor type
BF16
Β·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no library tag.

Model tree for Kaoeiri/MS-Inky-2409-22B

Merges
11 models
Quantizations
3 models