ToastyPigeon commited on
Commit
902c100
·
verified ·
1 Parent(s): ab95e15

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -36
README.md CHANGED
@@ -11,43 +11,12 @@ tags:
11
  - merge
12
 
13
  ---
14
- # ms-idk-v13
15
 
16
- This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
17
 
18
- ## Merge Details
19
- ### Merge Method
20
 
21
- This model was merged using the [task arithmetic](https://arxiv.org/abs/2212.04089) merge method using [unsloth/Mistral-Small-Instruct-2409](https://huggingface.co/unsloth/Mistral-Small-Instruct-2409) as a base.
22
 
23
- ### Models Merged
24
-
25
- The following models were included in the merge:
26
- * [unsloth/Mistral-Small-Instruct-2409](https://huggingface.co/unsloth/Mistral-Small-Instruct-2409) + [Alfitaria/mistral-small-fujin-qlora](https://huggingface.co/Alfitaria/mistral-small-fujin-qlora)
27
- * [unsloth/Mistral-Small-Instruct-2409](https://huggingface.co/unsloth/Mistral-Small-Instruct-2409) + [ToastyPigeon/mistral-small-springdragon-qlora](https://huggingface.co/ToastyPigeon/mistral-small-springdragon-qlora)
28
- * output/tempered-rp
29
-
30
- ### Configuration
31
-
32
- The following YAML configuration was used to produce this model:
33
-
34
- ```yaml
35
- base_model: unsloth/Mistral-Small-Instruct-2409
36
- merge_method: task_arithmetic
37
- slices:
38
- - sources:
39
- - layer_range: [0, 56]
40
- model: output/tempered-rp
41
- parameters:
42
- weight: 0.4
43
- - layer_range: [0, 56]
44
- model: unsloth/Mistral-Small-Instruct-2409+Alfitaria/mistral-small-fujin-qlora
45
- parameters:
46
- weight: 0.5
47
- - layer_range: [0, 56]
48
- model: unsloth/Mistral-Small-Instruct-2409+ToastyPigeon/mistral-small-springdragon-qlora
49
- parameters:
50
- weight: 0.1
51
- - layer_range: [0, 56]
52
- model: unsloth/Mistral-Small-Instruct-2409
53
- ```
 
11
  - merge
12
 
13
  ---
14
+ # Meadowlark - Alternate Version
15
 
16
+ This is an alternate version of [Meadowlark](https://huggingface.co/allura-org/MS-Meadowlark-22B). It's distinct/fun enough that I wanted to post it, but not really distinct enough from Meadowlark to be its own thing.
17
 
18
+ The recipe is very similar to that of Meadowlark, but I subbed out the train on Creative_Writing_Multiturn and replaced it with [Sunfall v0.7.0](https://huggingface.co/crestf411/MS-sunfall-v0.7.0) for a less-synthetic-feeling RP segment.
 
19
 
20
+ Sunfall does include some synthetic data, including some generated from Claude, so it's not exactly what I was looking for but it feels like it helps over what was there before.
21
 
22
+ For usage information see the model card for [Meadowlark](https://huggingface.co/allura-org/MS-Meadowlark-22B), same stuff here.