Triangle104 commited on
Commit
463bcde
·
verified ·
1 Parent(s): 3193031

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +32 -0
README.md CHANGED
@@ -16,6 +16,38 @@ base_model: DavidAU/L3-MOE-4X8B-Grand-Horror-25B
16
  This model was converted to GGUF format from [`DavidAU/L3-MOE-4X8B-Grand-Horror-25B`](https://huggingface.co/DavidAU/L3-MOE-4X8B-Grand-Horror-25B) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
17
  Refer to the [original model card](https://huggingface.co/DavidAU/L3-MOE-4X8B-Grand-Horror-25B) for more details on the model.
18
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
19
  ## Use with llama.cpp
20
  Install llama.cpp through brew (works on Mac and Linux)
21
 
 
16
  This model was converted to GGUF format from [`DavidAU/L3-MOE-4X8B-Grand-Horror-25B`](https://huggingface.co/DavidAU/L3-MOE-4X8B-Grand-Horror-25B) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
17
  Refer to the [original model card](https://huggingface.co/DavidAU/L3-MOE-4X8B-Grand-Horror-25B) for more details on the model.
18
 
19
+ ---
20
+ It is a LLama3 model, max context of 8192 (or 32k+ with rope) using mixture of experts to combine Dark/Horror models models of 8B each into one massive powerhouse at 25B parameters (equal to 32B - 4 X 8 B).
21
+
22
+ This model's instruction following, and output generation for creative writing, prose, fiction and role play are exceptional.
23
+
24
+ It excels at description, dialog, imagery, metaphors, and prose - and shows great variations in sentence / paragraph size, length, and composition.
25
+
26
+ It is also not afraid, and will not pull its punches.
27
+
28
+ And it has a sense of humor too.
29
+
30
+ It can do horror just as easily as it can do romance.
31
+
32
+ Most notably dialog is very "un-ai" like, combined with prose (short, and terse at times).
33
+
34
+ (lots of different examples below, including 2, 3 and 4 experts and different genres)
35
+
36
+ And it is fast: 34 t/s (2 experts) on a low end 16GB card, Q3KS.
37
+
38
+ Double this speed for standard/mid-range video cards.
39
+
40
+ Model can be used also for all genres (examples below showing this).
41
+
42
+ This model has been designed to be relatively bullet proof and operates with all parameters, including temp settings from 0 to 5.
43
+
44
+ It is an extraordinary compressed model, with a very low perplexity level (lower than Meta Llama3 Instruct).
45
+
46
+ It is for any writing, fiction or roleplay activity.
47
+
48
+ It requires Llama3 template and/or "Command-R" template.
49
+
50
+ ---
51
  ## Use with llama.cpp
52
  Install llama.cpp through brew (works on Mac and Linux)
53