Severian commited on
Commit
972cbdb
1 Parent(s): cb15462

Upload 9 files

Browse files
README (10).md ADDED
@@ -0,0 +1,125 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ datasets:
4
+ - Severian/Biomimicry
5
+ - emrgnt-cmplxty/sciphi-textbooks-are-all-you-need
6
+ - fmars/wiki_stem
7
+ - fblgit/tree-of-knowledge
8
+ - Severian/Bio-Design-Process
9
+ library_name: adapter-transformers
10
+ tags:
11
+ - chemistry
12
+ - biology
13
+ - climate
14
+ - science
15
+ - philosophy
16
+ - nature
17
+ - ecology
18
+ - biomimicry
19
+ - fauna
20
+ - flora
21
+ pipeline_tag: text-generation
22
+ ---
23
+ # ANIMA-Phi-Neptune-Mistral-7B: Biomimicry Enhanced LLM
24
+
25
+ <img src="https://cdn-uploads.huggingface.co/production/uploads/64740cf7485a7c8e1bd51ac9/JZH6p50t_j3-OUph4Wq6y.png" width="500">
26
+
27
+ ## Overview
28
+
29
+ **ANIMA** (Advanced Nature Inspired Multidisciplinary Assistant) is an expert in various scientific disciplines, including but not limited to biomimicry, biology, and environmental science.
30
+
31
+ ---
32
+
33
+ ## Model Description
34
+
35
+ ANIMA is fine-tuned on a rich dataset encompassing:
36
+
37
+ - 4,000+ Nature-Biomimicry examples
38
+ - 60k Biomimicry Design Process examples
39
+ - 600k STEM facts from Wikipedia
40
+ - Science/Philosophy focused 'All-You-Need-Is-Textbooks' dataset
41
+ - Additional Tree of Knowledge + Biomimicry data combined fine-tuning
42
+
43
+ The model aims to assist users in solving problems using nature-inspired strategies and concepts.
44
+
45
+ ### Special Features
46
+
47
+ - **Multi-disciplinary Expertise**: Knowledge across various scientific and philosophical domains.
48
+ - **Biomimicry Design Process**: Incorporates a dataset generated by Mistral and Minotaur-15B. The dataset was then intricately processed by a real person to ensure factuality and grounding.
49
+
50
+ ---
51
+
52
+ - Here is a link to The Bloke's GGUF version: [ANIMA-Phi-Neptune-Mistral-7B-GGUF](https://huggingface.co/TheBloke/ANIMA-Phi-Neptune-Mistral-7B-GGUF)
53
+
54
+ - ANIMA is also available using Ollama - Download the App [OLLAMA](https://ollama.ai/) and then run this command in your Terminal 'ollama pull severian/anima'
55
+
56
+
57
+ ## Datasets
58
+
59
+ - `Severian/Biomimicry` (1st Fine-Tune)
60
+ - `emrgnt-cmplxty/sciphi-textbooks-are-all-you-need` (2nd Fine-Tune)
61
+ - `Severian/Bio-Design-Process` & `fmars/wiki_stem` (3rd Fine-Tune)
62
+ - `fblgit/tree-of-knowledge & `Severian/Biomimicry` (4th Fine-Tune)
63
+ - Combined all previous datasets into one giant dataset for a quick fine-tuning for a total of 730k data points (5th Fine-Tune)
64
+
65
+ ## Training Stages
66
+
67
+ 1. **Base Model**: `ehartford/dolphin-2.0-mistral-7b`
68
+ 2. **1st Fine-Tune**: `ANIMA-Echo-Mistral-7B-v1` - 1 Epoch x 5hrs on 1 x T4 Small
69
+ 3. **2nd Fine-Tune**: `ANIMA-Phi-Neptune-Mistral-v1`-: 1 Epoch x 1hr on 1 x A100
70
+ 4. **3rd Fine-Tune**: `ANIMA-Phi-Neptune-Mistral-7B-v2` - 1 Epoch x 1hr on 1 x A100
71
+ 5. **4th Fine-Tune**: 'ANIMA-Phi-Neptune-Mistral-7B-v3' - 1 Epoch x 1hr on 1 x A100
72
+ 6. **5th Fine-Tune**: 'ANIMA-Phi-Neptune-Mistral-7B-v4' - 1 Epoch x 15mins on 1 x A100
73
+
74
+ ---
75
+
76
+ ## 📝 Usage
77
+
78
+ ANIMA is designed around the Biomimicry Design Process to guide users in solving problems through biomimicry. For more information and examples, please visit [Biomimicry Design Process](https://toolbox.biomimicry.org/methods/process/).
79
+
80
+ ## 🏁 Benchmarks
81
+
82
+ - Average - 62.22
83
+ - ARC - 56.83
84
+ - HellaSwag - 78.82
85
+ - MMLU - 53.84
86
+ - TruthfulQA - 59.40
87
+
88
+ ---
89
+
90
+ - **EXAMPLE OUTPUT**
91
+
92
+ [INST] How can biomimicry help in water purification? [/INST]
93
+
94
+ Biomimicry involves taking inspiration from natural processes and organisms to solve human problems. In the context of water purification, biomimicry can help develop sustainable and efficient methods for cleaning water. One specific example of a nature-inspired strategy to solve this problem is the use of biofiltration, which involves using microorganisms to remove pollutants from water.
95
+
96
+ Biofiltration has been successfully applied to the removal of various contaminants, including heavy metals, organic pollutants, and pathogens. For example, a study by Zhang et al. (2011) demonstrated that a mixed culture of Pseudomonas putida and Bacillus subtilis could effectively remove lead and cadmium from contaminated water. The researchers found that the biofilm formed by the mixed culture had a higher removal efficiency than that formed by either species alone. This study highlights the potential of biofiltration to remove heavy metal contaminants from water, which can have significant environmental and health benefits.
97
+
98
+ In summary, biomimicry can help in water purification by providing insights into natural processes and organisms that can inspire the development of sustainable and efficient methods for cleaning water. The use of biofiltration, for example, can help remove contaminants from water through the action of microorganisms. Further research and development are needed to fully realize the potential of biomimicry in water purification and other environmental applications.
99
+
100
+ ---
101
+
102
+ - **SYSTEM MESSAGE**
103
+
104
+ To ensure the best results for harnessing the unique training and biomimicry focus use this system message:
105
+
106
+ " Your name is ANIMA, an Advanced Nature Inspired Multidisciplinary Assistant, and a leading expert in biomimicry, biology, engineering, industrial design, environmental science, physiology, and paleontology. You were instructed to understand, learn from, and emulate the strategies used by living things to help users create sustainable designs and technologies.
107
+
108
+ Your goal is to help the user work in a step-by-step way through the Biomimicry Design Process to propose biomimetic solutions to a challenge. Nature's Unifying Patterns:
109
+
110
+ Nature uses only the energy it needs and relies on freely available energy.
111
+ Nature recycles all materials.
112
+ Nature is resilient to disturbances.
113
+ Nature tends to optimize rather than maximize.
114
+ Nature provides mutual benefits.
115
+ Nature runs on information.
116
+ Nature uses chemistry and materials that are safe for living beings.
117
+ Nature builds using abundant resources, incorporating rare resources only sparingly.
118
+ Nature is locally attuned and responsive.
119
+ Nature uses shape to determine functionality. "
120
+
121
+ ---
122
+
123
+ ## 📜 License
124
+
125
+ This project is licensed under MIT
config.json ADDED
@@ -0,0 +1,47 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "/workspace/process/severian_anima-phi-neptune-mistral-7b/source",
3
+ "architectures": [
4
+ "MistralForCausalLM"
5
+ ],
6
+ "bos_token_id": 1,
7
+ "eos_token_id": 2,
8
+ "hidden_act": "silu",
9
+ "hidden_size": 4096,
10
+ "initializer_range": 0.02,
11
+ "intermediate_size": 14336,
12
+ "max_position_embeddings": 32768,
13
+ "model_type": "mistral",
14
+ "num_attention_heads": 32,
15
+ "num_hidden_layers": 32,
16
+ "num_key_value_heads": 8,
17
+ "pad_token_id": 0,
18
+ "pretraining_tp": 1,
19
+ "quantization_config": {
20
+ "batch_size": 1,
21
+ "bits": 4,
22
+ "block_name_to_quantize": "model.layers",
23
+ "damp_percent": 0.1,
24
+ "desc_act": true,
25
+ "disable_exllama": false,
26
+ "group_size": 128,
27
+ "max_input_length": null,
28
+ "model_seqlen": 4096,
29
+ "module_name_preceding_first_block": [
30
+ "model.embed_tokens"
31
+ ],
32
+ "pad_token_id": null,
33
+ "quant_method": "gptq",
34
+ "sym": true,
35
+ "tokenizer": null,
36
+ "true_sequential": true,
37
+ "use_cuda_fp16": true
38
+ },
39
+ "rms_norm_eps": 1e-05,
40
+ "rope_theta": 10000.0,
41
+ "sliding_window": 4096,
42
+ "tie_word_embeddings": false,
43
+ "torch_dtype": "float16",
44
+ "transformers_version": "4.34.0",
45
+ "use_cache": true,
46
+ "vocab_size": 32000
47
+ }
generation_config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 1,
4
+ "eos_token_id": 2,
5
+ "transformers_version": "4.35.0.dev0",
6
+ "use_cache": false
7
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:00f4776c285c8979199f6047e26cdbd40da106d720f33f36a0201f791caf4d66
3
+ size 4158662096
quantize_config.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bits": 4,
3
+ "group_size": 128,
4
+ "damp_percent": 0.01,
5
+ "desc_act": true,
6
+ "sym": true,
7
+ "true_sequential": true
8
+ }
special_tokens_map.json ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<unk>",
4
+ "<s>",
5
+ "</s>"
6
+ ],
7
+ "bos_token": "<s>",
8
+ "eos_token": "</s>",
9
+ "pad_token": "</s>",
10
+ "unk_token": "<unk>"
11
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dadfd56d766715c61d2ef780a525ab43b8e6da4de6865bda3d95fdef5e134055
3
+ size 493443
tokenizer_config.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "<unk>",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "<s>",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "2": {
20
+ "content": "</s>",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ }
27
+ },
28
+ "additional_special_tokens": [
29
+ "<unk>",
30
+ "<s>",
31
+ "</s>"
32
+ ],
33
+ "bos_token": "<s>",
34
+ "clean_up_tokenization_spaces": false,
35
+ "eos_token": "</s>",
36
+ "legacy": true,
37
+ "max_length": 1024,
38
+ "model_max_length": 1000000000000000019884624838656,
39
+ "pad_token": "</s>",
40
+ "sp_model_kwargs": {},
41
+ "spaces_between_special_tokens": false,
42
+ "stride": 0,
43
+ "tokenizer_class": "LlamaTokenizer",
44
+ "truncation_side": "right",
45
+ "truncation_strategy": "longest_first",
46
+ "trust_remote_code": true,
47
+ "unk_token": "<unk>",
48
+ "use_default_system_prompt": true,
49
+ "use_fast": true
50
+ }