Text Generation
Safetensors
Danish
English
llama
peter-sk commited on
Commit
2eb2324
·
1 Parent(s): 6ecf91f
Files changed (1) hide show
  1. README.md +14 -1
README.md CHANGED
@@ -16,8 +16,13 @@ Munin-7B-open-pt is a 7 billion parameter language model continually pre-trained
16
 
17
  Munin-7B-open-pt is a base model that can be used as the starting point for fine-tuning and post-training. It has not been instruction-tuned and cannot directly be expected to function as a chat model.
18
 
 
 
 
 
19
  ## Training details
20
- Munin-7B-open-pt has been trained using the [maester](https://github.com/rlrs/maester/tree/main/3aca26960eaa1a16250b3feda40303c240ba4ca1) framework developed as part of the [Danish Foundation Models project](https://foundationmodels.dk/). All training was performed on a single 8x NVIDIA B200 node (the first of its kind in Denmark).
 
21
 
22
  The training was performed in three stages, with data mix (open-stageK.py) and maester (open-stageK.toml) configuration files available in each subfolder. The datasets can be created using the create_dataset.py script provided in this repository.
23
 
@@ -35,3 +40,11 @@ Munin-7B-Open-pt was trained only on Danish and English-language data and code f
35
  It will likely have poor performance on other languages or programming languages.
36
 
37
  As a base model, Munin-7B-Open-pt has not been aligned for safety and may, for example, reflect social biases present in its training data or potentially provide toxic or harmful information.
 
 
 
 
 
 
 
 
 
16
 
17
  Munin-7B-open-pt is a base model that can be used as the starting point for fine-tuning and post-training. It has not been instruction-tuned and cannot directly be expected to function as a chat model.
18
 
19
+ ## License
20
+
21
+ The model is made available under [Apache 2.0](https://www.apache.org/licenses/LICENSE-2.0) open source license. It may therefore be used, modified, distributed, and sublicensed for any purpose, including commercial use, without the licensee having to release their own derivative works under the same permissive terms, provided that users retain copyright and license notices and document any modifications they make.
22
+
23
  ## Training details
24
+
25
+ Munin-7B-open-pt has been trained using the [maester](https://github.com/rlrs/maester/tree/main/3aca26960eaa1a16250b3feda40303c240ba4ca1) framework developed as part of the [Danish Foundation Models project](https://foundationmodels.dk/). All training was performed on a single 8x NVIDIA B200 node (the first of its kind in Denmark) as part of the [SDU UCloud](https://cloud.sdu.dk/) research cloud.
26
 
27
  The training was performed in three stages, with data mix (open-stageK.py) and maester (open-stageK.toml) configuration files available in each subfolder. The datasets can be created using the create_dataset.py script provided in this repository.
28
 
 
40
  It will likely have poor performance on other languages or programming languages.
41
 
42
  As a base model, Munin-7B-Open-pt has not been aligned for safety and may, for example, reflect social biases present in its training data or potentially provide toxic or harmful information.
43
+
44
+ ## Project partners & funding
45
+ The development of Munin-7B-Open-pt was performed in a close collaboration between [Aarhus University](https://chc.au.dk/), the [Alexandra Institute](https://alexandra.dk/), and the [University of Southern Denmark](https://www.sdu.dk/en/forskning/machine-learning) as part of the [Danish Foundation Models project](https://foundationmodels.dk/).
46
+
47
+ Funding was provided by the [Danish Ministry of Digital Affairs](https://www.english.digmin.dk/) and the [Danish Ministry of Higher Education and Science](https://ufm.dk/en).
48
+
49
+ ## How to cite
50
+ Coming soon.