LHC88 commited on
Commit
2971ac5
1 Parent(s): d16f74b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -1
README.md CHANGED
@@ -13,12 +13,14 @@ datasets:
13
  ![image/png](https://huggingface.co/openaccess-ai-collective/DPOpenHermes-7B/resolve/main/assets/dpopenhermes.png)
14
 
15
  ## Laser Config
 
 
16
  top_k_layers: 32
17
  datasets: wikitext2, ptb, c4
18
 
19
  ## OpenHermes x Notus x Neural
20
 
21
- [<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
22
 
23
  This is a second RL fine tuned model of [Teknium](https://huggingface.co/teknium)'s [OpenHermes-2.5-Mistral-7B](https://huggingface.co/teknium/OpenHermes-2.5-Mistral-7B) using the [Intel/orca_dpo_pairs](https://huggingface.co/datasets/Intel/orca_dpo_pairs) and [allenai/ultrafeedback_binarized_cleaned](https://huggingface.co/datasets/allenai/ultrafeedback_binarized_cleaned) preference datasets for reinforcement learning using Direct Preference Optimization (DPO)
24
 
 
13
  ![image/png](https://huggingface.co/openaccess-ai-collective/DPOpenHermes-7B/resolve/main/assets/dpopenhermes.png)
14
 
15
  ## Laser Config
16
+
17
+ [**Lasered with AIDOcks**](https://github.com/l4b4r4b4b4/AIDocks)
18
  top_k_layers: 32
19
  datasets: wikitext2, ptb, c4
20
 
21
  ## OpenHermes x Notus x Neural
22
 
23
+ [<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
24
 
25
  This is a second RL fine tuned model of [Teknium](https://huggingface.co/teknium)'s [OpenHermes-2.5-Mistral-7B](https://huggingface.co/teknium/OpenHermes-2.5-Mistral-7B) using the [Intel/orca_dpo_pairs](https://huggingface.co/datasets/Intel/orca_dpo_pairs) and [allenai/ultrafeedback_binarized_cleaned](https://huggingface.co/datasets/allenai/ultrafeedback_binarized_cleaned) preference datasets for reinforcement learning using Direct Preference Optimization (DPO)
26