--- base_model: - v000000/L3.1-Celestial-Stone-2x8B - v000000/2x8B_DPO_TRAIN library_name: transformers tags: - merge - llama - mixtral - dpo --- # Llama-3.1-Celestial-Stone-2x8B-DPO Direct Preference Optimization - DPO [v000000/L3.1-Celestial-Stone-2x8B](https://huggingface.co/v000000/L3.1-Celestial-Stone-2x8B) Finetuned on Nvidia A100, 0.5 Epoch of [jondurbin/gutenberg-dpo-v0.1](https://huggingface.co/datasets/jondurbin/gutenberg-dpo-v0.1)