This model was fine-tuned using LoRA + DPO on the dockerNLcommands and Human-Like-DPO-Dataset datasets.
dockerNLcommands
Human-Like-DPO-Dataset