>>65627850>How big was the DPO dataset?They used 851k from pick-a-pic
https://huggingface.co/papers/2311.12908But I think a properly tagged booru imageset (as in NLP + not the human-assigned tags. EF-style but without the furry and brownness) finetune on sdxl would turn up a better result than DPO