Improve model card for PaCo-RL

#1
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +67 -0
README.md ADDED
@@ -0,0 +1,67 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ pipeline_tag: image-text-to-text
3
+ library_name: transformers
4
+ ---
5
+
6
+ # PaCo-RL: Advancing Reinforcement Learning for Consistent Image Generation with Pairwise Reward Modeling
7
+
8
+ **PaCo-RL** is a comprehensive framework designed for consistent image generation using reinforcement learning. It tackles the challenges of preserving identities, styles, and logical coherence across multiple images, which is crucial for applications such as storytelling and character design.
9
+
10
+ This model is presented in the paper [PaCo-RL: Advancing Reinforcement Learning for Consistent Image Generation with Pairwise Reward Modeling](https://huggingface.co/papers/2512.04784).
11
+
12
+ - **Project Page**: https://x-gengroup.github.io/HomePage_PaCo-RL/
13
+ - **Code Repository**: https://github.com/X-GenGroup/PaCo-RL
14
+ - **Data & Models Collection**: https://huggingface.co/collections/X-GenGroup/paco-rl
15
+
16
+ ## Overview
17
+
18
+ PaCo-RL argues that reinforcement learning offers a promising alternative for learning complex and subjective visual criteria in a data-free manner. The framework combines a specialized consistency reward model with an efficient RL algorithm.
19
+
20
+ ### Key Components
21
+
22
+ - **PaCo-Reward**: A pairwise consistency evaluator trained on a large-scale dataset constructed via automated sub-figure pairing. It evaluates consistency through a generative, autoregressive scoring mechanism enhanced by task-aware instructions and Chain-of-Thought (CoT) reasons.
23
+ - **PaCo-GRPO**: An efficient RL algorithm leveraging a novel resolution-decoupled optimization strategy to substantially reduce RL cost, alongside a log-tamed multi-reward aggregation mechanism that ensures balanced and stable reward optimization.
24
+
25
+ Extensive experiments show that PaCo-Reward significantly improves alignment with human perceptions of visual consistency, and PaCo-GRPO achieves state-of-the-art consistency performance with improved training efficiency and stability.
26
+
27
+ <div align="center">
28
+ <img src="https://github.com/X-GenGroup/PaCo-RL/raw/main/assets/readme_overview.png" alt="PaCo-RL Overview" width="800"/>
29
+ </div>
30
+
31
+ ## Quick Start
32
+
33
+ For detailed installation, training of the reward model (PaCo-Reward), and running the full RL training (PaCo-GRPO), please refer to the [official GitHub repository](https://github.com/X-GenGroup/PaCo-RL). The repository provides comprehensive documentation for each component.
34
+
35
+ ## Model Zoo
36
+
37
+ The PaCo-RL framework includes several models available on Hugging Face:
38
+
39
+ | Model | Type | HuggingFace |
40
+ |-------|------|-------------|
41
+ | **PaCo-Reward-7B** | Reward Model | [πŸ€— Link](https://huggingface.co/X-GenGroup/PaCo-Reward-7B) |
42
+ | **PaCo-Reward-7B-Lora** | Reward Model (LoRA) | [πŸ€— Link](https://huggingface.co/X-GenGroup/PaCo-Reward-7B-Lora) |
43
+ | **PaCo-FLUX.1-dev** | T2I Model (LoRA) | [πŸ€— Link](https://huggingface.co/X-GenGroup/PaCo-FLUX.1-dev-Lora) |
44
+ | **PaCo-FLUX.1-Kontext-dev** | Image Editing Model (LoRA) | [πŸ€— Link](https://huggingface.co/X-GenGroup/PaCo-FLUX.1-Kontext-Lora) |
45
+ | **PaCo-QwenImage-Edit** | Image Editing Model (LoRA) | [πŸ€— Link](https://huggingface.co/X-GenGroup/PaCo-Qwen-Image-Edit-Lora) |
46
+
47
+ ## Acknowledgement
48
+
49
+ Our work is built upon [Flow-GRPO](https://github.com/yifan123/flow_grpo), [LLaMA-Factory](https://github.com/hiyouga/LLaMA-Factory), [vLLM](https://github.com/vllm-project/vllm), and [Qwen2.5-VL](https://github.com/QwenLM/Qwen3-VL). We sincerely thank the authors for their valuable contributions to the community.
50
+
51
+ ## Citation
52
+
53
+ ```bibtex
54
+ @misc{ping2025pacorladvancingreinforcementlearning,
55
+ title={PaCo-RL: Advancing Reinforcement Learning for Consistent Image Generation with Pairwise Reward Modeling},
56
+ author={Bowen Ping and Chengyou Jia and Minnan Luo and Changliang Xia and Xin Shen and Zhuohang Dang and Hangwei Qian},
57
+ year={2025},
58
+ eprint={2512.04784},
59
+ archivePrefix={arXiv},
60
+ primaryClass={cs.CV},
61
+ url={https://arxiv.org/abs/2512.04784},
62
+ }
63
+ ```
64
+
65
+ <div align="center">
66
+ <sub>⭐ Star us on GitHub if you find PaCo-RL helpful!</sub>
67
+ </div>