CS492(C): Diffusion and Flow Models
Minhyuk Sung, KAIST, Fall 2025
Image Generation Challenge
Mid-Term Evaluation Submission Due: November 2 (Sunday), 23:59 KST
Final Submission Due: November 15 (Saturday), 23:59 KST
Where to submit: KLMS
What to Do
In this challenge, your task is to train an image diffusion/flow model beyond the previous 2D toy experiment setups from the assignments. After training, you are encouraged to explore and apply any techniques you find effective for achieving high-quality generation with only a few sampling steps.
Specifically, your tasks are as follows:
-
Implementing a Diffusion/Flow Wrapper
- We provide the architecture backbone (which should remain fixed), but everything around it is up to you.
- Your goal is to design and implement your own diffusion/flow model wrapper, including noise schedulers as well as the forward and reverse processes.
-
Improving Few-Step Generation
- Once your diffusion/flow wrapper is ready, the main challenge is to investigate and improve the generation quality with very few sampling steps.
- Your models will be evaluated with three Number of Function Evaluations (NFEs): 1, 2, and 4.
- You are encouraged to experiment with the latest techniques, such as Consistency Models, ReFlow, or any other advanced methods you find effective.
- Check out the Recommended Readings section, but you are not limited to implementing one of the algorithms introduced in those papers; they are provided only as references.
Important Notes
PLEASE READ THE FOLLOWING CAREFULLY! Any violation of the rules or failure to properly cite existing code, models, or papers used in the project in your write-up will result in a zero score.
- DO NOT use any pretrained diffusion model. You must train the model from scratch.
- DO NOT modify the provided U-Net architecture code.
- DO NOT modify the provided sampling code and evaluation script. These will be distributed to ensure consistent evaluation across all submissions.
- You are allowed to use open-source implementations, as long as they are clearly mentioned and cited in your write-up.
Dataset and Base Code
You are required to use the Simpsons Face image dataset for training and evaluation.

You do not need to download the dataset yourself. We will provide a script in the base code that automatically downloads the dataset.
Evaluation
-
This is a team-based competition. The performance of your image generative models will be evaluated quantitatively using Fréchet Inception Distance (FID) scores at Numbers of Function Evaluations (NFEs) = 1, 2, and 4.
-
The TAs will provide FID scores computed using their own implementation of diffusion and flow models as reference values (the code will not be released). You are expected to match or surpass these reference FIDs.
-
Additionally, to help everyone gauge progress, there will be a Mid-Term Evaluation where teams can submit intermediate results. Participation is optional, but the top team at each NFE in the mid-term evaluation that also outperforms the TAs’ FID scores will receive bonus credit toward the final grade. All submitted results will be shared anonymously with the class so that teams can see how others are performing.
-
Final grading will be determined relative to the best FID score achieved for each NFE. Specifically, the score for each NFE is calculated as follows: $$ \text{Score} = \max!\left( \frac{\text{TA's FID} - \text{Your FID}} {\text{TA's FID} - \text{Lowest FID}} \times 4 + 5,\; 0 \right) $$
- When Your FID=Lowest FID, you get 9 points for that NFE.
- When Your FID=TA′s FID, you get 5 points for that NFE.
-
Bonus credits per NFE:
- Mid-term Evaluation Bonus: Every team that outperforms the TA’s FID at the mid-term evaluation receives +0.5 point (for that NFE).
- Winner Bonus: If your team achieves the lowest FID for an NFE, you receive +0.5 point (for that NFE).
-
In total, the image generation challenge is worth a maximum of 30 points.
Mid-Term Evaluation Submission (Optional)
The purpose of the mid-term evaluation is to give all students a reference point for how other teams are progressing. Participation is optional, but the top team at each NFE in the mid-term evaluation that also outperforms the TAs’ FID scores will receive bonus credit toward the final grade.
- What to submit
- Self-contained source code
- Your submission must include the complete codebase necessary to run end-to-end from the TAs' side.
- TAs will run your code in their environment without additional modifications.
- For consistent evaluation, the files
sampling.pyandmeasure_fid.pywill be replaced with the official versions.
- A model checkpoint and config json file
- Self-contained source code
- Grading Procedure
- TAs will run your submitted code in their Python environment.
- The FID scores measured by TAs will be published on the leaderboard.
- Submissions that fail to run in the TA environment will be marked as failed on the leaderboard.
- Among the submissions exceeding the TAs' result, the top-k will earn bonus credit.
Final Submission
- What to submit:
- Self-contained source code
- A model checkpoint and config json file
- 2-page write-up.
- No template provided.
- Maximum of two A4 pages, excluding references.
- All of the following must be included:
- Technical details: One-paragraph description of the technical details of your few-step generation implementation.
- Training details: Training logs (e.g., training loss curves), and total training time.
- Qualitative evidence: ~8 sample images from the early training phases.
- Citations: All external code, and papers used must be properly cited.
- Missing any of these items will result in a penalty.
- If the write-up exceeds two pages, any content beyond the second page will be ignored, which may lead to missing required items.
Grading
There is no late day. Submit on time.
Late submission: Zero score.
Missing any required item in the final submission (samples, code/model, write-up): Zero score.
Missing items in the write-up: 10% penalty for each.
Recommended Readings
[1] Song et al., Consistency Models, ICML 2023.
[2] Kim et al., Consistency Trajectory Models: Learning Probability Flow ODE Trajectory of Diffusion, ICLR 2024.
[3] Liu et al., Flow Straight and Fast: Learning to Generate and Transfer Data with Rectified Flow, ICLR 2023.
[4] Frans et al., One Step Diffusion via Shortcut Models, ICLR 2025.
[5] Tong et al., Learning to Discretize Denoising Diffusion ODEs, ICLR 2025.