Allocating extra computation at inference time has recently improved sample quality in large language models and diffusion-based image generation. In parallel, Flow Matching (FM) has gained traction in language, vision, and scientific domains, but inference-time scaling methods for it remain under-explored. Concurrently, Kim et al., 2025 approach this problem but replace the linear interpolant with a non-linear variance-preserving (VP) interpolant at inference, sacrificing FM's efficient and straight sampling. Additionally, inference-time compute scaling for flow matching has only been applied to visual tasks, like image generation. We introduce novel inference-time scaling procedures for FM that preserve the linear interpolant during sampling. Evaluations of our method on image generation, and for the first time (to the best of our knowledge), unconditional protein generation, show that I) sample quality consistently improves as inference compute increases, and II) flow matching inference-time scaling can be applied to scientific domains.
翻译:在推理阶段分配额外计算资源近期已显著提升大型语言模型和基于扩散的图像生成样本质量。与此同时,流匹配方法在语言、视觉及科学计算领域获得广泛关注,但其推理时扩展方法的研究仍显不足。Kim等人(2025)虽对此问题展开研究,但在推理阶段采用非线性方差保持插值替代线性插值,牺牲了流匹配原有的高效直线采样特性。此外,现有流匹配推理计算扩展仅应用于图像生成等视觉任务。本文提出一种保持线性插值特性的新型流匹配推理扩展方法。通过在图像生成任务及(据我们所知)首次在无条件蛋白质生成任务上的实验评估表明:I)样本质量随推理计算量增加持续提升,II)流匹配推理扩展技术可成功应用于科学计算领域。