The generate-filter-refine (iterative) paradigm based on large language models (LLMs) has achieved progress in reasoning, programming, and program discovery in AI+Science. However, the effectiveness of search depends on where to search, namely, how to encode the domain prior into an operationally structured hypothesis space. To this end, this paper proposes a compact formal theory that describes and measures LLM-assisted iterative search guided by domain priors. We represent an agent as a fuzzy relation operator on inputs and outputs to capture feasible transitions; the agent is thereby constrained by a fixed safety envelope. To describe multi-step reasoning/search, we weight all reachable paths by a single continuation parameter and sum them to obtain a coverage generating function; this induces a measure of reachability difficulty; and it provides a geometric interpretation of search on the graph induced by the safety envelope. We further provide the simplest testable inferences and validate them via a majority-vote instantiation. This theory offers a workable language and operational tools to measure agents and their search spaces, proposing a systematic formal description of iterative search constructed by LLMs.
翻译:基于大语言模型(LLM)的生成-过滤-精炼(迭代)范式在AI+科学领域的推理、编程及程序发现方面取得了进展。然而,搜索的有效性取决于搜索的位置,即如何将领域先验编码为可操作的结构化假设空间。为此,本文提出了一种紧凑的形式化理论,用以描述和衡量由领域先验引导的LLM辅助迭代搜索。我们将智能体表示为输入与输出间的模糊关系算子,以捕捉可行状态转移;智能体由此被约束在一个固定的安全包络内。为描述多步推理/搜索,我们通过单一延续参数对所有可达路径进行加权并求和,得到覆盖生成函数;这引出了可达性难度的度量,并为安全包络所诱导的图上的搜索提供了几何解释。我们进一步给出了最简可检验的推断,并通过多数投票实例进行了验证。该理论提供了一套可操作的语言和工具来度量智能体及其搜索空间,为LLM构建的迭代搜索提出了系统化的形式化描述。