论文标题
关于学习线性dag的稀疏性和DAG约束的作用
On the Role of Sparsity and DAG Constraints for Learning Linear DAGs
论文作者
论文摘要
基于定向无环图(DAG)的学习图形结构是一个具有挑战性的问题,部分是由于可能的图形的较大搜索空间。最新的工作线将结构学习问题作为连续约束优化任务,使用最小二乘目标和DAG的代数表征。但是,该配方需要硬盘约束,并可能导致优化困难。在本文中,我们研究了线性高斯和非高斯病例中学习DAG模型的稀疏性和DAG限制的渐近作用,并研究了它们在有限样本状态中的有用性。基于理论结果,我们制定了基于可能性的分数功能,并表明只需要应用软稀疏性和DAG约束来学习与地面真实DAG等同的DAG。这导致了一个不受约束的优化问题,该问题更容易解决。使用基于梯度的优化和GPU加速度,我们的过程可以轻松处理数千个节点,同时保持高精度。广泛的实验验证了我们提出的方法的有效性,并表明dag-penatization-penation-penized-bigelieghile目标确实比具有硬dag约束的最小平方相比是有利的。
Learning graphical structures based on Directed Acyclic Graphs (DAGs) is a challenging problem, partly owing to the large search space of possible graphs. A recent line of work formulates the structure learning problem as a continuous constrained optimization task using the least squares objective and an algebraic characterization of DAGs. However, the formulation requires a hard DAG constraint and may lead to optimization difficulties. In this paper, we study the asymptotic role of the sparsity and DAG constraints for learning DAG models in the linear Gaussian and non-Gaussian cases, and investigate their usefulness in the finite sample regime. Based on the theoretical results, we formulate a likelihood-based score function, and show that one only has to apply soft sparsity and DAG constraints to learn a DAG equivalent to the ground truth DAG. This leads to an unconstrained optimization problem that is much easier to solve. Using gradient-based optimization and GPU acceleration, our procedure can easily handle thousands of nodes while retaining a high accuracy. Extensive experiments validate the effectiveness of our proposed method and show that the DAG-penalized likelihood objective is indeed favorable over the least squares one with the hard DAG constraint.