Learning Structure-Aware Representations of Dependent Types
Paper in proceeding, 2024

Agda is a dependently-typed programming language and a proof assistant, pivotal in proof formalization and programming language theory.This paper extends the Agda ecosystem into machine learning territory, and, vice versa, makes Agda-related resources available to machine learning practitioners.We introduce and release a novel dataset of Agda program-proofs that is elaborate and extensive enough to support various machine learning applications -- the first of its kind.Leveraging the dataset's ultra-high resolution, which details proof states at the sub-type level, we propose a novel neural architecture targeted at faithfully representing dependently-typed programs on the basis of structural rather than nominal principles.We instantiate and evaluate our architecture in a premise selection setup, where it achieves promising initial results, surpassing strong baselines.

Author

Konstantinos Kogkalidis

Aalto University

University of Bologna

Orestis Melkonian

Input/Output Global (IOG)

Jean-Philippe Bernardy

University of Gothenburg

Chalmers, Computer Science and Engineering (Chalmers), Computing Science

Advances in Neural Information Processing Systems

10495258 (ISSN)

Vol. 37

Advances in Neural Information Processing Systems
Vancouver, Canada,

Areas of Advance

Information and Communication Technology

Subject Categories (SSIF 2025)

Computer Sciences

Computer and Information Sciences

More information

Latest update

3/19/2025