Abstract
Large language models (LLMs) have caused a veritable revolution in the field of AI. However, LLMs do come with some considerable caveats including the lack of logical reasoning ability. This can make it challenging to use LLMs in environments where they need to give reliably correct answers. Recently, attempts have been made to alleviate this concern by generating a more transparent way of solving the problem using an LLM, instead of solving the problem directly with an LLM (so-called autoformalisation). Among others, answer set programs have been tried as a problem-solving intermediary in this context. However, current attempts at autoformalisation of answer set programs has been limited to toy examples or single, simple rules. In this work, we investigate the capabilities of LLMs in generating ASP that solve real-world scheduling problems, and identify techniques such as few-shot learning and chain-of-Thought as particularly succesful.
| Original language | English |
|---|---|
| Pages (from-to) | 157-168 |
| Number of pages | 12 |
| Journal | CEUR Workshop Proceedings |
| Volume | 4071 |
| Publication status | Published - 1 Jan 2025 |
| Event | 23rd International Workshop on Non-Monotonic Reasoning - Melbourne, aus, Melbourne, Australia Duration: 11 Nov 2025 → 13 Nov 2025 Conference number: 23 https://nmr.krportal.org/2025/ |