Use of Metamorphic Testing Techniques for Improving Causal Discovery using LLMs
Abstract
This thesis explores the use of Large Language Models (LLMs) in causal discovery, acritical task for understanding complex relationships in data. Traditional methods oftenfall short when dealing with high-dimensional datasets and confounders. LLMs, leveragingmetadata and contextual information, offer a novel approach by emulating expert domainreasoning. To address the inherent biases and limitations of LLMs, this research integratesmetamorphic testing and prompt engineering. Metamorphic testing ensures the reliabilityof causal inferences by systematically validating model outputs, while prompt engineeringoptimizes input queries for improved accuracy. The study employs datasets such as Asiaand Child to evaluate these methods. Results indicate that combining metamorphictesting and prompt engineering significantly enhances the robustness and precision ofLLM-based causal discovery, offering a promising direction for future AI applications. Thisframework not only advances the theoretical understanding but also provides practicaltools for researchers and practitioners in the field.