Skip to content

isVy08/ACCESS

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

33 Commits
 
 
 
 
 
 

Repository files navigation

ACCESS

This repo includes the demo benchmark and codes for reproducing the experiments in the paper ACCESS: A Benchmark for Abstract Causal Event Discovery and Reasoning to appear in NAACL 2025.

Dataset

ACCESS is a benchmark for discovery and reasoning over abstract causal events in everyday life. The folder benchmark/ provides an extraction of the ACCESS data for demonstrative purposes only.

To examine our benchmark, please refer to the following scripts

  • data_generator.py
  • metadata.py
  • metagraph.py.
  • test_graph.py.

The post-processed GLUCOSE data and other utilities are provided in this Google folder.

For requests to access the full benchmark, please contact Vy Vo at tran[dot]vo[at]monash[dot]edu.

Experiments

1. Clustering for Abstract Event Extraction

To do automatic clustering, you first need to create the data directory mkdir data, then download to data/ the related objects from the above folder. Then run the following command lines:

python run_clustering.py pivot
python tune_clustering.py pivot

Other algorithms for ablation are also provided, e.g., louvain, leiden, optics.

2. Statistical Structure Learning

You need to first install gcastle and causal-learn libraries to run the statistical causal discovery algorithms. Then, specify the graph size (e.g. $25$ nodes), method (e.g. NOTEARS) and run

python run_cd_alg.py 25 NOTEARS

3. Causal Reasoning with LLMs

ACCESS can be used to evaluate LLMs on the following tasks:

  • Abstract event identification:

    • refer to map_abstraction.py for how to map ACCESS abstractions to mentions in GLUCOSE.
  • Causal discovery:

    • refer to generate_causal_pairs.py for how to create a test set of non-contextual causal relations from ACCESS.
  • Causal QA reasoning:

    • refer to generate_causal_qa.py for how to automatically construct the GLUCOSE QA dataset and benchmark/ for an example.
    • note that the GLUCOSE question bank released in ACCESS has been subjected to human validation for true causality.

About

ACCESS: A Benchmark for Abstract Causal Event Discovery and Reasoning

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages