Hi there👋, this repo contains the PyTorch implementation for our paper:
Counterfactual-Enhanced Information Bottleneck for Aspect-Based Sentiment Analysis
Mingshan Chang, Min Yang, Qingshan Jiang, Ruifeng Xu. AAAI, 2024.
Great success in the ABSA task? We found that deep ABSA models are prone to learning 🫧spurious correlations🫧 between input features and output labels, resulting in poor robustness and generalization capability!
📌 An example of the spurious correlation problem in ABSA:
To reduce spurious correlations for ABSA, we propose a novel Counterfactual-Enhanced Information Bottleneck framework (called CEIB), which extends the information bottleneck (IB) principle to a factual-counterfactual balancing setting and incorporates augmented counterfactual data, to learn more robust and generalizable representations.
- We employ the IB principle to discard spurious features or shallow patterns while retaining sufficient information about the sentiment label.
- We devise a multi-pattern prompting method, which utilizes LLM to automatically generate counterfactual samples featuring identical spurious context words but opposite sentiment polarities for the original training data.
- We separate the mutual information in the original IB objective into factual and counterfactual parts. By balancing the predictive information of these two parts, we can learn more robust and generalizable representations against the dataset bias.
(a) a counterfactual data augmentation module that utilizes LLM to generate counterfactual data for the original training data.
(b) an information bottleneck module with a factual-counterfactual balancing setting to learn more robust and generalizable representations.
- Python 3.9.7
- PyTorch 1.11.0
- Transformers 4.18.0
- CUDA 11.0
-
Data
We have provided the generated counterfactual data in data/augmented_t5_xxl/.
You may also run the command:bash aug_data.sh
to generate the counterfactual data yourself. Before that, you should download t5-xxl and set the parameter--model_name_or_path
in aug_data.sh to your local directory. -
Models
Download the PyTorch version bert-base-uncased and set the parameter--model_dir
to your local directory.
-
Run the command:
bash run_CEIB_xxx.sh
, e.g. runbash run_CEIB_res14.sh
to train with REST14 dataset. -
More arguments can be found in run.py. Feel free to set parameters e.g.
--save_folder
(to save training results) and--data_dir
(to load training&testing data) to your customized path.
@article{Chang_Yang_Jiang_Xu_2024,
title={Counterfactual-Enhanced Information Bottleneck for Aspect-Based Sentiment Analysis},
volume={38},
url={https://ojs.aaai.org/index.php/AAAI/article/view/29726}, DOI={10.1609/aaai.v38i16.29726},
number={16},
journal={Proceedings of the AAAI Conference on Artificial Intelligence},
author={Chang, Mingshan and Yang, Min and Jiang, Qingshan and Xu, Ruifeng},
year={2024},
month={Mar.},
pages={17736-17744}
}
🤘Please cite our paper and kindly give a star if you find this repo helpful.