Retrieval Augmented Argumentation or Retrieval Augmented Argument Generation involves retrieving noisy evidence documents over the web and using them for subsequent argument generation. To facilitate RAG and computational argumentation research, we release ConQRet, a benchmark with popular controversial queries, paired with evidence documents retrieved and scraped over the public web, alongwith model-generated arguments.
- Retrieval Augmented Generation
- Evaluating Standalone Retrieval
- Evaluating RAG systems
Statistic | |
---|---|
Total topics | 98 |
Avg. docs per topic | 133 |
Avg. relevant docs per topic | 66 |
Avg. docs per stance | 33 |
Total documents retrieved & scraped | 6500 |
The total number of documents retrieved and scraped from the web are 6500.
python setup.py
Download the data from this Google Drive link, unzip it and copy it in the project home folder ("conqret-rag"). The password is provided at the end of the README.
This will run BM25 and BM25+GPT4o-mini reranker
python retriever.py
If you face a dlopen error, ensure that you set java home variables something like
import os
os.environ['JAVA_HOME'] = "...miniconda3/envs/conqret-rag/lib/jvm"
os.environ['JVM_PATH'] = "...miniconda3/envs/conqret-rag/lib/jvm/lib/server/libjvm.dylib"
Run the following script. It scrapes all the URLs present in the url_list.txt file.
python procon-parser.py
SaglyanchaVichaarVasudhaivaKutumbakam01293872
Do not publicly upload elsewhere. We are sharing the documents separately to mitigate the possibility of it being used for training, although we do not guarantee that many of them might already be used by popular models through other means.
@misc{dhole2024conqretbenchmarkingfinegrainedevaluation,
title={ConQRet: Benchmarking Fine-Grained Evaluation of Retrieval Augmented Argumentation with LLM Judges},
author={Kaustubh D. Dhole and Kai Shu and Eugene Agichtein},
year={2024},
eprint={2412.05206},
archivePrefix={arXiv},
primaryClass={cs.CL},
url={https://arxiv.org/abs/2412.05206},
}