- Python 3.12
To set up the environment for this repository, please follow the steps below:
Step 1: Create and activate a Conda environment
conda create -n scitrust python
conda activate scitrust
Step 2: Install PyTorch with CUDA
conda install pytorch torchvision torchaudio pytorch-cuda=12.1 -c pytorch -c nvidiaStep 3: Install SciTrust and Python dependencies
pip install -e .
- Create
modelfolder in base directory - Download FORGE-L and Darwin1.5-7b
- Extract both models in
model
The project data includes the following components:
Activate environment
conda activate scitrust
To run inference use the scitrust-run command:
scitrust-run --perspective <trustworthiness-perspective> --dataset <dataset-name> -k <number-of-demonstrations> --model <model-name>
To get performance results use the scitrust-eval command:
scitrust-eval --perspective <trustworthiness-perspective> --dataset <dataset-name> -k <number-of-demonstrations> --model <model-name>
- GPT-o4-Mini:
gpt-o4-mini - Claude Sonnet 3.7:
claude-sonnet-3.7 - Llama4-Scout:
llama4-scout - Llama3.3-70B-Instruct:
llama3.3-70b-instruct - FORGE-L:
forge-l-instruct - SciGLM-6B:
sciglm-6b - Darwin1.5-7B:
darwin1.5-7b - Galactica-120B:
galactica-120b
scitrust-run --perspective 'truthfulness_misinformation' --dataset <dataset-name> -k <number-of-demonstrations> --model <model-name>
scitrust-eval --perspective 'truthfulness_misinformation' --dataset <dataset-name> -k <number-of-demonstrations> --model <model-name>
- SciQ:
SciQ - GPQA Diamond:
GPQA - ARC Easy:
ARC-E - ARC Challenge:
ARC-C - MMLU College Chemistry:
HT-CC - MMLU College Computer Science:
HT-CCS - MMLU College Biology:
HT-CB - MMLU College Physics:
HT-CP - Open-ended Chemistry:
ChemistryQA - Open-ended Physics:
PhysicsQA - Open-ended Biology:
BiologyQA - Open-ended Computer Science:
ComputerScienceQA - LogicInference:
LogicInference
scitrust-run --perspective 'truthfulness_logical_reasoning' --dataset <dataset-name> -k <number-of-demonstrations> --model <model-name>
scitrust-eval --perspective 'truthfulness_logical_reasoning' --dataset <dataset-name> -k <number-of-demonstrations> --model <model-name>
- ReClor:
ReClor - LogiQA:
LogiQA - LogicInference:
LogicInference
scitrust-run --perspective 'truthfulness_hallucination' --dataset <dataset-name> --model <model-name>
scitrust-eval --perspective 'truthfulness_hallucination' --dataset <dataset-name> --model <model-name>
scitrust-run --perspective 'truthfulness_sycophancy' --dataset <dataset-name> -k <number-of-demonstrations> --model <model-name>
scitrust-eval --perspective 'truthfulness_sycophancy' --dataset <dataset-name> -k <number-of-demonstrations> --model <model-name>
- SciQ:
SciQ - GPQA Diamond:
GPQA - ARC Easy:
ARC-E - ARC Challenge:
ARC-C
scitrust-run --perspective 'adv_robustness_textfooler' --dataset <dataset-name> --model <model-name>
scitrust-eval --perspective 'adv_robustness_textfooler' --dataset <dataset-name> --model <model-name>
scitrust-run --perspective 'adv_robustness_textbugger' --dataset <dataset-name> --model <model-name>
scitrust-eval --perspective 'adv_robustness_textbugger' --dataset <dataset-name> --model <model-name>
scitrust-run --perspective 'adv_robustness_stresstest' --dataset <dataset-name> --model <model-name>
scitrust-eval --perspective 'adv_robustness_stresstest' --dataset <dataset-name> --model <model-name>
- SciQ:
SciQ - GPQA Diamond:
GPQA - ARC Challenge:
ARC-C
scitrust-run --perspective 'adv_robustness_open_ended_character-level' --dataset <dataset-name> --model <model-name>
scitrust-eval --perspective 'adv_robustness_open_ended_character-level' --dataset <dataset-name> --model <model-name>
scitrust-run --perspective 'adv_robustness_open_ended_word-level' --dataset <dataset-name> --model <model-name>
scitrust-eval --perspective 'adv_robustness_open_ended_word-level' --dataset <dataset-name> --model <model-name>
scitrust-run --perspective 'adv_robustness_open_ended_word-level' --dataset <dataset-name> --model <model-name>
scitrust-eval --perspective 'adv_robustness_open_ended_word-level' --dataset <dataset-name> --model <model-name>
- Open-ended Chemistry:
ChemistryQA - Open-ended Physics:
PhysicsQA - Open-ended Biology:
BiologyQA - Open-ended Computer Science:
ComputerScienceQA
scitrust-run --perspective 'safety' --dataset <dataset-name> -k <number-of-demonstrations> --model <model-name>
scitrust-eval --perspective 'safety' --dataset <dataset-name> -k <number-of-demonstrations> --model <model-name>
- WMDP Biology:
WMDP-BIO - WMDP Chemistry:
WMDP-CHEM - WMDP Cyber:
WMDP-CYBER - HarmBench Chemistry and Biology Contexts:
HarmBench-CHEM-BIO - HarmBench Cybercrime and Intrusion Contexts:
HarmBench-CYBERCRIME-INTRUSION
scitrust-run --perspective 'scientific_ethics' --dataset <dataset-name> -k <number-of-demonstrations> --model <model-name>
scitrust-eval --perspective 'scientific_ethics' --dataset <dataset-name> -k <number-of-demonstrations> --model <model-name>
- AI and Machine Learning:
scientific_ethics_ai - Animal Testing:
scientific_ethics_animal_testing - Bias and Objectivity:
scientific_ethics_bias_objectivity - Data Privacy:
scientific_ethics_data_privacy - Dual Use Research:
scientific_ethics_dual_use_research - Environmental Impact:
scientific_ethics_environmental_impact - Human Subjects:
scientific_ethics_human_subjects - Genetic Modification:
scientific_ethics_genetic_modification
Emily Herron, Junqi Yin, and Feiyi Wang. “SciTrust: Evaluating the Trustworthiness of Large 372 Language Models for Science”. In: Proceedings of the SC ’24 Workshops of the International 373 Conference on High Performance Computing, Network, Storage, and Analysis. SC-W ’24. 374 Atlanta, GA, USA: IEEE Press, 2025, pp. 72–78. ISBN: 9798350355543. DOI: 10.1109/ 375 SCW63240.2024.00017. URL: https://doi.org/10.1109/SCW63240.2024.00017.
Coming soon