Turning Logic Against Itself:
Probing Model Defenses Through Contrastive Questions

Jailbreak attack to bypass LLM safety mechanisms.
Jailbreak attack to bypass LLM safety mechanisms.
Fine-grained error evaluation of long-form LLM responses.
Evaluation of the hype around LLM abilities..
Counterfactual augmentation of data to improve out-of-domain generalization of models.
Platform to interpret and explain the predictions of machine learning models.
An online QA platform which allows users to query, compare and evaluate different models.