My efforts in the area of AI SEA (Safety, Ethics and Alignment) have broadly been in these two areas:
Red-Teaming + Adversarial attacks and Dataset auditing + bias mitigation and Model auditing.


A sampling of my research publications spanning these two areas : Model auditing, Red-Teaming and Adversarial attacks


PS: My more recent AI safety chronicles can be found here