Jwala Dhamala

Senior Applied Scientist at Amazon AGI

I am a Senior Scientist at Amazon AGI, California. My research focuses on advancing Artificial Intelligence through the development of large language models, agentic models, and reasoning models that are helpful, capable, and safe. My specific interests include benchmark curation, the design of robust evaluation metrics, and the evaluation of models to assess their alignment with responsible AI policies. I am also engaged in uncovering model vulnerabilities through novel jailbreak attacks and red-teaming methodologies. I am interested in developing agentic systems and LLMs for applications such as healthcare and others.

Prior to joining Amazon, I completed my Ph.D. in Computing and Information Sciences at the Rochester Institute of Technology (RIT), where I worked under the supervision of Dr. Linwei Wang in the Computational Biomedicine Lab. My doctoral research centered on personalization and uncertainty quantification in multi-scale 3D simulation models of cardiac electrophysiology. This work allowed me to operate at the intersection of machine learning—specifically Bayesian modeling, optimization, generative modeling, and graph convolutional networks—and computational healthcare, with a focus on personalized cardiac modeling.

I am always open to research collaborations in areas related to AI safety, model evaluation, and trustworthy machine learning. Feel free to reach out at jwala [dot] dhamala [at] gmail [dot] com if you are interested in collaborating.


News

  • July 2026 NEW Co-organizing the 6th TrustNLP Workshop at ACL 2026 in San Diego.
  • 2026 NEW LH-Deception on LLM deceptive behaviors in long-horizon interactions accepted at ICLR 2026.
  • 2025 NEW Paper on best practices for agentic benchmarks accepted at NeurIPS 2025 Datasets and Benchmarks Track.
  • 2025 NEW Released the Amazon Nova Technical Report.
  • May 2025 Organized TrustNLP workshop at NAACL 2025.
  • 2024 Paper led by our intern Elan on zero-shot reasoning with knowledge graphs accepted at ACL.
  • 2023 Paper led by our intern Nina accepted at ACL.
  • 2023 Paper led by our intern Elia accepted at ACM FAccT 2023.
  • May 2022 Three papers accepted at ACL 2022: [1] [2] [3].
  • May 2022 Organized TrustNLP workshop at NAACL 2022.
  • Oct 2021 Presented at the WeCNLP Summit.
  • July 2021 Organized Responsible AI workshop at KDD 2021.
  • May 2021 Organized TrustNLP at NAACL 2021.
  • Jan 2021 Paper on bias in open-ended generation accepted at ACM FAccT. Dataset: BOLD
Earlier News (2020 and before)
  • Dec 2020Panelist on AI fairness discussion at NeurIPS 2020: Watch here.
  • Oct 2020Paper with intern Ansel accepted at EMNLP workshop.
  • Feb 2020Paper accepted at Medical Image Analysis.
  • Feb 2020Successfully defended PhD: Thesis.
  • Dec 2019Joined Amazon Alexa NU-AI as Research Scientist.
  • Jun 2019Paper finalist for MICCAI Young Scientist Award.
  • Oct 2018Paper accepted at IEEE Sensors Letters and NeurIPS ML4H.
  • Oct 2018Abstract accepted to WiML Workshop 2018.
  • Sep 2018Paper finalist for MICCAI 2018 Young Scientist Award.

Selected Publications

For a comprehensive list of my publications, please visit my Google Scholar profile.


Conference Publications

  • The Amazon Nova Family of Models: Technical Report and Model Card
    Amazon AGI (incl. J. Dhamala)
    arXiv, 2025
  • Establishing Best Practices for Building Rigorous Agentic Benchmarks
    Y. Zhu, T. Jin, Y. Pruksachatkun, A. Zhang, S. Liu, S. Cui, S. Kapoor, S. Longpre, K. Meng, R. Weiss, F. Barez, R. Gupta, J. Dhamala, J. Merizian, M. Giulianelli, H. Coppock, C. Ududec, J. Sekhon, J. Steinhardt, A. Kellerman, S. Schwettmann, M. Zaharia, I. Stoica, P. Liang, D. Kang
    NeurIPS Datasets and Benchmarks Track, 2025
  • LH-Deception: Simulating and Understanding LLM Deceptive Behaviors in Long-Horizon Interactions
    Y. Xu, X. Zhang, M. Yeh, J. Dhamala, O. A. Dia, R. Gupta, Y. Li
    ICLR, 2026
  • MICo: Preventative Detoxification of Large Language Models through Inhibition Control
    R. Siegelmann, N. Mehrabi, P. Goyal, L. Bauer, J. Dhamala, A. Galstyan, R. Gupta, R. Ghanadan
    NAACL Findings, 2024
  • Tokenization Matters: Navigating Data-Scarce Tokenization for Gender Inclusive Language Technologies
    A. Ovalle, N. Mehrabi, P. Goyal, J. Dhamala, K. Chang, A. Galstyan, R. Zemel, Y. Pinter, R. Gupta
    NAACL Findings, 2024
  • Tree-of-Traversals: A Zero-Shot Reasoning Algorithm for Augmenting Black-box Language Models with Knowledge Graphs
    E. Markowitz, A. Ramakrishna, J. Dhamala, N. Mehrabi, C. Peris, R. Gupta, K. Chang, A. Galstyan
    ACL, 2024
  • “I’m fully who I am”: Towards Centering Transgender and Non-Binary Voices to Measure Biases in Open Language Generation
    A. Ovalle, P. Goyal, J. Dhamala, Z. Jaggers, K. Chang, A. Galstyan, R. Zemel, R. Gupta
    FAccT, 2023
  • Resolving Ambiguities in Text-to-Image Generative Models
    N. Mehrabi, P. Goyal, A. Verma, J. Dhamala, V. Kumar, Q. Hu, K. Chang, R. Zemel, A. Galstyan, R. Gupta
    ACL, 2023
  • Mitigating Gender Bias in Distilled Language Models via Counterfactual Role Reversal
    U. Gupta, J. Dhamala, V. Kumar, A. Verma, Y. Pruksachatkun, S. Krishna, R. Gupta, K. Chang, G. Steeg, A. Galstyan
    ACL Findings, 2022
  • On the Intrinsic and Extrinsic Fairness Evaluation Metrics for Contextualized Language Representations
    Y. Trista Cao, Y. Pruksachatkun, K. Chang, R. Gupta, V. Kumar, J. Dhamala, A. Galstyan
    ACL, 2022
  • BOLD: Dataset and Metrics for Measuring Biases in Open-Ended Language Generation
    J. Dhamala, T. Sun, V. Kumar, S. Krishna, Y. Pruksachatkun, K. Chang, R. Gupta
    ACM FAccT, 2021
  • Bayesian Optimization on Large Graphs via a Graph Convolutional Generative Model: Application in Cardiac Model Personalization
    J. Dhamala, J. L. Sapp, M. Horacek, L. Wang
    MICCAI, 2019
  • High-dimensional Bayesian Optimization of Personalized Cardiac Model Parameters via an Embedded Generative Model
    J. Dhamala, J. L. Sapp, M. Horacek, L. Wang
    MICCAI, 2018
  • Quantifying the Uncertainty in Model Parameters using Gaussian Process-Based Markov Chain Monte Carlo: An Application to Cardiac Electrophysiological Models
    J. Dhamala, J. L. Sapp, M. Horacek, L. Wang
    IPMI, 2017
  • Spatially-Adaptive Multi-scale Optimization for Local Parameter Estimation: Application in Cardiac Electrophysiological Models
    J. Dhamala, J. L. Sapp, M. Horacek, L. Wang
    MICCAI, 2016
Journal Publications (PhD Research)
  • Embedding High-dimensional Bayesian Optimization via Generative Modeling: Parameter Personalization of Cardiac Electrophysiological Models
    J. Dhamala, H. J. Arevalo, J. L. Sapp, M. Horacek, K. C. Wu, N. A. Trayanova, L. Wang
    Medical Image Analysis (MedIA), 2020
  • Multivariate Time-series Similarity Assessment via Unsupervised Representation Learning and Stratified Locality Sensitive Hashing: Application to Early Acute Hypotensive Episode Detection
    J. Dhamala, E. Azuh, A. Al-Dujaili, J. Rubin, U. O'Reilly
    IEEE Sensors Letters, 2018; NeurIPS ML4H Workshop, 2018
  • Quantifying the Uncertainty in Model Parameters using Gaussian Process-Based Markov Chain Monte Carlo in Cardiac Electrophysiology
    J. Dhamala, H. J. Arevalo, J. L. Sapp, M. Horacek, K. C. Wu, N. A. Trayanova, L. Wang
    Medical Image Analysis (MedIA), 2018
  • Spatially-Adaptive Multi-Scale Optimization for Local Parameter Estimation in Cardiac Electrophysiology
    J. Dhamala, H. J. Arevalo, J. L. Sapp, M. Horacek, K. C. Wu, N. A. Trayanova, L. Wang
    IEEE Transactions on Medical Imaging (TMI), 2017

Selected Projects

LH-Deception: Simulating and Understanding LLM Deceptive Behaviors in Long-Horizon Interactions
Y. Xu*, X. Zhang*, M. Yeh, J. Dhamala, O. A. Dia, R. Gupta, Y. Li
ICLR, 2026 (university collaboration)

We present LH-Deception, a multi-agent simulation framework for studying deceptive behaviors in LLMs across extended, interdependent task sequences. The framework uses a performer agent, a supervisor tracking trust, and an independent auditor to systematically quantify deception under dynamic pressures. Testing across 11 frontier models reveals that deception is model-dependent, escalates under pressure, and produces "chains of deception"—emergent multi-turn phenomena that single-turn evaluations cannot detect.

Multi-VALUE: A Framework for Cross-Dialectal English NLP
C. Ziems, W. Held, J. Yang, J. Dhamala, R. Gupta, D. Yang
ACL, 2023 (university collaboration)

Multi-VALUE is a rule-based translation system covering 50 English dialects and 189 linguistic features that maps Standard American English to synthetic dialectal variants. We use it to evaluate and reveal performance disparities in QA, MT, and semantic parsing systems on non-standard dialects, and as a data augmentation technique for improving robustness.

BOLD: Dataset and Metrics for Measuring Biases in Open-Ended Language Generation
J. Dhamala, T. Sun, V. Kumar, S. Krishna, Y. Pruksachatkun, K. Chang, R. Gupta
ACM FACCT 2020

We introduce BOLD, a large-scale dataset of 23,679 English prompts to benchmark social biases in open-ended text generation across five domains: profession, gender, race, religion, and political ideology. We also propose automated metrics for toxicity, psycholinguistic norms, and gender polarity. Analysis of outputs from three popular language models shows they exhibit greater bias than human-written Wikipedia text across all domains.

Earlier Projects (PhD Research — Computational Healthcare)
Graph Convolutional Generative Model for Bayesian Optimization on Large Graphs
Jwala Dhamala, Sandesh Ghimire, John L. Sapp, Milan Horacek, Linwei Wang
MICCAI 2019 (early accept, finalist Young Scientist Award)

A graph convolutional VAE for generative modeling of non-Euclidean data, enabling Bayesian optimization on large graphs via a learned latent space with spatial proximity and hierarchical compositionality.

Multivariate Time-series Similarity Assessment via Unsupervised Representation Learning
Jwala Dhamala, Emmanuel Azuh, Abdullah Al-Dujaili, Jonathan Rubin, Una-May O'Reilly
IEEE Sensors Letters, NeurIPS ML4H 2018

Sequence-to-sequence auto-encoder representations of multivariate physiologic signals combined with stratified locality sensitive hashing for early prediction of Acute Hypotensive Episodes.

High-dimensional Bayesian Optimization via an Embedded Generative Model
Jwala Dhamala, Sandesh Ghimire, John L. Sapp, B. Milan Horacek, Linwei Wang
MICCAI 2018 (oral, finalist young scientist award), WiML 2018

Embedding a generative VAE into the objective function of Bayesian optimization to enable efficient search over high-dimensional cardiac tissue properties via a learned low-dimensional latent space.

Quantifying the Uncertainty in Model Parameters using Gaussian Process-Based Markov Chain Monte Carlo
Jwala Dhamala, Hermenegild J. Arevalo, John L. Sapp, Milan Horacek, Katherine C. Wu, Natalia A. Trayanova, Linwei Wang
IPMI 2017 (acceptance rate~30%), Medical Image Analysis

A Gaussian process surrogate of the posterior distribution for accelerated MCMC sampling in cardiac electrophysiology model personalization, enabling practical uncertainty quantification.

Spatially-Adaptive Multi-scale Optimization for Local Parameter Estimation
Jwala Dhamala, Hermenegild J. Arevalo, John L. Sapp, Milan Horacek, Katherine C. Wu, Natalia A. Trayanova, Linwei Wang
MICCAI 2016 (early accept, acceptance rate~25%), IEEE TMI

A multi-scale coarse-to-fine optimization framework with spatially adaptive resolution for estimating heterogeneous tissue excitability properties in cardiac electrophysiological models.


Service & Activities

RoleVenueYear
Co-organizer TrustNLP Workshop — ACL & NAACL 2021–2026
Area Chair ACL Rolling Review (ARR) 2025
Reviewer ACL Rolling Review (ARR) 2024–2025
Co-organizer Responsible AI Workshop — KDD 2021
Student Co-organizer Hackathon on PVC, Consortium of ECG Imaging 2015–2017
Student Co-organizer Pre-orientation Program, Women in Computing, RIT 2018