The event will take place at Howard University. See Directions for more details.
| 9:00 - 9:30 | Registration + Breakfast (Light breakfast will be provided) |
| 9:30 - 9:45 | Opening Remark |
| 9:45 - 10:45 | Keynote #1 - {Dr. Mohit Iyyer} (University of Maryland) |
| 10:45 - 11:00 | Coffee Break |
| 11:00 - 12:00 | Poster Session #1 |
| 12:00 - 13:00 | Lunch (Lunch will be provided) |
| 13:00 - 14:00 | Keynote #2 - {Dr. Kianté Brantley} (Harvard University) |
| 14:00 - 14:15 | Coffee Break |
| 14:15 - 15:15 | Oral Paper Presentation (4 Oral Papers) |
| 15:15 - 16:30 | Poster Session #2 |
| 16:30 - 17:30 | Keynote #3 - {Dr. Gloria Washington} (Howard University) |
| 17:30 - 17:45 | Closing Remark & Award |
Assistant Professor of Computer Science
Harvard University
Dr. Kianté Brantley is an Assistant Professor of Computer Science at Kempner Institute and School of Engineering and Applied Sciences (SEAS) at Harvard University. His research focuses on problems at the intersection of machine learning and interactive decision-making, with the goal of improving the decision-making capabilities of foundation models. This involves studying, building, and improving techniques in reinforcement learning, imitation learning, and natural language processing.
Talk Title:
Regression as Policy Optimization: Advantages In, Policies Out
Abstract:
Reinforcement learning (RL) is an essential method for training large language models (LLMs), enabling better alignment with human preferences and enhanced reasoning capabilities. Nevertheless, RL post-training remains computationally demanding due to repeated rollouts, high-variance credit assignment, and the complexities of distributed systems that can introduce policy lag. A promising direction is to view KL-regularized policy optimization as a KL-prox (mirror-descent–style) step and solve it with a simple least-squares regression loss. This regression-based perspective addresses key RL challenges and enables more efficient post-training procedures.
In this talk, I will introduce a unified perspective encompassing three complementary approaches. A⋆-PO minimizes reliance on online sampling by utilizing offline value computation and optimal-advantage regression. OAPL supports scalable, fully off-policy regression, even when using stale rollouts and lagged policies in distributed settings. RDA2C offers a regularized dual-averaging approach that employs cumulative gradient information to stabilize updates and reduce variance, rather than relying on local, per-round mirror-descent steps. Although RDA2C has been primarily assessed on standard RL benchmarks rather than comprehensive LLM post-training, its focus on variance reduction and data reuse aligns closely with the stability challenges encountered in large-scale LLM alignment. Collectively, these methods provide an efficient toolkit for RL-based LLM post-training and present opportunities for further research and scalable deployment.
Associate Professor of Computer Science
University of Maryland, College Park
Dr. Mohit Iyyer is an associate professor in computer science at University of Maryland, College Park and a member of CLIP. Previously, he was an associate professor at UMass CS, a Young Investigator at AI2, and a PhD student at UMD CS. His research focuses on natural language processing and large language models. He is currently excited about; (a) long-form text generation, (b) long-context language understanding, (c) AI agents for collaborative writing, and (d) AI-generated text detection.
Talk Title:
Detecting and characterizing AI use in collaborative writings
Abstract:
Nowadays, lots of people are using AI to "help" write things. For instance, our recent study finds that ~9% of recently-published American newspaper articles are produced either partially or entirely by LLMs. However, most of this AI use is undisclosed to readers, raising important concerns of transparency and public trust. In this talk, I discuss the broad spectrum of AI-human writing, from minor edits for style/grammar to the generation of large blocks of content, in the context of what might be acceptable and unacceptable to a reader given lack of disclosure.
I'll then discuss EditLens, an automatic method that performs post-hoc AI detection: given a piece of text, can we tell to what extent AI was used to assist in its writing? Next, I'll turn to the frontier of AI writing: how good can fully AI-generated writing actually get? To explore this, I'll introduce autofiction.ai, our new web platform where readers can freely read, rate, and discuss full-length novels produced by frontier AI agents under transparent disclosure. Finally, I'll describe StoryScope, which conceptually moves AI detection beyond surface-level stylistic cues (which future models may not exhibit) by analyzing discourse-level narrative features such as character agency, temporal complexity, and moral ambiguity.
Associate Professor of Computer Science
Howard University
Dr. Gloria Washington is currently an Associate Professor of Computer Science at Howard University in Washington, DC. She was recently awarded the EBONY Magazine Power 100 Award for being a STEM Trailblazer 2025! She runs the Affective Biometrics Lab with her bright students. Her most notable project is Project Elevate Black Voices https://www.elevateblackvoiceshu.com. She is an empathetic technology researcher that focuses on the intersection of human-centered computing, affective computing, and biometrics. She likes to say her research seeks to give voices to everyone that felt silenced by asking questions like: how can technology impact positive human emotions while reducing barriers to entering technology and how can technology build lasting social impact through requiring persons to feel empathy...not just look away?
Talk Title:
The Origins of Our Discontent with Voice Assistant Technology
Abstract:
Despite their rapid diffusion into homes, appliances, and workplaces, voice assistant technologies continue to inspire widespread frustration, mistrust, and uneven adoption. Often framed as intuitive, neutral, and universally accessible interfaces to computation, these tools fail for many speakers of marginalized dialects and languages. This talk introduces the concept of Dialect Caste to explain the structural origins of our discontent with voice assistant technology found in everything.
Dialect Caste describes the hierarchical ordering of speech varieties within automated speech recognition (ASR) systems, in which dominant, standardized forms of language are treated as normative while other ways of speaking are systematically misrecognized or devalued. Drawing on sociolinguistics, critical race studies, and human–computer interaction, the talk situates contemporary voice assistants within longer histories of linguistic stratification tied to race, class, colonialism, and power.
It examines how core ASR techniques—statistical modeling, data-driven acoustic training, and lexicon design—have traditionally failed speakers of African American English and African creole–based languages. These failures found in the design of automatic speech recognition are not accidental: they stem from training data that underrepresents the unique morphosyntax, phonology, and prosody found in AAE and African creole-based languages. The talk highlights Project Elevate Black Voices, led by Howard University, as a critical intervention that challenges these norms by centering Black speech communities in the design, collection, and evaluation of voice technologies. Rather than treating dialectal variation as noise to be corrected, the project EBV reframes linguistic diversity as a design requirement and a site of epistemic authority. Finally, by analyzing everyday breakdowns in voice interaction through the lens of Dialect Caste, the paper argues that technical misrecognition produces social harm, reinforcing longstanding patterns of linguistic discrimination while shifting the burden of adaptation onto marginalized speakers. The talk concludes by calling for dialect-aware ASR paradigms and participatory governance models that treat voice technologies not merely as engineering artifacts, but as institutions that mediate whose voices are heard, valued, and understood.