Publications

ChainBuddy: An AI Agent System for Generating LLM Pipelines
As large language models (LLMs) advance, their potential applications have grown significantly. However, it remains difficult to evaluate LL… (voir plus)M behavior on user-specific tasks and craft effective pipelines to do so. Many users struggle with where to start, often referred to as the"blank page"problem. ChainBuddy, an AI assistant for generating evaluative LLM pipelines built into the ChainForge platform, aims to tackle this issue. ChainBuddy offers a straightforward and user-friendly way to plan and evaluate LLM behavior, making the process less daunting and more accessible across a wide range of possible tasks and use cases. We report a within-subjects user study comparing ChainBuddy to the baseline interface. We find that when using AI assistance, participants reported a less demanding workload and felt more confident setting up evaluation pipelines of LLM behavior. We derive insights for the future of interfaces that assist users in the open-ended evaluation of AI.
ChainBuddy: An AI-assisted Agent System for Generating LLM Pipelines
ChainBuddy: An AI-assisted Agent System for Generating LLM Pipelines
Development of small, cost‐efficient scintillating fiber detectors for automated synthesis of positron emission tomography radiopharmaceuticals
Hailey Ahn
Liam Carroll
Robert Hopewell
I-Huang Tsai
Dean Jolly
Gassan Massarweh
The Bifurcation Method: White-Box Observation Perturbation Attacks on Reinforcement Learning Agents on a Cyber Physical System
KIERNAN BRODA-MILIAN
Ranwa Al Mallah
Diagnostic tests for infections in critically ill immunocompromised patients
Adrien Joseph
Lara Zafrani
Dynamic HumTrans: Humming Transcription Using CNNs and Dynamic Programming
Isaac Neri Gomez-Sarmiento
Faez Amjed Mezdari
Enhancing Logical Reasoning in Large Language Models through Graph-based Synthetic Data
Jiaming Zhou
Abbas Ghaddar
Ge Zhang
Yaochen Hu
Soumyasundar Pal
Bin Wang
Yingxue Zhang
Jianye Hao
Despite recent advances in training and prompting strategies for Large Language Models (LLMs), these models continue to face challenges with… (voir plus) complex logical reasoning tasks that involve long reasoning chains. In this work, we explore the potential and limitations of using graph-based synthetic reasoning data as training signals to enhance LLMs' reasoning capabilities. Our extensive experiments, conducted on two established natural language reasoning tasks -- inductive reasoning and spatial reasoning -- demonstrate that supervised fine-tuning (SFT) with synthetic graph-based reasoning data effectively enhances LLMs' reasoning performance without compromising their effectiveness on other standard evaluation benchmarks.
Enhancing Logical Reasoning in Large Language Models through Graph-based Synthetic Data
Jiaming Zhou
Abbas Ghaddar
Ge Zhang
Yaochen Hu
Soumyasundar Pal
B. Wang
Yingxue Zhang
Jianye Hao
Explaining Network Decision Provides Insights on the Causal Interaction Between Brain Regions in a Motor Imagery Task
Davide Borra
Multi-modal Decoding of Reach-to-Grasping from EEG and EMG via Neural Networks
Davide Borra
Matteo Fraternali
Elisa Magosso
Relative biological effectiveness of clinically relevant photon energies for the survival of human colorectal, cervical, and prostate cancer cell lines
Joanna Li
N. Chabaytah
Joud Babik
Behnaz Behmand
H. Bekerat
Tanner Connell
Michael D C Evans
Russell Ruo
T. Vuong