Joshua Tom

Computer systems engineering

Hometown: Chandler, Arizona, United States

Graduation date: Spring 2028

Additional details: Honors student

Data icon, disabled. Four grey bars arranged like a vertical bar chart.

FURI | Fall 2025

Bias-proof Data: Evaluating LLM Generalization

Large language models (LLMs) are having considerable influence on real-world decisions and events. However, biases instilled from pre-training lead to unreliable performance. While other studies seek to reduce the impact of bias, the overall cause of bias remains widely unknown. The research team proposes using open-source LLMs and semantic word pairs to identify biased words or phrases that have a critical impact on model performance. This approach aims to identify how biases emerge through pre-training data. These findings may allow for future improvements in pre-training practices and model design, further enhancing their reliability, general robustness, and explainability for critical applications.

Mentor:

QR code for the current page

It’s hip to be square.

Students presenting projects at the Fulton Forge Student Research Expo are encouraged to download this personal QR code and include it within your poster. This allows expo attendees to explore more about your project and about you in the future. 

Right click the image to save it to your computer.

Additional projects from this student

Studying how subtle biases trigger LLM math errors will help develop more reliable AI for safer, smarter decision-making.

Mentor:

  • FURI
  • Spring 2025