Sophie L. Wang
MIT EECS | Philips Undergraduate Research and Innovation Scholar
Words That Make Language Models Perceive
2025–2026
Electrical Engineering and Computer Science; Mathematics
- AI and Machine Learning
Isola, Phillip
Large language models (LLMs) trained purely on text ostensibly lack any direct perceptual experience, yet their internal representations are implicitly shaped by multimodal regularities encoded in language. We test the hypothesis that explicit sensory prompting can surface this latent structure, bringing a text-only LLM into closer representational alignment with specialist vision and audio encoders. When a sensory prompt tells the model to ‘see’ or ‘hear’, it cues the model to resolve its next-token predictions as if they were conditioned on latent visual or auditory evidence that is never actually supplied. Our findings reveal that lightweight prompt engineering can reliably activate modality-appropriate representations in purely text-trained LLMs.
I’m excited to work on open-ended problems in representation learning and to continue growing as a researcher. Deep learning is evolving rapidly, and I’m grateful for the opportunity to explore this field through the SuperUROP program, with the support of Philips and my lab.
