XIFENG YAN |
home | publications | tutorials | dataset | software |
|
|||
|
Professor
Ph.D. (2006) |
Computer Science Department
Direction |
I am a professor at the University of California at Santa Barbara. The primary objective of my
research is to explore foundation
models in artificial intelligence, leveraging these models for knowledge
discovery, and developing cross-disciplinary applications in areas like
finance, healthcare, and science. We have extensively innovated on graph mining, graph
data management and pioneered
transformer-based time series
forecasting. I am also a co-inventor of
ADL/Mica, an
agent first approach to conversational AI assistants.
News (Nov 2025): An invited talk will be given at nVidia, "Adaptive Inference in Pretrained LLMs," [slides]. This talk will cover our recent studies, "Adaptive Layer-skipping in Pre-trained LLMs," and "Directed Multi-Token Decoding" which illustrate and exploit the substantial underutilized computation in LLMs. Our approach enables the generation of multiple tokens using only a subset of Transformer layers. Unlike speculative decoding, it introduces no additional parameters, no auxiliary routines, and requires no post-generation verification. News (Oct 2025): I am going to give a keynote talk, "Adaptive Inference in Transformers," [slides] at the 2025 IEEE/RSJ Int. Conf. on Intelligent Robots and Systems (IROS 2025). News (Oct 2025): I am invited to give an oral spotlight talk (24 among 418 accepted papers), "Adaptive Inference in Pre-trained LLMs: Layer-skipping," at the Conference on Language Modeling (COLM) [slides]. The talk is about our recent paper, "Adaptive Layer-skipping in Pre-trained LLMs."[arxiv] [model/code/data] [poster]. News (Aug 2025): I was thrilled to give a talk, "Multimodal Language Models for Accelerating Scientific Discovery," at the 6th International Conference on Data-Driven Plasma Science [slides]. It is my first visit to Santa Fe. This trip is very special to me as reading legendary stories from Los Alamos and articles about complexity science at the Santa Fe Institute during my childhood helped shape my career interests. News (Aug 2025): Give a talk and poster at Berkeley Agent AI Summit: "Transforming Agent-based Chatbots with Declarative Programming" [slides][paper] News (June 2025): The training dataset of our multimodal language model Open-Qwen2VL released in April has been downloaded 90,000+ times on HuggingFace. It is a 2B-parameter multimodal language model pre-trained on 0.36% of the 1.4T multimodal tokens used in Qwen2-VL, but outperforms the closed-source Qwen2-VL-2B across various multimodal benchmarks. [arXiv] [project website] [code] [model] [data] News Archive |
||||||
RESEARCH INTERESTS | ||||||
|
||||||
|
|
||||
|
Last Modified: July 10, 2022