XIFENG YAN |
home | publications | tutorials | dataset | software |
|
|||
|
Professor
Ph.D. (2006) |
Computer Science Department
Direction |
I am a professor at the University of California at Santa Barbara. The primary objective of my
research is to explore foundation
models in artificial intelligence, leveraging these models for knowledge
discovery, and developing cross-disciplinary applications in areas like
finance, healthcare, and science. We have extensively innovated on graph mining, graph
data management and pioneered
transformer-based time series
forecasting. I am also a co-inventor of
ADL/Mica, an
agent first approach to conversational AI assistants.
News (Oct 2025): I am going to give a keynote talk, "Adaptive Inference in Transformers," [slides] at the 2025 IEEE/RSJ Int. Conf. on Intelligent Robots and Systems (IROS 2025). This talk will present our recent studies that illustrate and leverage the underutilized computation in Transformer. News (Oct 2025): I am invited to give an oral spotlight talk (24 among 418 accepted papers), "Adaptive Inference in Pre-trained LLMs: Layer-skipping," at the Conference on Language Modeling (COLM) [slides]. The talk is about our recent paper, "Adaptive Layer-skipping in Pre-trained LLMs."[arxiv] [model/code/data]. News (Aug 2025): I was thrilled to give a talk, "Multimodal Language Models for Accelerating Scientific Discovery," at the 6th International Conference on Data-Driven Plasma Science [slides]. It is my first visit to Santa Fe. This trip is very special to me as reading legendary stories from Los Alamos and articles about complexity science at the Santa Fe Institute during my childhood helped shape my career interests. News (Aug 2025): Give a talk and poster at Berkeley Agent AI Summit: "Transforming Agent-based Chatbots with Declarative Programming" [slides][paper] News (June 2025): The training dataset of our multimodal language model Open-Qwen2VL released in April has been downloaded 66,000+ times on HuggingFace. It is a 2B-parameter multimodal language model pre-trained on 0.36% of the 1.4T multimodal tokens used in Qwen2-VL, but outperforms the closed-source Qwen2-VL-2B across various multimodal benchmarks. [arXiv] [project website] [code] [model] [data] News Archive |
||||||
RESEARCH INTERESTS | ||||||
|
||||||
|
|
||||
|
Last Modified: July 10, 2022