Eric Xing

Eric Xing

Carnegie Mellon University

H-index: 114

North America-United States

Professor Information

University

Carnegie Mellon University

Position

President at Mohamed bin Zayed University of AI Professor of Computer Science U

Citations(all)

57613

Citations(since 2020)

33465

Cited By

37934

hIndex(all)

114

hIndex(since 2020)

87

i10Index(all)

435

i10Index(since 2020)

340

Email

University Profile Page

Carnegie Mellon University

Research & Interests List

Machine Learning

ML Systems

Optimization

Statistics

Network Analysis

Top articles of Eric Xing

Learning to Prompt Segment Anything Models

Segment Anything Models (SAMs) like SEEM and SAM have demonstrated great potential in learning to segment anything. The core design of SAMs lies with Promptable Segmentation, which takes a handcrafted prompt as input and returns the expected segmentation mask. SAMs work with two types of prompts including spatial prompts (e.g., points) and semantic prompts (e.g., texts), which work together to prompt SAMs to segment anything on downstream datasets. Despite the important role of prompts, how to acquire suitable prompts for SAMs is largely under-explored. In this work, we examine the architecture of SAMs and identify two challenges for learning effective prompts for SAMs. To this end, we propose spatial-semantic prompt learning (SSPrompt) that learns effective semantic and spatial prompts for better SAMs. Specifically, SSPrompt introduces spatial prompt learning and semantic prompt learning, which optimize spatial prompts and semantic prompts directly over the embedding space and selectively leverage the knowledge encoded in pre-trained prompt encoders. Extensive experiments show that SSPrompt achieves superior image segmentation performance consistently across multiple widely adopted datasets.

Authors

Jiaxing Huang,Kai Jiang,Jingyi Zhang,Han Qiu,Lewei Lu,Shijian Lu,Eric Xing

Journal

arXiv preprint arXiv:2401.04651

Published Date

2024/1/9

Judging llm-as-a-judge with mt-bench and chatbot arena

Evaluating large language model (LLM) based chat assistants is challenging due to their broad capabilities and the inadequacy of existing benchmarks in measuring human preferences. To address this, we explore using strong LLMs as judges to evaluate these models on more open-ended questions. We examine the usage and limitations of LLM-as-a-judge, including position, verbosity, and self-enhancement biases, as well as limited reasoning ability, and propose solutions to mitigate some of them. We then verify the agreement between LLM judges and human preferences by introducing two benchmarks: MT-bench, a multi-turn question set; and Chatbot Arena, a crowdsourced battle platform. Our results reveal that strong LLM judges like GPT-4 can match both controlled and crowdsourced human preferences well, achieving over 80\% agreement, the same level of agreement between humans. Hence, LLM-as-a-judge is a scalable and explainable way to approximate human preferences, which are otherwise very expensive to obtain. Additionally, we show our benchmark and traditional benchmarks complement each other by evaluating several variants of LLaMA and Vicuna. The MT-bench questions, 3K expert votes, and 30K conversations with human preferences are publicly available at https://github. com/lm-sys/FastChat/tree/main/fastchat/llm_judge.

Authors

Lianmin Zheng,Wei-Lin Chiang,Ying Sheng,Siyuan Zhuang,Zhanghao Wu,Yonghao Zhuang,Zi Lin,Zhuohan Li,Dacheng Li,Eric Xing,Hao Zhang,Joseph E Gonzalez,Ion Stoica

Journal

Advances in Neural Information Processing Systems

Published Date

2024/2/13

MobiLlama: Towards Accurate and Lightweight Fully Transparent GPT

"Bigger the better" has been the predominant trend in recent Large Language Models (LLMs) development. However, LLMs do not suit well for scenarios that require on-device processing, energy efficiency, low memory footprint, and response efficiency. These requisites are crucial for privacy, security, and sustainable deployment. This paper explores the "less is more" paradigm by addressing the challenge of designing accurate yet efficient Small Language Models (SLMs) for resource constrained devices. Our primary contribution is the introduction of an accurate and fully transparent open-source 0.5 billion (0.5B) parameter SLM, named MobiLlama, catering to the specific needs of resource-constrained computing with an emphasis on enhanced performance with reduced resource demands. MobiLlama is a SLM design that initiates from a larger model and applies a careful parameter sharing scheme to reduce both the pre-training and the deployment cost. Our work strives to not only bridge the gap in open-source SLMs but also ensures full transparency, where complete training data pipeline, training code, model weights, and over 300 checkpoints along with evaluation codes is available at : https://github.com/mbzuai-oryx/MobiLlama.

Authors

Omkar Thawakar,Ashmal Vayani,Salman Khan,Hisham Cholakal,Rao M Anwer,Michael Felsberg,Tim Baldwin,Eric P Xing,Fahad Shahbaz Khan

Journal

arXiv preprint arXiv:2402.16840

Published Date

2024/2/26

Temporally Disentangled Representation Learning under Unknown Nonstationarity

In unsupervised causal representation learning for sequential data with time-delayed latent causal influences, strong identifiability results for the disentanglement of causally-related latent variables have been established in stationary settings by leveraging temporal structure. However, in nonstationary setting, existing work only partially addressed the problem by either utilizing observed auxiliary variables (eg, class labels and/or domain indexes) as side information or assuming simplified latent causal dynamics. Both constrain the method to a limited range of scenarios. In this study, we further explored the Markov Assumption under time-delayed causally related process in nonstationary setting and showed that under mild conditions, the independent latent components can be recovered from their nonlinear mixture up to a permutation and a component-wise transformation, without the observation of auxiliary variables. We then introduce NCTRL, a principled estimation framework, to reconstruct time-delayed latent causal variables and identify their relations from measured sequential data only. Empirical evaluations demonstrated the reliable identification of time-delayed latent causal influences, with our methodology substantially outperforming existing baselines that fail to exploit the nonstationarity adequately and then, consequently, cannot distinguish distribution shifts.

Authors

Xiangchen Song,Weiran Yao,Yewen Fan,Xinshuai Dong,Guangyi Chen,Juan Carlos Niebles,Eric Xing,Kun Zhang

Journal

NeurIPS 2023

Published Date

2023/10/28

Cappy: Outperforming and boosting large multi-task lms with a small scorer

Large language models (LLMs) such as T0, FLAN, and OPT-IML excel in multi-tasking under a unified instruction-following paradigm, where they also exhibit remarkable generalization abilities to unseen tasks. Despite their impressive performance, these LLMs, with sizes ranging from several billion to hundreds of billions of parameters, demand substantial computational resources, making their training and inference expensive and inefficient. Furthermore, adapting these models to downstream applications, particularly complex tasks, is often unfeasible due to the extensive hardware requirements for finetuning, even when utilizing parameter-efficient approaches such as prompt tuning. Additionally, the most powerful multi-task LLMs, such as OPT-IML-175B and FLAN-PaLM-540B, are not publicly accessible, severely limiting their customization potential. To address these challenges, we introduce a pretrained small scorer,\textit {Cappy}, designed to enhance the performance and efficiency of multi-task LLMs. With merely 360 million parameters, Cappy functions either independently on classification tasks or serve as an auxiliary component for LLMs, boosting their performance. Moreover, Cappy enables efficiently integrating downstream supervision without requiring LLM finetuning nor the access to their parameters. Our experiments demonstrate that, when working independently on 11 language understanding tasks from PromptSource, Cappy outperforms LLMs that are several orders of magnitude larger. Besides, on 45 complex tasks from BIG-Bench, Cappy boosts the performance of the advanced multi-task LLM, FLAN-T5, by a large margin …

Authors

Bowen Tan,Yun Zhu,Lijuan Liu,Eric Xing,Zhiting Hu,Jindong Chen

Journal

Advances in Neural Information Processing Systems

Published Date

2024/2/13

AttentionPert: Accurately Modeling Multiplexed Genetic Perturbations with Multi-scale Effects

Genetic perturbations (i.e. knockouts, variants) have laid the foundation for our understanding of many diseases, implicating pathogenic mechanisms and indicating therapeutic targets. However, experimental assays are fundamentally limited in the number of perturbation conditions they can measure. Computational methods can fill this gap by predicting perturbation effects under unseen conditions, but accurately predicting the transcriptional responses of cells to unseen perturbations remains a significant challenge. We address this by developing a novel attention-based neural network, AttentionPert, which accurately predicts gene expression under multiplexed perturbations and generalizes to unseen conditions. AttentionPert integrates global and local effects in a multi-scale model, representing both the non-uniform system-wide impact of the genetic perturbation and the localized disturbance in a network of gene-gene similarities, enhancing its ability to predict nuanced transcriptional responses to both single and multi-gene perturbations. In comprehensive experiments, AttentionPert demonstrates superior performance across multiple datasets outperforming the state-of-theart method in predicting differential gene expressions and revealing novel gene regulations. AttentionPert marks a significant improvement over current methods, particularly in handling the diversity of gene perturbations and in predicting out-of-distribution scenarios.

Authors

Ding Bai,Caleb Ellington,Shentong Mo,Le Song,Eric Xing

Journal

bioRxiv

Published Date

2024/2/7

Squeeze, recover and relabel: Dataset condensation at imagenet scale from a new perspective

We present a new dataset condensation framework termed Squeeze, Recover and Relabel (SRe L) that decouples the bilevel optimization of model and synthetic data during training, to handle varying scales of datasets, model architectures and image resolutions for efficient dataset condensation. The proposed method demonstrates flexibility across diverse dataset scales and exhibits multiple advantages in terms of arbitrary resolutions of synthesized images, low training cost and memory consumption with high-resolution synthesis, and the ability to scale up to arbitrary evaluation network architectures. Extensive experiments are conducted on Tiny-ImageNet and full ImageNet-1K datasets. Under 50 IPC, our approach achieves the highest 42.5\% and 60.8\% validation accuracy on Tiny-ImageNet and ImageNet-1K, outperforming all previous state-of-the-art methods by margins of 14.5\% and 32.9\%, respectively. Our approach also surpasses MTT in terms of speed by approximately 52(ConvNet-4) and 16(ResNet-18) faster with less memory consumption of 11.6 and 6.4 during data synthesis. Our code and condensed datasets of 50, 200 IPC with 4K recovery budget are available at https://github. com/VILA-Lab/SRe2L.

Authors

Zeyuan Yin*,Eric Xing,Zhiqiang Shen*

Journal

Advances in Neural Information Processing Systems, Spotlight

Published Date

2024/2/13

Generating, Reconstructing, and Representing Discrete and Continuous Data: Generalized Diffusion with Learnable Encoding-Decoding

The vast applications of deep generative models are anchored in three core capabilities -- generating new instances, reconstructing inputs, and learning compact representations -- across various data types, such as discrete text/protein sequences and continuous images. Existing model families, like Variational Autoencoders (VAEs), Generative Adversarial Networks (GANs), autoregressive models, and diffusion models, generally excel in specific capabilities and data types but fall short in others. We introduce generalized diffusion with learnable encoder-decoder (DiLED), that seamlessly integrates the core capabilities for broad applicability and enhanced performance. DiLED generalizes the Gaussian noising-denoising in standard diffusion by introducing parameterized encoding-decoding. Crucially, DiLED is compatible with the well-established diffusion model objective and training recipes, allowing effective learning of the encoder-decoder parameters jointly with diffusion. By choosing appropriate encoder/decoder (e.g., large language models), DiLED naturally applies to different data types. Extensive experiments on text, proteins, and images demonstrate DiLED's flexibility to handle diverse data and tasks and its strong improvement over various existing models.

Authors

Guangyi Liu,Yu Wang,Zeyu Feng,Qiyu Wu,Liping Tang,Yuan Gao,Zhen Li,Shuguang Cui,Julian McAuley,Eric P Xing,Zichao Yang,Zhiting Hu

Journal

arXiv preprint arXiv:2402.19009

Published Date

2024/2/29

Professor FAQs

What is Eric Xing's h-index at Carnegie Mellon University?

The h-index of Eric Xing has been 87 since 2020 and 114 in total.

What are Eric Xing's research interests?

The research interests of Eric Xing are: Machine Learning, ML Systems, Optimization, Statistics, Network Analysis

What is Eric Xing's total number of citations?

Eric Xing has 57,613 citations in total.

What are the co-authors of Eric Xing?

The co-authors of Eric Xing are Michael I. Jordan, Li Fei-Fei, David Blei, Noah A. Smith, Jun Zhu, Edoardo M Airoldi.

Co-Authors

H-index: 203
Michael I. Jordan

Michael I. Jordan

University of California, Berkeley

H-index: 144
Li Fei-Fei

Li Fei-Fei

Stanford University

H-index: 106
David Blei

David Blei

Columbia University in the City of New York

H-index: 104
Noah A. Smith

Noah A. Smith

University of Washington

H-index: 75
Jun Zhu

Jun Zhu

Tsinghua University

H-index: 51
Edoardo M Airoldi

Edoardo M Airoldi

Harvard University

academic-engine

Useful Links