Eduard Hovy

Eduard Hovy

Carnegie Mellon University

H-index: 104

North America-United States

Description

Eduard Hovy, With an exceptional h-index of 104 and a recent h-index of 64 (since 2020), a distinguished researcher at Carnegie Mellon University, specializes in the field of NLP, AI.

Professor Information

University

Carnegie Mellon University

Position

___

Citations(all)

59367

Citations(since 2020)

29983

Cited By

47210

hIndex(all)

104

hIndex(since 2020)

64

i10Index(all)

383

i10Index(since 2020)

225

Email

University Profile Page

Carnegie Mellon University

Research & Interests List

NLP

AI

Top articles of Eduard Hovy

A survey of data augmentation approaches for NLP

Data augmentation has recently seen increased interest in NLP due to more work in low-resource domains, new tasks, and the popularity of large-scale neural networks that require large amounts of training data. Despite this recent upsurge, this area is still relatively underexplored, perhaps due to the challenges posed by the discrete nature of language data. In this paper, we present a comprehensive and unifying survey of data augmentation for NLP by summarizing the literature in a structured manner. We first introduce and motivate data augmentation for NLP, and then discuss major methodologically representative approaches. Next, we highlight techniques that are used for popular NLP applications and tasks. We conclude by outlining current challenges and directions for future research. Overall, our paper aims to clarify the landscape of existing literature in data augmentation for NLP and motivate additional work in this area. We also present a GitHub repository with a paper list that will be continuously updated at https://github.com/styfeng/DataAug4NLP

Authors

Steven Y Feng,Varun Gangal,Jason Wei,Sarath Chandar,Soroush Vosoughi,Teruko Mitamura,Eduard Hovy

Journal

arXiv preprint arXiv:2105.03075

Published Date

2021/5/7

Self-training with noisy student improves imagenet classification

We present a simple self-training method that achieves 88.4% top-1 accuracy on ImageNet, which is 2.0% better than the state-of-the-art model that requires 3.5 B weakly labeled Instagram images. On robustness test sets, it improves ImageNet-A top-1 accuracy from 61.0% to 83.7%, reduces ImageNet-C mean corruption error from 45.7 to 28.3, and reduces ImageNet-P mean flip rate from 27.8 to 12.2. To achieve this result, we first train an EfficientNet model on labeled ImageNet images and use it as a teacher to generate pseudo labels on 300M unlabeled images. We then train a larger EfficientNet as a student model on the combination of labeled and pseudo labeled images. We iterate this process by putting back the student as the teacher. During the generation of the pseudo labels, the teacher is not noised so that the pseudo labels are as accurate as possible. However, during the learning of the student, we inject noise such as dropout, stochastic depth and data augmentation via RandAugment to the student so that the student generalizes better than the teacher.

Authors

Qizhe Xie,Minh-Thang Luong,Eduard Hovy,Quoc V Le

Published Date

2020

Unsupervised data augmentation for consistency training

Semi-supervised learning lately has shown much promise in improving deep learning models when labeled data is scarce. Common among recent approaches is the use of consistency training on a large amount of unlabeled data to constrain model predictions to be invariant to input noise. In this work, we present a new perspective on how to effectively noise unlabeled examples and argue that the quality of noising, specifically those produced by advanced data augmentation methods, plays a crucial role in semi-supervised learning. By substituting simple noising operations with advanced data augmentation methods such as RandAugment and back-translation, our method brings substantial improvements across six language and three vision tasks under the same consistency training framework. On the IMDb text classification dataset, with only 20 labeled examples, our method achieves an error rate of 4.20, outperforming the state-of-the-art model trained on 25,000 labeled examples. On a standard semi-supervised learning benchmark, CIFAR-10, our method outperforms all previous approaches and achieves an error rate of 5.43 with only 250 examples. Our method also combines well with transfer learning, eg, when finetuning from BERT, and yields improvements in high-data regime, such as ImageNet, whether when there is only 10% labeled data or when a full labeled set with 1.3 M extra unlabeled examples is used. Code is available at https://github. com/google-research/uda.

Authors

Qizhe Xie,Zihang Dai,Eduard Hovy,Thang Luong,Quoc Le

Journal

Advances in neural information processing systems

Published Date

2020

Professor FAQs

What is Eduard Hovy's h-index at Carnegie Mellon University?

The h-index of Eduard Hovy has been 64 since 2020 and 104 in total.

What are Eduard Hovy's top articles?

What are Eduard Hovy's research interests?

The research interests of Eduard Hovy are: NLP, AI

What is Eduard Hovy's total number of citations?

Eduard Hovy has 59,367 citations in total.

academic-engine

Useful Links