Concepedia

Concept

pre-trained models

Parents

1.8K

Publications

148.8K

Citations

7.1K

Authors

1.1K

Institutions

About

Pre-trained models is a methodological approach in machine learning involving the initial training of a model on a large, generic dataset to acquire foundational knowledge or representations. This concept investigates the efficacy of leveraging knowledge gained from this initial phase to improve performance on related, often more specialized, tasks through transfer learning. Key characteristics include the development of generalized feature extraction capabilities through training on extensive data, typically in an unsupervised or self-supervised manner. The significance of pre-trained models lies in their ability to substantially reduce the data and computational resources required for subsequent task-specific fine-tuning, leading to improved efficiency and state-of-the-art results across diverse applications.

Top Authors

Rankings shown are based on concept H-Index.

ZL

Tsinghua University

JG

Microsoft (United States)

ND

Microsoft Research Asia (China)

MS

Tsinghua University

XJ

Huawei Technologies (Sweden)

Top Institutions

Rankings shown are based on concept H-Index.

Google (United States)

Mountain View, United States

Cambridge, United Kingdom

Tsinghua University

Beijing, China

Pittsburgh, United States

Peking University

Beijing, China