Home

Awesome

Awesome-Foundation-Models

Awesome

A foundation model is a large-scale pretrained model (e.g., BERT, DALL-E, GPT-3) that can be adapted to a wide range of downstream applications. This term was first popularized by the Stanford Institute for Human-Centered Artificial Intelligence. This repository maintains a curated list of foundation models for vision and language tasks. Research papers without code are not included.

Survey

2024

Before 2024

Papers by Date

2024

2023

2022

<!-- * [Unified Vision and Language Prompt Learning](https://arxiv.org/pdf/2210.07225.pdf) (No code yet; from Nanyang Technological University and Apple) -->

2021

Before 2021

Papers by Topic

Large Language/Multimodal Models

Linear Attention

Large Benchmarks

Vision-Language Pretraining

Perception Tasks: Detection, Segmentation, and Pose Estimation

Training Efficiency

Towards Artificial General Intelligence (AGI)

AI Safety and Responsibility

Related Awesome Repositories