This is a list of awesome Foundation Model for X and X for Foundation Model related projects & papers.
- ImageBind: One embedding space to bind them all by Girdhar, Rohit, et al., CVPR 2023
- LM-Nav: Robotic Navigation with Large Pre-Trained Models of Language, Vision, and Action by Shah, Dhruv, Błażej Osiński, and Sergey Levine., PMLR 2023
- IoT in the Era of Generative AI: Vision and Challenges by Wang, Xin, et al., arxiv 2024
- TypeFly: Flying Drones with Large Language Model by Chen, Guojun, Xiaojing Yu, and Lin Zhong., arxiv 2023
- FlexGen: High-throughput Generative Inference of Large Language Models with a Single GPU by Sheng, Ying, et al., ICML 2023
- Tabi: An Efficient Multi-Level Inference System for Large Language Models by Wang, Yiding, et al. EuroSys 2023
- EFFICIENTLY SCALING TRANSFORMER INFERENCE by Pope, Reiner, et al., arxiv 2022
- SpecInfer: Accelerating Generative LLM Serving with Speculative Inference and Token Tree Verification by Miao, Xupeng, et al., arxiv 2023
- EnergonAI: An Inference System for 10-100 Billion Parameter Transformer Models by Du, Jiangsu, et al., arxiv 2022
- AlpaServe: Statistical Multiplexing with Model Parallelism for Deep Learning Serving by Li, Zhuohan, et al., OSDI 2023
- STI: Turbocharge NLP Inference at the Edge via Elastic Pipelining by Guo, Liwei, Wonkyo Choe, and Felix Xiaozhu Lin., ASPLOS 2023
- DeepSpeed-inference: enabling efficient inference of transformer models at unprecedented scale by Aminabadi, Reza Yazdani, et al., SC 2022
- PETALS: Collaborative Inference and Fine-tuning of Large Models by Borzunov, Alexander, et al., arxiv 2022
- Fairness in Serving Large Language Models by Sheng, Ying, et al., arxiv 2023
- Fast Distributed Inference Serving for Large Language Models by Wu, Bingyang, et al., arxiv 2023
- DISTRIBUTED INFERENCE AND FINE-TUNING OF LARGE LANGUAGE MODELS OVER THE INTERNET under review
- Orca: A Distributed Serving System for {Transformer-Based} Generative Models by Yu, Gyeong-In, et al., OSDI 2022
- TopoopT: Co-optimizing Network Topology and Parallelization Strategy for Distributed Training Jobs by Wang, Weiyang, et al., NSDI 2023
- Breadth-First Pipeline Parallelism by Lamy-Poirier, Joel., MLSys 2023
- On Optimizing the Communication of Model Parallelism by Zhuang, Yonghao, et al., MLSys 2023
- Galvatron: Efficient Transformer Training over Multiple GPUs Using Automatic Parallelism by Miao, Xupeng, et al., arxiv 2022
- Overlap Communication with Dependent Computation via Decomposition in Large Deep Learning Models by Wang, Shibo, et al., ASPLOS 2023
- Vcc: Scaling Transformers to 128K Tokens or More by Prioritizing Important Tokens by Zeng, Zhanpeng, et al., arxiv 2023
- Multimodal LLMs for health grounded in individual-specific data by Belyaeva, Anastasiya, et al., arxiv 2023
- Path to Medical AGI: Unify Domain-specific Medical LLMs with the Lowest Cost by Zhou, Juexiao, Xiuying Chen, and Xin Gao., arxiv 2023
- Decoding speech perception from non-invasive brain recordings by Défossez, Alexandre, et al., Nature Machine Intelligence 2023
- Large language models improve Alzheimer’s disease diagnosis using multi-modality data by Feng, Yingjie, et al., arxiv 2023
- Neuro-GPT: Developing A Foundation Model for EEG by Cui, Wenhui, et al., arxiv 2023
- From Classification to Clinical Insights: Towards Analyzing and Reasoning About Mobile and Behavioral Health Data With Large Language Models by Englhardt, Zachary, et al., arxiv 2023
- Conversational Health Agents: A Personalized LLM-Powered Agent Framework by Abbasian, Mahyar, et al., arxiv 2023
- UbiPhysio: Support Daily Functioning, Fitness, and Rehabilitation with Action Understanding and Feedback in Natural Language by Wang, Chongyang, et al., arxiv 2023
- GG-LLM: Geometrically Grounding Large Language Models for Zero-shot Human Activity Forecasting in Human-Aware Task Planning by Graule, Moritz A., and Volkan Isler., arxiv 2023
- Can Large Language Models Reason about Program Invariants? by Pei, Kexin, et al., ICML 2023
- The Hitchhiker's Guide to Program Analysis: A Journey with Large Language Models by Li, Haonan, et al., arxiv 2023
- Clover: Closed-Loop Verifiable Code Generation by Sun, Chuyue, et al., arxiv 2023
- Formalizing Natural Language Intent into Program Specifications via Large Language Models by Endres, Madeline, et al., arxiv 2023
- Ranking LLM-Generated Loop Invariants for Program Verification by Chakraborty, Saikat, et al., arxiv 2023
- Large language models for compiler optimization by Cummins, Chris, et al., arxiv 2023
- Magicoder: Source Code Is All You Need by Yuxiang Wei1 Zhe Wang2 Jiawei Liu1 Yifeng Ding1 Lingming Zhang, arxiv 2023
- Towards Free Data Selection with General-Purpose Models by Xie, Yichen, et al., arxiv 2023
- Prompt-aligned Gradient for Prompt Tuning by Zhu, Beier, et al., CVPR 2023
- MaPLe: Multi-modal Prompt Learning by Khattak, Muhammad Uzair, et al., CVPR 2023
- ClimaX: A foundation model for weather and climate by Nguyen, Tung, et al., arxiv 2023