Web: https://www.reddit.com/r/MachineLearning/comments/se6y9e/d_best_way_and_hassle_free_approach_to_loading_t5/

Jan. 27, 2022, 7:50 p.m. | /u/rirhun

Machine Learning reddit.com

Does anybody know how to load this large model (https://github.com/google-research/text-to-text-transfer-transformer) on multiple GPUs. So load some layers on some GPUs and other layers on other GPUs on the same machine. This model needs about 100GB of memory and though I'm trying to use model.parallelize(), I still run into Cuda out of memory error.

Any assistance would be greatly appreciated!

submitted by /u/rirhun
[link] [comments]

gpus machinelearning model on

Clinical Assistant or Associate Professor of Management Science and Systems

@ University at Buffalo | Buffalo, NY

Data Analyst

@ Colorado Springs Police Department | Colorado Springs, CO

Predictive Ecology Postdoctoral Fellow

@ Lawrence Berkeley National Lab | Berkeley, CA

Data Analyst, Patagonia Action Works

@ Patagonia | Remote

Data & Insights Strategy & Innovation General Manager

@ Chevron Services Company, a division of Chevron U.S.A Inc. | Houston, TX

Faculty members in Research areas such as Bayesian and Spatial Statistics; Data Privacy and Security; AI/ML; NLP; Image and Video Data Analysis

@ Ahmedabad University | Ahmedabad, India