May 5, 2022, 1:12 a.m. | Rishikesh Magar, Yuyang Wang, Amir Barati Farimani

cs.LG updates on arXiv.org arxiv.org

Machine learning (ML) models have been widely successful in the prediction of
material properties. However, large labeled datasets required for training
accurate ML models are elusive and computationally expensive to generate.
Recent advances in Self-Supervised Learning (SSL) frameworks capable of
training ML models on unlabeled data have mitigated this problem and
demonstrated superior performance in computer vision and natural language
processing tasks. Drawing inspiration from the developments in SSL, we
introduce Crystal Twins (CT): an SSL method for crystalline materials …

arxiv learning prediction self-supervised learning supervised learning twins

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

Data Management Assistant

@ World Vision | Amman Office, Jordan

Cloud Data Engineer, Global Services Delivery, Google Cloud

@ Google | Buenos Aires, Argentina