Feb. 15, 2024, 5:42 a.m. | Namkyeong Cho, Junseung Ryu, Hyung Ju Hwang

cs.LG updates on arXiv.org arxiv.org

arXiv:2402.09084v1 Announce Type: new
Abstract: This study investigates the impact of Sobolev Training on operator learning frameworks for improving model performance. Our research reveals that integrating derivative information into the loss function enhances the training process, and we propose a novel framework to approximate derivatives on irregular meshes in operator learning. Our findings are supported by both experimental evidence and theoretical analysis. This demonstrates the effectiveness of Sobolev Training in approximating the solution operators between infinite-dimensional spaces.

abstract arxiv cs.ai cs.lg derivatives framework frameworks function impact information loss meshes novel performance process research study training type

Doctoral Researcher (m/f/div) in Automated Processing of Bioimages

@ Leibniz Institute for Natural Product Research and Infection Biology (Leibniz-HKI) | Jena

Seeking Developers and Engineers for AI T-Shirt Generator Project

@ Chevon Hicks | Remote

Software Engineer for AI Training Data (School Specific)

@ G2i Inc | Remote

Software Engineer for AI Training Data (Python)

@ G2i Inc | Remote

Software Engineer for AI Training Data (Tier 2)

@ G2i Inc | Remote

Senior DevOps Engineer- Autonomous Database

@ Oracle | Reston, VA, United States