April 11, 2024, 4:42 a.m. | Taegyun Kwon, Dasaem Jeong, Juhan Nam

cs.LG updates on arXiv.org arxiv.org

arXiv:2404.06818v1 Announce Type: cross
Abstract: In recent years, advancements in neural network designs and the availability of large-scale labeled datasets have led to significant improvements in the accuracy of piano transcription models. However, most previous work focused on high-performance offline transcription, neglecting deliberate consideration of model size. The goal of this work is to implement real-time inference for piano transcription while ensuring both high performance and lightweight. To this end, we propose novel architectures for convolutional recurrent neural networks, redesigning …

abstract accuracy arxiv autoregressive autoregressive models availability cs.lg cs.sd datasets designs eess.as however improvements network neural network offline performance real-time scale transcription type work

Founding AI Engineer, Agents

@ Occam AI | New York

AI Engineer Intern, Agents

@ Occam AI | US

AI Research Scientist

@ Vara | Berlin, Germany and Remote

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne