Sept. 2, 2022, 1:15 a.m. | Yan Xia, Zhou Zhao, Shangwei Ye, Yang Zhao, Haoyuan Li, Yi Ren

cs.CL updates on arXiv.org arxiv.org

In this paper, we introduce a new task, spoken video grounding (SVG), which
aims to localize the desired video fragments from spoken language descriptions.
Compared with using text, employing audio requires the model to directly
exploit the useful phonemes and syllables related to the video from raw speech.
Moreover, we randomly add environmental noises to this speech audio, further
increasing the difficulty of this task and better simulating real applications.
To rectify the discriminative phonemes and extract video-related information
from …

arxiv curriculum curriculum learning learning video

Senior Machine Learning Engineer

@ GPTZero | Toronto, Canada

ML/AI Engineer / NLP Expert - Custom LLM Development (x/f/m)

@ HelloBetter | Remote

Doctoral Researcher (m/f/div) in Automated Processing of Bioimages

@ Leibniz Institute for Natural Product Research and Infection Biology (Leibniz-HKI) | Jena

Coding Data Quality Auditor

@ Neuberger Berman | Work At Home-Georgia

Post Graduate (Year-Round) Intern - Market Research Analyst and Agreement Support

@ National Renewable Energy Laboratory | CO - Golden

Retail Analytics Engineering - Sr. Manager (Data)

@ Axalta | Woonsocket-1 CVS Drive