Web: https://www.reddit.com/r/computervision/comments/ra1tyg/how_do_you_reduce_dimensionality_of_video_after/

Dec. 6, 2021, 7:45 a.m. | /u/temp12345124124

Computer Vision reddit.com

I'm building a video classifier for a school project (action classification on the UCF101 dataset). The constraint is that I'm not supposed to use CNNs - instead I'm to try extracting features and train with an MLP.

I applied the Canny edge detector to a subsample of video frames, which gives me some sparse images. But my question is how can I take advantage of this sparsity when concatenating everything together to feed into the MLP? Is there a way to reduce the dimensionality while maintaining information?

submitted by /u/temp12345124124 …

computervision video

Statistics and Computer Science Specialist

@ Hawk-Research | Remote

Data Scientist, Credit/Fraud Strategy

@ Fora Financial | New York City

Postdoctoral Research Associate - Biomedical Natural Language Processing and Deep Learning

@ Oak Ridge National Laboratory - Oak Ridge, TN | Oak Ridge, TN, United States

Senior Machine Learning / Computer Vision Engineer

@ Glass Imaging | Los Altos, CA

Research Scientist in Biomedical Natural Language Processing and Deep Learning

@ Oak Ridge National Laboratory | Oak Ridge, TN

W3-Professorship for Intelligent Energy Management

@ Universität Bayreuth | Bayreuth, Germany