March 14, 2022, 1:11 a.m. | Jun Qi, Chao-Han Huck Yang, Pin-Yu Chen, Javier Tejedor

cs.LG updates on arXiv.org arxiv.org

This work focuses on designing low complexity hybrid tensor networks by
considering trade-offs between the model complexity and practical performance.
Firstly, we exploit a low-rank tensor-train deep neural network (TT-DNN) to
build an end-to-end deep learning pipeline, namely LR-TT-DNN. Secondly, a
hybrid model combining LR-TT-DNN with a convolutional neural network (CNN),
which is denoted as CNN+(LR-TT-DNN), is set up to boost the performance.
Instead of randomly assigning large TT-ranks for TT-DNN, we leverage Riemannian
gradient descent to determine a TT-DNN …

arxiv gradient networks neural networks processing speech speech processing

Lead Developer (AI)

@ Cere Network | San Francisco, US

Research Engineer

@ Allora Labs | Remote

Ecosystem Manager

@ Allora Labs | Remote

Founding AI Engineer, Agents

@ Occam AI | New York

AI Engineer Intern, Agents

@ Occam AI | US

AI Research Scientist

@ Vara | Berlin, Germany and Remote