March 1, 2024, 5:47 a.m. | Kate Sanders, Nathaniel Weir, Benjamin Van Durme

cs.CV updates on arXiv.org arxiv.org

arXiv:2402.19467v1 Announce Type: cross
Abstract: It is challenging to perform question-answering over complex, multimodal content such as television clips. This is in part because current video-language models rely on single-modality reasoning, have lowered performance on long inputs, and lack interpetability. We propose TV-TREES, the first multimodal entailment tree generator. TV-TREES serves as an approach to video understanding that promotes interpretable joint-modality reasoning by producing trees of entailment relationships between simple premises directly entailed by the videos and higher-level conclusions. We …

abstract arxiv cs.ai cs.cl cs.cv current generator inputs language language models multimodal multimodal content neuro part performance question reasoning television tree trees type video

Lead Developer (AI)

@ Cere Network | San Francisco, US

Research Engineer

@ Allora Labs | Remote

Ecosystem Manager

@ Allora Labs | Remote

Founding AI Engineer, Agents

@ Occam AI | New York

AI Engineer Intern, Agents

@ Occam AI | US

AI Research Scientist

@ Vara | Berlin, Germany and Remote