no code implementations • 15 Jul 2019 • Zhongkai Sun, Prathusha K Sarma, William Sethares, Erik P. Bucy
This paper learns multi-modal embeddings from text, audio, and video views/modes of data in order to improve upon down-stream sentiment classification.