Search Results for author: Trung Dang

Found 8 papers, 4 papers with code

uaMix-MAE: Efficient Tuning of Pretrained Audio Transformers with Unsupervised Audio Mixtures

1 code implementation14 Mar 2024 Afrina Tabassum, Dung Tran, Trung Dang, Ismini Lourentzou, Kazuhito Koishida

Masked Autoencoders (MAEs) learn rich low-level representations from unlabeled data but require substantial labeled data to effectively adapt to downstream tasks.

Optimality in Mean Estimation: Beyond Worst-Case, Beyond Sub-Gaussian, and Beyond $1+α$ Moments

no code implementations21 Nov 2023 Trung Dang, Jasper C. H. Lee, Maoyuan Song, Paul Valiant

The state of the art results for mean estimation in $\mathbb{R}$ are 1) the optimal sub-Gaussian mean estimator by [LV22], with the tight sub-Gaussian constant for all distributions with finite but unknown variance, and 2) the analysis of the median-of-means algorithm by [BCL13] and a lower bound by [DLLO16], characterizing the big-O optimal errors for distributions for which only a $1+\alpha$ moment exists for $\alpha \in (0, 1)$.

A Multi-scale Graph Signature for Persistence Diagrams based on Return Probabilities of Random Walks

no code implementations28 Sep 2022 Chau Pham, Trung Dang, Peter Chin

Persistence diagrams (PDs), often characterized as sets of death and birth of homology class, have been known for providing a topological representation of a graph structure, which is often useful in machine learning tasks.

Graph Classification

A Study on Self-Supervised Object Detection Pretraining

no code implementations9 Jul 2022 Trung Dang, Simon Kornblith, Huy Thong Nguyen, Peter Chin, Maryam Khademi

In this work, we study different approaches to self-supervised pretraining of object detection models.

Object object-detection +2

Training Robust Zero-Shot Voice Conversion Models with Self-supervised Features

no code implementations8 Dec 2021 Trung Dang, Dung Tran, Peter Chin, Kazuhito Koishida

Unsupervised Zero-Shot Voice Conversion (VC) aims to modify the speaker characteristic of an utterance to match an unseen target speaker without relying on parallel training data.

Decoder Self-Supervised Learning +1

Revealing and Protecting Labels in Distributed Training

1 code implementation NeurIPS 2021 Trung Dang, Om Thakkar, Swaroop Ramaswamy, Rajiv Mathews, Peter Chin, Françoise Beaufays

Prior works have demonstrated that labels can be revealed analytically from the last layer of certain models (e. g., ResNet), or they can be reconstructed jointly with model inputs by using Gradients Matching [Zhu et al'19] with additional knowledge about the current state of the model.

Automatic Speech Recognition Automatic Speech Recognition (ASR) +4

Cannot find the paper you are looking for? You can Submit a new open access paper.