Search Results for author: Alexia Jolicoeur-Martineau

Found 16 papers, 12 papers with code

On Relativistic f-Divergences

1 code implementation ICML 2020 Alexia Jolicoeur-Martineau

We take a more rigorous look at Relativistic Generative Adversarial Networks (RGANs) and prove that the objective function of the discriminator is a statistical divergence for any concave function $f$ with minimal properties ($f(0)=0$, $f'(0) \neq 0$, $\sup_x f(x)>0$).

LoGAH: Predicting 774-Million-Parameter Transformers using Graph HyperNetworks with 1/100 Parameters

1 code implementation25 May 2024 Xinyu Zhou, Boris Knyazev, Alexia Jolicoeur-Martineau, Jie Fu

Unfortunately, predicting parameters of very wide networks relies on copying small chunks of parameters multiple times and requires an extremely large number of parameters to support full prediction, which greatly hinders its adoption in practice.

Generating and Imputing Tabular Data via Diffusion and Flow-based Gradient-Boosted Trees

2 code implementations18 Sep 2023 Alexia Jolicoeur-Martineau, Kilian Fatras, Tal Kachman

Through empirical evaluation across the benchmark, we demonstrate that our approach outperforms deep-learning generation methods in data generation tasks and remains competitive in data imputation.

Imputation

Diffusion models with location-scale noise

no code implementations12 Apr 2023 Alexia Jolicoeur-Martineau, Kilian Fatras, Ke Li, Tal Kachman

Diffusion Models (DMs) are powerful generative models that add Gaussian noise to the data and learn to remove it.

PopulAtion Parameter Averaging (PAPA)

1 code implementation6 Apr 2023 Alexia Jolicoeur-Martineau, Emy Gervais, Kilian Fatras, Yan Zhang, Simon Lacoste-Julien

Based on this idea, we propose PopulAtion Parameter Averaging (PAPA): a method that combines the generality of ensembling with the efficiency of weight averaging.

Gotta Go Fast with Score-Based Generative Models

no code implementations NeurIPS Workshop DLDE 2021 Alexia Jolicoeur-Martineau, Ke Li, Rémi Piché-Taillefer, Tal Kachman, Ioannis Mitliagkas

Score-based (denoising diffusion) generative models have recently gained a lot of success in generating realistic and diverse data.

Denoising

Gotta Go Fast When Generating Data with Score-Based Models

1 code implementation28 May 2021 Alexia Jolicoeur-Martineau, Ke Li, Rémi Piché-Taillefer, Tal Kachman, Ioannis Mitliagkas

For high-resolution images, our method leads to significantly higher quality samples than all other methods tested.

Ranked #8 on Image Generation on CIFAR-10 (Inception score metric)

Image Generation

Gradient penalty from a maximum margin perspective

2 code implementations15 Oct 2019 Alexia Jolicoeur-Martineau, Ioannis Mitliagkas

We present a unifying framework of expected margin maximization and show that a wide range of gradient-penalized GANs (e. g., Wasserstein, Standard, Least-Squares, and Hinge GANs) can be derived from this framework.

Image Generation

On Relativistic $f$-Divergences

1 code implementation8 Jan 2019 Alexia Jolicoeur-Martineau

Given the good performance of RGANs, this suggests that WGAN does not performs well primarily because of the weak metric, but rather because of regularization and the use of a relativistic discriminator.

GANs beyond divergence minimization

1 code implementation6 Sep 2018 Alexia Jolicoeur-Martineau

We observe that most loss functions converge well and provide comparable data generation quality to non-saturating GAN, LSGAN, and WGAN-GP generator loss functions, whether we use divergences or non-divergences.

The relativistic discriminator: a key element missing from standard GAN

10 code implementations ICLR 2019 Alexia Jolicoeur-Martineau

We show that this property can be induced by using a relativistic discriminator which estimate the probability that the given real data is more realistic than a randomly sampled fake data.

Generative Adversarial Network Image Generation

Cannot find the paper you are looking for? You can Submit a new open access paper.