More data is almost always beneficial for analysis and machine learning
tasks. In many realistic situations however, an enterprise cannot share its
data, either to keep a competitive advantage or to protect the privacy of the
data sources, the enterprise's clients for example. We propose a method for
data owners to share synthetic or fake versions of their data without sharing
the actual data, nor the parameters of models that have direct access to the
data. The method proposed is based on the privGAN architecture where local GANs
are trained on their respective data subsets with an extra penalty from a
central discriminator aiming to discriminate the origin of a given fake sample.
We demonstrate that this approach, when applied to subsets of various sizes,
leads to better utility for the owners than the utility from their real small
datasets. The only shared pieces of information are the parameter updates of
the central discriminator. The privacy is demonstrated with white-box attacks
on the most vulnerable elments of the architecture and the results are close to
random guessing. This method would apply naturally in a federated learning
setting.