S2IGAN: Speech-to-Image Generation via Adversarial Learning

Xinsheng Wang, Tingting Qiao, Jihua Zhu, Alan Hanjalic, Odette Scharenborg


An estimated half of the world’s languages do not have a written form, making it impossible for these languages to benefit from any existing text-based technologies. In this paper, a speech-to-image generation (S2IG) framework is proposed which translates speech descriptions to photo-realistic images without using any text information, thus allowing unwritten languages to potentially benefit from this technology. The proposed S2IG framework, named S2IGAN, consists of a speech embedding network (SEN) and a relation-supervised densely-stacked generative model (RDG). SEN learns the speech embedding with the supervision of the corresponding visual information. Conditioned on the speech embedding produced by SEN, the proposed RDG synthesizes images that are semantically consistent with the corresponding speech descriptions. Extensive experiments on datasets CUB and Oxford-102 demonstrate the effectiveness of the proposed S2IGAN on synthesizing high-quality and semantically-consistent images from the speech signal, yielding a good performance and a solid baseline for the S2IG task.


 DOI: 10.21437/Interspeech.2020-1759

Cite as: Wang, X., Qiao, T., Zhu, J., Hanjalic, A., Scharenborg, O. (2020) S2IGAN: Speech-to-Image Generation via Adversarial Learning. Proc. Interspeech 2020, 2292-2296, DOI: 10.21437/Interspeech.2020-1759.


@inproceedings{Wang2020,
  author={Xinsheng Wang and Tingting Qiao and Jihua Zhu and Alan Hanjalic and Odette Scharenborg},
  title={{S2IGAN: Speech-to-Image Generation via Adversarial Learning}},
  year=2020,
  booktitle={Proc. Interspeech 2020},
  pages={2292--2296},
  doi={10.21437/Interspeech.2020-1759},
  url={http://dx.doi.org/10.21437/Interspeech.2020-1759}
}