Published November 3, 2025 | Version v1
Conference paper Open

Pairing Disparate Datasets for Generative Outputs Across Music and Architecture

Description

Most multimodal models utilise semantically consistent data samples for model training and generative outputs. However, for design and composition work, semantic correlations are not the only means for cross-modal translations as suggested by synaesthetic phenomena. In this paper, we propose several methods to pair disparate datasets for training a model to translate between musical compositions and architectural design. We adopt and augment our own datasets to extract high-level features for both music and architectural samples. We utilise self-organising maps and agglomerative clustering strategies to pair data which is then used to train a conditional GAN. The techniques are evaluated through qualitative and quantitative assessment and are also applied in generating architectural compositions. Our experiments suggest promising initial results for feature pairing using hierarchical agglomerative clustering when generating images. We conclude by discussing the meaning and utilization of our proposed techniques as a means to systematically design with abstract features across modalities.

Files

CMMR2025_P2_4.pdf

Files (17.0 MB)

Name Size Download all
md5:03aebf2c2731d81ca96cdc492ff25593
17.0 MB Preview Download