Current physics-based synthesis techniques tend to synthesize the interaction between different functional elements of a sound generator by treating it as a single system. However, when dealing with the physical modeling of complex sound generators this choice raises questions about the resulting flexibility of the adopted synthesis strategy. One way to overcome this problem is to approach it by individually synthesizing and discretizing the objects that contribute to the generation of sounds. In this paper we address the problem of how to automatize the process of physically modeling the interaction between objects, and how to make it dynamical. We will show that this can be done through the automatic definition and implementation of a topology model that adapts to the contact and proximity conditions between the considered objects.