Abstract
In this paper, we propose a methodology to maximize the benefits of interdisciplinary cooperation in AI research groups. Firstly, we build the case for the importance of interdisciplinarity in research groups as the best means to tackle the social implications brought about by AI systems, against the backdrop of the EU Commission proposal for an Artificial Intelligence Act. As we are an interdisciplinary group, we address the multi-faceted implications of the mass-scale diffusion of AI-driven technologies. The result of our exercise lead us to postulate the necessity of a behavioural theory that standardizes the interaction process of interdisciplinary groups. In light of this, we conduct a review of the existing approaches to interdisciplinary research on AI appliances, leading to the development of methodologies like ethics-by-design and value-sensitive design, evaluating their strengths and weaknesses. We then put forth an iterative process theory hinging on a narrative approach consisting of four phases: definition of the hypothesis space, building-up of a common lexicon, scenario-building, interdisciplinary self-assessment. Finally, we identify the most relevant fields of application for such a methodology and discuss possible case studies.