1 Introduction

Model transformations are at the heart of model-driven engineering (MDE) (Sendall and Kozaczynski 2003; Metzger 2005). They provide a way to consistently and automatically derive a multitude of artefacts such as source code, simulation inputs or different views from system models (Schmidt 2006). Model transformations also allow to analyse system aspects on the basis of models (Schmidt 2006) and can provide interoperability between different modelling languages, e.g. architecture description languages like those described by Malavolta et al. (2010). Since the emergence of the MDE paradigm at the beginning of the century numerous dedicated model transformation languages (MTLs) have been developed to support engineers in their endeavours (Jouault et al. 2006; Arendt et al. 2010; OMG 2016b). Their appeal is driven by the promise of many advantages such as increased productivity, comprehensibility and domain specificity associated with using domain specific languages (Hermans et al. 2009; Johannes et al. 2009).

A recent literature study of us revealed, that, while a large number of such advantages and also disadvantages are claimed in literature, there exist only a few studies investigating to what extend these claims actually hold (Götz et al. 2021). The study presents 15 properties of MTLs for which literature claims advantages or disadvantages. In this context, a claimed positive effect on one of the properties means an advantage whereas a negative influence means a disadvantage. The properties identified in the study are: Analysability, Comprehensibility, Conciseness, Debugging, Ease of Writing (a transformation), Expressiveness, Extendability, (being) Just Better, Learnability, Performance, Productivity, Reuse & Maintainability, Tool Support, Semantics & Verification and Versatility.

Our study also revealed, that most claims in literature are made broadly and without much explanation as to where the claim originates from (Götz et al. 2021). Claims such as “Model transformation languages make it easy to work with models.” (Liepiņš 2012), “Declarative MTLs increase programmer productivity” (Lawley and Raymond 2007) or “Model transformation languages are more concise” (Hinkel and Goldschmidt 2019) illustrate this. We assume that authors make such claims while having certain context and background in mind, but choose to omit it for unspecified reasons. Some likely reason for omission of the context are, that authors believe it to not be worth mentioning or to preserve space which is often sparse in publications.

Regardless of the concrete reasons, a result of this practice is a lack of cause and effect relations in the context of model transformation languages that explain both why and when certain advantages or disadvantages hold. Claims are thus easily dismissed based on anecdotal evidence. Furthermore, setting up proper evaluation is also difficult because the claims do not provide the necessary background to do so.

To close this gap, we executed a large-scale empirical study using semi-structured interviews. It involved a total of 56 researchers and practitioners in the field of model transformations. The goal of our study was to compile a comprehensive list of influences on properties of model transformation languages guided by the following research questions:

RQ 1: What are the factors that influence properties of model transformation languages?

RQ 2: How do the identified factors influence MTL properties?

To concentrate our efforts and best utilize all available resources, we decided to focus on 6 of the 15 properties of model transformation languages identified by us in the preceding SLR (Götz et al. 2021). The 6 properties investigated in this study are: Comprehensibility, Ease of Writing, (practical) Expressiveness, Productivity, Reuse and Maintainability and Tool support. We have chosen these six because they all play a major role in providing reasons for the adoption of model transformation languages.

Interviewees were presented with a number of claims about MTLs from literature and asked to reveal their views on the matter, as well as assumptions and reasons that lead them to agree or disagree with the presented claims. We qualitatively analysed the interviews to understand the participants perceived influence factors and reasons for the advantages or disadvantages stated in the claims. The extracted data was then analysed to find commonalities between interviewees. This was done for single claims as well as for overarching factors and reasons that influence a variety of aspects of MTLs.

We present a comprehensive explanation of factors that, according to experts, play an essential role in the discussion of advantages and disadvantages of model transformation languages for the investigated properties. This is accompanied by a detailed exposition of how factors are relevant for the properties given above. Lastly, we discuss the most salient factors and argue actionable results for the community and further research.

As the first study of this type, we make the following contributions:

  1. 1.

    A comprehensive categorisation and listing of factors resulting in advantages or disadvantages of MTLs in the 6 properties studied.

  2. 2.

    A detailed description of why and how each identified factor exerts an influence on different properties.

  3. 3.

    Suggestions for how the presented information can be utilised to empirically investigate MTL properties.

  4. 4.

    Procedural proposals for improving current model transformation languages based on the presented data.

The results of our study show, that there is a large number of factors that influence properties of model transformation languages. There is also a number of factors on which this influence depends on, i.e. factors that have a moderation effect on the influence of other factors. These factors provide a solid basis that allows further studies to be conducted with more focus. They also enable precise decisions on where improvements and adjustments in or for model transformation languages can be made.

The remainder of this paper is structured as follows: Section 2 introduces model-driven engineering and model transformation languages, the context in which our study integrates. Section 3 will detail our methodology for preparing and conducting the interviews and the procedures used to analyse the data accumulated through the interviews. Afterwards Section 4 gives an overview over demographic data of our interview participants while 5 presents our code system and details the findings for each code based on the interviews and analysis thereof. In Section 6 we present overarching findings and in Section 7, we discuss actionable results that can be drawn from our study that indicate avenues to focus on for the research community. Section 8 contains a detailed discussion of the validity threats of this research, and in Section 9 related efforts are presented. Lastly, Section 10 draws a conclusion for our research and proposes future work.

2 Background

This section will provide the necessary background for the context in which our study is integrated in.

2.1 Model-Driven Engineering

The Model-Driven Architecture (MDA) paradigm was first introduced by the Object Management Group in 2001 (OMG 2001). It forms the basis for an approach commonly referred to as Model-driven development (MDD) (Brown et al. 2005), introduced as means to cope with the ever growing complexity associated with software development. At the core of it lies the notion of using models as the central artefact for development. In essence this means, that models are used both to describe and reason about the problem domain as well as to develop solutions (Brown et al. 2005). An advantage ascribed to this approach that arises from the use of models in this way, is that they can be expressed with concepts closer to the related domain than when using regular programming languages (Selic 2003).

When fully utilized, MDD envisions automatic generation of executable solutions specialized from abstract models (Selic 2003; Schmidt 2006). To be able to achieve this, the structure of models needs to be known. This is achieved through so called meta-models which define the structure of models. The structure of meta-models themselves is then defined through meta-models of their own. For this setup, the OMG developed a modelling standard called Meta-object Facility (MOF) (OMG 2016a) on the basis of which a number of modelling frameworks such as the Eclipse Modelling Framework (EMF) (Steinberg et al. 2008) and the .NET Modelling Framework (Hinkel 2016) have been developed.

2.2 Domain-Specific Languages

Domain-specific languages (DSLs) are languages designed with a notation that is tailored for a specific domain by focusing on relevant features of the domain (Van Deursen and Klint 2002). In doing so DSLs aim to provide domain specific language constructs, that let developers feel like working directly with domain concepts thus increasing speed and ease of development (Sprinkle et al. 2009). Because of these potential advantages, a well defined DSL can provide a promising alternative to using general purpose tools for solving problems in a specific domain. Examples of this include languages such as shell scripts in Unix operating systems (Kernighan and Pike 1984), HTML (Raggett et al. 1999) for designing web pages or AADL an architecture design language (SAEMobilus 2004).

2.3 Model Transformation Languages

The process of (automatically) transforming one model into another model of the same or different meta-model is called model transformation (MT). They are regarded as being at the heart of Model Driven Software Development (Sendall and Kozaczynski 2003; Metzger 2005), thus making the process of developing them an integral part of MDD. Since the introduction of MDE at the beginning of the century, a plethora of domain specific languages for developing model transformations, so called model transformation languages (MTLs), have been developed (Arendt et al. 2010; Balogh and Varró 2006; Jouault et al. 2006; Kolovos et al. 2008; Horn 2013; George et al. 2012; Hinkel and Burger 2019). Model transformation languages are DSLs designed to support developers in writing model transformations. For this purpose, they provide explicit language constructs for tasks involved in model transformations such as model matching. There are various features, such as directionality or rule organization (Czarnecki and Helsen 2006), by which model transformation languages can be distinguished. For the purpose of this paper, we will only be explaining those features that are relevant to our study and discussion in Sections 2.3.1 to 2.3.7. Table 1 provides an overview over the presented features.

Table 1 MTL feature overview

Please refer to Czarnecki and Helsen (2006), Kahani et al. (2019), and Mens and Gorp (2006) for complete classifications.

2.3.1 External and Internal Transformation Languages

Domain specific languages, and MTLs by extension, can be distinguished on whether they are embedded into another language, the so called host language, or whether they are fully independent languages that come with their own compiler or virtual machine.

Languages embedded in a host language are called internal languages. Prominent representatives among model transformation languages are FunnyQT (Horn 2013) a language embedded in Clojure, NMF Synchronizations and the .NET transformation language (Hinkel and Burger 2019) embedded in C#, and RubyTL (Cuadrado et al. 2006) embedded in Ruby.

Fully independent languages are called external languages. Examples of external model transformation languages include one of the most widely known languages such as the Atlas transformation language (ATL) (Jouault et al. 2006), the graphical transformation language Henshin (Arendt et al. 2010) as well as a complete model transformation framework called VIATRA (Balogh and Varró 2006).

2.3.2 Transformation Rules

Czarnecki and Helsen (2006) describe rules as being “understood as a broad term that describes the smallest units of [a] transformation [definition]”. Examples for transformation rules are the rules that make up transformation modules in ATL, but also functions, methods or procedures that implement a transformation from input elements to output elements.

The fundamental difference between model transformation languages and general-purpose languages that originates in this definition, lies in dedicated constructs that represent rules. The difference between a transformation rule and any other function, method or procedure is not clear cut when looking at GPLs. It can only be made based on the contents thereof. An example of this can be seen in Listing 1, which contains exemplary Java methods. Without detailed inspection of the two methods it is not apparent which method does some form of transformation and which does not.

figure a

In a MTL on the other hand transformation rules tend to be dedicated constructs within the language that allow a definition of a mapping between input and output (elements). The example rules written in the model transformation language ATL in Listing 2 make this apparent. They define mappings between model elements of type Member and model elements of type Male as well as between Member and Female using rules, a dedicated language construct for defining transformation mappings. The transformation is a modified version of the well known Families2Persons transformation case (Anjorin et al. 2017).

figure b

2.3.3 Rule Application Control: Location Determination

Location determination describes the strategy that is applied for determining the elements within a model onto which a transformation rule should be applied (Czarnecki and Helsen 2006). Most model transformation languages such as ATL, Henshin, VIATRA or QVT (OMG 2016b), rely on some form of automatic traversal strategy to determine where to apply rules.

We differentiate two forms of location determination, based on the kind of matching that takes place during traversal. There is the basic automatic traversal in languages such as ATL or QVT, where single elements are matched to which transformation rules are applied. The other form of location determination, used in languages like Henshin, is based on pattern matching, meaning a model- or graph-pattern is matched to which rules are applied. This does allow developers to define sub-graphs consisting of several model elements and references between them which are then manipulated by a rule.

The automatic traversal of ATL applied to the example from Listing 2 will result in the transformation engine automatically executing the Member2Male on all model elements of type Member where the function isFemale() returns false and the Member2Female on all other model elements of type Member.

The pattern matching of Henshin can be demonstrated using Fig. 1, a modified version of the transformation examples by Krause et al. (2014). It describes a transformation that creates a couple connection between two actors that play in two films together. When the transformation is executed the transformation engine will try and find instances of the defined graph pattern and apply the changes on the found matches.

Fig. 1
figure 1

Example Henshin transformation

This highlights the main difference between automatic traversal and pattern matching as the engine will search for a sub graph within the model instead of applying a rule to single elements within the model.

2.3.4 Directionality

The directionality of a model transformation describes whether it can be executed in one direction, called a unidirectional transformation or in multiple directions, called a multidirectional transformation (Czarnecki and Helsen 2006).

For the purpose of our study the distinction between unidirectional and bidirectional transformations is relevant. Some languages allow dedicated support for executing a transformation both ways based on only one transformation definition, while other require users to define transformation rules for both directions. General-purpose languages can not provide bidirectional support and also require both directions to be implemented explicitly.

The ATL transformation from Listing 2 defines a unidirectional transformation. Input and output are defined and the transformation can only be executed in that direction.

The QVT-R relation defined in Listing 3 is an example of a bidirectional transformation definition (For simplicity reasons the transformation omits the condition that males are only created from members that are not female). Instead of a declaration of input and output, it defines how two elements from different domains relate to one another. As a result given a Member element its corresponding Male elements can be inferred, and vice versa.

figure c

2.3.5 Incrementality

Incrementality of a transformation describes whether existing models can be updated based on changes in the source models without rerunning the complete transformation (Czarnecki and Helsen 2006). This feature is sometimes also called model synchronisation.

Providing incrementality for transformations requires active monitoring of input and/or output models as well as information which rules affect what parts of the models. When a change is detected the corresponding rules can then be executed. It can also require additional management tasks to be executed to keep models valid and consistent.

2.3.6 Tracing

According to Czarnecki and Helsen (2006) tracing “is concerned with the mechanisms for recording different aspects of transformation execution, such as creating and maintaining trace links between source and target model elements”.

Several model transformation languages, such as ATL and QVT have automated mechanisms for trace management. This means that traces are automatically created during runtime. Some of the trace information can be accessed through special syntax constructs while some of it is automatically resolved to provide seamless access to the target elements based on their sources.

An example of tracing in action can be seen in line 16 of Listing 2. Here the partner attribute of a Female element that is being created, is assigned to s.companion. The s.companion reference points towards a element of type Member within the input model. When creating a Female or Male element from a Member element, the ATL engine will resolve this reference into the corresponding element, that was created from the referred Member element via either the Member2Male or Member2Female rule. ATL achieves this by automatically tracing which target model elements are created from which source model elements.

2.3.7 Dedicated Model Navigation Syntax

Languages or syntax constructs for navigating models is not part of any feature classification for model transformation languages. However, it was often discussed in our interviews and thus requires an explanation as to what interviewees refer to.

Languages such as OCL (OMG 2014), which is used in transformation languages like ATL, provide dedicated syntax for querying and navigating models. As such they provide syntactical constructs that aid users in navigation tasks. Different model transformation languages provide different syntax for this purpose. The aim is to provide specific syntax so users do not have to manually implement queries using loops or other general purpose constructs. OCL provides a functional approach for accumulating and querying data based on collections while Henshin uses graph patterns for expressing the relationship of sought-after model elements.

3 Methodology

To collect data for our research question, we decided on using semi-structured interviews and a subsequent qualitative content analysis that follows the guidelines detailed by Kuckartz (2014). Semi-structured interviews were chosen as a data collection method because they present a versatile approach to eliciting information from experts. They provide a framework that allows to get insights into opinions, thoughts and knowledge of experts (Meyer and Booker 1990; Hove and Anda 2005; Kallio et al. 2016). The qualitative content analysis guidelines by Kuckartz (2014) were chosen because of their detailed descriptions for all steps of the analysis process. As such they provide a more detailed and modern framework compared to the procedures introduced by Mayring (1994), which have long been a gold standard in qualitative content analysis.

An overview over our complete study design can be found in Fig. 2. It shows the order of activities that were planned and executed as well as the artefacts produced and used throughout the study. Each activity is annotated with the section number in which we detail the activity. We split our approach into three main-phases: Preparation (detailed in Section 3.1), Operation (detailed in Section 3.2) and Coding & Analysis (detailed in Section 3.3).

Fig. 2
figure 2

Overview over the study design

For the preparation phase, we used a subset of the claimed properties of model transformation languages identified by us (Götz et al. 2021) to develop an interview guide. The guide focuses around asking participants whether they agree with a claim from one of the properties and then envisages the usage of why questions to gain a deeper understanding of their opinions on the matter. After identifying and contacting participants based on the publications considered during our previous literature review (Götz et al. 2021), we conducted 54 interviews with 55 interviewees (at the request of two participants, one interview was conducted with both of them together) and collected one additional written response. During the Coding & Analysis phase, we coded and analysed all 54 transcripts, as well as the written response, guided by the framework detailed by Kuckartz. In doing so, we focused first on factors and reasons for the individual properties and then on common factors and reasons between them.

The remainder of this section will describe in detail how each of the three phases of our study was conducted.

3.1 Interview Preparation

Our interview preparation phase consists of the creation of an interview guide plus selecting and contacting appropriate interview participants. We use the guidelines by Kallio et al. (2016) for the creation of our interview guide and expand the steps detailed there with steps for selecting and contacting participants. In addition, we use the guidance from Newcomer et al. (2015) to construct our study in the best possible way.

According to Kallio et al. (2016) the creation of an interview guide consists of five consecutive steps. First, researchers are urged to evaluate how appropriate semi-structured interviews are as a data collection method for the posed research questions. Then, existing knowledge about the topic should be retrieved by means of a literature review. Based on the knowledge from the knowledge retrieval phase, a preliminary interview guide can then be formulated and in another step be pilot tested. Lastly the complete interview guide can then be presented and used. As previously stated, we enhance these steps with two additional steps for selecting and contacting potential interview participants.

In the following, we detail how the presented steps were executed and the results thereof.

3.1.1 Identifying the Appropriateness of Semi-Structured Interviews

The goal of our study, outlined in our research questions, is to collect and analyse reasons and background information of why people believe claims about model transformation languages to be true. Data such as this is qualitative by nature and hence requires a research method capable of producing qualitative data. According to Hove and Anda (2005) and Meyer and Booker (1990) expert interviews are one of the most widely used research methodologies in the technical field for this purpose. They allow to ascertain qualitative data such as opinions and estimates. Interviews also enable qualitative interpretation of already available data (Meyer and Booker 1990) which perfectly aligns with our goal. Moreover the opportunity to ask further questions about specific statements made by the participants (Newcomer et al. 2015) fits the open ended nature of our research question. For these reasons, we believe semi-structured interviews to be a well suited to answer our research questions.

3.1.2 Retrieving Previous Knowledge

In our previous publication (Götz et al. 2021), we detailed the preparation, execution and results of an extensive structured literature review on the topic of claims about model transformation languages. The literature review resulted in a categorization of 127 claims into 15 different categories (i.e. properties of MTLs) namely Analysability, Comprehensibility, Conciseness, Debugging, Ease of Writing a transformation, Expressiveness, Extendability, Just better, Learnability, Performance, Productivity, Reuse & Maintainability, Tool Support, Semantics and Verification and lastly Versatility. These properties and the claims about them serve as the basis for the design of our interview study presented here.

3.1.3 Interview Guide

The interview guide involves presenting each interview participant with several claims on model transformation languages. We use claims from literature instead of formulating our own statements, to make them more accessible. This also prevents any bias from the authors to be introduced at this step. Participants are first asked to assess their agreement with a claim before transitioning into a discussion on what the reasons for their decision are based on an open-ended question. This style of using close-ended questions as a prerequisite for open-ended or probe questions has been suggested by multiple guides (Newcomer et al. 2015; Hove and Anda 2005).

We focus on a subset of six properties. This is due to the aim of keeping the length of interviews within an acceptable range for participants. According to Newcomer et al. (2015) semi-structured interviews should not exceed a maximum length of one hour. As a result, only a number of properties can be discussed per interview. In order to still talk with enough participants about each property, the number of properties examined must be reduced. The properties we discuss in the interviews and the reasons why they are relevant are as follows:

  • Comprehensibility: Is an important property when transformations are being developed as part of a team effort or evolve over time.

  • Ease of Writing: Is a decisive property that influences whether developers want to use a languages to write transformations in.

  • Expressiveness: Is one of the most cited properties in literature (Götz et al. 2021) and main selling point of domain specific languages in general.

  • Productivity: Is a property that is highly relevant for industrial adoption.

  • Reuse & Maintainability: Is another property that enables wider adoption of model transformation languages in project settings.

  • Tools: High-quality tools can provide huge improvements to the development.

The list consists of the 5 most claimed properties form the previous literature review (Götz et al. 2021) and is supplemented with Productivity, because we believe this attribute to be the most relevant for industry adoption.

To maximize the response rate of contacted persons, we aim for an interview length of 30 minutes. This decision is based on experiences from previous interview studies conducted at our research group (Groner et al. 2020; Juhnke et al. 2020) and fits within the maximum interview length suggested by Newcomer et al. (2015).

To best utilize the limited time per interview, the six properties are split into three sets of two properties each. In each interview one of the three sets is discussed.

For each property, one non-specific, one specific and one negative claim is used to structure all interviews involving this property around. A complete overview over all selected claims can be found in Table 2.

Table 2 Properties and Claims

We consider non-specific claims to be those that do not provide any rationale as to why the claimed property holds, e.g. “Model transformation languages ease the writing of model transformations.”. The non-specific claims chosen simply reflect the property itself. They serve the purpose of getting participants to state their assumptions and beliefs for the property without any influence exerted by the discussed claim.

We consider those claims as specific, that provide a rationale or reason for why the claimed property holds, e.g. “Model transformation languages, being DSLs, improve the productivity.”. And we consider negative claims to be those, that state a negative property of model transformation languages, e.g. “Model transformation languages lack sophisticated reuse mechanisms.”. Generally, we use claims where we believe the discussions about the reasons to provide useful insights.

There exist several reasons why we believe this setup of using the same three none-specific, specific and negative claims for each property to be appropriate. First, the non-specific claim allows participants to provide any and all factors or reasons that they believe influence a claimed property. The specific claim then allows us to introduce a reason, that participants might not have thought about. It also prompts a discussion about a particular reason or factor that is shared between all participants. This ensures at least one area for cross comparison between answers. The negative claim forces participants to also deliberate negative aspects, providing a counterbalance that counteracts bias. Furthermore, the non-specific claim provides an easy introduction into the discussion about a specific MTL property that can present the interviewer with an overview of the participants thoughts on the matter. It also allows participants to provide other influence factors not specifically covered through the discussed claims or even new factors and reasons not present in the collection of claims from our literature review (Götz et al. 2021).

The complete interview guide resulting from the aforementioned considerations can be seen in Fig. 3. After introductory pleasantries we start all interviews of with demographic questions. Although some sources discourage asking demographic questions early in the interview due to their sensitive nature (Newcomer et al. 2015), we use them to break the ice between the interviewer and interviewee because our demographic questions do not probe any sensitive information.

Fig. 3
figure 3

Interview guide

After this initial get-to-know each other phase, the interviewer then proceeds to explain the research intentions, goals and the procedure of the remaining interview. Depending on the property-set selected for the interview, participants are then presented with a claim about a property. They are asked to rate their agreement with the claim based on a 5-point likert scale (5: completely agree, 4: agree, 3: neither agree nor disagree, 2: disagree, 1: completely disagree). The likert scale is used to allow the interviewer to better assess the participants tendency compared to a simple yes or no question. This part of the interview is intended solely to get a first impression of the view of the participant and not for a quantitative analysis. It also creates a casual point of entry for the interviewee to think about the topic under consideration. We communicate this to all participants to reduce any pressure they might feel to answer the question correctly. Afterwards an open-ended question inquiring about the reasons for the interviewees assessment is asked.

Some terms used within the discussed claims have ambiguous definitions. We tried to ask participants to explain their understanding of such terms, to prevent errors in analysis due to interviewees having different interpretations thereof. This allows for better assessment during analysis. The terms we have deemed to be ambiguous are: ‘succinct syntax’, ‘mapping’, ‘specific skills’, ‘high-level abstractions’, ‘convenient facilities’, ‘sufficient tool support’, ‘powerful tool support’, ‘sophisticated reuse mechanisms’ and ‘expressiveness’. We provide a definition for the term expressiveness. This is, because we are only interested in a specific type of expressiveness, i.e. how concisely and readily developers can express something. We are not interested in expressiveness in a theoretical sense, i.e. the closeness to Turing completeness.

This process of presenting a claim, querying the participants agreement before discussing their reasons for the assessment is repeated for all 3 claims about both properties. After discussing all claims, it is explained to the participants that the formal part of the interview is finished and that they are allowed to make final remarks about all discussed topics or other properties they want to address. After this phase of the interview acknowledgements on the part of the interviewer are expressed before saying goodbye. The complete question catalogue for the interviews can be found in Appendix A.

The interview guide was tested in a pilot study by the main author with one co-author that was not involved in its creation. After pilot testing, we changed the question about agreement with a claim from a yes-no question to one that uses a likert scale. We also extended the question sets with non-specific claims that do not contain any reasoning. Before adding the non specific claim, discussions focused too much on the narrow view within the presented claims.

3.1.4 Selecting & Contacting Participants

The target population for our study consists of all users of model transformation languages. To select potential participants for our study we rely on data from our previous literature review (Götz et al. 2021). The literature review produced a list of publications that address the topic of model transformations and model transformation languages. Because search terms such as ‘model to text’ and the like were not used in the study, using this list limits our results to model to model transformation languages. We discuss this limitation more thoroughly in Section 8.2.

All authors of the resulting publications are deemed to be potential interview participants. We assume, that people using MTLs in industry do have some research background and thus have published work in the field. There is also no other systematic way to find industry users. We also assume that people who are still active in the field have published within the last 5 years. This limits outreach but makes the set of potential participants more manageable. For this reason, the list was shortened to publications more recent than 2015 before the authors of all publications was compiled. This resulted in a total of 645 potential participants.

After selection, the authors were contacted via mail. First, everyone was contacted once and then, after a week, everyone who had not responded by then was contacted again. The texts we use for both mails can be found in Appendix B. Ten potential participants, from the list of potential participants, were not contacted through this channel but via personalised emails, as they are personal contacts of the authors.

Within the contact mails, potential participants are asked to select a suitable date for the interview and fill out a data consent form allowing us to record and transcribe the interviews.

Overall of the 645 contacted authors, 55 agreed to participate in our interview study resulting in a response rate of 8.53%Footnote 1.

3.2 Interview Conduction and Transcription

All but one interview were conducted by the first author using the online conferencing tool WebEx and lasted between 20 and 80 minutes. Due to scheduling issues, one interview had to be conducted by the second author, who had a preparatory mock interview with the main interviewer. Additionally, at the request of two participants, one interview was conducted with both of them together. Since our main focus for all interviews is on discussions, we do not believe this to have any effect on its results. WebEx is the chosen conferencing tool, due to its availability to the authors and its integrated recording tool which is used to record all interviews. For data privacy reasons and for easier in-depth analysis later on, all recordings are transcribed by two authors. To increase the readability of heavily fragmented sentences they are shortened to only contain the actual content without interruptions. In case of audibility issues the transcribing authors consulted with each other to try and resolve the issue. Altogether the interviews produced just over 32 hours of audio and about 162.100 words of transcriptions.

Each day, the main author decided on which question sets to use for all participants that had agreed to partake in the interviews. The question sets had to be chosen daily, as many participants only responded to the invitation after interviews had already taken place.

The goal of the decision process was, to ensure an even spread of participants over the question sets based on relevant demographic backgrounds, namely research, industry, MTL developer and MTL user. We consider those relevant because each group has a different view point on model transformation languages and their usage for writing transformations. It is therefore important to have answers from each group for each set of questions, to reduce the risk of missing relevant opinions.

We were able to ensure that at least one representative for each demographic group provided answers for each question set. A complete uniform distribution was not possible due to overlaps in the demographic groups.

3.3 Coding & Analysis

Coding and analysing the interview transcripts is done in accordance with the guidelines for content structuring content analysis suggested by Kuckartz (2014). The guideline recommends a seven step process (depicted in Fig. 4) for coding and analysing qualitative data. All steps are carried out with tool support provided by the MAXQDAFootnote 2 software. In the following, we explain how each process step is conducted in detail. We will use the following statement as a running example to show how codes and sub-codes are assigned and how the coding of text segments evolved throughout the process: “Of course some MTLs use explicit traceability for instance. But even then you have a mechanism to access it. And if you have a MTL with implicit traceability where the trace links are created automatically then of course you gain a lot of expressivity because you don’t have to write something that you would otherwise have to write for almost every rule.” (P30)

Fig. 4
figure 4

Process of a content structuring content analysis as presented by Kuckartz (2014)

3.3.1 Initial Text Work

The initial text work step initiates our qualitative analysis. Kuckartz (2014) suggests to read through all the material and highlight important segments as well as to write notes for the transcripts using memos. Following these suggestions, we apply initial coding from constructivist grounded theory (Charmaz 2014; Vollstedt and Rezat 2019; Stol et al. 2016) to mark and summarize all text segments where interviewees reason about their beliefs on influence factors about the discussed properties. To do so, the two authors, which conducted and transcribed the interviews, read through all transcripts and mark all relevant text segments with codes that preferably represented the segment word for word. The codes allow for easier reference in later steps and, due to tooling, we are still able to quickly read the underlying text segment if necessary.

During this step, the example statement was labelled with the code automatic tracing increases expressiveness because no manual overhead.

3.3.2 Developing Thematic Main Codes

For developing the thematic main codes for our study we follow the common practice of inferring them from our research questions as suggested by Kuckartz (2014). Since the goal of our research is to investigate implicit assumptions, and factors that influence the assessment of experts about properties of model transformation languages three main codes arise:

  • Properties: Denoting which property is being discussed (e.g. Comprehensibility).

  • Factors: Denoting what influences a discussed property according to an interviewee (e.g. Bidirectional functionality of a MTL).

  • Factor assessment: Denoting an evaluation of how a factor influences a property (e.g. positive or negative or mixed depending on other factors).

The sub-codes for the property code can be directly defined based on the six properties from our previous literature review (Götz et al. 2021). As such they are deductive (a-priori) codes that are intended to mark text segments based on the properties that are being discussed in them.

3.3.3 Coding of All the Material with Main Codes

In order to code of all the material with the main codes one author analyses all interview transcripts. While doing so, the conversations about a discussed claim are marked with the code that is based on the property stated in the claim. To exemplify this, all discussions on the claim ”The use of MTLs increases the comprehensibility of model transformations.” are coded with the main code comprehensibility.

This realisation of the process step breaks with Kuckartz’s specifications in multiple ways. First, we do not code the material with the main codes Factors and Factor assessment, because all factors and factor assessments are already coded with the summarising initial codes. These will be refined into actual sub-codes of Factors and Factor assessment in a later step. Second, we directly code segments with the sub-codes for the Property main code, because the differentiation comes naturally with the structure of the interviews and delaying this refinement makes no sense. And third, this way of coding makes it possible that unimportant segments are also coded, something that Kuckartz suggests not to do. However, we actively decided in favour of this, because it accelerates the coding process enormously. Furthermore, only overlaps of the property codes with the other codes are considered, in later steps, thus automatically excluding unimportant text segments from consideration.

During this step, the coding for the example text segment was extended with the code Expressiveness. While this does not look like much of an enhancement on the surface, it is paramount to allow for systematic analysis in later steps.

After this step the example segment had its initial code, summarising the essence of the statement, and the explicit property sub-code Expressiveness, providing the first systematic categorisation of the segment.

3.3.4 Compilation of All Text Passages Coded with the Same Main Code

This step forms the basis for the subsequent iterative process of inductively developing sub-codes for each main code. Due to the use of the MAXQDA tool, this step is purely technical and does not require any special procedure outside of the selection of the main code that is being considered in the tool.

3.3.5 Inductive Development of Sub-Codes

The inductive development of sub-codes forms the most important coding step in our study. Inductive development here means that the sub-codes are developed based on the transcripts contents.

Kuckartz suggests to read through all segments coded with a main code to iteratively refine the code into several sub-codes that define the main category more precisely (Kuckartz 2014). We optimize this step by analysing all the initial codes from the Initial Text Work step, to construct concise and comprehensive codes for similar initial codes that could be used as sub-codes for the Factor or Factor assessment main codes. In doing so we follow the focused coding procedure of constructivist grounded theory to refine the initial code system.

All sub-codes of the Factor main code, that are refined using this process, are thematic codes, meaning they denote a specific topic or argument made within the transcripts. As a result, the sub-codes represent factors explicitly named by interviewees that influence the different properties. In contrast, all sub-codes of the Factor assessment main code, that are refined using this process, are evaluative codes, meaning they represent an evaluation, made by the authors, about an effect. More specifically, the codes represent an evaluation of how participants believe factors influence various properties.

Because of the importance of this coding step, the sub code refinement is created in a joint effort by three of the authors. First, over a period of three meetings, the authors develop comprehensive codes based on the initial codes of 18 interviews through discussions. Then the main author complements the resulting code system by analysing the remaining set of interview transcripts, while the two other authors each analyse half of them. In a final meeting any new sub code, devised by one of the authors, is discussed and a consensus for the complete code system is established.

During this step no code segment is extended with additional codes. Instead new codes derived from the initial codes are saved for usage in the following steps.

From the example code segment and its initial code, a sub-code automatic tracing for the Factors code was derived. The finalised sub-code Traceability was decided upon based on the combination with other derived codes of similar meaning, like traces.

3.3.6 Coding of All the Material with Complete Code System

After the final code system is established, the main author processes all transcripts to replace the initial codes with codes from the final code system. For this, each coded statement is re-coded with codes indicating the influence factors expressed by the interviewees as well as a factor assessment, if possible. This final coding step is done by the main author while all three co-authors each check 10 coded transcripts to validate the correct and consistent use of all codes and to make sure all relevant statements are considered. The results from the reviews are discussed in pairwise meetings between the main author and the reviewing co-author before being incorporated in a final coding approved by all authors.

During this step, the initial code for the example segment was dropped and replaced by the codes MTL advantage and Traceability.

The final codes assigned to the example text segment thus were: Expressiveness, Traceability and MTL advantage. The reasoning given within the statement as to why automatic tracing provides an expressiveness advantage, are manually extracted during analysis using tooling provided by MAXQDA.

3.3.7 Simple and Complex Analysis and Visualisation

The resulting coding and the coded text segments are then used as the basis for our analysis which, in accordance with our research question, focuses on identifying and evaluating factors that influence the properties of MTLs. As recommended by Kuckartz (2014), this is first done for each Property individually before analysis across all properties is conducted (as shown in Fig. 5).

Fig. 5
figure 5

Analysis forms in a content structuring content analysis as presented by Kuckartz (2014)

For analysing the influence factors of an individual property, we use the MAXQDA tooling to find segments coded with both a factor and the considered property. Using this approach we first compile a list of all factors relevant for a property, before then doing an in-depth analysis of all the gathered statements for each factor. Here the goal is to elicit commonalities and contradictions between the opinions of our interviewees that can be used to establish a theory on how each factor influences each property individually.

In terms of our example text segment, the segment and all other segments coded with Expressiveness and Traceability were read and analysed. The goal was to see if reduced overhead from implicit trace capabilities played a role in the argumentation of other participants and to gather all the other mentioned reasons.

For the analysis over all properties combined we apply the theoretical coding process of constructivist grounded theory (Charmaz 2014; Stol et al. 2016) to develop a model of influences. To do so, the Factor assessment s are used to examine how the factors influence the respective properties, what the commonalities between properties are and where the differences lie. The goal here is to develop a cohesive theory which explains the influences of factors on the individual properties but also on the properties as a whole and potential influences between the factors themselves.

In terms of our example text segment, the results from analysing Expressiveness and Traceability segments were compared to results from analysing segments coded with other property codes and Traceability. The goal was to find commonalities and differences between the analysed groups.

3.3.8 Privacy and Ethical concerns

All interview participants were informed of the data collection procedure, the handling of the data and their rights surrounding the process, prior to the interview.

During selection of potential participants the following data was collected and processed.

  • First & last name.

  • E-Mail address.

For participants that agreed to the partake in the interview study the following additional data was collected and processed during the course of the study.

  • Non anonymised audio recording of the interview.

  • Transcripts of the audio recordings.

All data collected during the study was not shared with any person outside of the group of authors. Audio recordings were handled only by the first and second author.

The complete information and consent form can be found in Appendix D. All participants have consented to having their interview recorded, transcribed and analysed based on this information. All interview recordings were stored on a single device with hardware encryption and deleted as soon as transcriptions were finalised. The interview transcripts were processed to prevent identification of participants. For this, identifying statements and names were removed.

Apart from the voice recordings and names, no sensitive information about the interviewees was collected.

The study design was not presented to an ethical board. The basis for this decision are the rules of the German Research Foundation (DFG) on when to use a ethical board in humanities and social sciencesFootnote 3. We refer to these guidelines because there are none specifically for software engineering research and humanities and social sciences are the closest related branch of science for our research.

4 Demographics

We interviewed, and got one written response, from a total of 56 experts from 16 different countries with varied backgrounds and experience levels and collected one comprehensive written response. Table 4 in Appendix C presents an overview of the demographic data about all interview participants. Experts and their statements are distinguished via an anonymous ID (P1 to P56).

4.1 Background

As evident from Fig. 6 participants with a research background constitute the largest portion of our interviewees. Overall there is an even split between participants solely from research and those that have at least some degree of industrial contact (either through industry projects or by working in industry). Only 3 participants stated to have used model transformations solely in an industrial context. This is in part offset by the fact that 25 of interviewees have executed research projects in cooperation with industry or have worked both in research and industry (22 and 3 respectively). While there is a definitive lack of industry practitioners present in our study, a large portion of interviewees are still able to provide insights into model transformations and model transformation languages with an industry view.

Fig. 6
figure 6

Distribution of participants background

Lastly, 10 of our participants are, in some capacity, involved in the development of model transformation languages. They can provide a different angle on advantages or disadvantages of MTLs compared to the 46 participants that use them solely for transformation purposes.

4.2 Experience

50 interviewees expressed to have 5 or more years of experience in using model transformations. Moreover, 24 of the participants have over 10 years of experience in the field. Lastly there was a single participant that had only used model transformations for a brief amount of time during their masters thesis.

4.3 Used Languages for Transformation Development

To better assess our participants and to qualify their answers with respect to their background we asked all interviewees to list languages they used to develop model transformations. Figure 7 summarises the answers given by participants while categorizing languages in one of three categories namely dedicated MTL, internal MTL and GPL. This differentiation is based on the classifications from Czarnecki and Helsen (2006) and Kahani et al. (2019).

Fig. 7
figure 7

Venn diagram depicting the language usage of participants

The distinction between GPL and dedicated/internal MTL is made, to gain an overview over how large the portion of users of general purpose languages for the development is, compared to the users of model transformation languages. Furthermore, it also allows for comprehending the viewpoint participants will take when answering questions throughout the interview, i.e. do they compare general purpose languages with model transformation languages based on their experience with both or do they give specific insights into their experiences with one of the two approaches. Internal MTL is separated from dedicated MTL because one claim within the interview protocol specifically explores the topic of internal model transformation languages.

52 participants have used dedicated model transformation languages such as ATL, Henshin or Viatra for transforming models. Only half as many (27) stated to have used general purpose languages for this goal. Lastly, only 5 indicated the use of internal MTLs.

When looking at the specific dedicated MTLs used ATL is by far the most prominent one used by interviewees. A total of 37 participants mention having used ATL. This is more than double the amount of the second most used language namely Henshin which is only mentioned by 17 interviewees. The QVT family then follows in third place with QvT-R having been used by 13 participants, QvT-O by 11. A complete overview over all dedicated model transformation languages used by our interviewees can be found in Fig. 8. Note that several interviewees mentioned using more than one language, making the total number of data points in this figure larger than 52.

Fig. 8
figure 8

Number of participants using a specific dedicated MTL

In the group of GPL languages used for model transformation (summarised in Fig. 9), Java is the most used language with 14 participants stating so. Note that several interviewees mentioned using more than one language, making the total number of data points in this figure larger than 27. Java is closely followed by Xtend which is mentioned by 12 interviewees. Then follows a steep drop of in popularity with Java Emitter Templates having been used by only four participants.

Fig. 9
figure 9

Number of participants using a specific GPL

Lastly, only four internal model transformation languages, namely RubyTL, NTL, NMF Synchronizations and FunnyQT, are mentioned. This shows a lack of prominence thereof. Moreover none of the languages is used by more than two interviewees.

5 Findings

Based on the responses of our interviewees and our analysis, we developed a framework to classify influence factors. It allows us to categorize how factors influence properties of MLTs and each other according to our interviewees. Note that we split the property Reuse & Maintainability into two properties for the purpose of reporting. This is done because interviewees chose to consider them separately. Thus reporting on them separately allows for presenting more nuanced results.

The factors themselves are split into six top-level factors namely GPL Capabilities, MTL Capabilities, Tooling, Choice of MTL, Skills and Use Case. The first factor, GPL Capabilities, encompasses sub-factors related to writing model transformations in general purpose languages. MTL Capabilities encompasses sub-factors that originate from transformation specific features of model transformation languages. Tooling contains factors surrounding tool support for MTLs. Choice of MTL details how the choice of language asserts its influence. The factor Skills encompasses sub-factors associated with skills. Lastly, the Use Cases factor contains sub-factors that relate to the involved use case an its influences.

Within the framework we differentiate between two kinds of factors. The first kind are factors, that have a positive or negative impact on properties of MTLs. These include the factors GPL Capabilities, MTL Capabilities and Tooling as well as their sub-factors. The second kind are factors that, depending on their characteristic, moderate how other factors influence properties, e.g. depending on the language, its syntax might have a positive or negative influence on the comprehensibility of written code. We call such factors moderating factors. These include the factors Choice of MTL, Skills and Use Case and their sub-factors.

Table 3 provides an overview over the answers given by our interviewees. The table shows factors on its rows and MTL properties on its columns. A + in a cell denotes, that interviewees mentioned the factor to have a positive effect on their view of the MTL property. A - means interviewees saw a negative influence and +/- describes that there have been mentions of both positive and negative influences. Lastly, a M in a cell denotes, that the factor represents a moderating factor for the MTL property, according to some interviewees. The detailed extent of the influence of each factor is described throughout Sections 5 and 6.

Table 3 Overview over quality attribute influences per factor

In the following we present all top-level factors and their sub-factors and describe their place within our framework. For each factor we detail its influence on properties of model transformation languages or on other factors, based on the statements made by our interviewees. All statements referred to in this section can be found verbatim in Table 5 in Appendix E.

5.1 GPL Capabilities

Using general purpose languages for developing model transformations, as an alternative to using dedicated languages was extensively discussed in our interviews. Interviewees mentioned both advantages and disadvantages that GPLs have compared to MTLs that made them view MTLs more or less favourably.

The disadvantages of GPLs compared to MTLs stem from additional features and abstractions that MTLs bring with them and will be discussed later in Section 5.2. The advantages of GPLs on the other hand can not be placed within the MTL Capability factors. These will instead be presented separately in this section.

According to our interviewees, advantages of GPLs are a relevant factor for all properties of MTLs.

General purpose languages are better suited for writing transformations that require lots of computations. This is because they were streamlined for these kinds of activities and designed for this task, with language features like streams, generics and lambdas. As a result, general purpose languages are far more advanced for such situations compared to model transformation languages, which sacrifice this for more domain expressiveness [Qgpl1].

Much like the language design for GPLs, their tools and ecosystems are mature and designed to integrate well with each other. Moreover, according to several interviewees, their tools are of high quality making developers feel more Productive [Qgpl2].

Lastly, multiple participants noted, that there are much more GPL developers readily available for companies to hire, thus making GPLs more attractive for them. This helps the Maintainability of existing code as such experts are more likely to Comprehend GPL code [Qgpl3]. Whether this aspect also improves the overall Productivity of transformation development in a GPL was disagreed upon, because it might be that developers trained in a MTL could produce similar results with less resources.

It was also mentioned, that much more training resources are available for GPL development, making it easier to start learning and using a new GPL compared to a MTL.

5.2 MTL Capabilities

The capabilities that model transformation languages provide that are not present in GPLs, are important factors that influence properties of the languages. This view is shared by our interviewees that raised many different aspects and abstractions present in model transformation languages.

The influence of capabilities specifically introduced in MTLs is diverse and depends on the concrete implementation in a specific language, the skills of the developers using the MTL and the use case in which the MTL is to be applied. We will discuss all the implications raised by our interviewees regarding the transformation specific capabilities of MTLs for the properties attributed to MTLs in detail, in this section.

5.2.1 Domain Focus

Domain Focus describes the fact that model transformation languages provide transformation specific constructs, abstractions or workflows. Interviewees remarked the domain focus, provided by MTLs, as influencing Comprehensibility, Ease of Writing, Expressiveness, Maintainability, Productivity and Tool Support. But the effects can differ depending on the specific MTL in question.

There exists a consensus that MTLs can provide better domain specificity than GPLs by introducing domain specific language constructs and abstractions. This increases Expressiveness by lifting the problem onto the same level as the involved models allowing developers to express more while writing less. MTLs allow developers to treat the transformation problem on the same abstraction level as the involved modelling languages [Qdf1]. This also improves the ease of development.

Several interviewees argued, that when moving towards domain specific concepts the Comprehensibility of written code is greatly increased. The reason for this is, that because transformation logic is written in terms of domain elements, unnecessary parts are omitted (compared to GPLs) and one can focus solely on the transformation aspect [Qdf2].

Having domain specific constructs was also raised as facilitating better Maintainability. Co-evolving transformations written in MTLs together with hardware-, technology-, platform- or model changes is said to be easier than in GPLs because “Once you have things like rules and helpers and things like left hand side and right hand side and all these patterns then [it is] easier to create things like meta-rules to take rules from one version to another version [...]” (P23).

Domain focus also enforces a stricter code structure on model transformations. This reduces the amount of variety in which they can be expressed in MTLs. As a result, developing Tool Support for analysing transformation scripts gets easier. Achieving similarly powerful tool support for general purpose languages, and even for internal MTLs, can be a lot harder or even impossible because much less is known solely based on the structure of the code. Analysis of GPL transformations has to deal with the complete array of functionality of general purpose language constructs [Qdf4]. While MTLs can be Turing complete too, they tend to limit this capability to specific sections of the transformation code. They also make more information about the transformation explicit compared to GPLs. This allows for easier analysis of properties of the transformation scripts which reduces the amount of work required to develop analysis tooling.

The influence of domain abstractions on Productivity was heavily discussed in our interviews. Interviewees agreed that, depending on the used language, Productivity gains are likely, due to their domain focus. However, one interviewee explained that precisely because of Productivity concerns companies in the industry might use general purpose languages. The reason for this boils down to the Use Case and project context. Infrastructure for general purpose languages might already be set up and developers do not need to be trained in other languages [Qdf5]. Moreover, different tasks might require different model transformation languages to fully utilise their benefits, which, from an organisational standpoint, does not make sense for a company. So instead one GPL is used for all tasks.

5.2.2 Bidirectionality

According to our interviewees bidirectional functionality in a model transformation language influences its Comprehensibility, Ease of Writing, Expressiveness and Maintainability and Productivity. Its effects on these properties then depends on the concrete implementation of the functionality in a MTL. It also depends on the Skills of the developers and the concrete Use Case.

Our interviewees mentioned that the problem of bidirectional transformations is inherently difficult and that high level formalisms are required to concisely define all aspects of such transformations. Many believe that because of this solutions using general purpose languages can never be sufficient. Statements in the vein of “in a general purpose programming language you would have to add a bit of clutter, a bit of distraction, from the real heart of the matter” (P42) were made several times. This, combined with having less optimal querying syntax, then shifts focus away from the actual transformation and decreases both the Comprehensibility and Maintainability of the written code.

Maintainability is also hampered because GPL solutions scatter the implementation throughout two or more rules (or methods or files) that have to be adapted in case of changes [Qbx2]. Expressive and high level syntax in MTLs helps alleviate these problems and increases the ease at which developers can write bidirectional model transformations.

Interviewees also commented on the fact that, thanks to bidirectional functionalities, consistency definitions and synchronisations between both sides of the transformation can be achieved easier. This improves the Maintainability of modelling projects as a whole and allows for more Productive workflows. Manual attempts to do so have been stated to be error-prone and labour-intensive.

It was also pointed out that the inherent complexity of bidirectionality leads to several problems that have to be considered. MTLs that offer syntax for defining bidirectional transformations are mentioned to be more complex to use as their unidirectional counterparts. They should thus only be used in cases where bidirectionality is a requirement. Moreover, one interviewee mentioned that developers are not generally used to thinking in bidirectional way [Qbx3].

Lastly, the models involved in bidirectional transformations also play a role regardless of the language used to define the transformation. Often the models are not equally powerful making it hard to actually achieve bidirectionality between them, because of information loss from one side to the other [Qbx4].

5.2.3 Incrementality

Dedicated functionality in MTLs for executing incremental transformations has been discussed as influencing Comprehensibility, Ease of Writing and Expressiveness. Similar to bidirectionality its influence is again heavily dependent on the Use Case in which incremental languages are applied as well as the Skills of the involved developers.

Declarative languages have been mentioned to facilitate incrementality because the execution semantics are already hidden and thus completely up to the execution engine. This increases the Expressiveness of language constructs. It can, however, hamper the Comprehensibility of transformation scripts for developers inexperienced with the language because there is no direct way of knowing in which order transformation steps are executed [Qinc1].

On the other hand interviewees also explained that writing incremental transformations in a GPL is unfeasible. Manual implementations are error-prone because too many kinds of changes have to be considered and chances are high that developers miss some specific kind. Due to the high level of complexity that the problem of incrementality inherently posses interviewees argued that writing such transformations in MTLs is much easier [Qinc2].

The same argumentation also applied for the Comprehensibility of transformations. All the additional code required to introduce incrementality to GPL transformations is argued to clutter the code so much that developers “[will be] in way over their head[s]” (P13).

As with bidirectionality interviewees agreed, that the Use Case needs to be carefully considered when arguing over incremental functionality. Only when ad-hoc incrementality is really needed should developers consider using incremental languages. In cases where transformations are executed in batches, maybe even over night, no actual incrementality is necessary and then “general purpose programming languages are very much contenders for implementing model transformations” (P42). It was also explained that using general purpose languages for simple transformations is common practice in industry as they are “very good in expressing [the required] control flow” (P42) and because none of the aforementioned problems for GPLs have a strong impact in these cases.

5.2.4 Mappings

The ability of a MTL to define mappings influences that languages Comprehensibility, Ease of Writing, Expressiveness, Maintainability and Reuse of model transformations. Developer Skills, the used Language and concrete Use Case also play an important role in the kind of influence.

Interviewees agreed, that the Expressiveness of transformation languages utilising syntax for mapping is increased due to them hiding low level operations [Qmap1]. However, as remarked by one participant, the semantic complexity of transformations can not be hidden by mappings, only the computational complexity.

According our interviewees mappings form a natural way of how people think about transformations. They impose a strict structure on how transformations need to be defined, making it easy for developers to start of writing transformations. The structure also aids general development, because all elements of a transformation have a predetermined place within a mapping. Being this restrictive has the advantage of directing ones thoughts and focus solely on the elements that should be transformed [Qmap2]. To transform an element, developers only need to write down the element and what it should be mapped to.

The simple structure expressed by mappings also benefits the Comprehensibility of transformations. It allows to easily grasp which elements are involved in a transformation, even by people that are not experienced in the used language. Trying to understand the same transformation in GPLs would be much harder because “[one] would not recognize [the involved elements] in Java code any more” (P32). Instead, they are hidden in between all the other instructions necessary to perform transformations in the language. Interviewees also mentioned that, due to the natural fit of mappings for transformations, it is much easier to find entry points from where to start and understand a transformation and to reconstruct relationships between input and output. This is aided by the fact that the order of mappings within a transformation does not need to conform with its execution sequence and thus enables developers to order them in a comprehensible way [Qmap3].

One interviewee explained that, from their experience, mappings lead to less code being written which makes the transformations both easier to comprehend and to maintain. However, they conceded that the competence of the involved developers is a crucial factor as well. According to them, language features alone do not make code maintainable. Developers need to have language engineering skills and intricate domain knowledge to be able to design well maintainable transformations [Qmap4]. Both are skills that too little developers posses.

Moreover, several interviewees raised the concern, that complex Use Cases can hamper the Comprehensibility of transformations. Understanding which elements are being mapped can be hard to grasp if several auxiliary functions are used for selecting the elements. Here one interviewee suggested that a standardized way of annotating such selections could help alleviate the problem.

It was also mentioned that, while mappings and other MTL features increase the Expressiveness of the language, they might make it harder for developers to start learning the languages. Because a lot of semantics are hidden behind keywords, developers need to first understand the hidden concepts to be able to utilise them correctly [Qmap5].

Other features that highlight how much Expressiveness is gained from mappings have also been mentioned. Mappings hide how relations between input and output are defined. This creates a formal and predictable correspondence between them and thus enables Tracing. Moreover, the correspondence between elements allows languages to provide functionality such as Bidirectionality and Incrementality [Qmap6].

Because many languages that utilise mappings can forgo definitions of explicit control flow, mappings allow transformation engines to do execution optimisations. However, one interviewee explained that they encountered Use Cases where developers want to influence the execution order, forcing them to introduce imperative elements into their code effectively hampering this advantage. It has also been mentioned that in complex cases the code within mappings can get complicated to the point where non experts are unable to comprehend the transformation again. This problem also exists for writing transformations as well. According to one interviewee mappings are great for linear transformations and are thus very dependent on the involved (meta-)models. Also in cases where complex interactions needs to be defined mappings do not present any advantage over GPL syntax and sometimes it can even be easier to define such logic in GPLs [Qmap7].

Lastly, mappings enable more modular code to be written. This in turn facilitates reuse, because reusing and changing code results in local changes instead of several changes throughout different parts of GPL code [Qmap8].

5.2.5 Traceability

The ability in model transformation languages to automatically create and handle trace information about the transformation has been discussed by our interviewees to influence Comprehensibility, Ease of writing, Expressiveness and Productivity. However, the concrete effect depends on the MTL and the skill of users.

All interviewees talking about automatic tracing agreed that it increases the Expressiveness of the language utilising it. In GPLs this functionality would need to be manually implemented using structures like hash maps. Code to set up traces would then also need to be added to all transformation rules [Qtrc1].

However, interviewees disagreed on how much this actually impacts the overall transformation development. Most interviewees felt like automatic trace handling Eases Writing transformations and even increases Productivity since no manual workarounds need to be implemented. This is because manual implementation requires developers to think about when and in which cases traces need to be created and how to access them correctly. It also enables languages that allow developers to define rules independent from the execution sequence. One interviewee however felt like this was not as effort intensive as commonly claimed and thus automatic trace handling to them is more of a nice to have feature than a requirement for writing transformations effectively. Moreover, for complex Use Cases of tracing such as QvTs late resolve, the Users are required to understand the principle of tracing [Qtrc2]. And according to another interviewee teaching how tracing and trace models work is hard.

Comprehending written transformations can also be aided by automatic trace management. Manual implementations introduce unnecessary clutter into transformation code that needs to be understood to be able to understand a whole transformation. This is especially true if effort has been put into making tracing work efficiently, according to one interviewee. Understanding a transformation is much more straight forward when only the established relationships between the input and output domains need to be considered, without any additional code to setup and use traces [Qtrc3].

Lastly, one interviewee raised the issue that manual trace handling might be necessary to write complex transformations involving multiple source and target models, as current engines are not intended for such Use Cases.

5.2.6 Automatic Model Traversal

According to our interviewees, the automatic traversal of the input model to apply transformations influences Ease of Writing, Expressiveness, Comprehensibility and Productivity. They also explain that depending on the implementation in a concrete MTL the effects can differ. Use Cases are also mentioned to be relevant to the influence of automatic traversal.

Automatic model traversal hides the traversal of the input model and how and when transformations are applied to the input model elements. Because of this many interviewees expressed that this feature in MTLs increases their Expressiveness. The reduced code clutter also helps with Comprehensibility.

It also Eases the Writing of transformations because developers do not need to worry about traversing the input and finding all relevant elements, a task that has been described as complicated by interviewees. This can be of significant help to developers. One interviewee explained, that they ran an experiment with several participants where they observed model traversal to be “one of the biggest problems for test persons” (P49).

Not having to manually define traversal reduces the amount of code that needs to be written and thus increases the overall Productivity of development, according to one interviewee. However, there can also be drawbacks from this practice. Hiding the traversal automatically leads to the context of model elements to be hidden from the developer. In cases where the context contains relevant information this can be detrimental and even mask errors that are hard to track down [Qtrv1].

Lastly, automatic input traversal enables transformation engines to optimize the order of execution in declarative MTLs. And MTLs where no automatic execution ordering can be performed have been described as being “close to plain GPLs” (P52).

5.2.7 Pattern-Matching

Some model transformation languages, such as Henshin, allow developers to define sub-graphs of the model graph, often using a graphical syntax, to be matched and transformed. This pattern-matching functionality influences the Comprehensibility, Expressiveness and Productivity, according to our interviewees. It is, however, strongly dependent on the specific language and Use Case. The feature is only present in a small portion of MTLs and brings with it its own set of restrictions depending on the concrete implementation in the language.

Pattern-matching functionality greatly increases the Expressiveness of MTLs. Similar to the basic model traversal no extra code has to be written to implement this semantic. However, the complexity of the abstracted functionality is even higher, since it is required to perform sub-graph matching to find all the relevant elements in a model. These patterns can also become arbitrarily complex and thus all interviewees talking about pattern-matching agreed that manual implementations are nearly impossible. Nevertheless one interviewee mentioned, that all languages they used that provided pattern-matching functionality (Henshin and TGG) had the drawback of providing no abstractions for resolving traces which takes away from its overall usefulness for certain Use Cases [Qpm1].

Not having to implement complex pattern-matching algorithms manually is also mentioned to increase the Productivity of writing transformations because this task is labour-intensive and error-prone.

Improvements for the Comprehensibility of transformations have also been recognized by some interviewees. They explained that the, often times graphical, syntax of languages with pattern-matching functionality allows to directly see the connection between involved elements. In GPLs this would be hidden behind all the code required to find and collect the elements. As such MTL code is “less polluted” (P52) than GPL code. Moreover, the Comprehensibility is also promoted by the fact that in some languages the graphical syntax shows the involved elements as they would be represented in the abstract syntax of the model.

5.2.8 Model Navigation

Dedicated syntax for expressing model navigation has influence on the Comprehensibility and Ease of Writing of model transformations as well as on the Expressiveness of the MTL that utilises it.

Having dedicated syntax for model navigation helps to ease development as it allows transformation engineers to simply express which elements or data they want to query from a model while the engine takes care of everything else. Furthermore, it has been mentioned that this has a positive effect on transformation development because developers do not need to consider the efficiency of the query compared to when defining such queries using nested loops in general purpose languages [Qnav1].

Because languages like OCL abstract from how a model is navigated to compute the results of a query, interviewees attributed a higher Expressiveness to them than GPL solutions and described code written in these languages as more concise. Several interviewees attribute a better Comprehensibility to OCL as a result of this conciseness, arguing that well designed conditions and queries written in OCL are easy to read [Qnav2].

OCL has however also been criticised by an interviewee. According to them, the language is too practically oriented, misses a good theoretical foundation and lacks elegance to properly express ones intent. They explain that because of this, the worth of learning such a language compared to using a more common language is uncertain.

5.2.9 Model Management

The impact of having to read and write models from and to files, i.e., model management, has been discussed by several interviewees. Automatic model management was discussed in our interviews as influencing the Comprehensibility, Ease of Writing and Expressiveness of model transformations in MTLs.

The argument for all three properties boils down to developers not having to write code for reading input models or writing output models, as well as the automatic creation of output elements and the order thereof. Interviewees agreed that implicit language functionality for these aspects raised the Expressiveness of languages. It reduces clutter when reading a written transformation and thus improving the Comprehensibility. Finally, developers do not have to deal with model management tasks, e.g. using the right format, that are not relevant to the actual transformation which helps with writing transformations [Qman1].

5.2.10 Reuse Mechanism

Mechanisms to reuse model transformations mostly influence the Reusability of model transformations in MTLs. Their concrete influence depends on the used Language and how reuse is handled in it. Interviewees also reported on cases where the users Skills with the language was relevant because novices might not be familiar with how the provided facilities can be utilised to achieve reuse.

There exists discourse between the interviewees about reuse mechanisms and their usefulness in model transformation languages. Several interviewees argued that MTLs do not have any reuse mechanisms that go beyond what is already present in general purpose languages. They believe that most, if not all, the reuse mechanisms that exist in MTLs are already present in GPLs and as such MTLs do not provide any reuse advantages [Qrm1]. According to them such reuse mechanisms include things like rule inheritance from languages like ATL or modules and libraries.

Other interviewees on the other hand suggested that while the aforementioned mechanisms stem from general purpose languages, they are still more transformation specific than their GPL counterpart. This is, because the mechanisms work on transformation specific parts in MTLs rather than generic constructs in GPLs [Qrm2, Qrm3]. Because of this focus, interviewees argue that they are more straight forward to use and thus improve Reusability in MTLs.

Interviewees also explained that there exist many languages that do not provide any useful reuse or modularisation mechanisms and that even in those that do it can be hard to achieve Reusability in a useful manner. However, one participant acknowledged that in their case, the reason for this might also relate to the inability of the Users to properly utilize the available mechanisms.

It has also been mentioned that reuse in model transformations is an inherently complex problem to solve. Transferring needs between two transformations which apply on different meta-models is difficult to do. As such, model transformation are often tightly tied to the domain in which they are used which makes reuse hard to achieve and most reuse between domains is currently done via copy & paste. This argument can present a reason why, as criticised by several interviewees, no advanced reuse mechanisms are broadly available.

The desire for advanced mechanisms has been expressed several times. One interviewee would like to see a mechanism that allows to define transformations to adapt to different input and output models to really feel like MTLs provide better reusability than GPLs. Another mentioned, that all reuse mechanisms conferred from object orientation rely on the tree like structure present in class structures while models are often more graph like and cyclic in nature. They believe that mechanisms that address this difference could be useful in MTLs.

Another disadvantage in some MTLs that was raised, is the granularity on which reuse can be defined. In languages like Henshin, for example, reuse is defined on a much coarser level than what is possible in GPLs.

Not having a central catalogue, similar to maven for Java, from which transformations or libraries can be reused, has also been critiqued as hindering reuse in model transformation languages.

5.2.11 Learnability

The learnability of model transformation languages has been discussed as influencing the Ease of Writing model transformations.

It has been criticised by several interviewees, that the learning curve for MTLs is steep. This is, in part, due to the fact that users not only need to learn the languages themselves, but also accompanying concepts from MDE which are often required to be able to fully utilise model transformation languages. The learning curve makes it difficult for users to get started and therefore hampers the Ease of Writing transformations [Qler1]. This effect could be observed among computer science students at several of the universities of our interviewees. The students were described to having difficulties adapting to the vastly different approach to development compared to more traditional methods. A potential reason for this could be that people come into contact with MDE principles too late, as noted by an interviewee [Qler2].

5.3 Tooling

While Tool Support is a MTL property that was investigated in our study, the tooling provided for MTLs, as well as several functional properties thereof, have been raised many times as factors that influence other properties attributed to model transformation languages as well. Most of the time this influence is negative, as tooling is seen as one of the greatest weak points of model transformation languages by our interviewees.

Many interviewees explained, that the most common languages do in fact have tools. The problem, however, lies in the fact that some helpful tools only exist for one language while others only exist for another language. As a result there is always some tool missing for any specific language. This leads people to feel like Tool Support for MTLs is bad compared to GPLs. Though there was one interviewee that explained that for their Use Cases, all tools required to be productive were present.

In the following, we will present several functional properties and tools that interviewees expressed as influential for Tool Support as well as other properties of MTLs.

5.3.1 Analysis Tooling

Analysis tools are seen as a strong suit of MTLs. Their existence in MTLs is said to impact Productivity, Comprehensibility and perceived Tool Support.

According to the interviewees, some analyses can only be carried out on MTLs, as the abstraction in transformations in GPLs is not high enough and too much information is represented by the program logic and not in analysable constructs. As one interviewee explained, this comes from the fact that for complex analysis, such as validating correctness, languages need to be more structured. Nevertheless, participants mainly mentioned analyses they would like to see, which is an indication that, while the potential for analysis tools for MTLs is high, they do not yet see usable solutions for it, or are unaware of it. This is highlighted by one interviewee that explained that they are missing ways to check properties of model transformations, even though such solutions exist for certain MTLs [Qdb1].

A desired analysis tool mentioned in the interviews is rule dependency analysis and visualisation. They believe that such a tool would provide valuable insights into the structure of written transformations and help to better comprehend them and their intent. “What I would need for QVT-R, for example, in more complex cases, would be a kind of dependency graph.” (P32). Moreover two interviewees expressed the desire for tools to verify that transformations uphold certain properties or preserve certain properties of the involved models.

5.3.2 Code Repositories

A gap in Tool Support that has been brought up several times, is a central platform to share transformation libraries, much like maven-central for Java or npm for JavaScript. This tool influences Tool Support and the Reusability of MTLs.

According to two interviewees, not having a central repository where transformations, written by other developers, can be browsed or downloaded, greatly hinders their view on the Reusability of model transformation languages. This is because it creates a barrier for reuse. For one thing, it is difficult to find model transformations that can be reused. Secondly, mechanisms that would simplify such reuse are then also missing. “I think what is currently missing is a catalogue or a tool like maven for having repositories for transformations so you can possibly find transformations to reuse.” (P14)

5.3.3 Debugging Tooling

Debuggers have been raised as essential tools that help with the Comprehensibility of written model transformations. The existence of a debugger for a given language therefore influences its Tool Support as well as its Comprehensibility.

One interviewee explained that, especially for declarative languages, where the execution deviates greatly from the definition, debugging tools would be a tremendous help in understanding what is going on. In this context, opinions were also expressed that more information is needed for debugging model transformations than for traditional programming and that the tools should therefore be able to do more. Interviewees mentioned the desire to be able to see how matchings arise or why expected matches are not produced as well as the ability to challenge their transformations with assertions to see when and why expressions evaluate to certain values. “Demonstrate to me that this is true, show me the section of the transformation in which this OCL constraint is true or false.” (P28).

Valuable debugging of model transformations is mainly possible in dedicated MTLs, according to one interviewee. They argue that debugging model transformations in GPLs is cumbersome because much of the code does not relate to the actual transformation thus hampering a developers ability to grasp problems in their code.

5.3.4 Ecosystem

The ecosystems around a language, as well as existing ecosystems, in which model transformations languages would have to be incorporated into, were remarked as mostly limiting factors for Productivity, Maintainability as well as the perceived amount of Tool Support.

One interviewee explained, that for many companies, adopting a model transformation language for their modelling concerns is out of the question because it would negatively impact their Productivity. The reason for this are existing ecosystems, which are designed for GPL usage. Moreover, it was noted that, to fully utilise the benefits of dedicated languages many companies would need to adopt several languages to properly utilise their benefits. This is seen as hard to implement as “people from industry have a hard time when they are required to use multiple languages” (P49) making it hard for them to maintain code in such ecosystems.

Ecosystems surrounding MTLs have also been criticised in hampering Productivity and perceived Tool Support. Several interviewees mentioned, that developers tend to favour ecosystems where many activities can be done in one place, something they see as lacking in MTL ecosystems. One interviewee even referred to this problem as the reason why they turned away from using model transformation languages completely [Qeco2].

This issue somewhat contrasts a concern raised by a different group of interviewees. They felt that the coupling of much of MDE tooling to Eclipse is a problem that hampers the adoption of MTLs and MDE. This coupling allows many tools to be available within the Eclipse IDE but, according to them, the problem lies in the fact that Eclipse is developed at a faster pace than what tool developers are able to keep up with, leaving much of the Tool Support for MTLs in an outdated state, limiting their exposure and usability [Qeco3].

5.3.5 IDE Tooling

One essential tool for Tool Support, Ease of Writing and Maintainability of MTLs are language specific editors in IDEs.

Several interviewees mentioned, that languages without basic IDE support are likely to be unusable, because developers are used to all the quality-of-life improvements, with autocompletion and syntax highlighting being the two most important features offered by such tools. Refactoring capabilities in IDEs, like renaming, have also been raised as crucial, especially for easing the Maintainability of transformations.

5.3.6 Interoperability

How well tools can be integrated with each other has been raised as a concern for the Tool Support of MTLs by several interviewees.

Interviewees see a clear advantage for GPLs when talking about interoperability between different MTL tools. They believe, that due to the majority of tools being research projects, little effort is spent into standardizing those in a way that allows for interoperability on the level that is currently provided for GPLs. “But the technologies, to combine them, it is difficult [...]” (P36). One interviewee described their first hand experience with this. They could not get a MTL to process models they generated with a software architecture tool because it produced non standard UML models which could not be used by the MTL. This problem has been echoed by another interviewee who explained that many MTLs do not work with non EMF compatible models.

5.3.7 Tooling Awareness

A few interviewees talked about the availability of information about tools and the general awareness of which tools for MTLs exist. According to them, this strongly influences the perceived lack of Tool Support for model transformation languages in general.

When starting out with model transformations it can be hard to find out which tools one should use or even which tools are available at all. Two interview participants mention experiencing this first hand. They further explain that there exists no central starting point when looking for tools and tools are generally not well communicated to potential users outside of research [Qawa1].

Another interviewee suspected that the same problem also happens the other way around. They believe that some well designed MTL tools are completely unknown outside of the companies that developed them for internal use.

5.3.8 Tool Creation Effort

The amount of effort, that is required to be put into the development of MTL tools, has been raised by many interviewees as a reason why Tool Support for MTLs is seen as lacking.

All interviewees talking about the effort involved in creating tools for MTLs agree that there is a lot of effort involved in developing tools. This is not a problem in and of itself but, when comparing tooling with GPLs interviewees felt like MTLs being at a disadvantage. The disadvantage stems from the community for MTLs being much smaller and thus having less man power to develop tools which limits the amount of tools that can be developed. Several interviewees noted, that the only solution they see for this problem is industrial backing or commercial tool vendors because “I am keenly aware of the cost to being able to develop a good programming language, the cost of maintaining it and the cost of adding debuggers and refactoring engines. It is enormous.” (P1).

When comparing the actual effort for creating transformation specific tools, some interviewees explained that their experience suggests easier tool development for MTLs than for GPLs. They explained that, extracting the transformation specific information necessary for such tools out of GPL code complicates the whole process, whereas dedicated MTLs with their small and focused language core provide much easier access to such information [Qtce2].

5.3.9 Tool Learnability

The learning curve for someone starting off with MTLs and MTL tools is discussed as a heavy burden to the perceived effectiveness of Tools and even influences Ease of Writing.

Several interviewees criticised the fact that when starting off with a new MTL and its accompanying tools there is little support for users. Many tools lack basic documentation on how to set them up properly and how to use them. As a result users feel lost and find it difficult to start off writing transformations [Qtle1].

5.3.10 Tool Usability

Related to the topic of learnability, the usability of tools for model transformation languages is discussed as influencing the quality of Tool Support for the languages as well as the Ease of Writing and Productivity.

To fully utilise the potential of MTLs useable tools are essential. Due to their higher level of abstraction, high quality tools are necessary to properly work with them and Write well rounded transformations [Quse1].

This is currently not the case when looking at the opinions of our interviewees talking about the topic of tool usability. There are tools available for people to start off with developing transformations but they are not well rounded and thus not ready for professional use, according to one interviewee. This is supported by several other interviewees opinions, many tools are faulty, which hinders the workflow and reduces Productivity [Quse2]. It has also been stated that if there were high quality useable tools available, they would be used. The reality for many users is, however, more in line with the experience of one interviewee who stated that they were unable to get many tools (for bidirectional languages) to even work at all.

5.3.11 Tool Maturity

A reason given for many of the criticised points surrounding MTL tools is their maturity. It is said to be a pivotal factor for everything related to Tool Support.

The maturity of tools for model transformation languages was commented on a lot. Tools need to be refined more in order to raze many of their current faults. The fact that this is not currently done relates back to the effort that is involved with it and the limited personnel available to do so. This is highlighted in an argument made by one of the interviewees who feels, that the community should not be hiding behind the argument of maturity [Qmat1].

5.3.12 Validation Tooling

Tools or frameworks to support the validation and testing of transformations written in MTLs have been discussed to influence the perceived Tool Support for nearly all MTLs.

Too much of the available tool support focuses solely on the writing phase of transformation development. There is little tool support for testing developed transformations, which has been raised as an area where much progress can, and has to be, made. Especially when comparing the current state of the art with GPLs, MTLs are seen as lacking [Qval1]. Not only are there little to no tools like unit testing frameworks, there is also too few transformation specific support such as tools to specifically verify binding or helper code in ATL.

5.4 Choice of MTL

The choice of MTL is an obvious factor that influences how other factors, such as the MTL Capabilities, influence the properties of model transformation languages. However, it should be explicitly mentioned, because it has been brought up countless times by interviewees while not often being considered in literature. Depending on the chosen model transformation language its capabilities and whole makeup changes, which has strong implications on all aspects of model transformation development.

A large number of the interviewees have commented on this. They either directly raised the concern, by prefacing a discussion with a statement such as “[...] it depends on the MTL”, or indirectly raised the concern, when comparing specific languages that do or do not exhibit certain capabilities and properties.

5.5 Skills

Skills of involved stakeholders is another group of factors that does not have a direct influence on how MTLs are perceived but instead plays a passive role. Many interviewees cited skills as a limiting factor to other influence factors. They argue that insufficient user skills could hinder advantages that MTLs can provide and might even create disadvantages compared to the more well-known and commonly used GPLs.

In this section we present the different types of skills mentioned by our interviewees as being relevant to the discussion of properties of model transformation languages.

5.5.1 Language Skills

The skill of developers in a specific model transformation language was raised by several interviewees as critical in facilitating many of the advantages provided through the languages capabilities. So much so that, according to them, the ability of developers to use and read a language can make or break any and all advantages and disadvantages of MTLs related to Comprehensibility, Ease of Writing, Maintainability and Reuseability.

Basic skills in any language are a prerequisite to being able to use it. They are also necessary to understand written code. There is no difference between GPLs and MTLs. It was however mentioned, that developers are generally more used to the development style in general purpose languages. Thus users need to learn how to solve a problem with the functionality of the model transformation language to be able to successfully develop transformations in a MTL [Qskl1]. This is especially relevant for complex transformations, where users are required to know of abstractions such as tracing or automatic traversal. Following on on this, one interviewee explained, that while learning the language is a requirement, using a new library, e.g. one for developing model transformations, in a GPL also entails learning and as such this must not be regarded as a disadvantage.

For reuse it is also paramount for users to know what elements of a transformation can be reused through language functionality. As a result the Reuseability is again limited by the knowledge of users in the specific language.

Lastly, being able to maintain a transformation written in an MTL also requires users to know the language to be able to understand where changes need to be made [Qskl2].

5.5.2 User Experience/Knowledge

Apart from mastering a used language, the amount of experience users have with said languages and techniques also play a vital role in bringing out the full potential of said languages. Our interviewees discussed this for Ease of Writing, Maintainability and Productivity.

One interviewee explained that, from their experience, the amount of experience developers have with a language greatly impacts their Productivity when using said language. The problem for MTLs that results from this is the fact that there is little incentive for a person that is trying to build up their CV to spend much time on dedicated languages such as MTLs [Qexp1]. Developers are more inclined to learning and accumulating experience in languages that are commonly used in different companies to improve their chances of landing jobs. As a result people tend to have little to no experience in using MTLs. This in turn results in them having a harder time developing transformations in these languages, and the final product being of lower quality than what they could achieve using a GPL in which they have more experience in.

The problem is further exacerbated in teaching. “Many MDSE courses are just given too late, when people are too acquainted with GPLs, and then its really hard for students to see the point of using models, modelling and MTLs, because it’s comparable with languages and stuff they have already learned and worked with.” (P6).

5.6 Use Case

Similar as the MTL itself and stakeholder Skills, the concrete Use Case in which model transformations are being developed is another factor that does not directly influence how properties of MTLs are being assessed. Instead, interviewees often mention that, depending on the Use Case, other influence factors could either have a positive or negative effect.

Use cases are distinguished along three dimensions. The complexity of involved models based on their structure, the complexity of the transformation based on the semantic gap between source and target, and the size of the transformation based on the use case. Depending on which differentiation is referred to by the interviewees, the considerations look differently.

5.6.1 Involved (Meta-) Models

The involved models and meta-models can have a large impact on the transformation and can hence heavily influence the advantages or disadvantages that MTLs exhibit.

Writing transformations for well behaved models, meaning models that are well structured and documented, can be immensely productive in a MTL while ‘badly’ behaved models bring out problems that require well trained experts to properly solve in a MTL. The UML meta-model was put forth as an example for such a badly behaved meta-model by one interviewee. According to them, transformations involving UML models can be problematic due to templates, which are model elements that are parametrized by other model elements, and eGenericType. The problem with these complex model elements is often worsened by low-quality documentation [Qmod1]. In cases where these badly behaved models are involved, many of the advantages from advanced features of MTLs can not be properly utilised without powerful tooling.

5.6.2 Semantic Gap Between Input and Output

Many interviewees formulate considerations based on the differentiation between ‘simple’ and ‘complex’ transformations in terms of the semantic gap that needs to be overcome. Transformations are considered simple when there is little semantic difference between the source and target models. Common comparisons read like: “transforming boxes into circles” (P32).

For simple transformations, model transformation languages are regarded as taking a lot of work off of the developers through the different language features discussed in Section 5.2. In more complex cases, transformations will get more complex and the developers experience gets more and more relevant, as more advanced language features need to be utilised, which can favour GPLs [Qgap1].

Others argue that the advantages of MTLs only really come into play in more complex cases or when high level features, such as bidirectionality or incrementality, are required. The reasoning for this argument is, that in simple cases the overhead of GPLs is not that prominent. Moreover, for writing complex transformations, dedicated query languages in MTLs are regarded by some to be much better than having to manually define complex conditions and loops in a GPL.

5.6.3 Size

The Size of the transformation based on the Use Case is considered by some interviewees to be a relevant factor as well. In cases with many rules that depend on each other, MTLs are seen as having advantages [Qsiz1]. The size of transformations has been said to be a limiting factor for the use of graphical languages as enormous transformations would make graphical notations confusing. Modularisation mechanisms of languages also become a relevant feature in these cases.

6 Cross-Factor Findings

Based on interview responses, we developed a structure model from structural equation modeling (Weiber and Mühlhaus 2021) that models interactions between the presented influence factors and the properties of model transformation languages.

Structure models depict assumed relationships between variables (Weiber and Mühlhaus 2021). They divide their components into endogenous and exogenous variables. The endogenous variables are explained by the causal influences assumed in the model. The exogenous variables serve as explanatory variables, but are not themselves explained by the causal model. Exogenous variables either directly influence a endogenous variable or they moderate an influence of another exogenous variable on an endogenous variable.

Structure models are therefore well suited to provide a theoretical framework for the findings of our work. Factors identified during analysis constitute exogenous variables while MTL properties constitute endogenous variables. Moderating factors also constitute exogenous variables, with the caveat of only having moderating influences on other influences.

A graphical overview over the influences identified by us can be found in Fig. 10. The detailed structure model is depicted in Fig. 11.

Fig. 10
figure 10

Graphical overview over factor influences and moderations

Fig. 11
figure 11

Structure model of influence and moderation effects of factors on MTL properties (Due to its size, the model has been made interactive. Please refer to the supplementary materials for the interactive version.)

The structure model depicts which MTL properties are influenced by which of the identified factors. For each MTL property the model also illustrates which factors moderate the influence on the property. Rectangles represent factors, rounded rectangles represent MTL properties. Below each MTL property the moderating factors for the property are displayed. Arrows between a factor and a MTL property represent the factor having an influence on the MTL property. Each influence on a MTL property is moderated by its moderating factors. The graphical representation deviates from standard presentation due to its size.

The capabilities of model transformation languages based on domain specific abstractions are aimed at providing advantages over general purpose languages. Whether these advantages are realised or whether disadvantages emerge is moderated by the Skills of the users, the concrete MTL chosen as well as the Use Case for which transformations are applied. Depending on these organisational factors the versatility provided by general purpose languages may overshadow advantages provided by MTLs.

Tooling can aid the usage of advanced features of MTLs by supporting developers in their endeavours beyond simple syntax highlighting. As a result, tools can further promote the advantages that stem from the domain specific abstractions. The biggest problem that tools for MTLs face is their availability and quality.

In the following we will present thorough discussions of the most salient observations based on our interviews and the presented structure model. Note that, as will be thoroughly discussed in Section 8.2, the observations have a limited applicability for industry use cases due to the lack of interviewees that use MTLs in an industry setting.

6.1 The Effects of MTL Capabilities

Capabilities of model transformation languages that go beyond what general purpose languages can offer, are regarded as opportunities for better support in development of transformations. The advantage often boils down to not having to manually implement the functionality in question when it is required. It also helps reduce clutter in transformation code, putting the mapping of input and output at the centre of attention. Moreover, they aid developers in handling problems specific to the transformation domain, such as synchronisations and the relationship of input and output values.

This does however come with its own set of limitations. Model transformation languages favour a different way of problem solving that is well suited to the problem at hand, but is unfamiliar for the common programmer. This is amplified by an education that is heavily focused on imperative programming and lacks deeper exposure to logical and functional programming. Knowledge and understanding of functional concepts would help developers when using query languages such as OCL, while logical concepts often find application in graph based transformation languages. The domain specific mechanisms in model transformation languages also make generalisations harder. This is highlighted in the discussions regarding reusability. Interviewees commonly referred to transformations as conceptionally hard to reuse because of their specificity that makes them applicable only to the use case for which they were developed.

6.2 Tooling Impact on Properties Other than Tool Support

Tooling, or the lack thereof, is a main factor that influences how people perceive the quality and availability of usable model transformation languages. However, our interviews show that tooling also facilitates many other properties. This is because tools are not developed as an end in themselves. Tools are intended to support developers in their efforts to develop and maintain their code.

As a result, the quality of available tools is a major factor that impacts all aspects of a MTL. “Basically all the good aids you see in a Java environment should be there even better in a MTL tool, because model transformation is so much more abstract and more relevant that you should be having tools that are again more abstract and more relevant.” (P28). Problems in the area of Usability, Maturity and Interoperability of tools have also been reported on in empirical studies on MDE in general (Whittle et al. 2013; Mohagheghi et al. 2013).

Herein also lies the biggest problem for model transformation languages. The quality of tools is inadequate. While there do exist good and useable tools, they are far and between, only exist for certain languages and are not integrated with each other. This greatly diminishes the potential of model transformation languages because, compared to general purpose languages, developing with them can often be scattered over multiple separate workflows and tools. There do exist many tools, but most of them are prototypical in nature and only available for individual languages. This makes it hard to fully utilise the capabilities of a MTL when suitable tools only exist in theory.

The lack of good tools can be attributed mainly to the amount of work required to develop them and the comparatively small community. Moreover, there are no large commercial vendors, that could put in the required resources to develop tools of a commercially viable quality.

6.3 The Importance of Moderating Factors

The saying “Use the right tool for the job” also applies to the context of model transformations. One of the most important things to note is that depending on the context, such as Use Case and developer Skills, the right language to use can differ greatly. This was highlighted time and time again in our interviews.

Interviewees insisted that the combination of use case and the concrete implementation of a language feature significantly change how well a feature supports properties such as Comprehensibility or Productivity, i.e. the influence of factors is moderated by ‘contextual’ factors. For one, the implementation of a feature in a language might not fit well for the problem that needs to be solved. Or the feature is not required at all and thus could impose effort on developers that is seen as unnecessary. In our opinion, this stems from the use cases language developers intended the language for. For example, a language such as Henshin is intended for cases where patterns of model elements need to be matched and manipulated. In such cases, the features provided by Henshin can bring significant advantages over implementing the intended transformation in general purpose languages. Other use cases, where these features are not required, bring no advantage. They can even have negative effects as the language design around them might hinder users from developing a straight forward solution.

The skills and background knowledge of users is relevant, as it can greatly influence how comfortable people are in using a language. This in turn reflects on how well they can perform. This is problematic for the adoption of model transformation languages, as programmers tend to be trained in imperative, general purpose languages. As a result, a gentle learning curve is essential and the initial costs of learning need to bear fruit in adequate time. The choice of using an MTL is therefore a long term investment that is not necessarily suited for only a single project.

The considerations around Use Case, Skills and the Choice of MTL are not novel, but they are rarely discussed explicitly. This is concerning because almost any decision process will come back to these three factors and their sub-factors, as seen in the fact that the influence on each MTL property in Fig. 11 is moderated by at least one of them. They provide organisational considerations that come into play before transformation development begins. Moreover, organisational concerns have already been identified as relevant factors for general MDE adoption (Whittle et al. 2013; Hutchinson et al. 2011; Hutchinson et al. 2011). As such they have to be at the centre of attention of researchers and language developers too.

7 Actionable Results

In this section we present and discuss actionable results that arise from the responses made by our interviewees and analysis thereof. Results will largely focus on actions that can be taken by researchers, because they make up the largest portion of our interview participants.

7.1 Evaluation and Development of MTL Capabilities

Our interviewees mentioned a large number of model transformation language capabilities and reasoned about their implications for the investigated properties of MTL. We believe the detailed results of the interviews can form a basis for further research into two key aspects:

  1. (I)

    backing up the expert opinions with empirical data

  2. (II)

    improving existing model transformation languages

7.1.1 Evaluation of MTL Capabilities and Properties

In our interviews, experts voiced many opinions on how and why factors influence the various MTL aspects examined in our study. The opinions were always based on personal experiences, experiences of colleagues and reasoning. We therefore believe, that our results provide a good insight into the communities sentiment and show that there exists consensus between the experts in many aspects. Model transformation language capabilities are considered largely beneficial, except for certain edge cases. However, empirical data to support this consensus is still missing. The lack of empirical studies into the topic of model transformation languages has already been highlighted in our preceding study (Götz et al. 2021).

We are firmly convinced that researchers within the community need to carry out extensive empirical studies, to back up the expert opinions and to explore the exact limits that interviewees hinted at.

We envision two main types of studies, experiments and case studies. Setting up experiments that consider real-world examples with a large number of suitable participants would be optimal but is hard to achieve. Introducing large transformation examples in experiments is very time-consuming and requires that all participants are experts in the languages used. In addition, recruiting appropriate participants is a generally difficult in software engineering studies (Rainer and Wohlin 2021). We recommend using existing transformations as the object of study in experiments instead. This enables the analysis of complex systems to generate quantitative data without involving human subjects.

If the assessments and experiences of developers are to be the central object of study, we recommend to set up case studies. This allows researchers to study effects in complex, real-world settings over a longer period of time. This is important because the exact effects of, for example, the use of a certain language feature often only become apparent to developers after a long period of use. Case studies of research projects or even industrial transformation systems can thus be used to obtain detailed information on the impact of the applied technologies.

To design such studies, our results can form an important basis.

(a) Empirical factor evaluation.

How and under which circumstances the factors we have identified affect MTL properties needs to be comprehensively evaluated. Here we envision both qualitative and quantitative studies that focus on the impact of a single factor or a group of related factors. These could, for example, make comparisons between cases where a factor does or does not apply. The results of such studies can help language developers make decisions about features to include in their MTLs.

Our interviews provide extensive context that should be taken into account in the study design and interpretation of results. For example, our interviews show that the semantic gap between input and output defines a relevant context that needs to be considered. For this reason, when investigating the advantages and disadvantages of mappings, transformations involving models with different levels of semantic gaps between input and output have to be used, to be able to fully evaluate all relevant use cases. Some transformations need to contain complicated selection conditions or complex calculations for attributes while others need to have less complicated expressions. Researchers can then evaluate how well mappings in a language fit the different scenarios to aid in providing a clear picture of their advantages and disadvantages.

(b) Empirical MTL property evaluation.

What advantages or disadvantages MTLs really have is still up for debate. We believe that the credibility of research efforts on MTL can be greatly improved with studies that provide empirical substantiation to the speculated properties. Advances like those made by Hebig et al. (2018) are rare and further ones, based on real world examples, must be carried out.

Our results can also make a valuable contribution to such studies. The factors we have identified as influencing a property can be taken into account in studies from the outset. They can be used to formulate null hypotheses on why a MTL is superior or inferior to a GPL when considering one specific property.

For example, a study that is interested in investigating the Comprehensibility of MTLs compared to GPLs can find a number of factors in our results that need to be taken into account. Such factors include tracing mechanisms, mappings or pattern matching capabilities. Researchers can consciously decide which of them are relevant for the transformations used in the study and what impact their presence or absence has on the study results. Based on these considerations hypotheses can be formed.

A recent study we conducted provides an example of how these considerations can be used to expand the body of empirical studies on this topic (Höppner et al. 2021). By focusing the investigation on Mappings, Model Navigation and Tracing we were able to present clear and focused results for comparing and explaining differences in the expressiveness of transformation code written in ATL and Java. We concentrated our analysis on these factors because they all influence Expressiveness according to our interviews.

Such considerations should of course be part of any proper study, but our results provide a basis that can be useful in ensuring that no relevant factors are overlooked.

(c) Influence Quantification.

Lastly, the results of this study should be quantified. The design of the reported study makes quantification of the importance of factors and their influence strengths impossible. However, such quantification is necessary to prioritise which factors to focus on first, both for assessment and for improvement. We intend to design and execute such a study as future work to this study.

We can use structural equation modelling methods (Weiber and Mühlhaus 2021) to quantify the factors and their influences because we already have a structure model. We plan to use an online survey to query users of MTLs from research and industry about the amount they use different language features, their perception of qualitative properties of their transformations and demographic data surrounding use-case, skills & experience and used languages. The responses are used as input for universal structure modelling (USM) (Buckler and Hennig-Thurau 2008) based on the structural equation model developed from the interview responses.

USM is used to estimate the influence and moderation weights of all variables within the structure model. We can therefore use it to produce quantified data on the influence and moderation effects of identified factors.

We are confident that the approach of using a survey to quantify interview results, can complement the current results, because several of the authors have had positive experiences applying it (Liebel et al. 2018; Juhnke et al. 2020).

7.1.2 Improving MTL Capabilities

To improve current model transformation languages the criticisms articulated by interviewees can be used as starting points for enhancements and innovation. There are several aspects that are considered to be problematic by our interviewees.

(d) Improve reuse mechanism adoption.

Reuse mechanisms in model transformation languages are one aspect where interviewees saw potential for improvement (see Section 5.2.10). Languages that do not currently possess mature reuse mechanisms can adopt them to become more usable. For the adoption of mature reuse mechanisms in MTLs we see the languages developers as responsible.

(e) Reuse mechanism innovation.

Innovation towards transformation specific reuse mechanisms, as has been requested by some participants (see Section 5.2.10), should also be advanced. This topic was discussed at length during the interviews on the statement “Having written several transformations, we have identified that current MTLs are too low a level of abstraction for succinctly expressing transformations between DSLs, because they demonstrate several recurring patterns that have to be reimplemented each time.” in Question Set 3.

Interviewees pointed out a need for reuse mechanisms that allow transformations to adapt to differing inputs and outputs. It would be conceivable to define transformation rules, or parts of them, independently of concrete model types, similar to generics in GPLs. This would allow development of generic transformation ‘templates’ of common transformation patterns. One pattern, for example, could be finding and manipulating specific model structures, like cliques, independent of the concrete model elements involved. Such templates can then be reused and adapted in all transformations where the pattern is required.

We believe, that innovating such new transformation specific reuse mechanisms is a community wide effort that needs to be taken on in order to make them more widely usable.

(f) Improving MDSE education.

The Learnability of MTLs has also been a point of criticism. We believe, that more effort needs to be put into the transfer of knowledge for MDSE and its techniques like model transformations and MTLs. This believe is supported by the findings of Hutchinson et al. (2011). They also identified the lack of MDSE knowledge as a limiting factor for the adoption of the approach.

People need to come into contact with the principles earlier so that the inhibition threshold to apply them is lower. This was also remarked by interviewees when discussing the Learnability (see Section 5.2.11). More focus needs to be given to modelling and modelling techniques in software engineering courses. This is especially important since the skill of users has been said to be a largely impactful factor upon which many of the advantages from other MTL capabilities rely. Furthermore, there exist studies such as the one by Dieste et al. (2017), which detected a connection between the experience of developers with a language and their productivity as well as the code quality of the resulting programs.

To achieve this, we believe, that the researchers from the community, in their role as higher education teachers and university staff, need to become active. They should advocate for teaching the concepts of MDSE and the advantages/disadvantages in undergraduate studies in computer science study programmes. This view is shared by Samiee et al. (2018). Particularly, it should be taught that models can be used for more than documentation purposes, e.g., code generation, simulations early in the development cycle, test case generation. These other uses are widely and successfully employed in the domain of cyber-physical systems according to Bucchiarone et al. (2021). Hence, it might be beneficial to include industrial modelling tools like Matlab/Simulink/Stateflow from this domain in addition to standard UML tools in undergraduate courses. Furthermore, we successfully used simulation frameworks for autonomous cars, like Carla (Dosovitskiy et al. 2017), in the past as targets for student projects when teaching courses on the development of modeling languages and model transformations. For example, the students devised a state machine language and code generator targeting the simulation framework to develop an automatic parking functionality. Model transformations were developed to flatten hierarchical state machines to non-hierarchical state machines prior to code generation.

(g) Increase knowledge retention.

It is also difficult to get to grips with the subject matter in general, as information on it is much harder to obtain than on general purpose programming (see Section 5.3.7). This starts with the fact that, we found websites on MTLs to often be outdated or unappealing and lack good tutorials and comprehensible documentation. These points need to be fixed, by the language developers, to provide potential users with better resources to combat the perceived steepness of the learning curve. More active community involvement is also conceivable here. Users of MTLs could invest time in creating documentation and keeping it up-to-date. The possibility of this working and producing good results can be seen in examples such as the arch-linux wikiFootnote 4.

(h) Improve community outwards presentation.

The model transformation community is small. In our opinion this leads to less innovation and poses the danger of entrenched practices. The problem is not limited to small communities as seen by, for example, the risk averse movie industry or low innovation automotive industry. An improved outwards presentation of the technology of model transformations can help alleviate the problem of limited human resources. The current hype surrounding low-code-platforms can be used to inspire young and aspiring researchers to contribute to its underlying concepts such as model transformations.

(i) Improve industry outreach and cooperation.

We think it is also paramount to pursue industry cooperation to gauge industrial needs in order to facilitate more industrial adoption of MTLs. Here ambitious studies are required that attempt to provide the community with clear requirements specific domains of industry have for MDE and transformation languages, as well as to show for which domains application is reasonable at all. There exist some field studies by Staron (2006), Mohagheghi and Dehlen (2008), and Mohagheghi et al. (2013) but they are far and in between and do not focus on the transformation languages involved. The research community can attempt to organize solutions for these requirements based on such field study and industry research. However, for such industry cooperation to be possible, a focused community outreach is required. There are notable advancements in this direction e.g. MDENetFootnote 5, but they are still in their infancy and require more involvement by the research community.

(j) Provide representative model transformation languages.

To provide reasonable evidence that model transformation languages can be competitive against GPLs there also needs to be heavy focus on providing less prototypical and more pragmatic and useable transformation languages (see Section 6.2). To that end only a few selected languages should be attempted to be made production ready, potentially through further industry cooperation. MTLs could be integrated into commercial modelling tools in order to be able to process models programmatically in the tool.

Alternatively, few modern standardised MTLs could be promoted by the community. Since such a decision has far-reaching effects, a central, community wide respected body is needed. The OMG could possibly take action for this as they are already deciding on community impacting standards.

The QVT standard was an ambitious push in this direction. However, we believe that the initiative needs a fresh take, given the findings of the last 20 years of research. This idea is supported by several interviewees who considered QVT to be bloated and outdated. Especially in the areas of bidirectional and incremental transformations we see huge potential. Furthermore, relying more on declarative approaches for defining uni-directional transformations should also be considered. This trend can also be observed in the field of GPLs with the introduction of more and more functional concepts into them.

Innovation in prototypical languages should then be thoroughly evaluated for its usefulness before adoption into one of the flagship languages. It is not the task of research to produce industry ready languages, but setting up the environment and using these languages should not be more complicated than for any general purpose programming language.

(k) Research legacy integration.

The integration of MTLs into existing legacy systems has been remarked as a huge entry barrier for industry adoption (see Sections 5.2.1 and 5.3.4). We believe this stems from a lack of techniques that facilitate gradual integration of modelling technology into existing systems and infrastructure. This is highlighted by the fact that basic literature such as that by Brambilla et al. (2017) does not contain any suggestions to this end. To combat this, we propose a dedicated branch of MDE research focused on developing tools and processes to integrate model driven techniques into legacy systems.

We envision distinct guidelines and processes on how to integrate transformations and transformation concepts into existing systems. There should be terms of reference as to which types of system components lend themselves to the use of model transformations. Furthermore, descriptions of which transformations and which transformation languages are suitable for which type of use case are also required. Having such guides can reduce the barrier of entry, because they provide a clear course of action when trying to (gradually) adopt the paradigm.

This also includes accessible GPL bindings for applying model transformation concepts. They can be used to gradually replace system components that can benefit from the use of transformations. This can be done without the overhead of integrating a new language and intermediate models. One example for this is DresdenOCL, a OCL dialect that can be used on Java code (Demuth and Wilke 2009).

7.2 Steps Towards Solving the Tooling Problem

From our interviews, we have to conclude that the biggest weak point of model transformation languages is their Tool Support.

The two biggest tooling gaps that we were able to identify are:

  1. (I)

    many necessary tools do not exist

  2. (II)

    existing tools lack user-friendliness and are not compatible with each other

We hope that our work can be a starting point in counteracting these drawbacks.

(l) Provide essential tooling.

In our view, tooling of flagship model transformation languages needs to be extended to include all the essential tools mentioned in the interviews to make MTLs production and industry ready. This includes useable Editors, Debuggers and Validation or Analysis tools. At best all such tools for a language should be useable within one IDE. One way language developers can help with this task is by implementing the Language Server Protocol (LSP) (Microsoft) or its graphical counterpart GLSP (Eclipse Foundation) for their MTL. This would greatly improve the ability of tool developers to create and distribute tooling.

(m) Develop transformation specific debugging.

As mentioned by our interviewees, for debuggers there is a need for model transformation specific techniques. Troya et al. (2022) showed that there are numerous advances in this area like those by Wimmer et al. (2009), Hibberd et al. (2007), and Ege and Tichy (2019) but none of them have led to well rounded debuggers yet. Further effort by researchers active in this area is therefore required. They should strive to develop their approaches to a point where they can be productively used to demonstrate their usefulness for a productive transformation development.

(n) Improve tool usability.

Most importantly, a lot of effort needs to be put into improving the usability of MTL tools. Our interviews have shown, that unusable tools are the most off putting factor that hampers wider adoption. To combat this, we believe usability studies to be essential. Studies to identify usability issues in the likeness of what is proposed by Pietron et al. (2018) can be used to gain insights into where problems originate from and how to improve them. Such studies have already been successfully utilised for other MDE related tooling (Stegmaier et al. 2019). We therefore need more researchers from the community to get involved in designing and conducting usability studies for tooling surrounding MTLs.

We think the results of usability studies can also provide useful lessons learned for tool developers to make tools more usable from the beginning. The overall goal must be to find out what needs to be changed or improved in MTL tools to make their adoption significant. Industrial efforts to provide proper tool support can then be based on these results and the existing, usable, tools. This adoption is necessary because, in our view, the human resources required for providing adequate long-term support for the tools can only be provided by commercially operating companies. Such long term support is necessary so that model transformation languages, and their accompanying tools, can gain a foothold in the fast-moving industrial world. The industrialisation of MTL tooling was also proposed during an open community discussion detailed by Burgueño et al. (2019).

The goal should be to provide well rounded, all-in-one solutions that integrate all necessary tooling in one place, to make development as seamless as possible. The appropriateness of this has been shown by Jonkers et al. (2006).

(o) Limit-test internal MTLs.

A different approach that should be further explored is the attempt to thoroughly embed an internal model transformation language in a main stream GPL as done by Hinkel and Goldschmidt (2019). The advantage of this approach is the ability to inherit tooling of the host language (Hinkel and Goldschmidt 2019) and it allows general purpose developers to apply their rich pool of experience. However, there are some drawbacks to this approach, as discussed in Section 5. The amount of tooling that can be properly integrated is limited and it is more difficult to develop transformation specific tooling for internal languages as it is hard to extract the required information from the code. For this reason, we think, the required tools should be known at design time and the language has to be designed to expose all the required information while not imposing this as an additional burden on developers. Researchers that plan to develop an internal model transformation language should therefore thoroughly asses the tool requirements for the use case for which they intend to develop their language.

8 Threats to Validity

Our interview study was carefully designed, and followed reputable guidelines for preparation, conduction and analysis. Nonetheless there are some threats to validity that need to be discussed to provide a complete picture of our study and its results.

8.1 Internal Validity

Internal validity describes the extent to which a casual conclusion based on the study is warranted. The validity is threatened by manual errors and biases of the involved researchers throughout the study process.

Errors could have been introduced during the transcription phase and during the analysis of the data since both steps were conducted by a single author at a time.

To prevent transcription errors, all transcripts were re-examined after completion to ensure consistency between the transcripts and audio recordings.

To minimize possible confirmation biases introduced during analysis and categorisation of interviewee statements, random samples were checked by other authors to find possible discrepancies between the authors assessments on statements. In cases where such discrepancies were encountered, thorough discussions between all authors were conducted to find a consensus that was then applied to all transcripts containing similar considerations.

Lastly there is the potential of misinterpretation of interviewees responses during analysis. While we carefully stuck to interpret statements literally during coding, there are words and phrases that have overloaded meanings. During the interviews, it would always be necessary to ask exactly what meaning interviewees used, but this was not always possible. Therefore the threat could not be mitigated completely as contextual information was required to interpret interviewees responses in some cases.

8.2 External Validity

External validity describes the extent to which the results of a study can be generalised. In our interview study this validity is threatened by our interview participant assortment ,which is a result of our sampling and selection method.

We utilise convenience sampling interviewing any and all people that respond to our emails. This can limit how representative the final group of interviewees is of the target population. The issue here is that we do not know much about the makeup of the target population. It is therefore difficult to assess how much the group of participants deviates from a representative set.

Using research publications as the starting point for participant selection also introduces a bias towards users from research. This can be clearly seen in Fig. 6. There is an apparent lack of participants from industry which limits the applicability of our results to industrial cases. This threat is somewhat mitigated by the fact that half of all participants do have at least some contact with industry, either through research projects in conjunction with industry or by having worked in industry.

Another threat to external validity relates to model to text (M2T) transformations. Only a few of our participants stated to have experience in applying M2T transformations. This is a result of how the initial set of potential participants was constructed. The search terms used in the SLR miss terms that relate to M2T such as ‘code generation’ or ‘model to text’. This limitation was opted into to avoid having to differentiate between the two transformation approaches during analysis. Moreover, the consensus during discussions was that we were talking about model to model transformations. As such, our results can not be applied to the field of model to text languages.

Lastly, there is the threat of participation bias. Participants may disproportionately posses a trait that reduces the generalisability of their responses. People that view model transformation languages positively might be more inclined to participate than critics. We can not preclude this threat, but, the amount of critique we were able to elicit from the interviews suggests the effects from this bias to be weak. Other impacts of this bias are discussed in Section 8.4.

8.3 Construct Validity

Construct validity describes the extent to which the right method was applied to find answers for the research question. This validity is threatened by an inappropriate method that allows for errors.

Prior to conducting our research much work went into designing a proper framework to use. Here we relied on reputable existing guidelines for both the interview and analysis parts of this work. We used open ended questions to facilitate an open space for participants to bring forth any and all their opinions and considerations for the topic at hand. The statements used as guidance can however present a potential threat since their wording could introduce an unconscious bias in our interviewees. To combat this we selected broad statements as well as used both a negative and a positive statement for each discussed property. However, there is a chance that these measures were not fully sufficient.

Lastly, it can not be excluded that some relevant factors have not been raised during our interviews. We have interviewed a large number of people, but this threat cannot be overcome because of the study design and the open nature of our research question.

8.4 Conclusion Validity

Conclusion validity describes the extent to which our results stem from the investigated variables and are reproducible. Here, the biases of our participants represent the biggest threat.

It is safe to assume that people who do research on a subject are more likely to see it in a positive light and less likely to find anything negative about it. As such there is the possibility that too little negative impact factors were considered and presented. However, we found that the people we interviewed were also able to deal with the topic in a very critical way. We therefore conclude that the statements may have been somewhat more positively loaded, but that the results themselves are meaningful.

9 Related Work

To the best of our knowledge, there exists no other interview study that focuses on influence factors on the advantages and disadvantages of model transformation languages. Nonetheless there exist several works that can be related to our study. The related work is divided into empirical studies on model transformation languages, empirical studies on model transformations in general and interview studies on MDE.

9.1 Empirical Studies on Model Transformation Languages

A structured literature review we conducted (Götz et al. 2021) forms the basis for the work presented in this paper. The goal of the reported literature review was to extract and categorize claims about the advantages and disadvantages of model transformation languages as well as to learn and report on the current state of evaluation thereof. The authors searched over 4000 publications to extract a total of 58 publications that directly claim properties of model transformation languages. In total the authors found 137 claims and categorized them into 15 properties. From their work the authors conclude that while many advantages and disadvantages are claimed little to no studies have been executed to verify them. They also point out a general lack of context and background information on the claimed properties that hinders evaluation and prompts scepticism.

Burgueño et al. report on a online survey, as well as a subsequent open discussion, at the 12th edition of the International Conference on Model Transformations (ICMT’2019) about the future of model transformation languages (Burgueño et al. 2019). Their goal for the survey was to identify reasons as to why developers decided for or against the use of model transformation languages and what their opinion on the future of these languages was. At ICMT’2019 where the results of the survey were presented they then moderated an open discussion on the same topic. The results of the study indicate that MTLs have fallen in popularity compared to at the beginning of the decade which they attribute to technical issues, tooling issues, social issues and the fact that general purpose languages have assimilated ideas from MTLs making GPLs a more viable option for defining model transformations. While their methodology differed from our interview study, the results of both studies support each other. However the results of our study are more detailed and provide a larger body of background knowledge that is relevant for future studies on the subject.

The notion of general purpose programming languages as alternatives to MTLs for writing model transformations has been explored by Hebig et al. (2018) and by us (Götz et al. 2021). Hebig et al. (2018) report on a controlled experiment where student participants had to complete three tasks involved in the development of model transformations. One task was to comprehend an existing transformation, one task involved modifying an existing transformation and one task required the participants to develop a transformation from scratch. The authors compare how the use of ATL, QVT-O and the general purpose language Xtend affect the outcome of the three tasks. Their results show no clear evidence of an advantage when using a MTL compared to a GPL but concede the narrow conditions under which the observation was made. The study provides a rare example of empirical evaluation of MTLs of which we suggest that more be made. The narrow conditions the authors struggled with could be alleviated by follow-up studies that draw from our results for defining their boundaries.

In a recent study by us (Götz et al. 2021) we put the value of model transformation language into a historical perspective and drew from the preliminary results of the interview study for the study setup. We compare the complexity of a set of 10 model transformations written in ATL with their counterparts written in Java SE5, which was current around 2006 when ATL was first introduced, and Java SE14. The Java transformations were translated from the ATL modules based on a predefined translation schema. The findings support the assumptions from Burgueño et al. (2019) in part. While we found that newer Java features such as Streams allow for a significant reduction in cyclomatic complexity and lines of code the relative amount of complexity of aspects that ATL can hide stays the same between the two Java versions.

Gerpheide et al. use an exploratory study with expert interviews, a literature review and introspection to formalize a quality model for the QVTo model transformation standard by the OMG (Gerpheide et al. 2016). They validate their quality model using a survey and afterwards use the quality model to identify tool support need of transformation developers. In a final step the authors design and evaluate a code test coverage tool for QVTo. Their study is similar to ours in that they also relied on expert interviews for their goal. The end goal of the study however differs from ours as they used the interviews to design a quality model for QvTo while we used it to formulate influence factors on quality attributes of model transformation languages

Lastly there are two study templates for evaluating model transformation languages which have yet to be used for executing actual studies. Kramer et al. present a template for a controlled experiment to evaluate the comprehensibility of model transformation languages (Kramer et al. 2016). Their approach suggests the use of a paper-based questionnaire to let participants prove their ability to understand what a transformation code snippet does. The influence of the language in which the code is written on comprehension speed and quality is then measured by comparing the average number of correct answers and the average time spent to fill out the questionnaires. Strüber and Anjorin propose a controlled experiment for comparing the benefits and drawbacks of the reusability mechanisms rule refinement and variability-based rules (Strüber and Anjorin 2016). They suggest that the value of the reusability of an approach can be measured by looking at the comprehensibility of the two mechanisms as well their changeability, which is measured through bug-fixing and modification tasks. The results of studies executed based on both study templates could draw from our results for their final design and would provide valuable empirical data, a gap we identified in this and the preceding literature review.

9.2 Empirical Studies on Model Transformations

Tehrani et al. executed an interview based study on requirements engineering for model transformation development (Tehrani et al. 2016). Their goal was to identify and understand the contexts and manner in which model transformations are applied as well as how requirements for them are established. To this end they interviewed 5 industry experts. From the interviews the authors found that out of 7 transformation projects only a single project was developed in an already existing project while all other projects were created from scratch. Their findings are relevant to our work since participants in our study agreed that it is hard to integrate MTLs in existing infrastructures. Whether the fact that MTLs are hard to integrate was an influence factor for the projects considered in the interview study by them is however not clear.

Groner et al. utilize an exploratory mixed method study consisting of a survey and subsequent interviews with a selection of the survey participants to try and evaluate how developers deal with performance issues in their model transformations (Groner et al. 2020). They also asses the causes and solutions that developers experienced. The survey results show that over half of all developers have experienced performance issues in their transformations. While the interviews allowed the authors to identify and categorize performance causes and solutions into 3 categories: Engine related, Transformation definition related and Model related. From the interviews they were also able to identify that tools such as useable profilers and static analyses would help developers in managing performance issues. The results of their study highlight that some of the factors identified by us are also relevant for other MTL properties not directly investigated in our study.

9.3 Interview Studies on Model Driven Software Engineering

There are numerous publications and several groups of researchers that have carried out large scale, in-depth empirical studies on model driven engineering as a whole. We focus on a selection of those that have relation to our study in terms of findings.

Whittle, Hutchinson, Rouncefield et al. used questionnaires (Hutchinson et al. 2011; Hutchinson et al. 2014) and interviews (Whittle et al. 2013; Hutchinson et al. 2011; Hutchinson et al. 2011; Hutchinson et al. 2014) to elicit positive and negative consequences of the usage of MDE in industrial settings. Apart from technical factors related to tooling they also found organisational and social factors that impact the adoption and efficacy of MDE. Several of their findings for MDE in general coincide with results from our study. Related to tooling they too found the factors of Interoperability, Maturity and Usability to be influential. Moreover, on the organisational side, the small amount of people that are knowledgable in MDE techniques and the problem of integrating into existing infrastructure are also results Whittle et al. found. Lastly, developers being more interested in using techniques that help build their CV was identified by them as a limiting factor too.

Staron analyses data collected from a case study of MDE adoption at two companies where one company withdrew from adopting MDE while the other was in the process of adoption (Staron 2006). Their findings suggest that legacy code was a main influence factor on whether a cost efficient MDE adoption was possible. This observation is consistent with our findings that integrating MTLs into existing infrastructures has a negative impact on the Productivity that can be achieved with MTLs.

The research group surrounding Mohagheghi also carried out multiple empirical studies on MDE, focusing on factors for and consequences of adoption thereof. They use surveys and interviews at several companies (Mohagheghi et al. 2013; Mohagheghi et al. 2013) as well as a literature review (Mohagheghi and Dehlen 2008) for this purpose. In addition to mature tooling, factors identified by the authors are usefulness, ease of use and compatibility with existing tools. Similar to statements by our interviewees, they also found that MDE is seen as a long term investment. It is not well suited for single projects.

Lastly, Akdur et al. report on a large online survey of people from the domain of embedded systems industry (Akdur et al. 2018). They too found tools surrounding MDE to be a major factor. Another interesting finding by them was that UML models are by far the most commonly used models. This is of relevance to our results since one of our interviewees pointed out, that the makeup of some UML models can have detrimental effects on the usefulness of MTLs.

The results of all presented research groups show, that many of the factors we identified for MTLs also apply to MDE in general which provides additional confidence in our results and shows that advancements in these areas would have a high impact.

10 Conclusion

There are many claims about the advantages and disadvantages of model transformation languages. In this paper, we presented and argued the detailed factors that play a role for such claims. Based on interviews with 56 participants from research and industry we present a structure model of relevant factors for the Ease of writing, Expressiveness, Comprehensibility, Tool Support, Productivity, Reuse and Maintainability of model transformation languages. For each factor we detail which properties they influence and how they influence them. We have identified two types of factors. There are factors that have a direct impact on said properties, e.g. different capabilities of model transformation languages like automatic trace handling. And there are factors that define a context whose characteristics moderate the the impact of the former factors, e.g. the Skills of developers.

Based on the interview results we suggest a number of tangible actions that need to be taken in order to convey the viability of model transformation languages and MDSE. For one, empirical studies need to be executed to provide proper substantiation to claimed properties. We also need to see more innovation for transformation specific reuse, legacy integration and need to improve outreach and presentation to both industry and academia. Lastly, efforts must be made to improve tool support and especially tool usability for MTLs.

For all of the suggested actions, our results can provide detailed data to draw from.