Abstract
App recommendations and data visualizations such as weather forecasts, navigation aids, or sleep tracking graphs play an increasingly important role in daily decisions. However, the apps’ underlying functioning often remains opaque, possibly resulting in a suboptimal user experience or inadequate reliance on recommendations. To approach design solutions for this, the paper investigates the effects of textual and graphic transparency cues on users’ mental model accuracy, user experience, and explanation satisfaction, using the example of a weather and a sleep tracking app. An online experiment with 293 participants showed that textual transparency cues (i.e., verbal explanations) led to higher felt and objectively measured mental model accuracy than graphic transparency cues (i.e., data visualizations). Textual cues were also more satisfying than graphic cues but did not result in significantly different ratings of user experience. Moreover, differences between textual and graphic cues in subjective mental model accuracy and explanation satisfaction were stronger for the weather than the sleep tracking app, implying context-specific differences in the impact of transparency cues. The results and limitations are discussed and linked to the challenge of finding a sweet spot for technology transparency design.
1 Introduction
People increasingly rely on data gathered, processed, and visualized by smartphone apps for making decisions in daily life. For instance, weather apps are used to decide which clothes to wear and navigation apps to find the most efficient travel route. However, to evaluate such recommendations for everyday life, it is important that users can assess how much weight they can give to the app’s recommendations: how likely is it that an umbrella will be needed if a weather app predicts rain, and should the planned route be changed if a navigation app detects the risk of a traffic jam? To make an informed choice, users need a sound understanding of how to interpret the app’s information correctly, when to follow the advice of an app and when to reject it. Yet, many systems do not provide sufficient explanations to users as to how the underlying processes work and how they lead to the respective outputs, 1 sometimes because the importance of the users’ understanding fades away in the design process. 2
Such an opacity is not only a practical but also an ethical problem. On the one hand, it might be annoying to choose the wrong clothes for specific weather conditions, but on the other hand, when it comes to more serious application areas, misunderstood health recommendations may have far-reaching consequences for the user’s well-being (e.g., if they are not tailored to the user’s conditions). Therefore, more transparent and explainable technology designs are necessary and advocated for. 3 , 4 In the scientific community, a field of research with a wide variety of terms emerged to investigate possible solutions. Relevant keywords for this field include technology transparency, explainability, interpretability, or understandability. 5 , 6 In this paper, we focus on the term of transparency as an overarching concept describing that users understand what the technology is currently doing, how it arrives at its outputs, and why. 1 , 7 Yet, in line with the lack of clear terminology and definitions, there is no consensus on how transparency cues should be designed. 8 , 9
Research so far has often investigated different contents of transparency cues, the so-called what aspects. 9 For instance, global explanations about how a technology in general works were compared to local explanations about why a particular output emerged. 10 However, how aspects, describing how a specific content should be presented to promote technology transparency, remained in the background. 9 Conceptually, text-/natural language-based presentation formats were often differentiated from multimedia formats (e.g., graphics, images, animation, audio). 8 , 11 Empirically, mostly natural language cues (e.g., phrases, texts) were examined and different designs of the same presentation format compared rather than different presentation formats themselves with each other (e.g., text-based cues with visualizations). 8 Yet, other formats than text-based cues that are more graphic-based may be more beneficial to implement because they can be used uniformly for different languages and do not have to be translated for diverse user groups. Furthermore, they may be easier to process and understand for users, not binding much cognitive capacity or overwhelming the user while delivering the same information. 12 In contrast, graphic-based cues might require more interpretation by users and accordingly be more prone to misunderstanding. From a cognitive capacity perspective, one could assume that there is a sweet spot where the user gets enough transparency to understand what is essential but is also not overwhelmed by too much complexity. In this regard, graphic-based versus text-based cues may be seen as two opposite poles of a continuum – both aiming to create transparency but using different channels. Previous research has investigated graphic and textual cues 13 but lacks a systematic understanding of which form of transparency cue is more effective given specific applications. Thus, it is important to find the most adequate form of transparency design and investigate when which presentation format is best suited to convey transparency.
Therefore, this work poses the question of how transparency design can support users in developing an appropriate understanding of the technology, while making them feel satisfied with the explanation and the overall user experience. We investigated this research question by comparing the influence of two largely applicable transparency formats (textual and graphic) in two mobile contexts (weather app and sleep tracking app). These specific examples were selected for several reasons: Everyday applications were chosen since they have not received much attention in previous research compared to, for example, expert systems, 14 , 15 although they are used frequently by most people and thus greatly impact their daily life. Additionally, we used apps for which the need for guidelines on transparency cue design has been highlighted in previous research. First, Vaughn et al. 16 argue that weather apps, contrary to television broadcasts or online articles, frequently lack contextualization of their forecast, which forces users to interpret the data by themselves. Given that over 75 % of their sample used weather apps as a primary source of information about the weather, it is imperative that users can correctly interpret the provided data and its limitations. Second, users draw similarly important conclusions from sleep apps (e.g., when to get up to be most awake). However, in a previous survey, over 60 % of the participants did not feel that current sleep apps enhanced their understanding of their sleep hygiene. 17 Since, to our knowledge, design guidelines for making both apps transparent, understandable, and satisfying are lacking, we aim to make a practically relevant contribution.
In the following sections, we first present current research on how transparency cues affect the mental model and understanding of users (effectiveness perspective) as well as the user experience and satisfaction (experience perspective), including our assumptions. Both perspectives are essential to consider because transparency cues can lead to conflicting results, improving outcomes of one perspective but impairing outcomes of the other. Likewise, Hoffman et al. 18 identified both the user’s satisfaction and mental model as crucial outcomes of technology transparency. 18 Then, the methods and results of the study are presented. Lastly, we discuss our findings and give an outlook on future research endeavors and practical implications.
1.1 Transparency cues as determinants of user mental model
A mental model is a conceptual representation of how a system functions, constructed by users and based on their understanding of the system. 19 A correct mental model allows users to adequately assess the reliability of a system, while insufficient mental models can result in wrong conclusions or evaluations from system outputs. They can lead to users distrusting or overtrusting a system instead of trusting it to an extent that matches its capabilities. 20 For instance, when users underestimate the capabilities of a research tool, they may distrust and take over the research themselves, even though they could use the technology as support and it could take work off their hands. Conversely, if they overestimate its capability, they may rely too much on the tool and not question the results when necessary. Thus, the accuracy of users’ mental models is crucial for an appropriate and adapted use of the technology.
Research suggested that transparency cues can improve users’ mental models. For example, Gunning et al. 21 postulated that an explainable system enables users to revise their mental model, which then can lead to appropriate trust, use, and better performance. Likewise, the Human-Automation System Oversight (HASO) Model 22 states that the system’s transparency influences the mental model and thereby promotes situation awareness. Situation awareness, in turn, helps users to monitor the system and intervene if necessary to avoid unwanted consequences. Such effects of transparency cues are also empirically supported. As an example, Ramaraj et al. 23 showed that natural language and visual explanations of a robot’s behavior led to more accurate mental models of users. Similarly, Anderson et al. 24 demonstrated that more extensive explanations were associated with better mental models of users who interacted with an AI during a game – though at the cost of higher cognitive loads and partly lower performance. In the context of mobile apps, higher transparency in a mobile sensing app significantly affected users’ understanding of data usage in the app but not their understanding of data logging. 25
Additionally, the felt and actual accuracy of the mental model can diverge: Users might subjectively feel like they understand how an app functions even if their explanations are objectively wrong. 18 , 26 This can lead again to inappropriate technology use since users rely only on their subjective understanding of the system in their judgements. Therefore, in our study, we investigate both users’ subjective and objective mental model accuracy to identify corresponding challenges and risks in transparency design. We thus aim to inform design solutions that support appropriate and concurring subjective and objective mental models. Yet, based on the research reviewed, 21 , 22 , 23 , 24 , 25 it can be generally assumed that technology outputs with transparency cues result in a more accurate mental model than outputs without transparency cues:
H1:
Technology outputs with a textual or graphic cue lead to a more accurate subjective and objective mental model than outputs without such a cue.
It is yet unclear how exactly explanations should be designed to benefit user understanding instead of being perceived as redundant or overly complicated. 27 , 28 Additionally, little is known yet about the generalizability or context-sensitivity of the effect of transparency cues on users’ mental model. We aim to contribute to a more nuanced understanding by comparing the effects of two different types of transparency cues across two different user scenarios.
1.2 Transparency cues as determinants of user experience and explanation satisfaction
User experience describes the users’ perceptions and responses resulting from interacting with a system, such as the perceived attractiveness, efficiency, perspicuity, dependability, stimulation, and novelty of a product. 29 Also focusing on the experience of users, explanation satisfaction represents how content users are with explanations respective transparency cues provided by a system. 30 In contrast to the mental model accuracy determining how a system may be used, the experience of users may influence if it is used at all, making this perspective equally relevant for the evaluation of transparency design. 31 , 32 , 33
Referring to Gunning et al.’s 21 model of explainability, transparency cues should affect users’ subjective evaluation of the system, including how satisfied users are with specific explanations. Moreover, Shin and Park 34 postulated in their FAT (fairness, accountability, and transparency) model that technology transparency increases satisfaction and supported this assumption empirically. Various empirical findings, also integrated through reviews, showed similar effects. 8 , 35 , 36 Furthermore, models of information processing, such as the Heuristic-Systematic Model/Elaboration Likelihood Model 37 , 38 or the cognitive effort perspective 39 , imply that information that is more easily processed by the respective users may be preferred by them. Likewise, Grice 40 not only developed maxims related to the content of messages but also one maxim dealing with their presentation to define good communication. According to this maxim of manner, communication can be improved by creating messages with high clarity for recipients. Applied to transparency cues, this may imply that the presentation format impacts the perceptions of users. Accordingly, participants of Vermeulen et al. 41 specified that they prefer visualizations over textual explanations because the former are easier to understand. Similarly, participants with little technological knowledge favored visual transparency cues over other cues like counterfactual explanations. Yet, Wastensteiner et al. 13 found that a textual explanation was better understood but less preferred by subjects. In sum, we assume that even though textual cues might be easier to understand, users may experience graphic cues more positively than textual cues, leading to the following hypotheses:
H2:
The graphic cue results in a higher user experience than the textual cue.
H3:
The graphic cue is perceived as more satisfying than the textual cue.
2 Methods
2.1 Participants
We initially received responses from 334 participants. After excluding participants who completed less than 80 % of the survey (n = 41), our final sample consisted of 293 participants. No further participants had to be removed for answering the attention check incorrectly (“For this question, please mark the answer ‘strongly agree’.”) or for responding more than 2 SD quicker than the average processing time. This number exceeded the required sample size of 147 participants based on a priori power analyses with G*Power 3 (see preregistration at https://osf.io/mh938/?view_only=cb4d737a87a445f7b549be570d436eea for details). 42 Participation requirements were an age of majority and proficiency in German. The participants were recruited via university mailing lists, social media, and snowball sampling. As an incentive, they could choose between taking part in a raffle of four coupons of €25 each and getting course credit.
The sample included subjects with ages from 18 to 67 years (M = 24.3, SD = 7.3). Of the 293 participants, 70.3 % identified as female, 27.7 % as male, and 2.1 % as having a diverse gender. Moreover, most participants were students (87.0 %) or employees (10.2 %) and had a high school diploma (65.5 %) or a university degree (29.0 %).
2.2 Design and material
The study was conducted as an online experiment. We implemented a mixed design with transparency cue as a between-subjects factor (no cue vs. textual cue vs. graphic cue, see Figure 1) and the smartphone app type as a within-subjects factor (weather app vs. sleep tracking app). The transparency cue was manipulated by a varying interface design of the two apps. To avoid careless responding due to fatigue, 43 order effect biases, 44 and carry-over effects of app understanding from one transparency cue condition to the next one, 45 each participant experienced both apps but only with one transparency condition. As a control variable, the suggested reliability of each app was also varied in the conditions with a cue to check the robustness of the results, independent of the transparency cue content. Therefore, the sample contained n = 62 participants in the no cue condition, n = 119 participants in the textual cue group, and n = 112 participants in the graphic cue group.

Example interfaces of the three transparency cue conditions for the weather app: (A) no cue condition, (B) textual cue condition, and (C) graphic cue condition. Text passages translated from German into English.
During the study, participants should imagine themselves using the apps. For the weather app, participants envisioned that a temperature-sensitive tree, which they are taking care of for a traveling friend, is standing on their balcony, and they are using the weather app to infer whether they should leave the tree out overnight. For the sleep tracking app, the scenario described waking up and checking the sleep data gathered by a sleep tracking ring worn the previous night. The app interface presented a recommendation to get up half an hour earlier in the morning based on the data collected.
Depending on the experimental condition, a different type of transparency cue about the app’s output was displayed (see Figure 1 for examples; all scenarios can be found in Appendix A):
In the no cue group, no information on the functioning of the app was given.
In the textual cue group, information on the functioning of the app was provided by a natural language, sentence-based explanation. This included a sentence informing users that the forecast is determined in such a way that the displayed temperature can deviate from the actual one by a certain number of degrees (for the weather app) and a sentence explaining sources of information that were recorded and used to compute the recommendation (for the sleep tracking app).
In the graphic cue group, information on the functioning of the app was provided by a visual, graphic-based explanation. This included a graph showing an area around the predicted temperature which might also be possible (for the weather app) and graphs for each source of information on which the app’s recommendation is based with symbolic markers indicating why the specific recommendation was made (for the sleep tracking app).
We developed the transparency cues in all conditions as an explanation of how the app output arises and aimed for a realistic design, like it might be implemented in real user settings. To this end, the textual cues were designed to be concrete, sound, and complete 46 while still being concise enough (two sentences) that real users would take the time to read them. The graphic cues were designed to contain the same information as the textual cues. Line graphs were chosen since this type is frequently used in weather and sleep tracking apps, is relatively simple and should therefore be familiar and understandable to participants. Similar graph types for both apps facilitate comparisons across the two different contexts.
Additionally, regarding the control variable, the transparency cue suggested that the app has either low or high reliability. In the weather app, this was done by varying how much the actual temperature might deviate from the prediction and, in the sleep app, by varying on how many types of recorded data the app based its recommendation. This should create the impression that the app output is not reliable, giving information on shortcomings of the app, versus that it is reliable, making it safe to consider. Since this is not the focus of this paper, we will not go into this in more detail below.
2.3 Procedure
The study was preregistered, approved by the university’s ethics committee (ethics vote: EK-MIS-2024-239), and implemented as an online survey. First, participants provided written informed consent to the privacy policy and were randomly assigned to one of the five conditions for the weather app (see Section 2.2). They had time to read the scenario and get familiar with the app interface and subsequently answered a questionnaire including the subjective and objective mental model accuracy, user experience, and explanation satisfaction (see Section 2.4). The questions regarding explanation satisfaction were not shown to the control group since they did not receive an explanation respective transparency cue. The same procedure was then repeated for the sleep tracking app, in that participants first received one of the five app interfaces and then responded to the same measures tailored to the sleep tracking app. In the end, questions about demographic information and covariates were posed (see Section 2.4).
2.4 Measures
Subjective mental model accuracy was measured by a single item (“I have understood the functioning of the app.”) on a 7-point scale from 1 (strongly disagree) to 7 (strongly agree). Objective mental model accuracy was assessed using the nearest neighbor method, 19 in that participants should select the response option that most precisely describes the functioning of the respective app. The four options to choose from varied between the two apps (see Table S1 of the Supplementary Material at https://osf.io/d3g4s/?view_only=967d302afa574792a52b80ee8284eef6). A correct response was coded with 1 and an incorrect one with 0.
As a variable of the experience perspective, user experience was evaluated via the short version of the User Experience Questionnaire (UEQ). 47 , 48 The short version contains eight semantic differential items with seven response options that can be divided into the pragmatic facet (including the perceived perspicuity, efficiency, and dependability) and hedonic facet of user experience (including the perceived stimulation and novelty; Cronbach’s α = 0.82, Cronbach’s αpragmatic = 0.82, Cronbach’s αhedonic = 0.78). For explanation satisfaction, the Explanation Satisfaction Scale 49 was used. One item was removed due to an inapplicability to the described scenarios, and the scale was adapted to the study content (e.g., “system” was replaced by “app”), resulting in six items rated from 1 (strongly disagree) to 7 (strongly agree; Cronbach’s α = 0.92).
Furthermore, covariates that were assumed to influence the results were surveyed. These included the affinity for technology interaction, 50 previous experiences with both apps, attitude towards both apps, frequency of smartphone use, and an attention check. All items were surveyed in German, and their English translations can be found in Table S1 of the Supplementary Material (see https://osf.io/d3g4s/?view_only=967d302afa574792a52b80ee8284eef6).
3 Results
In the following sections, we present descriptive analyses and effects of the transparency cue type, separated for the effectiveness and experience perspective. We conducted a one-way analysis of variance (ANOVA) with the transparency cue as a factor for each of the four outcomes (objective and subjective mental model accuracy, user experience, explanation satisfaction). Subsequently, Tukey’s honestly significant difference (HSD) tests were computed to identify differences between the three transparency cue groups and test the hypotheses. Afterwards, we compared the findings between the two app types, reflecting on possible differences depending on the app domain. For this, we calculated mixed-design ANOVAs using the app as a within-subjects and the transparency cue as a between-subjects independent variable. Also, we conducted analyses of covariance (ANCOVAs) to examine potential influences of interindividual differences.
While we report crucial statistics within the text, all calculations can be accessed via the Supplementary Material (see https://osf.io/d3g4s/?view_only=967d302afa574792a52b80ee8284eef6). They were performed with R and its integrated development environment RStudio. The data and analysis file are openly available in the study’s OSF project (see link above).
3.1 Descriptive analyses
Table 1 shows the descriptive values of all dependent variables. We found positive correlations between the subjective measures, namely user experience, its pragmatic and hedonic facets, explanation satisfaction, and to a lesser degree subjective mental model accuracy. Subjective mental model accuracy showed only a small correlation with objective mental model accuracy, which is in line with previous research 18 , 26 and highlights the relevance of separately investigating both constructs. Objective mental model accuracy did not correlate significantly with any other measures of subjective experience.
Means, standard deviations, and intercorrelations of dependent variables.
| M | SD | 1 | 2 | 3 | 4 | 5 | 6 | |
|---|---|---|---|---|---|---|---|---|
| 1 Subjective mental model accuracy | 6.05 | 1.00 | – | |||||
| 2 Objective mental model accuracy | 0.72 | 0.33 | 0.13* | – | ||||
| 3 User experience (UX) | 4.76 | 0.75 | 0.35*** | 0.04 | – | |||
| 4 UX – pragmatic | 5.36 | 0.96 | 0.45*** | 0.01 | 0.82*** | – | ||
| 5 UX – hedonic | 4.16 | 0.91 | 0.11 | 0.06 | 0.79*** | 0.30*** | – | |
| 6 Explanation satisfaction | 4.61 | 1.26 | 0.48*** | 0.08 | 0.55*** | 0.56*** | 0.35*** | – |
-
Values on all variables, except objective mental model accuracy, were measured on a scale from 1-7. Objective mental model accuracy was measured from 0–2. *p < 0.05, **p < 0.01, ***p < 0.001.
3.2 Effects of transparency cues
3.2.1 Effects of transparency cue on mental model accuracy
For subjective mental model accuracy, a significant effect of transparency cue was found (F(2, 290) = 4.12, p = 0.017). Tukey’s HSD tests showed that the textual cue resulted in a higher feeling of mental model accuracy than the graphic cue (−0.35, 95 % CI[−0.66; −0.04], p = 0.021). In contrast to H1, both cues did not receive higher ratings than the app interface without transparency cue (p = 0.111 for textual cue and p = 0.968 for graphic cue).
The transparency cue also had a significant influence on objective mental model accuracy (F(2, 290) = 12.09, p < 0.001). Again, the textual cue resulted in higher mental model scores than the graphic cue (−0.13, 95 % CI[−0.23; −0.03], p = 0.008). H1 was supported in that objective mental model accuracy was higher for the textual cue group than the no cue group (0.24, 95 % CI[0.12; 0.36], p < 0.001). Nevertheless, the difference between the graphic cue group and the no cue group was not significant, p = 0.073.
3.2.2 Effects of transparency cue on user experience and explanation satisfaction
Transparency cue type did not have a significant effect on user experience overall (F(2, 290) = 0.60, p = 0.549). For the pragmatic facet, we found no significant differences between the cues (F(2, 290) = 1.56, p = 0.212), whereas for the hedonic facet, we did (F(2, 290) = 6.87, p = 0.001). Tukey’s HSD tests indicated that the graphic cue resulted in higher hedonic user experience than the app interface without a transparency cue (0.52, 95 % CI[0.19; 0.86], p < 0.001). However, there was no significant difference between the textual cue group and no cue group (p = 0.076) and the graphic cue group and textual cue group (p = 0.156), not supporting H2.
Although the transparency cues differed regarding explanation satisfaction (F(1, 229) = 28.25, p < 0.001), the results contradicted H3 in that participants were more satisfied with the textual than the graphic cue (−0.84, 95 % CI[−1.14; −0.53], p < 0.001).
3.3 Comparison of apps
Mixed-design ANOVAs for all dependent variables showed that significant interactions of transparency cue and app occurred for subjective mental model accuracy (F(2, 290) = 3.63, p = 0.028) and explanation satisfaction (F(1, 229) = 15.57, p < 0.001). For both variables, the difference between transparency cues was larger for the weather than the sleep tracking app, with the graphic cue resulting in lower ratings than the textual cue (see Figures 2 and 3). Interaction effects were not present for objective mental model accuracy (p = 0.105) and user experience (p = 0.446, ppragmatic = 0.212, phedonic = 0.972).

Interaction of transparency cue and app on subjective mental model accuracy. Means and standard errors are shown for the six conditions resulting from transparency cue (no cue vs. textual vs. graphic) and app (weather vs. sleep tracking app).

Interaction of transparency cue and app on explanation satisfaction. Means and standard errors are shown for the four conditions resulting from transparency cue (textual vs. graphic) and app (weather vs. sleep tracking app), since explanation satisfaction was not measured in the no cue group.
3.4 Analyses of interindividual differences
Considering the covariates, attitude towards the apps had an influence on objective mental model accuracy (F(1, 56) = 10.94, p = 0.002) and on pragmatic user experience (F(1, 56) = 6.04, p = 0.017). Apart from that, just the frequency of smartphone use showed a main effect on pragmatic user experience (F(1, 56) = 4.12, p = 0.047), while all other interindividual differences did not significantly influence the dependent variables (see Tables S4 to S6 of the Supplementary Material at https://osf.io/d3g4s/?view_only=967d302afa574792a52b80ee8284eef6).
4 Discussion
This study investigated how explanatory transparency cues influence users’ understanding of and experience with two mobile apps (a weather app and a sleep tracking app). In detail, we analyzed to what degree textual and graphic cues impact subjective and objective mental model accuracy (effectiveness outcomes) as well as user experience and explanation satisfaction (experience outcomes). The results showed that users had a better subjective and objective understanding of the functioning of an app with textual cues than with graphic cues. Also, textual explanations were more satisfying to users than the graphic ones, but they resulted in similar ratings of user experience. The differences between textual and graphic transparency cues on subjective mental model accuracy and explanation satisfaction were more pronounced for the weather than the sleep tracking app. Furthermore, interindividual differences between participants (e.g., their affinity for technology interaction, experience with apps, frequency of smartphone use) had a minor influence on their app understanding and experience. In the following, we discuss single findings in more detail, reflect on contradictions to previous studies, and highlight limitations and implications for both future research and practical application.
4.1 Advantage of textual cues to users’ mental model
The results support existing models and previous research in that transparency cues had an impact on users’ mental models. 21 , 23 While previous research focused mainly on the objective mental model, 23 , 24 we demonstrate that both the felt and objectively measured understanding of the technology were influenced by the transparency design. Yet, only textual cues showed a positive effect on the participants’ mental models, while graphic cues did not result in significantly different mental model ratings than interfaces without transparency cues. This was against our assumptions since we proposed that both transparency cues should improve the users’ understanding of the apps. From a theoretical perspective, the result could be explained by different functions of graphic and textual explanations: A previous study on mental model formation found that participants consulted textual explanations as an initial “default” option to acquire a general subject understanding, whereas graphic explanations were only additionally reviewed “on demand” if specific knowledge was required. 51
Possibly, our results also depend on the concrete operationalization of the graphic cue, which could have been more difficult for participants to understand quickly than the textual cue. This explanation would align with guidelines developed by Xu et al., 52 stating that graphic cues should only be used if they are easy to understand. Yet, this suggests that differently designed graphic cues could have caused other responses from users.
As an alternative explanation, the ways to convey mental model information (e.g., textual, graphic) might be differently effective depending on the respective user. Thus, our found advantage of textual cues could imply that textual cues are generally well understood by most users in contrast to graphic ones, which might be better comprehensible only for specific user groups. It could also suggest that mental model information might be typically memorized in a verbal modality and that textual cues, which are closer to this modality, may generally convey such information better. Nevertheless, it cannot be stated at this point that textual cues are superior to graphic ones, making it necessary to investigate user responses to different designs of such cues by varying them, for instance, in the time needed for processing.
4.2 No found superiority of graphic cues on explanation satisfaction and user experience
Looking at the experience perspective, explanation satisfaction was significantly higher for textual than graphic cues. This finding contrasts with our expectations and previous evidence from other contexts where users were more satisfied with graphic cues, 13 showing that graphic cues are not per default superior to textual cues. Nevertheless, this may not directly imply that graphic cues cannot result in more satisfied users, since the results on explanation satisfaction corresponded to those on mental model accuracy. It could be that the graphic cues used were perceived as more complex than the textual cues, suggesting that other graphic realizations might lead to opposite effects. According to Kulesza et al., 46 a good explanation is concrete, sound, complete, and not overwhelming. While we aimed to portray these characteristics equally in the textual and graphic explanation, the results hint that our concrete graphic cues might have been more abstract or overwhelming, impairing users’ satisfaction.
Contrarily, the transparency cues did not significantly affect the overall and pragmatic user experience. There were also no differences between textual and graphic cues in terms of hedonic user experience, but graphic cues resulted in higher ratings on the hedonic facet than interfaces without transparency cues. Thus, user experience may be differently influenced by transparency cues than the other dependent variables and may depend more on other design factors. Moreover, against skepticism in previous works, 53 transparency cues appear not to have a negative effect on users’ experience and may even be hedonically useful. However, since interacting with the technology can be important for evaluating user experience, 54 it is also likely that user experience was not measured reliably and that, therefore, no differences in user experience were found.
4.3 Stronger difference between graphic and textual cues in the weather app
Comparing both apps investigated, the results pointed in a similar direction for the weather app and the sleep tracking app. However, the differences between the textual and graphic cues were more pronounced in the weather app than in the sleep tracking app scenario, due to low ratings of the graphic cues on subjective mental model accuracy and explanation satisfaction. This finding may support the above explanation that the graphic cues might have been more difficult for participants to understand immediately than the textual ones, since the graphic presented more complex information for the weather app than the sleep tracking app. For the weather app, a confidence interval of the temperature forecast was shown, whereas the sleep tracking app simply displayed graphs with smileys to make interpretation easier.
Alternatively, Hoffman et al.’s 18 model of the explanation process in explainable AI (XAI) might explain why the results were more pronounced for the weather than the sleep tracking app. The authors stated that users have an initial mental model of a system and that they revise this existing mental model based on transparency cues received. In our study, participants had less experience with sleep tracking apps than with weather apps, suggesting that previously existing mental models might have been more developed for the weather app. This, in turn, could have made them more prone to potential expectancy disconfirmations by the cues, causing the results.
Thus, the stronger differences between transparency cues may be either content-related, pointing at the concrete operationalizations of the cues, or context-related, pointing at differences between the apps. This signals that precise implications are not reasonable at this moment and that the need to search for the sweet spot for transparency design prevails.
4.4 Practical implications
While the inconsistency of our results does not allow for any definitive conclusions yet, some practical implications for app developments can still be derived:
Our findings suggest that the incorporation of transparency cues does not seem to generally have a negative effect on user experience, as discussed in previous research.
More precisely, textual cues were presented as a promising approach to integrate transparency into mobile apps, enhancing users’ mental model and experience.
However, special attention may be required for the detailed design of graphic cues. The fact that graphic transparency cues did not increase users’ mental model underlines the need for carefully and purposefully designed cues that correspond to users’ needs. This may apply especially to the portrayal of complex or multi-dimensional information, where graphic explanations are frequently used 55 but not necessarily promote user understanding.
The context-dependency of our results points to no “one size fits all” solution for enhancing user understanding in mobile apps. This highlights the importance of evaluating each design individually in its specific context to find out which transparency cue conveys relevant information most effectively. Therefore, a target-performance investigation with validated scales (e.g., the TechTra scales 56 ) is essential. Yet, it should be beneficial to investigate general factors that make a specific transparency design more effective (e.g., context characteristics, target group characteristics), as previously done for cross-technology factors influencing users’ transparency perception and need. 57
4.5 Limitations and future research
The study is constrained by several limitations, offering perspectives for future research. First, it was conducted online without real interaction with the app, which possibly impaired participants’ involvement and reduced the external validity of the results. 58 Yet, we chose this approach to enhance the internal validity and obtain meaningful results on the impact of transparency cues that can be further tested in a real setting where users directly interact with a technology and in other contexts (e.g., different app domains). Nonetheless, Babel et al. 59 have found that participants’ evaluation of robot characteristics did not differ significantly between online and laboratory contexts, thus providing evidence that external validity is not always impaired in online research on technology interaction.
Second, a cross-sectional design was applied, assessing short-term reactions of the participants. However, as Hoffman et al. 18 suggest by their explanation process model, effects of transparency cues might alter over time depending on the existing mental model of users. Thus, longitudinal studies are crucial to examine potential changes in users’ responses and identify optimal recommendations for transparency designs.
Third, comparisons between graphic and textual information are inherently challenging due to the amount of information that can be conveyed. For example, the information contained within a single graph might take many sentences to be accurately described in textual form. While we aimed to maximize comparability by keeping the content of explanations as similar as possible across both cue types, an optimal recommendation may be to combine multiple cue formats (e.g., textual and graphic). Previous studies implied that such a combination may yield better results in user understanding than just relying on one format. 60 , 61 Therefore, future studies should include an experimental condition in which the different cue types complement each other.
Fourth, as previously discussed, the results point to different factors that can be relevant for textual and graphic cue design, such as the cues’ complexity. 12 Future research should examine those and search for mechanisms that make some transparency cues more effective than others by implementing, for example, more gradual variations in cue designs (e.g., less complex to complex graphs, texts with neutral wording to evaluative content).
5 Conclusions
This paper investigated the effects of textual and graphic transparency cues on users’ mental models and perception in two mobile contexts, that is, a weather and a sleep tracking app. The findings contribute to the search for a sweet spot in transparency design by suggesting that additional transparency cues may not be detrimental for users’ experience and that textual cues could efficiently enhance mental model accuracy and user perceptions. Against our assumptions, graphic cues were not shown to be advantageous, which could be due to the transparency content being too complex for users to understand quickly. Therefore, graphic cues may only be beneficial under certain circumstances, such as easily understandable information, preserving the need to find the sweet spot in transparency design.
Funding source: Deutsche Forschungsgemeinschaft
Award Identifier / Grant number: 425412993
-
Research ethics: The study was approved by the ethics committee of the Faculty of Mathematics, Computer Science, and Statistics at LMU Munich on January 24, 2024, under the reference EK-MIS-2024-239.
-
Informed consent: Informed consent was obtained from all individuals included in this study.
-
Author contributions: All authors have accepted responsibility for the entire content of this manuscript and approved its submission.
-
Use of Large Language Models, AI and Machine Learning Tools: None declared.
-
Conflict of interest: The authors state no conflict of interest.
-
Research funding: This study was supported by the German Research Foundation (DFG), projects PerforM and TransforM, under grant number 425412993 as part of the Priority Program SPP2199 Scalable Interaction Paradigms for Pervasive Computing Environments.
-
Data availability: The raw data can be obtained from the OSF repository (https://osf.io/d3g4s/?view_only=967d302afa574792a52b80ee8284eef6).
Appendix A: Scenario descriptions
Vignette of the weather app :
Your best friend fulfilled her dream and went on a six-month trip. While she is away, she has entrusted you with the care of her beloved lemon tree. All the plants, including the lemon tree, are on your balcony. They can tolerate slightly colder weather, but not temperatures of 0° or colder. You look at your weather app, which shows you the following weather forecast: (see Figures 4–6)

Interface of the weather app without transparency cue. Text passages translated from German into English.

Interface of the weather app with high reliability: (A) textual and (B) graphic transparency cue. Text passages translated from German into English.

Interface of the weather app with low reliability: (A) textual and (B) graphic transparency cues. Text passages translated from German into English.
Vignette of the sleep tracking app :
You have bought a ring that analyzes your sleep and makes recommendations on this basis. You wore the ring last night. You wake up and look at your smartphone. The sleep app shows you the following recommendation: [see Figures 7–9]

Interface of the sleep tracking app without transparency cue. Text passages translated from German into English.

Interface of the sleep tracking app with high reliability: (A) textual and (B) graphic transparency cues. Text passages translated from German into English.

Interface of the weather app with low reliability: (A) textual and (B) graphic transparency cues. Text passages translated from German into English.
References
1. Miller, T. Explanation in Artificial Intelligence: Insights from the Social Sciences. Artif. Intell. 2019, 267, 1–38; https://doi.org/10.1016/j.artint.2018.07.007.Search in Google Scholar
2. Johnson, H.; Johnson, P. Explanation Facilities and Interactive Systems. In Proceedings of the 1st international conference on intelligent user interfaces – IUI ‘93; Gray, W. D.; Hefley, W. E.; Murray, D., Eds.; ACM Press: New York, NY, USA, 1993; pp. 159–166.10.1145/169891.169951Search in Google Scholar
3. Guidotti, R.; Monreale, A.; Ruggieri, S.; Pedreschi, D.; Turini, F.; Giannotti, F. Local Rule-Based Explanations of Black Box Decision Systems. ArXiv 2018, 1–10; https://doi.org/10.48550/arXiv.1805.10820.Search in Google Scholar
4. Guidotti, R.; Monreale, A.; Ruggieri, S.; Turini, F.; Giannotti, F.; Pedreschi, D. A Survey of Methods for Explaining Black Box Models. ACM Comput. Surv. 2019, 51 (5), 1–42; https://doi.org/10.1145/3236009.Search in Google Scholar
5. Brennen, A. What do People Really Want when they Say they Want “Explainable AI?” We Asked 60 Stakeholders. In Extended Abstracts of the 2020 CHI Conference on Human Factors in Computing Systems; Bernhaupt, R.; Mueller, F.; Verweij, D.; Andres, J.; McGrenere, J.; Cockburn, A.; Avellino, I.; Goguey, A.; Bjørn, P.; Zhao, S.; Samson, B. P.; Kocielnik, R., Eds.; ACM: New York, NY, USA, 2020; pp. 1–7.10.1145/3334480.3383047Search in Google Scholar
6. Vorm, E. S.; Combs, D. J. Y. Integrating Transparency, Trust, and Acceptance: The Intelligent Systems Technology Acceptance Model (ISTAM). Int. J. Hum.-Comput. Interact. 2022, 38 (18–20), 1828–1845; https://doi.org/10.1080/10447318.2022.2070107.Search in Google Scholar
7. Clinciu, M.-A.; Hastie, H. A Survey of Explainable AI Terminology. In Proceedings of the 1st Workshop on Interactive Natural Language Technology for Explainable Artificial Intelligence (NL4XAI 2019); Alonso, J. M.; Catala, A., Eds.; Association for Computational Linguistics: Stroudsburg, PA, USA, 2019; pp. 8–13.10.18653/v1/W19-8403Search in Google Scholar
8. Nunes, I.; Jannach, D. A Systematic Review and Taxonomy of Explanations in Decision Support and Recommender Systems. User Model. User-Adapted Interact. 2017, 27 (3–5), 393–444; https://doi.org/10.1007/s11257-017-9195-0.Search in Google Scholar
9. Hein, I.; Diefenbach, S.; Ullrich, D. Designing for Technology Transparency – Transparency Cues and User Experience. Usability Prof. 2023, 23, 1–5; https://doi.org/10.18420/muc2023-up-448.Search in Google Scholar
10. Mohseni, S.; Zarei, N.; Ragan, E. D. A Multidisciplinary Survey and Framework for Design and Evaluation of Explainable AI Systems. ACM Trans. Interact. Intell. Syst. 2021, 11 (3–4), 1–45; https://doi.org/10.1145/3387166.Search in Google Scholar
11. Gregor, S.; Benbasat, I. Explanations from Intelligent Systems: Theoretical Foundations and Implications for Practice. MIS Q. 1999, 23 (4), 497–530; https://doi.org/10.2307/249487.Search in Google Scholar
12. Kim, S.; Lombardino, L. J. Comparing Graphs and Text: Effects of Complexity and Task. J. Eye Mov. Res. 2015, 8 (3), 1–17; https://doi.org/10.16910/jemr.8.3.2.Search in Google Scholar
13. Wastensteiner, J.; Weiss, T. M.; Haag, F.; Hopf, K. Explainable AI for Tailored Electricity Consumption Feedback – An Experimental Evaluation of Visualizations. ArXiv 2022, 1–19; https://doi.org/10.48550/arXiv.2208.11408.Search in Google Scholar
14. Dai, X.; Keane, M. T.; Shalloo, L.; Ruelle, E.; Byrne, R. M. Counterfactual Explanations for Prediction and Diagnosis in XAI. In Proceedings of the 2022 AAAI/ACM Conference on AI, Ethics, and Society; Conitzer, V.; Tasioulas, J.; Scheutz, M.; Calo, R.; Mara, M.; Zimmermann, A., Eds.; ACM: New York, NY, USA, 2022; pp. 215–226.10.1145/3514094.3534144Search in Google Scholar
15. Das, D.; Nishimura, Y.; Vivek, R. P.; Takeda, N.; Fish, S. T.; Plötz, T.; Chernova, S. Explainable Activity Recognition for Smart Home Systems. ACM Trans. Interact. Intell. Syst. 2023, 13 (2), 1–39; https://doi.org/10.1145/3561533.Search in Google Scholar
16. Vaughn, C.; Sherman-Morris, K.; Brown, M. A Change in the Weather: Understanding Public Usage of Weather Apps. J. Oper. Meteorol. 2023, 11 (11), 140–160; https://doi.org/10.15191/nwajom.2023.1111.Search in Google Scholar
17. Karasneh, R. A.; Al-Azzam, S. I.; Alzoubi, K. H.; Hawamdeh, S.; Jarab, A. S.; Nusair, M. B. Smartphone Applications for Sleep Tracking: Rating and Perceptions About Behavioral Change Among Users. Sleep Sci. 2022, 15 (Spec 1), 65–73; https://doi.org/10.5935/1984-0063.20210007.Search in Google Scholar PubMed PubMed Central
18. Hoffman, R. R.; Miller, T.; Klein, G.; Mueller, S. T.; Clancey, W. J. Increasing the Value of XAI for Users: A Psychological Perspective. KI - Künstliche Intelligenz 2023, 37, 237–247; https://doi.org/10.1007/s13218-023-00806-9.Search in Google Scholar
19. Schraagen, J. M.; Elsasser, P.; Fricke, H.; Hof, M.; Ragalmuto, F. Trusting the X in XAI: Effects of Different Types of Explanations by a Self-Driving Car on Trust, Explanation Satisfaction and Mental Models. Proc. Hum. Factors Ergon. Soc. Annu. Meet. 2020, 64 (1), 339–343; https://doi.org/10.1177/1071181320641077.Search in Google Scholar
20. Visser, E. J. de; Cohen, M.; Freedy, A.; Parasuraman, R. A Design Methodology for Trust Cue Calibration in Cognitive Agents. In Virtual, Augmented and Mixed Reality. Designing and Developing Virtual and Augmented Environments; Springer: Cham, 2014; pp. 251–262.10.1007/978-3-319-07458-0_24Search in Google Scholar
21. Gunning, D.; Vorm, E.; Wang, J. Y.; Turek, M. Darpa’s Explainable AI (XAI) Program: A Retrospective. Appl. AI Lett. 2021, 2 (4), 1–11; https://doi.org/10.1002/ail2.61.Search in Google Scholar
22. Endsley, M. R. From Here to Autonomy. Hum. Factors 2017, 59 (1), 5–27; https://doi.org/10.1177/0018720816681350.Search in Google Scholar PubMed
23. Ramaraj, P.; Kumar, S.; Lasecki, W. Towards Using Transparency Mechanisms to Build Better Mental Models. Advances in cognitive systems: 7th goal reasoning workshop, 2019. https://www.semanticscholar.org/paper/Towards-using-transparency-mechanisms-to-build-Ramaraj-Kumar/386effc2d13e7970aae2873ad9600a417c2f5fb6 (accessed 01 16, 2025).Search in Google Scholar
24. Anderson, A.; Dodge, J.; Sadarangani, A.; Juozapaitis, Z.; Newman, E.; Irvine, J.; Chattopadhyay, S.; Olson, M.; Fern, A.; Burnett, M. Mental Models of Mere Mortals with Explanations of Reinforcement Learning. ACM Trans. Interact. Intell. Syst. 2020, 10 (2), 1–37; https://doi.org/10.1145/3366485.Search in Google Scholar
25. Bemmann, F.; Windl, M.; Erbe, J.; Mayer, S.; Hussmann, H. The Influence of Transparency and Control on the Willingness of Data Sharing in Adaptive Mobile Apps. Proc. ACM Hum.-Comput. Interact. 2022, 6 (MHCI), 1–26; https://doi.org/10.1145/3546724.Search in Google Scholar
26. Andrews, R. W.; Lilly, J. M.; Srivastava, D.; Feigh, K. M. The Role of Shared Mental Models in Human-AI Teams: A Theoretical Review. Theor. Issues Ergon. Sci. 2023, 24 (2), 129–175; https://doi.org/10.1080/1463922X.2022.2061080.Search in Google Scholar
27. Lu, H.; Ma, W.; Wang, Y.; Zhang, M.; Wang, X.; Liu, Y.; Chua, T.-S.; Ma, S. User Perception of Recommendation Explanation: Are Your Explanations what Users Need? ACM Trans. Inf. Syst. 2023, 41 (2), 1–31; https://doi.org/10.1145/3565480.Search in Google Scholar
28. Tuch, A. N.; Bargas-Avila, J. A.; Opwis, K.; Wilhelm, F. H. Visual Complexity of Websites: Effects on Users’ Experience, Physiology, Performance, and Memory. Int. J. Hum.-Comput. Stud. 2009, 67 (9), 703–715; https://doi.org/10.1016/j.ijhcs.2009.04.002.Search in Google Scholar
29. Laugwitz, B.; Held, T.; Schrepp, M. Construction and Evaluation of a User Experience Questionnaire. In HCI and Usability for Education and Work; Holzinger, A., Ed.; Springer Berlin Heidelberg, 2008; pp. 63–76.10.1007/978-3-540-89350-9_6Search in Google Scholar
30. Hoffman, R. R.; Klein, G.; Mueller, S. T. Explaining Explanation for “Explainable AI”. Proc. Hum. Factors Ergon. Soc. Annu. Meet. 2018, 62 (1), 197–201; https://doi.org/10.1177/1541931218621047.Search in Google Scholar
31. King, W. R.; He, J. A Meta-Analysis of the Technology Acceptance Model. Inf. Manag. 2006, 43 (6), 740–755; https://doi.org/10.1016/j.im.2006.05.003.Search in Google Scholar
32. Mlekus, L.; Bentler, D.; Paruzel, A.; Kato-Beiderwieden, A.-L.; Maier, G. W. How to Raise Technology Acceptance: User Experience Characteristics as Technology-Inherent Determinants. Gruppe. Interaktion. Organisation. Zeitschrift für Angewandte Organisationspsychologie (GIO) 2020, 51 (3), 273–283; https://doi.org/10.1007/s11612-020-00529-7.Search in Google Scholar
33. Partala, T.; Saari, T. Understanding the Most Influential User Experiences in Successful and Unsuccessful Technology Adoptions. Comput. Hum. Behav. 2015, 53, 381–395; https://doi.org/10.1016/j.chb.2015.07.012.Search in Google Scholar
34. Shin, D.; Park, Y. J. Role of Fairness, Accountability, and Transparency in Algorithmic Affordance. Comput. Hum. Behav. 2019, 98, 277–284; https://doi.org/10.1016/j.chb.2019.04.019.Search in Google Scholar
35. Bemmann, F.; Stoll, H.; Mayer, S. Privacy Slider: Fine-Grain Privacy Control for Smartphones. Proc. ACM Hum.-Comput. Interact. 2024, 8 (MHCI), 1–31; https://doi.org/10.1145/3676519.Search in Google Scholar
36. Chromik, M.; Eiband, M.; Völkel, S.; Buschek, D. Dark Patterns of Explainability, Transparency, and User Control for Intelligent Systems. IUI Workshops, 2019. https://www.semanticscholar.org/paper/Dark-Patterns-of-Explainability%2C-Transparency%2C-and-Chromik-Eiband/b289d835b8d8d29bbdc8529530a8919a98139e10 (accessed 01 16, 2025).Search in Google Scholar
37. Chaiken, S. Heuristic Versus Systematic Information Processing and the Use of Source Versus Message Cues in Persuasion. J. Pers. Soc. Psychol. 1980, 39 (5), 752–766; https://doi.org/10.1037//0022-3514.39.5.752.Search in Google Scholar
38. Petty, R. E.; Cacioppo, J. T. The Elaboration Likelihood Model of Persuasion. In Advances in experimental psychology; Berkowitz, L., Ed.; Academic Press, 1986; pp. 123–205.10.1016/S0065-2601(08)60214-2Search in Google Scholar
39. Payne, J. W.; Bettman, J. R.; Johnson, E. J. The Adaptive Decision Maker; Cambridge University Press, 2012.Search in Google Scholar
40. Grice, H. P. Logic and Conversation. In Syntax and Semantics Online; Speech, A.; Cole, P.; Morgan, J. L., Eds.; BRILL, Vol. 3, 1975; pp. 41–58; https://doi.org/10.1163/9789004368811_003.Search in Google Scholar
41. Vermeulen, J.; Kawsar, F.; Simeone, A. L.; Kortuem, G.; Luyten, K.; Coninx, K. Informing the Design of Situated Glyphs for a Care Facility. In 2012 IEEE Symposium on Visual Languages and Human-Centric Computing (VL/HCC); IEEE: New York, NY, USA, 2012; pp. 89–96.10.1109/VLHCC.2012.6344490Search in Google Scholar
42. Faul, F.; Erdfelder, E.; Buchner, A.; Lang, A.-G. Statistical Power Analyses Using G*Power 3.1: Tests for Correlation and Regression Analyses. Behav. Res. Methods 2009, 41 (4), 1149–1160; https://doi.org/10.3758/BRM.41.4.1149.Search in Google Scholar PubMed
43. Eisele, G.; Vachon, H.; Lafit, G.; Kuppens, P.; Houben, M.; Myin-Germeys, I.; Viechtbauer, W. The Effects of Sampling Frequency and Questionnaire Length on Perceived Burden, Compliance, and Careless Responding in Experience Sampling Data in a Student Population. Assessment 2022, 29 (2), 136–151; https://doi.org/10.1177/1073191120957102.Search in Google Scholar PubMed
44. Perreault, W. D. Controlling Order-Effect Bias. Public Opin. Q. 1975, 39 (4), 544–551.10.1086/268251Search in Google Scholar
45. Tourangeau, R.; Rasinski, K. A.; Bradburn, N.; D’Andrade, R. Carryover Effects in Attitude Surveys. Public Opin. Q. 1989, 53 (4), 495; https://doi.org/10.1086/269169.Search in Google Scholar
46. Kulesza, T.; Burnett, M.; Wong, W.-K.; Stumpf, S. Principles of Explanatory Debugging to Personalize Interactive Machine Learning. In Proceedings of the 20th International Conference on Intelligent User Interfaces; Brdiczka, O.; Chau, P.; Carenini, G.; Pan, S.; Kristensson, P. O., Eds.; ACM: New York, NY, USA, 2015; pp. 126–137.10.1145/2678025.2701399Search in Google Scholar
47. Schrepp, M.; Hinderks, A.; Thomaschewski, J. Design and Evaluation of a Short Version of the User Experience Questionnaire (UEQ-S). Int. J. Interact. Multimed. Artif. Intell. 2017, 4 (6), 103–108; https://doi.org/10.9781/ijimai.2017.09.001.Search in Google Scholar
48. Juric, M.; Pehar, F.; Pavlović, N. P. Translation, Psychometric Evaluation, and Validation of the Croatian Version of the User Experience Questionnaire (UEQ). Int. J. Hum.-Comput. Interact. 2024, 40 (7), 1644–1657; https://doi.org/10.1080/10447318.2022.2147713.Search in Google Scholar
49. Hoffman, R. R.; Mueller, S. T.; Klein, G.; Litman, J. Metrics for Explainable AI: Challenges and Prospects. ArXiv 2018, 1–50; https://doi.org/10.48550/arXiv.1812.04608.Search in Google Scholar
50. Wessel, D.; Attig, C.; Franke, T. ATI-S - An Ultra-short Scale for Assessing Affinity for Technology Interaction in User Studies. In Proceedings of Mensch und Computer 2019; Alt, F., Bulling, A., Döring, T., Eds.; ACM: New York, NY, USA, 2019; pp 147–154.10.1145/3340764.3340766Search in Google Scholar
51. Hochpöchler, U.; Schnotz, W.; Rasch, T.; Ullrich, M.; Horz, H.; McElvany, N.; Baumert, J. Dynamics of Mental Model Construction from Text and Graphics. Eur. J. Psychol. Educ. 2013, 28 (4), 1105–1126; https://doi.org/10.1007/s10212-012-0156-z.Search in Google Scholar
52. Xu, X.; Yu, A.; Jonker, T. R.; Todi, K.; Lu, F.; Qian, X.; Evangelista Belo, J. M.; Wang, T.; Li, M.; Mun, A.; Wu, T.-Y.; Shen, J.; Zhang, T.; Kokhlikyan, N.; Wang, F.; Sorenson, P.; Kim, S.; Benko, H.X. A. I. R. A Framework of Explainable AI in Augmented Reality. In Proceedings of the 2023 CHI Conference on Human Factors in Computing Systems; Schmidt, A.; Väänänen, K.; Goyal, T.; Kristensson, P. O.; Peters, A.; Mueller, S.; Williamson, J. R.; Wilson, M. L., Eds.; ACM: New York, NY, USA, 2023; pp. 1–30.10.1145/3544548.3581500Search in Google Scholar
53. Deters, H.; Droste, J.; Hess, A.; Klös, V.; Schneider, K.; Speith, T.; Vogelsang, A. The X Factor: On the Relationship Between User Experience and Explainability. In Nordic Conference on Human-Computer Interaction; ACM: New York, NY, USA, 2024; pp. 1–12.10.1145/3679318.3685352Search in Google Scholar
54. Bargas-Avila, J. A.; Hornbæk, K. Old Wine in New Bottles or Novel Challenges. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems; Tan, D.; Fitzpatrick, G.; Gutwin, C.; Begole, B.; Kellogg, W. A., Eds.; ACM: New York, NY, USA, 2011; pp. 2689–2698.10.1145/1978942.1979336Search in Google Scholar
55. Qin, X.; Luo, Y.; Tang, N.; Li, G. Making Data Visualization More Efficient and Effective: A Survey. VLDB J. 2020, 29 (1), 93–117; https://doi.org/10.1007/s00778-019-00588-3.Search in Google Scholar
56. Hein, I.; Ullrich, D.; Diefenbach, S. Development and Validation of the TechTra Scales: A Tool to Assess Perceptions of and Needs for Technology Transparency. In review.Search in Google Scholar
57. Hein, I.; Diefenbach, S. Towards a Comprehensive View on Technology Transparency: A Cross-Technology Investigation of Psychological and Design Factors Around Users’ Transparency Need and Perception. Int. J. Hum.-Comput. Interact. 2025, 1–18. https://doi.org/10.1080/10447318.2025.2502983.Search in Google Scholar
58. Collett, J. L.; Childs, E. Minding the Gap: Meaning, Affect, and the Potential Shortcomings of Vignettes. Soc. Sci. Res. 2011, 40 (2), 513–522; https://doi.org/10.1016/j.ssresearch.2010.08.008.Search in Google Scholar
59. Babel, F.; Kraus, J.; Hock, P.; Asenbauer, H.; Baumann, M. Investigating the Validity of Online Robot Evaluations: Comparison of Findings from an One-Sample Online and Laboratory Study. In Companion of the 2021 ACM/IEEE International Conference on Human-Robot Interaction; Bethel, C.; Paiva, A.; Broadbent, E.; Feil-Seifer, D.; Szafir, D., Eds.; ACM: New York, NY, USA, 2021; pp. 116–120.10.1145/3434074.3447141Search in Google Scholar
60. Hohman, F.; Srinivasan, A.; Drucker, S. M. TeleGam: Combining Visualization and Verbalization for Interpretable Machine Learning. In 2019 IEEE Visualization Conference (VIS); IEEE: New York, NY, USA, 2019; pp. 151–155.10.1109/VISUAL.2019.8933695Search in Google Scholar
61. Myers, B. A.; Weitzman, D. A.; Ko, A. J.; Chau, D. H. Answering Why and Why Not Questions in User Interfaces. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems; Grinter, R.; Rodden, T.; Aoki, P.; Cutrell, E.; Jeffries, R.; Olson, G., Eds.; ACM: New York, NY, USA, 2006; pp. 397–406.10.1145/1124772.1124832Search in Google Scholar
© 2025 the author(s), published by De Gruyter, Berlin/Boston
This work is licensed under the Creative Commons Attribution 4.0 International License.
Articles in the same Issue
- Frontmatter
- Editorial
- Social media and society
- Research Articles
- The role of social media in constructing meaning in life: a SEM analysis
- Sharing knowledge under pressure
- Social media in crisis communication: insights from peace operations on the African continent
- How to analyze cyberbullying on social media platforms
- Silenced voices: social media polarization and women’s marginalization in peacebuilding during the Northern Ethiopia War
- Chilling or resisting? Exploring the influence of technology-facilitated (gender-based) violence on female feminists in Colombia and Costa Rica
- Brief Report
- What kind of technology transparency do users appreciate? Comparison of textual and graphic cues in app design
Articles in the same Issue
- Frontmatter
- Editorial
- Social media and society
- Research Articles
- The role of social media in constructing meaning in life: a SEM analysis
- Sharing knowledge under pressure
- Social media in crisis communication: insights from peace operations on the African continent
- How to analyze cyberbullying on social media platforms
- Silenced voices: social media polarization and women’s marginalization in peacebuilding during the Northern Ethiopia War
- Chilling or resisting? Exploring the influence of technology-facilitated (gender-based) violence on female feminists in Colombia and Costa Rica
- Brief Report
- What kind of technology transparency do users appreciate? Comparison of textual and graphic cues in app design