Stuck with a difficult assignment? No time to get your paper done? Feeling confused? If you’re looking for reliable and timely help for assignments, you’ve come to the right place. We promise 100% original, plagiarism-free papers custom-written for you. Yes, we write every assignment from scratch and it’s solely custom-made for you.
Order a Similar Paper Order a Different Paper
See the attached pdf for article need
Select 2 article and 1 page of review for each article: Total 2 pages
- No specific format
- Should include brief summary, discussion, insights, implications to you, etc.
- Prepare in regular document format (word, pdf).
See the attached pdf for article need Select 2 article and 1 page of review for each article: Total 2 pages No specific formatShould include brief summary, discussion, insights, implications to you,
Full Terms & Conditions of access and use can be found at https://www.tandfonline.com/action/journalInformation?journalCode=hhci20 Human –Computer Interaction ISSN: (Print) (Online) Journal homepage: https://www.tandfonline.com/loi/hhci20 ML Lifecycle Canvas: Designing Machine Learning- Empowered UX with Material Lifecycle Thinking Zhibin Zhou, Lingyun Sun, Yuyang Zhang, Xuanhui Liu & Qing Gong To cite this article: Zhibin Zhou, Lingyun Sun, Yuyang Zhang, Xuanhui Liu & Qing Gong (2020) ML Lifecycle Canvas: Designing Machine Learning-Empowered UX with Material Lifecycle Thinking, Human –Computer Interaction, 35:5-6, 362-386, DOI: 10.1080/07370024.2020.1736075 To link to this article: https://doi.org/10.1080/07370024.2020.1736075 Published online: 30 Apr 2020.Submit your article to this journal Article views: 738View related articles View Crossmark dataCiting articles: 1 View citing articles ML Lifecycle Canvas: Designing Machine Learning-Empowered UX with Material Lifecycle Thinking Zhibin Zhou a,b,c , Lingyun Sun a,b,c , Yuyang Zhang b,c, Xuanhui Liu a,c, and Qing Gong a,b,c aInternational Design Institute, Zhejiang University, Hangzhou, China; bKey Laboratory of Design Intelligence and Digital Creativity of Zhejiang Province, Hangzhou, China; cInternational Design Institute, Zhejiang University, Hang Zhou, China ABSTRACTAs a particular type of artificial intelligence technology, machine learning (ML) is widely used to empower user experience (UX). However, designers, especially the novice designers, struggle to integrate ML into familiar design activities because of its ever-changing and growable nature. This paper proposes a design method called Material Lifecycle Thinking (MLT)that considers ML as a design material with its own lifecycle. MLT encourages designers to regard ML, users, and scenarios as three co-crea- tors who cooperate in creating ML-empowered UX. We have developed ML Lifecycle Canvas (Canvas), a conceptual design tool that incorporates visual representations of the co-creators and ML lifecycle. Canvas guides designersto organize essential information for the application of MLT. By involving design students in the “research through design ”process, the development of Canvas was iterated through its application to design projects. MLT and Canvas have been evaluated in design workshops, with completed propo- sals and evaluation results demonstrating that our work is a solid stepforward in bridging the gap between UX and ML. ARTICLE HISTORYReceived 16 June 2019 Revised 25 February 2020Accepted 25 February 2020 KEYWORDSDesign method; machinelearning; user experience; design tool 1. Introduction As a particular form of artificial intelligence (AI) technology, machine learning (ML) not only provides opportunities for the user experience (UX) community , but also raises several desi gn challenges because of its malleable nature (Gillies, Fiebrink, Garcia, et al., 2016 ; Kuniavsky & Churchill, 2017 ). ML can assist designers in extracting patterns from complex information (Witten et al., 2016 ), such as understanding semantics from speech (Deng & Li, 2013 ). Thus, ML can be used to empower the UX of various interactions. However, the intelligence of ML comes from the learn ing of massive labeled datasets like audio clips annotated with semantic tags. The learning/developm ent process of ML involves the annotation of data, the construction and training of an ML model, predicti on, and so on. This constant growth means that the performance of ML might evolve across the learning p rocess.Thus,wethinkofMLasbeinglikeagrowable “tree ”instead of “wood ”with fixed attributes. The static “wood ”is easily carved into a desired form. When faced with a “tree ”, however, the designers have to consider its entire lifecycle and growth process. The malleable nature of ML hinders designers, especially novice ones, from using it in their regular design activities (Dove et al., 2017 ;Yang, 2017 ). To cultivate the desired UX, a designer is supposed to behave like a qualified gardener who would influence the growth ofthetreebyfertilizing ,trimming,andotherwise caring for it. Designers are supposed to understand the growth process of ML and know how to create the final UX with proper resources and treatment. For example, the training data could be regarded as the “fertilizer ”that promotes the growth of ML, the application scenario could provide the required “sunlight and physical environment ”, and iterative training could be considered as the process of “trimming the CONTACT Lingyun Sun [email protected] Key Laboratory of Design Intelligence and Digital Creativity of Zhejiang Province, Hangzhou, China © 2020 Taylor & Francis Group, LLC HUMAN –COMPUTER INTERACTION 2020, VOL. 35, NOS. 5 –6, 362 –386 https://doi.org/10.1080/07370024.2020.1736075 leaves ”that crafts the performance of ML. The growth of ML is closely related to users and scenarios, so these three factors should be considered simultaneously following a specific method in the conception phase. However, there are currently few design methods that assist novice designers in fully understanding ML technology, users, and application scenarios to addr ess these development issues in UX design. Firstly, UX designers lack a technical background and are not well prepared to leverage ML capabilities (Dove et al., 2017 ), because ML attributes such as accuracy and speed constantly change during the development process. Secondly, the potential value of users is ignored to improve the performance of ML and bring users better experience (Zhu et al., 2018 ). When faced with the unfamiliar growable nature of ML, designers are unable to involve the available user inputs in the development process of ML (De Choudhury & Kiciman, 2018 ; Yang, Suh et al., 2018 ). Thirdly, designers are confused about the application scenarios of ML. For example, designers need to consider the restrictions of the envi ronment, laws, and regulations when proposing an ML solution in the real world (Amershi, Weld et al., 2019 ). As a result of the above challenges, designers rarely envision how ML might empower UX within current conceptual design activities, which leads to UX issues remaining unresolved. Therefore, designing ML-empowered UX demands methods that support designers in consider- ing the malleable nature of ML, users, and the scenario at the conceptual design stage. Although typical design concepts like service design thinking (Bitner et al., 2008 ) illustrate how all stake- holders/co-creators can be taken into account, conceptual design methods for empowering UX with ML should be tailored to the attributes of ML. This is because the malleable nature of ML makes it quite complex to consider all co-creators (ML, user, scenario) in a dynamic development process. Designers are supposed to participate in such development processes with design thinking that they are familiar with (Rozendaal et al., 2018 ; Yang, 2017 ). However, existing design methods barely consider the malleable nature of the material, which prevents designers from taking advantage of the familiar design thinking to conceive design proposals for ML-empowered UX. Our work aims to incorporate the development process of ML into typical conceptual design activities in which the designers systematically consider ML, use rs, and the scenario as a whole. As the performance of ML can evolve during its development process, the more involvement a designer has with this process, the more likely the design proposal of ML-empowered UX i s to suit the demands of scenarios and users. To this end, we present the basic concepts and mechanism of ML to designers instead of offering a black-box. We also adapt the concepts from design thinking (e.g., service design, mate rial driven design) to integrate the existing research on ML –human interaction, helping designers t o understand the malleable nature of ML and to consider ML, users, and the scenario in a holistic way. In this paper, we propose a design method called Material Lifecycle Thinking (MLT) that regards ML as a constantly growing material with an iterative “lifecycle. ”MLT involves designers in the development process of ML with the consideration of co-creators who cooperate in creating the UX. We also present ML Lifecycle Canvas, a conceptual design tool derived from MLT mainly for novice designers. The evaluation and reflection of MLT are conducted by applying Canvas to design activities. The Canvas creates a visual schematic incorporating the perspectives of ML, users, and the scenario, detailing the ML lifecycle from data annotation to ML model update. For this purpose, we first study the characteristics of ML by identifying associated UX issues and summarizing the related factors through a literature review (see Figure 1 ). We then present a prototype of Canvas and Figure 1. Roadmap of our work. HUMAN –COMPUTER INTERACTION 363 iterate it according to the collected feedback. The final version of Canvas is compared to a typical conceptual design tool in workshops involving 32 participants. Analysis of completed design proposals and evaluation results reveals that our system is effective in enhancing the understanding of ML and tackling ML-related UX issues. Our contribution consists of three parts. First, the proposed MLT method encourages designers to reflect on the whole lifecycle of a malleable material like ML and cultivate it into the desired form by considering the co-creators. Second, we developed Canvas, a conceptual design tool for applying MLT and inspiring novice designers to conceive ML-empowered UX proposals without extensive ML skills. By representing ML, users, and the scenario in a visual way, the interplay between them can be charted and analyzed throughout the entire ML lifecycle. Finally, we share the lessons learned in developing Canvas, which may offer insights for future research on the integration of other AI technologies with UX design. 2. Related work As an important branch of AI, ML is defined as a series of computer programs that learn from experience and improve performance without using explicit instructions when performing a specific task (Koza et al., 1996 ; Mitchell, 1999 ). A typical ML system is a data-driven deep neural network that includes data annotation, neural network building, model training, and model prediction. In recent years, the broad application of ML has promoted considerable research on democratizing ML in the field of design. 2.1. Lowering the barrier to ML for novice designers Existing practices aimed at the democratization of ML have provided programming tools, educa- tional material, and courses for novice designers, enabling them to understand basic concepts and practically construct ML prototypes. Current programming tools assist novice designers in creating ML prototypes with simple code. These tools can be categorized as MLaaS (machine learning as a service platform) (Ribeiro et al., 2016 ), such as IBM Watson; open-source toolkits like TensorFlow (Abadi et al., 2016 ) that require programming skills; and non-programming tools including Yale and Wekinator (Fiebrink & Cook, 2010 ). Additionally, hardware platforms like AIY (Google, 2019b ) provide easy-to-use tools to assemble physical ML artifacts. Platforms such as Google ’s autoML (Google, 2019a ) or IBM ’s autoAI (IBM, 2019 ) help engineers and designers to easily construct or train high-quality ML models according to their needs. These facilitate designers in prototyping through appropriate simplification of the abstract technical details and processes of ML. However, they do not really help designers with little technical background to understand the mechanisms of ML and ideate ML- empowered UX when faced with the unfamiliar growable ML technology. Existing practice in ML-related design courses provided us with possible course module settings (McCardle, 2002 ), design processes (Rozendaal et al., 2018 ), and a tangible tool platform (Vlist et al., 2008 ). Moreover, current educational material (Hebron, 2016 ) and text books for developing statistics literacy (Kruschke, 2011 ; McElreath, 2018 ) guide designers toward a preliminary under- standing of ML ’s capabilities and working mechanisms. Although these practices and tools inspire interesting design cases, there has been little attempt to understand the malleable nature of ML and consider the interests of the stakeholders/co-creators (i.e., users and scenarios). The studies mentioned above lower the barrier to accessing ML and simplify the workflow of manipulating ML technology. However, these practices do not focus on the conceptual design, which is indispensable for novice designers, and do not incorporate users and scenarios. Thus, there is a need for ML-related design methods and tools that support conceptual design activities while considering both users and scenarios. 364 Z. ZHOU ET AL. 2.2. Existing conceptual design tools Conceptual design is a series of activities that outline and draft the key characteristics of an experience (Atasoy & Martens, 2011 ; Hanington & Martin, 2012 ). An essential part of the early design process is to initiate creative reflection and plan subsequent phases (Kim & Ryu, 2014 ). To envision hypothetical experiences, designers tend to create novel assemblies of known things and conduct reflective con- versations with design materials, users, and scenarios (Schon & Desanctis, 2005 ). Typical conceptual design tools (Tschimmel, 2012 ) used in the conception phases cover empathy, problem definition, idea generation, and experimentation (see Figure 2 ). Some of them are used in virtually every stage of design. The above tools work well in terms of understanding users and scenarios when designers have a tacit understanding of the applied technologies, but they are not tailored to the unique malleable nature of ML. Hence, they might fail to guide designers to gather crucial ML-related information and conduct subsequent conceptual design activities. To overcome this deficiency, using the above foundation for transforming complex information into design insights through a visual approach (Snyder et al., 2014 ), we modify the existing conceptual design tools in the context of integrating ML and UX. 2.3. Research on integrating ML and UX Due to the malleable nature of ML, misalignments occur between the function and interpretation of ML systems (Baumer, 2017 ), which affects UX. Most designers know little about how a given technology functions and only use what they do know to extrapolate possibilities (Bucher, 2017 ), which might not align with the real situation. Thus, recent research integrating ML and UX regards ML as a design material and identifies its main challenges (Dove et al., 2017 ) and design values (Yang et al., 2018 ). Several theoretical attempts have been made to apply design thinking concepts like human- centered design with the aim of seeking novel design opportunities about ML. Some studies advocated “power to the people ”(Amershi, Cakmak et al., 2014 ; Hand, 1998 ; Patel et al., 2008 ) and pursued human-centered ML (Gillies, Fiebrink, Tanaka et al., 2016 ) and interactive ML (Yang, Phase Tool Brief summary Empathy: gather substantial information to gain an empathic understanding of the user and scenario Personas A tool to create a fictional character whose attributes can represent a group of users. It can make the abstract idea of the user more personal and reveal insights. (Pruitt & Adlin, 2010) Problem Definition: to identify the core problems by analyzing the gathered information User Journey Map A visual representation of a user’s journey through a service, showing all the touchpoints in the journey. Touchpoints are points of interaction where the user conducts transactions in the service, exchanges informatio n with products, and so on (Stickdorn et al., 2011) . Idea Generation:envision solutions based on the insights gained in the first two stages Brainstorm A collaborative tool whereby participants mainly generate ideas based on emotions and intuition ra ther than rati onal thinking. The main objective is to produc e a large number of ideas in a short time Experimentation: evaluate alternative solutions Service Blueprint A visualization tool that show s the relationship between all stakeholders, including the users, service offers, and so on. It consists of more stakeholders than just the user, so it can bridge cross-department efforts and identify opportunities (Bitner et al., 2008) . Figure 2. Typical conceptual design tools. HUMAN –COMPUTER INTERACTION 365 Suh et al., 2018 ). For instance, Amershi, Cakmak et al. ( 2014 ) and Wood ( 2014 ) allowed users to freely modify the personalized labels so that ML systems could better understand user-preferences. The Google UX community (Jess, 2017 ) encourages designers to involve users in the ML process as a way of gaining design insights. Other recent research has drawn inspiration from the typical design process (Rozendaal et al., 2018 ; Yang, 2017 ). Yang, Scuito et al. ( 2018 ) demand that designers identify the problem that ML is intended to solve, validate the technical feasibility, and then iterate the solution to craft the ML functionality. Different design strategies have also been proposed to improve the UX of malleable, unpredictable system. The theoretical strategy applies numerous theories from behavioral and social sciences to help understand how users behave (Murnane & Counts, 2014 ; Recasens et al., 2013 ). The speculative strategy involves extrapolating from existing circumstances to imagine possible futures and is less constrained by technical feasibility (Baumer, Ahn, Bie et al., 2014 ; Dunne & Raby, 2013 ). The participatory strategy involves people who might be impacted by the system in the design process (Kuhn & Muller, 1993 ; Muller, 2007 ). Eric Baumer ( 2017 ) developed a human-centered algorithm design process that incorporates the above strategies. Moreover, some design guidelines have investigated specific ML issues, such as how to design interactions that may produce unpredictable results. Amershi et al. ( 2019 ) developed 18 human –AI interaction design guidelines that are organized into categories based on when the user interacts with AI. These guidelines require designers to clarify the capabilities of AI, the context, relevant social norms, and so on. Additionally, Google ( 2019c ) provided suggestions for specific UX issues such as trust, failure, and data collection. Although the above guidelines are helpful, they are not well connected to the theoretical explorations or design strategies that inspire designers to make trade-offs between different guide- lines based on the real-life situation and to explore other feasible possibilities. To bridge this gap, designers require a methodological framework that allows them to use practical design guidelines while following the abovementioned design thinking concepts and strategies. 2.4. Summary Existing prototyping tools and design education have certainly lowered the barrier to ML. However, as current conceptual design tools are not tailored to the ML context, there is still no design method or tool that assist designers in ideating novel and practical approaches using ML. Moreover, recent research has suggested that there is a gap between detailed guidelines and theoretical recommenda- tions like “power to the people ”and “retain the typical design process ”. Moreover, there is a lack of research on encouraging designers to participate in the growing process of ML with design strategies and consider ML, users, and the scenario as a whole. Therefore, what is required is a method that can bridge the gap between theoretical foundations and detailed guidance. To construct such a method, we need to seriously consider the malleable nature of ML and holistically incorporate the demands of the users and scenarios. 3. Material lifecycle thinking In this section, we introduce the MLT conceptual design method, which considers malleable tech- nologies like ML as design materials with iterative lifecycles. MLT also encourages users and scenarios to collaborate with materials to co-create UX. In this way, designers are able to shape ever- changing materials through user engagement and scenario arrangement, rather than simply using the material as a static black box. 366 Z. ZHOU ET AL. 3.1. Why MLT? We originally intended to apply the “material driven design ”(MDD) (Karana et al., 2015 ) method to study ML as a literal material. MDD allows designers to firstly characterize the material both technically and experientially, and then envision the material ’s role when embodied in a product. The designers then imagine how the users experience the material, rather than using intuition. Finally, the designers integrate all findings into the final product concept. MDD and similar design methods work well with static physical materials whose characteristics might not change significantly over time. However, such methods ignore other resources involved in the development process and would be constrained when dealing with ever-changing materials. An ideal method for a malleable material should guide designers to understand the basic mechanisms of the material ’s lifecycle and the factors that influence its growth. MLT draws appropriate inspiration from existing design methods, as recent studies have suggested that the design method should not change significantly in the context of ML (Rozendaal et al., 2018 ; Yang, 2017 ). Essentially, it is an extension of existing design methods such as MDD, human-centered design, value-sensitive design, and green design. Most of these methods focus on design materials with fixed attributes, which can not cope with the challenges caused by the malleable attributes of ML. In contrast, MLT is tailored to malleable design ma terials and focuses on the development process of materials. For example, the idea of “lifecycle ”was borrowed from green design, such as lifecycle assessment (Russell et al., 2005 ), to assess the environmental impacts associated with all stages of a product ’slife,from raw material processing to maintenance and recycling. In MLT, the concept of “lifecycle ”implies that the material is “growable ”and describes the development process of a material whose attributes vary over different evolutionary cycles. Regarding ML, lifecyc le refers to the specific learning process because the performance of ML can change on each iteration, inevitably affecting the UX. 3.2. What is MLT? MLT encourages designers to participate in the main development process of the material while considering the users and scenarios involved in creating UX. Figure 3 illustrates MLT, with the five Figure 3. Brief introduction to MLT. HUMAN –COMPUTER INTERACTION 367 main steps presented in a cyclic manner. The steps are as follows: (1) Understanding the technical and experiential characteristics of the material, which helps to describe the design goal; (2) Identifying the co-creators and understanding their technical and experiential characteristics. Here, the term “co- creators ”is used to describe subjects that cooperate with the material to create UX during its lifecycle. The material itself can also be regarded as a co-creator; (3) Identifying touchpoints where the co- creators exchange resources with each other and have an impact on the material ’s development; (4) Envisioning possible UX by considering the trade-offs between co-creators with various characteristics to gain design insights, embodying the insights through the touchpoints of material for generating potential UX solutions; (5) Monitoring the material, as its characteristics might evolve because of the mutual influences among the co-creators and result in different UX. We use the breeding process of potted plants as an example to explain how MLT deals with the growable material (see Figure 3 ). The process starts with a malleable material (i.e., ML) and ends with a cycle that continually improves UX. In the case of ML-empowered UX, we regard the ML, users, and scenarios as co-creators with different characteristics and interests. We consider ML as a malleable material that requires the cooperation of users and scenarios to support UX. During each loop of the ML lifecycle, a human user “teaches ”ML through interaction inputs and their preferences, while the scenario provides resources to support the ML performance and brings restrictions. Designers gain design insights by resolving the conflicts resulting from the co-creators ’own demands. This process regards ML and the scenario as synergistic partners. The objective of MLT is to help designers propose feasible and suitable conceptual UX solutions, without requiring them to practically implement ML systems. 4. Development of ML lifecycle Canvas The ML Lifecycle Canvas, a conceptual design tool featuring the holistic visualization of cooperation among ML, users, and scenarios during the ML lifecycle, is intended to enhance design practices with the MLT method. The development process includes the “Research Through Design ” (Hanington & Martin, 2012 ) approach, which is a reflective process consisting of conception, prototype, evaluation, and iteration to discover insights that will subsequently be incorporated into the Canvas. As well as conducting a literature review, we invited designers to contribute their ideas for improving the Canvas in a participatory way. 4.1. Phase 1: identify the design issues and key factors The first step in applying MLT to grow ML-empowered UX is to understand ML as a malleable material. We identified the characteristics of ML by reviewing the literature on ML-related design challenges and influencing factors. 5. Design issues Using search terms such as “AI, ”“ ML, ”“ UX, ”and “challenges, ”we collected recent (past three years) publications that mentioned ML-related UX issues from the field of human-computer inter- action (HCI). We set the scope to the past three years because of time and resource constraints, but the summary of previous work in state-of-the-art publications also covered earlier work and offered inspirations. We searched in the HCI community instead of the broader computer science domain because almost all work in HCI has some relationship to people (though not only related to ML), and is thus connected to UX design (Yang et al., 2018 ). From the work of Holmquist ( 2017 ); Yang, Scuito, et al. ( 2018 ); Forlizzi, Zimmerman, Mancuso et al. (2007 ); Dove et al. ( 2017 ), and our previous work (Zhou, Gong et al., 2019 ), which cover most UX issues related to ML, we identified potential detailed challenges. Some of these issues were specific to a particular scenario, e.g., the distrust of automotive driving and unpredictable changes in adaptive UIs. These issues were extended to the universal scenario, e.g. the distrust to AI systems. Three of the 368 Z. ZHOU ET AL. authors extracted the above-mentioned issues from the collected publications, transformed descriptions of these issues into post-it notes and pasted them on the board to construct an affinity diagramming process. They sorted the descriptions into different clusters, and named each cluster with the most representative description. After an iterative process of reorganizing and merging, we obtained six clusters, each representing one of the typical issues listed in Figure 4 . Each issue represents several associated topics. For instance, transparency represents several concerns including explainable AI (Wang et al., 2019 ), the interpretability of results, and the visualization of decision-making processes. The factors that may influence these UX issues were determined through a literature review of HCI publications and social, cultural, and psychological studies. We also reviewed current work on design guidelines for human –AI interaction (Amershi, Weld et al., 2019 ; Google, 2019c ). After the same affinity diagramming process, we identified the nine key factors (excluding the user category) listed in Figure 5 . We divided the factors into the categories of ML and scenario. Regarding the user category, existing conceptual design toolboxes provide tools such as Personas and Empathy Maps to collect user demands and preferences. 5.1. Phase 2: conception of Canvas To apply MLT in conceptual design, we crafted Canvas to present a visual form. Canvas enables designers to complete the conceptual phase of t he design according to MLT in a cyclic manner. The conception of Canvas according to the fi ve specific steps of MLT is illustrated in Figure 6 . The first step of MLT is to understand the attributes of ML. Thus, we conceived Issue cards to represent the six specific issues caused by ML (see Figure 6a ), which can be considered as the material attributes. The factors related to ML (see Figure 5) offer tips for understanding the material and describing the design goal. Secondly, Canvas uses different rows to represent the co-creators (ML, user, and scenario; see Figure 6b ). The relationships among the three co-creators are somewhat ambiguous, but we clearly assigned distinct parts to each co-creator to avoid pre-defining relationships and restricting the flexibility of Canvas. Canvas does not intend to imply any relationships among the three co-creators, leaving designers with the freedom to mine potential relationships using the advantage of MLT. We used different approaches to help designers understand these co-creators. In particular, we custo- mized the Persona to promote empathy with users. Our Persona format contains the pain points, needs, and behavioral demographic information related to ML, which aids user studies in an ML Issues Brief summary Unpredictability The issue of managing the unpredictable conclusions given by ML systems according to given data no matter how well the ML system is trained. Transparency The issue of designing the product whose workings nobody can quite explain. And how does this affect qualities like trust and confidence in the system? Learning The issue of using available user input to not only evolve the performance of ML with unobtrusive interaction but also avoid the bias in data. Control The issue of designing ML systems that allow the co-control with the user, so that users can truly cooperate with ML system in decision making. Anthropomorphism The issue of determining the anthropomo rphic degree that can impact on how people perceive the intelligent system like perceived task suitability, engagement with agents, and perceived attractiveness. Interactivity The issue of supporting the interaction between user feedback and newly learned ML models. Many ML models are built to offer improvements in accuracy and may not run quickly enough for real-time interaction. Figure 4. Specific UX issues caused by ML. HUMAN –COMPUTER INTERACTION 369 context. As for the scenario and ML system, the related factors in Table 3 provide guidance for determining the questions worth considering. In stage 3, the different columns of Canvas are used to describe different touchpoints. The touchpoints at which the users, scenario, and ML system exchange resources are shown in Figure 6c . We used the stages of the ML process (i.e., data annotation, training, inference) to pre- define the touchpoints, reducing the cognitive burden on the designers. The most important step is to envision the potential UX and embody the design insights about users and scenarios in specific steps of the ML lifecycle (touchpoints). To this end, the information collected in stages 1 –3 is organized in a visual form in Canvas to enable the identification of design insights (see Figure 6d ). The three co-creators have their own goals and optimal interests, but sometimes they have to make compromises and concessions for the final experience. For example, acquiring more data can certainly help to train a better-performing ML application, but the data size might be limited by the user ’s willingness and regulations in the scenario. As discussed above, the mutual relationships of the co-creators vary greatly with the type of ML system, the user, and the scenario. Therefore, Canvas provides a template guiding designers to think about issues worth considering for such ambiguous relationships, instead of clearly mapping the factors in Table 3 to specific issues or touchpoints. Figure 6. Conception of Canvas. Category Brief summary of the factors Scenario Societal impact: Aspirations concerning culture, poli tical system and safety. (Cuevas et al., 2007) Ta s k t y p e s : Cognitive type, control type and perceptual aids type, etc. Physical environmen t:Light, humidity, noise, location etc. ML system Automation level: Predictability and stability of automation (Merritt et al., 2013) Capability: Types of tasks that the ML system can support. Aesthetics: Aesthetic appearance, which is related to automation level and instructions (Li & Yeh, 2010). Co-control Mode: Fixed, adjustable (e.g., set by the us er), or adaptive (e.g., accounts for user’s state) (Hancock & Chignell, 1988). Hardware: Internet, power supply (e.g., battery), computing capability (e.g., GPU), etc. Learning ability: Amount and quality of required data, the difficulty of iterative training, and the required computing resources etc. Figure 5. Factors influencing the scenario and ML system. 370 Z. ZHOU ET AL. Finally, the attributes of ML might change after each loop of the lifecycle following the involve- ment of the users and scenario. The design process supported by Canvas is iterative, and designers thus need to pay close attention to the evolution and growth of the ML system to revise the design proposal accordingly. 5.2. Phase 3: Canvas Prototype The Canvas Prototype consists of Canvas I, Canvas II, Issue cards, and Persona cards. Their usage is illustrated in Figure 7 . To seek potential improvements, the prototype was used in the conceptual phase of a design project, with essential technical support provided. 6. Participants Two instructors and 30 junior (third-year) undergraduate design students majoring in industrial design (M: 17, F: 13) were involved in a design project with the theme of “ML-empowered product design ”(see Figure 8 ). The project was part of a compulsory course and students received credit for the course. We selected the students who had long-term (at least 3 months) internship experience and had done at least 2 practical design projects to ensure that they could be regarded as novice designers. These design students had been trained in the same way. They had already mastered mainstream design tools, but lacked a deep understanding of ML, which helps us to obtain objective results. The participants were split into six groups, with the two instructors offering guidance to each group as required. The instructors explained the usage of Canvas and offered technical-related suggestions, but were not responsible for the design process. Figure 7. Prototype of Canvas. Figure 8. Working with the Canvas prototype. HUMAN –COMPUTER INTERACTION 371 7. Procedure First, the participants were introduced to UX issues, related factors, and the usage of Canvas. According to the topic assigned to each group, they then gathered essential information through user studies and online research. Furthermore, the participants entered the information into Canvas and discussed the trade-offs in the optional solutions to serve the application scenario and users. Finally, Canvas was used to exchange and evaluate design ideas. Building on the design tools discussed in Section 2 and other research on developing creative tools (Atasoy & Martens, 2016 ; Jacobs et al., 2017 ), we developed four evaluation criteria. Each participant took part in a 10 –15 minutes semi-structured interview. During the in-person interviews, we asked the laddering questions (Reynolds, & Gutman, 1998 ) that were adjusted for deeper investigation according to the participant’s answers. . Laddering refers to an in-depth interviewing technique for selecting the most appropriate level of question or researcher response to respondent dialogue, based on the premise that we arranged our questions in an order that starts with the least invasive and proceeds to deeper matters. The following were part of the laddering questions. Ease-of-use: Do you think the tool is easy to use? Why do you feel that it is (or is not) easy to use? What elements do you think are important for an easy-to-use tool? Why is this important to you? Creativity: Do you think the tool helps you to come up with diverse ideas? Why do you feel that it is (or is not) helpful for creativity? What kind of ideas does the tool help you to create? Why is the tool helpful for creating the mentioned ideas? Process: Do you think the tool helps you to integrate prior expertise in the design process? Why can you (or can you not) integrate prior expertise? Why do you feel that it is (or is not) helpful in the design process? How does this tool relate to the traditional design process? Understanding: Do you think the tool helps you improve your understanding of ML? Do you think the tool encourages you to think about issues and factors related to ML? What do you feel about that? What are the critical aspects of the tool in helping your understanding of ML? 8. Results At the end of the project, all six groups produced concept designs varying from modular devices for connecting distant families (Zhou, Jiang, et al., 2019 ) to a mobile app for assisting children in drawing comics. A completed Canvas is shown in Figure 9b , including detailed information and solutions for Cabe (see Figure 9a ), an AI robot that senses changes in emotion and reacts appro- priately to help users achieve a better learning experience. How the final version of Cabe dealt with the UX issues has been discussed in our previous work (Zhibin Zhou et al., 2019). We applied a content analysis process (Howitt & Cramer, 2010 ) to summarize the interview results. We took notes during the interviews and tagged them with the eight pre-defined codes related to the Figure 9. Cabe: one of the design outcomes of the six groups. 372 Z. ZHOU ET AL. strengths and weaknesses of four evaluation criteria. We used the audio recordings of interviews to avoid missing important information. One of the authors who is most familiar with the notes conducted the coding and other analysis processes. For example, “the cyclic nature of the ML lifecycle is not obvious in the Canvas (participant 5) ”was assigned the code of “negative comment to under- standing criterion ”while “it helps me learn more about the ML concepts and its process (participant 16) ”was assigned the code of “positive comment to understanding criterion. ”The coder was respon- sible for the whole project and designed all the questions for the interview, as well as taking the notes during the interviews. Thus, we considered him as the most suitable person for the coding task to ensure that the analyzer of the data understood the context of the entire conversation. From the analysis, we found that Canvas helped the majority of participants gain an increased understanding of ML-related factors and UX issues and imagine appropriate solutions. To identify the weaknesses of Canvas, the main negative comments are summarized in Figure 10 . 8.1. Phase 4: iteration of Canvas On the basis of the evaluation results, we made the following improvements to the Canvas. These improvements mainly addressed issues including overly simplified visualization layout, ML process (touchpoints) involving technical terms, and ambiguous factors. ● We transformed the tabular visualization layout implying a linear ML process into a iterativeCanvas is split into six fan-like sections, one to visually suggest a recyclable and realistic ML process. ● We combined Canvas I and Canvas II, isolated the Persona cards from Canvas, and rearranged its format to evoke empathy because users do not change their demands significantly at different stages of ML process. We also presented completed Canvas designs as examples and provided instructions that introduce the usage of Canvas. Evaluation criteria Brief summary Ease-of-use 1) Canvas had many ML technology terms and details that hinders participants from understanding touchpoints and the stages of ML process; 2) Participants learn little Creativity about the usage of Canvasthrough the Creativity 1) Some aspects of ML and scenario ar e ambiguous and complex so it was difficult for participants to ask appropriate questions; 2) Canvas offers large quantities of information combining ML technology with UX, while participants are accustomed to thinking about the design issues without much technical detail. Process 1) Canvas was kind of inconsistent with similar design tools because it was split into two parts; 2) Some participants complained that th e complexity of factors and ML process prevented them from ffreely using Canvas Understanding in design process. Understanding 1) ML process in Canvas was still quite abstract, especially when participants wanted to match the ML process to the corre sponding after stage o f product lifecycle; 2) The factors regarding the ML and scen ario were ambiguous so the participants might have different understandings; 3) Some participants ignored the cyclic nature of the ML process because the linear form of Canvas assumes a simple linear ML process. information transmitted by the tool itself without the explanation of attached cases. Figure 10. Interview results of the prototype Canvas. HUMAN –COMPUTER INTERACTION 373 ● We modified the description of touchpoints to make them less like the terminology of ML. We combined the data annotation and dataset processing stages into data collection and split the original inference stage into prediction and performance. ● We provided a Question list instead of an ambiguous description of factors, as it was difficult for designers to understand the given factors. In particular, the most typical questions of Question list were extracted from the information of the completed Canvas in phase 3 and existing guidebooks on human-AI interaction (Amershi, Weld et al., 2019 ; Google, 2019c ). The Question list contains problems worth considering in the conceptual phase. We provided a list because the evaluation showed that designers struggled to carry out the design process without explicit instructions. However, some questions might have implications for the ambiguous relation- ships among the three co-creators. Additionally, this list does not cover all aspects of ML technology. It serves as a self-checking list that can be freely modified whenever necessary, and designers are not required to consider every question. The aim is to inspire designers to think about the most valuable issues by answering questions. 8.2. Description of the final Canvas The final version of Canvas is split into six fan-l ike sections, each representing one stage of the ML lifecycle: data collection, construction, training, prediction, performance, and updating. Canvas can be filled in collaboratively, with gro ups of people using the Question list to sketch and model the various aspects of ML, users, and t he scenario. In each section, the green area (see Figure 11b ) of Canvas is used to record information ab out the scenario, while the pink area (see Figure11a ) is for ML-related information. Issue cards ( Figure 11f )arelocatedinthegrayarea (see Figure 11c ). The Persona card shown in Figure 11e allows designers to record users ’ attitudes and feedback regarding UX issues. Figure 11d presents the Question list, with questions categorized according to the relevant co-creator s and touchpoints. Brief usage instructions are presented below. (1) Answer questions related to ML in the list and add relevant information, such as the dataset, inference speed, and neural network types in the corresponding area. Identify the UX issues Figure 11. The final version of Canvas. 374 Z. ZHOU ET AL. to be addressed at each stage and put the corresponding Issue cards in the gray area, thus clarifying the design goal. (2) Conduct the user study and complete the Persona as required; identify the key features of potential users, fill in user preferences and core appeals for ML products. (3) Answer questions related to the scenario and add the corresponding information, thus constructing a basic understanding of the scenario in terms of the laws, regulations, noise, temperature and etc. Adjustments can be made to the Issue cards in this process. (4) Explore potential proposals based on the guidance of the MLT and collected information. To this end, identify the connections and relationships among the three co-creators and explore solutions to deal with the issues at different stages. At this point, the conflicts caused by the different interests of the co-creators should be resolved. (5) Summarize the final proposal and write it down on the Issue cards for iterative evaluation. 9. Evaluation of ML lifecycle canvas To test how well the iterated Canvas works in terms of the four criteria presented in the previous section, we organized two-week design workshops . As there were no conceptual tools specifically designed for ML, we compared Canvas with Service Blueprint (Blueprint) which has similar features and goals to those of Canvas. Moreover, we presented the design outcomes of Canvas to show its effectiveness in dealing with UX is suesrelatedtothemalleablenatureofMLin conceptual design. 9.1. Participants Thirty-two participants (M: 11, F: 21) were recruited to participate in the evaluation. All participants were junior (third-year) undergraduate design students with the same background as the participants in phase 3. They were randomly divided into six teams containing six or five participants. None of the participants had been involved in the project in phase 3; the instructors were the same. 9.2. Procedure We invited the 32 participants to an evaluation session to compare ML Canvas and Blueprint in two workshops (each lasted for two weeks) on the theme of ML-empowered UX. To ensure the participants would not be biased by the previous evaluated tool, there was a six-month interval between the two workshops. The group organization, design topics, and environment of the two workshops were the same. Thus, we regarded the collected data as two independent groups. Both the workshops were part of a compulsory course and students received credit for the course. Before each workshop, the design students were introduced to the malleable nature of ML technology, including the mechanisms of ML, its attributes of unpredictability, data-centric nature, the related UX issues and etc. In the first workshop, the participants were given a two-hour introduction to the MLT method and Canvas (see Figure 12a ), whereas in the second workshop, they were introduced to service design thinking and Blueprint. After the introduction, all the participants were asked to complete two assignments. First, they were asked to use Canvas/Blueprint to illustrate an existing design case, then they were required to improve this case using Canvas/Blueprint (see Figure 12c ). During the workshop, there were several discussions where the instructors answered questions on the use of Canvas and Blueprint (see Figure 12b ). After using Canvas/Blueprint in each workshop, participants evaluated the four aspects of Canvas/Blueprint on a 5-point Likert scale, with 1 indicating strong disagreement and 5 indicating strong agreement. Each workshop was followed by the same procedure of in-person semi-structured interviews as those in Section 4.3.2, each lasting for 10 –15 minutes. The laddering questions regarding the four criteria HUMAN –COMPUTER INTERACTION 375 were also the same as those in Section 4.3.2. We did not show any expectancies regarding the performance of the participants in the workshops and the evaluation part. 9.3. Evaluation results The Shapiro-Wilk test indicated that the data from the four dependent variables (ease of use, creativity, understanding, process) was not normal distribution. Thus, we conducted a Mann – Whitney U test to compare Canvas and Blueprint with respect to the above four criteria. The analysis results showed that Canvas outperforms Blueprint regarding creativity and understanding. However, there was no significant difference between the two design tools regarding ease of use and process. The Boxplot is presented in Figure 13 . Figure 12. Workshops where the participants used the final version of Canvas. a) Introduction to Canvas; b) discussion; c) design with Canvas and Blueprint. Figure 13. Comparison between Canvas and Blueprint across the four criteria. 376 Z. ZHOU ET AL. In the corresponding in-depth analysis, we analyzed the advantages and limitations of Canvas compared with Blueprint from the results of the semi-structured interviews, using the same content analysis process as in Section 4.3.2. The following descriptions explain the reasons for the differences. 9.3.1. Ease of use There was no significant difference between Canvas ( Mdn = 3) and Blueprint ( Mdn = 3.5), U = 544.0, p=.645, r= .06. Many participants thought that Canvas was well-structured and friendly. The reasons include the following: (1) The visualization layout of Canvas “assisted us to organize information in a clear manner ”(P17) because of the “well-arranged blocks corresponding to the different co-creators (P5) ” and “concise symbols indicating different stages of ML process (P11) ”. They regarded Canvas as a good tool to record thoughts, present ideas, and review the proposals. (2) Several participants described Issue cards as “flexible ”because they are “quite easy to make adjustments ”(P5). For example, someone thought that they could easily remove or add some issues cards (P8) and take notes on the cards (P4) when discussing with partners to make sure that everyone was on the same page. (3) They also mentioned that the Question list was friendly because “some questions directly reminded us of problems that we had not considered ”(P22). By answering questions, participants were motivated to “collect the related information ”(P17) and “figure out if the concept design is feasible ”(P19). Despite all this, the negative evaluations were mainly caused by unfamiliarity with the ML concepts and usage method of Canvas. The participants thought that (1) the layout was too different from the existing design tools that designers are familiar with; (2) the visual template of Canvas did not intuitively indicate the order in which information should be entered, which complicated the usage method. In contrast, the two-dimensional chart of Blueprint is easier to understand. 9.3.2. Creativity The score was significantly higher for Canvas ( Mdn =4) than for Blueprint ( Mdn = 4), U = 318.0, p= .004, r= .36. Most of the participants thought that Canvas helped them to conceive diverse and practical UX solutions in an ML context. (1) The majority of participants commended Canvas for concerns about the malleable nature of ML. It allowed them to have an “open mind with regard to the six issues ” (P7, P14, P16) and offered them definite objects instead of seeking solutions blindly. (2) The consideration of co-creators helped them “come up with multiple solutions to a single issue ”(P7) and “figure out not only ML-related issues, but also issues involving other aspects ”(P20). (3) The division into six ML stages made the identification of solutions more in-depth. P17 said that he learned to pay attention to cultivating “a system for the common progress of co-creators. ”Canvas reminded designers to try to “strike a balance among the three co-creators (P12) ”, thus avoiding paying too much attention to one co-creator and damaging the interests of others. However, we also received the following complaints: (1) Several participants felt that the ideas derived through Canvas were more like incremental innovation rather than breakthrough innova- tion. They thought the main reason would be that the Canvas aims to provide a feasible design scheme that can specify the next steps like prototype, manufacture and so on. (2) Canvas lacked a mechanism to link scattered solutions into a complete scheme, so the design ideas corresponding to the issues are difficult to integrate into a comprehensive solution. For instance, P5 and P6 pointed out that their ideas were limited to the given stages or issues. (3) Several participants thought that the semi-structured Question list somehow restricted their ideas. P2 and P9 suggested adding more open-ended questions to inspire divergent thinking. (4) P15 and P21 suggested adding a blank Issue card to describe those UX issues that are not caused by ML. Though Canvas is designed to address ML-related UX issues, providing a way to visualize other issues would help participants to have overall control of the design proposal. HUMAN –COMPUTER INTERACTION 377 9.3.3. Process There was no significant difference between Canvas ( Mdn =3) and Blueprint ( Mdn = 3), U= 591.0, p= .256, r= .14. Several participants reported that they could easily integrate other design tools into the use of Canvas: (1) Some found that ideating with Canvas was “similar to service design thinking that requires comprehensive consideration of all stakeholders ”(P14). P16 drew an analogy between Canvas and Blueprint: “Data collection, performance, and update can be regarded as the frontstage in service design, while other stages are like the backstage. ”(2) Most participants spontaneously integrated typical design tools like brainstorming, interviews, and 5W2H into the process of completing Canvas, which shows the compatibility of Canvas with other design tools. (3) Canvas offered guidelines for conducting traditional design activities in the context of ML. Some steps of the traditional design process need to be adjusted by the participants to incorporate ML. For example, user research should involve new content about the UX issues that are related to ML. P13 mentioned that “the Issue cards are helpful when creating the outline of user interviews. ” There were some suggestions on optimizing the process. (1) Several participants believed that the layout of Canvas should be more like those of other traditional tools to enable better collaboration. (2) Some hoped that Canvas would provide detailed and formal guidelines on how to integrate design knowledge into the process of using Canvas. 9.3.4. Understanding The score was significantly higher for Canvas ( Mdn =4) than for Blueprint ( Mdn = 3.5), U= 249.5, p< .001, r= .48. Most participants stated that their understanding of ML had been improved by working with Canvas. (1) Canvas not only reminded designers of issues that had been ignored (P1, P10), but also “set up a framework to learn about ML in the way of design thinking ”(P5). (2) Some students said that Canvas changed their perception on the role of designer in ML-empowered UX design: “I realized that UX design can extend into the ML process like data collection and prediction ”(P7). Through using Canvas, they learned to enhance the performance of ML by appropriate UX design, e.g., “helping the ML system to collect user data in a more pleasant way ”(P18). However, Issue cards might not cover all ML-related UX issues(P15, P21). Additionally, Canvas does not clearly show the relationship among users, the ML system, and the scenario and “it is quite difficult to image the mutual influence between them ”(P21). 9.4. Design outcomes The completed proposals (Canvas group) are presented in Figure 14. The design outcomes of the six groups are all conceptual design proposals and not the actual implemented algorithm/systems. Te a m Brief summary 1. Hello Barbie Chat robot for accompanying children through their preferences 2. Clips Intelligent camera that automati cally records memorable moments of and their pets 3. Makeup Camera App A mobile app that helps users to vi rtually beautify themselves with virtual make-up in real time 4. Language Translator A website for academic translation tha t can translate sentences and words into different languages 5. Echo Dot Smart speaker tailored to the needs of children 6. E-shopping recommendation Shopping recommendation app based on social relationships and browsing history Figure 14. Description of completed projects. 378 Z. ZHOU ET AL. To illustrate how the participants engaged in the process of growing the ML-empowered UX iteratively over time, we present part of the design activities of team 4 as follows. The presented activities are derived from the completed Canvas, the presentation of team 4, and instructors ’records. 9.4.1. Understand the material Understand the attributes of ML models, including key elements such as the required dataset, the type of the required ML model, and the inference speed (see Figure 15d ). The ML models for text translation use a Recurrent Neural Network (RNN) that requires a large amount of computation and storage, thus cloud computing is usually adopted as a solution. Similarly, several studies are addressing the interpretability of RNN models and are transformed into practical applications. In addition, such models require a large amount of high quality, annotated data to improve system performance. 9.4.2. Understand co-creators (1) Produce Persona (see Figure 15c ) for researchers and college students who require high-quality scientific translation. They expect the platform to interpret the translation results, which helps to evaluate the credibility. Meanwhile, they possess academic translation skills and can provide profes- sional revisions. (2) Understand the scenario that mainly includes laws and regulations, social rules and other restrictions in this case (see Figure 15d ). For example, we may need to give relevant tips during the translation process for sensitive words and terminologies. In addition, academic activities require high confidentiality of text, and are sensitive to data privacy. 9.4.3. Identify touchpoint Team 4 filled the above-collected informatio ninto the corresponding areas of the ML process (touchpoints) and identified how the co-creators could have an impact on the material ’s growth. In this case, users revise the transl ation result so that the ML can learn from the data provided by the users. As the ML model is re fined, the accuracy of translation will be improved so users can benefit from the platform. Also, existing interpretable practices in Figure 15. Part of the design proposal that engages in the process of growing the ML-empowered UX. HUMAN –COMPUTER INTERACTION 379 RNN models can meet the user ’s demand for reliable results. However, the data privacy requirements in this scenario conflict with the ML system that currently only provides online services. 9.4.4. Embody The design team tried to eliminate conflicts between co-creators ’dem ands and then filled in the solutions of issues in the corresponding area. For example, using different colors to present the different operating status of the system (see Figure 15b ); using example sentence and word frequencytopresenttheinterpretationoftranslationresults(see Figure 15a ). Finally, they produced a preliminary conceptual design pr oposal by embedding these design solutions to the ML process. 9.4.5. Start another loop The design team continued to monitor how the co-creators influence each other and ensure that UX issupported by the continuous growth of ML. For example, they generated the mechanism of the medal symbolizing professionalism (see Figure 15a ) that motivates users to produce data to con- tinuously improve the ML. The system also differentiates data providers for their professionalism to quickly select high-quality samples, and so on. We also present novel ideas for coping with six UX issues from these proposals (see Figure 16 ). These ideas demonstrate how designers tackled the malleable nature of ML through addressing the six UX issues. Some of the completed Canvas is shown in Figure 16d . 9.4.6. Unpredictability Participants explored several ways to alleviate the confusion caused by the uncertainty of malleable ML applications. (1) Team 5 designed guiding questions for when the smart speaker cannot under- stand what the user has said. For example, the robot will guide users to repeat themselves or use other terms to interpret what has been said when the voice quality is poor. (2) Team 1 explored reply strategies for when their robot is not confident about the answers. For example, the robot will seek help from parents by saying “Let ’s ask mom together ”to the child. (2) Team 6 used easily under- stood tags to describe recommendations (see Figure 16b ). This allows users to quickly understand why an item is recommended and somehow reduces the feeling of unpredictability. Figure 16. Outcomes of the design workshop. 380 Z. ZHOU ET AL. 9.4.7. Transparency Various methods were conceived to expose the complexity and malleable mechanism of ML to users, which might reduce unpleasant situations. (1) Team 3 envisaged that the ML system could visualize feature point annotation of the face, whereupon users could adjust misplaced feature points to enhance the performance of ML, thus improving the UX (see Figure 16c ). Moreover, manual annotation may help the ML optimize its learning dataset. (2) Team 4 showed the current working status (working/completed) through the color of the results box. When users choose a particular vocabulary, the interpretation and other information (such as frequency of occurrence, reference source) will be displayed (see Figure 16e ) to illustrate the credibility of the results. 9.4.8. Learning The participants applied novel strategies to produce labeled data and promote the growth of the ML application, although only with the authorization of the users. (1) Team 1 spontaneously stimulated emotional expressions and obtained data, such as capturing the child ’s happy expression by telling jokes. (2) Team 4 used rewards (medal symbolizing professionalism) to facilitate users making sug- gestions on the translation results. Based on the users ’proficiency, the translation platform assigns different weights to different users, thus attaching more importance to high-quality samples. (3) All design teams applied different approaches to protect the privacy of users. For example, teams 1 and 2 allowed training, prediction, and so on to be performed locally without connecting to the Internet. 9.4.9. Control Participants came up with different strategies to balance the control exerted by the ML application and the user over the products. (1) Team 2 listed the recognized subjects on a mobile app and enabled users to remove subjects that should no longer be recognized. (2) Team 5 offered different modes for children of different ages. The complexity of the voice interaction will increase as the child grows up. (3) Team 6 offered a seamless browsing mode, allowing users to remove some behavior data from the training dataset so that the ML system will not be misled. 9.4.10. Anthropomorphism The participants explored various anthropomorphic styles to avoid the “uncanny valley ”while providing a pleasant impression. (1) Team 1 increased the anthropomorphism of their robot by using body language and voice characteristics instead of facial expressions, thus avoiding the “uncanny valley. ”(2) In contrast, team 5 was cautious about anthropomorphism. They customized certain wake words, but reduced the anthropomorphism. This avoids cognitive problems in children who should not be over-dependent on such products. (3) Team 2 retained the non-anthropomorphic appearance of Clips because it clearly makes itself known as a camera. To some extent, this protects users ’privacy. 9.4.11. Interactivity Participants tried to make the interaction smooth, and thus provided improved UX and promoted the users ’contribution to the growth of ML. (1) Team 6 simplified the process of collecting user performances. For example, the option for the user to specify that they “do not like this recommen- dation ”was very easy to select. To further enhance reliability, users can easily state why they do not like a particular recommendation. (2) To reduce the computation time used for prediction, team 5 decreased the complexity of the voice content, e.g., setting modal words with no special meaning to respond to 1 –3-year-old children and to encourage them to keep talking. (3) Team 1 used an ML accelerator called a neural computing stick to decrease the computation time, as all the computation is performed within the local hardware. HUMAN –COMPUTER INTERACTION 381 10. Discussion and future work The proposed MLT method focuses on the malleable nature of ML-empowered UX design and enables designers to participate in the development of the material. During the design process of Canvas, we obtained several insights and lessons that may inspire further research into the design methodology and tools unifying AI and UX. 10.1. Contribution to the integration of AI and UX design In this paper, we have described the characteristics of ML from the perspective of conceptual design and summarized the design challenges caused by the growable nature of ML. This might provide a foundation for follow-up UX research that considers AI technology as a growable “tree ”instead of static “wood. ” The MLT method encourages designers to reflect on the whole lifecycle of a malleable material like ML and cultivate it into the desired form. As an extension of existing design methods, MLT provides specific features for human-AI interaction design. In contrast with Human-centered Design, MLT pays equal attention to factors besides human users, especially the design material, and supports co-creation among users, ML, and the scenario. Compared with Value-Sensitive Design and participatory design methods, MLT focuses on the development process of the design material and how to intervene in the process to gain the desired UX. The Canvas provides a conceptual design template for applying MLT in growing ML-empowered UX and developing design proposals. By discussing issues about ML, users, and the scenario in a visual way, the interplay between them can be analyzed throughout the entire ML lifecycle. Our evaluation result illustrates that Canvas performs better than Blueprint in terms of creativity and understanding. However, we do not observe significant difference between the two design tools regarding ease of use and process. The interviews reveals that Canvas introduces ML-related concepts in a way that is consistent with design thinking, which is easy understandable for designers. Besides, the forms and compositions of Canvas offer meaningful creative opportunities for designers by revealing the malleable nature of ML. Additionally, the Canvas framework can be extended to other malleable design materials, enabling designers to fulfil the five stages of the MLT process. The Issue cards could be used to describe any kinds of malleable material (see Figure 17a ).TherowsandcolumnsoftheCanvas format can be used to present the co-creators and touchpoints of any other materials (see Figure 17b,c ). Designers can insert information into each space representing the different co- creators at specific touchpoints. Finally, desi gners could summarize the design insights from collected information (see Figure 17d ). The extended Canvas could easily be transformed to be applied to a similar technology. 10.2. Future work on improving MLT and Canvas For now, MLT and Canvas do not violate the traditional design process; rather, they complement original design methods because they deal with a kind of technology with malleable attributes that Figure 17. The extended material lifecycle Canvas. 382 Z. ZHOU ET AL. has not been widely applied as a design material before. We chose to retain typical design activities because designers prefer to learn about an unfamiliar material in a familiar design process. However, the design thinking process might evolve as designers better understand the inner workings of ML and prepare for working with ML. For instance, the prototyping process might disappear when AI- empowered products can iterate themselves to form better versions. The evaluation of MLT and Canvas was conducted in a university, but no implementation in a working environment has yet been investigated. Although instructors played the role of technical teams, the design students lack experiences (e.g., cooperating with engineers, applying other emer- ging technology in design) that might facilitate considerations of technical feasibility and thus potentially influence the evaluation result. Besides, the construction of evaluation and recruitment of participants may inevitably bring some biases including the educational setting of the evaluation, order effects, and etc. Future work should eliminate the possibility of order effects and other biases. We would also try to gain feedback in an industrial environment. The current Canvas platform has several limitations and requires further improvements. (1) The Question list may constrain the creative design activities, as it cannot cover all possible questions and inevitably presupposes some relationships among the three co-creators. Future practice might examine a way of providing sufficient guidance without implying these relationships. (2) Canvas should not be confined to designers, as it could also serve as a bridge for connecting experts in different disciplines. For example, Canvas might help engineers learn about design thinking, or offer a way for product managers to gain a comprehensive view of the lifecycle of ML-empowered products, thus improving communication efficiency and promoting cooperation. However, experts without a design background might require more time to learn Canvas before using it in their tasks. (3) Within the scope of our current research, Canvas is for the conceptual phase and cannot really be used in prototyping, such as helping designers to implement the ML systems. We look forward to extending Canvas to a toolbox that is not constrained to the conceptual phase. In the future, we would like MLT to be used in various industries and be used by experts in different disciplines. This would provide us with insights that motivate the further unification of UX and AI and generate new design guidelines or principles in this filed. 11. Conclusion In this paper, we have introduced MLT as a means of helping designers, especially novice designers, to explore and appreciate the malleable nature of ML, and encouraging them to consider the ML system, users, and scenario as co-creators of ML-empowered UX. The ML Lifecycle Canvas, a visual schematic for applying MLT, was developed in an iterative design process involving designers. Canvas visually represents three co-creators (ML, user, and scenario) and outlines notable questions within the ML lifecycle to enable the identification of crucial design possibilities. Through a comparison with Service Blueprint, we found that our approach guides novice designers to tackle the identified UX issues throughout the whole ML lifecycle and gain an increased understanding of the co-creators of ML- empowered UX. Though our work is at an early stage, it takes us one step closer to unifying UX and ML. We expect that the MLT method and Canvas will be improved through future use in design practice. Funding Project supported by the National Natural Science Foundation of China (No. 61672451), the China Postdoctoral Science Foundation (2018M630658), Zhejiang Provincial Key Research and Development Plan of Zhejiang Province (No. 2019C03137), and the National Science and Technology Innovation 2030 Major Project of the Ministry of Scienceand Technology of China (2018AAA0100703). HUMAN –COMPUTER INTERACTION 383 Notes on contributor Lingyun Sun is a Professor of Design at Zhejiang University. He ’scurrently a deputy director of International Design Institute. He has interdisciplinary research experiences, including artificial intelligence, computer graphics, design cognition, interaction design, ergonomics and drama. He is taking part in the MIT-SUTD Teach-the-Teacher program and teaching courses regularly at SUTD. He is the author or coauthor of over 20 journal or conference papers on Industrial Design, Computer Aided Sketching and Interaction Design. He holds patents on inventions relating to interaction and sketching technology. He is a member of the Industrial Designers Society of America, a member of thecouncil of Chinese Mechanical Engineering Society Industrial Design Institution, a member of China Computer Federation. References Abadi, M., Barham, P., Chen, J., Chen, Z., Davis, A., Dean, J., Devin M, Ghemawat S, Irving G, Isard, M., & Kudlur M (2016 ).Tensorflow: A system for large-scale machine learning [Paper presentation], Savannah, GA, USA. OSDI. Amershi, S., Cakmak, M., Knox, W. B., & Kulesza, T. ( 2014 ). Power to the people: The role of humans in interactive machine learning. AI Magazine ,35 (4), 105 –120. https://doi.org/10.1609/aimag.v35i4.2513 . Amershi, S., Weld, D., Vorvoreanu, M., Fourney, A., Nushi, B., Collisson, P.,Suh J, Iqbal S, Bennett PN, Inkpen K, &Teevan J. ( 2019 ).Guidelines for human-AI interaction [Paper presentation]. Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems, Glasgow, Scotland, Uk. Atasoy, B., & Martens, J.-B. ( 2011 ).Crafting user experiences by incorporating dramaturgical techniques of storytelling [Paper presentation]. Procedings of the Second Conference on Creativity and Innovation in Design, Eindhoven, Netherlands.Atasoy, B., & Martens, J.-B. ( 2016 ). STORYPLY: Designing for user experiences using storycraft. In Collaboration in creative design (pp. 181 –210). Springer. Baumer, E. P. ( 2017 ). Toward human-centered algorithm design. Big Data & Society ,4(2), 2053951717718854. https:// doi.org/10.1177/2053951717718854 . Baumer, E. P. S., Ahn, J., Bie, M., Bonsignore, E. M., B, A., & Yip, J. ( 2014 ).CHI 2039: Speculative research visions [Paper presentation]. CHI ‘14 Extended Abstracts on Human Factors in Computing Systems, Toronto, Ontario, Canada. Bitner, M. J., Ostrom, A. L., & Morgan, F. N. ( 2008 ). Service blueprinting: A practical technique for service innovation. California Management Review ,50(3), 66 –94. https://doi.org/10.2307/41166446 Bucher, T. ( 2017 ). The algorithmic imaginary: exploring the ordinary affects of facebook algorithms. Information, Communication & Society ,20(1), 30-44. doi: 10.1080/1369118X.2016.1154086 Cuevas, H. M., Fiore, S. M., Caldwell, B. S., & Laura, S. (2007). Augmenting team cognition in human-automation teams performing in complex operational environments. Aviation Space & Environmental Medicine ,78(5), B63 –B70. https:// www.ingentaconnect.com/content/asma/asem/2007/00000078/A00105s1/art00010.De Choudhury, M., & Kiciman, E. ( 2018 ). integrating artificial and human intelligence in complex, sensitive problem domains: experiences from mental health. AI Magazine ,39(3), 69 –80. https://doi.org/10.1609/aimag.v39i3.2815 Deng, L., & Li, X. ( 2013 ). Machine learning paradigms for speech recognition: An overview. IEEE Transactions on Audio, Speech, and Language Processing ,21(5), 1060 –1089. https://doi.org/10.1109/TASL.2013.2244083 Dove, G., Halskov, K., Forlizzi, J., & Zimmerman, J. ( 2017 ).UX design innovation: challenges for working with machine learning as a design material [Paper presentation]. Proceedings of the 2017 CHI Conference on Human Factors in Computing Systems, Denver, Colorado, USA.Dunne, A., & Raby, F. ( 2013 ).Speculative everything: Design. Fiction, and social. The MIT press. Fiebrink, R., & Cook, P. R. ( 2010 ).The Wekinator: A system for real-time, interactive machine learning in music [Paper presentation]. Proceedings of The Eleventh International Society for Music Information Retrieval Conference (ISMIR 2010), Utrecht. Forlizzi, J., Zimmerman, J., Mancuso, V., & Kwak, S. ( 2007 ).How interface agents affect interaction between humans and computers [Paper presentation]. Proceedings of the 2007 conference on Designing pleasurable products and interfaces, Helsinki, Finland.Gillies, M., Fiebrink, R., Garcia, A. T., Bevilacqua, F., Heloir, A., Nunnari, F., Mackay W, Amershi S, Lee B, &d’Alessandro N. ( 2016 ).Human-centred machine learning [Paper presentation]. Proceedings of the 2016 CHI Conference Extended Abstracts on Human Factors in Computing Systems, San Jose, California, USA.Gillies, M., Fiebrink, R., Tanaka, A., Garcia, J., Bevilacqua, F., Heloir, A., Nunnari, F., Mackay W, Amershi S, Lee B, &d’Alessandro N. ( 2016 ).Human-centred machine learning [Paper presentation]. Proceedings of the 2016 CHI Conference Extended Abstracts on Human Factors in Computing Systems, San Jose, California, USA.Google. ( 2019a ).AI & machine learning products: Cloud AutoML . Retrieved from https://cloud.google.com/automl/ Google. ( 2019b ).Do-it-yourself artificial intelligence .https://aiyprojects.withgoogle.com/ 384 Z. ZHOU ET AL. Google. (2019c).People+ AI guidebook: Designing human-centered AI products .https://pair.withgoogle.com/intro/ Hancock, P. A., & Chignell, M. H. (1988). Mental workload dynamics in adaptive interface design. IEEE Transactions on Systems Man & Cybernetics ,18 (4), 647 –658.https://doi.org/10.1109/21.17382. Hand, D. J. ( 1998). Data mining: Statistics and more? The American Statistician,52(2), 112 –118. doi: 10.1080/ 00031305.1998.10480549. Hanington, B., & Martin, B. ( 2012). Universal methods of design: 100 ways to research complex problems. Develop Innovative Ideas, and Design Effective Solutions: Rockport Publishers . https://books.google.com/books/about/ Universal_Methods_of_Design.html?id=uZ8uzWAcdxEC Hebron, P. ( 2016).Machine learning for designers .O’Reilly Media. Holmquist, L. E. ( 2017). Intelligence on tap: Artificial intelligence as a new design material. Interactions,24 (4), 28–33. https://doi.org/10.1145/3085571 . Howitt, D, & Cramer, D. ( 2010). Introduction to qualitative methods in psychology (vol. In 1). UK: Pearson Education. IBM. ( 2019).AutoAI with IBM watson studio .https://www.ibm.com/cloud/watson-studio/autoai Jacobs, J., Gogia, S., Radom, M., & Brandt, J. R. ( 2017).Supporting expressive procedural art creation through direct manipulation [Paper presentation]. Proceedings of the 2017 CHI Conference on Human Factors in Computing Systems, Denver, Colorado, USA. Jess, H. ( 2017).Human-centered machine learning .https://medium.com/google-design/human-centered-machine- learning-a770d10562cd Karana, E., Barati, B., Rognoli, V., Der Laan, V., & Zeeuw, A. (2015 ).Material driven design (MDD): A method to design for material experiences. Kim, J., & Ryu, H. ( 2014). A design thinking rationality framework: Framing and solving design problems in early concept generation. Human–Computer Interaction ,29(5 –6), 516– 553.https://doi.org/10.1080/07370024.2014. 896706 Koza, J. R., Bennett, F. H., Andre, D., & Keane, M. A. ( 1996). Automated design of both the topology and sizing of analog electrical circuits using genetic programming. In Artificial intelligence in design’96 (pp. 151 –170). Springer. Kruschke, J. K. ( 2011).Doing Bayesian data analysis: A tutorial with R and BUGS . Academic Press/Elsevier. Kuhn, S., & Muller, M. J. ( 1993). Participatory design. Communications of the ACM ,36(6), 24–29. https://doi.org/10. 1145/153571.255960 Kuniavsky, M., & Churchill, E. ( 2017).Designing the user experience of machine learning systems SS-17-04 . Li, Y.-M., & Yeh, Y.-S. (2010). Increasing trust in mobile commerce through design aesthetics. Computers in Human Behavior ,26 (4), 673 –684.https://doi.org/10.1016/j.chb.2010.01.004 . McCardle, J. R. ( 2002). The challenge of integrating AI & smart technology in design education. International Journal of Technology & Design Education ,12 (1), 59 –76.https://doi.org/10.1023/A:101308940 . McElreath, R. ( 2018).Statistical rethinking: A Bayesian course with examples in R and Stan . Chapman and Hall/CRC. Merritt, S. M., Heather, H., Jennifer, L. C., & Deborah, L. (2013). I trust it, but I don ’t know why: Effects of implicit attitudes toward automation on trust in an automated system. Human Factors,55 (3), 520 –534.https://doi.org/10. 1177/0018720812465081. Mitchell, T. M. ( 1999). Machine learning and data mining. Communications of the ACM,42(11), 30–36. https://doi. org/10.1145/319382.319388 Muller, M. J. ( 2007). Participatory design: The third space in HCI. In The human-computer interaction handbook(pp. 1087 –1108). CRC press. Murnane, E. L., & Counts, S. (2014 ).Unraveling abstinence and relapse: Smoking cessation reflected in social media [Paper presentation]. Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, Toronto, Ontario, Canada. Patel, K., Fogarty, J., Landay, J. A., & Harrison, B. (2008 ).Examining difficulties software developers encounter in the adoption of statistical machine learning [Paper presentation]. Proceedings of the 23rd national conference on Artificial intelligence – Volume 3, Chicago, Illinois. Pruitt, J., & Adlin, T. (2010). The persona lifecycle: Keeping people in mind throughout product design . Elsevier. Recasens, M., Danescu-Niculescu-Mizil, C., & Jurafsky, D. (2013 ).Linguistic models for analyzing and detecting biased language [Paper presentation]. Proceedings of the 51st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). https://doi.org/10.1016/S2213-2600(13)70187-5 Reynolds, T. J, & Gutman, J. J. ( 1998). Laddering theory, method, analysis, and interpretation. Journal of Advertising Research ,28(1), 11-31. Ribeiro, M., Grolinger, K., & Capretz, M. A. M. ( 2016).MLaaS: machine learning as a service [Paper presentation]. IEEE International Conference on Machine Learning and Applications. Miami, FL, USA, the DOI is 10.1109/ ICMLA.2015.152. Rozendaal, M. C., Ghajargar, M., Pasman, G., & Wiberg, M. ( 2018). Giving form to smart objects: Exploring intelligence as an interaction design material. In M. Filimowicz & V. Tzankova (Eds.), New directions in third wave human-computer interaction: Volume 1 – technologies (pp. 25–42). Springer International Publishing. HUMAN–COMPUTER INTERACTION 385 Russell, A., Ekvall, T., & Baumann, H. (2005). Life cycle assessment introduction and overview. Journal of Cleaner Production, 13(13–14), 1207 –1210.c Schon, D. A., & Desanctis, V. (2005 ). The reflective practitioner: How professionals think in action. Proceedings of the IEEE , 73(4), 845 –846. https://books.google.com/books?hl=zh-CN&lr=&id=ceJIWay4- jgC&oi=fnd&pg=PR7&dq=The+reflective+practitioner:+How+professionals+thin k+in +action&ots=q88QS_DWum&sig=LrVDHtR4asDzR7DiTkl_cXJElk4#v=onepage&q=The%2 0reflective%20practi- tioner%3A%20How%20professionals%20think%20in%20action&f=false. Snyder, J., Baumer, E. P., Voida, S., Adams, P., Halpern, M., Choudhury, T., & Gay, G. (2014 ). Making things visible: Opportunities and tensions in visual approaches for design research and practice. Human–Computer Interaction ,29 (5 –6), 451 –486.c Stickdorn, M., Schneider, J., Andrews, K., & Lawrence, A. (2011). This is service design thinking: Basics, tools, cases (Vol. 1). Wiley Hoboken. Tschimmel, K. ( 2012).Design thinking as an effective toolkit for innovation [Paper presentation]. ISPIM Conference Proceedings, Barcelona, Spain. https://doi.org/10.1094/PDIS-11-11-0999-PDN Vlist, B., Westelaken, R., Bartneck, C., Hu, J., Ahn, R., Barakova, E., Delbressine F, & Feijs, L. (2008 ).Teaching machine learning to design students [Paper presentation]. Proceedings of the 3rd international conference on Technologies for E-Learning and Digital Entertainment, Nanjing, China. Wang, D., Yang, Q., Abdul, A., & Lim, B. Y. (2019 ).Designing theory-driven user-centric explainable AI [Paper presentation]. Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems, Glasgow, Scotland, Uk. Witten, I. H., Frank, E., Hall, M. A., & Pal, C. J. ( 2016).Data mining: Practical machine learning tools and techniques . Morgan Kaufmann. Wood, M. ( 2014).Hopes to serve up music in a novel way .http://nyti.ms/1fr8LY5 Yang, Q. ( 2017).The role of design in creating machine-learning-enhanced user experience [Paper presentation]. 2017 AAAI Spring Symposium Series on Designing the User Experience of Machine Learning Systems, Stanford, CA, USA. Yang, Q., Banovic, N., & Zimmerman, J. (2018 ).Mapping machine learning advances from HCI research to reveal starting places for design innovation [Paper presentation]. Proceedings of the 2018 CHI Conference on Human Factors in Computing Systems. Yang, Q., Scuito, A., Zimmerman, J., Forlizzi, J., & Steinfeld, A. (2018 ).Investigating how experienced ux designers effectively work with machine learning [Paper presentation]. Proceedings of the 2018 Designing Interactive Systems Conference, Hong Kong, China. Yang, Q., Suh, J., Chen, N.-C., & Ramos, G. ( 2018).Grounding interactive machine learning tool design in how non-experts actually build models [Paper presentation]. Proceedings of the 2018 Designing Interactive Systems Conference, Hong Kong, China. Zhou, Z., Gong, Q., Qi, Z., & Sun, L. ( 2019).ML-process Canvas: A design tool to support the UX design of machine learning-empowered products [Paper presentation]. Extended Abstracts of the 2019 CHI Conference on Human Factors in Computing Systems, Glasgow, Scotland, Uk. Zhou, Z., Jiang, H., Yang, C., Yang, J., Loy, Y. Y. W., & Sun, L. (2019 ).QianLi: A modular system for connecting distant family members through tacit interaction [Paper presentation]. International Conference on Human-Computer Interaction, HCII 2019, Orlando, Florida, USA. Zhu, H., Yu, B., Halfaker, A., & Terveen, L. ( 2018). Value-sensitive algorithm design: Method, case study, and lessons. Proceedings of the ACM on Human-Computer Interaction ,2 (CSCW), 1–23. https://doi.org/10.1145/3274463 386 Z. ZHOU ET AL.