How to build a gen AI application

Intro

As enterprises race towards building and deploying generative AI applications, there are many open questions about how to design the end user experience of these applications to take advantage of this powerful technology, while avoiding its pitfalls. The Cloud AI and Industry Solutions team has been conducting extensive design and user research activities to understand how users perceive and engage with generative AI applications. Here Google share our learnings in order to help enterprise customers build useful and delightful end user experiences for their generative AI applications. 

Though generative AI applications have exploded in popularity in the last few months, very little is understood today about users’ mental models, perceptions, preferences, and challenges while interacting with these applications. Our user experience (UX) team launched an organization-wide effort aimed at creating a collection of user-tested micro-interaction design patterns and research-backed guidance for using those design patterns in generative AI applications. 

For this, Google first launched a “design challenge” within our UX team to solicit a variety of micro-interaction designs, guided by generative AI design principles. Next, Google created prototypes of various chatbot applications that incorporated those designs. Finally, we collected qualitative feedback from 15 external users on these prototypes through various user research methods, like moderated and unmoderated user feedback sessions. 

Below, Google distill our takeaways in the form of design principles and design patterns to illustrate those principles. 

Key takeaways

1. Help users explore generative variability

One of Google key learnings was that we should help users explore generative variability, that is the ability of a generative AI application to produce a range of outputs for the same prompt or question. Additionally, Google need to help users understand triggers and end points of an interaction with a generative AI model. 

Fig (1)

Fig (3), Fig (4)

 2. Help users build trust 

It was important for users to feel like they can trust the generated responses. Google found that citing the source of the content helps increase user trust. Users appreciated citations in various formats. For example, when a conversational bot provided hotel recommendations in a table format (Fig 5), users liked having links to each hotel’s website so they could visit the hotel page, check out photos, and confirm that the price was correct. When presented with a response summarized from many sources (Fig 6), they wanted links to the individual sources that were used in arriving at the summary.

Fig (5), Fig (6)

3. Give users control over generated responses

Provide users control by letting them choose between multiple responses generated or by allowing them to modify the generated response based on their preferences. This helps ensure that users get the information that they need, in the format that they want. For example, in the chat application, users wanted to be able to edit the image that was generated based on the chat content to align it with their own preferences.

Another way of giving users control is to help them with prompt engineering. Users of generative AI applications may not understand how the quality of their prompt affects answers, and may be unaware of different prompting techniques that they can apply. In our studies, users liked “suggested prompt” chips (Fig 7) as a way to explore better and/or related prompts. They also liked the ability to edit prompts through dropdowns that helped select alternative values for prompt parameters (Fig 8).

Fig (7), Fig (8)

4. Improve results through user feedback

User feedback can be used to improve the quality of the generated information, and to make sure that it is aligned with the needs of the users. Google found that users cared about the placement and timing of the feedback collection. For example, when users saw the feedback request below (Fig 9) during a conversation with a travel chatbot, some thought that the chatbot was done providing information and was ready to move on to another topic. This was perhaps because the feedback collection was disjointed visually from the bot’s response. Google should design feedback collection to be “in-the-moment” by placing the feedback request next to the response Google need feedback on (Fig 10). Additionally, Google could collect reasons from the user to explain their rating, providing insight into how to make the generated responses better.

Fig (9)

Fig (10)

Conclusion

During this iterative research and design process, Google also learned best practices for conducting user research on interfaces that provide generative outputs. For example, user studies with a user research expert guiding the exploration worked much better than studies where we left users alone with the prototypes. But most critically Google learned that users care about transparency, and want to know more about what’s happening under the hood to see if they can trust the generated outputs. Providing users a sense of control and ways to give feedback also goes a long way to driving trust in the generated outputs.

Related posts

5 Do’s and Don’ts CSPs Should Know About Going Cloud-Native

by Cloud Ace Indonesia
3 years ago

How to integrate Gemini and Sheets with BigQuery

by Cloud Ace Indonesia
2 months ago

It’s official—Google Distributed Cloud Edge is generally available

by Kartika Triyanti
3 years ago