Pilot User Study
In the small pilot user study at the end of the project, in the last week, you will test an agent from another team and another team will test yours! Check out the Course Schedule for when this pilot study is planned.
Preparation: By this time, you should have an agent that is ready for final evaluation. For your agent to be ready for the pilot user study evaluation, you need to make sure that:
It runs…
It is able to filter recipes.
Furthermore, it provides brief instructions and an overview of the agent's features and capabilities (in particular, you should probably mention some of the extensions that you implemented) on the start page.
You also need to specify the procedure of your user study. See https://www.simplypsychology.org/research-report.html#method for more on this, and how to report your findings.
Organization: For the pilot user study day:
We will pair teams. The members of the team that you are paired with are the participants in your user study.
You need to invite team members from the team you are paired with, your participants, for your own user study:
Invite participants individually to interact with your agent at different time slots.
Each participant that you invite interacts with your agent. Explain the procedure of your user study.
Make sure they each have at least three interaction sessions (conversations) with your agent.
Collect and record the data of each interaction. Data can be log files of your agent, data that Dialogflow collects, and observations that you make of how a participant interacts with your agent.
Vice versa, make sure you comply with the procedure of the user study conducted by another team! In other words, take testing another team’s agent seriously! It is the proper thing to do. But we will also take how you conducted yourself as a participant into account in our final assessment.
Setting up your user study: Your team will run the pilot user study for your own agent and collect what you think is relevant data to analyze the performance of your agent. Consider some of the metrics that were already explained on the https://socialrobotics.atlassian.net/wiki/spaces/PM2/pages/2257616897 page: effectiveness, efficiency, robustness, and user satisfaction. So think about what you want to evaluate and how to do that. Also, consider that the greatest advantage of having another team test your agent is that they likely will interact differently and may have different conversations with your agent than you have seen before in your own agent testing (that’s why it makes sense to already involve people from outside your team during agent testing).
Reporting: In your Final Report you should focus on the analysis of the data that you collected. You should have collected data for close to 20 conversational interactions with your agent. For these conversations, you can report some of the more interesting and basic https://en.wikipedia.org/wiki/Descriptive_statistics. Descriptive statistics provide simple summaries about the sample (your participants). You do not need to provide figures or tables (you won’t have much space for that in your report; If you want to provide details you can add these in an appendix to your report) but should focus on the more interesting findings. Most importantly, briefly discuss and interpret the data that you collected to explain what the data can tell us about the performance of your agent.