Keith,
I am very interested in what you learn. I often do user studies with computer systems, and I am interested to learn what others are doing to evaluate user interactions with the UI.
I currently use TechSmith's Morae but it has been discontinued for several years. It is a great evaluation tool that basically mirrors the screen but will track clicks and mouse movement and allows real-time analysis with multiple observers who are able to place timestamp markers and notes, then combines the data. I believe they stopped supporting the software because of a third-party vendor issue and lack of sales (my assumptions). I hope to find a replacement.
Thanks for asking,
Burpee
------------------------------
Jerry Burpee
Human-Machine Interaction Sr. Principal Engineer
INTUITIVE Research & Technology Corp
Huntsville, AL
------------------------------
Original Message:
Sent: 03-10-2023 12:25
From: Keith Karn
Subject: Video capture of app interactions
Happy Friday all.
I'm planning a usability study that includes both hardware components and a companion mobile app that runs on iOS phones. In my audio and video recordings of the usability sessions I want to record the participant's interactions with the app on a mobile phone (iOS only). In past studies, I've tried various approaches for this including:
- Mounting a video camera to the phone and directing it toward the screen with a device called Mr. Tappy or similar custom-built devices.
- Screen mirroring / screen casting
- Taping off an area of a worksurface and asking the participant to interact with the phone / app keeping it on the worksurface within the bounds of the tape and capturing interaction with an overhead camera.
Each has it's pros and cons. Solution 1 (mounting the camera to the phone) allows video capture of the user touch interactions with the screen, as well as the app's display, but results in an unnatural feel when handing the phone as the camera and mounting bracket increase mass and rotational inertia. I recall that we had some reliability issues with the screen mirroring (Solution 2), but that was a few years ago, and things may be improved. In iOS, the screen mirroring does not display where the person touches the screen -- that can only be inferred by the app's response. As I recall, screen casting from Android devices allowed us to display a marker on the screen wherever the user touched it, but iOS did not. Solution 3 (limiting use of the phone to a restricted area on a surface) also results in unnatural use of the touch screen since the participant cannot hold the phone normally.
I haven't done any of these types of recordings in a few years and wonder if there are any better tools available now. Please share your thoughts and suggestions (preferably to the group, or separately to me and I can post a summary).
Thanks in advance.
Keith
------------------------------
Keith S. Karn, PhD
Human Factors in Context LLC
Philadelphia, PA
215-531-1521
Keith@HumanFactorsInContext.com
------------------------------