Sunday, November 13, 2011

Paper Reading #32- Taking advice from intelligent systems: the double-edged sword of explanations

Title:  Taking advice from intelligent systems: the double-edged sword of explanations.
Reference Information:
Kate Ehrlich, Susanna Kirk, John Patterson, Jamie Rasmussen, Steven Ross, and Daniel Gruen, "Taking advice from intelligent systems: the double-edged sword of explanations". IUI '11 Proceedings of the 16th international conference on Intelligent user interfaces. ACM New York, NY, USA. ©2011. ISBN: 978-1-4503-0419-1.
Author Bios:
Kate Ehlrich- Kate is Senior Technical Staff Member in the Collaborative User Experience group at IBM Research where she uses Social Network Analysis as a research and consulting tool to gain insights into patterns of collaboration in distributed teams.
Susanna Kirk- M.S., Human Factors in Information Design. Graduate Certificate, Business Analytics. Coursework in user-centered design, prototyping, user research, advanced statistics, data mining, data management and data visualization..
John Patterson- John Patterson is a Distinguished Engineer (DE) in the Collaborative User Experience Research Group.
Jamie Rasmussen- Jamie Rasmussen joined the Collaborative User Experience group in March 2007. He is working with John Patterson as part of a team that is exploring the notion of Collaborative Reasoning, the work a group of people does to collect, organize, and reason about information.
Steven Ross- Steve is presently working in the area of Collaborative Reasoning, using semantic technology to help individuals within an organization to think together more effectively and to enable them to discover and benefit from existing knowledge within the organization in order to avoid duplication of effort and permit better decisions to be made in a more timely fashion.
Daniel Gruen- Dan is currently working on the Unified Activity Management project, which is exploring new ways to combine informal work with business processes, recasting collaboration technologies in terms of the meaningful business activities in which people are engaged.
Summary:
  • Hypothesis: If the authors can investigate intelligent systems and its justifications, then maybe the accuracy of these systems will increase and users will not be "led astray" as much as they are being currently.
  • Methods: The authors decided to conduct a study on the effects of a user's response to a recommendation made by an intelligent system as well as the correctness of the recommendation. In this case, it was conducted on analysts engaged in network monitoring. The authors used a software called NIMBLE to help collect data for this study. 
  • Results: The users performed slightly better with a correct recommendation than without one. Results indicated that justifications grant benefits to users when a correct response is available. When there is no correct response available, neither suggestions nor justifications made a difference in performance. Most of the analysts seemed to discard the recommendations anyway, relying on their own inclinations. In the separate study concerning analyzing users' reactions, it was found that users typically follow the recommendations given and that the influence between the recommendation and the user's action is high. 
  • Content: The authors wanted to create a study to test the accuracy of recommendations and the relationship between that and the influence of those recommendations on users' actions. NIMBLE allowed the authors to do that, and they were accurately able to capture a system's interactions between users. By continually comparing results of studies with baseline values, the authors were able to figure out the benefit of providing correct recommendations versus the risk of negative actions associated with incorrect actions.
Discussion:
I am a little skeptical about this study. It definitely has good intentions and visions for future work, but there were too many variables that go into studies like these. The analysts could have just used their own judgement the entire time and completely discarded each recommendation, a user could have been completely biased by a an option on the screen, etc. I didn't understand a lot of some of the ways in which the authors went about interpreting their results. For example, they "divided" people into 3 fields: low, medium, and high? I wasn't sure what some of that jargon meant or what criteria was used to divide the users up. This kind of technology seems a bit limited for general purpose or release. The authors seemed happy about the findings and results of their studies, so I suppose that the authors achieved their goals. I'm not too sure, honestly.

Paper Reading #31- Identifying emotional states using keystroke dynamics

Title: Identifying emotional states using keystroke dynamics
Reference Information:
Clayton Epp, Michael Lippold, and Regan Mandryk, "Identifying emotional states using keystroke dynamics". CHI '11 Proceedings of the 2011 annual conference on Human factors in computing systems. ACM New York, NY, USA. ©2011. ISBN: 978-1-4503-0228-9.
Author Bios:
Clayton Epp- Senior Software Developer at University of Toronto, Saskatchewan, Canada. Computer Software.
Michael Lippold- Principal / Systems Engineer of Smartacus, Inc. Portland, Oregon.
Regan Mandryk- I’m an Assistant Professor in the Department of Computer Science at the University of Saskatchewan.
Summary:
  • Hypothesis: If the authors can conduct studies on keystrokes made by users and determine stroke patterns, then it is possible to determine emotional states of the user while stroking.
  • Methods: Two main areas of research are the focus for this paper: data collection and data processing to determine emotional state. The authors used ESM (experience-sampling methodology) for their studies. In ESM, users are asked to periodically record their experiences along with their every day activities.This allows data collection "in the moment" rather than retrospectively. Throughout their day, users would be prompted to type text and record their emotional state, as well as be prompted with keystrokes that they've made from the previous 10 minutes to the prompt. Users could potentially opt out of this collection if they didn't choose to participate at that time for whatever reason. Once data was collected, three features of the data were considered (keystoke/content, emotional state classes, and additional data points). Keystroke features included key press/release events as well as associated timestamps. Emotional state classes refers to the labeling of the users responses with a discrete emotional class via Likert scale responses. Additional data points refers to additional contextual information for each user (i.e. a user taking a break from the computer, a user switching to the mouse, finding out all process names the user is working on, etc).
  • Results: The authors used decision trees and a form of supervised learning to classify user data into emotional classifiers. As far as classifying, some content was thrown away because of skewed data (data belonging predominantly to extremes on the Likert scale). The authors showed successfully that their system can correctly classify at least two levels of seven emotional states.
  • Content: The authors wanted to create a system to classify user's emotional states simply be analyzing keystroke patterns and features. They wanted to make an inexpensive, unintrusive system as well. The authors were able to create a system to correctly classify emotional states on some level to a high degree of accuracy (77-84%). The authors had to throw away a lot of data because of a lack of responses from users, they had to agglomerate data because of skewed data, and they had to discard some classifiers in their results because of accuracy.
Discussion:
This paper was all right. I like the idea of being able to classify a user's emotion by keystrokes, but I doubt the accuracy of such a system because there are so many factors that go into classifying that. For example, I can type fast because I am angry or because I just naturally type fast and don't necessarily have any extreme emotion at the moment. This system could definitely use some improvement, but I suppose they are on the right track. The authors, in my opinion, achieved their goals somewhat because they found a foundation for classification, but their classification isn't anywhere near solid. They did, however, find a lot of areas for improvement. I'll believe this kind of technology and its success when I see it.

Paper Reading #29- Usable gestures for blind people: understanding preference and performance

Title: Usable gestures for blind people: understanding preference and performance
Reference Information:
Shaun Kane, Jacob Wobbrock, and Richard Ladner, "Usable gestures for blind people: understanding preference and performance". CHI '11 Proceedings of the 2011 annual conference on Human factors in computing systems. ACM New York, NY, USA. ©2011. ISBN: 978-1-4503-0228-9.
Author Bios:
Shaun Kane- I am an assistant professor in the Department of Information Systems at the University of Maryland, Baltimore County. My primary research interests are accessible user interfaces and mobile human-computer interaction. My work explores ways to make mobile devices easier to use, especially for people with disabilities and people in distracting environments.
Jacob Wobbrock- I am an Associate Professor in the Information School and an Adjunct Associate Professor in the Department of Computer Science and Engineering at the University of Washington.
Richard Ladner- Boeing Professor in Computer Science and Engineering in the Department of Computer Science and Engineering, the University of Washington.
Summary:

  • Hypothesis: If designers can better understand how blind people interact with a touch-screen device as opposed to people with sight, then products can be better suited and accommodated for blind people.
  • Methods: The authors conducted two studies to better understand the contrast between blind people and sighted people in terms of interactions with a touch-surface device. In the first, the authors asked 10 blind people and 10 sighted people to invent gestures to complete a task on a tablet PC. In the second, the authors asked the same participants to perform reference gestures to perform the tasks.
  • Results: From the first study, the participants each created 2 possible gestures for each task. They ended up rating their grstures based on "easyness" and "good match". There were no significant differences between the blind and the sighted here. A count revealed that the blind people's gestures contained a significantly higher gesture count than the sighted people's. Also, the blind people tended to make more abstract gestures for tasks whereas sighted people made more symbolic gestures. For the second study, some blind participants either opted to skip some gestures because of unfamiliararity or the results were discarded because they were not accurate whatsoever. Of the gestures performed, blind people and sighted people typically rated the "goodness" of the gestures around 5.7 of 10. The authors found that blind people tended to create larger areas or shapes for their gestures than sighted people. The size deviation also indicated that the variation of size was much greater with blind people than with sighted people. The authors found that the blind people took almost twice as long to complete the tasks on average as the sighted people. Blind people had generally a greater inaccuracy with locating specific spots on the screen than sighted people. It was also noted that blind people's gestures were not as smooth or fluid as sighted people's.
  • Content: The study that the authors conducted found that blind people prefer gestures involving the areas on the screen closer to the edges and that require multiple touches to complete. The authors also uncovered significant performance differences between gestures performed by the blind vs the sighted (noted above). The authors concluded with several notes for future designers: "avoid symbols used in print writing, favor edges, corners, and other landmarks, reduce demand for location accuracy, and limit time-based gesture processing".
Discussion:
I think this paper was really interesting. This becomes a real issue when a company shows that they are taking interest in incorporating as large of a population as possible and not punishing the blind for not being able to use one of their products. This is definitely one applicatory situation for the motivation behind this paper. The authors, in my mind, deifnitely achieved their goals by allowing future designers for touch-surface devices to see the differences between sighted and blind people's interactions. Some things need to be taken into account when designing new products that you wish to incorporate the blind (noted above). Some of the things, such as accuracy of the gestures being less for the blind, seemed really obvious to me and simply studying that could have been a little waste of time. But I guess in order to know for sure, it had to be studied. I would love to see a finished prototype of a device using these princicples in future work.

Paper Reading #28- Experimental Analysis of Touch-Screen Gesture Designs in Mobile Environments

Title: Experimental Analysis of Touch-Screen Gesture Designs in Mobile Environments
Reference Information:
Andrew Bragdon, Eugene Nelson, Yang Li, and Ken Hinckley, "Experimental Analysis of Touch-Screen Gesture Designs in Mobile Environments". CHI '11 Proceedings of the 2011 annual conference on Human factors in computing systems. ACM New York, NY, USA. ©2011. ISBN: 978-1-4503-0228-9.
Author Bios:
Andrew Bragdon- currently a second-year Ph.D student in Computer Science at Brown University. Andrew has worked at Microsoft, traveled around giving special lectures, and published multiple works in the field of Human-Computer Interactions within Computer Science.
Eugene Nelson- Professor, Department of Community, Family Medicine and The Dartmouth Institute, Dartmouth Medical School, Hanover, NH.
Yang Li- Yang is a Senior Research Scientist at Google. Before joining Google's research team, Yang was a Research Associate in Computer Science & Engineering at the University of Washington and helped found the DUB (Design:Use:Build), a cross-campus HCI community. He earned a Ph.D. degree in Computer Science from the Chinese Academy of Sciences, and then did a postdoctoral research in EECS at the University of California at Berkeley.
Ken Hinckley- a Principal Researcher at Microsoft. He has been a part of many publications of relevance in the past. His Ph.D work involved developing a props-based interface for neurosurgeons.
Summary:
  • Hypothesis:
  • Methods:
  • Results:
  • Content:
Discussion:

** I decided to use this blog as my "freebie" for doing the blog on autism earlier in the semester **

Wednesday, November 9, 2011

Paper Reading #30- Life "modes" in social media

Title: Life "modes" in social media
Reference Information:
Fatih Ozenc and Shelly Farnham, "Life "modes" in social media". CHI '11 Proceedings of the 2011 annual conference on Human factors in computing systems. ACM New York, NY, USA. ©2011. ISBN: 978-1-4503-0228-9.
Author Bios:
Fatih Ozenc- I have just completed my Ph.D. studies on interaction design, at Carnegie Mellon's School of Design.
Shelly Farnham- Shelly is a researcher in the FUSE labs at Microsoft. She has worked for Yahoo in the Communications and Communities Department.
Summary:
  • Hypothesis: If the authors can effectively manage a user's social media interactions across different "modes" of their lives (family, work, friends, college, sports, etc), then the experiences and services of the social media interactions will be maximized.
  • Methods: The authors performed in-depth, two-hour interviews to explore how people naturally mentally model different areas of their lives, how they incorporate communication technologies to support them, and how we might improve their online experiences of managing their social media streams. The authors created a UI that can aggregate all social media information sharing into a single mechanism that can further create information "boundaries" between facets of an individual's life as well as create transitionary models between "modes". The authors then conducted a study on separate individuals scoring highly on extroversion and having multi-faceted lives so that their feedback would be effective toward creating "division" mechanisms.
  • Results: Participants either sketched out their lives as "social memes" (resembling network graphs- having nodes and edges representing parts of life such as activities, groups, work, etc instead of individual interactions) or as "timeline memes" (such as a basic rundown on their daily routine or their activities as time has incremented. These resembled collages.).  "Family", "Friends", and "Work" were the three most dominantly present facets of life in the participant sketches. Users were found to have chosen different communication means with each "facet" of their lives based on the intimacy and closeness of the relationships in the different spaces (For example, someone would use their e-mail as a primary contact method for work as opposed to a cell phone for their family). Participants also described their transition "modes" as a sort of "ritual" or physical movement (i.e. "transitioning" from work to home involved a car ride). The participants favored having an organized and focused information sharing mechanism for their multi-faceted lifestyles.
  • Content: The authors wanted to create a way to have information boundaries for different "modes" of life for the user. Their results indicate that people with higher levels of faceted identity have the problem of organizing, sharing and consuming online, especially while managing and transitioning between family, work, and social life modes. They strategically use communication technologies to manage intimacy levels within these modes, and levels of permeability across the boundaries between these modes. For future efforts designing social media products, we recommend designers and researchers think about the user’s communication ecology holistically, consider life ‘modes’ as one of the organizing principles, prioritize focused sharing and the mobile medium, and incorporate cues and signals when designing for transitions.
Discussion:
I really enjoyed the ideas and concepts of this paper. I believe most people will actually enjoy using this sort of division mechanism. I myself don't care for it really. I'm not a private, "multi-personality" person at all. I don't mind who sees what about me. Mainly because I want people to know all about me. I enjoy letting people into my life and getting into other people's lives. Like I said, this is definitely a good idea, though. I also like HOW they went about their study. The only good way, in my opinion, that they were ever going to get the design part right is to place it in front of actual sample users that would be (potentially) benefitting from the proposed technology. I believe the authors achieved their goals and didn't simultaneously. They came up with good metrics to use for future designs, but proposed no good ways to go about doing it literally (in code or algorithms, i.e.). This paper was neat.

Wednesday, November 2, 2011

Paper Reading #27- Sensing cognitive multitasking for a brain-based adaptive user interface

Title: Sensing cognitive multitasking for a brain-based adaptive user interface
Reference Information:
Erin Solovey, Francine Lalooses, Krysta Chauncey, Douglas Weaver, Margarita Parasi, Matthias Scheutz, Angelo Sassaroli, Sergio Fantini, Paul Schermerhorn, Audrey Girouard, and Robert Jacob, " Sensing cognitive multitasking for a brain-based adaptive user interface". CHI '11 Proceedings of the 2011 annual conference on Human factors in computing systems. ACM New York, NY, USA. ©2011. ISBN: 978-1-4503-0228-9.
Author Bios:
Erin, Francine, Krysta, Douglas, Margarita, Matthias, Angelo, Sergio, and Robert are all associated with Tufts University in Massachusetts, USA.
Paul is associated with Indiana University.
Audrey is associated with Queen's University in Ontario, Canada.
Summary:
  • Hypothesis: If the authors can create a system to detect a user's "to-do list" and allow them to multitask on different things at once, then those tasks will be completed faster, the user will become "understood" by the system, and a new kind of technology will be effectively used.
  • Methods: While constructing their functional near-infrared spectroscopy (fNIRS), the authors took into account the three multitasking scenarios of the brain: branching, dual task, and delay. Branching is when you hold in mind goals while exploring and processing secondary goals. For example, you are working on your homework when a friend e-mails you and you begin to read the e-mail (while still remembering to return to your homework after you finish). Dual Task is when you have two tasks that require additional resources to complete. For example, a network technicial is fixing issues with his company's network while responding to important e-mails (possibly answering questions about the network). Finally, Delay is when a primary task is being worked on and a secondary task gets ignored. For example, you are watching a movie on your laptop when you see a notice come up that you have an e-mail. You ignore this notice. This is called delay because the secondary task gets delayed by the primary task. The authors were curious to see if they could tell the difference, cognitively, between the three kinds of multitasking. So they strapped some volunteers with gear and tested to see if they could while users performed given tasks (obviously multitasking was involved). After the preliminary study, the authors conducted a second study with the same methodologies in different spaces more relevant to HCI (for example, users were required to sort rocks by their type from Mars while keeping track of the position of a robot). The authors also launched a third study involving random vs predictive branching (the robot would move randomly in terms of the number of rock types were displayed vs the robot would move after every three presentations of rock types).
  • Results: The preliminary study returned a recognition accuracy of 68%. To the authors, this was promising. In the second study with the robot and the rocks, any result where the participant achieved less than a score of 70% were discarded because it was seen as the task being done incorrectly. In the last study, there was no significant statistical difference found between the random and predictive branching. The authors were able to construct a proof-of-concept model because they were able to differentiate between the three types of tasking and incorporate machine learning into it.
  • Content: The authors wanted to be able to create a system to measure and handle multitasking mechanisms. They were able to differentiate between three types of multitasking, create studies to measure efficiency for each kind of multitasking technique, test their system, and prove that it works as well.
Discussion:
I think these guys are geniuses. The technology is super advanced and the methodologies were complex when dealing with the proof-of-concept system. I don't know how much effect this will have in the HCI field (at least that I can see) because it didn't seem to me from reading this that there was much of a "new invention" or "new technology" here. They were able to recognize and "quantify" multitasking, but aside from that, I'm not sure how this could be applied. Maybe I missed it. I think the authors definitely achieved their goals, though. They said that all they wanted to do was be able to handle cognitive multitasking, which they were able to do. I'm indifferent about this article, honestly.

Monday, October 31, 2011

Paper Reading #26- Embodiment in brain-computer interaction

Title: Embodiment in brain-computer interaction
Reference Information:
Kenton O'Hara, Abigail Sellen, and Richard Harper, "Embodiment in brain-computer interaction". CHI '11 Proceedings of the 2011 annual conference on Human factors in computing systems. ACM New York, NY, USA. ©2011. ISBN: 978-1-4503-0228-9.
Author Bio:
Kenton O'Hara- Microsoft Researcher in Cambridge in the Department of Socio Digital Systems.
Abigail Sellen- Abigail Sellen is a Principal Researcher at Microsoft Research Cambridge in the UK and co-manager of Socio-Digital Systems, an interdisciplinary group with a focus on the human perspective in computing.
Richard Harper- Richard Harper is Principal Researcher at Microsoft Research in Cambridge and co-manages the Socio-Digital Systems group.
Summary:
  • Hypothesis: If BCI technology can be effectively used and harnassed, then it can be used outside of the realm of just video gaming and possibly with the combination of other technologies.
  • Methods: The authors decided to test the effectiveness of BCI with a video game called MindFlex. In the game, you keep a ball floating by concentrating (not just on the ball..on anything really. Higher concentration sends a higher "fan level" to the game. The higher the fan level, the higher the ball was kept up). The authors conducted a user study on participants (of at least two in a normal social setting) playing the game in a trial run to measure social relationships, interactions with the game, and coordination of the gameplay.
  • Results: The results of the video stream and data from gameplay indicate strategies were taken by participants to keep the ball up (they would hold certain poses, do certain things, etc...because they thought it would help them). Other things such as gaze, intent, vision, imagination, proximity to the ball, and gestures all played roles in gameplay. Spectators also played roles in how the players played the game. Sometimes they would joke with the player or sometimes they would help them play.
  • Content: The authors wanted to create an environment where the limits and applications of a BCI game were recordable. To do this, they took a game called MindFlex and gave it to participants to play for a week to later study how the game was played in terms of group size, strategy, focus, gestures, time spent playing, etc. They found that each user would approach the game uniquely and had a different style of play. Even the spectators of the game had a role to play in the gameplay of the player.
Discussions:
I thought this was kind of cool. I think I would have fun if I were able to play a game like this. I wonder if it would be difficult or not. I noticed that some players had some difficulty in controlling the ball sometimes. I wonder what kind of gestures / movements / strategies I would take to do well in the game. I don't think spectators would have an effect on me at all. I think that the authors definitely showed the the power of BCI technology is tangible and applicable to (at least) video gaming. Like the authors said, there are many other avenues of application for technologies like BCI. I think the authors achieved their goals for sure.

Paper Reading #25- Twitinfo: aggregating and visualizing microblogs for event exploration

Title: Twitinfo: aggregating and visualizing microblogs for event exploration
Reference Information:
Adam Marcus, Michael Bernstein, Osama Badar, David Karger, Samuel Madden, and Robert Miller, "Twitinfo: aggregating and visualizing microblogs for event exploration". CHI '11 Proceedings of the 2011 annual conference on Human factors in computing systems. ACM New York, NY, USA. ©2011. ISBN: 978-1-4503-0228-9.
Author Bio:
Adam Marcus- I am a graduate student at the MIT Computer Science and Artificial Intelligence Lab (CSAIL), where I am part of the Database group and the Haystack group. My advisors are Sam Madden and David Karger.
Michael Bernstein- I am a final-year graduate student focusing on human-computer interaction at MIT in the Computer Science and Artificial Intelligence Lab. I work with Professors David Karger and Rob Miller.
Osama Badar- Graduate student at MIT in the CSAIL.
David Karger- I am a member of the Computer Science and Artificial Intelligence Laboratory in the EECS. department at MIT.
Samuel Madden- Sam is an Associate Professor in the EECS department at MIT. He is also a part of the CSAIL group.
Robert Miller- I'm an associate professor in the EECS department at MIT, and leader of the User Interface Design Group in the Computer Science and Artificial Intelligence Lab.
Summary:
  • Hypothesis: If the authors can organize and effectively communicate the timeline-based display information of TwitInfo, then Twitter's existing implementation of information will be able to be more effectively manipulated and enhanced.
  • Methods: The authors create an algorithm to automatically "browse an event" and label the created event (which was done using a keyword) if a certain event is tweeted about over a certain amount of times per time unit. This will allow the activity on a timeline to "peak" to show that users are tweeting about an event. The authors were also able to implement a "sentiment analysis" about certain events (i.e. is this event "positive" or "negative" judging by user's comments and feedback about the event). TwitInfo also allows the creation and browsing of subevents. TwitInfo was evaluated against three soccer matches as well as a month-long collection of raw data. The authors then recruited 12 people to evaluate the UI of TwitInfo.
  • Results: TwitInfo is biased by the Twitter users' interests. It was effective in measuring events for the soccer matches as well as the earthquakes. Occassionally, TwitInfo returned false positives such as anyone tweeting about a soccer term that wasn't necessarily about the soccer match being observed. In the study of understanding the UI, the results concluded that TwitInfo is an effective source for news without even any prior knowledge of events. Users were able to find events on the timeline and read summaries of sub-events. Users didn't necessarily agree with the sentiment analysis, however.
  • Content: The authors wanted a better way to browse Twitter events and effectively get caught up on events and relevant news for a keyword event. The created a UI which incorporated a timeline-based display where users could browse events with "high peaks" of interest and get caught up on them. The authors also created an engine to determine whether an event was a positive thing or a negative thing for each part of the world.
Discussions:
I think this is very interesting, but as an avid anti-Twitter-er, I won't be using this technology. I have nothing against Twitter, I just got a Facebook first. The way the authors were able to take all of the information from Twitter and personal tweets and organize it all into meaningful, timelined events that users could browse is a very good idea though. I believe the authors weren't compltely happy with their results as far as the user evaluation and journalist evaluation of TwitInfo went, but I'd say that they achieved their goals. The technology that they created is definitely better than what was already in-place for Twitter, anyway.

Thursday, October 27, 2011

Paper Reading #24- Gesture avatar: a technique for operating mobile user interfaces using gestures

Title: Gesture avatar: a technique for operating mobile user interfaces using gestures
Reference Information:
Hao Lu and Yang Li, "Gesture avatar: a technique for operating mobile user interfaces using gestures". CHI '11 Proceedings of the 2011 annual conference on Human factors in computing systems. ACM New York, NY, USA. ©2011. ISBN: 978-1-4503-0228-9.
Author Bios:
Hao Lu- I am a graduate student at University of Washington Computer Science & Engineering and DUB Group. I work on technologies that change and improve the way how people interact with computers.
Yang Li- Yang is a Senior Research Scientist at Google. Before joining Google's research team, Yang was a Research Associate in Computer Science & Engineering at the University of Washington and helped found the DUB (Design:Use:Build), a cross-campus HCI community. He earned a Ph.D. degree in Computer Science from the Chinese Academy of Sciences, and then did a postdoctoral research in EECS at the University of California at Berkeley.
Summary:
  • Hypothesis: If using gesture methods to operate mobile interfaces performs optimally over finger-based touch input, then in a dynamic, mobile environment, it will prevail.
  • Methods: To correctly and accurately model a user's intent for interaction on a page, the shape of the gesture and the distance from the gesture to the objects aer taken into consideration. After the user draws a gesture, an "avatar" bounding box appears translucently behind the gesture. The user can then interact with that avatar the same way as they would their desired object. Only this avatar is larger than the object on the page so it will be easier to interact with. If the object that gets highlighted is not the desired object, the user may disgard that by using several techniques. The authors conducted a study among 20 participants involving half of them learning gesture avatar and then shift, and vice versa. They were asked to complete tasks using each software. For some tasks, users were sitting down, and for others, they were on a treadmill (to simulate using the device while walking).
  • Results: From the study, the authors found that Gesture Avatar was slower than Shift for 20px targets, but faster for 10px targets. Users using Shift while sitting were much faster than while they were walking on the treadmill, but for Gesture Avatar there was no significant difference in time reported. Gesture Avatar's error rates were lower than Shift's for all target sizes. 10 out of the 12 participants preferred using Gesture Avatar over Shift. The authors found that increasing the size of unique characters that would be drawn had no affect on Gesture Avatar's performance. Gesture Avatar supports one-shot interaction and acquiring moving targets also. There have been some integration issues with using Gesture Avatar in existing systems, however.
  • Content: The authors wanted to create a way to effectively use your mobile device accurately while in a mobile environment. Touch-activated UIs can be highly inaccurate because of "fat-finger" and "occulation" problems. To correct for these, the authors created a technique to allow the user to "call" a desired object to interact with by "describing it" with a gesture in order for Gesture Avatar to enlarge it so it can be mroe accurately interacted with (and to ensure the correct object is being interacted with). After studies, the authors showed that Gesture Avatar is superior to other similar existing systems.
Discussion:
I thought this was kind of a neat idea. At first, I thought "Oh man, Yang Li is at it with another dumb gesture recognition idea again..." but I was not entirely correct. This time, I believe the proposed system can be extremely useful- especially in a mobile environment. I also liked how the different ways to ensure the correct object is selected is implemented as well. Gesture Avatar allows for disregarding incorrect objects, finding the next best match, interacting more accurately with desired objects, and allowing existing features of a UI to continue to be used if the user didn't want to use Gesture Avatar techniques on a particular page. I might consider using something like this if I did have a smartphone / touch-screen phone. I think Yang and Hao achieved their goals. It always feels good to know something you created out-performs something in existance already.

Monday, October 24, 2011

Paper Reading #23- User-defined motion gestures for mobile interaction

Title: User-defined motion gestures for mobile interaction
Reference Information:
Jamie Ruiz, Yang Li, and Edward Lank. "User-defined motion gestures for mobile interaction". CHI '11 Proceedings of the 2011 annual conference on Human factors in computing systems. ACM New York, NY, USA. ©2011. ISBN: 978-1-4503-0228-9.
Author Bios:
Jamie Ruiz- I'm a fifth-year doctoral student in the Human Computer Interaction Lab in the Cheriton School of Computer Science at the University of Waterloo. My advisor is Dr. Edward Lank.
Yang Li- Yang is a Senior Research Scientist at Google. Before joining Google's research team, Yang was a Research Associate in Computer Science & Engineering at the University of Washington and helped found the DUB (Design:Use:Build), a cross-campus HCI community.
Edward Lank- I am an Assistant Professor in the David R. Cheriton School of Computer Science at the University of Waterloo.
Summary:
  • Hypothesis: If smartphones contain devices that allow for 3D tracking and sensing of the phone, then a set of optimal gestures to invoke commands with very natural mapping exists.
  • Methods: The authors conducted a "guessability study" on participants by having them perform tasks and asking them what gesture / motion allows for the optimal mapping. For the experiment, users were told to treat the smartphone as a "magic black brick" because the authors removed all recognition technology from it so the users wouldn't possibly be influenced by anything of that nature. They were told to create gestures for performing tasks from scratch. The participants were recorded via audio and video. Data was also collected from a software on the phone for a "what was the user trying to do?" perspective. The study was conducted over all users who have previous and relevant smartphone experience.
  • Results: Users most commonly used gestures not related to smartphone gestures. For example, viewing the homescreen for the smartphone had a very popular gesture as a result involving shaking the phone (as with an etch-a-sketch). Generally, the results gathered from the authors had a general agreement among gestures as well as the reasoning for the gestures. The authors therefore were allowed the luxury from the video "out-loud" process data to understand the user's thought process when creating the gesture. Tasks that were considered to be "opposite" of each other had similar gesture motions but were performed in "the opposite direction" for example. For example, zooming in and out from a map would involve moving the phone closer to you or farther away, respectively. The authors then went into a detailed study about what kind of gesture was created for performing tasks and what was involved / how the phone was treated for the gesture. "Agreement scores" were then calculated for each gesture to quantitatively find how "good" each individual gesture users made were.
  • Content: The authors wanted to be able to create more natural, easy-to-use gesture sets for interactions with motion for smartphones versus plain gesture interactions. To create such a set, the authors studied volunteers who use smartphones which created their own gestures for performing tasks on smartphones. These gestures were than individually taken and measured against every other participants to see how "good" each gesture created was. From here, the specific kind of interaction and what all was involved/manipulated was taken into account also to see what general kind of pattern was recognizable.
Discussion:
I love studies like this where there is no field-specific jargon, no technical processes, no high-level fancy vernacular to have to learn, and no difficult end-goal in mind- this was very straightforward. The authors wanted to create a way to have simple and optimal motion gestures to interact with a smartphone. What better way to do it than to let a sample set of smartphone users "create" such a set? I definitely believe the authors achieved their goals and I would definitely consider using a smartphone with motion-based capabilities of the motions were based from studies like these. I enjoyed reading this paper.

Paper Reading #22- Mid-air pan-and-zoom on wall-sized displays

Title: Mid-air pan-and-zoom on wall-sized displays
Reference Information:
Mathieu Nancel, Julie Wagner, Emmanuel Pietriga, Olivier Chapuis, and Wendy Mackay. "Mid-air pan-and-zoom on wall-sized displays". CHI '11 Proceedings of the 2011 annual conference on Human factors in computing systems. ACM New York, NY, USA. ©2011. ISBN: 978-1-4503-0228-9.
Author Bios:
Mathieu Nancel- I am a Ph.D. student in Human-Computer Interactions in the in | situ | team (INRIA/LRI) since september 2008. I work on distal interaction techniques for visualization platforms: precise pointing, navigation in very large datasets, menu techniques, etc.
My Ph.D. supervisors are Michel Beaudouin-Lafon and Emmanuel Pietriga.
Julie Wagner- She is a Ph.D Student at INRIA. Her supervisor is Wendy Mackay.
Emmanuel Pietriga- Chargé de Recherche - CR1, interim leader of INRIA team In Situ.
Full-time research scientist working for INRIA Saclay - ÃŽle-de-France.
Olivier Chapuis- Chargé de recherche (Research Scientist) CNRS at LRI (CNRS & Univ. Paris-Sud).
Member and team co-head (by interim) of the InSitu research team (LRI & INRIA Saclay Ile-de-France).
Wendy Mackay- She is currently a Research Director with INRIA Saclay in France, currently on sabbatical at Stanford University. She runs a research group called in|situ|. Their focus is on the design of innovative interactive systems that truly meet the needs of their users.

Summary:
  • Hypothesis: If the authors can study and trial different methods for large-scale wall-display-sized navigation, then they will create an optimal gesture set suitable for real and complex applications dealing with such a problem space. The authors had 7 separate hypotheses about each particular area of their design, but I combined this to make one general hypothesis.
  • Methods: The authors conducted a series of pilot tests, pursued research, and performed empirical studies to narrow down all possible gesture and input methods for this system down to 12. They took into account performance (cost and accuracy), fatigue over periods of use of the input, ease of use, and natural mapping. The authors conducted an experiment to evaluate each of the 12 factors they discussed using in their system to see which were optimal. The authors had their ideas about which of the 12 were optimal to begin with, but they tested their ideas to see if this was the case.
  • Results: After the experiment, the authors found that the two-handed gesture tasks were performed faster than the one-handed gesture tasks, involving smaller muscle groups for input interactions improves performance (providing higher guidance further improves it), and linearly-performed tasks were generally performed faster than circular ones. Circular gestures were slower because it was more often that participants overshot their target with circular gestures than with linear ones. After receiving feedback from the participants, they found that what the users were saying were (in general) in agreeance with their results from the data gathered.
  • Content: The authors wanted to study a field that has historically received little attention. They wanted to figure out a way to effectively explore a wall-sized interaction surface with respect to performance, ease, and minimal fatigue. The authors eventually found (via studies and pilot testing) an optimal set of interactions and gestures suitable for such an interaction space.
Discussion:
I thought this paper was pretty interesting. This field of study has a lot of real-world applications such as (as mentioned in the paper) crisis management/response, large-scale construction, and security management areas. I thought their studies were accurate, appropriate, and extensive enough to gather relevant and meaningful data for designing a suitable system for the display. They definitely achieved their goals, in my opinion. I would have loved to have been a participant in this particular experiment; I think it would have been fun. I would consider using a technology like this if I was a billionaire and wanted to play a video game on my wall where I was a controller or something.

Thursday, October 20, 2011

Paper Reading #21- Human model evaluation in interactive supervised learning

Title: Human model evaluation in interactive supervised learning
Reference Information:
Rebecca Fiebrink, Perry Cook, and Dan Trueman. "Human model evaluation in interactive supervised learning". CHI '11: Proceedings of the 2011 annual conference on Human factors in computing systems ACM. New York, NY, USA. ©2011. ISBN: 978-1-4503-0228-9.
Author Bios:
Rebecca Fiebrink- I will be joining Princeton as an assistant professor in Computer Science and affiliated faculty in Music in September, 2011. I have recently completed my PhD in Computer Science at Princeton, and I will be spending January through August 2011 as a postdoc at the University of Washington. I work at the intersection of human-computer interaction, applied machine learning, and music composition and performance.
Perry Cook- Professor Emeritus* (still researches but no longer teaches or accepts new graduate students) at Princeton University in the department of Computer Science and Dept. of Music.
Dan Trueman- Professor in the department of Music at Princeton University.
Summary:
  • Hypothesis: If the user can be allowed to iteratively update the current state of a working machine learning model, then the results and actions taken from that model will be improved (in terms of quality).
  • Methods: The authors conducted three studies of people applying supervised learning to their work in computer music. Study "A" was a user-centered design process, study "B" was an observatory study in which students were using the Wekinator in an assignment focused on supervised learning, and study "C" was a case study with a professional composer to build a gesture-recognition system.
  • Results: From the studies, the authors gathered results and analyzed all of the results. From study "A", the authors saw that participants iteratively re-trained the models (by editing the training dataset), for study "B", the students re-trained the algorithm an average of 4.1 times per task, and the professional from study "C" re-trained it an average of 3.7 times per task. Cross-validation wasn't used in study "A", but for studies "B" and "C", cross-validation was used an average of 1 and 1.8 times per task, respectively. Direct evaluation was also present in the evaluation metric of the system. This was used more frequently than cross-validation. Participants in "A" strictly used this measure, while the students and the professional in studies "B" and "C" used direct evaluation on an average of 4.8 and 5.4 times per task, respectively. Using cross-validation and direct evaluation, users were able to receieve feedback on how their actions affected the outcomes. The overall results were that users were able to fully understand and use the system effectively. The wekinator allowed users to create more expressive/intelligent/quality models than with other methods/techniques.
  • Conent: The authors wanted to create some method to allow users to provide feedback to the system iteratively while it is being used in order to hopefully create some iterative machine learning mechanism. The authors conducted user studies to test their hypothesis and collected results that suggested that the methods provided by the authors to perform tasks were superior to other techniques.
Discussion:
I thought this article was sort of interesting. It is a really good idea to have a system where you can tell the machine what is "good" or "bad" before it even spits out the final result. Being able to re-train your algorithm (mid-computation) is really beneficial to have. The cost-benefit for it seems reasonable, so I could see this idea becoming more widespread before long. The authors, in my opinion, definitely achieved their goals and proved their hypothesis to be true. I didn't understand cross-evaluation or direct evaluation in terms of the actual methods too much, but I know those factors were taken into consideration when collecting data for "satisfaction" of the system.

Monday, October 17, 2011

Paper Reading #20- The aligned rank transform for nonparametric factorial analyses using only anova procedures

Title: The aligned rank transform for nonparametric factorial analyses using only anova procedures
Reference Information:
Jacob Wobbrock. Leah Findlater, Darren Gergle, and James Higgins. "The aligned rank transform for nonparametric factorial analyses using only anova procedures". CHI '11: Proceedings of the 2011 annual conference on Human factors in computing systems. ACM, New York, NY, USA. ©2011. ISBN: 978-1-4503-0228-9.
Author Bios:
Jacob Wobbrock- He is an Associate Professor in the Information School and an Adjunct Associate Professor in the Department of Computer Science & Engineering at the University of Washington.
Leah Findlater- Is an Undergraduate Research Advisor for the Universitty of Washington. Has served on projects related to HCI.
Darren Gergle- He is an Associate Professor in the departments of Communication Studies and Electrical Engineering & Computer Science (by courtesy) at Northwestern University. He's also direct the CollabLab: The Laboratory for Collaborative Technology.
James Higgins- A professor in the department of statistics at Kansas State University.
Summary:
  • Hypothesis: If the Aligned Rank Transform (ART) method can be implemented for statistical analysis and nonparametric data, then studies concerning interactions and other complex factors can be made more quantifiable and easier to manage.
  • Methods: The ART method "aligns" the data before applying averaged ranks. Then, the common ANOVA procedures can be used. ARTool and ARTweb are also tools for assisting in using the ART method on relative studies. Unlike other statistical methods, ART doesn't violate ANOVA assumptions, inflate type I error rates, forego interaction effects, or disregard "too complex" of study data. the authors also made ART relatively easy to interpret- anyone familiar with ANOVA can interpret the data. The ART method can also consider "N" factors in its analysis. ART has 5 steps: 1) computing residuals 2) computing estimated effects for all main and interaction effects 3) compute aligned response Y' 4) assign averaged ranks Y'' 5) perform a full factorial ANOVA on Y''. Correctness is ensured on ART by ensuring that each column of data for Y' sums to 0 and by showing that the full-factorial ANOVA performed on Y' show all effects stripped out except for the ones which the data were aligned for initially. ARTool and ARTweb allow for simple interactions and clear data results.
  • Results: Findlater collected satisfaction ratings on accuracy and interface from 24 participants. The authors found that their creation was satisfactory among both fields with the users. The authors found some limitations for ART also (extreme skew, tied ranks, not randomized designs). 
  • Content: The authors wanted to create a way to handle complex data in experiments using familiar tests and data analysis already in place. ART is favored over traditional methods because of its simplicity and usability. ARTweb and ARTool allow for convenience of using ART on data and models with a computer and make ranking and alignment simple.
Discussion:
I know next to nothing about statistics. I took one course on it and forgot nearly everything relatively fast. I got lost pretty quickly in the paper, basically. It seems, however, that the field of HCI needed a unique way to interpret data in studies concerning (say) measuring if a device was "good" or not (whether users deemed it as "satisfactory" based on certain metrics). This seems to be a great idea and is also coupled with programs that allow for even easier usability with ART. This way, anyone with a basic knowledge of ANOVA and statistics can understand the ART results. I definitely believe that the authors achieved their goals because their method and programs are probably used commonly in these target types of experiments. It is good for authors to be able to see limitations on their own inventions, too. This way they can disclaimer it or they can release a "2.0" to correct some of these issues. If I needed to do some study concerning HCI interactions, I would definitely consider using ART.

Paper Reading #19- Reflexivity in digital anthropology

Title: Reflexivity in digital anthropology
Reference Information:
Jennifer Rode, "Reflexivity in digital anthropology". CHI '11: Proceedings of the 2011 annual conference on Human factors in computing systems ACM. New York, NY, USA. ©2011. ISBN: 978-1-4503-0228-9.
Author Bios:
Jennifer Rode- She is an Assistant Professor at Drexel's School of Information, as well as a fellow in Digital Anthropology at University College London.
Summary:
  • Hypothesis: If people can better understand  how to effectively use the different types of ethnographies, then the field of HCI will become clearer and more advanced as well as suited to target user audience.
  • Methods: Jennifer Rode really had no methods in her paper. She argued that there are different types of ethnographies and defined each of them. She also argues that in each scenario a certain type of ethnography is optimal. The results section outlines these in more detail.
  • Results: She defines reflexivity as having four main characteristics: intervention is seen as a data gathering opportunity, understanding how the method for data gathering actually impacts the data that is gathered, find structural patterns in what was observed, and extending theories. She claims that having reflexivity is crucial in ethnographies. She defines three types of ethnographies: realist, confessional, and impressionistic. A realistic ethnographical approach has four keys to it: experimental authority (in which the ethnographer, over a period of time, becomes familiar with his new environment and can no longer have to make inferences about it- rather make testable hypotheses), its typical forms, the native's point of view (from which the ethnographer hopes to study closely and represent accurately), and interpretive omnipotence (the author has the final say on what is written and recorded- basically he controls how the target group is perceived and represented). A confessional ethnographical approach broadly provides the ethnographer with doubts/faults concerning the study and allows them to answer them. Lastly, the impressionistic ethnographical approach recalls certain events uniquely with dramatic detail (like a well-told story). She also argues that there are several ethnographic conventions- discussing rapport, participant-observation, and the use of theory.
  • Content: Jennifer wanted to bring to the attention of anyone in the field of HCI the ethnographical approaches, conventions, and uses of each in modern-day society. She uses past examples, related work, and paradigm studies to provide examples for each.
Discussion:
This was a bit of a relatively unusual paper to read. I thought it was interesting and felt that a few of her points were definitely correct, but overall I feel like devoting an entire research paper to something along these lines is overkill. A simple "remember what you're doing out there- everything has a consequence. Everything!" will do. Like in my ethnography, I make a cautious effort to keep in mind how I am perceived, how I am gathering data, how I come across, etc. I got rather bored halfway through the paper because I felt her points were redundant or common sense. They were good to keep in mind for sure, but any GOOD ethnographer already knows to do/act certain ways that she addressed. There's no real "sure-fire" way to record if she achieved her goal or not because it is technically unclear to determine who all uses these proposed techniques or if they are actually referenced while in the field. I'm sure she believed she achieved her goals, though.

Wednesday, October 12, 2011

Paper Reading #18- Biofeedback game design: using direct and indirect physiological control to enhance game interaction

Title: Biofeedback game design: using direct and indirect physiological control to enhance game interaction
Reference Information:
Lennart Nacke, Michael Kalyn, Calvin Lough, and Regan Mandryk. " Biofeedback game design: using direct and indirect physiological control to enhance game interaction". CHI '11: Proceedings of the 2011 annual conference on Human factors in computing systems ACM New York, NY, USA ©2011.
Author Bios:
Lennart Nacke- He is an Assistant Professor for HCI and Game Science at the Faculty of Business and Information Technology at University of Ontario Institute of Technology (UOIT).
Michael Kalyn- A summer student working for Dr. Mandryk. He is a graduate in Computer Engineering and in his fourth year of Computer Science. His tasks this summer will be related to interfacing sensors and affective feedback.
Calvin Lough- Graduate student at the University of Saskatchewan.
Regan Mandryk- He's an Assistant Professor in the Department of Computer Science at the University of Saskatchewan. Obtained his B.Sc. in Mathematics from the University of Winnipeg in 1997, his M.Sc. in Kinesiology from Simon Fraser University in 2000, and his Ph.D. in Computing Science from Simon Fraser University in 2005.
Summary:
  • Hypothesis: If the authors can come up with either a direct or indirect method of physiological control sensors for users to control while immersed in a video game, then the user will enjoy the video game using these sensors much more than just with traditional controllers.
  • Methods: The authors created a game that featured many kinds of bonuses or power-ups that could be obtained if the user used their physiological inputs correctly (i.e. Medusa's Gaze to freeze enemies). The authors had some volunteers play through their game with randomly integrated physiological features to test how they were used.
  • Results: The results from the user studies and surveys indicated that the user's "fun" was directly associated with which game conditions were played under, that the users preferred the physiological sensors to play the game than without them, and that the kind of physiological conditions placed in the game had no effect on their "fun". Interestingly, however, users preferred some physiological controls over others. For example, one participant wasn't even sure that the EKG sensor controlled in the second game. Also, participants didn't like the fact that biting their lip was a control mechanism because it began to hurt after a few uses. Players did like the fact that flexing their leg allowed them to move faster or jump higher, however.
  • Content: Basically, the authors wanted to conduct a study based on a) if physiological controls/sensors made gameplay more interactive and "fun" and b) which sensors worked best. The authors found that the sensors did make the gameplay experience more enjoyable and that some sensors did work better than others. The sensors they found to be optimal were the ones that directly mapped the part of the body to gameplay (i.e. flexing your leg to move faster/jump higher) because it made the most conceptual sense to the player. The authors' surveys and questionnaires allowed them to have some insight into possibly designing a whole game solely based on physiological inputs.
Discussion:
I would have loved to have been a participant in this study. I think that some of these physiological sensors would definitely enhance gameplay (but I do agree with some of the participants that not all of the sensors were necessary. Some would definitely detract from the experience). This kind of study will more than likely breed the next generation of video games. I don't know how I feel about that. I like my games with the traditional controller and buttons and what not. The Kinect and WiiMote don't really interest me that much personally, but it is nonetheless an advancement in technology. I think that certain kinds of games could definitely have these kinds of sensors integrated phenomenally. In my opinion, the authors definitely achieved their goals because not only did they figure out if physiological sensors were "worth it", but they were accurately able to determine which kinds of sensors worked best for certain situations / gameplay environments.

Wednesday, October 5, 2011

Paper Reading #17- Privacy risks emerging from the adoption of innocuous wearable sensors in the mobile environment

Title: Privacy risks emerging from the adoption of innocuous wearable sensors in the mobile environment
Reference Information:
Andrew Raji, Animikh Ghosh, Santosh Kumar, and Mani Srivastava. "Privacy risks emerging from the adoption of innocuous wearable sensors in the mobile environment". CHI '11 Proceedings of the 2011 annual conference on Human factors in computing systems ACM New York, NY, USA ©2011. ISBN: 978-1-4503-0228-9.
Author Bios:
Andrew Raji- Works at the University of South Florida as a tenure-track Assistant Professor. He worked as a postdoctoral fellow from 2009-2010 and coordinated both the AutoSense and FieldStream projects.
Animikh Ghosh-  Joined SETLabs, Infosys after completing his M.S. in Comp Sc. from University of Memphis, Tennessee, USA. Now a research assistant under Dr. Santosh Kumar.
Santosh Kumar- I lead the Wireless Sensors and Mobile Ad Hoc Networks (WiSe MANet) Lab at the University of Memphis.
Mani Srivastava- Is a professor at UCLA in the engineering department. He received both the M.S. and Ph.D. degrees from the University of California, Berkeley, in 1987 and 1992, respectively.
Summary:
  • Hypothesis: If commercially-available mobile devices are going to continue to be worn, used, and distribtued, then their associated privacy leak concerns should be addressed immediately in order to prevent unintended information leaks about users through inferences of collected information from such technologies.
  • Methods: The authors studied groups of people to get feedback about privacy concerns.These groups of people used a technology called AutoSense. One group of people had no personal ties to data presented later on in the study while the other group was tested twice. They were tested after 3 days of using the device, and then again after all of the data was collected and presented back to them including inferences about each of them using the data from the technology to see if their concern level changed. The authors used 6 main points to focus their conclusions: measurements, behaviors, contexts, restrictions, abstractions, and privacy. Measurements are the raw data sets captured by a sensor. Behaviors are the inferred actions taken by the data producer with associated measurements. Context is any information that can be used to imply the situation of the behavior. Privacy threats are those that come with the data producer's identity being tied to the information. Restrictions remove data from a dataset before it can be shared to recude privacy threats. Abstractions provide a middle ground between complete restriction and complete sharing of data in order to render the device useful.
  • Results: The study found that users are most concerned with the release of information concerning conversation and commucing patterns. Obviously, the group with no personal stake in the data couldn't have a strong concern to the data from the second group of users using the AutoSense because they said that it didn't really bother them because it wasn't their own personal information. Upon data presentation sessnions, the authors also found (not surprisingly) that participants' concern levels rose as more and more information was released at a time. For example, if just timestamps of events or just events rather than a timestamp with an associated event all released together. This way, you couldn't really piece together information as easily as you could if the event and time were right there next to each other. The authors found that abstractions play a large role in determining the level of privacy threats.
  • Content: The authors conducted these studies to raise general awareness about privacy issues associated with general or specific-purpose mobile devices most people carry around today. They found that it is one issue to share personal information with someone you are directly talking to, but it is another issue entirely to share information across a wide network ("the web") so the general public may go access it.
Discussion:
This was an interesting read. It made me think about what all information I have released to the general public in the past and I will definitely be a little cautious when making decisions in the future. The authors had access toa LOT of information about the participants just because they were wearing a sensor around for 3 days. The inferences that can be made from the measurement collections are very strong as I have learned. These issues should be addressed more widespread when companies create such technologies, in my opinion. Based upon the results of their study, I believe the authors definitely achieved what they had set out to do initially. They are just trying to raise awareness about the things that the "leak" unintentionally to the "public", much like when you are visiting sites online and you "save your password" for a site.

Paper Reading #16- Classroom-based assistive technology: collective use of interactive visual schedules by students with autism

Title: Classroom-based assistive technology: collective use of interactive visual schedules by students with autism
Reference Information:
Meg Cramer, Sen Hirano, Monica Tentori, Michael Yeganyan, and Gillian Hayes. "Classroom-based assistive technology: collective use of interactive visual schedules by students with autism". CHI '11 Proceedings of the 2011 annual conference on Human factors in computing systems ACM New York, NY, USA ©2011. ISBN: 978-1-4503-0228-9.
Author Bios:
Meg Cramer- A graduate student at UC Irvine in the School of Information and Computer Sciences. Her advisor is Gillian Hayes.
Sen Hirano- He is a first-year PhD student in the department of  Informatics at the School of Information and Computer Science at the University of California, Irvine. His advisor is Dr. Gillian Hayes.
Monica Tentori- I am an assistant professor in computer science at UABC, Ensenada, México and a post-doctoral scholar at University of California, Irvine (UCI).
Michael Yeganyan-UC Irvine researcher for the Informatics STAR Research Group. Currently developing vSked, an assistive technology for children with autism, to be used in classrooms for child development.
Gillian Hayes- I am an Assistant Professor in Informatics in the School of Information and Computer Sciences and in the Department of Education at UC Irvine. My research interests are in human-computer interaction, ubiquitous computing,assistive and educational technologies and medical informatics.
Summary:
  • Hypothesis: If vSked is given a chance and becomes widely implemented, then it can promote student independence, reduce the quantity of educatorinitiated prompts, encourage consistency and predictability, reduce the time required to transition from one activity to another, and reduce the coordination required in the classroom.
  • Methods: The authors created a visual schedule on vSked for students using symbols or pictures to represent activities in the order in which the are scheduled for the day. For students with speaking impediments, communication boards allow those students to interact and participate in activities via voice or interaction with the functionality on the vSked device. A token-based reward system is also established on vSked to encourage students to stay on task, stay participating, and to stay involved. Each positive behavior displayed is recorded by vSked and tokens are added to the student's record. These tokens can add up to actual rewards that can be earned. The vSked system consists of a large touch screen monitor at the front of the class grouped with many miniature touch screen PCs for each of the students to interact with. All of the information within the system is connected. To accurately evaluate the effectiveness of vSked in the initial implementation stages, the authors conducted some experiments on sample sets of students.
  • Results: The results of the studies conducted by the authors on not only the students, but the teachers using the system as well, included a 2-week period before deployment as well as a 1-week period during deployment during the spring and summer semesters. The results showed that the students were more motivated to respond to questions and to progress to see what happens after each question is answered and also to pay attention to ask task, focusing on answering each specific question. It was also found that the students also preferred using the vSked because of its immediate rewarding of students with correct answers via firework graphics. There is no such capabilities with a simple pen-and-paper system. The authors also found that vSked allowed the students to basically "run their own day". The token-based system, each student's schedule, and trask transitions were now in the hands of vSked and of each individual student. The teachers didn't have to spend a lot of time bookkeeping these things and could focus more on other things that could be done. When rewards were earned, fireworks would display on the large screen in the front of the class so everyone could see. This would sometimes result in cheering from all the students (which is apparenlty rare among autistic students to be socially supportive/aware). Students are also now aware of how each other student is doing in the class. Often times, students were also found sharing tablets or looking at each other's tablets to see how a task was going.
  • Content: The authors of vSked wanted to create something to allow for a better, more enhanced learning environment for the specially educated children. So they had a vision to create a technology to integrate many individual features of a special education classroom and integrate them all. Through their implementation of a large TV screen centrally located in the classroom connected to many individual PCs for each student, it allows not only individuality and independence for each student to stay on top of things, but it also allows each student to see where they are relative to the whole class and to share experiences and skills with one another that was not previously seen prior to the vSked system.
Discussion:
I really liked this idea and how successful vSked was. Even though there were so many ways that were recorded in which vSked was successful, I believe there are still so many unexplored avenues in which vSked affects the students in social/psychological ways that were maybe unintended upon creation of the vSked (in positive ways). I believe the authors definitely met their goals and possibly even surpassed them when they conducted their studies and analyzed their feedback from trial classroom agents. I believe the authors were able to connect to the students more effectively with their technology then with traditional pen-and-paper systems. It seemed to intrigue the students more than usual for most activities. I like how the vSked is also allowing the students to develop their skills such as social interactions and maintaining their intelligence and independence, too.

Monday, October 3, 2011

Paper Reading #15-Madgets: actuating widgets on interactive tabletops

Title: Madgets: actuating widgets on interactive tabletops
Reference Information:
Malte Weiss, Florian Schwarz, Simon Jakubowski, and Jan Borchers. "Madgets: actuating widgets on interactive tabletops". UIST '10 Proceedings of the 23nd annual ACM symposium on User interface software and technology ACM New York, NY, USA ©2010. ISBN: 978-1-4503-0271-5.
Author Bios:
Malte Weiss- He is a PhD student at the Media Computing Group of RWTH Aachen University.
Florian Schwarz- Student assistant in the Media Computing Group of RWTH Aachen University.
Simon Jakubowski- Student assistant in the Media Computing Group of RWTH Aachen University.
Jan Borchers- Jan Borchers is full professor of computer science and head of the Media Computing Group at RWTH Aachen University. With his research group, he explores the field of human-computer interaction, with a particular interest in new post-desktop user interfaces for smart environments, ubiquitous computing, interactive exhibits, and time-based media such as audio and video.
Summary:
  • Hypothesis: If the authors's algorithm of recognition and manipulation can decompose the widgets into rigid bodies that can be actuated independently, then they will show that this actuation concept prepares the ground for a new class of actuated tabletop tangibles: magnetic widgets, or Madgets.
  • Methods: The authors's actuation algorithm triggers magnets attached to their design board to manipulate permanannt magnets attached to widgets for interactions. The different kinds of actuational directions that this algorithm deals with are tangential and normal (x-y and z, respectively). The movement algorithm takes into account "physical properties" of each interactive widget on the surface. Since their algorithm operates in real-time, constant update functions need to be called (refreshing and updating positions of widgets being interacted with, for instance). Also, the authors noticed that after about a minute, the magnets get hot individually. So if the temperature of a single magnet ever exceeds a certain temperature, it gets to "take a break" while another one is substituted in for it to do its work.
  • Results: While there were no studies conducted, trial runs, participant surveys, or anything of the sort, the authors pioneered ways to incorporate feedback mechanisms to ensure the proper use of the Madgets technology. Whenever a user drags an object, the Madgets board provides vibration feedback. Whenever a user turns a knob, switch, or adjusts a level for any widget, the Madget board provides resistance feedback (say if you tried to turn the volume past 100%). Whenever a user pushes a button down (seemingly), the area will provide vibration feedback.
  • Content: The authors set out to create a new technology by building on an existing one. They saw that interactive surfaces could be improved upon and so they took advantage of it. By providing details about the proper construction and implementation of their Madgets board, the authors showed what is possible for "upgraded widget technology".
Discussion:
I think this is kind of neat, but like with past papers dealing with similar issues regarding needing a specialized board for use, this can be cumbersome to construct/have around taking up space. I did like the unique interactions that this board allows you to make with the general purpose widgets, however. Also, the algorithms and thought processes going into the underlying algorithm into making the board work flawlessly was genius. I just would have never thought that magnets would overheat or markers would go out of bounds, etc...I give these guys some respect for that. Again, I don't see myself using this kind of technology ever, but there could be some fields where this could be used and expanded upon. The authors, in my opinion, definitely achieved their goals by creating this board and showcasing it to the public via their publication. I believe if they had tested it outside of their working group that they could have a more solid foundation for seeing if their product is satisfactory based on a user-satisfaction scale.

Paper Reading #14- TeslaTouch: electrovibration for touch surfaces

Title: TeslaTouch: electrovibration for touch surfaces
Reference Information:
Olivier Bau, Ivan Poupyrev, Ali Israr, and Chris Harrison. "TeslaTouch: electrovibration for touch surfaces". UIST '10 Proceedings of the 23nd annual ACM symposium on User interface software and technology ACM New York, NY, USA ©2010 ISBN: 978-1-4503-0271-5.
Author Bios:
Olivier Bau- He is currently a PostDoctoral Research Scientist at Disney Research in Pittsburgh in the Interaction Design group with Ivan Poupyrev. He received his PhD in Computer Science (HCI) at INRIA Saclay, working within the In|Situ| team with Wendy Mackay.
Ivan Poupyrev- He is currently a Senior Research Scientist at Disney Research Pittsburgh, a unit of Walt Disney Imagineering. There he directs an interaction technology group which focuses on inventing and exploring new interactive technologies for future entertainment and digital lifestyles
Ali Israr- He is currently working with the Interaction Design @ Disney group in Disney Research, The Walt Disney Company, USA.
Chris Harrison- He's a a fifth year Ph.D. student in the Human-Computer Interaction Institute at Carnegie Mellon University advised by Scott Hudson. He's also a Microsoft Research Ph.D. Fellow and editor-in-chief of XRDS, ACM's flagship magazine for students.
Summary:
  • Hypothesis: If the authors can integrate electrovibration into an interactive surface, then many features of tactible input will be unlocked for the user to experience with TeslaTouch.
  • Methods: In order to allow the transparent electrode to work, a basic circuit board is used to induce electricity for the interactions. No actual voltage is passed through the user (who is assumed to be grounded), and only a minimal current is passed (an amount considered safe for humans to be "charged" with). This TeslaTouch technology is just an "add-on" to current touchscreen / interactive devices already existing. The authors conducted experiments with trial users in order to obtain initial feedback results on the initial implementation on TeslaTouch.
  • Results: The studies conducted by the authors indicated that low frequency interactions were perceived as "rougher" relative to the higher frequency interactions. The effect of amplitude on the interactions depended on the frequency- the higher the frequency and amplitude simultaneously resulted in "smoother" interactions versus high amplitude, low frequency interactions being repotred as "stickier". The results were analyzed by the authors to determine the JNDs (just noticeable differences) between interactions and to determine what the borderline was between pleasurahle and discomforting levels of interactions.
  • Conents: The authors set out to create a new way of interacting on surfaces that would reap more benefits than traditional mechanical vibrotactile systems. They created "electrovibration" techniques that would differ based on the interaction type (dragging a file, erasing part of a picture, etc). After testing, the authors calculated what optimal levels of frequency and amplitude were appropriate for each interaction and also what appropriate levels of each were to obtain maximal pleasure from human touch with each interaction.
Discussion:
I like the idea of this technology, but I kind of got a little bored with the paper because this feels like the 5th paper in a row we had to report on multitouch surface devices. It is a cool idea, but I consider this technology an "add-on" to existing technologies if you want more options or different feedback mechanisms. What I got out of this paper was that TeslaTouch on an interactive surface is similar to playing Nintendo64 with a Rumble Pack. I believe the authors achieved their goals, but still (admittedly) had room to improve and could apply this technology to new applications for future projects. This is expandable for sure, and can definitely be used primarily instead of secondarily (from what I understood) on interactive surface devices in the future.

Thursday, September 29, 2011

Paper Reading #13- Combining multiple depth cameras and projectors for interactions on, above and between surfaces

Title: Combining multiple depth cameras and projectors for interactions on, above and between surfaces
Reference Information:
Andrew Wilson and Hrvoje Benko . "Combining multiple depth cameras and projectors for interactions on, above and between surfaces". UIST '10 Proceedings of the 23nd annual ACM symposium on User interface software and technology ACM New York, NY, USA ©2010 ISBN: 978-1-4503-0271-5.
Author Bios:
Andrew Wilson-I was a graduate student at the University of North Carolina at Chapel Hill's Department of Computer Science from August 1997 - December 2002. I've since graduated and moved on to Sandia National Laboratories where I'm working on GPU-accelerated visualization of gigabyte-sized data sets.
Hrvoje Benko- I am a researcher at Adaptive Systems and Interaction Microsoft Research. My research is on novel surface computing technologies and their impact on human-computer interaction. Prior to working at Microsoft, I received my PhD at Columbia University, working on augmented reality projects that combine immersive experiences with interactive tabletops.
Summary:

  • Hypothesis: "The rich, almost analog feel of a
    dense 3D mesh updated in real time invites an important
    shift in thinking about computer vision: rather than struggling
    to reduce the mesh to high-level abstract primitives,
    many interactions can be achieved by less destructive
    transformations and simulation on the mesh directly."
  • Methods: LightSpace allows users to interact with virtual objects within the realm of the environment. It supports through-body interactions, between-hands interactions, the picking up of objects, spatial interactions, and other features using unique depth analysis and algorithms. It is a hardware device suspended from the ceiling consisting of multiple depth cameras and projectors to track movements and interactions within the "smart room" environment. The cameras and projectors must be calibrated to share the same 3D environment.
  • Results: Via a user trial and feedback session lasting 3 days, the authors gathered that 6 users was the maximum (currently) number of users allowed by the "smart room" currently to still maintain accurate interaction representations. Also, the system seemed to slow down once more than 2 users were using it simultaneously. Occasionally, interactions from one user were never registered because another user was blocking a camera, so LightSpace never picked it up. Once they began to interact with the device, however, users felt that it was very natural and few had problems actually interacting.
  • Content: The authors presented a device that allowed users to live in a "virtual environment" and interact spatially with objects within a certain range. The authors explored different spatial interactions with new depth tracking techniques, and it seemed to produce fruitfully.
Discussion:
I think this is really cool to have and play around with. But just like recent papers we have read, I fail to see any applications in the real world for this technology so far. Their ideas with the depth tracking and figuring algorithms for interactions of different sorts was genius. I didn't like much how users could be blocked so easily from being registered, however. All it took was another user getting in the way of a camera angle to completely negate an action. I believe that the authors did achieve their goals. They set out to create a "smart room"- and that's what they got. If I had the resources I would definitely just buy one of these for a room in my house and just go nuts on it like I was working on my PC.

Monday, September 26, 2011

Paper Reading #12- Enabling beyond-surface interactions for interactive surface with an invisible projection

Title: Enabling beyond-surface interactions for interactive surface with an invisible projection
Reference Information:
Li-Wei Chen, Hsiang-Tao Wu, Hui-Shan Ko, Ju-Chun Ko, Home-Ru Lin, Mike Chen, Jane Hsu, Yi-Ping Hung, "Enabling beyond-surface interactions for interactive surface with an invisible projection". UIST '10 Proceedings of the 23nd annual ACM symposium on User interface software and technology ACM New York, NY, USA ©2010.
Author Bios:
Li-Wei Chen- Li-Wei's education: BSc Electrical Engineering, Queen's University, 1998. MASc Electrical Engineering, University of Toronto, 2000. PhD Electrical Engineering and Computer Science, Massachusetts Institute of Technology, 2005.
Hsiang-Tao Wu- is a Microsoft Researcher.
Ju-Chun Ko-  Ph. D (Candidate) student in Graduate Institute of Networking and Multimedia of National Taiwan University
Mike Chen- Professor in the Computer Science department at National Taiwan University.
Jane Hsu- Jane Hsu is a professor of Computer Science and Information Engineering at National Taiwan University
Yi-Ping Hung- Yi-Ping Hung received his B.S. in Electrical Engineering from the National Taiwan University, and his Master and Ph.D. degrees from Brown University. He is currently a professor in the Graduate Institute of Networking and Multimedia, and in the Department of Computer Science and Information Engineering, both at the National Taiwan University
Summary:
  • Hypothesis: If the authors can successfully create a programmable infrared (IR) technique that can interact with surfaces beyond a two-dimensional fashion, then enabling a spatially-aware multi-display  and multi-touch tabletop system is inherently possible.
  • Methods: The authors use a modified DLP projector, mirrors, an IR projection, and color projection to implement their creation. To handle multiple inputs from other devices, their device will create 4 smaller markers around projections from closer devices and give that projection priority while also creating 4 successively larger markers around remaining input projections to capture all of it at once. To handle multi-touch inputs, the marker patterns become dynamic to resize angles, sharpen resolution, and recognize all inputs. The authors conducted a small user study asking participants to use their map application by finding and investigating buildings.
  • Results: By processing one layer at a time and by also simulating the background, the authors are able to retrieve proper inputs, associations, and recognitions. The authors also have the i-m-flashlight as an exploratory tool, the i-m-lamp as a "desk lamp", and a i-m-camera for the 3D views. All of these devices interact with one another and are connected with the table. The results from the user study showed some flaws in the implementation of the device. For example, if a user wanted to explore an entire building, they couldn't. Only the bottom portion of the structure was displayed. When users tried to "scroll" up to see the top using the i-m-lamp, the views went out of range. Users found that the i-m-flashlight was more efficient in working on short-term tasks while the i-m-lamp was more efficient in working with long-term tasks. Moving the i-m-lamp around a lot in a short period of time produced less accurate results. The users found quickly that the different pieces of technology were meant for different kinds of interactions.
  • Content: The authors wanted to create something that would be spread across a tabletop and handle multiple inputs on the actual surface in three dimensions. By using the prototype that they made out of modified projectors, screens, cameras, and mirrors, they were able to do that. Their device can also handle invisible markers for detection on multi-inputs. Three devices, the i-m- lamp, i-m-camera, and i-m-flashlight all coordinate with each other to handle 3D imaging, processing, and dynamic inputs. They conducted a study on early user feedback and used the results to gather methods and ways to improve their creation for future versions.
Discussion:
I liked the diagrams and pictures of this new device. These people were really smart to be able to basically construct something that can do all of these things and can support all of these features from scratch. I see definite uses for this technology, particularly in the construction and architecture fields. I think the authors achieved, at least partially, what they set out to do. They created a device, got user feedback, and can now improve on it for future releases. At least they were able to catch problems users had with it before they mass commercialized it. I think this can be a building block, but seemed a bit complicated to make. I got a little lost when they began to describe methods for constructing and implementing the device.