Ben Lafreniere of the University of Saskatchewan visited the DGP to talk about rehearsal-based interfaces.
Ben Lafreniere is a Human-Computer Interaction researcher who specializes in the areas of learning and skill development with interactive systems, and the usability of feature-rich software. In 2014 he received a PhD in Computer Science from the University of Waterloo for his work on developing task-centric user interfaces. He is currently a Postdoctoral Fellow with the Interaction lab at the University of Saskatchewan.
Rehearsal-based interfaces are designed to enable a smooth transition from novice to expert performance by making the novice users visually-guided actions a physical rehearsal of the experts feedback-free actions. While a number of examples of these interfaces have been developed, including Marking Menus and FastTap menus, there is little published data on how skill development happens in real use of these interfaces. In this talk I will describe two studies we conducted on skill development in rehearsal-based interfaces: one in a game that directly rewards rapid menu selections, and another in a drawing application that has no particular need for urgency. Our results show very different patterns of adoption in these two applications, and suggest that rehearsal of physical actions alone does not guarantee that users will adopt expert methods. I will also discuss insights into what affects use of expert methods by users, and the implications of our findings for how rehearsal-based techniques should be employed in practice. Finally, I will discuss ongoing research that builds on this work.
Professor Sid Fels from the ECE department at UBC, a DCS PhD alumnus, visited the DGP to talk about his work on the intersections between HCI, human anatomy, and forms of expression.
Sid has been in the department of Electrical & Computer Engineering at the University of British Columbia since 1998. Sidney received his Ph.D. and M.Sc. in Computer Science at the University of Toronto in 1994 and 1990 respectively. He received his B.A.Sc. in Electrical Engineering at the University of Waterloo in 1988. He was recognized as a Distinguished University Scholar at UBC from 2004. He was a visiting researcher at ATR Media Integration & Communications Research Laboratories in Kyoto, Japan from 1996 to 1997. He also worked at Virtual Technologies Inc. in Palo Alto, CA. He is internationally known for his work in human-computer interaction, biomechanical modeling of human anatomy, and new interfaces for musical expression and interactive arts. He was a principal investigator on the Canadian Networks Centre of Excellence on Graphics, Animation and New Media (GRAND) from 2010-2014. He was the Director of the Media and Graphics Interdisciplinary Centre (MAGIC) from 2001-2012.
Design for Human Experience and Expression: Research at MAGIC and the HCT Laboratory
Research at the Media and Graphics Interdisciplinary Centre (MAGIC; www.magic.ubc.ca) and the Human Communications Technology (HCT) laboratory (hct.ece.ubc.ca) has been targeting design for human experience and expression. In this presentation, I’ll start with a discussion of gesture-to-speech and voice explorations, including Glove-TalkII and the Digital Ventriloquized Actors (DIVAs). I’ll connect these to other explorations of the new interfaces for musical and visual expression that we have created. I’ll then show some more work on new forms of personalized video viewing that we call the MyView system. MyView creates a new viewing experience though integrating human memory patterns into the navigation, browsing, annotation, authoring and sharing mechanisms. I will briefly discuss our work on modelling human anatomy (www.parametrichuman.org) and function, such as speaking, chewing, swallowing and breathing (www.magic.ubc.ca/OPAL.htm) with biomechanical models using our toolkit Artisynth (www.artisynth.org). This work is motivated by our quest to make a new vocal instrument that can be controlled by gesture. I’ll discuss some of the activities we have been doing on some new 3D displays: pCubee and Spheree. Finally, these investigations will be used to support a theory of designing for intimacy and discussions of perspectives on human computer interaction for new experiences and forms of expression.
Their work on “Energy-Efficient Structured Light Imaging” has been awarded with the Best Demo Award at ICCP 2015.
Prof. Michael Bernstein of Stanford visited the DGP to talk about his work in designing crowd-sourcing systems for experts.
Michael Bernstein is an Assistant Professor of Computer Science at Stanford University, where he co-directs the Human-Computer Interaction group and is a Robert N. Noyce Family Faculty Scholar. His research in human-computer interaction focuses on the design of crowdsourcing and social computing systems. This work has received Best Paper awards and nominations at premier venues in human-computer interaction and social computing (ACM UIST, ACM CHI, ACM CSCW, AAAI ISWSM). Michael has been recognized with the NSF CAREER award, as well as the George M. Sprowls Award for best doctoral thesis in Computer Science at MIT. He holds Ph.D. and M.S. degrees in Computer Science from MIT, and a B.S. in Symbolic Systems from Stanford University.
Crowdsourcing a Meeting of Minds
Crowdsourcing is an increasingly powerful method for combining amateurs’ efforts to recreate an expert’s abilities. However, across domains from design to engineering to art, few goals are truly the effort of just one person — even one expert. If we can now crowdsource simple tasks such as image labeling, how might we computationally coordinate many peoples’ abilities toward far more complex and interdependent goals? In this talk, I present computational systems for gathering and guiding crowds of experts — including professional programmers, designers, singers and artists. The resulting collectives tackle problems modularly and at scale, dynamically grow and shrink depending on task demands, and combine into larger organizations. I’ll demonstrate how computationally-enabled expert crowds can pursue goals such as designing new user experiences overnight, producing animated shorts in two days, and even pursuing novel research.
Prof. Chris Harrison of CMU visited the DGP to talk about his work in mobile sensing.
Dr. Harrison’s Lecture at DGP
Chris is an Assistant Professor of Human-Computer Interaction at Carnegie Mellon University. He broadly investigates novel sensing technologies and interaction techniques, especially those that empower people to interact with small devices in big ways. Harrison has been named a top 30 scientist under 30 by Forbes, a top 35 innovator under 35 by MIT Technology Review, a Young Scientist by the World Economic Forum, and one of six innovators to watch by Smithsonian. He has been awarded fellowships by Google, Qualcomm, Microsoft Research and the Packard Foundation. He is also the CTO of Qeexo, a touchscreen technology startup. When not in the lab, Chris can be found welding sculptures, visiting remote corners of the globe, and restoring his old house.
Interacting with Small Devices in Big Ways.
Eight years ago, multi-touch devices went mainstream, and changed our field, the industry and our lives. In that time, mobile devices have gotten much more capable, yet the core user experience has evolved little. Contemporary touch gestures rely on poking screens with different numbers of fingers: one-finger tap, two-finger pinch, three-finger swipe and so on. We often label these as “natural” interactions, yet the only place I perform these “gestures” is on my touchscreen device. We are also too quick to blame the “fat finger” problem for much of our touch interface woes – if a zipper or pen were too small to use, we would simply call that “bad design”. Fortunately, our fingers and hands are amazing, and with good technology and design, we can elevate touch interaction to new heights. I believe the era of multi-touch is coming to a close, and that we are on the eve of an exciting new age of “rich-touch” devices and experiences.
Jonathan Deber, Ricardo Jota, Clifton Forlines, and Daniel Wigdor learned that their paper was called-out for Honorable Mention at CHI 2015. Congratulations to the team! Paper details are below. The project will be presented at ACM CHI in Seoul in April.
How much Faster is Fast Enough? User Perception of Latency & Latency Improvements in Direct and Indirect Touch
, Ricardo Jota
, Clifton Forlines
, Daniel Wigdor
(2014). How much Faster is Fast Enough? User Perception of Latency & Latency Improvements in Direct and Indirect Touch. Proceedings of the 2015 SIGCHI conference on human factors in computing systems
(ACM CHI). 10 pages, in press.
Ricardo Jota was awarded the Postdoctoral Award for Outstanding Innovation for his research with Tactual Labs.
Congratulations to Jota for the well-deserved recognition!
Read more about the Mitacs Awards Reception and Jota’s work in this press release.
Photo Courtesy of Mitacs. Pictured Above: Ted Mao (Trojan Technologies), Rafael Falcon (University of Ottawa), Linda Gowman (Trojan Technologies), Daniela Tuchel (Royal Roads University), Minister Chris Alexander (Minister of Citizenship and Immigration), Minister Ed Holder (Minister of State, Science and Technology), Stephen Dugdale (Universite INRS), Minister Kerry-Lynne Findlay (Minister of National Revenue), Dr. Rob Annan (interim Chief Executive Officer, Mitacs), Professor Alan Fung (Ryerson University) and Ricardo Jota (University of Toronto).
Dr. Ali Mazalek, Associate Professor of Ryerson University and Georgia Tech will be visiting our lab on Thursday, December 4th. She will be presenting a talk between 11:00 am until 12:30 pm.
Welcome to DGP, Dr. Mazalek!
“Mind, material, and movement: embodying creativity in the digital era”
We are increasingly tethered to a range of pixelated boxes of varying shapes and sizes. These devices are ever present in our lives, transporting us daily into vast information and computational realms. And while our interactions with digital devices are arguably becoming more fluid and “natural”, they still make only limited use of our motor system and largely isolate us from our immediate physical surroundings. Yet a gradual shift in the cognitive sciences toward embodied paradigms of human cognition can inspire us to think about why and how computational media should engage our bodies and minds together. What is the role of physical movements and materials in the way we engage with and construct knowledge in the world? This talk will provide some perspectives on this question, highlighting research from the Synaesthetic Media Lab that supports creativity and expression across the physical and digital worlds.
Dr. Ali Mazalek has spent over 15 years trying to get digital technologies to fit better into her physical world and life, rather than letting them drag her into the pixelated depths of her computer screens. At the same time, she has a deep interest in how computational media can support and enhance creative practices and processes, supporting new forms of expression and new ways of thinking and learning. She is a Canada Research Chair in Digital Media and Innovation and Associate Professor in the RTA School of Media at Ryerson University, as well as Associate Professor of Digital Media at Georgia Tech. Her Synaesthetic Media Lab ( synlab.gatech.edu) is a playground where physical materials, analog sensors, and digital media happily co-exist and come together in novel ways to support creativity and expression across both science and art disciplines.
Professor Michael Terry will be presenting a talk on Thursday, November 20th from 11:00 am until 12:30 pm in room BA5187.
Please join us in welcoming Professor Terry!
“Interactive Systems Need to Know How to Read the Web and Watch YouTube”
In this talk, Professor Michael Terry will argue that there is great value in interactive systems that can learn how to accomplish tasks by “reading” web-based tutorials and “watching” how-to videos. He will focus primarily on text-based documents and search queries, and show how techniques from the fields of machine learning and information retrieval can be leveraged to extract streams of “how-to” information from web-based resources and instrumentation logs. These information sources enable a new class of interactive system that is more aware of the tasks it can perform, as well as how to accomplish these tasks. Importantly, this awareness continually evolves and tracks how the user community actually uses the system.
Michael Terry is an associate professor in the Cheriton School of Computer Science at the University of Waterloo, where he co-directs the HCI Lab. His research lies at the intersection of HCI, machine learning, and information retrieval. His current projects include machine understanding of instructional materials, task-centric user interfaces, and interactive machine learning systems designed to assist the digitization and cataloging of millions of biological specimens in London’s Natural History Museum.
As part of the Distinguished Lecture Series, Dr. Karen Myers is presenting a talk at the Bahen Centre this Tuesday, November 18th. The lecture is hosted in BA1170 at 11:00 am.
Learning from Demonstration Technology: A Tale of Two Applications
Learning from demonstration technology has seen increased focus in recent years as a means to endow computers with capabilities that might otherwise be difficult or time-consuming for a user to program. This talk describes two efforts that employ learning from demonstration technology to quite distinct ends. The first is to provide a capability that supports users with no programming experience in the creation of procedures that automate repetitive or time-consuming tasks. This capability has been operationally deployed within a collaborative planning environment that is used widely by the U.S. Army. The second is to support automated performance evaluation of students as they seek to acquire complex procedural skills through training in virtual environments. In this second case, instructional content developers employ learning from demonstration technology to create solution models for training exercises. An automated assessment capability employs soft graph matching to align a trace of a students response to an exercise with the solution models for that exercise, providing a flexible basis for evaluating student performance. In contrast to intelligent tutoring systems that force students to follow a pre-specified solution trajectory, our approach enables meaningful feedback in domains where solutions can have significant variability.
||Karen Myers is a Principal Scientist within the Artificial Intelligence Center at SRI International, where she leads a team focused on developing intelligent systems that facilitate man-machine collaboration. Myers has led the development of several AI technologies that have been successfully transitioned into operational use in areas that span collaborative systems, task management, and learning from demonstration. Her research interests include autonomy, multi-agent systems, automated planning, personalization, and mixed-initiative problem solving