CHI 2019

At CHI2019, HCI@KAIST presents 14 papers, 6 late breaking works and 3 workshops.
One best paper and one honorable mention are included. These works are from 11 labs of 4 different schools and departments at KAIST. We thank our outstanding colleagues and collaborators from industry, research centers and universities around the world. 

We will be hosting the KAIST HOSPITALITY NIGHT event on May 7th (Tue).
Join us at HCI@KAIST HOSPITALITY NIGHT!

DATE   May 7th (Tue) 2019
TIME  20:00 – 23:00
LOCATION  The Strip Joint Glasgow – pizza place & drink monger (956 Argyle street, Glasgow G3 8LU)

 

Paper & Notes

pickme
PicMe: Interactive Visual Guidance for Taking Requested Photo Composition

Monday 11:00 -12:20 | Session of On the Streets | Room: Boisdale 1

  • Minju Kim, Graduate School of Culture Technology, KAIST
  • Jungjin Lee, KAI Inc., Daejeon

PicMe is a mobile application that provides interactive onscreen guidance that helps the user take pictures of a composition that another person requires. Once the requester captures a picture of the desired composition and delivers it to the user (photographer), a 2.5D guidance system, called the virtual frame, guides the user in real-time by showing a three-dimensional composition of the target image (i.e., size and shape). In addition, according to the matching accuracy rate, we provide a small-sized target image in an inset window as feedback and edge visualization for further alignment of the detail elements. We implemented PicMe to work fully in mobile environments. We then conducted a preliminary user study to evaluate the effectiveness of PicMe compared to traditional 2D guidance methods. The results show that PicMe helps users reach their target images more accurately and quickly by giving participants more confidence in their tasks.

1-01
Co-Performing Agent: Design for Building User-Agent Partnership in Learning and Adaptive Services

Wednesday 16:00-17:20 | Session of The One with Bots | Room: Boisdale 1

  • Da-jung Kim, Department of Industrial Design, KAIST
  • Youn-kyung Lim, Department of Industrial Design, KAIST

Intelligent agents have become prevalent in everyday IT products and services. To improve an agent’s knowledge of a user and the quality of personalized service experience, it is important for the agent to cooperate with the user (e.g., asking users to provide their information and feedback). However, few works inform how to support such user-agent co-performance from a human-centered perspective. To fill this gap, we devised Co-Performing Agent, a Wizard-of-Oz-based research probe of an agent that cooperates with a user to learn by helping users to have a partnership mindset. By incorporating the probe, we conducted a two-month exploratory study, aiming to understand how users experience co-performing with their agent over time. Based on the findings, this paper presents the factors that affected users’ co-performing behaviors and discusses design implications for supporting constructive co-performance and building a resilient user–agent partnership over time.

ten
Ten-Minute Silence: A New Notification UX of Mobile Instant Messenger

Wednesday 12:00-12:20 | Session of UX Methods | Room: Forth

  • In-geon Shin, Department of Industrial Design, KAIST
  • Jin-min Seok, Department of Industrial Design, KAIST
  • Youn-kyung Lim, Department of Industrial Design, KAIST

People receive a tremendous number of messages through mobile instant messaging (MIM), which generates crowded notifications. This study highlights our attempt to create a new notification rule to reduce this crowdedness, which can be recognized by both senders and recipients. We developed an MIM app that provides only one notification per conversation session, which is a group of consecutive messages distinguished based on a ten-minute silence period. Through the two-week field study, 20,957 message logs and interview data from 17 participants revealed that MIM notifications affect not only the recipients’ experiences before opening the app but also the entire conversation experience, including that of the senders. The new notification rule created new social norms for the participants’ use of MIM. We report themes about the changes in the MIM experience, which will expand the role of notifications for future MIM apps.

skin
Like A Second Skin: Understanding How Epidermal Devices Affect Human Tactile Perception

Thursday 14:00 - 15:20 | Paper Session: Skin and Textiles | Hall 2

  • Aditya Shekhar Nittala, Saarland University
  • Klaus Kruttwig, INM-Leibniz Institute
  • Jaeyeon Lee, School of Computing, KAIST
  • Roland Bennewitz, INM-Leibniz Institute
  • Eduard Arzt, INM-Leibniz Institute
  • Jürgen Steimle, Saarland University

The emerging class of epidermal devices opens up new opportunities for skin-based sensing, computing, and interaction. Future design of these devices requires an understanding of how skin-worn devices affect the natural tactile perception. In this study, we approach this research challenge by proposing a novel classification system for epidermal devices based on flexural rigidity and by testing advanced adhesive materials, including tattoo paper and thin films of poly (dimethylsiloxane) (PDMS). We report on the results of three psychophysical experiments that investigated the effect of epidermal devices of different rigidity on passive and active tactile perception. We analyzed human tactile sensitivity thresholds, two-point discrimination thresholds, and roughness discrimination abilities on three different body locations (fingertip, hand, forearm). Generally, a correlation was found between device rigidity and tactile sensitivity thresholds as well as roughness discrimination ability. Surprisingly, thin epidermal devices based on PDMS with a hundred times the rigidity of commonly used tattoo paper resulted in comparable levels of tactile acuity. The material offers the benefit of increased robustness against wear and the option to re-use the device. Based on our findings, we derive design recommendations for epidermal devices that combine tactile perception with device robustness.

virtual
VirtualComponent: a Mixed-Reality Tool for for Designing and Tuning Breadboarded Circuits

Wednesday 10:00 - 10:20 | Session of Fabricating Electronics | Room: Hall 1

  • Yoonji Kim, Department of Industrial Design, KAIST
  • Youngkyung Choi, Department of Industrial Design, KAIST
  • Hyein Lee, Department of Industrial Design, KAIST
  • Geehyuk Lee, School of Computing, KAIST
  • Andrea Bianchi, Department of Industrial Design, KAIST

Prototyping electronic circuits is an increasingly popular activity, supported by the work of researchers, who developed toolkits to improve the design, debug and fabrication of electronics. While past work mainly dealt with circuit topology, in this paper we propose a system for determining or tuning the values of the circuit components. Based on the results of a formative study with seventeen makers, we designed VirtualComponent, a mixed-reality tool that allows to digitally place electronic components on a real breadboard, tune their values in software, and see these changes applied to the physical circuit in real-time. VirtualComponent is composed of a set of plug-and-play modules containing banks of components, and a custom breadboard managing the connections and components’ values. Through example usages and the results of an informal study with twelve makers, we demonstrate that VirtualComponent is easy to use and encourages users to test components’ value configurations with little effort.

lockn
LocknType: Lockout Task Intervention for Discouraging Smartphone App Use

Monday 14:00-15:20 | Paper Session: Human-Smartphone Interaction | Hall 2

  • Jaejeung Kim, Graduate School of Knowledge Service Engineering, KAIST
  • Joonyoung Park, Graduate School of Knowledge Service Engineering, KAIST
  • Hyunsoo Lee, Graduate School of Knowledge Service Engineering, KAIST
  • Minsam Ko, College of Computing, Hanyang
  • Uichin Lee, Graduate School of Knowledge Service Engineering, KAIST

Instant access and gratification make it difficult for us to self-limit the use of smartphone apps. We hypothesize that a slight increase in the interaction cost of accessing an app could successfully discourage app use. We propose a proactive intervention that requests users to perform a simple lockout task (e.g., typing a fixed length number) whenever a target app is launched. We investigate how a lockout task with varying workloads (i.e., pause only without number input, 10-digit input, and 30-digit input) influence a user’s decision making, by a 3-week, in-situ experiment with 40 participants. Our findings show that even the pause-only task that requires a user to press a button to proceed discouraged an average of 13.1% of app use, and the 30-digit-input task discouraged 47.5%. We derived determinants of app use and non-use decision making for a given lockout task. We further provide implications for persuasive technology design for discouraging undesired behaviors.

evaluating
Evaluating the Combination of Visual Communication Cues for HMD-based Mixed Reality Remote Collaboration

Tuesday 09:00 - 10:20 | Session of X Reality Evaluations | Room Dochart 2

  • Seungwon Kim, School of ITMS, University of South Australia
  • Gun Lee, School of ITMS, University of South Australia
  • Weidong Huang, Swinburne University of Technology
  • Hayun Kim, Graduate School of Culture Technology, KAIST
  • Woontack Woo, Graduate School of Culture Technology, KAIST
  • Mark Billinghurst, School of ITMS, University of South Australia

Many researchers have studied various visual communication cues (e.g. pointer, sketching, and hand gesture) in Mixed Reality remote collaboration systems for real-world tasks. However, the effect of combining them has not been so well explored. We studied the effect of these cues in four combinations: hand only, hand + pointer, hand + sketch, and hand + pointer + sketch, with three problem tasks: Lego, Tangram, and Origami. The study results showed that the participants completed the task significantly faster and felt a significantly higher level of usability when the sketch cue is added to the hand gesture cue, but not with adding the pointer cue. Participants also preferred the combinations including hand and sketch cues over the other combinations. However, using additional cues (pointer or sketch) increased the perceived mental effort and did not improve the feeling of co-presence. We discuss the implications of these results and future research directions.

theEffects
The Effects of Interruption Timing on Autonomous Height-Adjustable Desks that Respond to Task Changes

Tuesday 11:00 - 12:20 | Session of In the Office | Room Alsh 1

  • Bokyung Lee, Department of Industrial Design, KAIST
  • Sindy Wu, School of computer science and technology, KAIST
  • Maria Jose Reyes, Department of Industrial Design, KAIST
  • Daniel Saakes, Department of Industrial Design, KAIST

Actuated furniture, such as electric adjustable sit-stand desks, helps users vary their posture and contributes to comfort and health. However, studies have found that users rarely initiate height changes. Therefore, in this paper, we look into furniture that adjusts itself to the user’s needs. A situated interview study indicated task-changing as an opportune moment for automatic height adjustment. We then performed a Wizard of Oz study to find the best timing for changing desk height to minimize interruption and discomfort. The results are in line with prior work on task interruption in graphical user interfaces and show that the table should change height during a task change. However, the results also indicate that until users build trust in the system, they prefer actuation after a task change to experience the impact of the adjustment.

smart
SmartManikin: Virtual Humans with Agency for Design Tools

Tuesday 16:00 - 17:20 | Session of Design Tools | Room Alsh 1

  • Bokyung Lee, Department of Industrial Design, KAIST
  • Taeil Jin, Graduate school of culture and technology, KAIST
  • Sung-Hee Lee, Faculty of art and Design, University of Tsukuba
  • Daniel Saakes, Department of Industrial Design, KAIST

When designing comfort and usability in products, designers need to evaluate aspects ranging from anthropometrics to use scenarios. Therefore, virtual and poseable mannequins are employed as a reference in early-stage tools and for evaluation in the later stages. However, tools to intuitively interact with virtual humans are lacking. In this paper, we introduce SmartManikin, a mannequin with agency that responds to high-level commands and to real-time design changes. We first captured human poses with respect to desk configurations, identified key features of the pose and trained regression functions to estimate the optimal features at a given desk setup. The SmartManikin’s pose is generated by the predicted features as well as by using forward and inverse kinematics. We present our design, implementation, and an evaluation with expert designers. The results revealed that SmartManikin enhances the design experience by providing feedback concerning comfort and health in real time.

slow
Slow Robots for Unobtrusive Posture Correction

Tuesday 09:00 - 10:20 | Session of Weighty Interactions | Room Dochart 1

  • Joon-Gi Shin, Department of Industrial Design, KAIST
  • Eiji Onchi, Graduate school of comprehensive human science, University of Tsukuba
  • Maria Jose Reyes, Department of Industrial Design, KAIST
  • Junbong Song, TeamVoid
  • Uichin Lee, Graduate School of Knowledge Service Engineering, KAIST
  • Seung-Hee Lee, Faculty of Art and Design, University of Tsukuba
  • Daniel Saakes, Department of Industrial Design, KAIST

Prolonged static and unbalanced sitting postures during computer usage contribute to musculoskeletal discomfort. In this paper, we investigated the use of a very slow moving monitor for unobtrusive posture correction. In a first study, we identified display velocities below the perception threshold and observed how users (without being aware) responded by gradually following the monitor’s motion. From the result, we designed a robotic monitor that moves imperceptible to counterbalance unbalanced sitting postures and induces posture correction. In an evaluation study (n=12), we had participants work for four hours without and with our prototype (8 in total). Results showed that actuation increased the frequency of non-disruptive swift posture corrections and significantly.

howto
How to Design Voice Based Navigation for How-To Videos

Monday 11:00 - 12:20 | Interacting with Videos |

  • Minsuk Chang, School of Computing, KAIST
  • Anh Troung, Adobe Research, Stanford University
  • Oliver Wang, Adobe Research
  • Maneesh Agrawala, Stanford University
  • Juho Kim, School of Computing, KAIST

When watching how-to videos related to physical tasks, users’ hands are often occupied by the task, making voice input a natural fit. To better understand the design space of voice interactions for how-to video navigation, we conducted three think-aloud studies using: 1) a traditional video interface, 2) a research probe providing a voice controlled video interface, and 3) a wizard-of-oz interface. From the studies, we distill seven navigation objectives and their underlying intents: pace control pause, content alignment pause, video control pause, reference jump, replay jump, skip jump, and peek jump. Our analysis found that users’ navigation objectives and intents affect the choice of referent type and referencing approach in command utterances. Based on our findings, we recommend to 1) support conversational strategies like sequence expansions and command queues, 2) allow users to identify and refine their navigation objectives explicitly, and 3) support the seven interaction intents.

project website

diagonose
Diagnosing and Coping with Mode Errors in Korean-English Dual-language Keyboard

Monday 14:00 - 15:20 | Paper Session: Human-Smartphone Interaction | Hall 2

  • Sangyoon Lee, School of Computing, KAIST
  • Jaeyeon Lee, School of Computing, KAIST
  • Geehyuk Lee, School of Computing, KAIST

In countries where languages with non-LatIn characters are prevalent, people use a keyboard with two language modes namely, the native language and English, and often experience mode errors. To diagnose the mode error problem, we conducted a field study and observed that 78% of the mode errors occurred immediately after application switchIng. We implemented four methods (Auto-switch, Preview, Smart-toggle, and Preview & Smart-toggle) based on three strategies to deal with the mode error problem and conducted field studies to verify their effectiveness. In the studies considerIng Korean-English dual Input, Auto-switch was Ineffective. On the contrary, Preview significantly reduced the mode errors from 75.1% to 41.3%, and Smart-toggle saved typIng cost for recoverIng from mode errors. In Preview & Smart-toggle, Preview reduced mode errors and Smart-toggle handled 86.2% of the mode errors that slipped past Preview. These results suggest that Preview & Smart-toggle is a promisIng method for preventIng mode errors for the Korean-English dual-Input environment.

Torc
TORC: A Virtual Reality Controller for In-Hand High-Dexterity Finger Interaction

Thursday 11:00 - 12:20 | Paper Session: Unexpected interactions | Clyde Auditorium

  • Jaeyeon Lee, School of Computing, KAIST
  • Mike Sinclair, Microsoft Research
  • Mar Gonzalez-Franco, Microsoft Research
  • Eyal Ofek, Microsoft Research
  • Christian Holz, Microsoft Research

Recent hand-held controllers have explored a variety of haptic feedback sensations for users in virtual reality by producing both kinesthetic and cutaneous feedback from virtual objects. These controllers are grounded to the user’s hand and can only manipulate objects through arm and wrist motions, not using the dexterity of their fingers as they would in real life. In this paper, we present TORC, a rigid haptic controller that renders virtual object characteristics and behaviors such as texture and compliance. Users hold and squeeze TORC using their thumb and two fingers and interact with virtual objects by sliding their thumb on TORC’s trackpad. During the interaction, vibrotactile motors produce sensations to each finger that represent the haptic feel of squeezing, shearing or turning an object. Our evaluation showed that using TORC, participants could manipulate virtual objects more precisely (e.g., position and rotate objects in 3D) than when using a conventional VR controller.

geometric
Geometrically Compensating Effect of End-to-End Latency in Moving-Target Selection Games

Wednesday 16:00-17:20 | Session of Gameplay Analysis and Latency | Room Hall2

  • Injung Lee, Graduate school of culture and technology, KAIST
  • Sunjun Kim, Aalto University
  • Byungjoo Lee, Graduate school of culture and technology, KAIST

Effects of unintended latency on gamer performance have been reported. End-to-end latency can be corrected by post- input manipulation of activation times, but this gives the player unnatural gameplay experience. For moving-target selection games such as Flappy Bird, the paper presents a predictive model of latency on error rate and a novel compensation method for the latency effects by adjusting the game’s geometry design – e.g., by modifying the size of the selection region. Without manipulation of the game clock, this can keep the user’s error rate constant even if the end- to-end latency of the system changes. The approach extends the current model of moving-target selection with two additional assumptions about the effects of latency: (1) latency reduces players’ cue-viewing time and (2) pushes the mean of the input distribution backward. The model and method proposed have been validated through precise experiments.

Late Breaking Work

what
“What does your Agent look like?” A Drawing Study to Understand Users’ Perceived Persona of Conversational Agent

Tuesday - 10:20 - 11:00 | Session of Des | Room: Hall 4

  • Sunok Lee, Department of Industrial Design, KAIST
  • Sungbae Kim, Department of Industrial Design, KAIST
  • Sangsu Lee, Department of Industrial Design, KAIST

Conversational agents (CAs) become more popular and useful at home. Creating the persona is an important part of designing the conversational user interface (CUI). Since the CUI is a voice-mediated interface, users naturally form an image of the CA’s persona through the voice. Because that image affects users’ interaction with CAs while using a CUI, we tried to understand users’ perception via drawing method. We asked 31 users to draw an image of the CA that communicates with the user. Through a qualitative analysis of the collected drawings and interviews, we could see the various types of CA personas perceived by users and found design factors that influenced users’ perception. Our findings help us understand persona perception, and that will provide designers with design implications for creating an appropriate persona.

designing
Designing Personalities of Conversational Agents

Tuesday - 10:20 - 11:00 | Session of Des | Room: Hall 4

  • Hankyung Kim, Department of Industrial Design, KAIST
  • Dong Yoon Koh, Department of Industrial Design, KAIST
  • Gaeun Lee, Samsung Research
  • Jung-Mi Park, Samsung Research
  • Youn-kyung Lim, Department of Industrial Design, KAIST

Recent hand-held controllers have explored a variety of haptic feedback sensations for users in virtual reality by producing both kinesthetic and cutaneous feedback from virtual objects. These controllers are grounded to the user’s hand and can only manipulate objects through arm and wrist motions, not using the dexterity of their fingers as they would in real life. In this paper, we present TORC, a rigid haptic controller that renders virtual object characteristics and behaviors such as texture and compliance. Users hold and squeeze TORC using their thumb and two fingers and interact with virtual objects by sliding their thumb on TORC’s trackpad. During the interaction, vibrotactile motors produce sensations to each finger that represent the haptic feel of squeezing, shearing or turning an object. Our evaluation showed that using TORC, participants could manipulate virtual objects more precisely (e.g., position and rotate objects in 3D) than when using a conventional VR controller.

solvedeep
SolveDeep: A System for Supporting Subgoal Learning in Online Math Problem Solving

Tuesday - 10:20 - 11:00 | Session of Des | Room: Hall 4

  • Hyoungwook Jin, School of Computing, KAIST
  • Minsuk Chang, School of Computing, KAIST
  • Juho Kim, School of Computing, KAIST

Learner-driven subgoal labeling helps learners form a hierarchical structure of solutions with subgoals, which are conceptual units of procedural problem solving. While learning with such hierarchical structure of a solution in mind is effective in learning problem solving strategies, the development of an interactive feedback system to support subgoal labeling tasks at scale requires significant expert efforts, making learner-driven subgoal labeling difficult to be applied in online learning environments. We propose SolveDeep, a system that provides feedback on learner solutions with peer-generated subgoals. SolveDeep utilizes a learnersourcing workflow to generate the hierarchical representation of possible solutions, and uses a graph-alignment algorithm to generate a solution graph by merging the populated solution structures, which are then used to generate feedback on future learners’ solutions. We conducted a user study with 7 participants to evaluate the efficacy of our system. Participants did subgoal learning with two math problems and rated the usefulness of system feedback. The average rating was 4.86 out of 7 (1: Not useful, 7: Useful), and the system could successfully construct a hierarchical structure of solutions with learnersourced subgoal labels.

crowd
Crowdsourcing Perspectives on Public Policy from Stakeholders

Tuesday - 10:20 - 11:00 | Session of Des | Room: Hall 4

  • Hyunwoo Kim, School of Computing, KAIST
  • Eun-Young Ko, School of Computing, KAIST
  • Donghoon Han, School of Computing, KAIST
  • Sung-Chul Lee, School of Computing, KAIST
  • Simon T. Perrault, Singapore University of Technology and Design, Singapore
  • Jihee Kim, School of Business and Technology Management, KAIST​
  • Juho Kim, School of Computing, KAIST

Personal deliberation, the process through which people can form an informed opinion on social issues, serves an important role in helping citizens construct a rational argument in the public deliberation. However, existing information channels for public policies deliver only few stakeholders’ voices, thus failing to provide a diverse knowledge base for personal deliberation. This paper presents an initial design of PolicyScape, an online system that supports personal deliberation on public policies by helping citizens explore diverse stakeholders and their perspectives on the policy’s effect. Building on literature on crowdsourced policymaking and policy stakeholders, we present several design choices for crowdsourcing stakeholder perspectives. We introduce perspective-taking as an approach for personal deliberation by helping users consider stakeholder perspectives on policy issues. Our initial results suggest that PolicyScape could collect diverse sets of perspectives from the stakeholders of public policies, and help participants discover unexpected viewpoints of various stakeholder groups.

improving
Improving Two-Thumb Touchpad Typing in Virtual Reality

Wednesday 15:20 - 16:00 | Late-breaking Work: Poster Rotation 2 | Hall 4

  • Jeongmin Son, School of Computing, KAIST
  • Sunggeun Ahn, School of Computing, KAIST
  • Sunbum Kim, School of Computing, KAIST
  • Geehyuk Lee, School of Computing, KAIST

Two-Thumb Touchpad Typing (4T) using hand-held controllers is one of the common text entry techniques in Virtual Reality (VR). However, its performance is far below that of Two-Thumb Typing on a smartphone. We explored the possibility of improving its performance focusing on the following two factors: the visual feedback of hovering thumbs and the grip stability of the controllers. We examined the effects of these two factors on the performance of 4T in VR in user experiments. Their results show that hover feedback had a significant main effect on the 4T performance, but grip stability did not. We then investigated the achievable performance of the final 4T design in a longitudinal study, and its results show that users could achieve a Typing speed over 30 words per minute after two hours of practice.

fingmag
FingMag: Finger Identification Method for Smartwatch

Wednesday 15:20 - 16:00 | Late-breaking Work: Poster Rotation 2 | Hall 4

  • Keunwoo Park, School of Computing, KAIST
  • Geehyuk Lee, School of Computing, KAIST

Interacting with a smartwatch is difficult owing to its small touchscreen. A general strategy to overcome the limitations of the small screen is to increase the input vocabulary. A popular approach to do this is to distinguish fingers and assign different functions to them. As a finger identification method for a smartwatch, we propose FingMag, a machine-learning-based method that identifies the finger on the screen with the help of a ring. for this identification, the finger’s touch position and the magnetic field from a magnet embedded in the ring are used. In an offline evaluation using data collected from 12 participants, we show that FingMag can identify the finger with an accuracy of 96.21% in stationary geomagnetic conditions.

Workshops & Symposia

uc
User-Centered Graphical Models of Interaction

Workshop on Computational Modeling in Human-Computer Interaction

  • Minsuk Chang, School of Computing, KAIST
  • Juho Kim, School of Computing, KAIST

In this position paper, I present a set of data-driven techniques in modeling the learning material, learner workflow and the learning task as graphical representations, with which at scale can createn and support learning opportunities in the wild. I propose the graphical models resulting from this bottom-up approach can further serve as proxies for representing learnability bounds of an interface. I also propose an alternative approach which directly aims to “learn” the interaction bounds by modeling the interface as an agent’s sequential decision-making problem. Then I illustrate how the data-driven modeling techniques and algorithm modeling techniques can create a mutually beneficial bridge for advancing design of interfaces.

Readersourcing
Readersourcing an Accurate and Comprehensive Understanding of Health-related Information Represented by Media

Workshop on HCI for Accurate, Impartial and Transparent Journalism: Challenges and Solutions

  • Eun-Young Ko, School of Computing, KAIST
  • Ching Liu, National Tsing Hua University
  • Hyuntak Cha, Seoul National University
  • Juho Kim, School of Computing, KAIST

Health news delivers findings from health-related research to the public. As the delivered information may affect the public’s everyday decision or behavior, readers should get an accurate and compre- hensive understanding of the research from articles they read. However, it is rarely achieved due to incomplete information delivered by the news stories and a lack of critical evaluation of readers. In this position paper, we propose a readersourcing approach, an idea of engaging readers in a critical reading activity while collecting valuable artifacts for future readers to acquire a more accurate and comprehensive understanding of health-related information. We discuss challenges, opportunities, and design considerations in the readersourcing approach. Then we present the initial design of a web-based news reading application that connects health news readers via questioning and answering tasks.

playmaker
PlayMaker: A Participatory Design Method for Creating Entertainment Application Concepts Using Activity Data

Workshop on HCI for Accurate, Impartial and Transparent Journalism: Challenges and Solutions

  • Dong Yoon Koh, Department of Industrial Design, KAIST
  • Ju Yeon Kim, Department of Industrial Design, KAIST
  • Donghyeok Yun, Department of Industrial Design, KAIST
  • Youn-kyung Lim, Department of Industrial Design, KAIST

The public’s ever-growing interest in health has led the well-being industry to explosive growth over the years. This propelled activity trackers as one of the trendiest items among the current day wearable devices. Seeking new opportunities for effective data utilization, we present a participatory design method that explores the fusion of activity data with entertainment application. In this method we spur participants to design by mix-and-matching activity tracker data attributes to existing entertainment application features to produce new concepts. We report two cases of method implementation and further discuss the opportunities of activity tracker data as means for entertainment application design.

Let’s take KAIST HCI friends photos together! Take a photo with the HCI KAIST pose as shown in the example and upload it to the Instagram. With hash tags #HCIKAIST@CHI2019 or #KAISTNIGHT.​

CHI 2018

At CHI 2018, HCI@KAIST presents 19 papers, 11 late breaking works, 5 demonstrations and 1 video showcase.
These works are from 20 labs of 4 different schools and departments at KAIST. We thank our outstanding colleagues and collaborators from industry, research centers and universities around the world. 

Please come to the KAIST Night on April 25, 2018. Let’s take KAIST HCI Friends photo together!

HCI KAIST Reception
DATE  April 25, 2018
TIME  20:00 – 22:00
LOCATION  Joverse, Montreal

Paper & Notes

mechanism
Mechanism Perfboard: An Augmented Reality Environment for Linkage Mechanism Design and Fabrication

Monday 11:30-11:50 | Session of Craft, Fabrication, Making | Room 516C

  • Yunwoo Jeong, Department of Industrial Design, KAIST
  • Han-Jong Kim, Department of Industrial Design, KAIST
  • Tek-Jin Nam, Department of Industrial Design, KAIST

Prototyping devices with kinetic mechanisms, such as automata and robots, has become common in physical computing projects. However, mechanism design in the early-concept exploration phase is challenging, due to the dynamic and unpredictable characteristics of mechanisms. We present Mechanism Perfboard, an augmented reality environment that supports linkage mechanism design and fabrication. It supports the concretization of ideas by generating the initial desired linkage mechanism from a real world movement. The projection of simulated movement within the environment enables iterative tests and modifications in real scale. Augmented information and accompanying tangible parts help users to fabricate mechanisms. Through a user study with 10 participants, we found that Mechanism Perfboard helped the participant to achieve their desired movement. The augmented environment enabled intuitive modification and fabrication with an understanding of mechanical movement. Based on the tool development and the user study, we discuss implications for mechanism prototyping with augmented reality and computational support.

clight
c.light: A Tool for Exploring Light Properties in Early Design Stage

Monday 15:10-15:30 | Session of Tools for Designing | Room 514C

  • Kyeong Ah Jeong, Department of Industrial Design, KAIST
  • Eunjin Kim, Department of Industrial Design, KAIST
  • Taesu Kim, Department of Industrial Design, KAIST
  • Hyeon-Jeong Suk, Department of Industrial Design, KAIST

Although a light becomes an important design element, there are little techniques available to explore shapes and light effects in early design stages. We present c.light, a design tool that consists of a set of modules and a mobile application for visualizing the light in a physical world. It allows designers to easily fabricate both tangible and intangible properties of a light without a technical barrier. We analyzed how c.light contributes to the ideation process of light design through a workshop. The results showed that c.light largely expands designers’ capability to manipulate intangible properties of light and, by doing so, it facilitates collaborative and inverted ideation process in early design stages. It is expected that the results of this study could enhance our understanding of how designers manipulate light in a physical world in early design stages and could be a good stepping stone for future tool development

projective window
Projective Windows: Bringing Windows in Space to the Fingertip

Monday 15:10-15:30 | Session of Modelling AR & VR | Room 517A

  • Joon Hyub Lee, Department of Industrial Design, KAIST
  • Sang-Gyun An, Department of Industrial Design, KAIST
  • Yongkwan Kim, Department of Industrial Design, KAIST
  • Seok-Hyung Bae Department of Industrial Design, KAIST

In augmented and virtual reality (AR and VR), there may be many 3D planar windows with 2D texts, images, and videos on them. However, managing the position, orientation, and scale of such a window in an immersive 3D workspace can be difficult. Projective Windows strategically uses the absolute and apparent sizes of the window at various stages of the interaction to enable the grabbing, moving, scaling, and releasing of the window in one continuous hand gesture. With it, the user can quickly and intuitively manage and interact with windows in space without any controller hardware or dedicated widget. Through an evaluation, we demonstrate that our technique is performant and preferable, and that projective geometry plays an important role in the design of spatial user interfaces.

Impactive activation
Impact Activation Improves Rapid Button Pressing

Monday 17:10-17:30 | Session of Interaction Under Pressure | Room 516E

  • Sunjun Kim, Aalto University, Finland; School of Culture Technology, KAIST
  • Byungjoo Lee, School of Culture Technology, KAIST
  • Antti Oulasvirta, Aalto University, Finland

The activation point of a button is defined as the depth at which it invokes a make signal. Regular buttons are activated during the downward stroke, which occurs within the first 20 ms of a press. The remaining portion, which can be as long as 80~ms, has not been examined for button activation for reason of mechanical limitations. The paper presents a technique and empirical evidence for an activation technique called Impact Activation, where the button is activated at its maximal impact point. We argue that this technique is advantageous particularly in rapid, repetitive button pressing, which is common in gaming and music applications. We report on a study of rapid button pressing, wherein users’ timing accuracy improved significantly with use of Impact Activation. The technique can be implemented for modern push-buttons and capacitive sensors that generate a continuous signal.

agile 3d
Agile 3D Sketching with Air Scaffolding

Monday 16:10-16:30 | Session of Creativity, Sketching & Animation | Room 517D

  • Yongkwan Kim, Department of Industrial Design, KAIST
  • Sang-Gyun An, Department of Industrial Design, KAIST
  • Joon Hyub Lee, Department of Industrial Design, KAIST
  • Seok-Hyung Bae, Department of Industrial Design, KAIST

Hand motion and pen drawing can be intuitive and expressive inputs for professional digital 3D authoring. However, their inherent limitations have hampered wider adoption. 3D sketching using hand motion is rapid but rough, and 3D sketching using pen drawing is delicate but tedious. Our new 3D sketching workflow combines these two in a complementary manner. The user makes quick hand motions in the air to generate approximate 3D shapes, and uses them as scaffolds on which to add details via pen-based 3D sketching on a tablet device. Our air scaffolding technique and corresponding algorithm extract only the intended shapes from unconstrained hand motions. Then, the user sketches 3D ideas by defining sketching planes on these scaffolds while appending new scaffolds, as needed. A user study shows that our progressive and iterative workflow enables more agile 3D sketching compared to ones using either hand motion or pen drawing alone.

understanding the effect of
Understanding the Effect of In-Video Prompting on Learners and Instructors

Tuesday 9:00-9:20 | Session of Learning and training | Room 517D

  • Hyungyu Shin, School of Computing, KAIST
  • Eun-Young Ko, School of Computing, KAIST
  • Joseph Jay Williams, National University of Singapore, Singapore
  • Juho Kim, School of Computing, KAIST

Online instructional videos are ubiquitous, but it is difficult for instructors to gauge learners’ experience and their level of comprehension or confusion regarding the lecture video. Moreover, learners watching the videos may become disengaged or fail to reflect and construct their own understanding. This paper explores instructor and learner perceptions of in-video prompting where learners answer reflective questions while watching videos. We conducted two studies with crowd workers to understand the effect of prompting in general, and the effect of different prompting strategies on both learners and instructors. Results show that some learners found prompts to be useful checkpoints for reflection, while others found them distracting. Instructors reported the collected responses to be generally more specific than what they have usually collected. Also, different prompting strategies had different effects on the learning experience and the usefulness of responses as feedback.

moving target selection
Moving Target Selection: A Cue Integration Model

Tuesday 11:00-11:20 | Session of Buttons, Targets, Sliders | Room 518AB

  • Byungjoo Lee, School of Culture Technology, KAIST
  • Sunjun Kim, School of Culture Technology, KAIST; Aalto University, Finland
  • Antti Oulasvirta, Aalto University, Finland
  • JONG-IN LEE, School of Culture Technology, KAIST; Aalto University, Finland
  • Eunji Park, School of Culture Technology, KAIST

This paper investigates a common task requiring temporal precision: the selection of a rapidly moving target on display by invoking an input event when it is within some selection window. Previous work has explored the relationship between accuracy and precision in this task, but the role of visual cues available to users has remained unexplained. To expand modeling of timing performance to multimodal settings, common in gaming and music, our model builds on the principle of probabilistic cue integration. Maximum likelihood estimation (MLE) is used to model how different types of cues are integrated into a reliable estimate of the temporal task. The model deals with temporal structure (repetition, rhythm) and the perceivable movement of the target on display. It accurately predicts error rate in a range of realistic tasks. Applications include the optimization of difficulty in game-level design.

neuromechanics of a button
Neuromechanics of a Button Press

Tuesday 11:20-11:40 | Session of Buttons, Targets, Sliders | Room 518AB

  • Antti Oulasvirta, Aalto University, Finland
  • Sunjun Kim, Aalto University, Finland
  • Byungjoo Lee, School of Culture Technology, KAIST

To press a button, a finger must push down and pull up with the right force and timing. How the motor system succeeds in button-pressing, in spite of neural noise and lacking direct access to the mechanism of the button, is poorly understood. This paper investigates a unifying account based on neuromechanics. Mechanics is used to model muscles controlling the finger that contacts the button. Neurocognitive principles are used to model how the motor system learns appropriate muscle activations over repeated strokes though relying on degraded sensory feedback. Neuromechanical simulations yield a rich set of predictions for kinematics, dynamics, and user performance and may aid in understanding and improving input devices. We present a computational implementation and evaluate predictions for common button types.

enhancing online
Enhancing Online Problems Through Instructor-Centered Tools for Randomized Experiments

Tuesday 11:00-11:20 | Session of Automated and Crowd Supports for Learning | Room 518C

  • Joseph Jay Williams, National University of Singapore, Singapore
  • Anna N. Rafferty, Carleton College, USA
  • Dustin Tingley, Harvard, USA
  • Andrew Ang, Harvard, USA
  • Walter S. Lasecki, University of Michigan, USA
  • Juho Kim, School of Computing, KAIST

Digital educational resources could enable the use of randomized experiments to answer pedagogical questions that instructors care about, taking academic research out of the laboratory and into the classroom. We take an instructor-centered approach to designing tools for experimentation that lower the barriers for instructors to conduct experiments. We explore this approach through DynamicProblem, a proof-of-concept system for experimentation on components of digital problems, which provides interfaces for authoring of experiments on explanations, hints, feedback messages, and learning tips. To rapidly turn data from experiments into practical improvements, the system uses an interpretable machine learning algorithm to analyze students’ ratings of which conditions are helpful, and present conditions to future students in proportion to the evidence they are higher rated. We evaluated the system by collaboratively deploying experiments in the courses of three mathematics instructors. They reported benefits in reflecting on their pedagogy, and having a new method for improving online problems for future students.

collaborate dynamic-01
Collaborative Dynamic Queries: Supporting Distributed Small Group Decision-making

Tuesday 14:00-14:20 | Session of Distributed Work | Room 516C

  • Sungsoo (Ray) Hong, University of Washington, USA
  • Minhyang (Mia) Suh, University of Washington, USA
  • Nathalie Henry Riche, Microsoft Research, USA
  • Jooyoung Lee, School of Computing, KAIST
  • Juho Kim, School of Computing, KAIST
  • Mark Zachry, University of Washington, USA

Communication is critical in small group decision-making processes during which each member must be able to express preferences to reach consensus. Finding consensus can be difficult when each member in a group has a perspective that potentially conflicts with those of others. To support groups attempting to harmonize diverse preferences, we propose Collaborative Dynamic Queries (C-DQ), a UI component that enables a group to filter queries over decision criteria while being aware of others’ preferences. To understand how C-DQ affects a group’s behavior and perception in the decision-making process, we conducted 2 studies with groups who were prompted to make decisions together on mobile devices in a dispersed and synchronous situation. In Study 1, we found showing group preferences with C-DQ helped groups to communicate more efficiently and effectively. In Study 2, we found filtering candidates based on each member’s own filter range further improved a groups’ communication efficiency and effectiveness.

bebecode
BebeCode: Collaborative Child Development Tracking System

Wednesday 9:20-9:40 | Session of Children, Well-being, and Play | Room 518AB

  • Seokwoo Song, School of Computing, KAIST
  • Juho Kim, School of Computing, KAIST
  • Bumsoo Kang, School of Computing, KAIST
  • Wonjeong Park, Ewha Womans University
  • John Kim, KAIST

Continuous tracking young children’s development is important for parents because early detection of developmental delay can lead to better treatment through early intervention. Screening tests, often based on questions answered by a parent, are used to assess children’s development, but responses from only one parent can be subjective and even inaccurate due to limited memory and observations. In this work, we propose a collaborative child development tracking system, where screening test responses are collected through collaboration between parents or caregivers. We implement BebeCODE, a mobile system that encourages parents to independently answer all developmental questions for a given age and resolve disagreements through chatting, image/video sharing, or asking a third person. A 4-week deployment study of BebeCODE with 12 families found that parents had approximately 22% disagreements about questions regarding their children’s developmental and BebeCODE helped them reach a consensus. Parents also reported that their awareness of their child’s development, increased with BebeCODE.

too close and crowded
Too Close and Crowded: Understanding Stress on Mobile Instant Messengers based on Proxemics

Wednesday 11:40-12:00 | Session of Roads and Crowds | Room 516AB

  • In-geon Shin, Department of Industrial Design, KAIST
  • Jin-min Seok, Department of Industrial Design, KAIST
  • Youn-kyung Lim, Department of Industrial Design, KAIST

Nowadays, mobile instant messaging (MIM) is a necessity for our private and public lives, but it has also been the cause of stress. In South Korea, MIM stress has become a serious social problem. To understand this stress, we conducted four focus groups with 20 participants under MIM stress. We initially discovered that MIM stress relates to how people perceive the territory in MIM. We then applied proxemics—the theory of human use of space—to the thematic analysis as the rationale. The data revealed two main themes: too close and too crowded. The participants were stressed due to design features that let strangers or crowds into their MIM applications and forced them to interact and share their status with them. Based on this finding, we propose a set of implications for designing anti-stress MIM applications.

conceptscape collaborative-01
ConceptScape: Collaborative Concept Mapping for Video Learning

Wednesday 14:00-14:20 | Session of Learning 2 | Room 517D

  • Ching Liu, National Tsing Hua University, Taiwan
  • Juho Kim, School of Computing, KAIST
  • Hao-Chuan Wang, National Tsing Hua University, Taiwan; University of California, USA

While video has become a widely adopted medium for online learning, existing video players provide limited support for navigation and learning. It is difficult to locate parts of the video that are linked to specific concepts. Also, most video players afford passive watching, thus making it difficult for learners with limited metacognitive skills to deeply engage with the content and reflect on their understanding. To support concept-driven navigation and comprehension of lecture videos, we present ConceptScape, a system that generates and presents a concept map for lecture videos. ConceptScape engages crowd workers to collaboratively generate a concept map by prompting them to externalize reflections on the video. We present two studies to show that (1) interactive concept maps can be useful tools for concept-based video navigation and comprehension, and (2) with ConceptScape, novice crowd workers can collaboratively generate complex concept maps that match the quality of those by experts.

hapcube-01
HapCube: A Wearable Tactile Device to Provide Tangential and Normal Pseudo-Force Feedback on a Fingertip

Wednesday 16:00-16:20 | Session of Haptic Wearables | Room 517C

  • Hwan Kim, Department of Industrial Design, KAIST
  • HyeonBeom Yi, Department of Industrial Design, KAIST
  • Hyein Lee, Department of Industrial Design, KAIST
  • Woohun Lee, Department of Industrial Design, KAIST

Haptic devices allow a more immersive experience with Virtual and Augmented Reality. However, for a wider range of usage they need to be miniaturized while maintaining the quality of haptic feedback. In this study, we used two kinds of human sensory illusion of vibration. The first illusion involves creating a virtual force (pulling sensation) using asymmetric vibration, and the second involves imparting compliances of complex stress-strain curves (i.e. force-displacement curves of mechanical keyboards) to a rigid object by changing the frequency and amplitude of vibration. Using these two illusions, we developed a wearable tactile device named HapCube, consisting of three orthogonal voicecoil actuators. Four measurement tests and four user tests confirmed that 1) a combination of two orthogonal asymmetric vibrations could provide a 2D virtual force in any tangential directions on a finger pad, and 2) a single voicecoil actuator produced pseudo-force feedback of the complex compliance curves in the normal direction.

exploring multimodal
Exploring Multimodal Watch-back Tactile Display using Wind and Vibration

Wednesday 16:20-16:40 | Session of Haptic Wearables | Room 517C

  • Youngbo Aram Shim, School of Computing, KAIST
  • Jaeyeon Lee, School of Computing, KAIST
  • Geehyuk Lee, School of Computing, KAIST

A tactile display on the back of a smartwatch is an attractive output option; however, its channel capacity is limited owing to the small contact area. In order to expand the channel capacity, we considered using two perceptually distinct types of stimuli, wind and vibration, together on the same skin area. The result is a multimodal tactile display that combines wind and vibration to create “colored” tactile sensations on the wrist. As a first step toward this goal, we conducted in this study four user experiments with a wind-vibration tactile display to examine different ways of combining wind and vibration: Individual, Sequential, and Simultaneous. The results revealed the sequential combination of wind and vibration to exhibit the highest potential, with an information transfer capacity of 3.29 bits. In particular, the transition of tactile modality was perceived at an accuracy of 98.52%. The current results confirm the feasibility and potential of a multimodal tactile display combining wind and vibration.

pokering
PokeRing: Notifications by Poking Around the Finger

Wednesday 16:40-17:00 | Session of Haptic Wearables | Room 517C

  • Seungwoo Je, Department of Industrial Design, KAIST
  • Minkyeong Lee, Department of Industrial Design, KAIST
  • Yoonji Kim, Department of Industrial Design, KAIST
  • Liwei Chan, National Chiao Tung University, Taiwan
  • Xing-Dong Yang, Dartmouth College, USA
  • Andrea Bianchi, Department of Industrial Design, KAIST

Smart-rings are ideal for subtle and always-available haptic notifications due to their direct contact with the skin. Previous researchers have highlighted the feasibility of haptic technology in smart-rings and their promise in delivering noticeable stimulations by poking a limited set of planar locations on the finger. However, the full potential of poking as a mechanism to deliver richer and more expressive information on the finger is overlooked. With three studies and a total of 76 participants, we informed the design of PokeRing, a smart-ring capable of delivering information via stimulating eight different locations around the index finger’s proximal phalanx. We report our evaluation of the performance of PokeRing in semi-realistic wearable conditions, (standing and walking), and its effective usage for information transfer with twenty-one spatio-temporal patterns designed by six interaction designers in a workshop. Finally, we present three applications that exploit PokeRing’s notification usages.

recipescape
RecipeScape: An Interactive Tool for Analyzing Cooking Instructions at Scale

Thursday 9:00-9:20 | Session of Crowdsourcing, data mining, dealing with information | Room 516E

  • Minsuk Chang, School of Computing, KAIST
  • Leonore V. Guillain, ecole polytechnique fédérale de lausanne, Switzerland
  • Hyeungshik Jung, School of Computing, KAIST
  • Vivian M. Hare, Stanford University, USA; Chan Zuckerberg Initiative, USA
  • Juho Kim, School of Computing, KAIST
  • Maneesh Agrawala, Stanford University, USA

For cooking professionals and culinary students, understanding cooking instructions is an essential yet demanding task. Common tasks include categorizing different approaches to cooking a dish and identifying usage patterns of particular ingredients or cooking methods, all of which require extensive browsing and comparison of multiple recipes. However, no existing system provides support for such in-depth and at-scale analysis. We present RecipeScape, an interactive system for browsing and analyzing the hundreds of recipes of a single dish available online. We also introduce a computational pipeline that extracts cooking processes from recipe text and calculates a procedural similarity between them. To evaluate how RecipeScape supports culinary analysis at scale, we conducted a user study with cooking professionals and culinary students with 500 recipes for two different dishes. Results show that RecipeScape clusters recipes into distinct approaches, and captures notable usage patterns of ingredients and cooking actions.

thors hammer
Thor’s Hammer: An Ungrounded Force Feedback Device Utilizing Propeller-Induced Propulsive Force

Thursday 12:00-12:20 | Session of Force Feedback in VR | Room 516D

  • Seongkook Heo, University of Toronto, Canada
  • Christina Chung, University of Toronto, Canada
  • Geehyuk Lee, School of Computing, KAIST
  • Daniel Wigdor, University of Toronto, Canada

We present a new handheld haptic device, Thor’s Hammer, which uses propeller propulsion to generate ungrounded, 3-DOF force feedback. Thor’s Hammer has six motors and propellers that generates strong thrusts of air without the need for physical grounding or heavy air compressors. With its location and orientation tracked by an optimal tracking system, the system can exert forces in arbitrary directions regardless of the device’s orientation. Our technical evaluation shows that Thor’s Hammer can apply up to 4 N of force in arbitrary directions with less than 0.11 N and 3.9° of average magnitude and orientation errors. We also present virtual reality applications that can benefit from the force feedback provided by Thor’s Hammer. Using these applications, we conducted a preliminary user study and participants felt the experience more realistic and immersive with the force feedback.

to distrot or not
To Distort or Not to Distort: Distance Cartograms in the Wild

Thursday 14:00-14:20 | Session of Visualization of Space and Shape | Room 518C

  • Sungsoo (Ray) Hong, University of Washington, USA
  • Min-Joon Yoo, New York University, USA
  • Bonnie Chinh, University of Washington, USA
  • Amy Han, Swathmore, USA
  • Sarah Battersby, Tableau Software, USA
  • Juho Kim, School of Computing, KAIST

Distance Cartograms (DC) distort geographical features so that the measured distance between a single location and any other location on a map indicates absolute travel time. Although studies show that users can efficiently assess travel time with DC, distortion applied in DC may confuse users, and its usefulness “in the wild” is unknown. To understand how real world users perceive DC’s benefits and drawbacks, we devise techniques that improve DC’s presentation (preserving topological relationships among map features while aiming at retaining shapes) and scalability (presenting accurate live travel time). We developed a DC-enabled system with these techniques, and deployed it to 20 participants for 4 weeks. During this period, participants spent, on average, more than 50% of their time with DC as opposed to a standard map. Participants felt DC to be intuitive and useful for assessing travel time. They indicated intent in adopting DC in their real-life scenarios.

Demonstrations

hapcube-01
HapCube: A Wearable Tactile Device to Provide Tangential and Normal Pseudo-Force Feedback on a Fingertip

Monday 18:00-21:00 | D304 | Room 220BC

  • Hwan Kim, Department of Industrial Design, KAIST
  • HyeonBeom Yi, Department of Industrial Design, KAIST
  • Hyein Lee, Department of Industrial Design, KAIST
  • Woohun Lee, Department of Industrial Design, KAIST
exploring multimodal
Exploring Multimodal Watch-back Tactile Display using Wind and Vibration

Monday 18:00-21:00 | D305 | Room 220BC

  • Youngbo Aram Shim, School of Computing, KAIST
  • Jaeyeon Lee, School of Computing, KAIST
  • Geehyuk Lee, School of Computing, KAIST
thors hammer
Thor’s Hammer: An Ungrounded Force Feedback Device Utilizing Propeller-Induced Propulsive Force

Monday 18:00-21:00 | D110 | Room 220BC

  • Seongkook Heo, University of Toronto, Canada
  • Christina Chung, University of Toronto, Canada
  • Geehyuk Lee, School of Computing, KAIST
  • Daniel Wigdor, University of Toronto, Canada
agile 3d
Agile 3D Sketching with Air Scaffolding

Monday 18:00-21:00 | D411 | Room 220BC

  • Yongkwan Kim, Department of Industrial Design, KAIST
  • Sang-Gyun An, Department of Industrial Design, KAIST
  • Joon Hyub Lee, Department of Industrial Design, KAIST
  • Seok-Hyung Bae, Department of Industrial Design, KAIST
projective window
Projective Windows: Bringing Windows in Space to the Fingertip

Monday 18:00-21:00 | D412 | Room 220BC

  • Joon Hyub Lee, Department of Industrial Design, KAIST
  • Sang-Gyun An, Department of Industrial Design, KAIST
  • Yongkwan Kim, Department of Industrial Design, KAIST
  • Seok-Hyung Bae Department of Industrial Design, KAIST

Late Breaking Work

tnt
TNT: Exploring Pseudo Social Reminding for Effective Task Management

Tuesday 10:20-11:00 / 15:20-16:00 | LBW035 | Room 220BC

  • Wonyoung Shin, Graduate School of Knowledge Service Engineering, KAIST
  • Soowon Kang, Graduate School of Knowledge Service Engineering, KAIST
  • Inyeop Kim, Graduate School of Knowledge Service Engineering, KAIST
  • Mun Yong Yi, Department of Industrial&Systems Engineering, KAIST
  • Uichin Lee, Graduate School of Knowledge Service Engineering, KAIST
exprgram
Exprgram: A Language Learning Interface for Mastering Pragmatic Competence

Tuesday 10:20-11:00 / 15:20-16:00 | LBW057 | Room 220BC

  • Kyung Je Jo, School of Computing, KAIST
  • John Joon Young, School of Computing, KAIST
  • Chung
  • Juho Kim, School of Computing, KAIST

Mastering pragmatic competence, the ability to use language in a contextually appropriate way, is one of the most challenging parts of foreign language learning. Despite its importance, existing language learning systems often focus on linguistic components such as grammar, vocabulary, or pronunciation. Consequently, foreign language learners may generate grammatically flawless speech that is contextually inappropriate. With the diverse socio-cultural contexts captured in real-life settings, videos at scale can serve as a great material for learners to acquire pragmatic competence. We introduce Exprgram, a webbased video learning interface that assists learners to master pragmatic competence. With Exprgram, learners can raise their contextawareness, practice generating an alternative expression, and learn alternative expressions for the given context. Our user study with 12 advanced English learners shows potential in our learnersourcing approach to collecting descriptive context annotations and diverse alternative expressions.

enhancing storytelling
Enhancing Storytelling Experience with Story-Aware Interactive Puppet

Tuesday 10:20-11:00 / 15:20-16:00 | LBW076 | Room 220BC

  • Bogyeong Kim, Department of Industrial Design, KAIST
  • Jaehoon Pyun, Department of Industrial Design, KAIST
  • Woohun Lee, Department of Industrial Design, KAIST

Puppets are often used in storytelling, but there are few studies about puppets regarding the storytelling experience. In this paper, we introduce the concept of an ideal puppet for storytelling and discuss directions for puppet development. The ideal puppet is able to automatically animate itself in line with a story plot and positively influence the interactions in the storytelling dynamic. To see how children and parents would accept the concept, we created a preliminary prototype and conducted user study using Wizard-of-Oz method. Participants experienced enhanced immersion and increased communication between them by through the automatic movement of the puppet. They expected various roles from the puppet such as actor, support tool, and friend, which made various usage scenarios possible. The puppet should be developed in the direction of enhancing its advantages and including various usage scenarios, especially by combining the needs of both automation and manipulation.

designinghealth
Designing Health-Promoting Technologies with IoT at Home

Tuesday 10:20-11:00 / 15:20-16:00 | LBW083 | Room 220BC

  • Eulim Sull, Department of Industrial Design, KAIST
  • Youn-kyung Lim, Department of Industrial Design, KAIST

Health-related IT products (e.g., Fitbit) support persuasive technologies to reinforce an individual’s desired behaviors. While these products are dedicated to certain health behaviors, such as walking or specific types of sports, IoT at home can be integrated more broadly throughout one’s daily life. To address this opportunity, this paper aims to shed light on the use of domestic IoT that can foster changes toward healthy behaviors through a 3-week explorative field trial. This paper reports two major goals of health-promoting technologies using IoT as well as different persuasive techniques according to the temporal phases of before, during, and after the health behaviors.

button++
Button++: Designing Risk-aware Smart Buttons

Tuesday 10:20-11:00 / 15:20-16:00 | LBW116 | Room 220BC

  • Eunji Park, School of Culture Technology, KAIST
  • Hyunju Kim, School of Culture Technology, KAIST
  • Byungjoo Lee, School of Culture Technology, KAIST

Buttons are the most commonly used input devices. So far the goal of the designers was to provide a passive button that can accept user input as easily as possible. Therefore, based on Fitts’ law, they maximize the size of the button and make the distance closer. This paper proposes Button++, a novel method to design smart buttons that actively judge user’s movement risk and selectively trigger input. Based on the latest model of moving target selection, Button++ tracks the user’s submovement just before the click and infers the expected error rate that can occur if the user repeatedly clicks with the same movement. This allows designers to make buttons that actively respond to the amount of risk in the user’s input movement.

detecting personality unobtrusive
Detecting Personality Unobtrusively from Users' Online and Offline Workplace Behaviors

Wednesday 10:20-11:00 / 15:20-16:00 | LBW515 | Room 220BC

  • Seoyoung Kim, School of Computing, KAIST
  • Jiyoun Ha, School of Computing, KAIST 
  • Juho Kim, School of Computing, KAIST

Personality affects various social behaviors of an individual, such as collaboration, group dynamics, and social relationships within the workplace. However, existing methods for assessing personality have shortcomings: self-assessed methods are cumbersome due to repeated assessment and erroneous due to a self-report bias. On the other hand, automatic, data-driven personality detection raises privacy concerns due to a need for excessive personal data. We present an unobtrusive method for detecting personality within the workplace that combines a user’s online and offline behaviors. We report insights from analyzing data collected from four different workplaces with 37 participants, which shows that complementing online and offline data allows a more complete reflection of an individual’s personality. We also present possible applications of unobtrusive personality detection in the workplace.

micro-ngo
Micro-NGO: Tackling Wicked Social Problems with Problem Solving and Action Planning Support in Chat

Wednesday 10:20-11:00 / 15:20-16:00 | LBW559 | Room 220BC

  • Joonyoung Park, Graduate School of Knowledge Service Engineering, KAIST
  • Jin Yong Sim, Graduate School of Knowledge Service Engineering, KAIST
  • Jaejeung Kim, Graduate School of Knowledge Service Engineering, KAIST
  • Mun Yong Yi, Graduate School of Knowledge Service Engineering, KAIST
  • Uichin Lee, Graduate School of Knowledge Service Engineering, KAIST

Health-related IT products (e.g., Fitbit) support persuasive technologies to reinforce an individual’s desired behaviors. While these products are dedicated to certain health behaviors, such as walking or specific types of sports, IoT at home can be integrated more broadly throughout one’s daily life. To address this opportunity, this paper aims to shed light on the use of domestic IoT that can foster changes toward healthy behaviors through a 3-week explorative field trial. This paper reports two major goals of health-promoting technologies using IoT as well as different persuasive techniques according to the temporal phases of before, during, and after the health behaviors.

InteractionRestraint1
Interaction Restraint: Enforcing Adaptive Cognitive Tasks to Restrain Problematic User Interaction

Wednesday 10:20-11:00 / 15:20-16:00 | LBW553 | Room 220BC

  • Sung-Chul Lee, College of Business, KAIST
  • Jihee Kim, College of Business, KAIST
  • Juho Kim, School of Computing, KAIST

When a group of citizens wants to tackle a social problem online, they need to discuss the problem, possible solutions, and concrete actions. Instant messengers are a common tool used in this setting, which support free and unstructured discussion. But tackling complex social problems often calls for structured discussion. In this paper, we present Micro-NGO, a chat-based online discussion platform with built-in support for (1) the problem-solving process and (2) the action planning process. To scaffold the process, Micro-NGO adopts a question prompting strategy, which asks relevant questions to users in each stage of the problem-solving process. Users can answer the questions and vote for the best answer while they freely discuss in the chat room. For an informal evaluation, we conducted a pilot study with two groups (n=7). The participants held a discussion while collectively answering the question prompts and reached consensus to send a petition letter about campus issues to the related personnel.

identifying everyday objects
Identifying Everyday Objects with a Smartphone Knock

Wednesday 10:20-11:00 / 15:20-16:00 | LBW606 | Room 220BC

  • Taesik Gong, School of Computing, KAIST
  • Hyunsung Cho, School of Computing, KAIST
  • Bowon Lee, Inha University
  • Sung-Ju Lee, School of Computing, KAIST

We use smartphones and their apps for almost every daily activity. For instance, to purchase a bottle of water online, a user has to unlock the smartphone, find the right e-commerce app, search the name of the water product, and finally place an order. This procedure requires manual, often cumbersome, input of a user, but could be significantly simplified if the smartphone can identify an object and automatically process this routine. We present Knocker, an object identification technique that only uses commercial off-theshelf smartphones. The basic idea of Knocker is to leverage a unique set of responses that occur when a user knocks on an object with a smartphone, which consist of the generated sound from the knock and the changes in accelerometer and gyroscope values. Knocker employs a machine learning classifier to identify an object from the knock responses. A user study was conducted to evaluate the feasibility of Knocker with 14 objects in both quiet and noisy environments. The result shows that Knocker identifies objects with up to 99.7% accuracy.

actuating a monitor
Actuating a Monitor for Posture Changes

Wednesday 10:20-11:00 / 15:20-16:00 | LBW606 | Room 220BC

  • Joongi Shin, Department of Industrial Design, KAIST
  • Woohyeok Choi
  • Uichin Lee, Graduate School of Knowledge Service Engineering, KAIST
  • Daniel Saakes, Department of Industrial Design, KAIST

The position and orientation of a monitor affects users’ behavior at their desk. In this study, we explored and designed six types of interactions between an actuated monitor and a user to induce posture changes. We built a virtual monitor that simulates the motions of an actuated monitor and slowly moved in the opposite direction of unbalanced sitting postures. We conducted an explorative study with eight participants. The study showed participants’ responses and step by step posture changes toward balanced sitting postures. As contribution, we share considerations for designing monitor actuations that induce posture intervention.

thermal interaction
Thermal Interaction with a Voice-based Intelligent Agent

Wednesday 10:20-11:00 / 15:20-16:00 | LBW631 | Room 220BC

  • Seyeong Kim, Department of Industrial Design, KAIST
  • Yea-kyung Row, Department of Industrial Design, KAIST
  • Tek-Jin Nam, Department of Industrial Design, KAIST
Recently, voice-based intelligent agents (VIAs), such as Alexa, Siri, and Bixby, are becoming popular. One of the interaction challenges with VIAs is that it is difficult to deliver rich information, experience, and meaning via voice-only communication channels. We propose interactive thermal augmentation to address VIA’s interaction challenges. We developed a prototype system and conducted a user study to investigate the effects of thermal interaction in a VIA interaction context. The preliminary study results revealed that: 1) the thermal interface helped the participants to understand the information better; 2) the integration of heat and sound sensation provided an immersive and engaging experience; 3) a thermal stimulus worked as an additional feedback channel that supplements the voice interface. We discuss potentials and considerations when adopting thermal interaction to enrich people’s experiences with a VIA.

Video Showcase

rolling graphics
Rolling Graphics: Create Graphics on the Cross Section of a Roll Cake

Wednesday 17:30-18:30 | Video Showcase | Room 517D

  • Joongi Shin, Department of Industrial Design, KAIST
  • Maria Jose, Department of Industrial Design, KAIST
  • Su Ah Han, Department of Industrial Design, KAIST
  • Moojin Joh, Department of Industrial Design, KAIST
  • Daniel Saakes, Department of Industrial Design, KAIST

Let’s take KAIST HCI friends photos together! Take a photo with the HCI KAIST pose as shown in the example and upload it to the Instagram. With hash tags #HCIKAIST@CHI2019 or #KAISTNIGHT.​