See only:


[C17] Investigating the Necessity of Delay in Marking Menu Invocation.
Henderson J., et al.Malacria S., Nancel M., & Lank E. (2020). ACM CHI '20.
Henderson J., Malacria S., Nancel M., & Lank E.

Delayed display of menu items is a core design component of marking menus, arguably to prevent visual distraction and foster the use of mark mode. We investigate these assumptions, by contrasting the original marking menu design with immediately-displayed marking menus. In three controlled experiments, we fail to reveal obvious and systematic performance or usability advantages to using delay and mark mode. Only in very constrained settings—after significant training and only two items to learn—did traditional marking menus show a time improvement of about 260 ms. Otherwise, we found an overall decrease in performance with delay, whether participants exhibited practiced or unpracticed behaviour. Our final study failed to demonstrate that an immediately-displayed menu interface is more visually disrupting than a delayed menu. These findings inform the costs and benefits of incorporating delay in marking menus, and motivate guidelines for situations in which its use is desirable.

[C16] AutoGain: Gain Function Adaptation with Submovement Efficiency Optimization.
Lee B., et al.Nancel M., Kim S., & Oulasvirta A. (2020). ACM CHI '20.
Lee B., Nancel M., Kim S., & Oulasvirta A.

A well-designed control-to-display gain function can improve pointing performance with indirect pointing devices like trackpads. However, the design of gain functions is challenging and mostly based on trial and error. AutoGain is an unobtrusive method to individualize a gain function for indirect pointing devices in contexts where cursor trajectories can be tracked. It gradually improves pointing efficiency by using a novel submovement-level tracking+optimization technique that minimizes aiming error (undershooting/overshooting) for each submovement. We first show that AutoGain can produce, from scratch, gain functions with performance comparable to commercial designs, in less than a half-hour of active use. Second, we demonstrate AutoGain’s applicability to emerging input devices (here, a Leap Motion controller) with no reference gain functions. Third, a one-month longitudinal study of normal computer use with AutoGain showed performance improvements from participants’ default functions.


[C15] A Comparative Study of Pointing Techniques for Eyewear Using a Simulated Pedestrian Environment.
Roy Q., et al.Zakaria C., Perrault S., Nancel M., Kim W., Misra A., & Cockburn A. (2019). In Interact '19. IFIP - pdf - BibTEX
Roy Q., Zakaria C., Perrault S., Nancel M., Kim W., Misra A., & Cockburn A.

Eyewear displays allow users to interact with virtual content displayed over real-world vision, in active situations like standing and walking. Pointing techniques for eyewear displays have been proposed, but their social acceptability, efficiency, and situation awareness remain to be assessed. Using a novel street-walking simulator, we conducted an empirical study of target acquisition while standing and walking under different levels of street crowdedness. We evaluated three phone-based eyewear pointing techniques: indirect touch on a touchscreen, and two in-air techniques using relative device rotations around forward and a downward axes. Direct touch on a phone, without eyewear, was used as a control condition. Results showed that indirect touch was the most efficient and socially acceptable technique, and that in-air pointing was inefficient when walking. Interestingly, the eyewear displays did not improve situation awareness compared to the control condition. We discuss implications for eyewear interaction design.

[NS1] Interfaces utilisateurs – Dispositions de clavier bureautique français (NF Z 71-300).
In Norme AFNOR '19 AFNOR - BibTEX
Roy Q., Zakaria C., Perrault S., Nancel M., Kim W., Misra A., & Cockburn A.

Le présent document définit la disposition du clavier informatique à 105 touches dans sa version bureautique et 72 touches dans sa version compacte. Il est particulièrement adapté à la saisie de la langue française dans un contexte multilingue où le recours à d’autres langues latines est fréquent. [...] Conçu pour les utilisateurs français, il pourrait néanmoins présenter un intérêt pour d’autres pays francophones. [...] Les travaux de normalisation à l’origine de ce document font suite à l’édition 2015 du rapport au Parlement sur l’emploi de la langue française (Délégation générale à la langue française et aux langues de France, ministère de la Culture et de la Communication) puis d’une publication intitulée « Vers une norme française pour les claviers informatiques » qui font état de la difficulté de saisir certains caractères courants en français tels que les capitales accentuées ou la ligature « œ ». [...] Les principaux objectifs qui ont concouru à la rédaction de ce document sont les suivants :
– homogénéiser le parc de claviers informatiques en France et notamment réduire les disparités dans la disposition des caractères entre différents fabricants de matériels et de systèmes d’exploitation ;
– améliorer l’ergonomie du clavier pour la saisie du français tout en s’inscrivant dans la continuité des dispositifs existants (dispositions de clavier dites « AZERTY », non normalisées, établies par l’usage), de façon à ne pas provoquer de difficultés d’appropriation de la nouvelle disposition par les utilisateurs ;
– permettre la saisie de l’ensemble des caractères des langues régionales de France dont la liste3 figure sur le site internet de la Délégation générale à la langue française et aux langues de France3 ;
– permettre la saisie de l’ensemble des caractères des langues à alphabet latin présentes sur le continent européen, avec comme priorité les caractères usuels des grandes langues de communication en Europe telles que l’allemand, l’espagnol ou le portugais ;
– rendre plus accessibles de nouveaux jeux de caractères et symboles pouvant être utiles lors de la rédaction de documents spécifiques ou techniques (lettres de l’alphabet grec, symboles mathématiques par exemple).


[C14*] Next-Point Prediction for Direct Touch Using Finite-Time Derivative Estimation.
Nancel M., et al.Aranovskiy S., Ushirobira R., Efimov D., Poulmane S., Roussel N., & Casiez G. (2018). In ACM UIST '18. ACM - pdf - BibTEX
Nancel M., Aranovskiy S., Ushirobira R., Efimov D., Poulmane S., Roussel N., & Casiez G.

End-to-end latency in interactive systems is detrimental to performance and usability, and comes from a combination of hardware and software delays. While these delays are steadily addressed by hardware and software improvements, it is at a decelerating pace. In parallel, short-term input prediction has shown promising results in recent years, in both research and industry, as an addition to these efforts. We describe a new prediction algorithm for direct touch devices based on (i) a state-of-the-art finite-time derivative estimator, (ii) a smoothing mechanism based on input speed, and (iii) a post-filtering of the prediction in two steps. Using both a pre- existing dataset of touch input as benchmark, and subjective data from a new user study, we show that this new predictor outperforms the predictors currently available in the literature and industry, based on metrics that model user-defined negative side-effects caused by input prediction. In particular, we show that our predictor can predict up to 2 or 3 times further than existing techniques with minimal negative side-effects.

[TR5] Historique et méthodologie de la nouvelle disposition de clavier AZERTY.
Nancel M., et al. (2018). Inria TechReport Inria - pdf - BibTEX
Nancel M.,

Document de travail utilisé dans la rédaction de l’annexe H “Historique et méthodologie” de la norme AFNOR Z 71-300 : “Dispositions de clavier bureautique français”.

[TR4] Élaboration de la disposition AZERTY modernisée.
Feit A., et al.Nancel M., Weir D., Bailly G., John M., Karrenbauer A., & Oulasvirta A. (2018). Inria TechReport Inria - pdf - BibTEX
Feit A., Nancel M., Weir D., Bailly G., John M., Karrenbauer A., & Oulasvirta A.

Document de travail utilisé dans la rédaction de l'annexe F “Élaboration de la disposition AZERTY modernisée” de la norme AFNOR Z 71-300 : “Dispositions de clavier bureautique français”.

[C13] Introducing Transient Gestures to Improve Pan and Zoom on Touch Surfaces.
Avery J., et al.Malacria S., Nancel M., Casiez G., & Lank E. (2018). In ACM CHI '18. ACM - pdf - video - BibTEX
Avery J., Malacria S., Nancel M., Casiez G., & Lank E.

Despite the ubiquity of touch-based input and the availability of increasingly computationally powerful touchscreen devices, there has been comparatively little work on enhancing basic canonical gestures such as swipe-to-pan and pinch-to-zoom. In this paper, we introduce transient pan and zoom, i.e. pan and zoom manipulation gestures that temporarily alter the view and can be rapidly undone. Leveraging typical touchscreen support for additional contact points, we design our transient gestures such that they co-exist with traditional pan and zoom interaction. We show that our transient pan-and-zoom reduces repetition in multi-level navigation and facilitates rapid movement between document states. We conclude with a discussion of user feedback, and directions for future research.

[C12] Pointing at a Distance with Everyday Smart Devices.
Siddhpuria S., et al.Malacria S., Nancel M., & Lank E. (2018). In ACM CHI '18. ACM - pdf - video - BibTEX
Siddhpuria S., Malacria S., Nancel M., & Lank E.

Large displays are becoming commonplace at work, at home, or in public areas. However, interaction at a distance -- anything greater than arms-length -- remains cumbersome, restricts simultaneous use, and requires specific hardware augmentations of the display: touch layers, cameras, or dedicated input devices. Yet a rapidly increasing number of people carry smartphones and smartwatches, devices with rich input capabilities that can easily be used as input devices to control interactive systems. We contribute (1) the results of a survey on possession and use of smart devices, and (2) the results of a controlled experiment comparing seven distal pointing techniques on phone or watch, one- and two-handed, and using different input channels and mappings. Our results favor using a smartphone as a trackpad, but also explore performance tradeoffs that can inform the choice and design of distal pointing techniques for different contexts of use.


[C11] Modeling User Performance on Curved Constrained Paths.
Nancel M., et al.Lank E. (2017). In ACM CHI '17. ACM - pdf - BibTEX
Nancel M., Lank E.

In 1997, Accot and Zhai presented seminal work analyzing the temporal cost and instantaneous speed profiles associated with movement along constrained paths. Their work posited and validated the steering law, which described the relationship between path constraint, path length and the temporal cost of path traversal using a computer input device (e.g. a mouse). In this paper, we argue that the steering law fails to correctly model constrained paths of varying, arbitrary curvature, propose a new form of the law that accommodates these curved paths, and empirically validate our model.


[PP1] AutoGain: Adapting Gain Functions by Optimizing Submovement Efficiency.
Lee B., et al.Nancel M., & Oulasvirta A. (2016). arXiv PrePrint. arXiv - pdf - BibTEX
Lee B., Nancel M., & Oulasvirta A.

A well-designed control-to-display (CD) gain function can improve pointing performance with an indirect pointing device such as a trackpad. However, the design of gain functions has been challenging and mostly based on trial and error. AutoGain is an unobtrusive method to obtain a gain function for an indirect pointing device in contexts where cursor trajectories can be tracked. It gradually improves pointing efficiency by using a novel submovement-level tracking+optimization technique. In a study, we show that AutoGain can produce gain functions with performance comparable to commercial designs in less than a half hour of active use. This is attributable to reductions in aiming error (undershooting/overshooting) for each submovement. Our second study shows that AutoGain can be used to obtain gain functions for emerging input devices (here, a Leap Motion controller) for which no good gain function may exist yet. Finally, we discuss deployment in a real interactive system.

[C10*] Next-Point Prediction Metrics for Perceived Spatial Errors.
Nancel M., et al.Vogel D., De Araujo B., Jota R., & Casiez G. (2016). In ACM UIST '16. ACM - pdf - video - BibTEX
Nancel M., Vogel D., De Araujo B., Jota R., & Casiez G.

Touch screens have a delay between user input and corresponding visual interface feedback, called input “latency” (or “lag”). Visual latency is more noticeable during continuous input actions like dragging, so methods to display feedback based on the most likely path for the next few input points have been described in research papers and patents. Designing these “next-point prediction” methods is challenging, and there have been no standard metrics to compare different approaches. We introduce metrics to quantify the probability of 7 spatial error “side-effects” caused by next-point prediction methods. Types of side-effects are derived using a thematic analysis of comments gathered in a 12 participants study covering drawing, dragging, and panning tasks using 5 state-of-the-art next-point predictors. Using experiment logs of actual and predicted input points, we develop quantitative metrics that correlate positively with the frequency of perceived side-effects. These metrics enable practitioners to compare next-point predictors using only input logs.

[W1] Hands Up: Who Knows Something About Performance and Ergonomics of Mid-Air Hand Gestures.
Feit A., et al.Nancel M. (2016). ACM CHI Workshop '16. - pdf - BibTEX
Feit A., Nancel M.

Advances in markerless and un-instrumented hand tracking allow us to make full use of the hands' dexterity for interaction with computers. However, the biomechanics of hand movements remain to be thoroughly studied in HCI. The large number of degrees of freedom of the hand (25) presents us with a huge design space of possible gestures, which is hard to fully explore with traditional methods like elicitation studies or design heuristics. We propose an approach to develop a model of fatigue and stress of manual mid-air input, inspired by prior work on the ergonomics of arm movements and on the performance of multi-finger gestures. Along with our vision of the incoming challenges in mid-air interaction, we describe a design framework for mid-air input that, given such models, can be used to automatically evaluate any given gesture set, or propose an optimal gesture vocabulary for a given set of tasks.

[C9] The Performance and Preference of Different Fingers and Chords for Pointing, Dragging, and Object Transformation.
Goguey A., et al.Nancel M., Vogel D., & Casiez G. (2016). In ACM CHI '16. ACM - pdf - video - BibTEX
Goguey A., Nancel M., Vogel D., & Casiez G.

The development of robust methods to identify which finger is causing each touch point, called “finger identification,” will open up a new input space where interaction designers can associate system actions to different fingers. However, relatively little is known about the performance of specific fingers as single touch points or when used together in a “chord.” We present empirical results for accuracy, throughput, and subjective preference gathered in five experiments with 48 participants exploring all 10 fingers and 7 two-finger chords. Based on these results, we develop design guidelines for reasonable target sizes for specific fingers and two-finger chords, and a relative ranking of the suitability of fingers and two-finger chords for common multi-touch tasks. Our work contributes new knowledge regarding specific finger and chord performance and can inform the design of future interaction techniques and interfaces utilizing finger identification.


[C8*] Gunslinger: Subtle Arms-down Mid-air Interaction.
Liu M., et al.Nancel M., & Vogel D. (2015). In ACM UIST '15. ACM - pdf - video - BibTEX
Liu M., Nancel M., & Vogel D.

We describe Gunslinger, a mid-air interaction technique using barehand postures and gestures. Unlike past work, we explore a relaxed arms-down position with both hands interacting at the sides of the body. It features novel ‘hand-cursor’ feedback to communicate recognized hand posture, command mode and tracking quality; and a simple, but flexible hand posture recognizer. Although Gunslinger is suitable for many usage contexts, we focus on integrating mid-air gestures with large display touch input. We show how the Gunslinger form factor enables an interaction language that is equivalent, coherent, and compatible with large display touch input. A four-part study evaluates Midas Touch, posture recognition feedback, fundamental pointing and clicking, and general usability.

[J2] Mid-air Pointing on Ultra-Walls.
Nancel M., et al.Pietriga E., Chapuis O., & Beaudouin-Lafon M. (2015). In ACM ToCHI '15 ACM - pdf - BibTEX
Nancel M., Pietriga E., Chapuis O., & Beaudouin-Lafon M.

Ultra-high-resolution wall-sized displays (“ultra-walls”) are effective for presenting large datasets, but their size and resolution make traditional pointing techniques inadequate for precision pointing. We study mid-air pointing techniques that can be combined with other, domain-specific interactions. We first explore the limits of existing single-mode remote pointing techniques and demonstrate theoretically that they do not support high-precision pointing on ultra-walls. We then explore solutions to improve mid-air pointing efficiency: a tunable acceleration function and a framework for dual-precision techniques, both with precise tuning guidelines. We designed novel pointing techniques following these guidelines, several of which outperform existing techniques in controlled experiments that involve pointing difficulties never tested prior to this work. We discuss the strengths and weaknesses of our techniques to help interaction designers choose the best technique according to the task and equipment at hand. Finally, we discuss the cognitive mechanisms that affect pointing performance with these techniques.

[C7*] Clutching Is Not (Necessarily) the Enemy.
Nancel M., et al.Vogel D., & Lank E. (2015). In ACM CHI '15. ACM - pdf - video - BibTEX
Nancel M., Vogel D., & Lank E.

Clutching is usually assumed to be triggered by a lack of physical space and detrimental to pointing performance. We conduct a controlled experiment using a laptop trackpad where the effect of clutching on pointing performance is dissociated from the effects of control-to-display transfer functions. Participants performed a series of target acquisition tasks using typical cursor acceleration functions with and without clutching. All pointing tasks were feasible without clutching, but clutch-less movements were harder to perform, caused more errors, required more preparation time, and were not faster than clutch-enabled movements.

[C6] Myopoint: Pointing and Clicking Using Forearm Mounted EMG and Inertial Motion Sensors.
Haque F., et al.Nancel M., & Vogel D. (2015). In ACM CHI '15. ACM - pdf - video - BibTEX
Haque F., Nancel M., & Vogel D.

We describe a mid-air, barehand pointing and clicking interaction technique using electromyographic (EMG) and inertial measurement unit (IMU) input from a consumer armband device. The technique uses enhanced pointer feedback to convey state, a custom pointer acceleration function tuned for angular inertial motion, and correction and filtering techniques to minimize side-effects when combining EMG and IMU input. By replicating a previous large display study using a motion capture pointing technique, we show the EMG and IMU technique is only 430 to 790 ms slower and has acceptable error rates for targets greater than 48 mm. Our work demonstrates that consumer-level EMG and IMU sensing is practical for distant pointing and clicking on large displays.


[C5*] Causality – A Conceptual Model of Interaction History.
Nancel M., et al.Cockburn A. (2014). In ACM CHI '14. ACM - pdf - video - BibTEX
Nancel M., Cockburn A.

Simple history systems such as Undo and Redo permit retrieval of earlier or later interaction states, but advanced systems allow powerful capabilities to reuse or reapply combinations of commands, states, or data across interaction contexts. Whether simple or powerful, designing interaction history mechanisms is challenging. We begin by reviewing existing history systems and models, observing a lack of tools to assist designers and researchers in specifying, contemplating, combining, and communicating the behaviour of history systems. To resolve this problem, we present CAUSALITY, a conceptual model of interaction history that clarifies the possibilities for temporal interactions. The model includes components for the work artifact (such as the text and formatting of a Word document), the system context (such as the settings and parameters of the user interface), the linear timeline (the commands executed in real time), and the branching chronology (a structure of executed commands and their impact on the artifact and/or context, which may be navigable by the user). We then describe and exemplify how this model can be used to encapsulate existing user interfaces and reveal limitations in their behaviour, and we also show in a conceptual evaluation how the model stimulates the design of new and innovative opportunities for interacting in time.


[C4] Body-centric Design Space for Multi-surface Interaction.
Wagner J., et al.Nancel M., Gustafson S., Huot S., & Mackay W. (2013). In ACM CHI '13. ACM - pdf - video - BibTEX
Wagner J., Nancel M., Gustafson S., Huot S., & Mackay W.

We introduce BodyScape, a body-centric design space for both analyzing existing multi-surface interaction techniques and suggesting new ones. We examine the relationship between users and their environment, specifically how different body parts enhance or restrict movement in particular interaction techniques. We illustrate the use of BodyScape by comparing two free-hand techniques, on-body touch and mid- air pointing, separately and in combination. We found that touching the torso is faster than touching the lower legs, since it affects the user’s balance; individual techniques outperform compound ones; and touching the dominant arm is slower than other body parts because the user must compensate for the applied force. The latter is surprising, given that most recent on-body touch techniques focus on touching the dominant arm.

[C3*] High-Precision Pointing on Large Wall Displays using Small Handheld Devices.
Nancel M., et al.Chapuis O., Pietriga E., Yang X., Irani P., & Beaudouin-Lafon M. (2013). In ACM CHI '13. ACM - pdf - video - BibTEX
Nancel M., Chapuis O., Pietriga E., Yang X., Irani P., & Beaudouin-Lafon M.

Rich interaction with high-resolution wall displays is not limited to remotely pointing at targets. Other relevant forms of interaction include virtual navigation, text entry, and direct manipulation of control widgets. However, most techniques for remotely acquiring targets with high precision have studied remote pointing in isolation, focusing on pointing efficiency, and ignoring the need to support these other forms of interaction. We investigate high-precision pointing techniques capable of acquiring targets as small as 4 millimeters on a 5.5 meters wide display while leaving up to 93 % of a typical tablet device's screen space available for task-specific widgets. We compare these techniques to state-of-the-art distant pointing techniques and show that two of our techniques, a purely relative one and one that uses head orientation, perform as well or better than the best pointing-only input techniques while using a fraction of the interaction resources.


[D1*] Designing and Combining Interaction Techniques in Large Display Environments.
Nancel M., et al. (2012). PhD Université Paris-Sud - pdf - BibTEX
Nancel M.,

Large display environments (LDEs) are interactive physical workspaces featuring one or more static large displays as well as rich interaction capabilities, and are meant to visualize and manipulate very large datasets. Research about mid-air interactions in such environments has emerged over the past decade, and a number of interaction techniques are now available for most elementary tasks such as pointing, navigating and command selection. However these techniques are often designed and evaluated separately on specific platforms and for specific use-cases or operationalizations, which makes it hard to choose, compare and combine them. In this dissertation I propose a framework and a set of guidelines for analyzing and combining the input and output channels available in LDEs. I analyze the characteristics of LDEs in terms of (1) visual output and how it affects usability and collaboration and (2) input channels and how to combine them in rich sets of mid-air interaction techniques. These analyses lead to four design requirements intended to ensure that a set of interaction techniques can be used (i) at a distance, (ii) together with other interaction techniques and (iii) when collaborating with other users. In accordance with these requirements, I designed and evaluated a set of mid-air interaction techniques for panning and zooming, for invoking commands while pointing and for performing difficult pointing tasks with lim- ited input requirements. For the latter I also developed two methods, one for calibrating high-precision techniques with two levels of precision and one for tuning velocity-based transfer functions. Finally, I introduce two higher-level design considerations for combining interaction techniques in input-constrained environments. Designers should take into account (1) the trade-off between minimizing limb usage and performing actions in parallel that affects overall performance, and (2) the decision and adaptation costs incurred by changing the resolution function of a pointing technique during a pointing task.

[J1] Multisurface Interaction in the WILD Room.
Beaudouin-Lafon M., et al.Chapuis O., Eagan J., Gjerlufsen T., Huot S., Klokmose C., Mackay W., Nancel M., Pietriga E., Pillias C., Primet R., & Wagner J. (2012). In IEEE Comp. IEEE - pdf - BibTEX
Beaudouin-Lafon M., Chapuis O., Eagan J., Gjerlufsen T., Huot S., Klokmose C., Mackay W., Nancel M., Pietriga E., Pillias C., Primet R., & Wagner J.

The WILD room (wall-sized interaction with large datasets) serves as a testbed for exploring the next generation of interactive systems by distributing interaction across diverse computing devices, enabling multiple users to easily and seamlessly create, share, and manipulate digital content.


[C2] Rapid Development of User Interfaces on Cluster-Driven Wall Displays with jBricks.
Pietriga E., et al.Huot S., Nancel M., & Primet R. (2011). In ACM EICS '11. ACM - pdf - BibTEX
Pietriga E., Huot S., Nancel M., & Primet R.

Research on cluster-driven wall displays has mostly focused on techniques for parallel rendering of complex 3D models. There has been comparatively little research effort dedicated to other types of graphics and to the software engineering issues that arise when prototyping novel interaction techniques or developing full-featured applications for such displays. We present jBricks, a Java toolkit that integrates a high-quality 2D graphics rendering engine and a versatile input configuration module into a coherent framework, enabling the exploratory prototyping of interaction techniques and rapid development of post-WIMP applications running on cluster-driven interactive visualization platforms.

[TR3] Precision Pointing for Ultra-High-Resolution Wall Displays.
Nancel M., et al.Pietriga E., & Beaudouin-Lafon M. (2011). Inria TechReport Inria - pdf - BibTEX
Nancel M., Pietriga E., & Beaudouin-Lafon M.

Ultra-high-resolution wall displays have proven useful for displaying large quantities of information, but lack appropriate interaction techniques to manipulate the data efficiently. We explore the limits of existing modeless remote pointing techniques, originally designed for lower resolution displays, and show that they do not support high-precision pointing on such walls. We then consider techniques that combine a coarse positioning mode to approach the target's area with a precise pointing mode for acquiring the target. We compare both new and existing techniques through a controlled experiment, and find that techniques combining ray casting with relative positioning or angular movements enable the selection of targets as small as 4 millimeters while standing 2 meters away from the display.

[C1*] Mid-air Pan-and-Zoom on Wall-sized Displays.
Nancel M., et al.Wagner J., Pietriga E., Chapuis O., & Mackay W. (2011). In ACM CHI '11. ACM - pdf - video - BibTEX
Nancel M., Wagner J., Pietriga E., Chapuis O., & Mackay W.

Very-high-resolution wall-sized displays offer new opportunities for interacting with large data sets. While pointing on this type of display has been studied extensively, higher-level, more complex tasks such as pan-zoom navigation have received little attention. It thus remains unclear which techniques are best suited to perform multiscale navigation in these environments. Building upon empirical data gathered from studies of pan-and-zoom on desktop computers and studies of remote pointing, we identified three key factors for the design of mid-air pan-and-zoom techniques: uni- vs. bimanual interaction, linear vs. circular movements, and level of guidance to accomplish the gestures in mid-air. After an extensive phase of iterative design and pilot testing, we ran a controlled experiment aimed at better understanding the influence of these factors on task performance. Significant effects were obtained for all three factors: bimanual interaction, linear gestures and a high level of guidance resulted in significantly improved performance. Moreover, the interaction effects among some of the dimensions suggest possible combinations for more complex, real-world tasks.


[TR2] Push Menu: Extending Marking Menus for Pressure-Enabled Input Devices.
Huot S., et al.Nancel M., & Beaudouin-Lafon M. (2010). Inria TechReport Inria - pdf - BibTEX
Huot S., Nancel M., & Beaudouin-Lafon M.

Several approaches have been proposed to increase the breadth of standard Marking Menus over the 8 item limit, most of which have focused on the use of the standard 2D input space (x-y). We present Push Menu, an extension of Marking Menu that takes advantage of pressure input as a third input dimension to increase menu breadth. We present the results of a preliminary experiment that validates our design and shows that Push Menu users who are neither familiar with pen-based interfaces nor continuous pressure control can handle up to 20 items reliably. We also discuss the implications of these results for using Push Menu in user interfaces and for improving its design.

[P1] 131 millions de pixels qui font le mur.
Beaudouin-Lafon M., et al.Pietriga E., Mackay W., Huot S., Nancel M., Pillias C., & Primet R. (2010). In Plein Sud '10. Université Paris-Sud - pdf
Beaudouin-Lafon M., Pietriga E., Mackay W., Huot S., Nancel M., Pillias C., & Primet R.

Imaginez un mur d’écrans qui affiche des images en haute définition. Imaginez que par des gestes simples, vous puissiez interagir avec lui… Nous ne sommes pas dans le film «Minority Report », mais face à la concrétisation d’un projet unique d’interaction homme-machine (IHM), la plate-forme Wild, qui permet d’interagir avec des masses de données complexes.


[F1*] Un espace de conception fondé sur une analyse morphologique des techniques de menus.
Nancel M., et al.Huot S., & Beaudouin-Lafon M. (2009). In ACM IHM '09. ACM - pdf - BibTEX
Nancel M., Huot S., & Beaudouin-Lafon M.

Cet article présente un espace de conception basé sur une analyse morphologique des mécanismes de structuration des menus et de sélection des items. Son but est de faciliter l'exploration de nouveaux types de menus afin notamment d'augmenter leur capacité sans détériorer leurs performances. L'article démontre l'aspect génératif de cet espace de conception grâce à quatre nouveaux designs de menus, basés sur des combinaisons de dimensions pas ou peu explorées. Pour deux d'entre eux, des expérimentations contrôlées montrent qu'ils offrent des performances comparables aux menus de la littérature.


[TR1] Extending Marking Menus With Integral Dimensions: Application to the Dartboard Menu.
Nancel M., et al.Beaudouin-Lafon M. (2008). Inria TechReport - pdf - BibTEX
Nancel M., Beaudouin-Lafon M.

Marking menus have many benefits, including fast selection time, low error rate and fast transition to expert mode, but these are mitigated by a practical limit of 8 items per menu. Adding hierarchical levels increases capacity, but at the expense of longer selection times and higher error rates. In this paper we introduce Extended Marking Menus, a variant of marking menus that increases their width without sacrificing performance. Extended marking menus organize the items in several rings or layers. Selection is achieved by simultaneous control of direction, as in traditional marking menus, and another dimension such as distance, speed or pressure. We examine the design space of these new menus and study the Distance Extended Marking Menu, or Dartboard Menu, in more detail. We report on two experiments, one to calibrate the sizes of the rings, the other showing that it performs faster than the Zone and Flower menus but is less accurate than the Zone menu.

ACM SUI'20 (Spatial User Interaction)
CHCCS GI'20 (Graphics Interface)
ACM CHI'17 – Interaction Techniques
ACM CHI'16 – Interaction Techniques
ACM ITS'14 (Interactive Tabletops and Surfaces)
ACM AUIC'14 (Australasian User Interface Conference)

CONEX-Plus (2019)
Inria Lille – Nord Europe (2018-2020)
FWO Research Foundation (2018)

2 Outstanding Reviews distinction, ACM CHI 2020.

2 Outstanding Reviews distinction, ACM CHI 2019.

Exceptional Reviewer distinction, ACM UIST 2014.

Exceptional Reviewer distinction, ACM UIST 2012.

HCI Journals

ACM ToCHI ACM Transactions on Computer‑Human Interaction ( 2013, 2016, 2018 )
IJHCS International Journal of Human-Computer Studies ( 2013 – 2016 )
BIT Behaviour and Information Technology ( 2018 )

International Conferences in HCI

ACM CHI ACM SIGCHI Conference on Human Factors in Computing Systems ( 2008 – 2020 )
ACM UIST ACM Symposium on User Interface Software and Technology ( 2008, 2012 – 2019 )
ACM CSCW ACM Conference on Computer-Supported Cooperative Work and Social Computing ( 2014 )
Interact International Conference on Human-Computer Interaction ( 2015 )
MobileHCI International Conference on Human-Computer Interaction with Mobile Devices and Services ( 2015 – 2017 )
ACM ISS ACM International Conference on Interactive Surfaces and Spaces ( 2016 – 2019 )
ACM GI Graphics Interface ( 2015, 2017, 2019 )
ACM NordiCHI ACM Nordic Conference on Human-Computer Interaction ( 2012, 2014, 2016, 2018 )
ACM SUI ACM Symposium on Spatial User Interaction ( 2016, 2019 )
ACM AUIC ACM Australasian User Interface Conference ( 2015 )
ACM DIS ACM Conference on Designing Interactive Systems ( 2012, 2014, 2016, 2019 )
ACM ICMI ACM International Conference on Multimodal Interaction ( 2017 )

Research Project Grants

ANR Projet ANR ( 2015 )

Book Chapters

Book Designing with the Mind in Mind (2nd Edition). Jeff Johnson ( 2014 )

French Conferences in HCI

ACM IHM Conférence Francophone sur l'Interaction Homme-Machine ( 2009, 2011, 2013 – 2015, 2018, 2019 )

Non-HCI Journals

IEEE TNSRE IEEE Transactions on Neural Systems and Rehabilitation Engineering ( 2013, 2014 )
Ergonomics The Official Journal of the Institute for Ergonomics and Human Factors ( 2015 )

Non-HCI Conferences

ACM SIGGRAPH ACM special interest group on Computer GRAPHics and Interactive Techniques ( 2017 )
IEEE PacificVis IEEE Pacific Visualization Symposium ( 2012, 2013 )
2020: (Ontario).
2019: (Paris, France).
2018: [C14*] UIST 2018 (Berlin, Germany).
2016: [C10*] UIST 2016 (Tokyo, Japan).
2015: [C8*] UIST 2015 (Charlotte, NC, USA).
2015: [C7*] CHI 2015 (Seoul, South Korea).
2014: [C5*] CHI 2014 (Toronto, Canada).
2013: NUS HCI Lab (National University of Singapore).
[C3*] CHI 2013 (Paris, France).
2012: (Orsay, France).
2011-2012: (Orsay, France).
2011: [C1*] CHI 2011 (Vancouver, Canada).
2009: [F1*] IHM°09 (Grenoble, France).
2018, 2019
Master IVI, Univ. Lille M2
Univ. Lille M1
2011 - 2012
Univ. Paris-Sud XI L1
Polytech Paris-Sud L3
Polytech Paris-Sud L3
Master Informatique - Univ. Paris-Sud XI M1
2010 - 2011
L2 Info - Univ. Paris-Sud XI L2
Polytech Paris-Sud L3
Polytech Paris-Sud M2
2009 - 2010
TER (MIAGE & L3 Info) - Univ. Paris-Sud XI L3
2008 - 2009
Processing & Arduino Mastère Nouveaux Médias - ENSCI M2
TER (MIAGE & L3 Info) - Univ. Paris-Sud XI L3

2015 - 2016 Post Doctoral Fellow User Interfaces Lab, Aalto University
2014 - 2015 Post Doctoral Fellow HCI Lab, University of Waterloo
2013 Post Doctoral Fellow History Management Systems HCI Lab, University of Canterbury
2008 - 2012 Designing and Combining Interaction Techniques in Large Display Environments Université Paris-Sud XI
2007 - 2008 Université Paris-Sud XI
2003 - 2008 IFIPS