The limits of Proxemics

A wave of recent papers draw on Hall’s Proxemics to design ubicomp systems*. The paper by Greenberg et al. “Proxemic Interactions: The New Ubicomp?” even made the cover of Interactions Magazine a few months ago, that’s quite a statement.

The ideas of proxemics are attractive, as they seem easy to operationalize**. But proxemic systems, like many ubicomp ones, rely heavily on sensing and this will always be tricky to manage. Just to give a practical example: the new macbooks, adjust the screen luminosity to the presence/closeness of a user. They measure it through variations from the light sensor placed close to the camera. I find it super frustrating to have your screen luminosity drop out of nowhere, and looking at the forum threads on the topic, I’m not the only one annoyed.

Since proxemic systems won’t work perfectly in most real-life situations, they will end up being complemented with micro-mobility management, reasoning on traces of user activity, general heuristics and fine tuning. And boom, we’re back to the sensing issues [1] described 10 years ago by Belloti et al. (2002):

  • “When I address a system, how does it know I am addressing it?
  • When I ask a system to do something how do I know it is attending?
  • When I issue a command, how does the system know what it relates to?
  • How do I know the system understands my command and is correctly executing my intended action?
  • How do I recover from mistakes?”

Notes:

* Back at CHI in 2009 (not sure of the year), one of the Keynote actually presented proxemics as an inspiration for designing interaction.

** I’m not sure it’s the right operationalization though. Proxemic circles are super coarse. This might be ok for spatial interaction between people, but it sucks for interaction between people and objects, where we know that things get subtle and will be really hard to pick with sensors (see for instance Luff and Heath on micro-mobility [2]). And yes I’m aware of Nick’s great work tackling the problem.

[1] Victoria Bellotti, Maribeth Back, W. Keith Edwards, Rebecca E. Grinter, Austin Henderson, and Cristina Lopes. 2002. Making sense of sensing systems: five questions for designers and researchers. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (CHI ’02). ACM, New York, NY, USA, 415-422.

[2] Paul Luff and Christian Heath. 1998. Mobility in collaboration. In Proceedings of the 1998 ACM conference on Computer supported cooperative work (CSCW ’98). ACM, New York, NY, USA, 305-314.

Building incentives for replication

David Karger recently posted some thoughts on replication at CHI and WWW, suggesting a mandatory replication paragraph at the end of every paper. It’s a nice idea and while it’s kind of implied in the post, a special section in the review form of reviewers and ACs might be a more manageable thing to push forward. This could offer a smoother transition and progressively help articles supporting replication stand out in PC meetings. It might also be easier to push something in the review system than asking for a specific paragraph in papers. Then the different subcommittees could decide to deal with this differently depending on their research practices.

A review section dedicated to replication could help the authors and the reviewers (who are likely authors) build a discourse on replication and data sharing (beyond people going to workshops and SIGs on replication). Hopefully over time, this would also lead to higher standards on replication and data-set quality.

I submitted a note to CHI last year with all the data and scripts (ended up at ITS). The didn’t receive any comment on the data (even critical). Sadly I ended up improving the paper but not the data which I consider to be at least half of the contribution.

Just for reference here is the text above the review field on PCS (for CHI’13), not a lot on replication:

Write your review of the paper here. Please address each of the following issues:

  • Significance of the paper’s contribution to HCI and the benefit that others can gain from the contribution: why do the contribution and benefit matter?
  • Originality of the work: what new ideas or approaches are introduced? We want to emphasize that an acceptable paper must make a clear contribution to Human-Computer Interaction;
  • Validity of the work presented: how confidently can researchers and practitioners use the results?
  • Presentation clarity;
  • Relevant previous work: is prior work adequately reviewed?

If you have concerns about the methodological or statistical approaches taken by the authors, or its level of advancement over prior work, please cite a source for your objection (e.g., a definitive paper, a set of professional guidelines or a standard textbook). This is needed to help authors improve their submissions and to enable the Associate Chair to evaluate potentially conflicting reviews.

Please consider making any other recommendations that you think might be of use to the author(s).