Article très intéressant de John Pavius sur MIT Technology Review, à propos de Google Glass :
Don’t worry, though–in a couple years, we’ll be apparently able to use future iterations of Glass much less weirdly. A Redditor discovered some code implying that we’ll be able to snap photos merely by winking. What could be more natural and effortless than that? Designers at Fjord speculate that these kinds of body-based micro-interactions are the future of interface design. “Why swipe your arm when you can just rub your fingers together,” they write. “What could be more natural than staring at something to select it, nodding to approve something?… For privacy, you’ll be able to use imperceptible movements, or even hidden ones such as flicking your tongue across your teeth.”
These designers think that the difference between effortless tongue-flicking and Glass’s crude chin-snapping is simply one of refinement. I’m not so sure. To me they both seem equally alienating–I don’t think we want our bodies to be UIs.
We want our technology to be ready-to-hand: we want to act through it, not on it. And our bodies don’t have to become marionettes to that technology. If anything, it should be the other way around.
Bon sujet pour le bac philo de cette année, vous avez 2 heures…