Future Interfaces Group: The next phase of computer-human interaction

Engadget
17 Dec 201806:53

Summary

TLDRThe Future Interfaces Group at Carnegie Mellon University develops innovative human-computer interaction technologies, exploring new ways to communicate beyond traditional inputs like keyboards and touchscreens. Their projects include transforming smartwatches into advanced input devices through high-speed accelerometers and leveraging environmental sound recognition for contextual awareness. They also experiment with camera-based systems for real-time monitoring in smart environments. These cutting-edge innovations aim to create more intuitive, assistive technology. However, the team acknowledges the challenges of security, privacy, and practicality in the adoption of these technologies at scale.

Takeaways

  • πŸ“± Over 100 million devices can now distinguish between knuckle and finger touches and detect when they are lifted to the ear.
  • πŸ€– The Future Interfaces Group (FIG) at Carnegie Mellon University, established in 2014, explores new modes of human-computer interaction.
  • πŸ’‘ FIG is sponsored by major tech companies like Google, Intel, and Qualcomm, focusing on speculative and experimental technologies.
  • πŸ–₯️ The lab's vision includes creating intelligent environments where smart devices have contextual awareness and can assist users more naturally.
  • πŸ‘‚ One project, Ubiquoustics, enables devices to listen to ambient sounds, like chopping vegetables or blending, to understand their context.
  • ⌚ Another innovation involves transforming smartwatches into versatile input devices using high-speed accelerometers to detect micro-vibrations.
  • πŸ–οΈ Gesture-based interaction is being explored, such as snapping fingers to control lights or clapping to activate devices.
  • πŸ“Έ The lab also explores turning cameras into sensors, enabling smart environments to recognize objects or people without active human monitoring.
  • πŸš— FIG is testing real-time parking solutions using camera-based technology to reduce congestion and pollution in cities.
  • βš–οΈ FIG balances technological innovation with privacy concerns, recognizing that no system is 100% secure and that trust hinges on the perceived value of new technologies.

Q & A

  • What is the Future Interfaces Group (FIG) and where is it located?

    -The Future Interfaces Group (FIG) is a research lab at Carnegie Mellon University in Pittsburgh, Pennsylvania, that focuses on human-computer interaction. It was founded in 2014 and works on speculative projects to improve communication between humans and machines beyond traditional methods like keyboards and touchscreens.

  • What are some examples of projects developed by FIG?

    -Examples of FIG projects include touchscreens that can detect if you're using a knuckle or finger, and devices that can recognize if you're lifting the phone to your ear. The lab works on ideas that expand how machines interact with humans, using sensors, sound recognition, and other contextual cues.

  • What is the grand vision of the FIG lab?

    -The grand vision of the FIG lab is to create intelligent environments where devices, like smart speakers or watches, are aware of their surroundings and can interact with humans using nonverbal cues, such as gestures, gaze, and sounds, similar to how humans communicate with each other.

  • How does the FIG lab increase implicit input bandwidth in devices?

    -FIG increases implicit input bandwidth by enhancing devices' ability to understand contextual information. For example, they use sound recognition to determine activities in a room, such as distinguishing between chopping vegetables or running a blender, so that devices can better assist users.

  • What is the uBaku Stiix project and how does it work?

    -The uBaku Stiix project uses sound recognition to understand what is happening in an environment. By training computers to recognize distinctive sounds, like chopping vegetables or using a blender, the project explores how devices can use microphones to gather contextual information about their surroundings.

  • How does the lab use smartwatches in its research?

    -The lab experiments with smartwatches by increasing their sensitivity. They overclock the accelerometer in a smartwatch to detect micro-vibrations, which allows the watch to sense subtle interactions, such as finger taps, transforming the watch into an input platform for controlling devices like lights or TVs using gestures.

  • What is Sensors, and how does it utilize camera feeds?

    -Sensors is a startup that uses camera-based technology to turn existing cameras in public places into sensor feeds. This system can recognize actions like counting the number of people on a sofa or identifying objects like laptops or phones, helping automate monitoring in places like libraries, restaurants, or streets.

  • How is FIG’s camera-based technology applied in real-world scenarios?

    -FIG’s camera-based technology is applied in real-world scenarios like real-time parking management, where cameras count available parking spaces and guide drivers to open spots, helping reduce congestion and pollution. This technology uses existing infrastructure to provide practical solutions.

  • What challenges do technologies developed by FIG face when moving from research to commercialization?

    -Technologies developed by FIG face challenges such as practicality, feasibility, and ensuring security and privacy when transitioning from research to commercialization. These technologies must balance innovation with real-world impact and user concerns, especially in terms of data privacy and usability.

  • How does FIG address privacy and security concerns in its projects?

    -FIG acknowledges that no technology is 100% secure or privacy-preserving. The lab focuses on making technologies that strike the right balance between innovation and privacy, ensuring that users understand and accept the potential trade-offs when adopting new devices with microphones, sensors, or cameras.

Outlines

00:00

πŸ“± The Future of Human-Computer Interaction

This paragraph introduces how modern smartphones have advanced to detect different inputs, such as distinguishing between a knuckle or finger touch and recognizing when the phone is lifted to the ear. These innovations are part of ongoing projects at the Future Interfaces Group Lab at Carnegie Mellon University. Sponsored by tech giants like Google, Intel, and Qualcomm, the lab explores futuristic ways humans can interact with machines, beyond traditional interfaces like keyboards and touchscreens.

05:01

πŸ”¬ Creating the Future Interfaces Lab

The founder of the Future Interfaces Group, who joined Carnegie Mellon University (CMU) five years ago, reflects on setting up the lab. His research focused on using the human body as an interactive computing surface. With the support of students and researchers, the lab continues to push the boundaries of human-computer interaction. The grand vision involves creating intelligent environments where devices like smartphones and smart speakers understand contextual human interactions, similar to how humans use non-verbal cues in communication.

πŸ‘‚ Increasing Implicit Input for Devices

The lab's focus is on increasing devices' ability to understand their environment, often referred to as 'implicit input bandwidth.' One project, called Ubiquitous Acoustics (uBaku Stiix), enables devices to recognize environmental sounds to infer what’s happening, like detecting if someone is chopping vegetables or blending food. Using sensors like microphones already built into devices, they explore ways to collect and process contextual data affordably.

⌚ Transforming Devices with Enhanced Sensors

Smartwatches, which are highly capable computers, are the focus of another lab project. By overclocking the accelerometer in a smartwatch, the lab increases its ability to capture detailed micro-vibrations, enabling the watch to interpret subtle gestures and movements. This could allow users to control home devices through gestures, like snapping to turn on lights or clapping to control a TV. The lab develops hundreds of similar ideas each year, with a few transforming into startups.

πŸ“Έ Turning Cameras into Sensors for Smart Environments

Some of the lab’s projects, such as 'sensors,' leverage existing technology, like cameras, to create smarter environments. They explore using video feeds to automatically analyze environments in real-time, for example, detecting how many people are present in a room or recognizing objects like laptops on a table. This approach could be used in public spaces or homes to offer smarter, real-time monitoring without human intervention.

πŸš— Smart Cities and the Future of Parking

The lab collaborates with cities, such as using camera-based sensors to count cars and help solve real-world problems like parking. By utilizing existing infrastructure like traffic cameras, the system could potentially guide drivers to available parking spots, reducing congestion and air pollution. However, the success of such technologies depends on balancing practicality, feasibility, and cost with long-term value.

πŸ” Ethical Considerations in Technology Adoption

The closing discussion addresses the ethical implications of advancing technologies, such as using surveillance cameras for sensing. While the research offers exciting possibilities, privacy and security concerns are inevitable. The lab believes that no technology can be 100% secure, but by making the right trade-offs and involving users in the design process, people may accept potential privacy risks if they see clear benefits.

Mindmap

Keywords

πŸ’‘Future Interfaces Group

The Future Interfaces Group (FIG) is a research lab at Carnegie Mellon University, focused on developing new forms of human-computer interaction. Founded in 2014, the lab creates speculative and innovative ideas that explore how we communicate with machines, aiming to go beyond traditional input methods like keyboards and touchscreens. The lab is sponsored by major tech companies like Google, Intel, and Qualcomm.

πŸ’‘Human-computer interaction

Human-computer interaction (HCI) refers to how humans communicate and interact with computers and other digital systems. The video emphasizes new ways to enhance this interaction, not just through conventional means like voice commands or touchscreens, but by developing systems that understand contextual and environmental cues, such as sound or gestures, to improve the user experience.

πŸ’‘Implicit input bandwidth

Implicit input bandwidth refers to the ability of a device to gather contextual information from its surroundings to enhance interaction without direct input from the user. This concept is central to the video, where it is discussed in terms of expanding how devices like smartphones or smartwatches interpret environmental cues (e.g., sound or motion) to offer more natural and responsive interactions.

πŸ’‘Contextual understanding

Contextual understanding in the video refers to a device's ability to recognize and interpret its environment, such as sound or motion, to anticipate and respond to user needs. This is key to advancing human-computer interaction beyond current methods. For example, a device that can identify sounds like chopping vegetables or blending in a kitchen could offer relevant assistance or features without being explicitly prompted.

πŸ’‘Ubiquitous computing

Ubiquitous computing, also known as pervasive computing, is the concept of embedding computational capability into everyday objects and environments, making them 'smart.' The video explores how future devices will integrate seamlessly into daily life, collecting information from surroundings and reacting accordingly, such as smartwatches detecting micro-vibrations or sound to perform tasks.

πŸ’‘Sensor fusion

Sensor fusion refers to combining data from multiple sensors to make more accurate assessments of a situation or environment. In the video, this is demonstrated with smartwatches that use various sensors to detect gestures, micro-vibrations, and other contextual cues, improving interaction. The technology is used to enhance how machines interpret complex human behavior and environmental signals.

πŸ’‘Smart environments

Smart environments are physical spaces equipped with devices and sensors that gather and respond to data in real time to improve the user's experience. The Future Interfaces Group explores how such environments can assist users by gathering contextual information, such as using cameras in public spaces to monitor occupancy or integrating gesture-based controls for home devices like lighting and TVs.

πŸ’‘Speculative design

Speculative design involves creating prototypes or concepts that push the boundaries of current technology to explore future possibilities. The Future Interfaces Group engages in speculative design by developing hundreds of ideas each year, some of which become startups, such as the technology behind touchscreen improvements or computer vision systems like Sensors, which aim to enhance smart environments.

πŸ’‘Gesture recognition

Gesture recognition technology allows devices to interpret and respond to human gestures as a form of input. In the video, gesture recognition is exemplified by the use of smartwatches that can detect movements like snapping or clapping to control lights or other home devices. This is part of a broader effort to expand how users interact with machines in more intuitive ways.

πŸ’‘Computer vision

Computer vision is a field of artificial intelligence that enables machines to interpret and process visual data from the world. The video discusses how the Sensors startup, developed by the Future Interfaces Group, uses cameras to detect and count objects or people in public spaces. This technology can be applied to real-time tasks like monitoring parking availability or determining the occupancy of a room.

Highlights

Over a hundred million phones can now detect if you're using your knuckle or finger to touch the screen, as well as whether you're lifting the device to your ear.

The Future Interfaces Group Lab at Carnegie Mellon University in Pittsburgh, Pennsylvania, has been pioneering advancements in human-computer interaction since 2014.

The lab is backed by major sponsors like Google, Intel, and Qualcomm and develops hundreds of speculative ideas each year to enhance communication between humans and machines.

The group’s focus is on creating new modes of interaction beyond keyboards, touchscreens, mice, or even voice commands.

Founder of the lab envisions intelligent environments where devices like Google Home or smartwatches can have full contextual awareness, similar to human assistants.

A key area of research is enhancing 'implicit input bandwidth,' enabling devices to gather contextual information about their environment, like sound-based understanding.

The Ubakustix project trains computers to use microphones to understand environmental sounds, such as identifying kitchen activities like chopping vegetables or running a blender.

Research in the lab shows how smartwatches can be transformed into high-precision devices by overclocking their accelerometers, detecting micro-vibrations and enabling gesture-based controls.

Innovative smart gestures, such as snapping fingers or twisting a wrist, allow users to control lights or navigate menus through gestures alone.

Several of the lab’s projects have led to real startups, such as Kiko (touchscreen technology) and Sensors (a computer vision startup).

Sensors uses cameras in public environments, like restaurants or streets, to turn video feeds into sensor data, identifying objects, people, or even parking spaces.

Real-time parking systems are being piloted using existing city cameras, aiming to direct drivers to available spaces, reducing congestion and pollution.

The lab’s approach emphasizes practical and scalable solutions that can transition from research into real-world applications.

Challenges remain with security and privacy as cameras and microphones gain contextual awareness; however, the lab prioritizes designing technologies that balance benefits and privacy concerns.

The lab actively engages users in testing and feedback, ensuring that the value proposition of new technologies is clear, increasing the likelihood of adoption.

Transcripts

play00:00

there are over a hundred million phones

play00:02

that can tell if you're using your

play00:04

knuckle or finger to touch the screen as

play00:05

well as whether you're lifting the

play00:07

device to your ear there are examples of

play00:09

projects that started here at the future

play00:10

interfaces group lab at Carnegie Mellon

play00:12

University in Pittsburgh Pennsylvania

play00:14

the lab has been around since 2014 and

play00:17

counts Google Intel and Qualcomm among

play00:19

its sponsors every year they develop

play00:22

hundreds of speculative ideas all to do

play00:24

with how we communicate with machines

play00:25

beyond the mode of keyboard touchscreen

play00:27

mouse or even voice we came here to see

play00:30

some of their latest ideas and what they

play00:31

might have to say about the future of

play00:33

human-computer interaction

play00:36

[Music]

play00:39

I came to CMU as faculty about five

play00:43

years ago and founded the future

play00:45

interfaces group and we set up shop in

play00:47

this building a little bit off campus so

play00:48

we had lots of space to build crazy

play00:50

prototypes and put things together I

play00:51

wanted to build on my PhD thesis

play00:53

research which was looking at how to use

play00:55

the human body as a like an interactive

play00:57

computing surface and so we extended a

play00:59

lot of those themes and obviously I took

play01:01

on master students and undergraduates

play01:03

and PhD student researchers to extend

play01:05

that vision and help them sort of

play01:07

explore new frontiers in human-computer

play01:08

interaction a grand vision that the

play01:11

whole lab has has bought into is the

play01:12

notion of having intelligent

play01:14

environments you know right now if you

play01:16

have a Google home or an Alexa or one of

play01:18

these smart assistants sitting on your

play01:20

kitchen countertop it's totally

play01:21

oblivious to what's going on around and

play01:23

that's true of your Smart Watch and

play01:24

that's true of your smartphone they want

play01:26

to make them truly assistive and they

play01:28

can fill in all of a context like a good

play01:29

humanist system would be able to do they

play01:31

need to have that awareness like when

play01:35

humans communicate there's these verbal

play01:37

and nonverbal cues that we use like you

play01:39

know gaze and gesture and all these

play01:41

different things to enrich that

play01:42

conversation in human-computer

play01:44

interaction you don't really have that a

play01:46

lot of my current work is all about

play01:48

increasing implicit input bandwidth so

play01:50

what I mean by that is increasing the

play01:52

ability for these devices to have

play01:54

contextual understanding about what's

play01:56

happening around them so a good example

play01:58

of this is sound we have this project

play02:01

called u baku stiix that listens to the

play02:03

environment and tries to guess what's

play02:04

going on

play02:11

if I teleported you into my kitchen but

play02:14

I blindfolded you and I started blending

play02:16

something or chopping vegetables you'd

play02:18

be able to know that Chris's chopping

play02:20

vegetables are running the blender or

play02:21

turning on a stove or running the

play02:22

microwave and so we just ask ourselves

play02:23

well if sound is is so distinctive that

play02:25

humans can do it

play02:26

can we not train computers to use the

play02:28

microphones that almost all of them have

play02:30

you know whether it's a smart speaker or

play02:31

even a smart watch you have all these

play02:33

sensors that other people have created

play02:34

that are at your disposal and the

play02:35

question is how do you put them together

play02:36

to do this in a low-cost impractical way

play02:39

I think of smart watches it's like

play02:45

really capable computers it's they

play02:47

should be able to almost like transform

play02:49

the hand until like an arm to point oh I

play02:51

supposed to just extensions of the phone

play02:53

typically accelerometers in the watch

play02:54

are around 100 Hertz so here what we did

play02:57

is we overclocked the accelerometer on

play03:00

the watch so it becomes high-speed so

play03:03

you can see here when I interact with

play03:06

this coffee grinder

play03:07

you can actually see the micro

play03:09

vibrations that are propagating from my

play03:11

hand to the watch you can't see that

play03:14

effect from the 100 Hertz accelerometer

play03:16

because it's to course the vibrations

play03:18

when I tap here and when I tap here are

play03:20

actually quite different so I can

play03:22

basically transform this area around the

play03:24

watch into like an input platform you

play03:26

can also combine this with the motion

play03:27

data so when I like snap I can basically

play03:31

either snap to turn on the lights then I

play03:33

can do this gesture and then twist to

play03:35

you know adjust the lighting in that

play03:37

house and then I could do like a clap

play03:40

gesture to turn on the TV and do like

play03:41

these types of gestures to navigate up

play03:43

and down these are only a few of the

play03:46

hundreds of ideas that pop up at the lab

play03:48

every year a couple of them turn into

play03:50

real startups one of them is Kiko which

play03:53

is behind the touchscreen technology we

play03:54

saw at the beginning another newer one

play03:57

is a computer vision startup called

play03:58

sensors one of the technologies that we

play04:01

did for smart environments was a camera

play04:02

based approach we noticed that in a lot

play04:04

of settings like in you know restaurants

play04:06

or libraries or airports or even out on

play04:08

the street there's a lot of cameras

play04:09

these days and what we asked know could

play04:11

we turn these into a sensor feed so you

play04:13

don't have to have someone in a back

play04:14

room is looking at 50 screens but can we

play04:16

smell action-wise and that's what we did

play04:18

in sensors here's an example of how we

play04:21

can go and make a question so we have a

play04:23

camera

play04:23

actually right above us you can see us

play04:24

here right now this updates no once

play04:27

every 30 seconds or once every minute so

play04:29

the first thing you do is we select a

play04:31

region of interest in this case these

play04:34

two sofas it's going to be a let's say

play04:37

how many and now Lily's gonna ask how

play04:39

many people are here that's it and right

play04:44

now it's saying there's three people

play04:45

here and we're not just limited to be

play04:48

sofas I could ask is there a laptop or

play04:50

phones on this table is there food on

play04:53

this table anything you can ask you can

play04:55

do it so like I think though the model

play04:57

of the company kind of is if you can see

play04:58

it we can sense it so we're doing a

play05:00

real-time parking pilot right now with

play05:02

the city and and what we're using is

play05:04

existing cameras along a stretch to

play05:06

basically count cars so we can use that

play05:08

as a real-time model potentially like

play05:10

real-time parking but also does help

play05:12

people find parking spots if you can

play05:13

direct them to adjacent parking give me

play05:16

much more efficient and reduce

play05:17

congestion and air pollution and so on

play05:18

deploying that sort of technology city

play05:20

scale requires a huge capital investment

play05:22

at the end as a number doesn't matter if

play05:24

it's produced by a video camera or by

play05:26

physical sensors in the pavement so in

play05:28

order for technologies to be adopted

play05:29

downstream passed the research phase

play05:31

into the engineering and

play05:32

commercialization phase is they have to

play05:34

be practical feasibility is obviously

play05:37

critical we like to tackle problems that

play05:39

we know we can make progress on and then

play05:41

will we balance that with its impact on

play05:43

value the research is undoubtedly

play05:45

exciting but what else happens when a

play05:47

security camera doesn't just see but

play05:49

understands any technology can be

play05:52

misused what happens to an idea after it

play05:54

leaves the lab it is a gray area sort of

play05:57

like cars you're never gonna make the

play05:59

100% safe car but that doesn't mean we

play06:01

should eliminate all cars and we should

play06:04

think about that for technology that no

play06:05

technology is ever gonna be 100% secure

play06:08

or 100 percent privacy preserving it and

play06:10

so we always try to think about how to

play06:12

make these technologies that make the

play06:13

right trade-off because we have a vision

play06:15

of how they're gonna exist we can think

play06:17

about in our mind oh this would be so

play06:18

cool if I had this in my kitchen but

play06:20

we're too close to that domain we think

play06:21

everything is cool all of the

play06:23

technologies that we build are put in

play06:25

front of users and if you can get people

play06:27

to buy into the vision then maybe

play06:29

they'll accept that oh but there's a

play06:30

microphone on this thing that could be

play06:31

listening to me in my kitchen and if you

play06:33

make that value proposition right

play06:34

they'll accept it if you get that value

play06:37

property

play06:37

wrong then it'll dispel Terr and it

play06:39

won't be adopted

Rate This
β˜…
β˜…
β˜…
β˜…
β˜…

5.0 / 5 (0 votes)

Related Tags
Human-ComputerGesture ControlContextual ComputingSmart EnvironmentsCMU LabInnovative TechSmart DevicesAI InterfacesFuture ComputingMachine Learning