Inside The Laughing Room
An AI-powered laugh track amuses and unsettles in interactive installation
A work by Jonny Sun in collaboration with Hannah Davis, Christopher Sun, and associate professor Stephanie Frampton, head of the MIT ARTificial Intelligence project
Stills of The Laughing Room," an installation at the Cambridge Public Library evoking a TV sitcom set with an algorithm-controlled laugh track. The installation was livestreamed online and shown on monitors at MIT's Hayden Library. Photos: metaLAB at Harvard
“I wanted the installation to resemble a sitcom set from the 1980s—a private, familial space,” said Sun. “I wanted to explore how AI is changing our conception of private space, with things like the Amazon Echo or Google Home, where you’re aware of this third party listening.”
By Brigham Fay, MIT Libraries
The Laughing Room, an interactive art installation by author and illustrator Jonny Sun in collaboration with artist and developer Hannah Davis, looks like a typical living room — couches, armchairs, coffee table, soft lighting. This cozy scene, however, sits in a glass-enclosed space, flanked by bright lights and a microphone, with a bank of laptops and a video camera positioned across the room. People wander in, take a seat, begin chatting. After a pause in the conversation, a riot of canned laughter rings out, prompting genuine giggles from the group.
Presented at the Cambridge Public Library November 16-18, The Laughing Room was an artificially intelligent room programmed to play an audio laugh track whenever participants said something that its algorithm deemed funny.
Sun, a PhD candidate at MIT, affiliate at the Berkman Klein Center for Internet and Society at Harvard, and creative researcher at the metaLAB at Harvard, created the project to explore the increasingly social and cultural roles of technology in public and private spaces, users’ agency within and dependence on such technology, and the issues of privacy raised by these systems.
The installations were presented as part of ARTificial Intelligence, an ongoing program led by MIT associate professor of literature Stephanie Frampton to foster public dialogue about the emerging ethical and social implications of artificial intelligence (AI) through art and design.
The Laughing Room team, from left: Stephanie Frampton, MIT associate professor of literature; Christopher Sun, project manager and collaborator; Sarah Newman, creative researcher at metaLAB; Nikhil Dharmaraj, research assistant; Hannah Davis, artist, programmer, and project collaborator; Matthew Battles, associate director at metaLAB; and Jonny Sun, artist, MIT doctoral student. Photo: Stephanie Mitchell/Harvard Gazette.
“Cambridge is the birthplace of artificial intelligence, and this installation gives us an opportunity to think about the new roles that AI is playing in our lives every day. The installations were presented as part of ARTificial Intelligence, an ongoing MIT program that fosters public dialogue about the emerging ethical and social implications of artificial intelligence."
— Stephanie Frampton, associate professor of literature at MIT
Setting the Scene
“Cambridge is the birthplace of artificial intelligence, and this installation gives us an opportunity to think about the new roles that AI is playing in our lives every day,” said Frampton. “It was important to us to set the installations in the Cambridge Public Library and MIT Libraries, where they could spark an open conversation at the intersections of art and science.”
“I wanted the installation to resemble a sitcom set from the 1980s — a private, familial space,” said Sun. “I wanted to explore how AI is changing our conception of private space, with things like the Amazon Echo or Google Home, where you’re aware of this third party listening.”
The Control Room, a companion installation located in Hayden Library, displayed a live stream of the action in The Laughing Room, while another monitor showed the algorithm evaluating people’s speech in real time. Live streams were also shared online via YouTube and Periscope.
“It’s an extension of the sitcom metaphor, the idea that people are watching,” said Sun. The artist was interested to see how people would act, knowing they had an audience. Would they perform for the algorithm? Sun likened it to Twitter users trying to craft the perfect tweet so it will go viral.
“Almost all machine learning starts from a dataset,” said Hannah Davis, an artist, musician, and programmer who collaborated with Sun to create the installation’s algorithm. She described the process at an “Artists Talk Back” event held Saturday, November 17, at Hayden Library. The panel discussion included Davis, Sun, Frampton, collaborator Christopher Sun, research assistant Nikhil Dharmaraj, Reinhard Engels, manager of technology and innovation at Cambridge Public Library, Mark Szarko, librarian at MIT Libraries, and Sarah Newman, creative researcher at metaLAB. The panel was moderated by metaLAB founder and director Jeffrey Schnapp.
Davis explained how, to train the algorithm, she scraped stand-up comedy routines from YouTube, selecting performances by women and people of color to avoid programming misogyny and racism into how the AI identified humor. “It determines what is the setup to the joke and what shouldn’t be laughed at, and what is the punchline and what should be laughed at,” said Davis. Depending on how likely something is to be a punchline, the laugh track plays at different intensities.
The Laughing Room algorithm at work; Image courtesy Jeffrey Schnapp/metaLAB
"Davis explained how, to train the algorithm, she scraped stand-up comedy routines from YouTube, selecting performances by women and people of color to avoid programming misogyny and racism into how the AI identified humor."
Fake Laughs, Real Connections
Sun acknowledged that the reactions from The Laughing Room’s participants have been mixed: “Half of the people came out saying ‘that was really fun,’” he said. “The other half said ‘that was really creepy.’”
That was the impression shared by Colin Murphy, a student at Tufts University who heard about the project from following Sun on Twitter: “This idea that you are the spectacle of an art piece, that was really weird.”
“It didn’t seem like it was following any kind of structure,” added Henry Scott, who was visiting from Georgia. “I felt like it wasn’t laughing at jokes, but that it was laughing at us. The AI seems mean.”
While many found the experience of The Laughing Room uncanny, for others it was intimate, joyous, even magical.
“There’s a laughter that comes naturally after the laugh track that was interesting to me, how it can bring out the humanness,” said Newman at the panel discussion. “The work does that more than I expected it to.”
Frampton noted how the installation’s setup also prompted unexpected connections: “It enabled strangers to have conversations with each other that wouldn’t have happened without someone listening.”
Detail, Listening Room poster
"While many found the experience of The Laughing Room uncanny, for others it was intimate, joyous, even magical. 'There’s a laughter that comes naturally after the laugh track that was interesting to me, how it can bring out the humanness,' said Sarah Newman of metaLAB. 'The work does that more than I expected it to.'"
Continuing his sitcom metaphor, Sun described these first installations as a “pilot,” and is looking forward to presenting future versions of The Laughing Room. He and his collaborators will keep tweaking the algorithm, using different data sources, and building on what they’ve learned through these installations. The Laughing Room returns to MIT in the Wiesner Student Art Gallery in May 2019, and the team is planning further events at MIT, Harvard, and Cambridge Public Library throughout the coming year.
“This has been an extraordinary collaboration and shown us how much interest there is in this kind of programming and how much energy can come from using the libraries in new ways,” said Frampton.
The Laughing Room and The Control Room were funded by Harvard's metaLAB, the MIT SHASS DeFlorez Fund for Humor, the MIT Council of the Arts, and MIT Center For Art, Science & Technology (CAST), and presented by the MIT-based ARTificial Intelligence project in partnership with the Cambridge Public Library and the MIT Libraries.