“The Creepy Study” – Surveillance Cameras Analyze Emotions

Activist Post – by Nicholas West

By now, nearly everyone is familiar with just how massive the State surveillance apparatus has become. The question is no longer about whether or not we are under surveillance, but how much further will the intrusion expand.

As facial recognition technology becomes more sophisticated, and the ease of camera and sensor installation makes all barriers to entry a non-issue, we are beginning to see the advent of real-time tracking that is looking for emotion as well as our data.  

facialrecognitionsThis concept recently went to outrageous lengths when it was uncovered that politicians across the world are already using facial recognition embedded in billboards to judge and manipulate the emotions of those who visually respond to the message displayed.  It even has its own branch of study: neuropolitics.

The word “creepy” has become an overused catch-all that actually does a disservice to what is being employed – kind of a nervous laughter reaction to the reality of completely losing one’s privacy and freedom to engage in natural behavior without being tracked and judged for it. Nevertheless, a new program at the University of Iowa actually has been informally – but officially – labeled as “The Creepy Study.” Let it not be said that the people in charge of implementing these technologies don’t know what they are up to.

Tucked into the white ceiling tiles, the ceiling camera he (Bill Hedgcock, Associate Professor of Marketing) had installed at the Pappajohn Business Building at the University of Iowa scans the faces of all who pass under it and instantly calculates their moods — collecting readings for joy, frustration, confusion, fear, anger and sadness.

Hedgcock believes the University of Iowa is the world’s only business school with this kind of real-time software that converts images of people’s faces into readings for different moods. If it works, experts see huge potential for facial encoding in the worlds of marketing, advertising and political campaigns. (emphasis added)

Here is a perfect example of why creepy doesn’t encapsulate what this really is.  What it really is, is using students as guinea pigs in an experiment to eradicate privacy and monetize human beings, while convincing the students that they are positioning themselves to be leaders in a new industry that is built on the manipulation of human behavior. That’s not creepy, that’s a technocratic takeover.

Naturally, there are privacy assurances that the system just spits out raw data and does not assign an actual human name to what is being collected, but I think we are done with believing that as leakafter leak, government and private, has told us that no data is currently safe or private.

For now, the system is claiming no more than a 35% accuracy rate, as most of what is observed is being categorized as “neutral” emotion.

But the final statements from Dan Hill (president of one company working with this tech) and Professor Hedgcock about what is intended, is an open admission that manipulation is the end goal and there is no remorse about it. Emphasis added:

“I think it’s a good application because politics is very emotional,” Hill said. “There are a lot of so-called independents who really tilt one way or the other, but they like the idea that they’re independent.”

Hedgcock says political strategists are already interested in neuromarketing, the marketing field that uses measurements of the human body — brain scans, respiratory rates and changes in the skin — to measure a consumer’s reaction.

But not many campaigns are interested in talking publicly about those ideas, Hedgcock says, because they can be viewed as too Big Brotherly or manipulative.

For now, the professor is putting aside any concerns about the “creepy” factor.

“Every academic I know in this area of neuromarketing would say the last thing we need to worry about is over manipulating,” he said. “The bigger concern is does it even work?”

Main article source:
http://college.usatoday.com/2016/01/04/univ-of-of-iowa-scanner-computes-emotions/

Image Credit

Nicholas West writes for ActivistPost.com

http://www.activistpost.com/2016/01/the-creepy-study-surveillance-cameras-analyze-emotions.html

2 thoughts on ““The Creepy Study” – Surveillance Cameras Analyze Emotions

  1. “Nevertheless, a new program at the University of Iowa actually has been informally – but officially – labeled as “The Creepy Study.”

    Merely calling a spade a spade.

    “But the final statements from Dan Hill (president of one company working with this tech) and Professor Hedgcock about what is intended, is an open admission that manipulation is the end goal and there is no remorse about it.”

    Remorse? From a commie???

    That’ll be the day.

  2. While I find the very effort to develop and publicly use such technology to be damnable, I have strong doubts that it will ever work.

    Most people walk around wearing poker faces and don’t grin, scowl, or otherwise noticeably change their facial expressions when viewing a billboard. And even if they did, who can judge the REASON for the change?

    For example, I might look at a picture of Edward Snowden and scowl slightly, but it wouldn’t be because I didn’t like him; it would be because seeing his photo would remind me of NSA surveillance. Or maybe I’d look at a photo of Clinton, recall a dirty joke about her, and smile a little. Would the system interpret that as a positive reaction?

    Besides, something like this has already been tried. The FAST system that was intended to detect terrorists in airports collected a lot more data than just facial expression, and it was a total flop.

    Anyway, this continual push for more intrusive surveillance ends whenever We The People put a stop to it.

Join the Conversation

Your email address will not be published. Required fields are marked *


*