Tags: , , , , , , , , , , , ,

CBT/ PSA Psychotherapy Research

Greetings CBT Subreddit,

Many, many thanks to those who have responded considering my research on experiences of beneficial altered in Cognitive-Behavioral Therapy and Psychodynamic Psychotherapy! It is greatly appreciated!

I am re-posting here as I am continuing to look for people who would like to be interviewed about their experiences of change in CBT or Psychodynamic Psychotherapy. I am offering a $25.00 Amazon e-gift card to participants.

If interested, please email me at cbell5 @luxsci. net( HIPPA compliant email ), and I can send you the full survey recruitment letter and such studies description. From that point, we can arrange to talk about such studies by phone and determine if you’d like to participate.

All the best,


Read more: www.reddit.com

No tags for this post.

It’s a riot: the stressful AI simulation built to understand your emotions

Inspired by global unrest, Riot utilizes artificial intelligence, film and gaming technologies to help unpick how people react in stressful situations

An immersive film project is attempting to understand how people react in stressful situations by using artificial intelligence( AI ), film and gaming technologies to place participants inside a simulated riot and then detecting their emotions in real time.

Called Riot, the project is the result of a collaboration between awarding winning multidisciplinary immersive filmmaker Karen Palmer and Professor Hongying Meng from Brunel University. The two have worked together previously on Syncself2, a dynamic interactive video installing.

Riot was inspired by global unrest, and was specifically inspired by Palmers experience of watching live footage of the Ferguson protests in 2015. I felt a big sense of annoyance, rage and helplessness. I needed to create a piece of work that would encourage dialogue around these types of social issues. Riots all over the world now seem to be[ the] last sort of[ community] expression, she said.

Whereas Syncself2 used an EEG headset to place the user in the action, with Riot Palmer wanted to try and achieve a more seamless interface. Hongying and I discussed AI and facial recognition; the tech came from creating an experience which simulated a riot it needed to be as though you were there.

Designed as an immersive social digital experience, the objective is to get through a simulated riot alive. This is achieved through interacting with a variety of characters who can help you reach home. The video narrative is controlled by the emotional state of the user, which is monitored through AI software in real hour.

Machine learning is the key technology for feeling detection systems. From the dataset collected from audiences, AI techniques are used to learn from the data and build the computational model which can be integrated into the interactive film system and see the emotions in real-time, explained Meng.

The programme in development at Brunel can read seven emotions, but not all are appropriate for the experience created by the Riot team. Currently, Riots pilot interface can recognise three emotional state: anxiety, rage and pacify.

Karen Palmer introduces Dr Erinma Ochu to the Riot interface. Photo: Katy Vans

I tried it along with Dr Erinma Ochu, a lecturer in science communication and future media at the University of Salford, whose PhD was in applied neuroscience.

Riot is played out on a large screen, with 3D audio audio surrounding us as a camera watches our facial expressions and calculates in real hour how we are reacting. Based on this feedback, the algorithm decides how the story unfolds.

We see looters, anarchists and police playing their parts and interacting immediately with us. What happens next is up to us: our reactions and answers determine the story, and as the screen is not enclosed in a headset, but open for others to assure, the committee is also generates a public narrative.

Ochu reacted with jumps and gasps to what was happening around her and ultimately didnt make it home. Its interesting to try something you wouldnt do in real life so you can explore a part of your character that you might suppress if you were going to get arrested, she said.

As a scientist and storyteller she felt Riot was ahead of the curve: This has leapfrogged virtual reality, she said.

According to the Riot team, virtual reality( VR) developers have struggled to create fulfill tales in an environment in which, unlike film, you cant control where the user appears or what road they take through the narrative.

In order to overcome these issues and create a coherent, persuading storyline, the team from Brunel re-trained their software versions of facial recognition technology to work for Riot.[ This] offer a perfect platform to demonstrate our the investigations and development. Art builds our run easier to understand. We have been doing research in feeling detection from facial expression, voice, body gesture, EEG, etc for many years, said Meng. He hopes the projects success will induce people ensure the benefits of AI, leading to the development of smart homes, house and cities.

For now, the feeling detection tool being worked on at Brunel can be used in clinical defines to measure ache and emotional states such as depression in patients. Similar tech has already been used in a therapeutic situate; a study last year at the University of Oxford used VR to help those with persecutory hallucinations. Those who trialed real life scenarios combined with cognitive therapy saw significant improvement in their symptoms.

Onlookers observing a players journey through Riot. Palmer hopes technology will advance so that future versions of Riot will support multiple players. Photo: Katy Vans

But can Riots current AI facial recognition tech work for everyone? People with Parkinsons, sight or hearing issues might need an EEG headset and other physical monitors to gain the same immersive experience unless tech development rapidly catches up with Palmers ultimate vision of a 360 degree screen, which would also let a group of participants to play together.

Perhaps Riot and its tech could herald a new empathetic, responsible and responsive future for storytelling and gaming in which the viewer or player is encouraged to bring about change both in the narrative and in themselves. After all, if you could genuinely ensure a story from the another persons point of view what might you learn about them and yourself? How might you carry those insights into the real world to make a difference?

The V& A is likely to be exhibiting Riot as part of the Digital Design Weekend September 2017. The project is currently shortlisted for the Sundance New Frontier Storytelling Lab .

Read more: www.theguardian.com