Robot opera: Robert Gyorgyi interviews Ron Chrisley

Robert Gyorgyi, a Music student here at Sussex, recently interviewed me for his dissertation on robot opera.  He asked me about my recent collaborations, in which I programmed Nao robots to perform in operas composed for them.  Below is the transcript.

Robot-opera-1024x576

Interview with Dr Ron Chrisley, 20 April 2018, 12:00, University of Sussex

Bold text: Interviewer (Robert Gyorgyi), [R]: Dr Ron Chrisley

NB: The names ‘Ed’ and ‘Evelyn’ often come up within the interview. ‘Ed’ refers to Ed Hughes, the composer of Opposite of Familiarity (2017) and Evelyn to ‘Evelyn Ficarra’, composer of O, One (2017)

How did you hear about the project? Was it a sort of group brainstorming or was the idea proposed to you?

[R] -Evelyn approached me, then we had a meeting when she explained her vision to me.

These NAO robots are social robots designed to speak, not to sing. Was the assignment of their new task your main challenge? How did you do that? Continue reading

Advertisements

Robot Opera Mini-Symposium Video

47309

Last June I participated in the Robot Opera Mini Symposium organised by the Centre for Research in Opera and Music Theatre (CROMT) at Sussex.  A video of all the talks, and the robot opera performances themselves, is available below.  My 17-minute talk can be found at 08:40 into the video.

Robot Opera coverage in “Viva Lewes”

The September 2017 issue of Viva Lewes magazine features a two-page spread by Jacqui Bealing on the robot opera project that Evelyn Ficarra, Ed Hughes and I have been collaborating on (as detailed in earlier updates on this blog).  The article is available at:

 https://www.yumpu.com/en/document/view/59398533/viva-lewes-issue-132-september-2017/122

For convenience, I include a copy of the article below.

Screen Shot 2017-10-05 at 14.00.01

Machine consciousness at the Brighton Digital Festival

image-exmachina-800x500

Next Tuesday I’ll be giving a brief talk on machine consciousness prior to a screening of the film Ex Machina, as part of the Brighton Digital Festival. Sackler colleagues Keisuke Suzuki and David Schwartzman will be giving consciousness-illuminating VR demos involving our Nao robots as well. The event is being organised in conjunction with the British Science Association.  More info at http://theoldmarket.com/shows/toms-film-club-ex_machina-2015/


Update, 4 October 2017:

Here are some photos of the event, courtesy of Amber John.  As you can see, the title I settled on was “Turing Tests and Machine Consciousness”.

IMG_4889.JPGIMG_4890.JPGIMG_4892.JPGIMG_4896.JPGTOM1.pngTOM2.jpg

Hands-on learning with social robots in schools

img_1347I’ve been working with student assistant Deepeka Khosla to design hands-on social robotics curricula for school students. We delivered three sessions for year 7 and 8 students on January 12th using AiBO and NAO robots, which involved some of the students doing some (very-limited) coding of the robots, and inspection of their program and sensory states, a basic form of increasing “transparency” of social robots.
A key component of making robots more intelligibile is the development of “roboliteracy”: a good understanding of what can and what cannot be (currently) done/expected to be done by social robots. Familiarity can be a key component of de-mystification/anxiety reduction.
img_4691Current plans are underway to develop a more advanced, coding-based 3-hour learning session for year 9 students, for delivery over 2017-1018, starting in May. This will be marketed exclusively to girls. During my recent visit to the UAE I was inspired by what I saw, and the reports I heard, concerning the strong representation of women and girls in robotics education in that part of the world. Just letting girls here know about that, showing them photos of female robotics teams from there, etc., might be an example of a way to make the course content match that marketing aim.
Any suggestions/examples concerning robot curriculum in schools would be very welcome!
Support for development and delivery of these sessions has been provided by the Widening Participation initiative at Sussex.

Ethically designing robots without designing ethical robots

robot_ethicsNext Thursday, November 17th, at 13:00 I’ll be leading the E-Intentionality seminar in Freeman G22. I’ll be using this seminar as a dry run for the first part of my keynote lecture at the UAE Social Robotics meeting next week. It builds on work that I first presented at Tufts in 2014.

Abstract:

Since robots will not, in the near future, be responsible agents, avoiding some moral hazards (e.g., that of abdication of responsibility) will require designs that assist in tracing complex lines of responsibility backwards from outcomes, through the robot, and back to the appropriate humans and/or social institutions. I look at one approach to ethically designing robots, that of designing ethical robots – robots that are given a set of rules that are intended to encode an ethical system, and which are to be applied by the robot in the generation of its behaviour. I argue that this approach will in many cases obfuscate, rather than clarify, the lines of responsibility involved (resulting in “moral murk”), and can lead to ethically adverse situations. After giving an example of such cases, I offer an alternative approach to ethical design of robots, one that does not presuppose that notions of obligation and permission apply to the robot in question, thereby avoiding the problems of moral murk and ethical adversity.

Artificial social agents in a world of conscious beings

I forgot to mention in the update posted earlier today that fellow PAICSer, Steve Torrance, will also be a keynote speaker at the 2nd Joint UAE Symposium on Social Robotics.  Here are his title and abstract.logo

Artificial social agents in a world of conscious beings.

Steve Torrance

Abstract

It is an important fact about each of us that we are conscious beings, and that the others we interact with in our social world are also conscious beings. Yet we are appear to be on the edge of a revolution in new social relationships – interactions and intimacies with a variety of non-conscious artificial social agents (ASAs) – both virtual and physical. Granted, we often behave, in the company of such ASAs as though they are conscious, and as though they are social beings. But in essence we still think of them, at least in our more reflective moments, as “tools” or “systems” – smart, and getting smarter, but lacking phenomenal awareness or real emotion.

In my talk I will discuss ways in which reflection on consciousness – both natural and (would-be) artificial – impacts on our intimate social relationships with robots. And I will propose some implications for human responsibilities in developing these technologies.

I will focus on two questions: (1) What would it take for an ASA to be conscious in a way that “matters”? (2) Can we talk of genuine social relationships or interactions with agents that have no consciousness?

On question (1), I will look at debates in the fields of machine consciousness and machine ethics, in order to examine the range of possible positions that may be taken. I will suggest that there is a close relation between thinking of a being as having a conscious phenomenology, and adopting a range of ethical attitudes towards that being. I will also discuss an important debate between those who take a “social-relational” approach to phenomenological and ethical attributions, and those who take an “objectivist” approach. I will offer ways to resolve that debate. This will help provide guidance, I hope, to those who are developing the technologies for smarter ASAs, which possibly may have stronger claims to be taken as conscious. On (2), I will suggest that, even for ASAs that are acknowledged not to be conscious, it is possible that there could be a range of ethical roles that they could come to occupy, in a way that would justify our talking of “artificial social agents” in a rich sense, one that would imply that they had both genuine ethico-social responsibilities and ethico-social entitlements.

The spread of ASAs – whether or not genuinely conscious, social or ethical – will impose heavy responsibilities upon technologists, and those they work with, to guide the social impacts of such agents in acceptable directions, as such agents increasingly inter-operate with us and with our lives. I will thus conclude by pointing to some issues of urgent social concern that are raised by the likely proliferation of ASAs in the coming years and decades.