If you are interested in AI Safety, come visit the AI Safety Reading Group. The AI Safety reading group meets on Skype Wednesdays at 18:45 UTC, discussing new and old articles on different aspects of AI Safety. We start with a presentation round, then a summary of the article is presented, followed by discussion both on the article and in general.Sometimes we have guests. On Wednesday the 14th, Stuart Armstrong will be giving a presentation on his research agenda in the reading group: https://www.alignmentforum.org/posts/CSEdLLEkap2pubjof/research-agenda-v0-9-synthesising-a-human-s-preferences-into Join us by Skype, by adding ‘soeren.elverlin’.Previous guests include Eric Drexler, Rohin Shah, Matthijs Maas, Scott Garrabrant, Robin Hanson, Roman Yampolskiy, Vadim Kosoy, Abram Demski and Paul Christiano. A full list of articles read can be found at https://aisafety.com/reading-group/
Is this reading group still running? I’m wondering whether to point people to it.
Comment
Yes, we are still running, though at a bi-weekly schedule. We will discuss Paul Christiano’s "Another (Outer) Alignment failure story" on the 8th of July.
I’m sad to have missed Eric Drexler’s recent Q&A session. The slides for that session don’t seem to contain Eric’s answers and there is no linked recording. Is there any chance someone kept notes, or can write a summary from their memory of Eric’s answers?
Comment
Eric Drexler requested that I did not upload a recording to YouTube. Before the session, I compiled this document with most of the questions: https://www.dropbox.com/s/i5oqix83wsfv1u5/Comprehensive_AI_Services_Q_A.pptx?dl=0 We did not get to post the last few questions. Are there any questions from this list you would like me to try to remember the answers to?
Comment
Do you have a recording of the session? If so, can you send it to me via PM or email?
I’m interested in answers to pretty much all of the questions. If no recording is available, any chance you could write up as many answers as you can remember? (If not, I’ll try harder to narrow down my interest. :)
I’m also curious why Eric Drexler didn’t want you to upload a recording to YouTube. If the answers contain info hazards, it seems like writing up the answers publicly would be bad too. If not, what could outweigh the obvious positive value of releasing the recording? If he’s worried about something like not necessarily endorsing the answers that he gave on the spot, maybe someone could prepare a transcript of the session for him to edit and then post?
I’m very interested in his responses to the following questions:
The question addressing Gwern’s post about Tool AIs wanting to be Agent AIs.
The question addressing his optimism about progress without theoretical breakthroughs (related to NNs/DL).
Comment
Comment
Thanks!