The difficulty in predicting AI, in three lines

https://www.lesswrong.com/posts/srQFeGqgrZZXto8Ht/the-difficulty-in-predicting-ai-in-three-lines

An over-simplification, but an evocative one:

Comment

https://www.lesswrong.com/posts/srQFeGqgrZZXto8Ht/the-difficulty-in-predicting-ai-in-three-lines?commentId=522TGRmR3kPSzHJ3o

To summarise the argument further.

"A lot of people talk rubbish about AI. Therefore most existing predictions are not very certain."

That doesn’t in itself mean that it’s hard to predict AI—merely that there are many existing predictions which aren’t that good. Whether we could do better if we (to take the given example) used the scientific method isn’t something the argument covers.

Comment

https://www.lesswrong.com/posts/srQFeGqgrZZXto8Ht/the-difficulty-in-predicting-ai-in-three-lines?commentId=9RpLFZqmQx2t7bZgY

Whether we could do better if we (to take the given example) used the scientific method

I don’t really see how we could do that. Yes, most predictions are rubbish—but a lot are rubbish because predicting AI is not something we have good ways of doing.

https://www.lesswrong.com/posts/srQFeGqgrZZXto8Ht/the-difficulty-in-predicting-ai-in-three-lines?commentId=rqcTvKes8paxDxy4d

I don’t see how the third proposition follows from the first two.

Comment

https://www.lesswrong.com/posts/srQFeGqgrZZXto8Ht/the-difficulty-in-predicting-ai-in-three-lines?commentId=ZEkQdqSLttjayt2HR

Clarified the second line.

Comment

That at least makes sense.

Comment

What did it originally say?

Comment

it didn’t have the "; AI predictions generally don’t."

I’ve been working with these predictions for such a long time, I forgot not everyone had this at the forefront of their minds.

https://www.lesswrong.com/posts/srQFeGqgrZZXto8Ht/the-difficulty-in-predicting-ai-in-three-lines?commentId=BRaE3duKYW5PAQG33

The social sciences are contentious, their predictions questionable.

And yet social sciences use the scientific method; mathematics doesn’t

Hence statements involving math should be treated as less certain than any prediction in the social sciences.

Comment

https://www.lesswrong.com/posts/srQFeGqgrZZXto8Ht/the-difficulty-in-predicting-ai-in-three-lines?commentId=fGxmf3E7twYdJLW8D

:-)

You’re right—there is one area whose methods are even better than science. If only more problems could be solved like math problems!

https://www.lesswrong.com/posts/srQFeGqgrZZXto8Ht/the-difficulty-in-predicting-ai-in-three-lines?commentId=PbrpHnb3zCqmS3Tgf

I’ve started giving AI timelines as between 10 years and 100 years.

Comment

https://www.lesswrong.com/posts/srQFeGqgrZZXto8Ht/the-difficulty-in-predicting-ai-in-three-lines?commentId=FN4XKAtKviJtPSTYx

That seems reasonable. I give the 5-100 year range myself.

Comment

Figures.

https://www.lesswrong.com/posts/srQFeGqgrZZXto8Ht/the-difficulty-in-predicting-ai-in-three-lines?commentId=7HCuSFh3Gw36x8gEc

That implicitly assumes that there aren’t reasons why social sciences are contentious which don’t also apply to AI predictions, but I don’t think that’s terribly unreasonable (EDIT: where by "I don’t think that’s terribly unreasonable" I mean that the reasons why the social sciences are contentious despite using the scientific method that I can think off the top of my head would also kind-of apply to AI predictions).

https://www.lesswrong.com/posts/srQFeGqgrZZXto8Ht/the-difficulty-in-predicting-ai-in-three-lines?commentId=mRYPF3x9KjS46CAvD

And yet social sciences use the scientific method; AI predictions generally don’t.

Can you please clarify this point?

Comment

https://www.lesswrong.com/posts/srQFeGqgrZZXto8Ht/the-difficulty-in-predicting-ai-in-three-lines?commentId=zmTmMB7oowjj4sYDB

The social sciences are sciences; AI predictions are mainly speculative thinking by people who just put on their thinking caps and think really really hard about the future (see some of the examples in http://​​lesswrong.com/​​lw/​​e79/​​ai_timeline_prediction_data/​​).

Comment

Are you saying that these predictions are unscientific because they are based on untestable models? Or because the models are testable for "small" predictions, but the AI predictions based on them are wild extrapolations beyond the models’ validity?

Comment

Most predictions don’t use models; most models aren’t tested; and AI predictions based on tested models are generally wild extrapolations.

Comment

It does sound pretty bad if that’s the case. My suspicion is that the models are there, just implicit and poor-quality. Maybe trying to explicate, compare and critique them would be worthwhile.

https://www.lesswrong.com/posts/srQFeGqgrZZXto8Ht/the-difficulty-in-predicting-ai-in-three-lines?commentId=yW9Xw8n85fHs6FCM8

Yes, people say all sorts of unjustified stuff about AI as if their musings were true, out of excitement and carelessness. But the line of thought in the post is ultimately destructive because it sets low expectations for no good reason.

To use the scientific method just means to make falsifiable predictions. So any arbitrary hypothesis counts, no matter how outlandish, so long as it’s predictive. On the other hand, you don’t need to use science in order to reason, and since "human-level AI" is not available for experimental study, we can only reason about it. But it’s a pretty sure thing that such an AI will think that 1+1 equals 2…

There are no details here e.g. about the methodologies used to produce futurological predictions of the "time until X", or about the premises employed in reasoning about AI dispositions and capabilities; and that means there’s no argument about the degree of reliability or usefulness that can be obtained when reasoning about AI; just the bare assertion, "not even as good as the worst of social science". Also, there’s no consideration of the power of intention. A lot of the important statements in LW’s AI futurology are about designing an AI to have desired properties.

Comment

https://www.lesswrong.com/posts/srQFeGqgrZZXto8Ht/the-difficulty-in-predicting-ai-in-three-lines?commentId=tJ7WZzk4xjFAAEdDa

I’m constructing a detailed analysis of all these points for my "How to Predict AI" paper.

And there are few details about methodologies, yes—because the vast majority of predictions have no methodologies. The quality of predictions is really, really low, and there are reasons to suspect that even when the methodologies are better, the prediction is still barely better than guesswork.

My stub was an unjustified snark, but the general sentiment behind it—that AI predictions (especially timeline predictions) are less reliable that social science results—is, as far as I can tell, true.