Posts Tagged‘neural nets’

Models of our mind and communities

by reestheskin on 18/12/2018

Comments are disabled

Google’s AI Guru Wants Computers to Think More Like Brains | WIRED

This is from an interview with Geoffrey Hinton who — to paraphrase Peter Medawar’s comments about Jim Watson — has something to be clever about. The article is worth reading in full, but here are a few snippets.

Now if you send in a paper that has a radically new idea, there’s no chance in hell it will get accepted, because it’s going to get some junior reviewer who doesn’t understand it. Or it’s going to get a senior reviewer who’s trying to review too many papers and doesn’t understand it first time round and assumes it must be nonsense. Anything that makes the brain hurt is not going to get accepted. And I think that’s really bad…

What we should be going for, particularly in the basic science conferences, is radically new ideas. Because we know a radically new idea in the long run is going to be much more influential than a tiny improvement. That’s I think the main downside of the fact that we’ve got this inversion now, where you’ve got a few senior guys and a gazillion young guys.

I would make a few comments:

  1. First the history of neural nets is long: even people like me had heard about them in the late 1980s. The history of ideas is often like that.
  2. The academy is being sidetracked into thinking it should innovate or develop ideas that whilst important are not revolutionary. Failure should be the norm, rather than the continued treadmill of grant income and papers.
  3. Scale and genuine discovery — for functioning of peer groups — seldom go together.
  4. Whilst most of the really good ideas are still out there, it is possible to create structures that stop people looking for them.
  5. Hinton makes a very important point in the article with broad relevance. He argues that you cannot judge (or restrict the use of) AI on the basis of whether or not it can justify its behaviour in terms of rules or logic — you have to judge it on it ability to work, in general. This is the same standard we apply to humans, or at least we did, until we thought it wise or expedient to create the fiction that much of human decision making is capable of conscious scrutiny. This applies to medicine, to the extent that clinical reasoning is often a fiction that masters like to tell novices about. Just-so stories, to torment the young with. And elsewhere in the academy for the outlandish claims that are made for changing human behaviour by signing up for online (“human remains”)courses (TIJABP).

All has been said before, I know, but no apology will be forthcoming.