Posts

Showing posts from 2023

"Pre-Galilean" Foolishness

 I am currently reading The Master and His Emissary , which appears to be an excellent book. ("Appears" because I don't know the neuroscience literature well enough to say for sure, yet.) But then on page 186 I find: "Asking cognition, however, to give a perspective on the relationship between cognition and affect is like asking astronomer in the pre-Galilean geocentric world, whether, in his opinion, the sun moves round the earth of the earth around the sun. To ask a question alone would be enough to label one as mad." OK, this is garbage. First of all, it should be pre-Copernican, not pre-Galilean. But much worse is that people have seriously been considering heliocentrism for many centuries before Copernicus. Aristarchus had proposed a heliocentric model in the 4th-century BC. It had generally been considered wrong, but not "mad." (And wrong for scientific reasons: Why, for instance, did we not observe stellar parallax?) And when Copernicus propose...

"Machine Learning"

 The name is a misnomer. And a harmful one, because it interferes with understanding the process that is really occuring. What is really occurring is a search of a constrained program space. Let's say you want to be able to identify images of hot dogs . You begin with a plausible program for doing so, that is able to also search the space of nearby programs that might get better results on the problem. You then (in "supervised learning") provide scores that indicate how well one of these possible programs has done on solving the problem. After doing this for some time you settle upon a program that solves the problem "well enough." This is a great technique that can produce truly impressive results on a wide class of problems, such as identifying images of hot dogs. But notice that, except for the phrase in scare quotes, there is no "learning" in the description. Calling this "learning" is importing ideological baggage that just obscures what...