In Science, What Matters?

And how do you learn what matters?

When I was a grad student, I read Stanislav Ulam’s memoir Adventures of a Mathematician. I was impressed by something Ulam said about John von Neumann: that he grasped the difference between the trunk of the tree of mathematics and the branches. Between core issues and lesser ones. Between what matters more and what matters less. I wanted to make similar distinctions within psychology. Nobody talked about this, however. Not even other books.

Some research will be influential, will be built upon. Some won’t. To put it bluntly, some research will matter, some won’t. I once thought of teaching a graduate course where students learn to predict how many citations an article will receive. You take a 10-year-old journal issue, for example, and try to predict how many citations each article will receive. I like to think it would have been a helpful class: The key to a successful scientific career is writing articles that are often cited. I even had a title: “What Will You Do After You Stop Imitating Your Advisor?”

When I was a grad student the short answer to “what matters?” in experimental psychology was clear enough:

1. New methods. The Skinner box, for example, was a new way to study instrumental learning. Skinner didn’t discover or create the first laboratory demonstration of instrumental learning; he simply made it easier to study.

2. New effects. New cause-and-effect linkages. For example, John Garcia discovered that if you make a rat sick after it experiences a new flavor it will avoid foods with that flavor.

My doctoral dissertation was about a new way to study animal timing.

A few months ago I had coffee with Glen Weyl, a graduate student in economics at Princeton. We discussed his doctoral research, which is about how to test theories. One of Glen’s advisors had told him about a paper by Hal Pashler and me on the subject. Hal and I argued that fitting a model to data is a poor way to test the model because there is no allowance for the model’s flexibility. The first reviewers of our paper didn’t like it. “You don’t realize how hard it is to find a model that fits,” one of them wrote.

Glen’s interest in this question began during a seminar in Italy, when he realized the speaker was more or less ignoring the problem. The speaker was comparing how well two different theories could explain the same data without taking into account their different amounts of flexibility. Glen’s thesis proposes a Bayesian framework that allows you to do this. His main example uses data of Charness and Rabin from choice experiments. (Matt Rabin is a MacArthur Fellow.) Taking flexibility into account, he reaches a different conclusion than they did.

I wondered how Glen decided this was important. (It’s a method, yes, but a highly abstract one.) I asked him. He replied:

Sadly, despite my interest in the history of economic thought, I don’t have a lot of insight about why I came upon these thoughts. But one thing: my interests are very interdisciplinary . . . My work is based on drawing connections between economics, philosophy of science, and computer science (and meta-analysis from psychology and bio-statistics). Most of my work takes this form: as you’ll see on my website, I’ve used theoretical insights from economics and computer science as well as evidence from neuroscience, psychology and biology to critique the individualist foundations of liberal rights theory; I’ve used ideas from decision theory to lay firmer foundations for goals set out by computer scientists designing algorithms; I’ve used tools from information theory to instantiate insights from psychology to help understand the design of auctions; and I’ve used computational neuroscience to model biases in economic information processing. Broad interests are hard to have, because they limit the time for learning a particular area in depth, but I prefer to read broadly and draw connections rather than to read deeply and chip away at open questions.

That was interesting. I read broadly, and so does Hal, who knows more about the philosophy of science than I do. I wrote to Glen:

The usual comment about interdisciplinary knowledge is that it’s good because you can bring ideas from one area, including solutions and methods, to solve problems in another area. . . . But maybe it’s also good because by learning about different areas you absorb a range of different value systems and this makes you less sensitive to fads (which vary from field to field), more sensitive to longer-lasting and more broadly-held values.

The more trees you know, the easier it is to see the forest.

Evaluating new product ideas.

My Theory of Human Evolution (Civil Rights Movement edition)

From Eyes on the Prize, about an Easter boycott of Nashville stores:

Easter was a most important time to buy. All blacks had to have a full, brand new outfit at Easter, no matter how poor you were, right? You may start three months ahead of time paying for that Easter outfit, and you may be paying for it for three months later.

There is a similar tradition in China: At the start of the new year you buy new clothes. I’ve blogged before about how rituals, ceremonies, and holidays promoted technological development: They increased the demand for high-end items. This helped skilled craftspeople make a living.

Amy Winehouse and Nassim Taleb

Will Amy Winehouse — who won five Grammys last night — help or hurt the music industry? A few years ago, I went to a tasting event called The Joy of Sake. There were about 100 of the best sakes from Japan. A pre-event talk for retailers discussed the decline of sake in Japan. (Soju is cool; sake is old-fashioned.) That was the reason for the show. I loved tasting 30-odd high-quality sakes but the overall effect on me was the opposite of what the promoters wanted. I quickly became a connoisseur. I no longer liked the cheap stuff — ugh! But the stuff I did like was too expensive. I stopped buying sake.

Before last night I had heard of Amy Winehouse and I had heard Rehab, but hadn’t put the two together. Her Grammy performance blew me away. I watched a bunch of YouTubes of her. Back at the Grammys, I listened to an orchestra play Rhapsody in Blue. I used to like it; now it sounded awful. I listened to a few more group performances; they too sounded bad. Just as The Joy of Sake had made me no longer enjoy cheap sake, listening to a lot of Amy Winehouse had made me no longer enjoy “average” music — music where several individual performances are combined.

I thought of The Black Swan by Nassim Taleb. Taleb defined Mediocristan as situations where no one datum can have a big effect on the result. The average height of 100 people, for example. In Extremistan, by contrast, a single datum can make a big difference. The average wealth of 100 people, for example — one person can have much more money than the other 99 put together. Orchestras are Mediocristan, I realized; individual singers are Extremistan. In art, emotional impact is everything. Extremistan allows really big impact; Mediocristan does not. Maybe this is why classical music is dying.

I felt like throwing away half my CDs. I could use the space. Thanks, Amy!

How to Be Wrong (continued)

I asked a friend of mine why she was a good boss. “I was nurturing,” she said. A big study of managers reached essentially the same conclusion: Good managers don’t try to make employees fit a pre-established box, the manager’s preconception about how to do the job. A good manager tries to encourage, to bring out, whatever strengths the employee already has. This wasn’t a philosophy or value judgment, it was what the data showed. The “good” managers were defined as the more productive ones — something like that. (My post about this.)

The reason for the study, as Veblen might say, was the need for it. Most managers failed to act this way. I posted a few days ago about a similar tendency among scientists: When faced with new data, a tendency to focus on what’s wrong with it and ignore what’s right about it. To pay far more attention to limitations than strengths. Here are two examples:

1. Everyone’s heard “correlation does not imply causation”. I’ve never heard a parallel saying about what correlation does imply. It would be along the lines of “something is better than nothing.”

2. Recently I attended a research group meeting in which a postdoc talked about new data she had gathered. The entire discussion was about the problems with it — what she couldn’t infer from it. There could have been a long discussion about how it added to what we already know, but there wasn’t a word about this.

Some of the comments considered this behavior a kind of Bayesian resistance to change in beliefs. But it occurs regardless of whether the new data support or contradict prior beliefs. There’s nothing about prior beliefs in “correlation does not imply causation.” The post-doc wasn’t presenting data that contradicted what anyone knew. Also, similar behavior occurs in other areas besides science (e.g., how managers manage) in which the Bayesian explanation doesn’t fit so well.

I think it’s really strong. I was guilty of it myself when discussing it! I made very clear how this tendency is a problem, giving the analogy of a car that could turn left but not right. Obviously bad. I said nothing about the opportunities this tendency gives everyone. My self-experimentation is an example. The more that others reject useful data, the more likely it is that useful data is lying around and doesn’t require much effort to find. I have called this behavior dismissive; I could have called it generous. It’s like leaving money lying on the ground.

A related discussion at Overcoming Bias. What should “correlation does not imply causation” be replaced with?

Addendum. Barry Goldwater weighs in: “I’m frankly sick and tired of the political preachers across this country telling me as a citizen that if I want to be a moral person, I must believe in ‘A,’ ‘B,’ ‘C,’ and ‘D.’” Indeed, preachers spend far more time on what we are doing wrong (and should do less of) than on what we are doing right (and should do more of). The preacher Joel Osteen has taken great advantage of this tendency. “I think most people already know what they are doing wrong,” he told 60 Minutes.

Science in Action: Flavor-Calorie Learning (another simple example)

At the heart of the Shangri-La Diet is the idea that we learn to associate flavors (smells) with calories. This learning was first shown in rat experiments. There’s some human evidence, but not much. If I could discover more about what controls this learning, I might be able to improve the diet. For example, maybe I could say more about what the flavor-free window should be.

My earlier self-experimentation on this subject – I used tea for flavor and sugar for calories — was helpful. To my surprise, I found that really small changes in flavor made a noticeable difference. If I switched from one canister of Peet’s Gunpowder Tea to a new canister, the ratings went down, although everything else stayed the same. From this came the notion of ditto food: Foods with exactly the same flavor each time are especially fattening. I hadn’t realized what a difference it would make if you kept the flavor exactly the same each time.

It’s been hard to learn more. After Christmas dinner, my mom gave me the leftover brandy (A. R. Murrow). I used it for a very simple experiment in which I learned to like it. I’ve never drunk brandy in any quantity and I started off not liking it. Every day for a few weeks, I drank one tablespoon. I drank it in a few sips over a few minutes. I didn’t eat anything else for at least 30 minutes. I rated how good it tasted on a 0-100 scale where 10 = very bad, 20= quite bad, 25 = bad, 30 = somewhat bad, 40 = slightly bad, 50 = neutral, 60 = slightly good, 70 = somewhat good, 75 = good, 80 = quite good, 90 = very good. The overall rating was the maximum of the ratings of the several sips. (The first sip usually tasted the best.)

Here are the results.

learning to like brandy

I’ve observed similar results five or six times. They are more support for the most basic conclusions: 1. The effect is very clear. One tablespoon of brandy has only 30 calories. 2. A really simple experiment is easy.

That’s a promising start but then it gets hard, or at least non-obvious. As a way to study flavor-calorie learning, this little example has several flaws: 1. Slow learning. 2. Expensive materials. 3. Little control of flavor. The best I can do is choose which liquor to buy. Soon I will run out of ones I haven’t used. 4. No way to separate flavor and calories in time. 5. No way to change the calorie source.

An earlier demonstration used a soft drink. It’s really Science in Inaction: I’ve made zero progress in a year.

Ranjit Chandra: A New Position

The Indian health tourism company Indicure has appointed Dr. Ranjit Chandra, whose story is told here, to be one of its panel of experts. Few scientists have a more impressive resume:

Dr. Chandra has received 16 honorary degrees including DSc honoris causa recently from Panjab University. He has received over 100 awards worldwide. In 2003, he was given the Jubilee Gold Medal by the Queen and the title of Honorable Baron of Blackburn. Prof. Chandra is an Officer of the Order of Canada, the highest award given to Canadian citizens.

More about Chandra. His work remains influential.

The Lessons of Bilboquet

There are lots of omega-3-related self-experiments I’d like to do: 1. What about fish oil? 2. Is omega-6 bad for the brain? As my olive-oil results suggested. 3. “Blind” experiments where I don’t know what I’ve ingested. I wanted to use a design that involved many tests/day. This would be easy if the tests were fun, hard if they weren’t. Games are fun–could I figure out why and make a mental test that was like playing a game?

After talking with Greg Niemeyer, I decided that color, variety, feedback, and appropriate difficulty (not too little, not too much) were possible reasons games are fun. I constructed a letter-counting task with all of these attributes — and it wasn’t fun. I had to push myself to do it. These attributes may help, but not a lot.

Then, as I’ve posted, a friend gave me a bilboquet. For such a simple object, it was surprisingly fun and slightly addictive. Thinking about other addictive games, such as Tetris (I once played a lot of Tetris), I guessed that the crucial features of a game that make it addictive are: 1. Success is sharply defined. 2. Not too easy. 3. Hand-eye coordination. (Not any eye-body coordination: I did thousands of balancing tests but had no trouble stopping.)

I constructed a new task with these attributes: Click the Circle. A circle appears on the screen, you move the pointer to the circle and click on it; a new circle appears somewhere else, you move the pointer to click on it, etc. At the end there’s a little feedback: how long it took. Very simple.

This task, at least so far, is addictive. I think something else may be going on in addition to the three factors: we enjoy completion, especially visual completion. (Which Tetris had a lot of.) In this case the visual completion is the blank space that appears when I click on a circle. If I have a few dishes to do, it’s easy to do them–the promise of an empty sink (= visual completion) draws me to the task. In contrast, if there are a lot of dishes to do, it’s much harder to do a few of them. I’ll probably do none of them or all of them. If you have 20 dishes to do, doing them will generate a lot more pleasure (and thus will be easier to do in the future) if you can manage to create 20 completion moments than if they get piled up and there is only one completion moment.