88
$\begingroup$

In cargo cult science Feynman writes:

"Millikan measured the charge on an electron by an experiment with falling oil drops, and got an answer which we now know not to be quite right. It's a little bit off, because he had the incorrect value for the viscosity of air....Why didn't they discover that the new number was higher right away? It's a thing that scientists are ashamed of--this history--because it's apparent that people did things like this: When they got a number that was too high above Millikan's, they thought something must be wrong--and they would look for and find a reason why something might be wrong. When they got a number closer to Millikan's value they didn't look so hard. And so they eliminated the numbers that were too far off, and did other things like that. We've learned those tricks nowadays, and now we don't have that kind of a disease."

What tricks is he talking about specifically? In fact, in general, what tricks do physicists learn for doing experiments and avoiding fooling themselves?

Second: it's my strong belief that Feynman is referring to something here which isn't in textbooks but instead built into the culture of physics, but I don't understand what exactly it is, and I suspect it's passed along in the culture of physics labs. If someone can explain with some stories, examples or general comments, what is that culture like?

$\endgroup$
11
  • 14
    $\begingroup$ I have fooled myself once rather badly, saw a temperature effect in a substance where I expected it. But colleagues wanted to double-check and saw that my effect was due to the sample rod getting shorter at low temperature. Awfully embarrassing, I was lucky that this was caught before the thing went into print. $\endgroup$
    – user137289
    Commented Oct 28, 2019 at 11:12
  • 5
    $\begingroup$ According to Skeptics SE, Feynman's comment may actually be a bit misleading. skeptics.stackexchange.com/questions/44092/… $\endgroup$
    – JMac
    Commented Oct 28, 2019 at 16:32
  • 44
    $\begingroup$ As a Physics teaching assistant, I once gave top marks to a student who had performed a standard experiment to test conservation of momentum -- and gotten a strange result. Rather than try to cover it up by lengthening his error bars or waving his hands about mistakes he may have made in the setup, he gave a very careful and detailed account of his procedure, and concluded that momentum was not conserved. I spent a few minutes of the next lab session praising that report. $\endgroup$
    – Beta
    Commented Oct 29, 2019 at 4:18
  • 18
    $\begingroup$ In a high-school physics class, I remember a lab experiment to compute g by rolling a ball down an inclined plane. My lab group came up with some odd value of g like 6.4m/s^2 and, well, that was our data so we wrote-up the lab report like that. Other groups looked at their data and said "umm, that value is wrong...let's just fix it" and all their lab reports reported g =9.8m/s^2. It turns out that the lab was a bit contrived for that reason, and my group was the only one who did it correctly. This stuff happens all the time: when you "know" the right answer, it's easy to "discover" it. $\endgroup$ Commented Oct 29, 2019 at 13:01
  • 14
    $\begingroup$ @Beta I don't really agree that "concluding that momentum was not conserved" is a commendable conclusion. One should attempt to evaluate which is more likely: that momentum is indeed not conserved, or that the experimentalist made a mistake in their analysis. In the example, anyone honest would say the second case is more likely. IMO There's nothing wrong with recognizing this and concluding that more work must be done to resolve the obvious discrepancy one way or the other. No progress can be made if we are always arguing over momentum conservation. $\endgroup$ Commented Oct 30, 2019 at 5:37

6 Answers 6

65
$\begingroup$

There are lots of different strategies that are employed by the scientific community to counteract the kind of behavior Feynman talks about, including:

  • Blind analyses: In many experiments, it is required for the data analysis procedure to be chosen before the experimenter actually sees the data. This "freezing" of methodology ensures that nothing about the data itself changes the way it's analyzed, and a methodology that changes once data starts coming in is a red flag that physicists check for in peer review.

  • Statistical literacy: The more you know about statistics, the easier it is to spot data that has been manipulated. Much effort has been devoted to increasing the knowledge of proper statistical practices among physicists so that publishing a flawed analysis (whether intentional or not) is difficult. For example, most experimental physics courses nowadays include training on the basics of statistics and data analysis (for example, mine extensively used Data Reduction and Error Analysis for the Physical Sciences by Bevington and Robinson).

  • Independent collaborations: It's common nowadays for multiple detectors to perform tests of the same hypotheses independently of each other. The procedure, data, and results are all carefully kept as separate as possible before the respective analyses are published. This increases the likelihood that bias or manipulation will be detected, since it will usually cause a difference in results between multiple studies of the same hypothesis.

  • Verification of old results with new data: This is probably less common than it should be in science, due to a cultural preference for performing novel tests, but still occurs at a reasonable frequency in physics. Even in the huge detectors and giant collaborations of high-energy physics, new data is often cross-checked with old data as a side effect of some analyses. For example, an analysis trying to measure the mass of a new particle will often utilize already-known particles to detect its signature, and in the process of characterizing the dataset, will end up confirming earlier measurements of those particles as a "sanity check" of the integrity of the dataset.

  • Incentivization of properly-done disproof: Typically the case where a measurement disagrees with existing hypotheses/theories is met with as much or even more excitement as one that confirms existing hypotheses/theories. This is especially true in high-energy physics, where the majority of the community is eagerly awaiting the first statistically-significant experimental disagreement with the Standard Model. This excitement also brings intense scrutiny of the experiments claiming to measure a disagreement, which helps filter out improperly-done analyses.

This is, of course, a partial list.

$\endgroup$
1
  • 7
    $\begingroup$ We also see independent analysis teams inside a single collaboration. These are not as independent as separate collaborations and check back with each other periodically with the aim of catching errors and convergins on a single well validatied approach. But they really force each team to be prepared to defend its choices as their counterparts will ask the hard questions after the next collaboration meeting. $\endgroup$ Commented Oct 28, 2019 at 17:15
42
$\begingroup$

My favorite story (which I learned about recently) is about Frank Dunnington and his measurements of electron properties in about 1930.

He was measuring the ratio $e/m_e$. Experiments took quite a long time (four years!). When the experimental device was constructed he asked the person who helped him to construct not to tell him some key attribute of the device. He asked: please make two slits here, the distance between them should be somewhere between 18 and 22 degrees. Although this distance was very important, he did not know it until all the experiments were finished.

When he was done with experiments he disassembled the device, measured the distance, put the actual value into formulas in his paper, and published it.

$\endgroup$
3
  • 12
    $\begingroup$ This catagory of percautions is called "blinding the analysis", and it can be done in software as well as in hardware. It is a common procedure in much of particle physics these days. $\endgroup$ Commented Oct 28, 2019 at 15:12
  • 4
    $\begingroup$ Yeah, the simplest way to do it these days is to add a random number with a fixed seed to the result of your analysis as you're running the experiment, then remove that part of the analysis when you're finished taking data and checking every systematic you can think of and ready to publish $\endgroup$
    – llama
    Commented Oct 28, 2019 at 15:56
  • 3
    $\begingroup$ The manipulation to use may be addition or something more complicated depending on the nautre of the analysis. It has to be something that commutes with whatever transformations you are performing on the data during the analysis. For an asymmetry analysis experiments a multiplicative factor applied to the raw rates can be easily backed out, but an addition would be disasterous. $\endgroup$ Commented Oct 28, 2019 at 16:31
16
$\begingroup$

What Feynman is talking about is not particular to physicists, its particular to human nature and it is called "confirmation bias"

Confirmation bias is the tendency to search for, interpret, favor, and recall information in a way that affirms one's prior beliefs or hypotheses. It is a type of cognitive bias and a systematic error of inductive reasoning. People display this bias when they gather or remember information selectively, or when they interpret it in a biased way. The effect is stronger for desired outcomes, for emotionally charged issues, and for deeply-entrenched beliefs.

This is no longer happening as at the time of the electron measurements , because people have become aware of this bias generally. In addition , there exists in particle physics research the opposite bias, people looking for deviations of expected values and theories which would indicate new physics. A recent example is the faster than light neutrino which was finally shown to be a measurement error.

$\endgroup$
2
  • 5
    $\begingroup$ While working in industry in the past, I've seen the research guys recommend a computer calculation as a slight "upgrade". I had my own opinion that the current calculation was more correct. Because I realized that, I color coded the data from both calculations, such that I couldn't recognize which calculation was which. When I chose the better method, it turned out to be the one that I favored, but I didn't know that when I made the selection. As far as one can, it pays to be aware of your own biases, and take steps to guard against them. $\endgroup$ Commented Oct 28, 2019 at 20:14
  • 1
    $\begingroup$ I think the important part was that Millikan was aware of this, and of all human's inability to completely compensate for it. This is why we have the term 'outliers' and why the scientific method requires their inclusion in the data set. You can see this in his own notes where it says "publish" in the margin for data he liked. But if you look at the entire data set, the outliers are all there. $\endgroup$
    – Mazura
    Commented Oct 30, 2019 at 3:49
14
$\begingroup$

The Particle Data Group (PDG) which every other year summarizes the knowledge of Particle Physics prefixes their tome with the figure I include below (from the 2016 edition). It shows how our knowledge of a few select values evolves over time. You can see how some of the values jump at various points in time. That doesn't mean that nature suddenly changed, it is due to our measurements becoming more and more refined and (hopefully) correcting past mistakes. When the values change significantly sometimes the explanation is benign (say, the mass of a particle was measured relative to another particle for which a better measurement became available at that point in time), sometimes invalid assumptions in the measurement were corrected (say, a negligible quantity turned out to not be negligible), sometimes the effect described by Feynman will have played a role (perhaps mediated by another quantity which was used for calibration).

Not only because of my interest in this particle (please read my thesis), I find the most intriguing of these plots to be the one showing the mass of the $\eta(547)$ (second from bottom, leftmost column), one of the light mesons. Here a number of measurements up until 1990 fell into the same ballpark, but with a lot of tension (indicated by the green error bars which tell us that the PDG applied their scaling technique). Then around 1990 a new measurement appeared which shifted the mean value by several error bars and, after some experimental tension in the mid-noughts, the value jumped again to the now best value. Basically, the tension could twice only be resolved by much more precise experiments that shifted the central value significantly.

So it seems like even in the most fundamental measurements there can be a reluctance to change, but precision wins.

Particle properties over time

$\endgroup$
0
$\begingroup$

You have exactly reversed Feynman's meaning in his use of the word "tricks". His use of "tricks" references discarding values far from Millikan's and keeping values close to Millikan's. Thus, his use of "tricks" is to methods of fooling ourselves.

You then use "tricks" in exactly the opposite sense. Consequently, there is no such thing as what you request: "What tricks is he talking about specifically. In fact in general what tricks do Physics major learn for doing experiments and avoiding fooling yourselves." is explicitly describing an empty set of tricks.

$\endgroup$
0
0
$\begingroup$

There is a nice article from Robert MacCoun and Saul Perlmutter: "Blind analysis: Hide results to seek the truth". They suggest that one should use random numbers to

  1. add noise to measured data-points,
  2. add artificial biases, and/or
  3. scramble the categories to which the data-points belong

during the pre-data-analysis. The central idea of this approach is that we debug the programs for the data analysis, draw conclusion for the next steps of the data analysis, and analyse the influence of a possible outlier while we are still "blind folded" and thus unbiased. Only when one agrees that the data analysis is done, the random numbers are removed and the final results are investigated.

$\endgroup$

Not the answer you're looking for? Browse other questions tagged or ask your own question.