Pruning the tree of science

Pruning the tree of science

Photo by veeterzy on Unsplash

Originally published 21 January 1991

More than half of sci­en­tif­ic research con­tributes noth­ing to the growth of knowledge.

This aston­ish­ing state­ment just might be true, but before I give the evi­dence, let me explore a famil­iar metaphor.

Sci­en­tif­ic knowl­edge grows organ­i­cal­ly, like a tree.

Every piece of pub­lished research is like a new bud on a twig. The bud is con­nect­ed to every oth­er bud on the tree. Two buds may be very close togeth­er, on the same twig, or very far apart, so that to trace the con­nec­tion one would have to fol­low twigs and branch­es all the way back to the trunk and out again along oth­er branch­es and twigs. Ulti­mate­ly, all sci­en­tif­ic knowl­edge is one. The con­nect­ed­ness of sci­ence gives us con­fi­dence in its integrity.

The amount of research being pub­lished today is so great that no one can claim to know the over­all shape of the tree. There are present­ly near­ly 75,000 sci­en­tif­ic jour­nals list­ed in the Bowker/Ulrich cat­a­log of inter­na­tion­al peri­od­i­cals, and thou­sands of new jour­nals are added each year. It is vir­tu­al­ly impos­si­ble for any one sci­en­tist to become famil­iar with more than a sin­gle branch of the tree.

Or to put it anoth­er way, we can’t see the tree of sci­ence for the for­est of pub­lished words. How­ev­er, help may be at hand.

The citation rule

A rule of sci­ence requires that every pub­lished paper cite all pre­vi­ous­ly pub­lished work that bears upon the same sub­ject. The cita­tion rule assures that new research is firm­ly con­nect­ed to the tree of science.

Com­put­er-based cita­tion index­es may make it pos­si­ble to sketch out the out­lines of the tree, and com­put­er ana­lysts have begun to do just that, by trac­ing net­works of citations.

At the request of the jour­nal Sci­ence, David Pendle­bury of the Philadel­phia-based Insti­tute for Sci­en­tif­ic Infor­ma­tion (ISI) recent­ly searched a cita­tion data­base for papers pub­lished between 1981 and 1985. He dis­cov­ered that 55 per­cent of papers sur­veyed had not been cit­ed even once in the five years after publication.

The ISI data­base includ­ed only the top 4,500 sci­ence and social sci­ence jour­nals, or about 6 per­cent of the total num­ber list­ed in the Bowker/Ulrich cat­a­log of peri­od­i­cals. These are the jour­nals that are most like­ly to be read and cit­ed, so the cita­tion rate for papers pub­lished in oth­er jour­nals is almost cer­tain­ly lower.

Because of the cita­tion rule, if a work is not cit­ed, it pre­sum­ably has no influ­ence on sub­se­quent research. Appar­ent­ly, half of the buds on the tree of sci­ence are dead ends. They could be snipped away and the tree would grow as robust­ly as ever. Or at least that’s what the com­put­er study seems to indicate.

At first glance, this might seem promis­ing. Good hor­ti­cul­tur­al prac­tice sug­gests that a tree grows bet­ter with judi­cious prun­ing. Snip a branch here and there, and the oth­er branch­es will grow more vig­or­ous­ly. This might be espe­cial­ly true when the avail­able nutri­ents for growth — in the form of gov­ern­ment fund­ing — are becom­ing increas­ing­ly meager.

The prob­lem is know­ing which branch­es to prune.

Cita­tion index­es, if wide­ly avail­able on an inter­na­tion­al com­put­er net­work, might pro­vide the answer. Fund­ing, pro­mo­tion, and aca­d­e­m­ic tenure could be direct­ed to those researchers whose works are cit­ed. Uncit­ed work would be allowed to with­er away. The fecun­di­ty of a line of research, not a mere list of pub­li­ca­tions, would become the mea­sure of quality.

There is, in fact, a strong cor­re­la­tion between cita­tion rates and the pub­lic per­cep­tion of qual­i­ty. A typ­i­cal Har­vard sci­en­tif­ic paper, for exam­ple, is cit­ed 25 times over a 15-year peri­od, which puts Har­vard at the top of the Ivy League cita­tion sweep­stakes, with Yale and Prince­ton not far behind. The Ivy League’s cita­tion rate is sig­nif­i­cant­ly high­er than the aver­age paper in ISI’s database.

It is tempt­ing to think that by direct­ing lim­it­ed resources only to cit­ed lines of research, waste would be elim­i­nat­ed and the growth knowl­edge would not be impaired. But there are dan­gers with using cita­tion index­es to direct the growth of science.

New branches and deadwood

Who can tell which sci­en­tif­ic research being done today will be the start of a fruit­ful new branch of the tree and which is des­tined to be dead­wood? Sig­nif­i­cant work may go unrec­og­nized for years fol­low­ing its pub­li­ca­tion. With­out some pos­si­bly waste­ful fund­ing sup­port, promis­ing ideas may fade before they have a chance to estab­lish them­selves in a cita­tion index.

More wor­ri­some, sci­ence might become a self-per­pet­u­at­ing aris­toc­ra­cy, rather than an open mer­i­toc­ra­cy. It may be nec­es­sary to tol­er­ate a sub­stan­tial amount of utter­ly fruit­less research in order to insure that sci­ence remains open to the gift­ed young, and to oth­ers work­ing out­side of elite research establishments.

It may be, after all, that sci­ence grows best waste­ful­ly and wild, with­out the hor­ti­cul­tur­ist’s tidy­ing hand.

Share this Musing: