[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]
Re: RECENT MANUAL MEASUREMENTS OF OA AND OAA
- To: liblicense-l@lists.yale.edu
- Subject: Re: RECENT MANUAL MEASUREMENTS OF OA AND OAA
- From: David Goodman <dgoodman@Princeton.EDU>
- Date: Fri, 20 Jan 2006 05:56:27 EST
- Reply-to: liblicense-l@lists.yale.edu
- Sender: owner-liblicense-l@lists.yale.edu
I agree with you, and I now explain why. What you are asking for is the implications of the posting, and we deliberately did not discuss it there. I now give my understanding of the meaning for what we did. The measurements we analyzed were of both OA and OAA. The OA measurement is significant, because such a number can obviously be used as a mark of progress towards OA, and I think both supporters and opponents would want to have accurate measurements of this. We found, as posted, that for all the sets of previous measurements we were trying to verify, the determination was done by a robot program which is not able to distinguish accurately between OA and non-OA. Thus, none of the %OA data based on that robot can be validly used. We did not discuss earlier %OA measurements from the same group, because we did not analyze them. The OA Advantage is not a straightforward measure of anything. It is probably meaningless on several levels. First, it is based on the robot's correct identification of OA and non-OA articles, and the robot can not do this even approximately right. Second, we have not yet discussed the details of the method by which that number is computed, but it may suffice to say that the method does not count anything in OA journals, or in any journal which included no OA articles whatsoever. { We copied the method used in order to have comparative data.) This is not yet important, because the robot can't make the correct distinctions in any case. Third, to the extent that there is any OA advantage or other advantage at the journal level, it will always have been taken into account by the Impact Factor for the journal But you are proposing a more important question. You are asking, even if the measurements were correct, even if the OAA number were correctly determined, even if it were separated from the Impact Factor, what is being measured? So, fourth, OAA is composed of many parts. Kurtz has shown best how to distinguish some of them in a closed universe of both citers and citations, but his analysis by its nature applies only in astronomy, because of the distinctive data system astronomers use. One would expect that publishing more universally available journals will result in increasing readings, and that some of the readings may eventuate as citations. But of course this cannot be predicted for any given article. The quality of the article will certainly increase both readership and citations, and authors might very appropriately pay to give this increased readership for their very best papers. This has been termed the Quality Advantage, and I think that appropriate. It again cannot be predicted for any individual article. One would want to measure this, as a functions of all the relevant variables. You have listed several; I suspect there are even more. But there does not yet seem any sound way of distinguishing the variables or making these measurements. I agree that it seems necessary to use controlled samples, and I doubt this can be done. For example, in 1953, Watson & Crick published two papers in Nature, the first giving the structure of DNA, the second in the next issue, giving the implication of the structure. The first one has been cited more than twice the second. To me, this reduces the likelihood of finding true matched pairs of articles for a controlled measurement. Perhaps someone more clever than I will find a way. But this is all hypothetical until the basic determinations can be carried out correctly. I am curious about whether there shall be any forthcoming papers or postings using these results without at least a warning of their probable invalidity. In conclusion, "OA can increase the impact of an article by > 50-250%!" is indeed not a good argument. The instruments used are worthless. The calculation used does not include all the data, however determined. The impact is a function of many variables, which have not been distinguished. There are many real reasons why all authors should publish OA, such as the public appreciation and interest gained by the easy access to the material, the greater availability to students at small institutions, and scientists' basic responsibility to publish in the widest practical manner. Dr. David Goodman (and Nisa Bakkalbasi and Kristin Antelman) Palmer School of Library and Information Science Long Island University and formerly Princeton University Library dgoodman@liu.edu ----- Original Message ----- From: Phil Davis <pmd8@cornell.edu> Date: Thursday, January 19, 2006 7:16 pm Subject: Re: RECENT MANUAL MEASUREMENTS OF OA AND OAA To: liblicense-l@lists.yale.edu > David, > > Your work on the validity of using automated robots to detect OA > articles is very important if we are to better understand the > effect of Open Access on article impact. Many of us appreciate > your work on this topic. > > The thing that I'm troubled by is that the term "Open Access > Advantage" is both loaded and vague. There are many different > types of open access (full, partial, delayed, open to developing > countries, publisher archived, self-archived, institutional > archived, etc.), so to talk about an Open Access Advantage gives > credit to many different publishing and access models at the same > time. > > Because studies of this sort are not controlled experiments, this > means that the best a researcher can do is come up with likely > causal explanations, and hope to rule others out. In the case of > higher citations for articles found as OA, article quality and > early notification are possible explanations, as are editor > selection bias, self-selection bias, and the Matthew Effect. > Hundreds of Emerald articles that were republished among two or > more journals demonstrates that simple article duplication can > also explain increased impact. All of these may be partial > causes that explain higher impact (the "Open Access Advantage"), > yet they are not limited to Open Access. > > As a consequence, I am worried about people making unqualified > statements like "OA can increase the impact of an article by > 50-250%!" The answer may be more like, "author republishing > (online and in print) may increase citation impact in some > fields, especially among highly prestigious journals and > authors". Although this is not as simple as declaring that Open > Access increases citation impact, it may be much more precise. > > --Phil Davis
- Prev by Date: ALA Midwinter: ALCTS Scholarly Communications DG
- Next by Date: Standardized metadata tags for peer reviewed and free.
- Previous by thread: Re: RECENT MANUAL MEASUREMENTS OF OA AND OAA
- Next by thread: Re: RECENT MANUAL MEASUREMENTS OF OA AND OAA
- Index(es):