Not signed in (Sign In)

Vanilla 1.1.9 is a product of Lussumo. More Information: Documentation, Community Support.

  1.  
    I am currently in the process of writing up a paper based on the survey of motivations to use MathOverflow. In the mean time I wanted to share with you the results of another study I did on MO as well as some basic demographic data from the survey.

    http://www.utpsyc.org/MathOverflowSurvey/MathOverflowStudies.php

    Thanks again to everyone who participated.
    • CommentAuthorAngelo
    • CommentTimeJan 10th 2011
     
    The total number of published papers seems to be very dubious measure of "Offline reputation". (Not that I have a better one to suggest.)
  2.  
    Thanks for the update!
  3.  

    Indeed, quite interesting. Thanks for the info.

    @Angelo- I guess you're just suggesting that the name of that variable be changed to "Number of published papers" ?

  4.  
    @Ben: my interpretation of Angelo's comment is quite different---we'll have to let him clarify, but my guess is that his point is that if you take someone like Andrew Wiles, and count his published papers, you'll get something like 20. On the other hand, I know at least one number theorist who has published over 100 papers. Am I to conclude that these other number theorists are "five times better than Wiles"?
  5.  
    On published papers:

    Yes, it is not a perfect measure of offline reputation. But as angelo says it is not obvious how to get a better measure.

    One alternative would be to measure impact factor for an author (http://en.wikipedia.org/wiki/H-index). It would have increased the amount of work considerably and probably only would have improved my measure for those who have a lot of math papers. At the moment it did not seem to be worth the effort given the potential benefit.

    I log normalized offline reputation because the bulk of individuals fall into having very few published math papers. Therefore I am not making a large distinction between authors with 20 vs 100 papers, it is roughly equivalent to 1 vs 2.
  6.  

    @Yla,

    I'd say the number of published papers is such a wildly imperfect measure of offline reputation that it is irresponsible to use the phrase "offline reputation" to refer to this number.

  7.  
    @Scott

    Can you explain why you think it is such a poor measure? Also, do you think number of published math papers is a poorer measure of offline reputation than MO Points is a measure of online reputation?
    • CommentAuthorDL
    • CommentTimeJan 10th 2011
     
    @Yla: While I would not go far as to call the phrase "offline reputation" irresponsible, I'd argue that the burden of defending metrics in your survey falls on you, not on Scott Morrison. Indeed, I would expect such a defense to be part of any publication on this issue.

    Regardless, that's an interesting survey! Though I am curious about what status those who claimed to be "status seeking" on MO hope to achieve.
  8.  
    @Yla, it's a poor measure because there's little relation between number of papers published and the esteem (most) mathematicians hold each other in. When I say "little" I mean that in a pretty strong sense -- some of the most prolific paper-producers create papers that are largely never read.

    I think the main thing people are concerned about is the language that you use. "Number of papers published" is just that, it has almost nothing to do with reputation so it's not worthwhile to use that language as it only serves to confuse your readers. So why not just call it "Number of papers published" since that's what it is you're talking about?
  9.  
    @DL I agree that it is my responsibility to defend my measure. Here is why I think it is a good enough measure for the purposes of this study.

    I think there is a misunderstanding about what kinds of MO users we are trying to distinguish between. "Number of published papers" is not an attempt to tell us whether Andrew Wiles or Grigori Perelman is the more accomplished mathematician. Professional mathematicians might be interested in such a thing, but this study is interested in the attributes of the bulk of the users of the site, not the upper tail.

    I hope we can all agree that there is a clear difference in the amount of offline reputation that an undergraduate student has, and that a full professor has, and that the number of papers published will show that. In fact, I think it's likely that we could run the same analysis with offline reputation being a binary measure of whether or not the user had any published papers, and get similar results.

    As far as "irresponsibility", it is common practice within psychology to try to measure a (ill defined and difficult to measure) feature of the real world by making an operational definition, and then naming the measure by the feature of the real world. Others can then argue about how good that operational definition is, but nobody thinks or claims that an operational definition completely captures the real world feature, or that it should be interpreted as the real world feature.

    In the context of communities of user generated content this is the first paper with any kind of measure of offline reputation. Perhaps it only explains 10% of the "Platonic ideal offline reputation", but within my field it is standard to call it "offline reputation" even in this case. Apologies for any misunderstandings. I am open to any suggestions of other ways to measure offline reputation.

    @Ryan Budney "offline reputation" is something people care about in the context of other communities (e.g. Yahoo! Answers, Wikipedia). "Number of papers" is not generalizable to these other contexts so it is less interesting within my field.
  10.  

    If you wanted, you could just sort people on MO by their status (undergraduate, graduate student, post-doc, tenure-track faculty, full professor, interested outsider, etc.). I think the point is that quality is much more important than the number of papers (e.g. in some fields people tend to publish many more papers than others by nature of the field rather than by being better mathematicians).

    There are plenty of undergraduates that publish papers (e.g. through REUs), but they probably don't have the "offline reputation" (or experience, or ability to answer MO questions) that a fourth-year graduate student at a top school that (for whatever reason) never published would.

  11.  

    @Yla: I think people's concern here (as well as mine) is the following: the term "offline reputation" is evocative enough to connote a certain meaning: namely the degree to which a certain MO user is esteemed by the mathematical community. But in your work, you do not give any precise (or imprecise!) definition of it other than equating it with the number of papers published. This does seem problematic.

    "I hope we can all agree that there is a clear difference in the amount of offline reputation that an undergraduate student has, and that a full professor has, and that the number of papers published will show that."

    No, not yet. First we need to agree on what offline reputation means. If it actually means "number of published papers", yes, it will show that. But as I said above, you seem to be relying on some more vague idea of what the term means. One can find undergraduates with more published papers than full professors. For instance, Sophie Morel is a full professor at Harvard University and has, according to MathSciNet, one published paper. So no, your implicit equation is really not so clear.

    "In fact, I think it's likely that we could run the same analysis with offline reputation being a binary measure of whether or not the user had any published papers, and get similar results."

    Then perhaps you should try it and see.

    "As far as "irresponsibility", it is common practice within psychology to try to measure a (ill defined and difficult to measure) feature of the real world by making an operational definition, and then naming the measure by the feature of the real world. Others can then argue about how good that operational definition is, but nobody thinks or claims that an operational definition completely captures the real world feature, or that it should be interpreted as the real world feature."

    The word "irresponsibility" is a little bit strong here. But I don't entirely agree with what you say: that which is ill-defined deserves more efforts at attempted definitions and descriptions, not less. Similarly with things which are difficult to measure. In other words, it seems professional to explicitly address and argue for your equation of "offline reputation" and number of published papers. Moreover, I think the portion of your paper when you set off a red boxed number one and write "Offline reputation: Number of published math papers" is definitely a claim that your operational definition captures the real world feature.

    I do think it is facile to present "measurements" without carefully addressing the extent to which the quantity that you're recording is actually a measurement of your phenomenon. To say "others can then argue" about it is kind of a copout, because that's not how things work in real life or even in academic life: academics have been told that "they can argue about" the actual meaning of student evaluation scores, impact factors and NRC rankings. Well, anyone can argue about anything but that doesn't change the fact that these statistics are now being used in making the most important decisions: hiring, promotion, tenure, funding, and so forth.

    "Perhaps it only explains 10% of the "Platonic ideal offline reputation", but within my field it is standard to call it "offline reputation" even in this case."

    Again, you seem to be presuming (perhaps facetiously, but that's kind of the point) that there is some Platonic ideal of offline reputation. What does this term actually mean to you? Moreover, in what academic field does the "offline reputation" have a meaning so standard that it appears in papers without definition or question? [I would be interested to see references.] Could this term really be more than a few years old? You're truly okay with reporting a measurement that might explain only 10% (or less?) of what it's desired to measure? Shouldn't you at least try to do better?

    Here is a suggestion: why don't you measure offline reputation by a survey? But be sure to tell the surveyants what the term means!

    • CommentAuthorHJRW
    • CommentTimeJan 10th 2011 edited
     

    If this really is standard practice in the discipline of psychology, then I see very little purpose in browbeating Yla about it. It's important to tread very carefully when criticising another field, particularly if one isn't familiar with its usual methodology.

    • CommentAuthorAngelo
    • CommentTimeJan 10th 2011
     
    I would not have dreamed of using the term "irresponsible". However, I agree with Kevin: all that the number of your papers measures is how many papers you have published. This has, I suppose, a positive, but very weak, correlation with your reputation as a mathematician. Indeed, publishing too much can hurt your reputation as much as publishing too little. To me, "At most moderate correlations between offline and online reputation" sounds very different from "At most moderate correlations between number of published papers and online reputation". The latter is, at the very least, far less surprising.

    Now, of course, coming with a better measure is hard; but I hope I don't sound snotty if I say that if something can not be reasonably done within the constraints of a study (in this case, measuring offline reputation), perhaps should not be done.
  12.  
    In case there is any confusion, the webpage is not my paper itself, it is just a subset of the results.

    The process of operationalization is a standard method in psychology and can be found in most psychology research methods textbooks. Here is one textbook:

    http://books.google.com/books?id=j7aawGLbtEoC&lpg=PA6&dq=psychological%20research%20methods%20operationalization&pg=PA6#v=onepage&q=psychological%20research%20methods%20operationalization&f=false

    Here is a relevant quote:

    "Most social psychological researchers accept the philosophy that the specific operations and measures employed in a given research study are only partial 'representations' of the theoretical constructs of interest--and imperfect representations at that."

    Here is an explanation via Wikipedia:

    http://en.wikipedia.org/wiki/Operationalization

    For example, a common operational definition of anger is how many hot sauce drops a participant is willing to put on food for another participant.

    @angelo "Now, of course, coming with a better measure is hard; but I hope I don't sound snotty if I say that if something can not be reasonably done within the constraints of a study (in this case, measuring offline reputation), perhaps should not be done."

    I think this a difference between fields. The philosophy in psychology is to make do with the measure you have until you can find a better measure. Number of papers clearly is measuring something, because it predicts some amount of the ratings of submissions on MathOverflow. It has been helpful to hear that you think that number of papers does a poor job at distinguishing between reputations of mathematicians. Perhaps its lack of explanatory power in perceived quality of answers is because number of papers is not capturing very much of the reputation of mathematicians in the upper tail (these seem to be the people generally answering questions).

    Incidentally, I think that the other measures are also imperfect. In particular I have thought a lot about whether ratings of questions and answers are actually measuring the perceived quality of submissions. More specifically to what degree is the score for a given question or answer influenced by factors above the actual quality of the submission?
  13.  

    Number of papers clearly is measuring something, because it predicts some amount of the ratings of submissions on MathOverflow.

    Sorry to be pugnacious about this, but why then don't you call number of published papers "online reputation", rather than "offline reputation"? It seems you have a better case for this misnomer than the one you're using!

    Of course this is being a devil's advocate, it would be silly to actually call number of published papers "online reputation". It's just that it seems to me that calling it "offline reputation" is equally inappropriate as it would be to call it all sorts of other things. Surely there are some criteria for deciding when "operationalization" has gone too far? (I admit of course, that I am totally ignorant of psychologists' practice here.)

  14.  

    Reading your link to Wikpedia, I found

    One of the main critics of operationalism in social science argues that "the original goal was to eliminate the subjective mentalistic concepts that had dominated earlier psychological theory and to replace them with a more operationally meaningful account of human behavior. But, as in economics, the supporters ultimately ended up "turning operationalism inside out" (Green 2001, 49). "Instead of replacing 'metaphysical' terms such as 'desire' and 'purpose'" they "used it to legitimize them by giving them operational definitions." Thus in psychology, as in economics, the initial, quite radical operationalist ideas eventually came to serve as little more than a "reassurance fetish" (Koch 1992, 275) for mainstream methodological practice."
    

    and indeed it seems to me that that's what's going on here. Rather than replacing the fuzzy concept of "offline reputation" that one might initially think was of interest with the "operationally defined" concept of "number of published papers", you're conflating the two.

  15.  

    @Yla,

    responding to your earlier question

    Also, do you think number of published math papers is a poorer measure of offline reputation than MO Points is a measure of online reputation?

    could I first ask for clarification of what you mean by "online reputation" and "offline reputation" in the first place? I'm not really sure how I would try to define these separately. Perhaps we could try to do this as the respect held for someone on account of the actions performed in either an "offline" or "online" context. Thus while I'm deeply impressed by, say, T, as a mathematician, in my regard he has accrued some "offline" reputation when I've seen him give excellent public lectures, and some "online" reputation when he's posted great articles on his blog. But if I read one of his papers via the arxiv, and like it, does that contribute to his "online" or "offline" reputation? And if a trusted colleague tells me via email "Wow, T's latest result is fabulous", which is that?

    I suspect that with a definition of online and offline reputations along these lines, I would argue that MO reputation is an extremely good proxy for "online reputation", but merely because the main other available source of it is writing blog posts, which relatively very few mathematicians do.

  16.  

    I have studied little psychology, but enough to know that "anger" is a standard theoretical concept in psychology and that "hot sauce" does not appear in the discussion of anger as a theoretical concept. I agree with Scott Morrison that there seems to be a conflation of theoretical concepts and measurable phenomena going on here.

    I repeat my request for pointers to literature on the theoretical concept of "offline reputation", which Ms. Tausczik said is a standard term in her field. Without this, I don't see how it is possible to discuss the validity of counting papers as a measuring tool.

  17.  

    "Number of papers clearly is measuring something, because it predicts some amount of the ratings of submissions on MathOverflow."

    Yes, number of papers is measuring something: the number of papers. You have collected this data and analyzed its correlation with ratings of MO submissions: sounds good. This is what has actually been done. I don't see the added value in equating "number of papers" with "offline reputation", as neither a theoretical framework for the latter is being presented nor an argument for the validity of the former as a measurement for the latter. You could explain that you chose to measure and analyze the number of papers out of a feeling that is capturing something about the status of the participants in the mathematical community, but that's about it.

    You mention the problem that offline reputation is a concept which is being used in more general contexts where "number of papers" doesn't make sense. I understand and respect that concern, but I don't see how that just picking something and calling it offline reputation contributes to the solution.

    • CommentAuthorDL
    • CommentTimeJan 10th 2011
     
    I have read papers which use operational definitions of things like "friendliness" or other vague terms, essentially in the way that Yla seems to be using "number of published papers" as a proxy for whatever "offline reputation" is supposed to be. So while I share some of the philosophical qualms stated by many of the others here, my (very limited) personal experience is that Yla correctly describes the usual practice of her field.

    Perhaps I am wrong (and Pete, Scott, feel free to correct me if I am) but my understanding of these objections is that they in large part stem from the fact that few mathematicians (I suspect) care very much about the metric, "number of published papers." So it's probably not a good metric for whatever "offline reputation" might mean in a community of mathematicians.

    In any case, perhaps this discussion would be better served by moving it to email or another less public forum? It seems to at this point be a discussion about experimental methodology and the community standards of mathematicians and psychologists, rather than about MO. In particular the vast number of people on one side of the issue (regardless of my general agreement with that side), makes me worry that this forum is being somewhat uncharitable to Yla, who seems to have good intentions toward MO--I certainly found this survey interesting.
    • CommentAuthorEmerton
    • CommentTimeJan 11th 2011
     

    Dear DL,

    Thanks for making this comment. In particular, I support the sentiment of your third paragraph.

    Regards,

    Matthew

    • CommentAuthorAngelo
    • CommentTimeJan 11th 2011
     
    I fully agree with DL.
  18.  
    I also strongly agree with DL.
  19.  
    I think Yla does a good job of justifying her practices. I agree with DL and Henry Wilton. I can imagine all sorts of situations where it would be extremely inappropriate to use [number of papers published] as a stand-in for [offline reputation], but this isn't one of them.
  20.  

    Point taken! :-)

  21.  

    I'm way too late to the party. The point I was making and I thought Angelo was as well is that it's not harder or substantially longer to write "number of published papers" instead of "offline reputation." Maybe this is just the mathematician in me speaking, but why be vague when you can be precise with no more effort?

    • CommentAuthorEmerton
    • CommentTimeJan 12th 2011
     

    Dear Ben,

    I think that Yla already explained that the desired variable is "offline reputation", and that "number of published papers" is being used an an easily measurable proxy for this.

    Cheers,

    Matt

  22.  

    Yes, but "number of published papers" is what was measured. Why not let your readers decide how good a proxy for offline reputation that is?

    • CommentAuthorEmerton
    • CommentTimeJan 13th 2011
     

    Dear Ben,

    My impression is that this would be moving in the direction of presenting pure raw data, and leaving all interpretation and analysis to the reader. I think the goal of a study like this is to provide interpretation and analysis, not simply to provide underlying raw data. Obviously, one can question the interpretation and analysis (which is what you are doing); I am just trying to posit a reason (based on my own, limited, impressions of what is happening here) why Yla may not want to do what you are suggesting (perhaps at the expense of leaving the interpretation and analysis open to some criticism on this point).

    Cheers,

    Matt

    • CommentAuthormarkvs
    • CommentTimeJan 13th 2011
     
    Instead of total number of publications one can count the total number of citations (say, from google scholar). It is a little more accurate than the number of publications, although google scholar gives incorrect results sometimes. There are also lots of indices like the h-index, hc-index, g-index, and so on which you can get using Harzing's "Publish or perish" software. These are far from perfect but are certainly used in promotion and hiring cases.
    • CommentAuthorRyan Budney
    • CommentTimeJan 13th 2011 edited
     
    A problem with Google Scholar is it does not always find the right person. For example, there are two people with the name John Harper that are both algebraic topologists. Some names are particularly bad for these kinds of issues, like Cohen.
  23.  

    Also, this was a short survey in which people self-reported their number of publications. You can't really expect people to go compute their h-index. Not to mention that the survey is already done. Suggestions about how to design the survey are way too late. Suggestions about presenting the results of said survey at least have some chance of being constructive.

    • CommentAuthormarkvs
    • CommentTimeJan 13th 2011
     
    I think people may want to compute their own h-indices because those are really used in promotional cases. Google scholar and "Publish or perish" are not easy to use, but one can in principle compute the h-index using these tools. As for the current survey, I agree that the total number of publications is not a reliable indicator of the off-line reputation. Moreover, there is no such thing as an objective "off-line reputation" because the reputation order is only partial and depends on who is ordering. Nevertheless there are many pairs of mathematicians $(A,B)$ for which almost everybody agrees that $A\gg B$ in any off-line reputation order. If you consider those pairs and the corresponding MO reputations you can easily conclude that there is very little correlation between MO and off-line reputations, as the survey claims. For example, M. Gromov and J.-P. Serre have combined MO reputation 0. This is not a big surprise: I think that in applied statistics, it is quite common that unreliable methods produce reliable results.