Oddbean new post about | logout
 One of the most remarkable things to me about all these studies that turn out to have misused data, etc., is that most researchers apparently keep incredibly poor records; there's no audit history for changes; Excel is somehow the gold standard. I guess that's true in business, but I somehow thought there would be university, journal, and professional standards that required very specific one-way archiving of data, etc. 
 @bb1a99d2 😬 🤐 
 @cef5b5fc I am surely naive but after this many years, it's incredible 
 @bb1a99d2 From my experience in industry, keeping intermediate copies of data is prohibitively expensive, even if you only keep one copy per day. 
 @eab3a90d I don't think most research data in social sciences is bulky or changes day to day! 
 @bb1a99d2 
Every PI receiving a grant is their own little small business and it is a free for all. Depending on the nature of the grant, their may be some rudimentary training in ethics (i.e. a training grant), but most training on data retention, figure production, and writing is left to the PI. Some journals do have articles explaining acceptable image adjustments, but they are few and far between. 
 @bb1a99d2 The university standard for specific, one-way archiving of data last I was there was called "Final_Final_Final_V2". 
 @bb1a99d2 @ed044e85 ditto journalism. I’ve been thinking about data libraries, version control and reproducibility a bit lately (and for my whole career). The combination of scant interest, requirement for *some* specialist skill, and poor tooling makes it a challenging task https://elvery.net/drzax/building-and-managing-a-personal-professional-data-library/ 
 @bb1a99d2 we're trying to improve things but there's a lot of work to do
https://osf.io/4pd9n/ 
 @bb1a99d2 
There are a few journals that actually check code for reproducibility - this is one of the pioneers: https://odum.unc.edu/archive/journal-verification/