Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
We generally refer to "evaluation" instead of "refereeing" because The Unjournal does not publish work; it only links, rates, and evaluates it.
For more information about what we are asking evaluators to do, see:
We follow standard procedures, considering complementary expertise, interest, and cross-citations, as well as checking for conflicts of interest. (See our internal guidelines for .)
We aim to consult those who have first.
We favor evaluators with a track record of careful, in-depth, and insightful evaluation — while giving ECRs a chance to build such a record.
For several reasons... (for more discussion, see )
It's equitable, especially for those not getting "service credit" for their refereeing work from their employer.
Paying evaluators can reduce adverse selection and conflicts of interest —arguably inherent to the traditional process where reviewers work for free.
Yes, we allow evaluators to choose whether they wish to remain anonymous or "sign" their evaluations. See .
To limit this concern:
You can choose to make your evaluation anonymous. You can make this decision from the outset (this is preferable) or later, after you've completed your review.
Your evaluation will be shared with the authors before it is posted, and they will be given two weeks to respond before we post. If they cite what they believe are any major misstatements in your evaluation, we will give you the chance to correct these.
It is well-known that referee reports and evaluations are subject to mistakes. We expect most people who read your evaluation will take this into account.
We will put your evaluation on and give it a DOI. It cannot be redacted in the sense that this initial version will remain on the internet in some format. But you can add an addendum to the document later, which we will post and link, and the DOI can be adjusted to point to the revised version.
See the FAQ as well as the .
We have two main ways that papers and research projects enter the Unjournal process:
Authors submit their work; if we believe the work is relevant, we assign evaluators, and so on.
We select research that seems potentially influential, impactful, and relevant for evaluation. In some cases, we request the authors' permission before sending out the papers for evaluation. In other cases (such as where senior authors release papers in the prestigious
For either track, authors are invited to be involved in several ways:
Authors are informed of the process and given an opportunity to identify particular concerns, request an embargo, etc.
Evaluators can be put in touch with authors (anonymously) for clarification questions.
Authors are given a two-week window to respond to the evaluations (this response is published as well) before the evaluations are made public. They can also respond after the evaluations are released.
If you are writing a signed evaluation, you can share it or link it on your own pages. Please wait to do this until after we have given the author a chance to respond and posted the package.
Otherwise, if you are remaining anonymous, please do not disclose your connection to this report.
Going forward:
We may later invite you to write and evaluate more about this piece of research . . .
. . . and to help us judge prizes (e.g., the ).
We may ask if you want to be involved in replication exercises (e.g., through the
The evaluations provide at least three types of value, helping advance several paths in our :
For readers and users: Unjournal evaluations assess the reliability and usefulness of the paper along several dimensions—and make this public, so other researchers and policymakers can learn from them.
For careers and improving research: Evaluations provide metrics of quality. In the medium term, these should provide increased and accelerated career value, improving the research process. We aim to build metrics that are credibly comparable to the current "tier" of journal a paper is published in. But we aim to do this better in several ways:
See "what
See our .
We still want your evaluation and ratings. Some things to consider as an evaluator in this situation.
A paper/project is not only a good to be judged on a single scale. How useful is it, and to who or what? We'd like you discuss its value in relation to previous work, it’s implications, what it suggests for research and practice, etc.
Even if the paper is great...
Would you accept it in the “top journal in economics”? If not, why not?
Would you hire someone based on this paper?
Would you fund a major intervention (as a government policymaker, major philanthropist, etc.) based on this paper alone? If not, why not
We can use payment as an incentive for high-quality work, and to access a wider range of expertise, including people not interested in submitting their own work to The Unjournal.
You can add an addendum or revision to your evaluation later on (see below).
As a general principle, we hope and intend always to see that you are fairly compensated for your time and effort.
More quickly, more reliably, more transparently, and without the unproductive overhead of dealing with journals (see '')
Allowing flexible, , thus improving the research process, benefiting research careers, and hopefully improving the research itself in impactful areas.
Feedback and suggestions for authors: We expect that evaluators will provide feedback that is relevant to the authors, to help them make the paper better.
Can you quantify your confidence in these 'crucial' results, and their replicability and generalizability to other settings? Can you state your probabilistic bounds (confidence or credible intervals) on the quantitative results (e.g., 80% bounds on QALYs/DALYs/or WELLBYs per $1000)
Would any other robustness checks or further work have the potential to increase your confidence (narrow your belief bounds) in this result? Which?
Do the authors make it easy to reproduce the statistical (or other) results of the paper from shared data? Could they do more in this respect?
Communication: Did you understand all of the paper? Was it easy to read? Are there any parts that could have been better explained?
Is it communicated in a way that would it be useful to policymakers? To other researchers in this field, or in the general discipline?
See sections below
: An overview of what we are asking; payment and recognition details
: The Unjournal's evaluation guidelines, considering our priorities and criteria, the metrics we ask for, and how these are considered.
Other sections and subsections provide further resources, consider future initiatives, and discuss our rationales.
The Unjournal is an independent 501(c)(3) nonprofit organization based in the United States.
Evaluators: To claim your payment for evaluation work, please complete .
You will receive your payment via a transfer (they may ask you for your bank information if you don't have an account with them).
We aim to process all payments within one week.
Confidentiality: Your identity will only be visible to The Unjournal administrators for the purposes of making this payment. If you have concerns about the payment process, please contact us at .
Team members and contractors: Please refer to the internal Coda workspace for expense submission procedures.
is making research better by evaluating what really matters. We aim to make rigorous research more impactful and impactful research more rigorous.
The academic journal system is out-of-date. It wastes resources, discourages innovation, and encourages rent-seeking.
The Unjournal is not a journal. We don’t publish research. Instead, we commission (and pay for) open, rigorous expert evaluation of publicly-hosted research. We make it easier for researchers to get feedback and credible ratings of their work, so they can focus on doing better research rather than journal-shopping.
We currently focus on quantitative work that informs global priorities, especially in economics, policy, and social science. Our evaluation packages help policymakers and practitioners understand which research to trust and how to use it. These provide open critique and discussion to benefit students and the research community.
Click on the cards below to find out more about our mission, organizational structure, and ways to collaborate, or 'ask or search our page' for answers to your questions.
You can also press ⌘k or control-k to search or query our site.
For the 2024-25 cycle, we are focusing on evaluator recognition, with $6,500 in prizes for evaluators who provided exceptionally credible, insightful, and detailed evaluations. We are currently finalizing the winners and will announce them soon.
Given limited funds and a relatively small pool of author-submitted work during 2024-25, we are shifting from monetary "author prizes" to recognition-based awards highlighting two dimensions:
Flowing Water Scholar: Recognizing the most valuable author engagement and response—openness, epistemic modesty, willingness to update, growth mindset, and constructive dialogue with evaluators.
Polaris Research: Recognizing the strongest and most useful research, based on highly credible and detailed evaluation manager reports, careful evaluations, and high "merited journal tier" ratings.
We aim to establish an ongoing author prize with a monetary component for 2026 and beyond. This will particularly reward authors who submit their work and engage meaningfully with the evaluation process. The structure and scale will depend on our funding situation.
To be considered for future recognition:
for Unjournal evaluation
Engage with the evaluation process by responding substantively to evaluator feedback
Sign up as an
Updates will be announced on and our . Contact with questions.
Our inaugural Impactful Research Prize was awarded in early 2024. See for the full announcement.
Winners:
First Prize ($2,500): Takahiro Kubo and co-authors for ""
Second Prize ($1,000): Johannes Haushofer and co-authors for ""
We also recognized exceptional evaluators: Phil Trammell, Hannah Metzler, Alex Bates, and Robert Kubinec.
The prize winners were selected through a multi-step, collaborative process involving both the management team and the advisory board. Selection criteria included research quality and credibility, potential for real-world impact, and authors' engagement with The Unjournal's evaluation process.
Initial Evaluation: All papers evaluated by The Unjournal were eligible. Evaluations and ratings from external evaluators informed the initial shortlisting.
The pilot prize was announced in late 2022 as part of The Unjournal's first year. First-prize winners received $2,500 (increased from the original $2,000 thanks to a contribution from EAecon), and runners-up received $1,000.
Prize winners had the opportunity to present their work at an online seminar and prize ceremony co-hosted by The Unjournal, Rethink Priorities, and EAecon.
Eligibility: All work evaluated by The Unjournal was eligible. Engagement with the evaluation process, including responding to evaluator comments, was a factor in determining winners.
Note: In a subsection below, , we outline the basic requirements for submissions to The Unjournal.
Meeting and Consensus: A prize committee of four volunteers discussed shortlisted papers and allocated points among 10 candidates to narrow to a shortlist of five.
Point Voting: The shortlist was shared with all management and advisory board members, who allocated up to 100 points among the finalists.
Special Considerations: At least one winner needed to be author-submitted or show substantial engagement. Early-career researchers received slight preference.
Final Selection: Prizes went to the papers with the most points.
A "curated guide" to this GitBook; updated July 2025
You can now ask questions of this GitBook using a chatbot: click the search bar or press cmd-k and choose "ask Gitbook."
For authors, evaluators, and others interested in collaboration.
Write-ups of the main points for a few different audiences
Important benefits of journal-independent public evaluation and The Unjournal's approach, with links to deeper commentary
Description of our evaluation process
Groups we work with; comparing approaches
What research are we talking about? What will we cover?
These are of more interest to people within our team; we are sharing these in the spirit of transparency.
A "best feasible plan" for going forward
Successful proposals (ACX, SFF), other applications, initiatives
Key resources and links for managers, advisory board members, staff, team members and others involved with The Unjournal project.
Note: we have moved some of this "internal interest content" over to our Coda.io knowledge base.
The Unjournal encourages better research by making it easier for researchers to get feedback and credible ratings. We coordinate and fund public journal-independent expert evaluation of hosted papers and dynamically presented projects. We publish evaluations, ratings, manager summaries, author responses, and links to evaluated research on our PubPub page.
As the name suggests, we are not a journal!
We work independently of traditional academic journals. We're building an open platform and a sustainable system for feedback, ratings, and assessment. We're currently focusing on quantitative work that informs global priorities in economics, social science, and policy.
How to get involved?
We're looking for research to evaluate, as well as evaluators. You can submit your own research here, or suggest research using this form. We offer financial prizes for suggesting research we end up evaluating. If you want to be an evaluator, apply here. You can use the same form to express your interest in joining our management team, advisory board, or reviewer pool. For more information, see our how to get involved guide.
Why The Unjournal? Peer review is great, but conventional academic publication processes are wasteful, slow, and rent-extracting. They discourage innovation and prompt researchers to focus more on "gaming the system" than on the quality of their research. We will provide an immediate alternative, and at the same time, offer a bridge to a more efficient, informative, useful, and transparent research evaluation system.
Does The Unjournal charge any fees?
No. We're a US-registered tax-exempt 501(c)(3) nonprofit, and we don't charge fees for anything. We compensate evaluators for their time and we even award prizes for strong research and evaluation work, in contrast to most traditional journals. We do so thanks to funding from the Effective Altruism Infrastructure Fund, Long-Term Future Fund and Survival and Flourishing Fund.
At some point in the future, we might consider sliding-scale fees for people or organizations submitting their work for Unjournal evaluation, or for other services. If we do this, it would simply be a way to cover the compensation we pay evaluators and to cover our actual costs. Again, we are a nonprofit and we will stay that way.
Research submission/identification and selection: We identify, solicit, and select relevant research work to be hosted on any open platform in any format that can gain a time-stamped DOI. Authors are encouraged to present their work in the ways they find most comprehensive and understandable. We support the use of dynamic documents and other formats that foster replicability and open science. (See: the benefits of dynamic docs).
Paid evaluators (AKA "reviewers"): We compensate evaluators (essentially, reviewers) for providing thorough feedback on this work. (Read more: Why do we pay?)
Eliciting quantifiable and comparable metrics: We aim to establish and generate credible measures of research quality and usefulness. We benchmark these against traditional previous measures (such as journal tiers) and assess the reliability, consistency, and predictive power of these measures. (Read more: )
Public evaluation: We publish the evaluation packages (including reports, ratings, author responses, and manager summaries) on our . Making evaluation public facilitates dialogue, and supports transparency, impact, understanding, and
Linking, not publishing: Our process is not "exclusive." Authors can submit their work to a journal (or other evaluation service) at any time. This approach also allows us to benchmark our evaluations against traditional publication outcomes.
Prizes: We award financial prizes and hold public events to recognize the most credible, impactful, useful, and insightful research, as well as strong engagement with our evaluation process.
Transparency: We aim for maximum transparency in our processes and judgments.
For example, this is closely related to ELife's "Publish, Review, Curate" model; see their updated (Oct 2022) model here. COS is also building a "lifecycle journal". PREReview promotes public journal-independent evaluation. However, we cover a different research focus and make some different choices, discussed below. We also discuss other Parallel/partner initiatives and resources, many of whom we are building partnerships with. However, we think we are the only group funded to do this in this particular research area/focus. We are also taking a different approach to previous efforts, including funding evaluation (see Why pay evaluators (reviewers)?) and asking for quantified ratings and predictions (see Guidelines for evaluators).
The Unjournal is funded by grants from philanthropic organizations focused on effective research and global priorities. We are actively applying for grants to sustain and expand our work.
Current and past funders include:
(LTFF) — our initial ACX grant was passed to LTFF
EA Funds (EAIF; Pivotal Questions project)
We share our grant applications in the spirit of transparency — see our section, including both successful and unsuccessful applications.
Academics and funders have complained about this stuff for years and continue to do so every day on social media . We are fairly confident our critiques of the traditional review and publication process will resonate with most readers.
So why haven't academia and the research community been able to move to something new? There is a difficult collective action problem. Individual researchers and universities find it risky to move unilaterally. But we believe we have a good chance of finally changing this model and moving to a better equilibrium. We will:
Take risks: Many members of The Unjournal management are not traditional academics; we can stick our necks out. We are also recruiting established senior academics who are less professionally vulnerable.
Bring in new interests, external funding, and incentives: There are a range of well-funded and powerful organizations—such as the Sloan Foundation and Open Philanthropy—with a strong inherent interest in high-impact research being reliable, robust, and reasoning-transparent. This support can fundamentally shift existing incentive structures.
Allow less risky "bridging steps": As noted above, The Unjournal allows researchers to submit their work to traditional journals. In fact, this will provide a benchmark to help build our quantitative ratings and demonstrate their value.
Communicate with researchers and stakeholders to make our processes easy, clear, and useful to them.
Make our output useful, in the meantime: It may take years for university departments and grant funders to incorporate journal-independent evaluations as part of their metrics and reward systems. The Unjournal can be somewhat patient: our evaluation, rating, feedback, and communication are already providing a valuable service to authors, policymakers, and other researchers.
Leverage new technology: A new set of open-access and AI-powered tools makes what we are trying to do easier, and makes formats other than static PDFs more useful every day.
Reward early adopters with prizes and recognition: We can replace "fear of standing out" with "fear of missing out." In particular, authors and research institutions that commit to publicly engaging with evaluations and critiques of their work should be commended and rewarded. And we are doing this.
This GitBook is a knowledge base that supplements our main public page, unjournal.org. It serves as a platform to organize our ideas and resources and track our progress towards our dual objectives:
Making "peer evaluation and rating" of open projects into a standard high-status outcome in academia and research, specifically within economics and social sciences. This stands in contrast to the conventional binary choice of accepting or rejecting papers to be published as PDFs and other static formats.
Building a cohesive and efficient system for publishing, accruing credibility, and eliciting feedback for research aligned with effective altruism and global priorities. Our ultimate aim is to make rigorous research more impactful, and impactful research more rigorous.
Explore our evaluation data and research through these resources:
Read our published evaluations, ratings, and author responses
Explore evaluation metrics and trends interactively
Detailed data analysis and methodology documentation
Browse research we've prioritized or are considering for evaluation
See Content overview
You can also search and query this Gitbook (press control-K or command -k)
(see our In a nutshell) wants your involvement, help, and feedback. We offer recognize and compensate people for their time and effort.
Read our evaluation packages here, add comments, cite the evaluations in your work, let us know how you found them useful and how we can improve.
Make Unjournal evaluations part of your process, using them to consider your funding and policy choices, encouraging your team or department to submit work for public evaluation, and make this part of your hiring, promotion, and research funding criteria.
Submit your research for Unjournal to evaluate , or by contacting . Suggest research for us to assess using .
Join our team: Complete (about 3–5 min) to apply for our...
Evaluator pool: to be eligible to be commissioned and paid to evaluate and rate research, mainly in quantitative social science and policy
Do an Independent Evaluation to build your portfolio, receive guidance, and be eligible for promotion and prizes. See details
Suggest for us to focus on
Give us feedback: What did you find useful and how did you use the work? Is anything unclear? What could be improved? Email contact@unjournal.org, and we will respond. We will offer rewards for the most useful suggestions.
is the founder and co-director of The Unjournal. The organization is currently looking for field specialists and evaluators, as well as suggestions for relevant work for The Unjournal to evaluate.
The Unjournal is building a system for credible, public, journal-independent feedback and evaluation of research.
Identify, invite, or select contributions of relevant research that is publicly hosted on any open platform or archive in any format.
We maintain an open call for participants for several different roles:
(involving honorariums for time spent)
members (no time commitment)
(who will often also be on the Advisory Board)
You can express your interest (and enter our database) .
We're interested in researchers and research-users who want to help us prioritize work for evaluation, and manage evaluations, considering
... research in any social science/economics/policy/impact-assessment area, and
... research with the potential to be among the most globally-impactful.
Some particular areas where we are hoping to expand our expertise (as of Sept. 2024) include:
- Biological & pandemic risk
We will reach out to evaluators (a.k.a. "reviewers") on a case-by-case basis, appropriate for each paper or project being assessed. This is dependent on expertise, the researcher's interest, and a lack of conflict of interest.
Time commitment: Case-by-case basis. For each evaluation, for the amount of time to spend.
Compensation: We pay evaluators for their time and effort, targeting an average of $350 per evaluation including prizes and incentives (and aim to return to $450+ with future funding). .
Who we are looking for: We are putting together a list of people interested in being an evaluator and doing paid referee work for The Unjournal. We generally prioritize the pool of evaluators who signed up for our database before reaching out more widely.
Interested? Please fill out (about 3–5 min, same form for all roles or involvement).
Ready to get started doing evaluations and building a track record? See our new initiative, offering prizes and recognition for the best work. You can evaluate work in our , or suggest and evaluate work.
We are looking for high-quality, globally pivotal research projects to evaluate, particularly those embodying open science practices and innovative formats. We are putting out a call for relevant research. Please suggest research . (We offer bounties and prizes for useful suggestions – see note.) For details of what we are looking for, and some potential examples, and accompanying links.
You can also put .
If you are interested in discussing any of the above in person, please email us () to arrange a conversation.
Note: This is under continual refinement; see our for more details.
Field specialist teams: help identify, prioritize, and manage research evaluation in a particular field or cause area. A related lower-commitment role: help suggest, prioritize, and discuss research
Management team or advisory board, to be part of our decision-making
Pay evaluators to give careful feedback on this work, with prizes and incentives for strong evaluation work.
Elicit quantifiable and comparable metrics of research quality as credible measures of value (see: ). Synthesize the results of these evaluations in useful ways.
Publicly post and link all reviews of the work. Award financial prizes for the work judged strongest.
Allow evaluators to choose if they wish to remain anonymous or to "sign" their reviews.
Aim to be as transparent as possible in these processes.
A pool of Evaluators (who will be paid for their time and their work; we also draw evaluators from outside this pool)
- Long-term trends, demography
- Macroeconomics/growth/(public) finance
- Quantitative political science (voting, lobbying, etc.)
- Social impact of new technology (including AI)
January 2026: We are not currently hiring for specific staff positions. However, we hope to hire for similar roles in the future, depending on grant funding and other considerations.
People are welcome to submit their resume and expression of interest for future consideration using the form below.
We continue to recruit evaluators and field specialists on an ongoing basis. See our .
To indicate your potential interest in roles at The Unjournal, please fill out and link (or upload) your CV or webpage.
If you already filled out this form for a role that has changed titles, don't worry. You will still be considered for relevant and related roles in the future.
If you add your name to this form, we may contact you to offer you the opportunity to do paid project work.
If you are interested in conducting paid research evaluation for The Unjournal, or in joining our advisory board, please complete the form linked .
Feel free to contact contact@unjournal.org with any questions.
Not looking for a job but want to participate? See our page for other opportunities including:
Submitting or suggesting research for evaluation
Becoming an evaluator
Joining as a field specialist
The Unjournal, a US-registered 501(c)(3) nonprofit, is an equal-opportunity employer and contractor. We are committed to creating an inclusive environment for all employees, volunteers, and contractors. We do not discriminate on the basis of race, color, religion, gender, gender identity or expression, sexual orientation, national origin, genetic information, disability, age, or veteran status.
See our data protection statement .
Communications, Writing, and Public Relations Specialist
See Research & operations-linked roles & projects for details on:
Research and Evaluation Specialist (RES)
See Standalone project: Impactful Research Scoping (temp. pause) for details on:
Impactful Research Scoping project
Dec 2024: We are still looking to bring in more field specialists to build our teams in a all areas, but particularly in the quantitative social science and economics/behavior modeling of catastrophic risks, AI governance and safety.
In addition to the "work roles," we are looking to engage researchers, research users, meta-scientists, and people with experience in open science, open access, and management of initiatives similar to The Unjournal.
We are continually looking to enrich our general team and board, including our , These roles come with some compensation and incentives.
(Please see links and consider submitting an expression of interest).
Our strategic direction and goals
What is The Unjournal? See our summary.
The Unjournal's pilot phase was completed in August 2023. We are now focused on:
Scaling evaluation coverage — Expanding the volume and range of research we evaluate
Building field specialist teams — Recruiting experts to monitor and prioritize research in key areas
Improving evaluation processes — Refining our metrics, guidelines, and workflow
See the vision and broad plan presented :
I (David Reinstein) am an economist who left UK academia after 15 years to pursue a range of projects (). One of these is :
The Unjournal (with funding from the and the Survival and Flourishing Fund) organizes and funds public-journal-independent feedback and evaluation, paying reviewers for their work. We focus on research that is highly relevant to global priorities, especially in economics, social science, and impact evaluation. We encourage better research by making it easier for researchers to get feedback and credible ratings on their work.
We are looking for your involvement...
We can provide credible reviews and ratings that have value as measures of research quality comparable to (or better than) traditional journal systems.
We identify important work that informs global priorities.
We boost work in innovative and transparent and replicable formats (especially dynamic documents).
Authors engage with our process and find it useful.
(As a push) Universities, grantmakers, and other arbiters assign value to Unjournal ratings.
See our original grant proposal for the detailed 12-month plan we followed during the pilot.
Key elements completed:
✅ Built founding committee of experienced researchers
✅ Defined scope and principles
✅ Set up platforms (moved from Kotahi/Sciety to PubPub)
✅ Created submission and evaluation rules
✅ Completed pilot evaluations (10 papers, 21 evaluations)
We want researchers who are interested in doing evaluation work for The Unjournal. We pay an average of $200-$400 per complete and on-time evaluation, and we award monetary prizes for the strongest work. Right now we are particularly looking for economists and people with quantitative and policy-evaluation skills. We describe what we are asking evaluators to do here: essentially a regular peer review with some different emphases, plus providing a set of quantitative ratings and predictions. Your evaluation content would be made public (and receive a DOI, etc.), but you can choose if you want to remain anonymous or not.
To sign up to be part of the pool of evaluators or to get involved in The Unjournal in other ways, please fill out this brief form or email contact@unjournal.org.
We welcome suggestions for particularly impactful research that would benefit from (further) public evaluation. We choose research for public evaluation based on an initial assessment of methodological strength, openness, clarity, relevance to global priorities, and the usefulness of further evaluation and public discussion. We sketch these criteria here, and discuss some potential examples here (see research we have chosen and evaluated at unjournal.pubpub.org, and a larger list of research we're considering here).
If you have research—your own or others—that you would like us to assess, please submit it here or contact contact@unjournal.org. Authors of evaluated papers are eligible for recognition and prizes.
We are looking for both feedback on and involvement in The Unjournal. Feel free to reach out at .
View our data protection statement
14 Jan 2025: The Unjournal is still looking to build our team and evaluator pool. Please consider the roles below and express your interest here or contact us at contact@unjournal.org.
Activities of those on the management committee may involve a combination of the following (although you can choose your focus):
Contributing to the decision-making process regarding research focus, reviewer assignment, and prize distribution.
Collaborating with other committee members on the establishment of rules and guidelines, such as determining the metrics for research evaluation and defining the mode of assessment publication.
Helping plan The Unjournal’s future path.
Time commitment: A minimum of 15–20 hours per year.
Compensation: We have funding for a $57.50 per hour honorarium for the first 20 hours, with possible compensation for additional work. will be further compensated (at roughly $300–$450 per paper).
Who we are looking for: All applicants are welcome. We are especially interested in those involved in global priorities research (and related fields), policy research and practice, open science and meta-science, bibliometrics and scholarly publishing, and any other academic research. We want individuals with a solid interest in The Unjournal project and its goals, and the ability to meet the minimal time commitment. Applying is extremely quick, and those not chosen will be considered for other roles and work going forward.
Beyond direct roles within The Unjournal, we're building a larger, more passive advisory board to be part of our network, to offer occasional feedback and guidance. There is essentially no minimum time commitment for advisory board members—only opportunities to engage. We sketch some of the expectations in the fold below.
As an AB member...
you agree to be listed on our page as being on the advisory board.
you have the option (but not the expectation or requirement) to join our Slack, and to check in once in a while.
Goals: Support open evaluation and impactful, credible research, helping The Unjournal keep on top of the latest, most relevant, influential, and globally-consequential work. Help raise awareness of The Unjournal, fostering a transition away from traditional journal peer review towards a more transparent, efficient, and rigorous approach.
Focus on a particular area of research, policy, or impactful outcome, in collaboration with other members of our area-focused teams.
Keep track of new or under-considered research with potential for impact and explain and assess the extent to which The Unjournal can add value by commissioning its evaluation. Provide a brief explanation of why each paper matters and how it aligns with The Unjournal’s approach.
Scope: Perhaps three papers per academic term; FS are encouraged to suggest more. This comes with compensation and incentives
Help track and curate this list of prioritized research, considering and discussing specific areas and issues that merit evaluation. We maintain a public database of research of potential interest .
Write and share "second opinion" assessments of this research, considering the importance and relevance for our evaluation. Vote on a short list of papers in your area to determine whether we should commission these for evaluation. Attend (very occassional) 'field group meetings' to discuss this.
Evaluation management (FS)
Field Specialists may serve as 'evaluation managers' (with additional compensation). Evaluation managers help us identify and commission evaluators for work we have prioritized, help manage this evaluation process, and write a summary report on the evaluations
Understand our mission and be ready to explain it to colleagues and (optionally) on social media.
We are organizing several teams of field specialists (and management and advisory board members). These teams will hold occasional online meetings (perhaps every 3 months) to discuss research to prioritize, and to help coordinate 'who covers what'. If team members are interested, further discussions, meetings, and seminars might be arranged, but this is very much optional.
As of Dec 2024, we have the following teams (organized around fields and outcomes)
Development economics and global health and development
You will be asked to fill out to let us know what fields, topics, and sources of research you would like to "monitor" or dig into to help identify and curate work relevant for Unjournal evaluation, as well as outlining your areas of expertise (the form takes perhaps 5–20 minutes).
This survey helps us understand when to contact you to ask if you want to be an evaluation manager on a paper we have prioritized for evaluation.
Guided by this survey form (along with discussions we will have with you, and coordination with the team), we will develop an “assignment” that specifies the area you will cover. We will try to divide the space and not overlap between field specialists. This scope can be as broad or focused as you like.
The Unjournal's field specialists choose an area they want to monitor. By this we mean that a field specialist will
Keep an eye on designated sources (e.g., particular working paper series) and fields (or outcomes or area codes), perhaps every month or so; consider new work, dig into archives
Public Leadership: It publicly signals your commitment to open science principles, innovation, and research impact.
Professional Credibility & Recognition: URAs are chosen through a selective process, and will be publicly acknowledged. URAs gain valuable experience, strengthen their CV, and position themselves for potential future opportunities with The Unjournal (e.g., Field Specialist or management roles).
Networking & Community: Build contacts with others in your field and with related interests. Stay on top of cutting-edge research, discuss and understand its value and connection to the field and to policy/impact.
See the page.
Time commitment: There is no specific time obligation—only opportunities to engage. We may also consult you occasionally on your areas of expertise. Perhaps 1–4 hours a month is a reasonable starting expectation for people already involved in doing or using research, plus potential additional paid assignments. Our document also provides some guidance on the nature of work and the time involved.
Compensation: We aim to fairly compensate people for time spent on work done to support The Unjournal, and to provide incentives for suggesting and helping to prioritize research for evaluation. See the page. will be compensated at roughly $300–$450 per project.
Who we are looking for: For the FS roles, we are seeking active researchers, practitioners, and stakeholders with a strong publication record and/or involvement in the research and/or research-linked policy and prioritization processes. For the AB, also people with connections to academic, governmental, or relevant non-profit institutions, and/or involvement in open science, publication, and research evaluation processes. People who can offer relevant advice, experience, guidance, or help communicate our goals, processes, and progress.
Interested? (about 3–5 min, using the same form for all roles).
Application criteria: Applicants will selected based on (1) their ability to assess, discuss, and communicate the global impact of research, (2) their understanding of The Unjournal approach, (3) their research skills, interests, and experience.
Low Time Commitment: This role is designed to fit into a busy academic or research schedule. You’ll stay involved at a comfortable level without overextending your commitments.
Compensation This role comes with no baseline compensation, but URAs are eligible for incentive compensation for suggesting papers, for doing high-value assessment for prioritization, and more (see ).
Expectations for URAs
Recommend 2–3 papers each academic term that you think have the potential to make a substantial impact. Provide a brief explanation of why each paper matters and how it aligns with The Unjournal’s approach. Earn compensation for your recommendations.
Provide Second Opinions & Prioritization: Complete three “second opinion” assessments each term, offering a concise evaluation of a paper’s relevance and importance. Contribute to broader discussions about research prioritization.
Vote & Engage: Participate in monthly voting on a short list of papers. Stay semi-active on Slack and Coda (with no strict requirement to check in regularly, beyond voting).
Why are we offering this role? Early career researchers and PhD students may be interested in getting involved but worried about making commitments. But getting busy students and researchers involved even minimally could keep us in touch with the cutting edge of research and help us forge collaborations in academia. So we're offering the “Unjournal Research Affiliate” role, which is similar to the Field Specialist role, but with less responsibility and time commitment.
for this and other roles.
Self-contained public release:
If you are interested in discussing any of the above in person, please email us () to arrange a conversation.
We invite you to (the same as that linked above) to leave your contact information and outline which parts of the project interest you.
Note: These descriptions are under continual refinement; see our for more details.
January 2026: We are not currently hiring for these positions. The role descriptions below are preserved as examples of positions we may hire for in the future. See the for how to express interest.
These are principally not research roles, but familiarity with research and research environments will be helpful, and there is room for research involvement depending on the candidate's interest, background, and skills/aptitudes.
The Unjournal is looking to work with a proficient writer who is adept at communicating with academics and researchers (particularly in economics, social science, and policy), journalists, policymakers, and philanthropists. As we are in our early stages, this is a generalist role. We need someone to help us explain what The Unjournal does and why, make our processes easy to understand, and ensure our outputs (evaluations and research synthesis) are accessible and useful to non-specialists. We seek someone who values honesty and accuracy in communication; someone who has a talent for simplifying complex ideas and presenting them in a clear and engaging way.
The work is likely to include:
Promotion and general explanation
Spread the word about The Unjournal, our approach, our processes, and our progress in press releases and short pieces, as well as high-value emails and explanations for a range of audiences
Make the case for The Unjournal to potentially skeptical audiences in academia/research, policy, philanthropy, effective altruism, and beyond
Keeping track of our progress and keeping everyone in the loop
Help produce and manage our external (and some internal) long-form communications
Help produce and refine explanations, arguments, and responses
Help provide reports to relevant stakeholders and communities
Making our rules and processes clear to the people we work with
Explain our procedures and policies for research submission, evaluation, and synthesis; make our systems easy to understand
Help us build flexible communications templates for working with research evaluators, authors, and others
Other communications, presentations, and dissemination
Write and organize content for grants applications, partnership requests, advertising, hiring, and more
Potentially: compose non-technical write-ups of Unjournal evaluation synthesis content (in line with interest and ability)
Most relevant skills, aptitudes, interests, experience, and background knowledge:
Understanding of The Unjournal project
Strong written communications skills across a relevant range of contexts, styles, tones, and platforms (journalistic, technical, academic, informal, etc.)
Familiarity with academia and research processes and institutions
Familiarity with current conversations and research on global priorities within government and policy circles, effective altruism, and relevant academic fields
Willingness to learn and use IT, project management, data management, web design, and text-parsing tools (such as those mentioned below), with the aid of GPT/AI chat
Further desirable skills and experience:
Academic/research background in areas related to The Unjournal’s work
Operations, administrative, and project management experience
Experience working in a small nonprofit institution
Experience with promotion and PR campaigns and working with journalists and bloggers
Proposed terms:
Project-based contract "freelance" work
$30–$55/hour USD (TBD, depending on experience and capabilities). Hours for each project include some onboarding and upskilling time.
Our current budget can cover roughly 200 hours of this project work. We hope to increase and extend this (depending on our future funding and expenses).
This role is contract-based and supports remote and international applicants. We can contract people living in most countries, but we cannot serve as an immigration sponsor.
you will be looped in for your input on some decisions surrounding The Unjournal's policies and direction. Such communications might occur once per month, and you are not obligated to respond.
you may be invited to occasional video meetings (again optional).
you are “in our system” and we may consult you for other work.
you will be compensated for anything that requires a substantial amount of your time that does not overlap with your regular work.
Economics, welfare, and governance
Psychology, behavioral science, and attitudes
Innovation and meta-science, impact of emerging technologies, catastrophic risks
Animal welfare: markets, attitudes
Environmental economics
Other teams are being organized or considered
We will compensate you for the time you spend on this process (details tbd), particularly to the extent that the time you spend does not contribute to your other work or research. (See incentives and norms here.)
Suggest/Input work into our database … papers/projects/research that seems relevant for The Unjournal to evaluate. Give some quick ‘prioritization ratings’
If you have time, give a brief on why this work relevant for UJ (impactful, credible, timely, open presentation, policy-relevant, etc) and what areas need particular evaluation and feedback
Influencing the Agenda: URAs help shape what research The Unjournal evaluates and promotes, furthering our mission of global impact, and advocating for your own priorities.
Compensation: Incentive pay for recommending and evaluating papers. Be compensated ($) for research surveying and scoping you may are already be doing. See compensation details here.\
Represent The Unjournal: Understand our mission and be ready to explain it to colleagues and on social media. Attend at least one Unjournal meeting (one hour) per year to stay connected.
Optional: Consult on evaluation management Occasionally advise on evaluator suggestions for papers in your field (additional compensation may apply).
Several expositions for different audiences, fleshing out ideas and plans
See/subscribe to
See slide deck (Link: ; offers comment access)
(Link: ; offers comment access)
Earlier presentation materials from 2022 are preserved here for reference:
Nov 2022: Version targeted towards OSF/Open Science
Earlier discussion document, aimed at EA/global priorities, academic, and open-science audiences
2021 A shorter outline posted on
January 2026: We are not currently hiring for these positions. The role descriptions below are preserved as examples of positions we may hire for in the future. See the for how to express interest.
The potential roles discussed below combine research-linked work with operations and administrative responsibilities. Overall, this may include some combination of:
Interacting with authors, recruiting, and overseeing evaluators
Synthesizing and disseminating the results of evaluations and ratings
Aggregating and benchmarking these results
Helping build and improve our tools, incentives, and processes
Curating outputs relevant to other researchers and policymakers
Doing "meta-science" work
See also our field specialist team pool and evaluator pool. Most of these roles involve compensation/honorariums. See Advisory/team roles (research, management)
Potential focus areas include global health; development economics; markets for products with large externalities (particularly animal agriculture); attitudes and behaviors (altruism, moral circles, animal consumption, effectiveness, political attitudes, etc.); economic and quantitative analysis of catastrophic risks; the economics of AI safety and governance; aggregation of expert forecasts and opinion; international conflict, cooperation, and governance; etc.
Work (likely to include a combination of):
Identify and characterize research (in the area of focus) that is most relevant for The Unjournal to evaluate
Summarize the importance of this work, its relevance to global priorities and connections to other research, and its potential limitations (needing evaluation)
Help build and organize the pool of evaluators in this area
Assist evaluation managers or serve as evaluation manager (with additional compensation) for relevant papers and projects
Synthesize and communicate the progress of research in this area and insights coming from Unjournal evaluations and author responses; for technical, academic, policy, and intelligent lay audiences
Participate in Unjournal meetings and help inform strategic direction
Liaise and communicate with relevant researchers and policymakers
Help identify and evaluate prize winners
Meta-research and direct quantitative meta-analysis (see "Project" below)
Desirable skills and experience:
Note: No single skill or experience is necessary independently. If in doubt, we encourage you to express your interest or apply.
Understanding of the relevant literature and methodology (to an upper-postgraduate level) in this field or a related field and technical areas, i.e., knowledge of the literature, methodology, and policy implications
Research and policy background and experience
Strong communication skills
Proposed terms:
300 hours (flexible, extendable) at $25–$55/hour USD (TBD, depending on experience and skills)
This is a contract role, open to remote and international applicants. However, the ability to attend approximately weekly meetings and check-ins at times compatible with the New York timezone is essential.
Length and timing:
Flexible; to be specified and agreed with the contractor.
Extensions, growth, and promotions are possible, depending on performance, fit, and our future funding.
Statistics, data science and "aggregation of expert beliefs"
We welcome independent evaluations and offer prizes for the strongest submissions. Contact us at contact@unjournal.org if you're interested.
Disambiguation
Note on other versions of this content.
is seeking academics, researchers, and students to submit structured evaluations of the most impactful research emerging in the social sciences. Strong evaluations will be posted or linked on our , offering readers a perspective on the implications, strengths, and limitations of the research. These evaluations can be submitted using for academic-targeted research or for more applied work; evaluators can publish their name or maintain anonymity; we also welcome collaborative evaluation work. We will facilitate, promote, and encourage these evaluations in several ways, described below.
We are particularly looking for people with research training, experience, and expertise in quantitative social science and statistics including cost-benefit modeling and impact evaluation. This could include professors, other academic faculty, postdocs, researchers outside of academia, quantitative consultants and modelers, PhD students, and students aiming towards PhD-level work (pre-docs, research MSc students etc.) But anyone is welcome to give this a try — when in doubt, piease go for it.
We are also happy to support collaborations and group evaluations. There is a good track record for this — see: “”, ASAPBio’s, and for examples in this vein. We may also host live events and/or facilitate asynchronous collaboration on evaluations
Instructors/PhD, MRes, Predoc programs: We are also keen to work with students and professors to integrate ‘independent evaluation assignments’ (aka ‘learn to do peer reviews’) into research training.
Your work will support The Unjournal’s core mission — improving impactful research through journal-independent public evaluation. In addition, you’ll help research users (policymakers, funders, NGOs, fellow researchers) by providing high quality detailed evaluations that rate and discuss the strengths, limitations, and implications of research.
Doing an independent evaluation can also help you. We aim to provide feedback to help you become a better researcher and reviewer. We’ll also give prizes for the strongest evaluations. Lastly, writing evaluations will help you build a portfolio with The Unjournal, making it more likely we will commission you for paid evaluation work in the future.
We focus on rigorous, globally-impactful research in quantitative social science and policy-relevant research. (See for details.) We’re especially eager to receive independent evaluations of:
Research we publicly prioritize: see our we've prioritized or evaluated. (Also...)
Research we previously evaluated (see , as well as )
Work that other people and organizations suggest as having high potential for impact/value of information (also see)
You can also suggest research yourself and then do an independent evaluation of it.
We’re looking for careful methodological/technical evaluations that focus on research credibility, impact, and usefulness. We want evaluators to dig into the weeds, particularly in areas where they have aptitude and expertise. See our.
The Unjournal’s structured evaluation forms: We encourage evaluators to do these using either:
Our : If you are evaluating research aimed at an academic journal or
Our : If you are evaluating research that is probably not aimed at an academic journal. This may include somewhat less technical work, such as reports from policy organizations and think tanks, or impact assessments and cost-benefit analyses
Other public evaluation platforms: We are also open to engaging with evaluations done on existing public evaluation platforms such as. Evaluators: If you prefer to use another platform, please let us know about your evaluation using one of the forms above. If you like, you can leave most of our fields blank, and provide a link to your evaluation on the other public platform.
Academic (~PhD) assignments and projects: We are also looking to build ties with research-intensive university programs; we can help you structure academic assignments and provide external reinforcement and feedback. Professors, instructors, and PhD students: please contact us ().
We will encourage all these independent evaluations to be publicly hosted, and will share links to these. We will further promote the strongest independent evaluations, potentially re-hosting them on our platforms (such as )
However, when we host or link these, we will keep them clearly separated and signposted as distinct from the commissioned evaluations; independent evaluations will not be considered official, and their ratings won’t be included in our (see dashboard; see discussion).
Bounties: We offer prizes for the 'most valuable independent evaluations'. Strong submissions may receive prizes of $500 or more.
As a reference...
All evaluation submissions will be eligible for these prizes and “grandfathered in” to any prizes announced later. We will announce and promote the prize winners (unless they opt for anonymity).
Evaluator pool: People who submit evaluations can elect to join our evaluator pool. We will consider and (time-permitting) internally rate these evaluations. People who do the strongest evaluations in our focal areas are likely to be commissioned as paid evaluators for The Unjournal.
We’re also moving towards a two-tiered base compensation for evaluations. We will offer a higher rate to people who can demonstrate previous strong review/evaluation work. These independent evaluations will count towards this ‘portfolio’.
Our provides examples of strong work, including the.
We will curate guidelines and learning materials from relevant fields and from applied work and impact-evaluation. For a start, see We plan to build and curate more of this...
We are reaching out to PhD programs and pre-PhD research-focused programs. Some curricula already involve “mock referee report” assignments. We hope professors will encourage their students to do these through our platform. In return, we’ll offer the incentives and promotion mentioned above, as well as resources, guidance, and some further feedback
5. Fostering a positive environment for anonymous and signed evaluations
We want to preserve a positive and productive environment. This is particularly important because we will be accepting anonymous content. We will take steps to ensure that the system is not abused. If the evaluations have an excessively negative tone, have content that could be perceived as personal attacks, or have clearly spurious criticism, we will ask the evaluators to revise this, or we may decide not to post or link it.
Crowdsourced feedback can add value in itself; encouraging this can enable some public evaluation and discussion of work that The Unjournal doesn’t have the bandwidth to cover
Improving our evaluator pool and evaluation standards in general.
Students and ECRs can practice and (if possible) get feedback on independent evaluations
commissions public evaluations of impactful research in quantitative social sciences fields. We are an alternative and a supplement to traditional academic peer-reviewed journals – separating evaluation from journals unlocks a . We ask expert evaluators to write detailed, constructive, critical reports. We also solicit a set of structured ratings focused on research credibility, methodology, careful and calibrated presentation of evidence, reasoning transparency, replicability, relevance to global priorities, and usefulness for practitioners (including funders, project directors, and policymakers who rely on this research). While we have mainly targeted impactful research from academia, our covers impactful work that uses formal quantitative methods but is not aimed at academic journals. As of January 2026, we've commissioned over 100 evaluations of 53 papers, and published these evaluation packages , linked to academic search engines and bibliometrics. See our for current statistics.
They can demonstrate their ability this publicly, enabling us to recruit and commission the strongest evaluators
Examples will help us build guidelines, resources, and insights into ‘what makes an evaluation useful’.
This provides us opportunities to engage with academia, especially in Ph.D programs and research-focused instruction.
I was in academia for about 20 years (PhD Economics, UC Berkeley; Lecturer, University of Essex; Senior Lecturer, University of Exeter). I saw how the journal system was broken.
Academics constantly complain about it (but don't do anything to improve it).
Most conversations are not about research, but about 'who got into what journal' and 'tricks for getting your paper into journals'
Open science and replicability are great, and dynamic documents make research a lot more transparent and readable. But these goals and methods are very hard to apply within the traditional journal system and its 'PDF prisons'.
Now I'm working outside academia and can stick my neck out. I have the opportunity to help fix the system. I work with research organizations and large philanthropists involved with effective altruism and global priorities. They care about the results of research in areas that are relevant to global priorities. They want research to be reliable, robust, reasoning-transparent, and well-communicated. Bringing them into the equation can change the game.
Archived: This project was paused in late 2023 to focus on field specialist positions. Similar research scoping work is now integrated into field specialist and evaluation manager roles. The description below is preserved for reference.
We planned to hire 3–7 researchers for a one-off paid project.
There are two opportunities: Contracted Research (CR) and Independent Projects (IP).
Project Outline
What specific research themes in economics, policy, and social science are most important for global priorities?
What projects and papers are most in need of further in-depth public evaluation, attention, and scrutiny?
Where does "Unjournal-style evaluation" have the potential to be one of the most impactful uses of time and money? By impactful, we mean in terms of some global conception of value (e.g., the well-being of living things, the survival of human values, etc.).
This is an initiative that aims to identify, summarize, and conduct an in-depth evaluation of the most impactful themes in economics, policy, and social science to answer the above questions. Through a systematic review of selected papers and potential follow-up with authors and evaluators, this project will enhance the visibility, understanding, and scrutiny of high-value research, fostering both rigorous and impactful scholarship.
Contracted Research (CR) This is the main opportunity, a unique chance to contribute to the identification and in-depth evaluation of impactful research themes in economics, policy, and social science. We’re looking for researchers and research users who can commit a (once-off) 15–20 hours. CR candidates will:
Summarize a research area or theme, its status, and why it may be relevant to global priorities (~4 hours).
We are looking for fairly narrow themes. Examples might include:
The impact of mental health therapy on well-being in low-income countries.
We will compensate you for your time at a rate reflecting your experience and skills ($25–$65/hour). This work also has the potential to serve as a “work sample” for future roles at The Unjournal, as it is highly representative of what our and are commissioned to do.
We are likely to follow up on your evaluation suggestions. We also may incorporate your writing into our web page and public posts; you can choose whether you want to be publicly acknowledged or remain anonymous.
Independent Projects (IP)
We are also inviting applications to do similar work as an “Independent Project” (IP), a parallel opportunity designed for those eager to engage but not interested in working under a contract, or not meeting some of the specific criteria for the Contracted Research role. This involves similar work to above.
If you are accepted to do an IP, we will offer some mentoring and feedback. We will also offer prize rewards/bounties for particularly strong IP work. We will also consider working with professors and academic supervisors on these IP projects, as part of university assignments and dissertations.
You can apply to the CR and IP positions together; we will automatically consider you for each.
Get Involved!
If you are interested in involvement in either the CR or IP side of this project, please let us know .
An important part of making this a success will be to spread the word, to get positive attention for this project, to get important players on board, network externalities, and change the equilibrium. We are also looking for specific feedback and suggestions from "mainstream academics" in Economics, Psychology, and policy/program evaluation, as well as from the Open Science and EA communities.
See
The impact of cage-free egg regulation on animal welfare.
Public attitudes towards AI safety regulation.
Identify a selection of papers in this area that might be high-value for UJ evaluation (~3 hours).
Choose at least four of these from among NBER/"top-10 working paper" series (or from work submitted to the UJ – we can share – or from work where the author has expressed interest to you).
For a single paper, or a small set of these papers (or projects) (~6 hours)
Read the paper fairly carefully and summarize it, explaining why it is particularly relevant.
Discuss one or more aspects of the paper that need further scrutiny or evaluation.
Identify 3 possible evaluators, and explain why they might be particularly relevant to evaluate this work. (Give a few sentences we could use in an email to these evaluators).
Possible follow-up task: email and correspond with the authors and evaluators (~3 hours).
The "Unjournal" is happening, thanks to ACX and the LTFF! We will be organizing and funding:
Journal-independent peer review and rating,
... of projects (not just "pdf-imprisoned papers"),
focusing on Economics, Psychology, and Impact Evaluation research,
relevant to the world's most pressing problems and most effective solutions.
Target: Academics, not necessarily EA aligned. But I don’t think this is deceptive because the funders should give a tipoff to anyone who digs, and ultimately The Unjournal might also go beyond EA-relevant stuff.
Tone: Factual, positive
Do you love for-profit journals
taking your labor and selling it back to your university library?
making you jump through arcane hoops to "format your article"?
forcing you through inscrutable sign-in processes?
Then please don't bother with The Unjournal.
Target: Academics, not necessarily EA aligned who are frustrated with this stuff.
Tone: Sarcastic, irreverent, trying to be funny
Journals: Rent-extracting, inefficient, pdf-prisons, gamesmanship. But no researcher can quit them.
Until The Unjournal: Rate projects, shared feedback, pay reviewers.
No trees axed to print the latest "Journal of Fancy Manuscripts." We just evaluate the most impactful work.
Target, Tone: Same as above, but less sarcastic, using language from Economics … maybe also appealing to library and university admin people?
Traditional academic journals: Rent-extracting, inefficient, delaying innovation. But no researcher or university can quit them.
Or maybe we do have some escape bridges. We can try to Unjournal. Projects get rated, feedback gets shared, reviewers get paid. No trees get chopped down to print the latest "Journal of Fancy Manuscripts." We are starting small, but it only takes one domino.
Your paper got rejected after two glowing reviews? Up for tenure? How many more journals will you have to submit it to? Will you have to make the same points all over again? Or will the new referees tell you the exact opposite of the last ones?
Don't worry, there's a new game in town: The Unjournal. Submit your work. Get it reviewed and rated. Get public feedback. Move on . . . or continue to improve your project and submit it wherever else you like.*
*And we are not like the "Berkeley Electronic Press". We will never sell out, because we have nothing to sell.
Aim, tone: Similar to the above
Tired of the 'pdf prison'? Got...
a great web interface for your project, with expandable explanations
an R-markdown dynamic document, with interactive tools, data, code.
or your software or data is the project.
Can't submit it to a journal but need feedback and credible ratings? Try The Unjournal.
Target: More open-science and tech-savvy people
Referee requests piling up? You better write brilliant reviews for that whopping $0, so the author can be annoyed at you and they can disappear into the ether. Or you can help The Unjournal, where you get paid for your work, and reviews become part of the conversation.
Aim tone: similar to 2–3
Social science research:
builds methods of inferring evidence from data;
builds clear logical arguments;
helps us understand behavior, markets, and society; and
informs "policy" and decision making . . . but for whom and for what goal?
The US government and traditional NGOs are often the key audience (and funders). "It's easier to publish about US data and US policy," they say. But most academics think more broadly than that. And Economics as a field has historically aimed at "the greatest social good." The Unjournal will prioritize research that informs the most effective interventions and global priorities, for humanity (and animals) now and in the future.
Target: EAs and EA-aligned researchers, researchers who might be "converted"
Tone: Straightforward, idealistic
You are a researcher at an organization trying to find the most effective ways to improve the world, reduce suffering, prevent catastrophic risks, and improve the future of humanity. You, your team, your funders, and the policymakers you want to influence . . . they need to know if your methods and arguments are strong, and if your evidence is believable. It would be great if academic experts could give their honest feedback and evaluation. But who will evaluate your best work, and how will they make this credible? Maybe The Unjournal can help.
Target: Researchers and research-related ops people at EA and EA-adjacent orgs. Perhaps OP in particular.
Tone: Casual but straightforward
ACX will announce this, I shared some text
Post on ACX substack
The Unjournal is in large part about shifting the equilibrium in academia/research. As I said in the application, I think most academics and researchers are happy and ready for this change but there's a coordination problem to resolve. (Everyone thinks "no one else will get on this boat," even though everyone agrees it's a better boat). I would love to let ACX readers (especially those in research and academia) know there's a "new game in town." Some further key points (please let me know if you think these can be stated better):
The project space is unjournal.org, which I'd love to share with the public ... to make it easy, it can be announced as "" as in "bitly dot com EA unjournal"... and everyone should let me know if they want editor access to the gitbook; also, I made a quick 'open comment space' in the Gdoc .
I'm looking for feedback and for people interested in being part of this, and for 'nominations' of who might be interested (in championing this, offering great ideas, being part of the committee)
We will put together a committee to build some consensus on a set of workable rules and standards (especially for "how to choose referees," "what metrics should they report," and "how to define the scope of EA-relevant work to consider"). But we won't "hold meetings forever"; we want to build an MVP soon.
Social media
Twitter: Academia (esp. Econ, Psych, Global Health), Open science, EA
EA Forum post (and maybe AMA?)
EA orgs
Open science orgs (OSF, BITSS, ...)
Academic Economics (& other fields) boards/conferences/groups?
Universities/groupings of universities
Slack groups
Global EA
EA Psychology
Open science MooC?
I think this could be a big win for EA and RP "getting more relevant research," for improving academia (and ultimately replacing the outdated system of traditional journals), and for building stronger ties between the two groups.
Researchers should know:
We will pay reviewers to offer feedback, assessment, and metrics, and reviews will be public (but reviewers might be anonymous -- this is a discussion point).
We will offer substantial cash prizes for the best projects/papers, and will likely ask the winners to present their work at an online seminar
You'll be able to submit your research project/paper to the unjournal (or recommend others' work) at any point in the "publication process"; it is not exclusive, and will not prevent you from 'publishing elsewhere'
You're encouraged to submit (time-stamped) 'projects' including dynamic documents connected to data, and interactive presentations
Archived content: This page contains historical updates from 2022-2023.
For current news, visit our main website news page or follow us on Bluesky (@unjournal.bsky.social).
We partitioned research monitoring amongst "Field Specialists" who focus on particular areas of research, policy, or impactful outcomes.
We increased the base compensation for evaluations to $400, and set aside $150 per evaluation for incentives, rewards, and prizes.
Gary Charness, Professor of Economics, UC Santa Barbara
Nicolas Treich, Associate Researcher, INRAE, Toulouse School of Economics
Anca Hanea, Associate Professor, University of Melbourne
Jordan Dworkin, Program Lead, Impetus Institute for Meta-science
Michael Wiebe, Data Scientist, Economist Consultant
As of this update, we had begun evaluation processes for nine papers, completed six full evaluation packages, and learned valuable lessons about our workflow.
We received notification of our SFF grant award.
Platform selection and setup
First papers identified for evaluation
The Unjournal is delighted to announce the winners of our inaugural Impactful Research Prize. We are awarding our first prize to Takahiro Kubo (NIES Japan and Oxford University) and co-authors for their research titled "Banning wildlife trade can boost demand". The paper stood out for its intriguing question, the potential for policy impact, and methodological strength. We particularly appreciated the authors’ open, active, and detailed engagement with our evaluation process.
The second prize goes to Johannes Haushofer (NUS Singapore and Stockholm University) and co-authors for their work "The Comparative Impacts of Cash Transfers and a Psychotherapy Program on Psychological and Economic Wellbeing". Our evaluators rated this paper among the highest across a range of metrics. It was highly commended for its rigor, the importance of the topic, and the insightful discussion of cost-effectiveness.
We are recognizing exceptional evaluators for credible, insightful evaluations. Congratulations to Phil Trammell (Global Priorities Institute at the University of Oxford), Hannah Metzler (Complexity Science Hub Vienna), Alex Bates (independent researcher), and Robert Kubinec (NYU Abu Dhabi).
We would like to congratulate all of the winners on their contributions to open science and commitment to rigorous research. We also thank other authors who have submitted their work but have not been selected at this time - we received a lot of excellent submissions, and we are committed to supporting authors beyond this research prize.
Please see the full press release, as well as award details, below and :
\
Canonical source: For the most up-to-date team information, see our Team page at unjournal.org.
The information below may lag behind the main site. Team members can update details in our internal Coda workspace.
The Unjournal was founded by David Reinstein, who maintains this GitBook and other resources.
See also: Governance of The Unjournal
— Founding Director
— Co-Director (Monk Prayogshala)
Davit Jintcharadze — Manager of Operations
— Columbia University
— London School of Hygiene and Tropical Medicine
— University of Melbourne
We have 60+ field specialists and research affiliates organized by expertise areas including development economics, global health, behavioral science, environmental economics, AI governance, and animal welfare.
The table below shows team members taking on research-monitoring roles (see ):
Over 200 people have signed up to assess research for The Unjournal, including approximately 90 PhD holders and 40 professors.
Interested in becoming an evaluator? .
— Staff (University of South Carolina)
— Staff (Monk Prayogshala)
— Staff (University of Guelph)
— Staff (Nordic Africa Institute)
— Staff (Founders Pledge)
— Staff (Cologne Institute for Information Systems)
Andrei Potlogea — Staff (University of Edinburgh)
— London Business School
— University of Chieti
— Coefficient Giving
— Asterisk Magazine
— University of Minnesota
— Graz University of Technology
— Toulouse School of Economics
— Economist
— Haas School of Business, UC Berkeley
— UC Berkeley
— University of Cambridge
— University of Cambridge
— Good Science Project
— University of Sydney
— Imperial College London
— INRAE
Eva Vivalt — University of Toronto
Cooper Smout — Free Our Knowledge
Brian Nosek — Center for Open Science
Ted Miguel — BITSS, Berkeley
Daniel Saderi — PREreview
January 2026: We now publish updates primarily on our main website news page and on social media.
Follow us on Bluesky (@unjournal.bsky.social) for the latest news and discussions.
Research evaluation is changing: New approaches go beyond the traditional journal model, promoting transparency, replicability, open science, open access, and global impact. You can be a part of this.
Join us on March 25 for an interactive workshop, featuring presentations from Macie Daley (Center for Open Science), (The Unjournal), (UC Santa Barbara), and The Unjournal's Impactful Research Prize and Evaluator Prize winners.
See the for all details.
With the completed set of evaluations of and ," our pilot was complete:
10 research papers evaluated
21 evaluations
5 author responses
You can see all our published evaluations at .
For continuously updated statistics, see our and .
For earlier updates, see .
Sign up for our newsletter to receive occasional updates:
Alternatively, fill out this to get this newsletter and tell us some things about yourself and your interests. The data protection statement is linked .
TLDR: Unjournal promotes research replicability/robustness
Unjournal evaluations aim to support the "Reproducibility/Robustness-Checking" (RRC) agenda. We are directly engaging with the Institute for Replication (I4R) and the repliCATS project (RC), and building connections to Replication Lab/TRELiSS and Metaculus.
We will support this agenda by:
Promoting data and code sharing: We request pre-print authors to share their code and data, and reward them for their transparency.
: Breaking out of "PDF prisons" to achieve increased transparency.
Encouraging detailed evaluations: Unjournal evaluators are asked to:
highlight the key/most relevant research claims, results, and tests;
propose possible robustness checks and tests (RRC work); and
Implementing computational replication and robustness checking: We aim to work with I4R and other organizations to facilitate and evaluate computational replication and robustness checking.
Advocating for open evaluation: We prioritize making the evaluation process transparent and accessible for all.
While the in psychology is well known, economics is not immune. Some very prominent and influential work has , depends on is not, or uses . Roughly 40% of experimental economics work . Prominent commenters have that the traditional journal peer-review system does a poor job of spotting major errors and identifying robust work.
My involvement with the shed light on a key challenge (see posts): The effectiveness of replication depends on the claims chosen for reproduction and how they are approached. I observed that it was common for the chosen claim to miss the essence of the paper, or to focus on a statistical result that, while likely to reproduce, didn't truly convey the author's message.
Simultaneously, I noticed that many papers had methodological flaws (for instance, lack of causal identification or the presence of important confounding factors in experiments). But I thought that these studies, if repeated, would likely yield similar results. These insights emerged from only a quick review of hundreds of papers and claims. This indicates that a more thorough reading and analysis could potentially identify the most impactful claims and elucidate the necessary RRC work.
Indeed, detailed, high-quality referee reports for economics journals frequently contain such suggestions. However, these valuable insights are often overlooked and rarely shared publicly. Unjournal aims to change this paradigm by focusing on three main strategies:
Identifying vital claims for replication:
We plan to have Unjournal evaluators help highlight key "claims to replicate," along with proposing replication goals and methodologies. We will flag papers that particularly need replication in specific areas.
Public evaluation and author responses will provide additional insight, giving future replicators more than just the original published paper to work with.
By concentrating on NBER papers, we increase the likelihood of overlap with journals targeted by the Institute for Replication, thus enhancing the utility of our evaluations in aiding replication efforts.
We can rely on and build a shared talent pool: UJ evaluators may be well-suited—and keen—to become robustness-reproducers (of these or other papers) as well as repliCATS participants.
We see the potential for synergy and economies of scale and scope in other areas, e.g., through:
sharing of IT/UX tools for capturing evaluator/replicator outcomes, and statistical or info.-theoretic tools for aggregating these outcomes;
As a "journal-independent evaluation" gains career value, as replication becomes more normalized, and as we scale up:
This changes incentive systems for academics, which makes rewarding replication/replicability easier than with the traditional journals’ system of "accept/reject, then start again elsewhere."
Encouraging author-assisted replication:
The Unjournal's platform and metrics, promoting dynamic documents and transparency, simplify the process of reproduction and replication.
By emphasizing replicability and transparency at the working-paper stage (Unjournal evaluations’ current focus), we make authors more amenable to facilitate replication work in later stages, such as post-traditional publication.
Predicting replicability and recognizing success:
We aim to ask Unjournal evaluators to make predictions about replicability. When these are successfully replicated, we can offer recognition. The same holds for repliCATS aggregated/IDEA group evaluations: To know if we are credibly assessing replicability, we need to compare these to at least some "replication outcomes."
The potential to compare these predictions to actual replication outcomes allows us to assess the credibility of our replicability evaluations. It may also motivate individuals to become Unjournal evaluators, attracted by the possibility of influencing replication efforts.
sharing of protocols for data, code, and instrument availability (e.g., );
communicating the synthesis of "evaluation and replication reports"; or
encouraging institutions, journals, funders, and working paper series to encourage or require engagement.
More ambitiously, we may jointly interface with prediction markets. We may also jointly integrate into platforms like OSF as part of an ongoing process of preregistration, research, evaluation, replication, and synthesis.
Public communication of Unjournal evaluations and responses may encourage demand for replication work.
In a general sense, we see cultural spillovers in the willingness to try new systems for reward and credibility, and for the gatekeepers to reward this behavior and not just the traditional "publication outcomes".
Our theory of change is shown above as a series of possible paths; we indicate what is arguably the most "direct" path in yellow. All of these paths begin with our setting up, funding, communicating, and incentivizing participation in a strong, open, efficient research evaluation system (in green, at the top). These processes all lead to impactful research being more in-depth, more reliable, more accessible, and more useful, and thus better informing decision-makers and leading to better decisions and outcomes (in green, at the bottom).
Interactive presentation: Walk through the Theory of Change step-by-step in our slide presentation (use arrow keys to navigate). Great for conferences and talks!
You can zoom in on below
(Yellow) Faster and better feedback on impactful research improves this work and better informs policymakers and philanthropists.
(Blue) Our processes and incentives will foster ties between (a) mainstream and prominent academic and policy researchers and (b) global-priorities or EA-aligned researchers. This will improve the rigor, credibility, exposure, and influence of previously "EA niche" work while helping mainstream researchers better understand and incorporate ideas, principles, and methods from the EA and rationalist research communities (such as counterfactual impact, cause-neutrality, reasoning transparency, and so on.) This process will also nudge mainstream academics towards focusing on impact and global priorities, and towards making their research and outputs more accessible and useable.
(Pink) The Unjournal’s more efficient, open, and flexible processes will become attractive to academics and stakeholders. As we become better at "predicting publication outcomes," we will become a replacement for traditional processes, improving research overall—some of which will be highly impactful research.
Rigorous quantitative and empirical research in economics, business, public policy, and social science has the potential to improve our decision-making and enable a flourishing future. This can be seen in the research frameworks proposed by 80,000 Hours, Open Philanthropy, and The Global Priorities Institute (see ). This research is routinely used by effective altruists working on global priorities or existential risk mitigation. It informs both philanthropic decisions (e.g., those influenced by , whose inputs are largely based on academic research) and . Unfortunately, the academic publication process is notoriously slow; for example, in economics, it between the first presentation of a research paper and the eventual publication in a peer-reviewed journal. Recent reforms have sped up parts of the process by encouraging researchers to put working papers and preprints online.
However, working papers and preprints often receive at most only a cursory check before publication, and it is up to the reader to judge quality for themselves. Decision-makers and other researchers rely on peer review to judge the work’s credibility. This part remains slow and inefficient. Furthermore, it provides very noisy signals: A paper is typically judged by the "prestige of the journal it lands in"’ (perhaps after an intricate odyssey across journals), but it is hard to know why it ended up there. Publication success is seen to depend on personal connections, cleverness, strategic submission strategies, good presentation skills, and relevance to the discipline’s methods and theory. These factors are largely irrelevant to whether and how philanthropists and policymakers should consider and act on a paper’s claimed findings. Reviews are kept secret; the public never learns why a paper was deemed worthy of a journal, nor what its strengths and weaknesses were.
We believe that disseminating research sooner—along with measures of its credibility—is better.
We also believe that publicly evaluating its quality before (and in addition to) journal publication will add substantial additional value to the research output, providing:
a quality assessment (by experts in the field) that can decisionmakers and other researchers can read alongside the preprint, helping these users weigh its strengths and weaknesses and interpret its implications; and
faster feedback to authors focused on improving the rigor and impact of the work.
Various initiatives in the life sciences have already begun reviewing preprints. While economics took the lead in sharing working papers, public evaluation of economics, business, and social science research is rare. The Unjournal is the first initiative to publicly evaluate rapidly-disseminated work from these fields. Our specific priority: research relevant to global priorities.
The Unjournal will encourage and incentivize substantive and helpful feedback and careful quantitative evaluation. We will publish these evaluations in a carefully curated space, and clearly aggregate and communicate this output.
This will help us achieve our focal, most tangible "theory of change" pathway (mapped in our "Plan for Impact"):
Better (faster, public, more thorough, more efficient, quantified, and impact-minded) evaluation of pivotal research
The Unjournal’s open feedback should also be valuable to the researchers themselves and their research community, catalyzing progress. As the Unjournal Evaluation becomes a valuable outcome in itself, researchers can spend less time "gaming the journal system." Shared public evaluation will provide an important window to other researchers, helping them better understand the relevant cutting-edge concerns. The Unjournal will permit research to be submitted in a wider variety of useful formats (e.g., dynamic documents and notebooks rather than "frozen pdfs"), enabling more useful, replicable content and less time spent formatting papers for particular journals. We will also allow researchers to improve their work in situ and gain updated evaluations, rather than having to spin off new papers. This will make the literature more clear and less cluttered.
Some of the paths will take longer than others; in particular, it will be hard to get academia to change, particularly because of entrenched systems and a collective action problem. We discuss how we hope to overcome this In particular, we can provide leadership and take risks that academics won’t take themselves:
Bringing in new interests, external funding, and incentives can change the fundamental incentive structure.
makes this research better (both the evaluated work and adjacent work) and encourages more such work
and makes it easier for decision makers to evaluate and use the work, leading to better decisions and better outcomes,
thus reducing X-risk and contributing to long-term survival and flourishing.
We can play a long game and build our processes and track record while we wait for academia to incorporate journal-independent evaluations directly into their reward systems. Meanwhile, our work and output will be highly useful to EA and global-priorities longtermist researchers and decision makers as part of their metrics and reward systems.
The Unjournal’s more efficient, open, and flexible processes will become attractive to academics and stakeholders. As we become better at "predicting publication outcomes," we will become a replacement for traditional processes, improving research overall—some of which will be highly impactful research.
This process will also nudge mainstream academics towards focusing on impact and global priorities, and towards making their research and outputs more accessible and useable.
Should research projects be improved and updated 'in the same place', rather than with 'extension papers'?
Small changes and fixes: The current system makes it difficult to make minor updates – even obvious corrections – to published papers. This makes these papers less useful and less readable. If you find an error in your own published work, there is also little incentive to note it and ask for a correction, even if this were possible.
In contrast, a 'living project' could be corrected and updated in situ. If future and continued evaluations matter, they will have the incentive to do so.
Lack of incentives for updates and extensions
But we recognize there may also be downsides to _'_all extensions and updates in a single place'...
PS (?): Are these things like 'living' google docs that keep getting updated? If so I'd consider using workarounds to replicate their benefits on the forum for a test run (e.g., people add a version to paper title or content or post a new version for each major revision). More generally, I'd prefer the next publication norm for papers to be about making new 'versions' of prior publications (e.g, a 'living review' paper on x is published and reviewed each year) than creating live documents (e.g., a dynamic review on x is published on a website and repeatedly reviewed at frequent and uncertain intervals when the authors add to it). I see huge value in living documents. However, I feel that they wouldn't be as efficient/easy to supervise/review as 'paper versions'.
@GavinTaylor: I don’t think living documents need to pose a problem as long as they are discretely versioned and each version is accessible. Some academic fields are/were focused on books more than papers, and these were versioned by edition. Preprinting is also a form of versioning and combining the citations between the published paper and its preprint/s seems to be gaining acceptance (well, google scholar may force this by letting you combine them) - I don’t recall ever seeing a preprint citation indication a specific version (on preprint servers that support this) but its seems possible.
Some discussion follows. Note that the Unjournal enables this but does not require it.
In contrast, a 'living project' can be improved and extended in situ.
Clutter and confusion in the literature: Because of the above, researchers often try to spin an improvement to a previous paper as very new and different. They do sometimes publish a range of papers getting at similar things and using similar methods, in different papers/journals. This makes it hard for other researchers and readers to understand which paper they should read.
In contrast, a 'living project' can keep these in one place. The author can lay out different chapters and sections in ways that make the full work most useful.
DR: I mainly agree with @gavintaylor, but I appreciate that 'changing everything at the same time' is not always the best strategy.
The main idea is that each version is given a specific time stamp, and that is the object that is reviewed and cited. This is more or less already the case when we cite working papers/drafts/mimeos/preprints.
Gavin, on the latter 'past version accesibility' issue, This could/should be a part of what we ensure with specific rules and tech support, perhaps.
I think the issue with the current citing practice for live documents like webpages is that even if a ‘version’ is indicated (e.g. access date) past versions aren’t often very accessible.
They might also not be ideal for citing as they would be an ever-changing resource. I can imagine the whole academic system struggling to understand and adapt to such a radical innovation given how focused it is on static documents. With all of this considered, I'd like 'dynamic/living work' to be incentivised with funding and managed with informal feedback and comments rather than being formally reviewed (at least for now). I'd see living review work as sitting alongside and informing 'reviewed' papers rather than supplanting them. As an example, you might have a website that provides a 'living document' for lay people about how to promote charity effectively and then publish annual papers to summarise the state of the art for an academic/sophisticated audience.
DR: I can see the arguments on both sides here. I definitely support replications and sometimes it may make sense for the author to “start a new paper” rather than make this an improvement if the old one. I also think that the project should be time stamped, evaluated and archived at particular stages of its development.
But I lean to thinking that in many to most cases a single project with multiple updates will make the literature clearer and easier to navigate than the current proliferation of “multiple very similar papers by the same author in different journals”. It also seems a better use of researcher time, rather than having to constantly restate and repackage the same things
Did you just write a brilliant peer review for an economics (or social science, policy, etc.) journal? Your work should not be wasted, there should be a way to share your insights and get credit!
Consider transforming these insights into a public "independent evaluation" for . This will benefit the community and help make research better and more impactful. And we can share your work and provide you feedback. This will help you build a portfolio with The Unjournal, making it more likely we'll hire you for paid work and compensate you at the higher rate. And we offer prizes for the best work.
You can do this either anonymously or sign your name.
To say this in a fancier way:
Journal peer review is critical for assessing and improving research, but too often these valuable discussions remain hidden behind closed doors. By publishing a version of your review, you can: (1) Amplify the impact of your reviewing efforts by contextualizing the research for a broader audience, (2) Facilitate more transparent academic discussions around the strengths and limitations of the work, (3) Get public recognition for your peer review contributions, which are often unseen and unrewarded (4) Reduce overall reviewing burdens by allowing your assessment to be reused, (5) Support a culture of open scholarship by modeling constructive feedback on public research
According to a : Who “owns” peer reviews (emphasis added)
While the depth of commentary may vary greatly among reviews, given the minimal thresholds set by copyright law, it can be presumed that most reviews meet the requirements for protection as an “original work of authorship”. As such, in the absence of an express transfer of copyright or a written agreement between the reviewer and publisher establishing the review as a “work for hire”, it may be assumed that, by law, the reviewer holds copyright to their reviewer comments and thus is entitled to share the review however the reviewer deems fit...
The COPE council notes precisely the benefits we are aiming to unlock. They mention an 'expectation of confidentiality' that seems incompletely specified.
For example, reviewers may wish to publish their reviews in order to demonstrate their expertise in a subject matter and to contribute to their careers as a researcher. Or they may see publication of their reviews as advancing discourse on the subject and thus acting for the benefit of science as a whole. Nevertheless, a peer reviewer’s comments are significantly different from many other works of authorship in that they are expressly solicited as a work product by a journal and—whatever the peer review model—are subject to an expectation of confidentiality. However, without an express agreement between the journal and the reviewer, it is questionable whether such obligation of confidentiality should be considered to apply only until a final decision is reached on the manuscript, or to extend indefinitely.
Several journals explicitly agree that reviewers are welcome to publish the content of their reviews, with some important caveats. The gathered public statements from several journals and publishers confirming that they support reviewers posting their comments externally. However, they generally ask reviewers to remove any confidential information before sharing their reviews. This includes: the name of the journal, the publication recommendation (e.g., accept, revise, or reject), and any other details the journal or authors considered confidential, such as unpublished data.
For these journals, we are happy to accept and share/link the verbatim content as part of an independent Unjournal evaluation.
But even for journals that have not signed onto this, as the COPE mentioned Your peer review is your intellectual property, it is not owned by the journal!
There may be some terms and conditions you agreed to as part of submitting a referee report. Please consult these carefully.
However, you are still entitled to share your own expert opinions on publicly-shared research. You may want to rewrite the review somewhat. You should make it clear that it refers to the publicly-shared (working paper/preprint) version of the research, not the one the journal shared with you in confidence. As above, you should probably not mention the journal name, the decision, or any other sensitive information. You don't even need to mention that you did review the paper for a journal.
Even if a journal considers the specific review confidential, this doesn't prevent the reviewer from expressing their independent assessment elsewhere.
As an expert reviewer, you have unique insights that can improve the quality and impact of research. Making your assessment available through The Unjournal amplifies the reach and value of your efforts. You can publish evaluations under your name or remain anonymous.
Ready to make your peer reviews work harder for science? Consider submitting an , for recognition, rewards, and to improve research. Contact us anytime at contact@unjournal.org for guidance... We look forward to unlocking your valuable insights!
\
What do we offer? How does it improve upon traditional academic review/publishing?
Visual overview: See our benefits landing page for an illustrated summary of what makes The Unjournal different.
Reshaping academic evaluation: Beyond accept/reject: The Unjournal's process reduces the high costs and "gaming" associated with standard journal publication mechanisms.
Promoting open and robust science: We promote research replicability and robustness in line with the RRC agenda.
: We prioritize impactful work. Expert evaluators focus on reliability, robustness, and usefulness. This fosters a productive bridge between high-profile mainstream researchers and global-impact-focused organizations, researchers, and practitioners.
: We open up the evaluation process, making it more timely and transparent, and providing valuable public metrics and feedback for the benefit of authors, other researchers, and policymakers who may want to use the research..
: By separating evaluation from journal publication, free research from the static 'PDF prisons'. This enables "dynamic documents/notebooks" that boost transparency and replicability, improved research communication through web-based formats. It also facilitates "living projects": research that can continuously grow, improving in response to feedback and incorporating new data and methods in the same environment.
'Dynamic Documents' are projects or papers that are developed using prominent tools such as R-markdown or JuPyTer notebooks (the two most prominent tools).
The salient features and benefits of this approach include:
Integrated data analysis and reporting means the data analysis (as well as math/simulations) is done and reported in the same space that the results and discussion are presented. This is made possible through the concealment of 'code blocks'.
Transparent reporting means you can track exactly what is being reported and how it was constructed:
Making the process a lot less error-prone
Helping readers understand it better (see )
Helping replicators and future researchers build on it
Other advantages of these formats (over PDFs for example) include:
Convenient ‘folding blocks’
Margin comments
Elife's 'editable graphics'... Brett Viktor?
see corrigendum in journals Reinhart and Rogoff error?
Some quick examples from my own work in progress (but other people have done it much better)
Integrating interactive tools
OSF and all of their training/promo materials in OS
Traditional peer review is a closed process, with reviewers' and editors' comments and recommendations hidden from the public.
In contrast, all Unjournal evaluations* (along with authors' responses and evaluation manager summaries) are made public and easily accessible. We give each of these a separate DOI and work to make sure each enters the literature and bibliometric databases. We aim further to curate these, making it easy to see the evaluators' comments in the context of the research project (e.g., with sidebar/hover annotation).
Open evaluation is more useful:
to other researchers and students (especially those early in their careers). Seeing the dialogue helps them digest the research itself and understand its relationship to the wider field. It helps them understand the strengths and weaknesses of the methods and approaches used, and how much agreement there is over these choices. It gives an inside perspective on how evaluation works.
to people using the research, providing further perspectives on its value, strengths and weaknesses, implications, and applications.
Publicly posting evaluations and responses may also lead to higher quality and more reliability. Evaluators can choose whether or not they wish to remain anonymous; there are , but in either case, the fact that all the content is public may encourage evaluators to more fully and transparently express their reasoning and justifications. (And where they fail to do so, readers of the evaluation can take this into account.)
The fact that we are asking for evaluations and ratings of all the projects in our system—and not using "accept/reject"—should also drive more careful and comprehensive evaluation and feedback. At a traditional top-ranked journal, a reviewer may limit themselves to a few vague comments implying that the paper is "not interesting or strong enough to merit publication." This would not make sense within the context of The Unjournal.
We do not "accept or reject" papers; we are evaluating research, not "publishing" it. But then, how do other researchers and students know whether the research is worth reading? How can policymakers know whether to trust it? How can it help a researcher advance their career? How can grantmakers and organizations know whether to fund more of this research?
As an alternative to the traditional measure of worth—asking, "what tier did a paper get published in?"—The Unjournal provides metrics: We ask evaluators to provide a specific set of ratings and predictions about aspects of the research, as well as aggregate measures. We make these public. We aim to synthesize and analyze these ratings in useful ways, as well as make this quantitative data accessible to meta-science researchers, meta-analysts, and tool builders.
Feel free to check out our and .
These metrics are separated into different categories designed to help researchers, readers, and users understand things like:
How much can one believe the results stated by the authors (and why)?
How relevant are these results for particular real-world choices and considerations?
Is the paper written in a way that is clear and readable?
We also request overall ratings and predictions . . . of the credibility, importance, and usefulness of the work, and to help benchmark these evaluations to each other and to the current "journal tier" system.
However, even here, the Unjournal metrics are also precise in a sense that "journal publication tiers" are not. There is no agreed-upon metric of exactly how journals rank (e.g., within economics' "top-5" or "top field journals"). More importantly, there is no clear measure of the relative quality and trustworthiness of the paper within particular journals.
In addition, there are issues of lobbying, career concerns, and timing, discussed elsewhere, which make the "tiers" system less reliable. An outsider doesn't know, for example:
Was a paper published in a top journal because of a special relationship and connections? Was an editor trying to push a particular agenda?
Was it published in a lower-ranked journal because the author needed to get some points quickly to fill their CV for an upcoming tenure decision?
In contrast, The Unjournal requires evaluators to give specific, precise, quantified ratings and predictions (along with an explicit metric of the evaluator's uncertainty over these appraisals).
Of course, our systems will not solve all problems associated with reviews and evaluations: power dynamics, human weaknesses, and limited resources will remain. But we hope our approach moves in the right direction.
See also .
We want to reduce the time between when research is done (and a paper or other research format is released) and when other people (academics, policymakers, journalists, etc.) have a credible measure of "how much to believe the results" and "how useful this research is."
Here's how The Unjournal can do this.
Early evaluation: We will evaluate potentially impactful research soon after it is released (as a working paper, preprint, etc.). We will encourage authors to submit their work for our evaluation, and we will the evaluation of work from the highest-prestige authors.
We will pay evaluators with further incentives for timeliness (as well as carefulness, thoroughness, communication, and insight). that these incentives for promptness and other qualities are likely to work.
Public evaluations and ratings: Rather than waiting years to see "what tier journal a paper lands in," the public can simply consult The Unjournal
See
→ Evaluators may be more motivated to be careful and complete.
On the other hand . . .
For public evaluations, people might defer to being overly careful.
At standard journals, referees do want to impress editors, and often (but not always) leave very detailed comments and suggestions.
By “Dynamic Documents” I mean papers/projects built with Quarto, R-markdown, or JuPyTer notebooks (the most prominent tools) that do and report the data analysis (as well as math/simulations) in the same space that the results and discussion are presented (with ‘code blocks’ hidden).
I consider some of the benefits of this format, particularly for EA-aligned organizations like Open Philanthropy: Benefits of Dynamic Documents
“Continually update a project” rather than start a “new extension paper” when you see what you could have done better.
The main idea is that each version is given a specific time stamp, and that is the object that is reviewed and cited. This is more or less already the case when we cite working papers/drafts/mimeos/preprints.
See , further discussing the potential benefits.
FAQ for authors of research that The Unjournal selected for public evaluation, and for authors considering submitting their work to The Unjournal for evaluation
Quick overview: See our author landing page for a visual summary of benefits, the submission process, and FAQ.
You can fill out this form to submit your work, or email contact@unjournal.org with questions .
We generally seek two evaluators (aka 'reviewers') with research interests in your area and with complementary expertise. You, the author, can suggest areas you want to get feedback and suggestions on.
The evaluators write detailed and helpful evaluations, and submit them either "signed" or anonymously. Using our evaluation forms, they provide quantitative ratings on several dimensions, such as methods, relevance, and communication. They predict what journal tier the research will be published in, and what tier it should be published in. Here are the .
These evaluations and ratings are typically made public (see ), but you will have the right to respond before (or after) these are posted.
To consider your research we only need a link to a publicly hosted version of your work, ideally with a DOI. We will not "publish" your paper. The fact that we are handling your paper will not limit you in any way. You can submit it to any journal before, during, or after the process.
You can request a conditional embargo by emailing us at , or via the submission form. Please explain what sort of embargo you are asking for, and why. By default, we would like Unjournal evaluations to be made public promptly. However, we may make exceptions in special circumstances, particularly for very early-career researchers whose career prospects are particularly vulnerable.
If there is a vulnerable early-career researcher on the authorship team, we may allow authors to "embargo" the publication of the evaluation until a later date. Evaluators (referees) will be informed of this. This date can be contingent, but it should not be indefinite.
For further details on this, and some examples, see .
We may ask for some of the below, but these are mainly optional. We aim to make the process very light touch for authors.
A link to a non-paywalled, hosted version of your work (in any format) which can be given a DOI.
Responses to
We also allow you to respond to evaluations, and we give your response its own DOI.
By submitting your research and engaging with public evaluation, you send a powerful public signal that you are confident in your work, open to constructive criticism, and motivated to seek the truth!
And there are further benefits to authors:
Substantive feedback will help you improve your research. Substantive and useful feedback is often very hard to get, especially for young scholars. It's hard to get anyone to read your paper – we can help!
Being evaluated by The Unjournal is a sign of impact. We .
Ratings are markers of credibility for your work that could help your career.
Your work may be "under-published" because of:
Time pressure: Perhaps you, or a co-author were in a hurry and submitted it to a "safe" but low-ranked journal.
Interdisciplinary approaches: It may have been considered by a journal in one field, but it needs feedback and credibility from other fields (e.g., theory vs. empirics, etc.).
Improvements: You may have substantially improved and extended the work since its publication. The current journal system can only handle this if you claim it is a 'brand new paper'. We aim to fix this (see
There are risks and rewards to any activity, of course. Here we consider some risks you may weigh against the benefits mentioned above.
Exclusivity
Public negative feedback
Some traditional journals might have restrictions on the public sharing of your work, and perhaps they might enforce these more strongly if they fear competition from The Unjournal.
However, The Unjournal is not exclusive. Having your paper reviewed and evaluated in The Unjournal will not limit your options; you can still submit your work to traditional journals.
See Brian Nosek's response to this question , echoing the same points.
Our evaluations are public. While there has been some movement towards open review, this is still not standard. Typically when you submit your paper, reviews are private. With The Unjournal, you might get public negative evaluations.
We think this is an acceptable risk. Most academics expect that opinions will differ about a piece of work, and everyone has received negative reviews. Thus, getting public feedback — in The Unjournal or elsewhere — should not particularly harm you or your research project.
Brian Nosek, : I certainly understand the fear. I shared it myself until I had the experience of realizing that's how people advance their careers. In scholarship, we are engaged in a community of skeptical inquiry. To advance one's career, you have to have people critiquing your work. If everyone is ignoring your work, you are not actually making progress in the field. Productive engagement with skeptical inquiry is a way that people improve their reputation and advance their careers.
Nonetheless, we are planning some exceptions for early-career researchers (see ).
Unjournal evaluations should be seen as . Like all such signals, they are noisy. But submitting to The Unjournal shows you are confident in your work, and not afraid of public feedback.
Authors might worry that "a bad signal will hurt a lot, and a good signal will only help a little."
But if getting any public feedback was so damaging, why would academics be eager to present their work at seminars and conferences?
The main point: Unbiased signals cannot systematically lead to beliefs updating in one direction.
For fancy and formal ways of saying this and related concepts, see , , Rational Expectations, and the Law of Iterated Expectations.
Within our , The Unjournal directly chooses papers (from prominent archives, well-established researchers, etc.) to evaluate. We don't request authors' permission here.
As you can see in our , on this track, we engage with authors especially at two points:
Informing the authors that the evaluation will take place, requesting they engage, and giving them the opportunity to request a or specific types of feedback.
Of particular interest: are we looking at the most recent version of the paper/project, or is there a further revised version we should be considering instead?
After the evaluations have been completed, the authors are given two weeks to respond, and have their response posted along with our 'package'. (Authors can also respond after we have posted the evaluations, and we will put their response in the same 'package', with a DOI etc.)
Once we receive unsolicited work from an author or authors, we keep it in our database and have our team decide on prioritization. If your paper is prioritized for evaluation, The Unjournal will notify you.
At present, we do not have a system to automatically share the status of author submissions with authors. We hope to put one in place. You can email us for clarification and updates.
You can still submit your work to any traditional journal.
The Unjournal aims to evaluate the most recent version of a paper. We reach out to authors to ensure we have the latest version at the start of the evaluation process.
If substantial updates are made to a paper during the evaluation process, authors are encouraged to share the updated version. We then inform our evaluators and ask if they wish to revise their comments.
If the evaluators can't or don't respond, we will note this and still link the newest version.
Authors are encouraged to respond to evaluations by highlighting major revisions made to their paper, especially those relevant to the evaluators' critiques. If authors are not ready to respond to evaluations, we can post a placeholder response indicating that responses and/or a revised version of the paper are forthcoming.
Re-evaluation: If authors and evaluators are willing to engage, The Unjournal is open to re-evaluating a revised version of a paper after publishing the evaluations of the initial version.
We share evaluations with the authors and give them a chance to respond before we make the evaluations public (and again afterward, at any point). We add these to our . Evaluation manager's (public) reports and our further communications incorporate the paper, the evaluations, and the authors' responses.
Authors' responses could bring several benefits...
Personally: a chance to correct misconceptions and explain their approach and planned steps. If you spot any clear errors or omissions, we give evaluators a chance to adjust their reports in response. Your response can also help others have a more accurate and positive view of the research. This includes the evaluators, as well as future journal referees and grant funders.
For research users, to get an informed balanced perspective on how to judge the work
For other researchers, to better understand the methodological issues and approaches. This can serve to start a public dialogue and discussion to help build the field and research agenda. Ultimately, we aim to facilitate a back-and-forth between authors, evaluators, and others.
Evaluations may raise substantive doubts and questions, and make some specific suggestions, and ask followup questions about (e.g.) data, context, or assumptions. There's no need to respond to every evaluator point. Only respond where you have something substantive: clarifying doubts, explaining the justification for your particular choices, and giving your thoughts on the suggestions (which will you incorporate, or not, and why?).
A well-written author response might have a clear narrative and group responses into themes.
Try to have a positive tone (no personal attacks etc.) but avoid formality, over-politeness, or flattery. Revise-and-resubmit responses at standard journals sometimes begin each paragraph with "thank you for this excellent suggestion". Feel free to skip this; we want to focus on the substance.\
Examples: We've received several detailed and informative author responses, such as:
We may ask for a link to data and code, if possible. Note that our project is not principally about replication, and we are not insisting on this. However, sharing code, data, and materials is highly encouraged.
The chance to publicly respond to criticism and correct misunderstandings.
Increasing the visibility of your work, which may lead to additional citations. We publicize our evaluations and the original papers on our social media feed, and occasionally in notebook and longer-form content.
A connection to the Open Science/Open Access and EA/Global Priorities communities. This may lead to grant opportunities, open up new ambitious projects, and attract strong PhD students to your research groups.
A reputation as an early adopter and innovator in open science.
Prizes: You may win an (publicity, reputation, and substantial financial prizes). These prizes are tied, in part, to your engagement with The Unjournal.
Undervalued or updated work: Your paper may have been "under-published". Perhaps there are a limited set of prestigious journals in your field. You now see ways you could improve the research. The Unjournal can help; we will also consider 'post-peer-review publication' evaluation.
Innovative formats: Journals typically require you to submit a LaTeX or MS Word file, and to use their fussy formats and styles. You may want to use tools like that integrate your code and data, allow you to present dynamic content, and enhance reproducibility. The Unjournal encourages this, and we can evaluate research in virtually any format.
Limited journal opportunities: You may have "used up" the good journals in your field, or otherwise fallen through the cracks.
Your focus on reliability and impact: Perhaps your work paper is very impactful and the empirics are strong, but the field doesn't see it as innovative or clever. You need another opportunity.
If a reviewer evaluates a paper without much information on how others rated it, they might suspect that there have been some substantial and important criticisms and that the paper has flaws. This is particularly the case if it has been around a while, they are submitting to a second-tier journal, etc.
On the other hand, if I see that an author has submitted their work to a transparent assessment service like The Unjournal I would update positively, all else being equal. I would consider the public critiques, of course, but I will take into account that this paper has been held up to more scrutiny than other work I might have seen.
You can now ask questions of this GitBook using a chatbot! Click the search bar and choose 'ask gitbook'.
We organize and fund public, journal-independent feedback, rating, and evaluation of academic work. We focus on work that is highly relevant to global priorities, especially in economics, social science, and impact evaluation. We encourage better research by making it easier for researchers to get credible feedback. See for more details.
No. The Unjournal does not charge any fees. In fact, unlike most traditional journals, we compensate evaluators for their time, and award prizes for strong work.
We are a nonprofit organization. We do not charge fees for access to our evaluations, and work to make them as open as possible. In future, we may consider sliding-scale fees for people submitting their work for Unjournal evaluation. If so, this would simply be to cover our costs and compensate evaluators. We are a nonprofit and will stay that way.
No. We do not publish research. We just commission public evaluation and rating of relevant research that is already publicly hosted. Having your work evaluated in The Unjournal should not limit you from submitting it to any publication.
We have from philanthropists and organizations who are interested in our priority research areas. We hope that our work will provide enough value to justify further direct funding. We may also seek funding from governments and universities supporting the open-access agenda.
Sure! Please contact us at .
This page is a work-in-progress
15 Dec 2023: Our main current process involves
Submitted and (internally/externally) suggested research
Prioritization ratings and discussion by Unjournal field specialists
Feedback from field specialist area teams
A final decision by the management team, guided by the above
See (also embedded below) for more details of the proposed process.
This discussion is a work-in-progress
We are targeting global priorities-relevant research...
With the potential for impact, and with the potential for Unjournal evaluations to have an impact (see our high-level considerations and our prioritization ratings discussions).
Our initial focus is quantitative work that informs , especially in economics, policy, and social science, informing our .
We give a data presentation of the work we have already covered and the work we are prioritizing , which will be continually updated.
But what does this mean in practice? What specific research fields, topics, and approaches are we likely to classify as 'relevant to evaluate'?
We give some lists and annotated examples below.
The Unjournal focuses on ...
Research where the fundamental question being investigated involves human behavior and beliefs and the consequences of these. This may involve markets, production processes, economic constraints, social interactions, technology, the 'market of ideas', individual psychology, government processes, and more. However, the main research question should not revolve around issues outside of human behavior, such as physical science, biology, or computer science and engineering. These areas are out of our scope (at least for now).
Research that is fundamentally quantitative and uses
This to generally involves the academic fields:
Economics
Applied Statistics (and some other applied math)
Psychology
Political Science
These discipline/field boundaries are not strict; they may adapt as we grow
These were chosen in light of two main factors:
Our founder and our team is most comfortable assessing and managing the consideration of research in these areas.
These fields seem to be particularly amenable to, and able to benefit from our journal-independent evaluation approach. Other fields, such as biology, are already being 'served' by strong initiatives like .
To do: We will give and explain some examples here
The Unjournal's mission is to prioritize
research with the strongest potential for a positive impact on global welfare
where public evaluation of this research will have the greatest impact
Given this broad goal, we consider research into any cause, topic, or outcome, as long as the research involves fields, methods, and approaches within our domain (see above), and as long as the work meets our other requirements (e.g., research must be publicly shared without a paywall).
While we don't have rigid boundaries, we are nonetheless focusing on certain areas:
We have mainly commissioned evaluations of work involving development economics and health-related outcomes and interventions in low- and middle-income countries.
As well as research involving
Environmental economics, conservation, harm to human health
The social impact of AI and emerging technologies
Economics, welfare, and governance
We are currently prioritizing further work involving
Psychology, behavioral science, and attitudes: the spread of misinformation; other-regarding preferences and behavior; moral circles
Animal welfare: markets, attitudes
Methodological work informing high-impact research (e.g., methods for impact evaluation)
We are also considering prioritizing work involving
AI governance and safety
Quantitative political science (voting, lobbying, attitudes)
Political risks (including authoritarian governments and war and conflict)
To do: We will give and explain some examples here
Research that targets and addresses a single specific question or goals, or a small cluster. It should not mainly be a broad discussion and overview of other research or conceptual issues.
Other quantitative social science fields (perhaps Sociology)
Applied "business school" fields: finance, accounting, operations, etc.
Applied "policy and impact evaluation" fields
Life science/medicine where it targets human behavior/social science
The economics of innovation; scientific progress and meta-science
The economics of health, happiness, and wellbeing
Long-term growth and trends; the long-term future of civilization; forecasting
See links below accessing current policies of The Unjournal, accompanied by discussion and including templates for managers and editors.
Video overview: Two minute overview of our 6-step evaluation process.
People and organizations submit their own research or suggest research they believe may be high-impact. The Unjournal also directly monitors key sources of research and research agendas. Our team then systematically prioritizes this research for evaluation. See the link below for further details.
We choose an evaluation manager for each research paper or project. They commission and compensate expert evaluators to rate and discuss the research, following our evaluation template and . The original research authors are given a chance to publicly respond before we post these evaluations. See the link below for further details.
We make all of this evaluation work public on , along with an evaluation summary. We create DOIs for each element and submit this work to scholarly search engines. We also present a summary and analysis of our .
We outline some further details in the link below.
See for a full 'flowchart' map of our evaluation workflow
We are also piloting several initiatives that involve a different process. See:
Our initial focus is quantitative work that informs global priorities (see linked discussion), especially in economics, policy, and social science. We want to see better research leading to better outcomes in the real world (see our 'Theory of Change').
See (earlier) discussion in public call/EA forum discussion .
To reach these goals, we need to select "the right research" for evaluation. We want to choose papers and projects that are highly relevant, methodologically promising, and that will benefit substantially from our evaluation work. We need to optimize how we select research so that our efforts remain mission-focused and useful. We also want to make our process transparent and fair. To do this, we are building a coherent set of criteria and goals, and a specific approach to guide this process. We explore several dimensions of these criteria below.
Management access only: General discussion of prioritization in Gdoc . Private discussion of specific papers in our Coda resource. We incorporate some of this discussion below.
When considering a piece of research to decide whether to commission it to be evaluated, we can start by looking at its general relevance as well as the value of evaluating and rating it.
Our prioritization of a paper for evaluation should not be seen as an assessment of its quality, nor of its 'vulnerability'. Furthermore, 'the prioritization is not the evaluation', it is less specific and less intensive.
Why is it relevant and worth engaging with?
We consider (and prioritize) the importance of the research to global priorities; its relevance to crucial decisions; the attention it is getting, the influence it is having; its direct relevance to the real world; and the potential value of the research for advancing other impactful work. We de-prioritize work that has already been credibly (publicly) evaluated. We also consider the fit of the research with our scope (social science, etc.), and the likelihood that we can commission experts to meaningfully evaluate it. As noted , some 'instrumental goals' (, , driving change, ...) also play a role in our choices.
Some features we value, that might raise the probability we consider a paper or project include the commitment and contribution to open science, the authors' engagement with our process, and the logic, communication, and transparent reasoning of the work. However, if a prominent research paper is within our scope and seems to have a strong potential for impact, we will prioritize it highly, whether or not it has these qualities.
2. Why does it need (more) evaluation, and what are some key issues and claims to vet?
We ask the people who suggest particular research, and experts in the field:
What are (some of) the authors’ key/important claims that are worth evaluating?
What aspects of the evidence, argumentation, methods, and interpretation, are you unsure about?
What particular data, code, proofs, and arguments would you like to see vetted? If it has already been peer-reviewed in some way, why do you think more review is needed?
Put broadly, we need to consider how this research allows us to achieve our own goals in line with our , targeting "ultimate outcomes." The research we select and evaluate should meaningfully drive positive change. One way we might see this process: “better research & more informative evaluation” → “better decision-making” → “better outcomes” for humanity and for non-human animals (i.e., the survival and flourishing of life and human civilization and values).
As we weigh research to prioritize for evaluation, we need to balance directly having a positive impact against building our ability to have an impact in the future.
Below, we adapt the (popular in effective altruism circles) to assess the direct impact of our evaluations.
Importance
What is the direct impact potential of the research?
This is a massive question many have tried to address (see sketches and links below). We respond to uncertainty around this question in several ways, including:
Consulting a range of sources, not only EA-linked sources.
EA and more or less adjacent: and overviews.
Non-EA, e.g., .
Neglectedness
Where is the current journal system failing GP-relevant work the most . . . in ways we can address?
Tractability
“Evaluability” of research: Where does the UJ approach yield the most insight or value of information?
Existing expertise: Where do we have field expertise on the UJ team? This will help us commission stronger evaluations.
"Feedback loops": Could this research influence concrete intervention choices? Does it predict near-term outcomes? If so, observing these choices and outcomes and getting feedback on the research and our evaluation can yield strong benefits.
Consideration/discussion: How much should we include research with indirect impact potential (theoretical, methodological, etc.)?
Moreover, we need to consider how the research evaluation might support the sustainability of The Unjournal and the broader general project of open evaluation. We may need to strike a balance between work informing the priorities of various audences, including:
Relevance to stakeholders and potential supporters
Clear connections to impact; measurability
Support from relevant academic communities
Consideration/discussion: What will drive further interest and funding?
Finally, we consider how our choices will increase the visibility and solidify the credibility of The Unjournal and open evaluations. We consider how our work may help drive positive institutional change. We aim to:
Interest and involve academics—and build the status of the project.
Commission evaluations that will be visibly useful and credible.
‘Benchmark traditional publication outcomes’, track our predictiveness and impact.
We are aware of possible pitfalls of some elements of our vision.
We are pursuing a second "high-impact policy and applied research" track for evaluation. This will consider work that is not targeted at academic audiences. This may have direct impact and please SFF funders, but, if not done carefully, this may distract us from changing academic systems, and may cost us status in academia.
A focus on topics perceived as niche (e.g., the economics and game theory of AI governance and AI safety) may bring a similar tradeoff.
On the other hand, perhaps a focus on behavioral and experimental economics would generate lots of academic interest and participants; this could help us benchmark our evaluations, etc.; but this may also be less directly impactful.
We hope we have identified the important considerations (above); but we may be missing key points. We continue to engage discussion and seek feedback, to hone and improve our processes and approaches.
We present and analyze the specifics surrounding our current evaluation data in
Below: An earlier template for considering and discussing the relevance of research. This was/is provided both for our own consideration and for sharing (in part?) with evaluators, to give them some guidance. Think of these as bespoke evaluation notes for a "research overview, prioritization, and suggestions" document.
Get a selection of seminal GP publications; look back to see what they are citing and categorize by journal/field/keywords/etc.
Increase public visibility and raise public interest.
Bring in supporters and participants.
Achieve substantial output in a reasonable time frame and with reasonable expense.
Maintain goodwill and a justified reputation for being fair and impartial.
As mentioned under High level considerations, consider factors including importance to global priorities, relevance to the field, the commitment and contribution to open science, the authors’ engagement, and the transparency of data and reasoning. You may consider the ITN framework explicitly, but not too rigidly.
Why does it need (more) review, and what are some key issues and claims to vet?
What are (some of) the authors’ main important claims that are worth carefully evaluating? What aspects of the evidence, argumentation, methods, interpretation, etc., are you unsure about? What particular data, code, proof, etc., would you like to see vetted? If it has already been peer-reviewed in some way, why do you think more review is needed?
What sort of reviewers should be sought, and what should they be asked?
What types of expertise and background would be most appropriate for the evaluation? Who would be interested? Please try to make specific suggestions.
How well has the author engaged with the process?
Do they need particular convincing? Do they need help making their engagement with The Unjournal successful?
As we are paying evaluators and have limited funding, we cannot evaluate every paper and project. Papers enter our database through:
submission by authors;
our own searches (e.g., searching syllabi, forums, working paper archives, and white papers); and
suggestions from other researchers, practitioners, and members of the public, and recommendations from high-impact research-users. We have posted more detailed instructions for .
Our management team rates the suitability of each paper according to the criteria discussed below and .
We have followed a few procedures for finding and prioritizing papers and projects. In all cases, we require more than one member of our research-involved team (field specialist, managers, etc.) to support a paper before prioritizing it.
We are building a grounded systematic procedure with criteria and benchmarks. We also aim to give managers and field specialists some autonomy in prioritizing key papers and projects. As noted elsewhere, we are considering targets for particular research areas and sources.
See our basic process (as of Dec. 2023) for prioritizing work:
Through October 2022: For the papers or projects at the top of our list, we contacted the authors and asked if they wanted to engage, only pursuing evaluation if agreed.
In our under certain conditions, we inform authors but do not request permission. For this track, we have largely focused on working papers.
In deciding which papers or projects to send out to paid evaluators, we have considered the following issues. We aim to communicate the team's answers for each paper or project to evaluators before they write their evaluations.
Consider: , field relevance, open science, authors’ engagement, data and reasoning transparency. In gauging this relevance, the team may consider the , but not too rigidly.
What are (some of) the authors’ main claims that are worth carefully evaluating? What aspects of the evidence, argumentation, methods, interpretation, etc., is the team unsure about? What particular data, code, proof, etc., would they like to see vetted? If it has already been peer-reviewed in some way, why do they think more review is needed?
How well has the author engaged with the process? Do they need particular convincing? Do they need help making their engagement with The Unjournal successful?
See for further discussion of prioritization, scope, and strategic and sustainability concerns.
Research can be "submitted" by authors (here) or "suggested" by others. For a walk-through on suggesting research, see this video example.
There are two main paths for making suggestions: through our survey form or through Airtable.
Anyone can suggest research using the survey form at . (Note, if you want to "submit your own research," go to .) Please include the following steps:
Begin by reviewing to get a sense of the research we cover and our priorities. Look for high-quality research that 1) falls within our focus areas and 2) would benefit from (further) evaluation.
When in doubt, we encourage you to suggest the research anyway.
Navigate to The Unjournal's . Most of the fields here are optional. The fields ask the following information:
Who you are: Let us know who is making the suggestion (you can also choose to stay anonymous).
If you leave your contact information, you will be eligible for financial "bounties" for strong suggestions.
If you are already a member of The Unjournal's team, additional fields will appear for you to link your suggestion to your profile in the Unjournal's database.
Complete all the required fields and submit your suggestion. The Unjournal team will review your submission and consider it for future evaluation. You can reach out to us at with any questions or concerns.
People on our team may find it more useful to suggest research to The Unjournal directl via the Airtable. See for a guide to this. (Please request document permission to access this explanation.)
Aside on setting the prioritization ratings: In making your subjective prioritization rating, please consider “What percentile do you think this paper (or project) is relative to the others in our database, in terms of ‘relevance for The UJ to evaluate’?” (Note this is a redefinition; we previously considered these as probabilities.) We roughly plan to commission the evaluation of about 1 in 5 papers in the database, the ‘top 20%’ according to these percentiles. Please don’t consider the “publication status or the “author's propensity to engage” in this rating. We will consider those as separate criteria.
Please don’t enter only the papers you think are ‘very relevant’; please enter in all research that you have spent any substantial time considering (more than a couple minutes). If we all do this, we should all aim for our percentile ratings to be approximately normally distributed; evenly spread over the 1-100% range.
Research Label: Provide a short, descriptive label for the research you are suggesting. This helps The Unjournal quickly identify the topic at a glance.
Research Importance: Explain why the research is important, its potential impact, and any specific areas that require thorough evaluation.
Research Link: Include a direct URL to the research paper. The Unjournal prefers research that is publicly hosted, such as in a working paper archive or on a personal website.
Peer Review Status: Inform about the peer review status of the research, whether it's unpublished, published without clear peer review, or published in a peer-reviewed journal.
"Rate the relevance": This represents your best-guess at how relevant this work is for The Unjournal to evaluate, as a percentile relative to other work we are considering.
Research Classification: Choose categories that best describe the research. This helps The Unjournal sort and prioritize suggestions.
Field of Interest: Select the outcome or field of interest that the research addresses, such as global health in low-income countries.
As noted in Process: prioritizing research, we ask people who suggest research to provide a numerical 0-100 rating:
We also ask people within our team to act as 'assessors' to give as second and third opinions on this. This 'prioritization rating' is one of the criteria we will use to determine whether to commission research to be evaluated (along with author engagement, publication status, our capacity and expertise, etc.) Again, see the previous page for the current process.
We are working on a set of notes on this, fleshing this out and giving specific examples. At the moment this is available to members of our team only (ask for access to "Guidelines for prioritization ratings (internal)"). We aim to share a version of this publicly once it converges, and once we can get rid of arbitrary sensitive examples.
I. This is not the evaluation itself. It is not an evaluation of the paper's merit per se:
Influential work, and prestigious work in influential areas may be highly prioritized regardless of its rigor and quality
The prioritization rating might consider quality for work that seems potentially impactful, which does not seem particularly prestigious or influential. Here aspects like writing clarity, methodological rigor, etc., might put it 'over the bar'. However, even here these will tend to be based on rapid and shallow assessments, and should not be seen as meaningful evaluations of research merit.
II. These ratings will be considered along with the discussion by the field team and the management. Thus is helpful if you give a justification and explanation for your stated rating.
Define/consider the following ‘attributes’ of a piece of research:
Global decision-relevance/VOI: Is this research decision-relevant to high-value choices and considerations that are important for global priorities and global welfare?\
Prestige/prominence: Is the research already prominent/valued (esp. in academia), highly cited, reported on, etc?\
Influence: Is the work already influencing important real-world decisions and considerations?
Obviously, these are not binary factors; there is a continuum for each. But for the sake of illustration, consider the following flowcharts.
If the flowcharts do not render, please refresh your browser. You may have to refresh twice.
"Fully baked": Sometimes prominent researchers release work (e.g., on NBER) that is not particularly rigorous or involved, which may have been put together quickly. This might be research that links to a conference they are presenting at, to their teaching, or to specific funding or consulting. It may be survey/summary work, perhaps meant for less technical audiences. The Unjournal tends not to prioritize such work, or at least not consider it in the same "prestigious" basket (although there will be exceptions). In the flowchart above, we contrast this with their "fully-baked" work.
Decision-relevant, prestigious work: Suppose the research is both ‘globally decision-relevant’ and prominent. Here, if the research is in our domain, we probably want to have it publicly evaluated. This is basically the case regardless of its apparent methodological strength. This is particularly true if it has been recently made public (as a working paper), if it has not yet been published in a highly-respected peer-reviewed journal, and if there are non-straightforward methodological issues involved. \
Prestigious work that seems less globally-relevant: We generally will not prioritize this work unless it adds to our mission in other ways (see, e.g., our ‘sustainability’ and ‘credibility’ goals ). In particular we will prioritize such research more if:
It is presented in innovative, transparent formats (e.g., dynamic documents/open notebooks, sharing code and data)
The research indirectly supports more globally-relevant research, e.g., through…
Providing methodological tools that are relevant to that ‘higher-value’ work
(If the flowchart below does not render, please refresh your browser; you may have to refresh twice.)
Decision-relevant, influential (but less prestigious) work: E.g., suppose this research might be cited by a major philanthropic organization as guiding its decision-making, but the researchers may not have strong academic credentials or a track record. Again, if this research is in our domain, we probably want to have it publicly evaluated. However, depending on the rigor of the work and the way it is written, we may want to explicitly class this in our ‘non-academic/policy’ stream.\
Decision-relevant, less prestigious, less-influential work: What about for less-prominent work with fewer academic accolades that is not yet having an influence, but nonetheless seems to be globally decision-relevant? Here, our evaluations seem less likely to have an influence unless the work seems potentially strong, implying that our evaluations, rating, and feedback could boost potentially valuable neglected work. Here, our prioritization rating might focus more on our initial impressions of things like …
Methodological strength (this is a big one!)
Rigorous logic and communication
Open science and robust approaches
Again: the prioritization process is not meant to be an evaluation of the work in itself. It’s OK to do this in a fairly shallow way.
In future, we may want to put together a loose set of methodological ‘suggestive guidelines’ for work in different fields and areas, without being too rigid or prescriptive. (To do: we can draw from some existing frameworks for this [ref].)
\
We acknowledge the potential for "information hazards" when research methods, tools, and results become more accessible. This is of particular concern in the context of direct physical and biological science research, particularly in (although there is a case that specific ). ML/AI research may also fall into this category. Despite these potential risks, we believe that the fields we plan to cover—detailed above—do not primarily present such concerns.
In cases where our model might be extended to high-risk research—such as new methodologies contributing to terrorism, biological warfare, or uncontrolled AI—the issue of accessibility becomes more complex. We recognize that increasing accessibility in these areas might potentially pose risks.
While we don't expect these concerns to be raised frequently about The Unjournal's activities, we remain committed to supporting thoughtful discussions and risk assessments around these issues.
See also...
Claim: Rating and feedback is better than an ‘all-or-nothing’ accept/reject process. Although people like to say “peer review is not binary”, the consequences are.
“Publication in a top journal” is used as a signal and a measuring tool for two major purposes. First, policymakers, journalists, and other researchers look at where a paper is published to assess whether the research is credible and reputable. Second, universities and other institutions use these publication outcomes to guide hiring, tenure, promotion, grants, and other ‘rewards for researchers.’
Did you know?: More often than not, of the "supply of spaces in journals” and the “demand to publish in these journals”. Who is the consumer? Certainly not the perhaps-mythical creature known as the ‘reader’.
Perhaps not. We are accustomed to using ratings as filters in our daily lives. Readers, grantmakers, and policymakers can set their own threshold. They could disregard papers and projects that fail to meet, for instance, a standard of at least two peer reviews, an average accuracy rating above 3, and an average impact rating exceeding 4.
In the field of economics, between the ‘first working paper’ that is publicly circulated and the final publication. During that time, the paper may be substantially improved, but it may not be known to nor accepted by practitioners. Meanwhile, it provides little or no career value to the authors.
As a result, we see three major downsides:
Time spent gaming the system:
Researchers and academics spend a tremendous amount of time 'gaming' this process, at the expense of actually doing better research.
Randomness in outcomes, unnecessary uncertainty and stress
Wasted feedback, including reviewer's time
I (Reinstein) have been in academia for about 20 years. Around the departmental coffee pot and during research conference luncheons, you might expect us to talk about theories, methods, and results. But roughly half of what we talk about is “who got into which journal and how unfair it is”; “which journal should we be submitting our papers to?”; how long are their “turnaround times?”; “how highly rated are these journals?”; and so on. We even exchange on how to
There is a lot of pressure, and even bullying, to achieve these “publication outcomes” at the expense of careful methodology.
The current system can sideline deserving work due to unpredictable outcomes. There's no guarantee that the cream will rise to the top, making research careers much more stressful—even driving out more risk-averse researchers—and sometimes encouraging approaches that are detrimental to good science.
A lot of ‘feedback’ is wasted, including the . Some reviewers write ten-page reports critiquing the paper in great detail, even when they reject the paper. These reports are sometimes very informative and useful for the author and would also be very helpful for the wider public and research community to understand the nature of the debate and issues.
However, researchers often have a very narrow focus on getting the paper published as quickly and in as high-prestige a journal as possible. Unless the review is part of a 'Revise and Resubmit' that the author wants to fulfill, they may not actually put the comments into practice or address them in any way.
Of course, the reviews may be misinformed, mistaken, or may misunderstand aspects of the research. However, if the paper is rejected (even if the reviewer was positive about the paper), the author has no opportunity or incentive to respond to the reviewer. Thus the misinformed reviewer may remain in the dark.
The other side of the coin: a lot of effort is spent trying to curry favor with reviewers who are often seen as overly fussy and not always in the direction of good science.
John List (Twitter : "We are resubmitting a revision of our study to a journal and the letter to the editor and reporters is 101 pages, single-spaced. Does it have to be this way?"
Paola Masuzzo; “I was told that publishing in Nature/Cell/Science was more important than everything else.”
Anonymous; "This game takes away the creativity, the risk, the ‘right to fail’. This last item is for me, personally, very important and often underestimated. Science is mostly messy. Whoever tells us otherwise, is not talking about Science.”
Here's a rough sketch of the process and timings at top journals in economics. report an average of over 24 months between initial submisson and final acceptance (and nearly three years until publication).
David Reinstein, Nov 2024: Over the last six months we have considered and evaluated a small amount of work under this “Applied & Policy Stream”. We are planning to continue this stream for the forseeable future.
Much of the most impactful research is not aimed at academic audiences and may never be submitted to academic journals. It is written in formats that are very different from traditional academic outputs, and cannot be easily judged by academics using the same standards. Nonetheless, this work may use technical approaches developed in academia, making it important to gain expert feedback and evaluation.
The Unjournal can help here. However, to avoid confusion, we want to make this clearly distinct from our main agenda, which focuses on impactful academically-aimed research.
Our “Applied & Policy Stream” will be clearly labeled as separate from our main stream. This may constitute roughly 10 or 15% of the work that we cover. Below, we refer to this as the “applied stream” for brevity.
Our considerations for prioritizing this work are generally the same as for our academic stream – is it in the fields that we are focused on, using approaches that enable meaningful evaluation and rating? Is it already having impact (e.g., influencing grant funding in globally-important areas)? Does it have the potential for impact, and if so, is it high-quality enough that we should consider boosting its signal?
We will particularly prioritize policy and applied work that uses technical methods that need evaluation by research experts, often academics.
This could include a range of applied research from EA/GP/LT linked organizations such as GPI, Rethink Priorities, Open Philanthropy, FLI, HLI, Faunalytics, etc., as well as EA-adjacent organizations and relevant government white papers. \
Ratings/metrics: As in the academic stream, this work will be evaluated for its credibility, usefulness, communication/logic, etc. However, we are not seeking to have this work assessed by the standards of academia in a way that yields a comparison to traditional journal tiers. Evaluators: Please ignore these parts of our interface; if you are unsure if it is relevant feel free to ask.
Evaluator selection, number, pay: Generally we want to continue to select academic research experts or non-academic researchers with strong academic and methodological background to do these evaluations. A key purpose of this applied stream is largely to bring research expertise, particularly from academia, to work that is not normally scrutinized by such experts.
The compensation may be flexible as well; in some cases the work may be more involved than for the academic stream and in some cases less involved. As a starting point we will begin by offering the same compensation as for the academic stream.
Careful flagging and signposting: To preserve the reputation of our academic-stream evaluations we need to make it clear, wherever people might see this work, that it is not being evaluated by the same standards as the academic stream and doesn't “count” towards those metrics.
This research is more likely to fall into the category of , "already influencing a substantial amount of funding in impact-relevant areas, or substantially influencing policy considerations".
If the research itself is being funded by a global-impact focused foundation or donor, this will also constitute a strong prima facie reason to commission an evaluation (without requiring the authors' consent). See .
You can request a conditional embargo by emailing us at contact@unjournal.org, or via the submission/response form. Please explain what sort of embargo you are asking for, and why. By default, we'd like Unjournal evaluations to be made public promptly.
However, we may make exceptions in special circumstances particularly
for very early-career researchers who are not clearly ,
where the research is not obviously already influencing a substantial amount of funding in impact-relevant areas, or substantially influencing policy considerations
If there is an early-career researcher on the authorship team, we may allow authors to "embargo" the publication of the evaluation until a later date. Evaluators (referees) will be informed of this. This date can be contingent, but it should not be indefinite.
For example, we might grant an embargo that lasts until after a PhD/postdoc’s upcoming job market or until after publication in a mainstream journal, with a hard maximum of 14 months. (Of course, embargoes can be ended early at the request of the authors.)
In exceptional circumstances we may consider granting a "conditional indefinite embargo."
We will invite 2 or 3 relevant experts to evaluate and rate this work, letting them know about the following embargo
When the evaluations come back, we will ask if you want to respond/revise. If you commit to responding (please let us know your plan within 1 week):
We will invite 2 or 3 relevant experts to evaluate and rate this work, letting them know about the following embargo
When the evaluations come back..., we will ask if you want to respond.
We will invite 2 or 3 relevant experts to evaluate and rate this work, letting them know about the following embargo
When the evaluations come back. If all evaluators gave a 4.5 rating or higher as their middle rating on the "" rating (basically suggesting they think it's at the level meriting publication in a top-5+ journal) we will give you 3 weeks to respond before posting the package. (This is roughly our usual policy)
Note: the above are all exceptions to our regular rules, examples of embargos we might or might not agree to.
Journal-independent review allows work to be rated separately in different areas: theoretical rigor and innovation, empirical methods, policy relevance, and so on, with separate ratings in each category by experts in that area. As a researcher in the current system, I cannot both submit my paper and get public evaluation from (for example) JET and the Journal of Development Economics for a paper engaging both areas.
The Unjournal, and journal-independent evaluation, can enable this through
commissioning a range of evaluators with expertise in distinct areas, and making this expertise known in the public evaluations;
we will make it public that the evaluations are complete, and you have committed to revise and respond.
We will give you 8 weeks to revise the paper, to write a response note how you have revised,
We will give the evaluators additional time to adjust their evaluations and ratings in response to your revision/response
After this we will publish the evaluation package
If you do not commit to responding, we will post the evaluation package
If you are happy with the evaluations, we can post them at any time, by your request.
If all evaluators gave a 4.5 rating or higher as their middle rating on the "" rating (basically suggesting they think it's at the level meriting publication in a top-5+ journal) we will give you 3 weeks to respond before posting the package. (This is roughly our usual policy)
Otherwise (if any rate below 4.5 but none rate it below 3.25) we will give you 8 weeks to revise the paper in response to this, to write a response noting how you have responded. We will give the evaluators further time to adjust their evaluations and ratings in turn, before posting the evaluation package.
If any evaluators rate the paper 'fairly negatively' (below 3.25) on this measure, we will grant a six month embargo from this point, before posting the package. During this time you will also have the opportunity to revise and respond, as in the previous case (case 2.2).
If you are happy with the evaluations, we can post them at any time, by your request.
Otherwise we will wait to post the evaluations until June 15, or until all PhD student or Post-doc authors have found a new job (as reported on social media, LinkedIn etc)
During the intervening time, you have the opportunity to revise and respond, and if you do we give the evaluators time to update their evaluations and ratings in turn.
If you are happy with the evaluations, we can post them at any time, by your request.
allowing authors to gain evaluation in particular areas in addition to the implicit value of publication in specific traditional field journals.
We will consider research in almost any format — it does not need to be a frozen pdf or even a linear paper: see Promoting 'Dynamic Documents' and 'Living Research Projects'
We will only evaluate research objects that are openly accessible without paywalls or logins. The NBER working paper series is a borderline case, as there are some limits to access, but we're covering it for now because of it's importance and because there are workarounds to the access limits.
For some fields, the peer review process is rapid and perhaps lower intensity; here 'post-peer-review evaluation' makes sense. In fields like economics, with very slow and high-intensity peer-review, we mostly consider evaluating work before it's been accepted in a peer-reviewed journal. I.e., 'working papers'. In such cases, our evaluations may contribute less to some parts of our (efficiency, author feedback and revision, early credibility signals). See
But we sometimes consider (economics) papers already published in peer-reviewed journals, if we still see substantial value in having this work publicly evaluated and rated.
In particular, we may evaluate a 'journal-published paper' if
The paper significantly influences policy or funding decisions and therefore merits additional, transparent scrutiny.
Our evaluation process supports broader goals, such as fostering an impactful research community or refining our evaluation methods.
We have reasons to suspect that the journal's focus or field may be less likely to consider critical aspects of the paper, such as empirical robustness or policy relevance. E.g., a natural-science focused journal may neglect issues of causal inference that are more familiar to quantitative social scientists/economists.
31 Aug 2023: Our present approach is a "working solution" involving some ad-hoc and intuitive choices. We are re-evaluating the metrics we are asking for as well as the interface and framing. We are gathering some discussion in this linked Gdoc, incorporating feedback from our pilot evaluators and authors. We're also talking to people with expertise as well as considering past practice and other ongoing initiatives. We plan to consolidate that discussion and our consensus and/or conclusions into the present (Gitbook) site.
Ultimately, we're trying to replace the question of "what tier of journal did a paper get into?" with "how highly was the paper rated?" We believe this is a more valuable metric. It can be more fine-grained. It should be less prone to gaming. It aims to reduce randomness in the process, through things like 'the availability of journal space in a particular field'. See our discussion of .
To get to this point, we need to have academia and stakeholders see our evaluations as meaningful. We want the evaluations to begin to have some value that is measurable in the way “publication in the AER” is seen to have value.
While there are some ongoing efforts towards journal-independent evaluation, these tend not use comparable metrics. Typically, they either have simple tick-boxes (like "this paper used correct statistical methods: yes/no") or they enable descriptive evaluation without an overall rating. As we are not a journal, and we don’t accept or reject research, we need another way of assigning value. We are working to determine the best way of doing this through quantitative ratings. We hope to be able to benchmark our evaluations to "traditional" publication outcomes. Thus, we think it is important to ask for both an overall quality rating and a journal ranking tier prediction.
In addition to the overall assessment, we think it will be valuable to have the papers rated according to several categories. This could be particularly helpful to practitioners who may care about some concerns more than others. It also can be useful to future researchers who might want to focus on reading papers with particular strengths. It could be useful in meta-analyses, as certain characteristics of papers could be weighed more heavily. We think the use of categories might also be useful to authors and evaluators themselves. It can help them get a sense of what we think research priorities should be, and thus help them consider an overall rating.
However, these ideas have been largely ad-hoc and based on the impressions of our management team (a particular set of mainly economists and psychologists). The process is still being developed. Any feedback you have is welcome. For example, are we overemphasizing certain aspects? Are we excluding some important categories?
We are also researching other frameworks, templates, and past practice; we hope to draw from validated, theoretically grounded projects such as .
In eliciting expert judgment, it is helpful to differentiate the level of confidence in predictions and recommendations. We want to know not only what you believe, but how strongly held your beliefs are. If you are less certain in one area, we should weigh the information you provide less heavily in updating our beliefs. This may also be particularly useful for practitioners. Obviously, there are challenges to any approach. Even experts in a quantitative field may struggle to convey their own uncertainty. They may also be inherently "poorly calibrated" (see discussions and tools for ). Some people may often be "confidently wrong." They might state very narrow "credible intervals", when the truth—where measurable—routinely falls outside these boundaries. People with greater discrimination may sometimes be underconfident. One would want to consider and potentially correct for poor calibration. As a side benefit, this may be interesting for research in and of itself, particularly as The Unjournal grows. We see 'quantifying one's own uncertainty' as a good exercise for academics (and everyone) to engage in.
2 Oct 2023 -- We previously suggested 'weightings' for individual ratings, along with a note
We give "suggested weights" as an indication of our priorities and a suggestion for how you might average these together into an overall assessment; but please use your own judgment.
We included these weightings for several reasons:
We have removed suggested weightings for each of these categories. We discuss the rationale at some length .
Evaluators working before October 2023 saw a previous version of the table, which you can see .
We previously gave evaluators two options for expressing their confidence in each rating:
Either:
The 90% Confidence/Credible Interval (CI) input you see below (now a 'slider' in PubPub V7) or
We had included the note:
We give the previous weighting scheme in a fold below for reference, particularly for those reading evaluations done before October 2023.
As well as:
Suggested weighting: 0. Why 0?
Elsewhere in that page we had noted:
As noted above, we give suggested weights (0–5) to suggest the importance of each category rating to your overall assessment, given The Unjournal's priorities. But you don't need, and may not want to use these weightings precisely.
The weightings were presented once again along with each description in the section .
[FROM PREVIOUS GUIDELINES:]
You may feel comfortable giving your "90% confidence interval," or you may prefer to give a "descriptive rating" of your confidence (from "extremely confident" to "not confident").
Quantify how certain you are about this rating, either giving a 90% / interval or using our . (We prefer the 90% CI. Please don't give both.
5 = Extremely confident, i.e., 90% confidence interval spans +/- 4 points or less
4 = Very confident: 90% confidence interval +/- 8 points or less
3 = Somewhat confident: 90% confidence interval +/- 15 points or less
2 = Not very confident: 90% confidence interval, +/- 25 points or less
1 = Not confident: (90% confidence interval +/- more than 25 points)
[Previous...] Remember, we would like you to give a 90% CI or a confidence rating (1–5 dots), but not both.
The example in the diagram above (click to zoom) illustrates the proposed correspondence.
And, for the 'journal tier' scale:
From "five dots" to "one dot":
5 = Extremely confident, i.e., 90% confidence interval spans +/– 4 points or less*
4 = Very confident: 90% confidence interval +/– 8 points or less
3 = Somewhat confident: 90% confidence interval +/– 15 points or less
[Previous guidelines]: The description folded below focuses on the "Overall Assessment." Please try to use a similar scale when evaluating the category metrics.
95–100: Among the highest quality and most important work you have ever read.
90–100: This work represents a major achievement, making substantial contributions to the field and practice. Such work would/should be weighed very heavily by tenure and promotion committees, and grantmakers.
For example:
This work represents a strong and substantial achievement. It is highly rigorous, relevant, and well-communicated, up to the standards of the strongest work in this area (say, the standards of the top 5% of committed researchers in this field). Such work would/should not be decisive in a tenure/promotion/grant decision alone, but it should make a very solid contribution to such a case.
60–74.9: A very strong, solid, and relevant piece of work. It may have minor flaws or limitations, but overall it is very high-quality, meeting the standards of well-respected research professionals in this field.
40–59.9: A useful contribution, with major strengths, but also some important flaws or limitations.
20–39.9: Some interesting and useful points and some reasonable approaches, but only marginally so. Important flaws and limitations. Would need substantial refocus or changes of direction and/or methods in order to be a useful part of the research and policy discussion.
5–19.9: Among the lowest quality papers; not making any substantial contribution and containing fatal flaws. The paper may fundamentally address an issue that is not defined or obviously not relevant, or the content may be substantially outside of the authors’ field of expertise.
0–4: Illegible, fraudulent, or plagiarized. Please flag fraud, and notify us and the relevant authorities.
The previous categories were 0–5, 5–20, 20–40, 40–60, 60–75, 75–90, and 90–100. Some evaluators found the overlap in this definition confusing.
This page explains the value of the metrics we are seeking from evaluators.
We are considering asking evaluators, with compensation, to assist and engage in the process of "robustness replication." This may lead to some interesting follow-on possibilities as we build our potential collaboration with the Institute for Replication and others in this space.
We might ask evaluators discussion questions like these:
What is the most important, interesting, or relevant substantive claim made by the authors, (particularly considering global priorities and potential interventions and responses)?
What statistical test or evidence does this claim depend on, according to the authors?
How confident are you in the substantive claim made?
"Robustness checks": What specific statistical test(s) or piece(s) of evidence would make you substantially more confident in the substantive claim made?
If a robustness replication "passed" these checks, how confident would you be then in the substantive claim? (You can also express this as a continuous function of some statistic rather than as a binary; please explain your approach.)
Background:
The Institute for Replication is planning to hire experts to do "robustness-replications" of work published in a top journal in economics and political science. Code- and data sharing is now being enforced in many or all of these journals and other important outlets. We want to support their efforts and are exploring collaboration possibilities. We are also considering how to best guide potential future robustness replication work.
We wanted to make the overall rating better defined and thus more useful to outsiders and comparable across raters
Emphasizing what we think is important (in particular, methodological reliability)
We didn't want evaluators to think we wanted them to weigh each category equally … some are clearly more important
However, we decided to remove these weightings because:
Reduce clutter in an already overwhelming form and guidance doc. ‘More numbers’ can be particularly overwhelming
These weights were ad-hoc, and they may suggest we have a more grounded ‘model of value’ than we already do. (And there is also some overlap in our categories anyways, something we are working on addressing.)
Some people interpreted what we intended incorrectly (e.g., they thought we were saying ‘relevance to global priorities’ is not an important thing)
A five-point 'Likert style' measure of confidence, which we described qualitatively and explained how we would convert it into CIs when we report aggregations.
To make this process less confusing, to encourage careful quantification of uncertainty, and to enable better-justified aggregation of expert judgment, we are de-emphasizing the latter measure.
Still, to accommodate those who may not be familiar with or comfortable stating "90% CIs on their own beliefs" we offer further explanations, and we are providing tools to help evaluators construct these. As a fallback, we will still allow evaluators to give the 1-5 confidence measure, noting the correspondence to CIs, but we discourage this somewhat.
The previous guidelines can be seen here; these may be useful in considering evaluations provided pre-2024.
5
45, 55
4
10, 35
3
40, 70
2
30,46
0**
21,65
10, 35
40, 70
30,46
21,65
1 = Not confident: 90% confidence interval +/– 25 points
This paper is substantially more rigorous or more insightful than existing work in this area in a way that matters for research and practice.
The work makes a major, perhaps decisive contribution to a case for (or against) a policy or philanthropic intervention.
Why these guidelines/metrics?(holistic, most important!)
39, 52
5
47, 54
Why these guidelines/metrics?(holistic, most important!)
39, 52
47, 54


45, 55
In addition to soliciting research submissions by authors, we have a process for sourcing and prioritizing unsubmitted research for evaluation. For some of this research we ask for author engagement but do not require their permission.
Choose a set of "top-tier working paper series" and medium-to-top-tier journals, as well as research posted in other exclusive working paper archives and to work where all authors seem to be prominent, secure, and established. See .
Identify relevant papers in this series, following our stated criteria (i.e., , strength, ). For NBER this tends to include
recently released work in the early stages of the journal peer-review process, particularly if it addresses a timely subject; as well as
Public benefit: Working papers (especially NBER) are already influencing policy and debate, yet they have not been peer-reviewed and may take years to go through this process, if ever (e.g., many NBER papers). However, it is difficult to understand the papers' limitations unless you happen to have attended an academic seminar where they were presented. Evaluating these publicly will provide a service.
1. Negative backlash: Some authors may dislike having their work publicly evaluated, particularly when there is substantial criticism. Academics complain a lot about unfair peer reviews, but the difference is that here the evaluations will be made public. This might lead The Unjournal to be the target of some criticism.
Responses:
Public engagement in prominent and influential work is fair and healthy. It is good to promote public intellectual debate. Of course, this process needs to allow constructive criticism as well as informative praise.
Aside: in the future, we hope to work directly with working paper series, associations, and research groups to get their approval and engagement with Unjournal evaluations. We hope that having a large share of papers in your series evaluated will serve as a measure of confidence in your research quality. If you are involved in such a group and are interested in this, please reach out to us ().
Dec 7, 2024: We have updated some of our rules and guidelines on this page. These will be applied going forward (in future contacts with authors) but not retroactively.
All NBER working papers are generally eligible, with rare exceptions.
We treat these on a case-by-case basis and use discretion. All CEPR members are reasonably secure and successful, but their co-authors might not be, especially if these co-authors are PhD students they are supervising.
In some areas and fields (e.g., psychology, animal product markets) the publication process is relatively rapid or it may fail to engage general expertise. In general, all papers that are already published in peer-reviewed journals are eligible for our direct track.
(We discuss when we consider post-pub evaluation in )
These are eligible (without author permission) if all authors/all lead authors "have high professional status" or are otherwise less career-sensitive to the consequences of this evaluation.
We define this (at least for economics) as:
having a tenured or ‘long term’ positions at well-known, respected universities or other research institutions, or
having a tenure-track positions at a top universities (e.g., top-20 globally by some credible ranking) and having published one or more papers in a "top-five-equivalent" journal, or
clearly not pursuing an academic career (e.g., the "partner at the aid agency running the trial").
On the other hand, if one or more authors is a PhD student close to graduation or an untenured academic outside a "top global program,’’ then we will ask for permission and potentially offer an embargo.
If the PhD student or untenured academic is otherwise clearly extremely high-performing by conventional metrics; e.g., an REStud "tourist" or someone with multiple published papers in top-5 journals. In such cases the paper might be considered eligible for direct evaluation.
See also
Under review/R&R at a journal? The fact that the paper is under submission or in "revise and resubmit" at a top journal does not preclude us from evaluating it. In some cases it may be particular important and helpful to evaluate work at this stage. But we'd like to be aware of this, as it can weigh into our considerations and timing.
We will also evaluate work directly, without requiring author permission, where it is clear that this research is already influencing a substantial amount of funding in impact-relevant areas, or substantially influencing policy considerations. Much of this work will be evaluated as part of our .
The Unjournal is a nonprofit organization started in mid-2022. We commission experts to publicly evaluate and rate research. Read more about us here.
Write an evaluation of a specific research paper or project: essentially a standard, high-quality referee report.
Give quantitative ratings and predictions about the research by filling in a structured form.
Answer a short questionnaire about your background and our processes.
See Guidelines for Evaluators for further details and guidance.
Why use your valuable time writing an Unjournal evaluation? There are several reasons: helping high-impact research users, supporting open science and open access, and getting recognition and financial compensation.
The Unjournal's goal is to make impactful research more rigorous, and rigorous research more impactful, while supporting open access and open science. We encourage better research by making it easier for researchers to get feedback and credible ratings. We evaluate research in high-impact areas that make a difference to global welfare. Your evaluation will:
Help authors improve their research, by giving early, high-quality feedback.
Help improve science by providing open-access, prompt, structured, public evaluations of impactful research.
Inform funding bodies and meta-scientists as we build a database of research quality, strengths and weaknesses in different dimensions. Help research users learn what research to trust, when, and how.
For more on our scientific mission, see here.
Your evaluation will be made public and given a DOI. You have the option to be identified as the author of this evaluation or to remain anonymous, as you prefer.
You will be given a $200-$400 honorarium for providing a prompt and complete evaluation and feedback ($100-$300 base + $100 'promptness bonus') in line with our expected standards.
Our current baseline compensation has two tiers, aimed to reward strong previous work doing public evaluations and reviews for us and for others. These tiers are not about academic seniority or credentials.
Minimum base compensation tiers:
$100 + $100 for first-time evaluators without demonstrated public review experience
$200-300 + $100 for return Unjournal evaluators and those with previous strong public review experience (for The Unjournal or through other initiatives).
We will be integrating other incentives and prizes into this, and are committed to $450 in average compensation per evaluation, including prizes.
Aug 21, 2025: We are still committed to a $450 target in the medium-run. However, because of temporary funding concerns, we are adjusting this to a $350 target for (on-time and complete) evaluations that we are commissioning in the near future. Watch this space for updates.
You will also be eligible for monetary prizes for "most useful and informative evaluation," plus other bonuses.
See also "submitting claims and expenses".
In addition public evaluations and referee reports, we can accept critical syntheses and literature review papers and essays as example of evaluation experience. You can also share with us an example of a previous strong referee report you have written, that would be suitable for making public given the required permissions. (Also see Reviewers from previous journal submissions for a discussion of publicly sharing these).
If you have been invited to be an evaluator and want to proceed, simply respond to the email invitation that we have sent you. You will then be sent a link to our evaluation form.
To sign up for our evaluator pool, see 'how to get involved'
To learn more about our evaluation process, seeGuidelines for evaluators. If you are doing an evaluation, we highly recommend you read these guidelines carefully
12 Feb 2024: We are moving to a hosted form/interface in PubPub. That form is still somewhat a work-in-progress, and may need some further guidance; we try to provide this below, but please contact us with any questions. If you prefer, you can also submit your response in a Google Doc, and share it back with us. Click here to make a new copy of that directly.
work that has been around for many years, is widely cited and influential, yet has never been published in a peer-reviewed journal.
We do this systematically and transparently; authors shouldn't feel singled out nor left out.
Notify the work's authors that The Unjournal plans to commission evaluations. We're not asking for permission, but
making them aware of The Unjournal, the process, the benefits to authors, and the authors' opportunities to engage with the evaluation and publicly respond to the evaluation before it is made public;
letting us know if we have the most recent version of the paper, and if updates are coming soon;
letting the authors complete our forms if they wish, giving further information about the paper or e.g. adding a "permalink" to updated versions;
asking if there are authors in sensitive career positions justifying a; and
asking the authors if there is specific feedback they would like to receive.
Reaching out to and commissioning evaluators, as in our regular process. Considerations:
Evaluators should be made aware that the authors have not directly requested this review, but have been informed it is happening.
As this will allow us to consider a larger set of papers more quickly, we can reach out to multiple evaluators more efficiently.
Fear of public evaluation (safety in numbers): There may be some shyness or reluctance to participate in The Unjournal evaluation process (for reasons to do so, see our benefits to authors discussion). It is scary to be a first mover, and it may feel unfair to be among the few people to have an evaluation of your work out there in public (in spite of the Bayesian arguments presented in the previous link). There should be "safety" in numbers: having a substantial number of prominent papers publicly evaluated by The Unjournal will ease this concern.
Passive evaluation may be preferred to active consent: Academics (especially early-career) may also worry that they will seem weird or rebellious by submitting to The Unjournal, as this may be taken as "rejecting mainstream system norms." Again, this will be less of a problem if a substantial number of public evaluations of prominent papers are posted. You will be in good company. Furthermore, if we are simply identifying papers for evaluation, the authors of these papers cannot be seen as rejecting the mainstream path (as they did not choose to submit).
Piloting and building a track record or demonstration: The Unjournal needs a reasonably large set of high-quality, relevant work to evaluate in order to help us build our system and improve our processes. Putting out a body of curated evaluation work will also allow us to demonstrate the reasonableness and reliability of this process.
We will work to ensure that the evaluations we publish involve constructive dialogue, avoid unnecessary harshness, and provide reasons for their critiques. We also give authors the opportunity to respond.
We are focusing on more prominent papers, with authors in more secure positions. Additionally, we offer a potential "embargo" for sensitive career situations, e.g., those that might face early-career researchers.
2. Less author engagement: If authors do not specifically choose to have their work evaluated, they are less likely to engage fullly with the process.
Response: This is something we will keep an eye on, weighing the benefits and costs.
3. Evaluator/referee reluctance: As noted above, evaluators may be more reluctant to provide ratings and feedback on work where the author has not instigated the process.
Response: This should largely be addressed by the fact that we allow evaluators to remain anonymous. A potential cost here is discouraging signed evaluations, which themselves have some benefits (as well as possible costs).
4. Slippery-slope towards "unfairly reviewing work too early": In some fields, working papers are released at a point where the author does not wish them to be evaluated, and where the author is not implicitly making strong claims about the validity of this work. In economics, working papers tend to be released when they are fairly polished and the authors typically seek feedback and citations. The NBER series is a particularly prominent example.
Response: We will be careful with this. Initially, we started this 'direct evaluation' process only for the NBER series. We later extended this, as noted above.
This page describes The Unjournal's evaluation guidelines, considering our priorities and criteria, the metrics we ask for, and how these are considered.
These guidelines apply to the publicly visible forms currently hosted in Coda here (academic stream) and here (applied stream), as well as to our (legacy) PubPub forms. You can also preview our draft evaluation interface (in development, not yet in use).
Please see for an overview of the evaluation process, as well as details on compensation, public recognition, and more.
Write an evaluation of the paper or project. To some extent, this resembles a high-quality referee report for a traditional journal without the binary focus on 'should we accept or reject?'. Below, we describe some of our values and emphases. We also value insights for less-technical practitioners, especially in your evaluation 'abstract'.
In writing your evaluation and providing ratings, please consider the following.
In many ways, the written part of the evaluation should be similar to a report an academic would write for a traditional high-prestige journal (e.g., see some 'conventional guidelines' ). Most fundamentally, we want you to use your expertise to critically assess the main claims made by the authors. Are they well-supported? Are the assumptions believable? Are the methods appropriate and well-executed? Explain why or why not.
However, we'd also like you to pay some consideration to our priorities, including
Advancing our knowledge and supporting practitioners
Justification, reasonableness, validity, and robustness of methods
Logic and communication, intellectual modesty, transparent reasoning
See our and the corresponding 'ratings' for more details on each of these. You don't need to structure your review according to these metrics, but please pay some attention to them.
Please pay attention to anything our managers and editors specifically suggested that to focus on. We may ask you to focus on specific areas of expertise. We may also forward specific feedback requests from authors.
Unless you were advised otherwise, this evaluation, including the review and quantitative metrics, will be given a DOI and, hopefully, will enter the public research conversation. Authors will be given two weeks to respond to the evaluations (and evaluators can adjust if any obvious oversights are found) before the evaluations, ratings, and responses are made public. You can choose whether you want to be identified publicly as an author of the evaluation.
If you have questions about the authors’ work, you can ask them anonymously: we will facilitate this.
We want you to evaluate the most recent/relevant version of the paper/project that you can access. If you see a more recent version than the one we shared with you, please let us know.
We may give early-career researchers the right to veto the publication of very negative evaluations or to embargo the release of these for a defined period. We will inform you in advance if this will be the case for the work you are evaluating.
You can reserve some "sensitive" content in your report to be shared with only The Unjournal management or only the authors, but we hope to keep this limited.
For a model of what we are looking for, see examples of Unjournal evaluations that we thought were particularly strong ("Prize winning and commended evaluations").
We designed this process to balance three considerations with three target audiences. Please consider each of these:
Crafting evaluations and ratings that help researchers and policymakers judge when and how to rely on this research. For Research Users.
Ensuring these evaluations of the papers are comparable to current journal tier metrics, to enable them to be used to determine career advancement and research funding. For Departments, Research Managers, and Funders.
Providing constructive feedback to Authors.
We discuss this, and how it relates to our impact and "theory of change", .
We accept that in the near-term an Unjournal evaluation may not be seen to have substantial career value.
Furthermore, work we are considering may tend be at an earlier stage. authors may submit work to us, thinking of this as a "pre-journal" step. The papers we select (e.g., from NBER) may also have been posted long before authors planned to submit them to journals.
This may make the 'feedback for authors' and 'assessment for research users' aspects more important, relative to traditional journals' role. However, in the medium-term, a positive Unjournal evaluation should gain credibility and career value. This should help make our evaluations an "endpoint" and an important "career goal" for a research paper, replacing or supplementing the current "which journal tier" metric.
We ask for a set of nine quantitative metrics. For each metric, we ask for a score and a 90% credible interval. We describe these in detail below. (We explain .)
For some questions, we ask for a percentile ranking from 0-100%. This represents "what proportion of papers in the reference group are worse than this paper, by this criterion". A score of 100% means this is essentially the best paper in the reference group. 0% is the worst paper. A score of 50% means this is the median paper; i.e., half of all papers in the reference group do this better, and half do this worse, and so on.
Here* the population of papers should be all serious research in the same area that you have encountered in the last three years.
For the applied and policy stream the reference group should be "all applied and policy research you have read that is aiming at a similar audience, and that has similar goals".
Here, we are mainly considering research done by professional researchers with high levels of training, experience, and familiarity with recent practice, who have time and resources to devote months or years to each such research project or paper. These will typically be written as 'working papers' and presented at academic seminars before being submitted to standard academic journals. Although no credential is required, this typically includes people with PhD degrees (or upper-level PhD students). Most of this sort of research is done by full-time academics (professors, post-docs, academic staff, etc.) with a substantial research remit, as well as research staff at think tanks and research institutions (but there may be important exceptions).
This is a judgment call. Some criteria to consider... First, does the work come from the same academic field and research subfield, and does it address questions that might be addressed using similar methods? Second, does it deal with the same substantive research question, or a closely related one? If the research you are evaluating is in a very niche topic, the comparison reference group should be expanded to consider work in other areas.
We are aiming for comparability across evaluators. If you suspect you are particularly exposed to higher-quality work in this category, compared to other likely evaluators, you may want to adjust your reference group downwards. (And of course vice-versa, if you suspect you are particularly exposed to lower-quality work.)
For each metric, we ask you to provide a 'midpoint rating' and a 90% credible interval as a measure of your uncertainty. Our interface provides slider bars to express your chosen intervals:
for more guidance on uncertainty, credible intervals, and the midpoint rating as the 'median of your belief distribution'.
The table below summarizes the percentile rankings.
Percentile ranking (0-100%)
Judge the quality of the research heuristically. Consider all aspects of quality, credibility, importance to future impactful applied research, and practical relevance and usefulness, importance to knowledge production, and importance to practice.
Do the authors do a good job of (i) stating their main questions and claims, (ii) providing strong evidence and powerful approaches to inform these, and (iii) correctly characterizing the nature of their evidence?
Percentile ranking (0-100%)
Are the methods used well-justified and explained; are they a reasonable approach to answering the question(s) in this context? Are the underlying assumptions reasonable?
Are the results and methods likely to be robust to reasonable changes in the underlying assumptions? Does the author demonstrate this?
Avoiding bias and (QRP): Did the authors take steps to reduce bias from opportunistic reporting and QRP? For example, did they do a strong pre-registration and pre-analysis plan, incorporate multiple hypothesis testing corrections, and report flexible specifications?
Percentile ranking (0-100%)
To what extent does the project contribute to the field or to practice, particularly in ways that are relevant to global priorities and impactful interventions?
(Applied stream: please focus on ‘improvements that are actually helpful’.)
Originality and cleverness should be weighted less than the typical journal, because The Unjournal focuses on impact. Papers that apply existing techniques and frameworks more rigorously than previous work or apply them to new areas in ways that provide practical insights for GP (global priorities) and interventions should be highly valued. More weight should be placed on 'contribution to GP' than on 'contribution to the academic field'.
Do the paper's insights inform our beliefs about important parameters and about the effectiveness of interventions?
Does the project add useful value to other impactful research?
We don't require surprising results; sound and well-presented null results can also be valuable.
Percentile ranking (0-100%)
Are the goals and questions of the paper clearly expressed? Are concepts clearly defined and referenced?
Is the reasoning "transparent"? Are assumptions made explicit? Are all logical steps clear and correct? Does the writing make the argument easy to follow?
Are the conclusions consistent with the evidence (or formal proofs) presented? Do the authors accurately state the nature of their evidence, and the extent it supports their main claims?
Are the data and/or analysis presented relevant to the arguments made? Are the tables, graphs, and diagrams easy to understand in the context of the narrative (e.g., no major errors in labeling)?
Percentile ranking (0-100%)
This covers several considerations:
Would another researcher be able to perform the same analysis and get the same results? Are the methods explained clearly and in enough detail to enable easy and credible replication? For example, are all analyses and statistical tests explained, and is code provided?
Is the source of the data clear?
Is the data made as available as is reasonably possible? If so, is it clearly labeled and explained??
Consistency
Do the numbers in the paper and/or code output make sense? Are they internally consistent throughout the paper?
Useful building blocks
Do the authors provide tools, resources, data, and outputs that might enable or enhance future work and meta-analysis?
Are the paper’s chosen topic and approach likely to be useful to
Does the paper consider real-world relevance and deal with policy and implementation questions? Are the setup, assumptions, and focus realistic?
Do the authors report results that are relevant to practitioners? Do they provide useful quantified estimates (costs, benefits, etc.) enabling practical impact quantification and prioritization?
Do they communicate (at least in the abstract or introduction) in ways policymakers and decision-makers can understand, without misleading or oversimplifying?
Percentile ranking (0-100%)
Could the paper's topic and approach potentially help inform
Most work in our will not be targeting academic journals. Still, in some cases it might make sense to make this comparison; e.g., if particular aspects of the work might be rewritten and submitted to academic journals, or if the work uses certain techniques that might be directly compared to academic work. If you believe a comparison makes sense, please consider giving an assessment below, making reference to our guidelines and how you are interpreting them in this case.
To help universities and policymakers make sense of our evaluations, we want to benchmark them against how research is currently judged. So, we would like you to assess the paper in terms of journal rankings. We ask for two assessments:
a normative judgment about 'how well the research should publish';
a prediction about where the research will be published.
Journal ranking tiers are on a 0-5 scale, as follows:
0/5: "Won't publish/little to no value". Unlikely to be cited by credible researchers
1/5: OK/Somewhat valuable journal
2/5: Marginal B-journal/Decent field journal
We give some example journal rankings , based on SJR and ABS ratings.
We encourage you to consider a non-integer score, e.g. 4.6 or 2.2.
As before, we ask for a 90% credible interval.
PubPub note
Journal ranking tier (0.0-5.0)
Assess this paper on the journal ranking scale described above, considering only its merit, giving some weight to the category metrics we discussed above.
Equivalently, where would this paper be published if:
the journal process was fair, unbiased, and free of noise, and that status, social connections, and lobbying to get the paper published didn’t matter;
journals assessed research according to the category metrics we discussed above.
Journal ranking tier (0.0-5.0)
If this work has already been published, and you know where, please report the prediction you would have given absent that knowledge.
We want policymakers, researchers, funders, and managers to be able to use The Unjournal's evaluations to update their beliefs and make better decisions. To do this well, they need to weigh multiple evaluations against each other and other sources of information. Evaluators may feel confident about their rating for one category, but less confident in another area. How much weight should readers give to each? In this context, it is useful to quantify the uncertainty.
But it's hard to quantify statements like "very certain" or "somewhat uncertain" – different people may use the same phrases to mean different things. That's why we're asking for you a more precise measure, your credible intervals. These metrics are particularly useful for meta-science and meta-analysis.
You are asked to give a 'midpoint' and a 90% credible interval. Consider this as the smallest interval that you believe is 90% likely to contain the true value. See the fold below for further guidance.
You may understand the concepts of uncertainty and credible intervals, but you might be unfamiliar with applying them in a situation like this one.
You may have a certain best guess for the "Methods..." criterion. Still, even an expert can never be certain. E.g., you may misunderstand some aspect of the paper, there may be a method you are not familiar with, etc.
Your uncertainty over this could be described by some distribution, representing your beliefs about the true value of this criterion. Your "'best guess" should be the central mass point of this distribution.
You are also asked to give a 90% credible interval. Consider this as
We also ask for the 'midpoint', the center dot on that slider. Essentially, we are asking for the median of your belief distribution. By this we mean the percentile ranking such that you believe "there's a 50% chance that the paper's true rank is higher than this, and a 50% chance that it actually ranks lower than this."
If you are "", your 90% credible intervals should contain the true value 90% of the time. To understand this better, assess your ability, and then practice to get better at estimating your confidence in results. will help you get practice at calibrating your judgments. We suggest you choose the "Calibrate your Judgment" tool, and select the "confidence intervals" exercise, choosing 90% confidence. Even a 10 or 20 minute practice session can help, and it's pretty fun.
We are now asking evaluators for “claim identification and assessment” where relevant. This is meant to help practitioners use this research to inform their funding, policymaking, and other decisions. It is not intended as a metric to judge the research quality per se. This is not required but we will reward this work.
.
Lastly, we ask evaluators about their background, and for feedback about the process.
For the two questions below, we will publish your responses unless you specifically ask these questions to be kept anonymous.
How long have you been in this field?
If you prefer, you can submit your response in a Google Doc, and share it back with us. Click to make a new copy of that directly.
Length/time spent: This is up to you. We welcome detail, elaboration, and technical discussion.
recommends a 2–3 page referee report; suggest this is relatively short, but confirm that brevity is desirable. , economists report spending (median and mean) about one day per report, with substantial shares reporting "half a day" and "two days." We expect that reviewers tend spend more time on papers for high-status journals, and when reviewing work that is closely tied to their own agenda.
We have made some adjustments to this page and to our guidelines and processes; this is particularly relevant for considering earlier evaluations. See .
If you still have questions, please contact us, or see our FAQ on .
Our data protection statement is linked .
The flowchart below focuses on the evaluation part of our process in detail. See Evaluation workflow – Simplified for a more condensed flowchart.
Video overview: Two minute overview of our 6-step evaluation process.
Interactive presentation: For a visual walk-through of our evaluation process, see our step-by-step slide presentation (use arrow keys to navigate).
(Section updated 1 August 2023)
Submission/selection (multiple routes)
Author (A) submits work (W), creates new submission (submits a URL and DOI), through our platform or informally.
Author (or someone on their behalf) can complete a submission form; this includes a potential "request for embargo" or other special treatment.
Note that we intend to automate and integrate many of the process into an editorial-management-like system in PubPub.
In our current (8 Feb 2023 pilot) phase, we have the evaluators consider the paper "as is," frozen at a certain date, with no room for revisions. The authors can, of course, revise the paper on their own and even pursue an updated Unjournal review; we would like to include links to the "permanently updated version" in the Unjournal evaluation space.
After the pilot, we may consider making minor revisions part of the evaluation process. This may add substantial value to the papers and process, especially where evaluators identify straightforward and easily-implementable improvements.
If "minor revisions" are requested:
... the author has four (4) weeks (strict) to make revisions if they want to, submit a new linked manuscript, and also submit their response to the evaluation.
We don't want to replicate the slow and inefficient processes of the traditional system. Essentially, we want evaluators to give a report and rating as the paper stands.
We also want to encourage papers as projects. The authors can improve it, if they like, and resubmit it for a new evaluation.
Identify the paper's main claims and carefully assess their validity, leveraging your own background and expertise.
Answer a short questionnaire about your background and our processes.
Relevance to global priorities, usefulness for practitioners
0 - 100%
Do the authors communicate their work in ways policymakers and decision-makers can understand, without misleading or oversimplifying?
Do the authors present practical impact quantifications, such as cost-effectiveness analyses? Do they report results that enable such analyses?
3/5: Top B-journal/Strong field journal
4/5: Marginal A-Journal/Top field journal
5/5: A-journal/Top journal
For some questions, the "true value" refers to something objective, e.g. will this work be published in a top-ranked journal? In other cases, like the percentile rankings, the true value means "if you had complete evidence, knowledge, and wisdom, what value would you choose?"
For more information on credible intervals, this Wikipedia entry may be helpful.
If you are "well calibrated", your 90% credible intervals should contain the true value 90% of the time.
How many proposals and papers have you evaluated? (For journals, grants, and other peer review.)
Answers to the questions below will not be made public:
How would you rate this template and process?
Do you have any suggestions or questions about this process or The Unjournal? (We will try to respond to your suggestions, and incorporate them in our practice.) [Open response]
Would you be willing to consider evaluating a revised version of this project?
Overall assessment
0 - 100%
Claims, strength and characterization of evidence:
0 - 100%
Methods: Justification, reasonableness, validity, robustness
0 - 100%
Advancing knowledge and practice
0 - 100%
Logic and communication
0 - 100%
Open, collaborative, replicable science
0 - 100%
What journal ranking tier should this work be published in?
0.0-5.0
lower, upper
What journal ranking tier will this work be published in?
0.0-5.0
lower, upper

Managers and field specialists select work (or the project is submitted independently of authors) and the management team agrees to prioritize it.
For either of these cases (1 or 2), authors are asked for permission.
Alternate Direct Evaluation track: "Work enters prestige archive" (NBER, CEPR, and some other cases).
Managers inform and consult the authors but permission is not needed. (Particularly relevant: we confirm with author that we have the latest updated version of the research.)
Prioritization
Following author submission ...
Manager(s) (M) and Field Specialists (FS) prioritize work for review (see Project selection and evaluation).
Following direct evaluation selection...
M or FS may add additional (fn1) "evaluation suggestions" (see ) explaining why it's relevant, what to evaluate, etc., to be shared later with evaluators.
If requested (in either case), M decides whether to grant embargo or other special treatment, notes this, and informs authors.
M assigns an Evaluation Manager (EM – typically part of our management team or advisory board) to selected project.
EM invites evaluators (aka "reviewers") and shares the paper to be evaluated along with (optionally) a brief summary of why The Unjournal thinks it's relevant, and what we are asking.
Potential evaluators are given full access to (almost) all information submitted by the author and M, and notified of any embargo or special treatment granted.
EM may make special requests to the evaluator as part of a management policy (e.g., "signed/unsigned evaluation only," short deadlines, extra incentives as part of an agreed policy, etc.).
EM (also, optionally) may add "evaluation suggestions" to share with the evaluators.
Evaluator accepts or declines the invitation to review, and if the former, agrees on a deadline (or asks for an extension).
If the evaluator accepts, the EM shares full guidelines/evaluation template and specific suggestions with the evaluator.
Evaluator completes an evaluation form.
Evaluator submits evaluation including numeric ratings and predictions, plus "CI's" for these.
Possible addition (future plan): Reviewer asks for minor revisions and corrections; see "How revisions might be folded in..." in the fold below.
EM collates all evaluations/reviews, shares these with Author(s).
Evaluator must be very careful not to share evaluators' identities at this point.
This includes caution to avoid accidentally-identifying information, especially where evaluators chose anonymity.
Even if evaluators chose to "sign their evaluation," their identity should not be disclosed to authors at this point. However, evaluators are told they can reach out to the authors if they desire.
Evaluations are shared with the authors as a separate doc, set of docs, file, or space; which the evaluators do not have automatic access to. (Going forward, this will be automated.)
It is made clear to authors that their responses will be published (and given a DOI, when possible).
Author(s) read(s) evaluations, given two working weeks to submit responses.
If there is an embargo, there is more time to do this, of course.
EM creates evaluation summary and "EM comments."
EM or UJ team publishes each element on our PubPub space as a separate "pub" with a DOI for each (unless embargoed):
Summary and EM comments
With a prominent section for the "ratings data tables"
Each evaluation, with summarized ratings at the top
The author response
All of the above are linked in a particular way, with particular settings;
Authors and evaluators are informed once elements are on PubPub; next steps include promotion, checking bibliometrics, etc.
("Ratings and predictions data" to enter an additional public database.)
The Unjournal Evaluators have the option of remaining anonymous (see Evaluation ('refereeing')). Where evaluators choose this, we will carefully protect this anonymity, aiming at a high standard of protection, as good as or better than traditional journals. We will give evaluators the option to take extra steps to safeguard this further. We are offering anonymity in perpetuity to those who request it. (As well as anonymity on other terms to those who request it, on explicitly mutually agreed upon terms.)
If they choose to stay anonymous, there should be no way for authors to be able to ‘guess’ who has reviewed their work.
We will take steps to keep private any information that could connect the identity of an anonymous evaluator and their evaluation/the work they are evaluating.
We will take extra steps to make the possibility of accidental disclosure extremely small (this is never impossible of course, even in the case of conventional journal reviews). In particular, we will use pseudonyms or ID codes for these evaluators in any discussion or database that is shared among our management team that connects individual evaluators to research work.
If we ever share a list of Unjournal’s evaluators this will not include anyone who wished to remain anonymous (unless they explicitly ask us to be on such a list).
We will do our best to warn anonymous evaluators of ways that they might inadvertently be identifying themselves in the evaluation content they provide.
We will provide platforms to enable anonymous and secure discussion between anonymous evaluators and others (authors, editors, etc.) Where an anonymous evaluator is involved, we will encourage these platforms to be used as much as possible. In particular, see .
Aside: In future, we may consider , and these tools will also be useful.
It's a norm in academia that people do reviewing work for free. So why is The Unjournal paying evaluators?
From a recent survey of economists:
We estimate that the average (median) respondent spends 12 (9) working days per year on refereeing. The top 10% of the distribution dedicates 25 working days or more, which is quite substantial considering refereeing is usually unpaid.
The peer-review process in economics is widely-argued to be too slow and lengthy. But there is evidence that payments may help improve this.
, they note that few economics journals currently pay reviewers (and these payments tend to be small (e.g., JPE and AER paid $100 at the time). However, they also note, citing several papers:
The existing evidence summarized in Table 5 suggests that offering financial incentives could be an effective way of reducing turnaround time.
notes that the work of reviewing is not distributed equally. To the extent that accepting to do a report is based on individual goodwill, the unpaid volunteer model could be seen to unfairly penalize more generous and sympathetic academics. Writing a certain number of referee reports per year is generally considered part of "academic service". Academics put this on their CVs, and it may lead to being on the board of a journal which is valued to an extent. However, this is much less attractive for researchers who are not tenured university professors. Paying for this work would do a better job of including them in the process.
'Payment for good evaluation work' may also lead to fair and more useful evaluations.
In the current system academics may take on this work in large part to try to impress journal editors and get favorable treatment from them when they submit their own work. They may also write reviews in particular ways to impress these editors.
For less high-prestige journals, to get reviewers, editors often need to lean on their personal networks, including those they have power relationships with.
Reviewers are also known to strategically try to get authors to cite and praise the reviewer's own work. They maybe especially critical to authors they see as rivals.
To the extent that reviewers are doing this as a service they are being paid for, these other motivations will be comparatively somewhat less important. The incentives will be more in line with doing evaluations that are seen as valuable by the managers of the process, in order to get chosen for further paid work. (And, if evaluations are public, the managers can consider the public feedback on these reports as well.)
We are not ‘just another journal.’ We need to give incentives for people to put effort into a new system and help us break out of the old inferior equilibrium.
In some senses, we are asking for more than a typical journal. In particular, our evaluations will be made public and thus need to be better communicated.
We cannot rely on 'reviewers taking on work to get better treatment from editors in the future.' This does not apply to our model, as we don't have editors make any sort of ‘final accept/reject decision’
Our ‘paying evaluators’ brings in a wider set of evaluators, including non-academics. This is particularly relevant to our impact-focused goals.
A simplified rendering, skipping some steps and possibilities.
Interactive presentation: Walk through the evaluation process step-by-step in our slide presentation (use arrow keys to navigate). Includes timelines and comparison with traditional journals.
Cite evidence and reference specific parts of the research when giving feedback.
Justify your critiques and claims in a reasoning-transparent way, rather than merely ‘"passing judgment." Avoid comments like "this does not pass the smell test".
Provide specific, actionable feedback to the author where possible.
Try to restate the authors’ arguments, clearly presenting the most reasonable interpretation of what they have written. See .
Be collegial and encouraging, but also rigorous. Criticize and question specific parts of the research without suggesting criticism of the researchers themselves.
We're happy for you to use whichever process and structure you feel comfortable with when writing your evaluation content.
Core
Briefly summarize the work in context
Highlight positive aspects of the paper and its strengths and contributions, considered in the context of existing research.
Remember: The Unjournal doesn’t “publish” and doesn’t “accept or reject.” So don’t give an Accept, Revise-and-Resubmit', or Reject-type recommendation. We ask for quantitative metrics, written feedback, and expert discussion of the validity of the paper's main claims, methods, and assumptions.
Economics
Semi-relevant:
Report:
Open Science
(Conventional but open access; simple and brief)
(Open-science-aligned; perhaps less detail-oriented than we are aiming for)
(Journal-independent “PREreview”; detailed; targets ECRs)
General, other fields
(Conventional; general)
(extensive resources; only some of this is applicable to economics and social science)
‘the 4 validities’ and \
Most importantly: Identify and assess the paper's most important and impactful claim(s). Are these supported by the evidence provided? Are the assumptions reasonable? Are the authors using appropriate methods?
Note major limitations and potential ways the work could be improved; where possible, reference methodological literature and discussion and work that models what you are suggesting.
Optional/desirable
Offer suggestions for increasing the impact of the work, for incorporating the work into global priorities research and impact evaluations, and for supporting and enhancing future work.
Discuss minor flaws and their potential revisions.
Desirable: formal 'claim identification and assessment'
Please don't spend time copyediting the work. If you like, you can give a few specific suggestions and then suggest that the author look to make other changes along these lines.
We describe the nature of the work we are looking to evaluate, along with examples, in this forum post. Update 2024: This is now better characterized under "What research to target?" and "What specific areas do we cover?".
If you are interested in submitting your work for public evaluation, we are looking for research which is relevant to global priorities—especially quantitative social sciences—and impact evaluations. Work that would benefit from further feedback and evaluation is also of interest.
Your work will be evaluated using our evaluation guidelines and metrics. You can read these here before submitting.
Important Note: We are not a journal. By having your work evaluated, you will not be giving up the opportunity to have your work published in a journal. We simply operate a system that allows you to have your work independently evaluated.
If you think your work fits our criteria and would like it to be publicly evaluated, please submit your work through this form.
If you would like to submit more than one of your papers, you will need to complete a new form for each paper you submit.
By default, we would like Unjournal evaluations to be made public. We think public evaluations are generally good for authors, as explained here. However, in special circumstances and particularly for very early-career researchers, we may make exceptions.
If there is an early-career researcher on the author team, we will allow authors to "embargo" the publication of the evaluation until a later date. This date is contingent, but not indefinite. The embargo lasts until after a PhD/postdoc’s upcoming job search or until it has been published in a mainstream journal, unless:
the author(s) give(s) earlier permission for release; or
until a fixed upper limit of 14 months is reached.
If you would like to request an exception to a public evaluation, you will have the opportunity to explain your reasoning in the submission form.
See "Conditional embargos & exceptions" for more detail, and examples.
The Unjournal presents an additional opportunity for evaluation of your work with an emphasis on impact.
Substantive feedback will help you improve your work—especially useful for young scholars.
Ratings can be seen as markers of credibility for your work that could help your career advancement at least at the margin, and hopefully help a great deal in the future. You also gain the opportunity to publicly respond to critiques and correct misunderstandings.
You will gain visibility and a connection to the EA/Global Priorities communities and the Open Science movement.
You can take advantage of this opportunity to gain a reputation as an ‘early adopter and innovator’ in open science.
You can win prizes: You may win a “best project prize,” which could be financial as well as reputational.
Entering into our process will make you more likely to be hired as a paid reviewer or editorial manager.
We will encourage media coverage.
If we consider your work for public evaluation, we may ask for some of the items below, although most are optional. We will aim to make this a very light touch for authors.
A link to a non-paywalled, hosted version of your work (in any format—PDFs are not necessary) that can be given a Digital Object Identifier (DOI). Again, we will not be "publishing" this work, just evaluating it.
A link to data and code, if possible. We will work to help you to make it accessible.
Assignment of two evaluators who will be paid to assess your work. We will likely keep their identities confidential, although this is flexible depending on the reviewer. Where it seems particularly helpful, we will facilitate a confidential channel to enable a dialogue with the authors. One person on our managing team will handle this process.
Have evaluators publicly post their evaluations (i.e., 'reviews') of your work on our platform. As noted above, we will ask them to provide feedback, thoughts, suggestions, and some quantitative ratings for the paper.
By completing the submission form, you are providing your permission for us to post the evaluations publicly unless you request an embargo.
You will have a two-week window to respond through our platform before anything is posted publicly. Your responses can also be posted publicly.
For more information on why authors may want to engage and what we may ask authors to do, please see For researchers/authors.
Evaluations and author response are given DOI's, enter the bibliometric record
Future consideration:
"publication tier" of authors' responses as a workaround to encode aggregated evaluation
Hypothes.is annotation of hosted and linked papers and projects (aiming to integrate: see: )
Sharing evaluation data on public Github repo (see data reporting here)
We aim to elicit the experiment judgment from Unjournal evaluators efficiently and precisely. We aim to communicate this quantitative information concisely and usefully, in ways that will inform policymakers, philanthropists, and future researchers.
In the short run (in our pilot phase), we will attempt to present simple but reasonable aggregations, such as simple averages of midpoints and confidence-interval bounds. However, going forward, we are consulting and incorporating the burgeoning academic literature on "aggregating expert opinion." (See, e.g., Hemming et al, 2017; Hanea et al, 2021; McAndrew et al, 2020; Marcoci et al, 2022.)
We are working on this in our public data presentation (Quarto notebook) here.
We are considering...
Syntheses of evaluations and author feedback
Input to prediction markets, replication projects, etc.
Less technical summaries and policy-relevant summaries, e.g., for the EA Forum, Asterisk magazine, or mainstream long-form outlets
We invite academic journals to recognize The Unjournal's in their editorial process. We envision a lightweight partnership that can reduce journals' reviewer and editorial burden, improve decision-making, increase transparency, and support open science, while preserving full journal independence.
When authors submit research that the Unjournal has evaluated, journal editors could
Encourage reviewers to read these evaluations, setting explicit policies for how previous assessments should be considered.
Further possibilities: While this page focuses on a low-commitment consideration-based partnership, we're also happy to discuss more formal and deeper integrations. These could include an automatic send-to-review threshold, ratings-based fast-track, Unjournal as first-round-review, or a conditional acceptance framework.
Structured, public evaluation packages including:
Referee-style written assessments from expert evaluators
Authors' responses
Quantitative ratings on research quality dimensions (methods, communication, etc.), benchmarked as percentiles relative to an external reference group, as well as easy to compare across our evaluations (see our ratings dashboard)
Identification and assessment of the main claims, recommended robustness checks
"Journal tier" recommendations and predictions (e.g., "top-5," "top field journal," ...)
"Evaluation manager's" syntheses of this work and its implications
All evaluations are commissioned independently, conducted by domain and field experts, and published with DOIs, openly accessible and permanently linked to the research.
The steps below are similar to existing cascading and portable peer review agreements among journals, and "author-mediated transfer" policies (e.g., see The Economic Journal's policies).
When an Unjournal evaluation package exists for a submitted manuscript:
Consider the package, perhaps focusing on the "merited journal tier" assessment and the "Overall assessment" percentile score in desk/triage decisions.
Use this as one input among your standard criteria for this decision
Consider our reports in deciding whom to send it to and what guidance to give
Retain full editorial discretion — the evaluation does not bind your decision
Encourage reviewers to consult the Unjournal package when appropriate
Editors or journal policies set the terms and guidance for how reviewers should consider previous evaluations in their writing and recommendations
Reviewers should form independent judgments and explicitly note any aspects of their report that was influenced by previous evaluations
We would appreciate if you could recognize this partnership publicly, both for transparency and to boost impact. We give some examples below.
Add to your author guidelines:
Use of Public Evaluation Packages Our editors may consider public evaluation packages from The Unjournal when deciding whether to send a manuscript for external review. Where available, we take into account structured metrics and "merited journal tier" recommendations alongside our standard criteria. Final decisions remain at the editors' discretion.
Guidance for Reviewers Reviewers may consult The Unjournal's evaluation package for context but should form independent judgments. If referenced, cite and link to the package in your report.
For accepted papers that benefited from Unjournal evaluations, include a statement like the one below on the hosted page and in a footnote in the paper.
This review process was informed by The Unjournal's evaluation package: [link our package]. However, the reviewers take full responsibility for their judgments and recommendations and the final decisions were made by the editorial board independently of previous assessments.
Permit The Unjournal to list your journal as a partner and acknowledge the relationship on our website and communications (e.g., similar to PCI-friendly journal listings). We encourage you to do the same.
Many journals publicly state they consider Peer Community In (PCI) evaluations and are listed as "PCI-friendly." Our proposal follows this established model of transparent recognition while maintaining editorial independence. peercommunityin.org
The Economic Journal explicitly encourages authors to include editor letters and referee reports from previous submissions (EJ Instructions §5.1 Supporting Information—Past Reports), allowing editors to use them in decisions. OUP Academic
Many funding agencies allow preliminary assessments to guide full applications.
Reduced review burden: Use existing expert assessments to inform desk decisions
Faster turnaround: Expedite triage with pre-existing evaluations
Enhanced transparency: Public Unjournal evaluations provide context and boost understanding of methodological considerations and the peer review process
Quality signals: Authors with strong Unjournal evaluations will self-select to submit their work to journals that value them
Community leadership: Join journals advancing scholarly communication. Demonstrate openness to open science practices.
The Unjournal promotes public evaluation of hosted research as an alternative to "accept/reject" journal publications as a more efficient process and a more informative and measure research quality and usefulness. But we recognize that moving from existing systems is difficult, people will work within current incentives, and journal-based peer review also adds value. We're working to leverage existing systems to facilitate a gradual transition.
We want to raise awareness of our work, boost our credibility and demonstrate impact. Partnering with journals can do this and will encourage authors to submit their work with us and engage with our evaluations further.
We're happy to work with traditional journals in other mutually-beneficial ways, such as:
(As noted above) Deeper integrations are possible where journals could commit to automatic decisions or interim processes based on Unjournal ratings.
Sharing resources and insights about the review/evaluation process, guidelines for reviewers, etc. to help promote efficient, credible, transparent peer review and support open science.
Editors may want to recommend that authors submit particularly Unjournal-relevant work to us, either as an endpoint or as an initial step for further consideration.\
Q: Does this compromise our editorial independence? A: No. Editors retain complete discretion. The Unjournal evaluation is one input among many, just as prior referee reports or working paper citations might be.
Q: What if we disagree with the Unjournal evaluation? A: Disagreement is expected and appropriate. Your editors' judgment supersedes any external assessment. You can desk-reject papers with positive Unjournal evaluations or send for papers with lukewarm evaluations out for review.
Q: Will this create author expectations or appeals? A: Authors understand that Unjournal evaluations are informative, not determinative. Your submission guidelines should make clear that editors consider multiple factors and retain full discretion.
Q: What about evaluator bias or conflicts of interest? A: The Unjournal has rigorous COI policies and transparency standards. All evaluator identities and potential conflicts are disclosed. Editors can weigh this information as they would with any referee report.
You may wish to consider
a full partnership
a partial implementation trial (e.g., start with editorial triage only for 6-12 months; assess outcomes before expanding)
a field-specific trial, limited to specific subfields where Unjournal coverage is strongest
Review this proposal with your editorial board
Discuss any questions or concerns with us (contact@unjournal.org, book a chat)
Pilot with a small number of submissions to test the workflow
Formalize this by adding the policy language to your guidelines
Joint promotion and resource-sharing
Suggest a pivotal question using this form (or express your interest as an organization).
Keep track of of our progress: see this 'forum sequence' and this public database of PQs.
commissions public evaluations of impactful research in quantitative social sciences fields. We are seeking pivotal questions to guide our choice of research papers to commission for evaluation. We're contacting organizations that aim to use evidence to do the most good, and asking:
Which open questions most affect your policies and funding recommendations?
For which questions would research yield the highest ‘value of information’?\
The Unjournal has focused on finding research papers that seems relevant to impactful questions and crucial considerations, and then commissioning experts to publicly evaluate them. (For more about our process, see ). Our field specialist teams search and monitor prominent research archives (like ), and consider , while keeping an eye on forums and social media.
We're now exploring turning this on its head and identifying pivotal questions first and identifying evaluating a cluster of research that informs these. This could offer a more efficient and observable path to impact. (For context, see our .)\
The Unjournal will ask impact-focused research-driven organizations such as Open Philanthropy and Charity Entrepreneurship to identify specific quantifiable questions that impact their funding, policy, and research-direction choices. For example, If GiveWell is considering recommending a charity running a CBT intervention in West Africa, they’d like to know “how much does a 16 week course of non-specialist psychotherapy increase self-reported happiness, compared to the same amount spent on direct cash transfers?” We’re looking for the questions with the highest value-of-information (VOI) for the organization’s work over the next few years.
We have some requirements — the questions should relate to The Unjournal’s coverage areas and engage rigorous research in economics, social science, policy, or impact quantification. Ideally, organizations will identify at least one piece of publicly-available research that relates to their question. But we are doing this mainly to help these organizations, so we will try to keep it simple and low-effort for them.
If The Center for Humane Technology is considering a political campaign for AI safety in California, they could consider “how much does television and social media advertisements increase the vote share for ballot initiatives supporting the regulation of technology and business for safety reasons?”
OP might be considering funding organizations that promote democracy, largely because they think democracies may be more resilient to global catastrophies. As a tractable proxy, they may want to know “by what percentage does a country being a democracy reduce the loss of life in a natural disaster on the scale of a 7+ magnitude earthquake”?
We will work to minimize the effort required from these organizations; e.g., by leveraging their existing writings and agendas to suggest potential high value-of-information questions. We will also crowdsource questions (via EA Forum, social media, etc.), offering bounties for valuable suggestions.\
The Unjournal team will discuss the suggested questions, leveraging our field specialists’ expertise. We’ll rank these questions, prioritizing at least one for each organization.
We’ll work with the organization to specify the priority question precisely and in a useful way. We want to be sure that (1) evaluators will interpret these questions as intended, and (2) the answers that come out are likely to actually be helpful. We’ll make these lists of questions public and solicit general feedback — on their relevance, on their framing, on key sub-questions, and on pointers to relevant research.
Where practicable, we will operationalize the target questions as a claim on a prediction market (for example, Metaculus) to be resolved by the evaluations and synthesis below.
Where feasible, post these on public prediction markets (such as Metaculus)
If the question is well operationalized, and we have a clear approach to 'resolving it' after the evaluations and synthesis, we will post it on a reputation-based market like or . Metaculus is offering 'minitaculus' platforms such as to enable these more flexible questions.
We will ask (and help) the organizations and interested parties to specify their own beliefs about these questions, aka their 'priors'. We may adapt the Metaculus interface for this.
Once we’ve converged on the target question, we’ll do a variation of our usual evaluation process.
For each question, we will prioritize roughly two to five relevant research papers. These may be suggested by the organization that proposed the question, sourced by The Unjournal, or discovered through community feedback (see note).
As we normally do, we’ll have evaluation managers recruit expert evaluators to assess each paper. However, we’ll ask the evaluators to focus on the target question, and to consider the target organization’s priorities.
We’ll also enable phased deliberation and discussion among evaluators. This is inspired by the, and some evidence suggesting that the (mechanistically aggregated) estimates of experts after deliberations perform better than their independent estimates (also mechanistically aggregated). We may also facilitate collaborative evaluations and ‘live reviews’, following the examples of , , and others.
We will contact both the research authors (as per our standard process) and the target organizations for their responses to the evaluations, and for follow-up questions. We’ll foster a productive discussion between them (while preserving anonymity as requested, and being careful not to overtax people’s time and generosity)
We’ll commission one or more evaluation managers to write a report as a summary of the research investigated.
These reports should synthesize “What do the research, evaluations, and responses say about the question/claim?” They should provide an overall metric relating to the truth value of the target question (or similar for the parameter of interest). In cases where we integrate prediction markets, they should decisively resolve the market claim.
Next, we will share these synthesis reports with authors and organizations for feedback.
We’ll put up each evaluation on our page, bringing them into academic search tools, databases, bibliometrics, etc. We’ll also curate them, linking them to the relevant target question and to the synthesis report.
We will produce, share, and promote further summaries of these packages. This could include forum and blog posts summarizing the results and insights, as well as interactive and visually appealing web pages. We may also produce less technical content, perhaps submitting work to outlets like, , or .
At least initially, we’re planning to ask for questions that could be definitively answered and/or measured quantitatively. We will help organizations and other suggesters refine their questions to make this the case. These should resemble questions that could be posted on forecasting platforms such as or . These should also resemble the we currently request from evaluators.
We give detailed guidance with examples below:
Why do we want these pivotal questions to be 'operationalizable'?
We’re still refining this idea, and looking for your suggestions about what is unclear, what could go wrong, what might make this work better, what has been tried before, and where the biggest wins are likely to be. We’d appreciate your feedback! (Feel free to email to make suggestions or arrange a discussion.)
If you work for an impact-focused research organization and you are interested in participating in our pilot, please reach out to us at contact@unjournal.org to flag your interest and/or complete . We would like to see:
A brief description of what your organization does (linking your ‘about us’ page is fine)
A specific, , high-value claim or research question you'd like to be evaluated, that falls within our scope (~quantitative social science, economics, policy, and impact measurement)
A brief explanation of why this question is particularly high-value for your organization or your work, and, if applicable, how you have tried to answer it
Please also let us know how you would like to engage with us on refining this question and addressing it. Do you want to follow up with a 1-1 meeting? How much time are you willing to put in? Who, if anyone, should we reach out to at your organization?
Remember that we plan to make all of this analysis and evaluation public. However, we will not make any of your input public without your consent.
If you don’t represent an organization, we still welcome your suggestions, and will try to give feedback. (Note on 'bounties'.)
Again, please remember that we currently focus on quantitative ~social sciences fields, including economics, policy, and impact modeling (see for more detail on our coverage). Questions surrounding (for example) technical AI safety, microbiology, or measuring animal sentience are less likely to be in our domain.
If you want to talk about this first, or if you have any questions, please send an email or with David Reinstein, our co-founder and director.
On this page we link to and discuss some takes on answers to the questions, Which research is most impactful? Which research should be prioritized?
At The Unjournal, we are open to various approaches to the issues of "what is the most impactful research"? Perhaps looking at the research we have already evaluated and research we are prioritizing will give you some insights. However, it seems fair that we should give at least one candidate description or definition.
"The direct global impact of a work of research is determined by the value of the information that it provides in helping individuals, governments, funders, and policymakers make better decisions. While research may not definitively answer key questions it should leave us more informed (in a Bayesian sense, 'more concentrated belief distributions') about these. We will measure the value of these 'better choices' in terms of the extent these improve the welfare of humans and other sentient beings, and the survival and flourishing of life and human civilization and values."
The above comes close to how some people on The Unjournal team think about research impact and prioritization, but we don't plan to adopt an official guiding definition.
Note the above definition is meant to exclude more basic research, which may also be high value, but which mainly serves as a building block for other research. In fact, The Unjournal does consider the value of research as an input into other research, particularly when it directly influences direct policy-relevant research, e.g., see .
It also excludes the value of "learning the truth" as an intrinsic good; we have tended not to make this a priority.
For more guidance on how we apply this, see our .
Syllabi and course outlines that address global prioritization
Those listed below are at least somewhat tied to Effective Altruism.
" page in "Economics for EA and vice versa" Gitbook
Stafforini's list of EA syllabi
We welcome suggestions for additional syllabi to include here. Contact us at .
We next consider organizations that take a broad focus on helping humans, animals, and the future of civilization. Some of these have explicitly set priorities and research agendas, although the level of specificity varies. Most of the organizations below have some connections to Effective Altruism; over time, we aim to also look beyond this EA focus.
GPI focuses on prioritization research—what to prioritize and why; how to make these decisions. They focus less on how to implement improvements and interventions.
The agenda is divided into "The longtermism paradigm" and "General issues in global prioritisation."
The agenda focuses largely on formal theory (in philosophy, economics, and decision science) and, to a lesser extent., methodology. They aim to identify and inform "crucial considerations," and rarely focus on specific impact assessments.
Nonetheless, the agenda cites some empirical and directly policy-relevant work, and there are suggestions (e.g., from Eva Vivalt) that they might move more towards this in the future.
Below, I (Reinstein) list the categories from GPI's 2020 agenda. I give a first-pass impression of the relevance of these categories for The Unjournal, in something like descending order (bold = most clearly relevant).
🎉
More relevant to The Unjournal:
; posted on the EA Forum as ""
posted on the EA Forum as
is a fairly brief discussion and overview linking mostly to OP-funded research.
("Research Priorities," 2021): A particularly well organized discussion. Each section has a list of relevant academic literature, some of which is recent and some of which is applied or empirical.
: Their "" and "Research briefs" are particularly helpful, and connect to a range of academic and policy research
: simple discussions of the cause they prioritize, backed by numbers and links/citations
: Some directional suggestions in the "Our current plans" section under "Our research going forward is expected to focus on:"
: Not easy to link to research; they have a large number of priorities, goals, and principles; see infographic:
Their "" page considering relative cost-effectiveness; generally a shallow review/BOTEC spreadsheet approach. "CEARCH attempts to identify a cause’s marginal expected value (MEV)."
: This page is particularly detailed and contains a range of useful links to other agendas!
"
(Gainsburg et al, 2021)
"
: The survey, as reported in the paper, does not suggest a particular agenda, but it does suggest a direction . . . economists would generally like to see more work in certain applied areas.
): This seems extremely relevant. . . . NSF
If a CE project is considering promoting farmed fish welfare legislation in India, they might ask “as the price of India-farmed fish increases by 10%, how much will consu
If possible, a link to at least one research paper that relates to this question
Optionally, your current beliefs about this question (your ‘priors’)
Less relevant to The Unjournal: "Intergenerational governance", "The value of the future of humanity", "Articulation and evaluation of longtermism", "Other ways of leveraging the size of the future", "Moral uncertainty for longtermists"
More relevant to The Unjournal:
Less relevant: Decision-theoretic issues, Epistemological issues

Other unsuccessful inquiries and applications include (details may be shared)
Templeton (inquiry)
AI for Human Reasoning
Emergent Ventures
Peter Slattery: on EA Forum, fork moved to .
Other comments, especially post-grant, in discussion space (embedded below) will be integrated back.
At least initially, we’re planning to ask for questions that could be definitively answered and/or measured quantitatively, and we will help organizations and other suggesters refine their questions to make this the case. These should approximately resemble questions that could be posted on forecasting platforms such as or . These should also somewhat resemble the we currently request from evaluators.
is particularly relevant. As
ACX Grants
Navigation Fund (inquiry, no response given)
if you handed your question to a genuine clairvoyant, could they see into the future and definitively tell you [the answer]? Some questions like ‘Will the US decline as a world power?’...‘Will an AI exhibit a goal not supplied by its human creators?’ struggle to pass the Clairvoyance Test… How do you tell one type of AI goal from another, and how do you even define it?... In the case of whether the US might decline as a world power, you’d want to get at the theme with multiple well-formed questions such as ‘Will the US lose its #1 position in the IMF’s annual GDP rankings before 2050?’.... These should also somewhat resemble the 'claim identification' we currently request from evaluators.
Metaculus and Manifold: claim resolution.\
Some questions are important, but difficult to make specific, focused, and operationalizable. For example (from 80,000 Hours’ list of “research questions”):
“What can economic models … tell us about recursive self improvement in advanced AI systems?”
“How likely would catastrophic long-term outcomes be if everyone in the future acts for their own self-interest alone?”
“How could AI transform domestic and mass politics?” \
Other questions are easier to operationalize or break down into several specific sub-questions. For example (again from 80,000 Hours’ “research questions”):
Could advances in AI lead to risks of very bad outcomes, like suffering on a massive scale? Is it the most likely source of such risks?
I rated this a 3/10 in terms of how operationalized it was. The word “could” is vague. “Could” might suggest some reasonable probability outcome (1%, 0.1%, 10%), or it might be interpreted as “can I think of any scenario in which this holds?” “Very bad outcomes” also needs a specific measure.
However, we can reframe this to be more operationalized. E.g., here are some fairly well-operationalized questions:
What is the risk of a catastrophic loss (defined as the death of at least 10% of the human population over any five year period) occurring before the year 2100?
How does this vary depending on the total amount of money invested in computing power for building advanced AI capabilities over the same period?
Here are some highly operationalizable questions developed by the Farm Animal Welfare team at Open Phil:
What percentage of plant-based meat alternative (PBMA) units/meals sold displace a unit/meal of meat?
What percentage of people will be [vegetarian or vegan] in 20, 50, or 100 years?
And a few more posed and addressed by Our World in Data:
How much of global greenhouse gas emissions come from food? (full article)
What share of global CO₂ emissions come from aviation? (full article)
However, note that many of the above questions are descriptive or predictive. We are also very interested in causal questions such as
What is the impact of an increase (decrease) in blood lead level by one “natural log unit” on children’s learning in the developing world (measured in standard deviation units)?
Why are we seeking these pivotal questions to be 'operationalizable'?\
This is in line with our own focus on this type of research
The Unjournal mainly focuses on evaluating (largely empirical) research that clearly poses and answers specific impactful questions, rather than research that seeks to define a question, survey a broad landscape of other research, open routes to further inquiry, etc.
I think this will help us focus on fully-baked questions, where the answer is likely to provide actual value to the target organization and others (and avoid the old ‘ trap).
It offers potential for benchmarking and validation (e.g., using prediction markets), specific routes to measure our impact (updated beliefs, updated decisions), and informing the we’re asking from evaluators (see footnote above).
However, as this initiative progresses we may allow a wider range of questions, e.g., more open-ended, multi-outcome, non-empirical (perhaps ‘normative), and best-practice questions.
Our progress and ongoing work
See Plan of action for our current strategy.
See Updates (earlier) for recent developments.
Rethink Priorities will act as fiscal sponsor for this, to help administer payments. They will also receive $5,000 to cover roughly two hours/week of Reinstein's time on this project.
Administering payments to referees, researchers, etc.
We will need to make small payments to (say) 20–50 different referees, 5–10 committee members and "editorial managers," 5–10 research prize winners, as well as clerical and IT assistants.
LTFF:
Please let us know how you would like your grant communicated on the ACX blog, e.g., if you'd like Scott to recommend that readers help you in some way (see this post for examples).
See __
Passed on to LTFF and funding was awarded
frozen version as Dropbox paper here
Passed on to LTFF and funding was awarded
Start date = ~21 February 2022
The "Unjournal" will organize and fund 'public journal-independent evaluation’ of EA-relevant/adjacent research, encouraging this research by making it easier for academics and EA-organization researchers to get feedback and credible ratings.
Peer review is great, but academic publication processes are wasteful, slow, rent-extracting, and they discourage innovation. From
Academic publishers extract rents and discourage progress. But there is a coordination problem in ‘escaping’ this. Funders like Open Philanthropy and EA-affiliated researchers are not stuck, we can facilitate an exit.
The traditional binary ‘publish or reject’ system wastes resources (wasted effort and gamesmanship) and adds unnecessary risk. I propose an alternative, the “Evaluated Project Repo”: a system of credible evaluations, ratings, and published reviews (linked to an open research archive/curation). This will also enable more readable, reliable, and replicable research formats, such as dynamic documents; and allow research projects to continue to improve without “paper bloat”. (I also propose some ‘escape bridges’ from the current system.)
Global priorities and EA research organizations are looking for ‘feedback and quality control’, dissemination, and external credibility. We would gain substantial benefits from supporting, and working with the Evaluated Project Repo (or with related peer-evaluation systems), rather than (only) submitting our work to traditional journals. We should also put some direct value on results of open science and open access, and the strong impact we may have in supporting this.
I am asking for funding to help replace this system, with EA 'taking the lead'. My goal is permanent and openly-hosted research projects, and efficient journal-independent peer review, evaluation, and communication. (I have been discussing and presenting this idea publicly for roughly one year, and gained a great deal of feedback. I return to this in the next section).
I propose the following 12-month Proof of Concept: Proposal for EA-aligned research 'unjournal' collaboration mechanism
Build a ‘founding committee’ of 5-8 experienced and enthusiastic EA-aligned/adjacent researchers at EA orgs, research academics, and practitioners (e.g., draw from speakers at recent EA Global meetings).
Update 1 Aug 2022, mainly DONE, todo: consult EAG speakers
I will publicly share my procedure for choosing this group (in the long run we will aim at transparent and impartial process for choosing ‘editors and managers’, as well as aiming at decentralized forms of evaluation and filtering.)
2. Host a meeting (and shared collaboration space/document), to come to a consensus/set of principles on
A cluster of EA-relevant research areas we want to start with
A simple outreach strategy
How we determine which work is 'EA-interesting’
Update 1 Aug 2022: 2 meetings so far, agreed on on going-forward policies for most of the above
3. Post and present our consensus (on various fora especially in the EA, Open Science, and relevant academic communities, as well as pro-active interviews with key players). Solicit feedback. Have a brief ‘followup period’ (1 week) to consider adjusting the above consensus plan in light of the feedback.
Update 1 Aug 2022: Done somewhat; waiting to have 2+ papers assessed before we engage more
4. Set up the basic platforms, links
Note: I am strongly leaning towards as the main platform, which has indicated willingness to give us a flexible ‘experimental spac\
Update 1 Aug 2022: Going with Kotahi and Sciety as a start; partially setup
5. Reach out to researchers in relevant areas and organizations and ask them to 'submit' their work for 'feedback and potential positive evaluations and recognition', and for a chance at a prize.
The unjournal will *not be an exclusive outlet.* Researchers are free to also submit the same work to 'traditional journals' at any point.
Their work must be publicly hosted, with a DOI. Ideally the 'whole project' is maintained and updated, with all materials, in a single location. We can help enable them to host their work and enable DOI's through (e.g.) Zenodo; even hosted 'dynamic documents' can be DOI'd.
Update 1 Aug 2022: Did a 'bounty' and some searching of our own, plan a 'big public call' afrter pilot evaluations of 2+ papers
Researchers are encouraged to write and present work in 'reasoning transparent' (as well as 'open science' transparent) ways. They are encouraged to make connections with core EA ideas and frameworks, but without being too heavy-handed. Essentially, we are asking them to connect their research to 'the present and future welfare of humanity/sentient beings'.
Reviews will, by default, be made public and connected with the paper. However, our committee will discuss I. whether/when authors are allowed to withdraw/hide their work, and II. when reviews will be ‘signed’ vs anonymous. In my conversations with researchers, some have been reluctant to ‘put themselves out there for public criticism’, while others seem more OK with this. We aim to have roughly 25 research papers/projects reviewed/evaluated and 'communicated' (to EA audiences) in the first year.
Update July 2022: scaled back to 15 papers
My suggestions on the above, as a starting point...
Given my own background, I would lean towards ‘empirical social science’ (including Economics) and impact evaluation and measurement (especially for ‘effective charitable interventions’)
Administration should be light-touch, to also be attractive to aligned academics
We should build "editorial-board-like" teams with subject/area expertise
Laying these out; I have responses to some of these, others will require further consideration \
Will researchers find it useful to submit/share their work? From my experience (i) as an academic economist and (ii) working at Rethink Priorities, and my conversations with peers, I think people would find this very useful. I would have (and still would).
i. FEEDBACK IS GOLD: It is very difficult to get anyone to actually read your paper, and to get actual useful feedback on your work. The incentive is to publish, not to read, papers are dense and require specific knowledge, and people may be reluctant to criticize peers, and economists tend to be free-riders. It is hard to engage seminar audiences on the more detailed aspects of the work, and then one gets feedback on the ‘presentation’ not the ‘paper’. We often use ‘submission to journal’ as a way to get feedback, but this is slow, not the intended use of the journal process (I’ve been told), and often results in less-useful feedback. (A common perception is that the referee ‘decides what decision to make and then fleshes out a report to justify it.)
ii. ACADEMICS NEED SOURCES OF TIMELY VALIDATION: The publication process is extremely slow and complicated in Economics (and other fields, in my experience), requiring years of submissions and responses to multiple journals. This imposes a lot of risky for an academic’s career, particularly pre-tenure. Having an additional credible source validating the strength of one’s work could help reduce this risk. If we do this right, I think hiring and tenure committees would consider it as an important source of quality information.
iii. EA ORGS/FUNDERS need both, but the traditional journal process is costly in time and hassle. I think researchers and research managers at RP would be very happy to get feedback through this, as well as an assessment of the quality of their work, and suggestions for alternative methods and approaches. We would also benefit from external signals of the quality of our work, in justifying this to funders such as Open Philanthropy. (OP themselves would value this greatly, I believe. They are developing their own systems for asse_s_sing the quality of their funded work, but I expect they would prefer an external source.) However, it is costly for us at RP to submit to academic journals: the process is slow and bureaucratic and noisy, and traditional journals will typically not evaluate work with EA priorities and frameworks in mind. (Note that I suggest the unjournal make these priorities a factor while also assessing the work’s rigor in ways that invoke justifiable concerns in academic disciplines.)
I assume that similar concerns apply to other EA research organizations.
iv. OPEN SCIENCE AND DYNAMIC FORMATS
IMO the best and most transparent way to present data-driven work (as well as much quantitative work) is in a dynamic document, where narrative, code, and results are presented in concert. Readers can ‘unfold for further details’. The precise reasoning, data, and generation of each result can be traced. These can also be updated and improved with time. Many researchers, particularly those involved in Open Science, find this the most attractive way to work and present their work. However, ‘frozen pdf prison’ and ‘use our bureaucratic system’ approaches makes this very difficult to use in traditional journals. As the ‘unjournal’ does not host papers, but merely assesses work with DOI’s (which can be, e.g. a hosted web page, as frozen at a particular point in time of review), we can facilitate this. Will researchers find it 'safe' to share their work?
A large group of Economists and academics tend to be conservative, risk-averse, and leader-following. But there are important exceptions and also substantial groups that seek to be particular innovative and iconoclastic.
The key concerns we will need to address (at least for some researchers). i. Will my work be ‘trashed publicly in a way that hurts my reputation’? I think this is more for early-career; more experienced researchers will have a thicker skin and realize that it’s common-knowledge that some people disagree with their approaches. ii. Will this tag me as ‘weird or non-academic’. This might be addressed by our making connections to academic bodies and established researchers. How to get quality reviews and avoid slacking/free-riding by reviewers? Ideas:
compensation and rewarding quality as an incentive,
recruiting reviewers who seem to have intrinsic motivations,
publishing some ‘signed’ reviews (but there are tradeoffs here as we want to avoid flattery)
QUANTIFY and CALIBRATE
We will ask referees to give a set of quantitative ratings in addition to their detailed feedback and discussion. These should be stated in ways that are made explicitly relative to other work they have seen, both within the Unjournal, and in general. Referees might be encouraged to ‘calibrate’; first given a set of (previously traditionally-published) papers to rank and rate. They should be later reminded about how the distribution of the evaluation they have given.
Within our system, evaluations themselves could be stated ‘relative to the other evaluations given by the same referee.’
BENCHMARK We also will encourage or require referees provide a ‘a predicted/equivalent “traditional publication outcome” and possibly incentivize these predictions. (And we could consider running public prediction markets on this in the longer run, as has been done in other contexts). This should be systematized. It could be stated as “this project is of the sufficient quality that it has a 25% probability of being published in a journal of the rated quality of Nature, and a 50% probability of being published in a journal such as the Journal of Public Economics or better … within the next 3 years.” (We can also elicit statements about the impact factor, etc.)
I expect most/many academics who submit their work will also submit it to traditional journals at least in the first year or so of this project. (but ultimately we hope this 'unjournal' system of journal-independent evaluation provides a signal of quality that will supercede The Musty Old Journal.) This will thus provide us a way to validate the above predictions, as well as independently establish a connection between our ratings and the ‘traditional’ outcomes. PRIZE as a powerful signal/scarce commodity The “prize for best submissions” (perhaps a graded monetary prize for the top 5 submissions in the first year) will provide a commitment device and a credible signal, to enhance the attractiveness and prestige of this.
We may try to harness and encourage additional tools for quality assessment, considering cross-links to prediction markets/Metaculus, the coin-based 'ResearchHub', etc.
Will the evaluations be valued by gatekeepers (universities, grantmakers, etc.) and policy-makers? This will ultimately depend on the credibility factors mentioned above. I expect they will have value to EA and open-science-oriented grantmakers fairly soon, especially if the publicly-posted reviews are of a high apparent quality.
I expect academia to take longer to come on board. In the medium run they are likely to value it as ‘a factor in career decisions’ (but not as much as a traditional journal publication); particularly if our Unjournal finds participation and partnership with credible established organizations and prominent researchers.
I am optimistic because of my impression that non-traditional-journal outcomes (arXiv and impact factors, conference papers, cross-journal outlets, distill.pub) are becoming the source of value in several important disciplines How will we choose referees? How to avoid conflicts of interest (and the perception of this)?
This is an important issue. I believe there are ‘pretty good’ established protocols for this. I’d like to build specific prescribed rules for doing this, and make it transparent. We may be able to leverage tools, e.g., those involving GPT3 like elicit.org.
COI: We should partition the space of potential researchers and reviewers, and/or establish ‘distance measures’ (which may themselves be reported along with the review). There should be specified rules, e.g., ‘no one from the same organization or an organization that is partnering with the author’s organization’. Ideally EA-orgresearchers’ work should be reviewed by academic researchers, and to some extent vice-versa.
How to support EA ideas, frameworks, and priorities while maintaining (actual and perceived) objectivity and academic rigor
(Needs discussion)
Why hasn't this been done before? I believe it involves a collective action problem, as well as a coordination/lock-in problem that can be solved by bringing together the compatible interests of two groups. Academic researchers have expertise, credibility, but they are locked into traditional and inefficient systems. EA organizations/researchers have a direct interest in feedback and fostering this research, and have some funding and are not locked into traditional systems.
Yonatan Cale restating my claim:
Every Econ researcher (interested in publishing) pays a price for having the system set up badly, the price isn't high enough for any one researcher to have an incentive to fix the system for themselves, but as a group, they would be very happy if someone would fix this systematic problem (and they would in theory be willing to "pay" for it, because the price of "fixing the system" is way lower then the sum of the prices that each one of them pays individually)
‘Sustainability’ … Who will pay for these reviews in the longer run
Once this catches on…Universities will pay to support this; they will save massively on journal subscriptions. Governments supporting Open Science will fund this. Authors/research orgs will pay a reasonable submission fee to partly/fully cover the cost of the reviews. EA-aligned research funders will support this.
But we need to show a proof-of-concept and build credibility. The ACX grant funds can help make this happen.\
My CV should make this clear\
I have been an academic economist for 15-20 years, and I have been deeply involved in the research and publication process, with particular interests in open science and dynamic documents. (PhD UC Berkeley Lecturer University of Essex, Senior Lecturer, University of Exeter). My research has mainly been in Economics, but also involving other disciplines (especially Psychology).
I’m a Senior Economist at Rethink Priorities, where I’ve worked for the past year, engaging with a range of researchers and practitioners at RP and other EA groups
My research has involved EA-relevant themes since the latter part of my PhD. I’ve been actively involved with the EA community since about 2016, when I received a series of ESRC ‘impact grants’ for the innovationsinfundraising.org and giveifyouwin.org projects, working with George Howlett and the CEA
I have had long 1-1 conversations on this idea with a range of knowledgable and relevant EAs, academics, and open-science practitioners, and technical/software developers including
Cooper Smout, head of ‘, which I’d like to ally with (through their pledges, and through an open access journal Cooper is putting together, which the Unjournal could feed into, for researchers needing a ‘journal with an impact factor’)
Participants in the GPI seminar luncheon
Daniela Saderi of PreReview
, my online CV has links to almost everything else@givingtools on twitter david_reinstein on EA forum; see post on this: I read/discuss this on my podcast, e.g., see \
Feel free to give either a simple number, or a range, a complicated answer, or a list of what could be done with how much
Over a roughly one-year ‘pilot’ period, I propose the following. Note that most of the costs will not be incurred in the event of the ‘failure modes’ I consider. E.g., if we can’t find qualified and relevant reviewers and authors, these payments will not be made
$15k: Pay reviewers for their time for doing 50 reviews of 25 papers (2 each), at 250 USD per review (I presume this is 4-5 hours of concentrated work) --> 12,500 USD
$5k to find ways to ’buy off” 100 hours of my time (2-3 hours per week over some 50 weeks) to focus on managing the project, setting up rules/interface, choosing projects to review, assigning reviewers, etc. I will do this either through paying my employer directly or ‘buying time’ by getting delivery meals, Uber rides, etc.)
$5k to ’buy off” 100 hours of time from other ‘co-editors’ to help, and for a board to meet/review the initial operating principles
$5k: to hire about 100 hours technical support for 1 year to help authors host and format their work, to tailor the ‘experimental’ space that PreReview has promiosed us, and potentially working with the EA forum and other interfaces
$2.5k: Hire clerical/writing/copy editing support as needed
$7.5k: rewards for ‘authors of the best papers/projects’ (e.g., 5 * 1000 USD … perhaps with a range of prizes) … and/or additional incentives for ‘best reviews’ (e.g., 5 * 250 USD)
We have an action plan (mainly for EA organizations) and a workspace in the GitBook here: This also nests several essays discussing the idea, including the collaborative document (with many comments and suggestions) at \
Most of the measures of ‘small success’ are scaleable; the funds I am asking for, for referee payments, some of my time, etc., will not be spent/will be returned to you if we do not recieve quality submissions and commitments to review and assist in the management
My own forecast (I’ve done some calibration training, but these are somewhat off-the-cuff) 80% that we will find relevant authors and referees, and this will be a useful resource for improving and assesing the credibility of EA-relevant research
60% that we will get the academic world substantially involved in such a way that it becomes reasonably well known, and quality academic researchers are asking to ‘submit their work’ to this without our soliciting their work.
50% that this becomes among the top/major ways that EA-aligned research organizations seek feedback on their work (and the work that they fund — see OpenPhil), and a partial alternative to academic publication
10-25% that this becomes a substantial alternative (or is at the core of such a sea-change) to traditional publication in important academic fields and sub-fields within the next 1-3 years. (This estimate is low in part because I am fairly confident a system along these lines will replace the traditional journal, but less confident that it will be so soon, and still less confident my particular work on this will be at the center of it.) \
Yes
Yes
How we will evaluate, rate, rank, and give feedback on work
The platforms we will work with
How to promote and communicate the research work (to academics, policymakers, and the EA community)
Create a set of rules for 'submission and management', 'which projects enter the review system' (relevance, minimal quality, stakeholders, any red lines or 'musts'), how projects are to be submitted (see above, but let's be flexible), how reviewers are to be assigned and compensated (or 'given equivalent credit')
Rules for reviews/assessments
Reviews to be done on the chosen open platform (likely Prereview) unless otherwise infeasible
Share, advertise, promote the reviewed work
Establish links to all open-access bibliometric initiatives to the extent feasible
Each research paper/project should be introduced in at least one EA Forum post
I’ve been considering and discussing this proposal for many years with colleagues in Economics and other fields, and presenting it publicly and soliciting feedback over the past year— mainly through , social media, EA and open science Slack groups and conferences (presenting this at a GPI lunch and at the COS/Metascience conference, as well as in an EA Forum post and the post mentioned above).
Cecilia Tilli, Foundation to Prevent Antibiotics Resistance and EA research advocate
Sergey Frolov (Physicist), Prof. J.-S. Caux, Physicist and head of
Peter Slattery, Behaviourworks Australia
Alex Barnes, Business Systems Analyst,
Gavin Taylor and Paola Masuzzo of IGDORE (biologists and advocate of open science)
William Sleegers (Psychologist and Data Scientist, Rethink Priorities)
Nathan Young
Edo Arad (mathematician and EA research advocate)
Hamish Huggard (Data science, ‘literature maps’)
Yonatan Cale, who helped me put this proposal together through asking a range of challengin questions and offering his feedback.
Content from this grant application is linked and embedded below, verbatim.
Our brief application to a regranting program (first round) is and embedded below, verbatim.
This comparison is auto-generated from our internal database. Last updated: January 2026
These initiatives share significant overlap with The Unjournal's mission or are active partners.
Know of an initiative we should add? .
🔧 Platform/tool: reviewing/ratin
Sciety: 'curation' ... make it easier to access preprints and reviewers across platforms. We used their platform to some extent and may do so in the future. Some considerations, e.g., only able to in
📄 Open public evaluation initiat
Crowdsourced evaluations? Suggested by Stuart Buck. I love how Arcadia Science envisions the research project not just the standalone paper.https://research.arcadiascience.com/ https://research.arcadi
📄 Open public evaluation initiat
We're in contact/collaboration
📄 Open public evaluation initiat
We’re in contact/collaboration
🤝 Open science collaboration
🔧 Platform/tool: reviewing/ratin
We use PubPub to host our evaluations at unjournal.pubpub.org. It has many strong features and we continue to collaborate with Knowledge Futures on development.
📄 Open public evaluation initiat
"eLife's is/was a fairly well respected (?) journal in Life Sciences. Their New Model (originally called 'publish, review, curate') was big news. Their three month update seems fairly stable/successf
📄 Open public evaluation initiat
Mostly CS
📄 Open public evaluation initiat
"PLOS is a nonprofit, Open Access publisher empowering researchers to accelerate progress in science and medicine by leading a transformation in resea
🔧 Platform/tool: reviewing/ratin
Plaudit is a third-party tool. OSF seems to encourage it (https://help.osf.io/hc/en-us/articles/360038856834-Endorse-a-Preprint). You can endorse the
🔧 Platform/tool: reviewing/ratin
Hub to upload/link DOI'd research, by subject. Comments and 'costly' upvotes/downvotes (?), no explicit 'reviewing' process. For-profit? "... a modern
📄 Open public evaluation initiat
Making good inroads in physics+ Distinctions/considerations: physics+-focused, exclusive publishing, may not allow dynamic formats or continuous impro
📄 Open public evaluation initiat
Has most of Unjournal's features, but no quantitative measures, no external funding for evaluators, and seems very small. A platform for disseminatin
📄 Open public evaluation initiat
📄 Open public evaluation initiat
Optional transparent peer review (author opt-in, for accepted work)
🔧 Platform/tool: reviewing/ratin
Review Commons is a platform for high-quality journal-independent peer-review in the life sciences." Pre-prints reviewed, streams into (regular and OA
Science Open
🔧 Platform/tool: reviewing/ratin
Working with traditional and open journals and trying to bring in stakeholders. Seems to foster academic editor-based curation. But tied to traditiona
📄 Open public evaluation initiat
Nascent? Econometrics focused, not sure if it will get off the ground
🤝 Open science collaboration
Focus on medicine?
📄 Open public evaluation initiat
Open review/evaluation. Unjournal is an evaluation provider. Emphasis on multiple points of evaluation throughout the research life cycle
🔄 replication/reproducibility/ro
Goal: Robustness 'replication' of all top published papers in Econ, Poli Sci, etc. Strong template and lots of credible backers. A possible partner for unjournal.
🔧 Platform/tool: reviewing/ratin
PCI and the Peer Community Journal; See especially the PCI Psychology Community, and to a lesser extent the nascent Economics one
🔧 Platform/tool: reviewing/ratin
We provide ways for feedback to preprints to be done openly, rapidly, constructively, and by a global community of peers." Distinctions/limitations: More of a platform than an initiative. Mostly bot
🔄 replication/reproducibility/ro
"the first certification agency for scientific code & data" -- a possible source of worthy 'open' work
📄 Open public evaluation initiat
At least some detailed public evaluations, although they don’t make these as prominent as Unjournal. Founded 2007, owned by DeGruyter from 2020. Archi
🔧 Platform/tool: reviewing/ratin
Connected to Sciety, Elife etc. They were once hosting our editorial management platform and linking the output with a Sciety group; something we tho
🔧 Platform/tool: reviewing/ratin
Hosts (?) interfaces with multiple archives. "Pre or post-moderation". Facilitates and encourages (?) Plaudit and PREreview
Organization/discussion following a thread in...
Some conversation highlights:
Kris Gulati: Recently I've been talking to more global priorities-aligned researchers to get to know what people are working on. I noticed they're somewhat scattered around (Stanford, PSE, Chicago, Oxford etc.). Additionally, sometimes established academics don't always entirely grasp global priorities-focused work and so it can be tough to get feedback on ideas from supervisors or peers when it's pretty different to the more orthodox research many academics focus on. One way of remedying this is to have an informal seminar series where people interested in GP work present early stages ideas and can receive feedback on their ideas, etc.
David Mannheim: Yes, this seems very promising. And I think that it would be pretty easy to get a weekly seminar series together on this on Zoom.
Robin Hanson: Why limit it to PhD students? All researchers can gain from feedback, and can offer it.
Eva Vivalt: Sounds great. GPI also has seminars on global priorities research in philosophy and economics that might be of interest. . . . [followed by some notes of caution] I'm just worried about stretching the senior people too thin. I've been attending the econ ones remotely for a while, and only this semester did I finally feel like it was really mature; a senior person less would be a setback. I fully think there should be many groups even within econ, and at different institutions; that would be a healthy ecosystem.
Kris, responding to Dave Rhys-Bernard: If GPI's seminar series is meant to be private, then it's worth running something additional, given we can get a decent critical mass of attendance and some senior people are happy to attend.
DR: I think a focus on empirical economics, social science, and program evaluation would be most promising (and I could help with this). We could also incorporate "applications of economic theory and decision theory." Maybe I would lean away from philosophy and "fundamental theory," as GPI's seminar seems to concentrate on that.
Rethink Priorities would probably (my guess) be willing to attach our name to it and a decent number of our research staff would attend. I think GPI might be interested, hopefully Open Philanthropy and other organizations. Robin Hanson and other academics have expressed interest.
We could try to get a MWE/PoC online seminar 1x per month, for example
Start with...
Presentations of strong, solid working papers and research projects from reputable authors
DR: I would not publicize this until we get off the ground . . . just do it informally, perhaps.
15-20 minute presentations
Provide or link writeup for follow-up comments and
Such a seminar needs association with credible people and orgs to get participation.
Do we need any funding for small honorariums or some such?
Do we want to organize "writeups" and publicity? Should we gain support for this?
See draft above
I think a really neat feature, as a companion to the seminar, could be that the author would (ideally) post a working paper or project website and everyone would leave collaborative annotation comments on it.
This kind of feedback could be golden for the author.
GPI lunchtime seminar (not public)
EA Global talks
RP has internal talks; hope to expand this
"Job-markety PhD students"
Make it desirable to present
Selective choices, "awards" with small stipends? Or "choose a donation"?
Guarantee of strong feedback and expertise
Communication/coverage of work
Encourage experts to attend and give concrete feedback
Open and saved chat window
Write up feedback; consider drawing future presenters from the attending crowd
Sciety is essentially a hub for curating the sort of evaluations that Unjournal aims to do. Users can access research works that have been publicly evaluated.
There are several initiatives of public—and sometimes journal-independent—peer evaluation, including around two dozen , such as the , , and . However, these are nearly exclusively in biology and related areas.
Sciety’s mission is to grow a network of researchers who evaluate, curate and consume scientific content in the open. In doing so, we will support several long-term changes to scientific communication:
is a fairly well respected (?) journal in Life Sciences. Their New Model (originally called "publish, review, curate") was big news. Their three-month update seems fairly stable and successful. Here's their FAQ. Their model is similar to ours in many ways, but it's mainly or exclusively for life sciences. They use for curation.
They don't have explicit quantitative metrics, but an "eLife assessment . . . is written with the help of a common vocabulary to ensure consistency," which may proxy this.
Evaluators (reviewers) are not compensated. ("We offer remuneration to our editors but not to our peer reviewers.")
The Peer Commuities In organization (PCI) and the Peer Community Journal, a diamond open access journal, have considerable overlap with The Unjournal model. They started out (?) as a "recommendation system" but now have established the PCI Journal to "publish unconditionally, free of charge, exclusively, immediately (as soon as possible) [and on an opt-in basis] . . . any article recommended by a PCI." Each PCI has its own policies.
Especially relevant to The Unjournal are these aspects of their program:
The standard "recommender" model has an approved recommender volunteer to play the role of managing editor for a paper and make the decisions; authors are consulted to recommend reviewers.
This might bring up concerns about conflict of interest, e.g., I become "recommender" for a friend or for the stuff that supports my agenda.
There are 17 "Peer Communities in" (i.e., research areas)—mainly in life sciences (some seem to have just started; there are no public preprints up).
Authors must submit work.
They (opt-in) "publish" the article rather than being an "overlay journal," to improve their indexing possibilities (but this is opt-in; you can also submit elsewhere and there are "PCI-friendly" journals).
They depend on volunteer evaluations.
Their evaluation is 0/1 and descriptive rather than quantitative.
Shift the publishing decision from editors to authors
Move evaluation and curation activity from before to after publication
Our community-driven technology effort is producing an application that can support the changes in behaviour required to secure this future.
Reviewers' names are not displayed. ("All public comments posted alongside a preprint will be signed by eLife and not by individuals, putting the onus on eLife as an organisation and community to ensure that the outputs of our peer-review process are of the highest standard.")
They charge a $2,000 APC. Presumably, this is true for all "reviewed preprints" on the eLife website, whether or not you request it become a "version of record."
The evaluation is non-exclusive unless you request that the reviewed preprint be a "'Version of Record' that will be sent to indexers like PubMed and can be listed in funding, job applications and more."
Some share of the work they cover are registered reports.
Ben West on how to make the EA Forum involve more rigorous processes
Ozzy at QURI — I will contact him about "how to do better evaluation"
Stefan Torges post:
Link: "Asterisk is a quarterly magazine of clear writing and clear thinking about things that matter."
Asterisk is a new quarterly journal of ideas from in and around Effective Altruism. Our goal is to provide clear, engaging, and deeply researched writing about complicated questions. This might look like a superforecaster giving a detailed explanation of the reasoning they use to make a prediction, a researcher discussing a problem in their work, or deep-dive into something the author noticed didn’t quite make sense. While everything we publish should be useful (or at least interesting) to committed EAs, our audience is the wider penumbra of people who care about improving the world but aren't necessarily routine readers of, say, the EA forum.
Includes "Speculative pieces with 'epistemic signposts'"
Possible scope for collaboration or sharing
Followup on crucial research—I will share non-sensitive parts of Airtable
Sharing a database/CRM of
Experts to vouch
Interested academics and good writers
Shared thinking on "what is relevant" and "how to classify things"
Unjournal could '"feed in" to Asterisk: Academic article, then you do a writeup; they have funding, can pay authors ~$4,000 for 4000 words; can't guarantee that academic work will feed into Asterisk
Passing back and forth relevant work and directions to go in
Some shared cross-promotion (e.g., at universities and in policy circles, where both Unjournal and Asterisk are relevant)
In a lot of newsy media, it's very hard to find big-picture reporting without a confidence threshold.
Future Perfect is about popularizing EA ideas; Asterisk is about writing about EA ideas in ways that non-EAs can understand but EAs will also be interested in.
Bulletin of Atomic Scientists as a comparison example ... international control of nuclear weapons .. clarify their thinking for people who did not have the tech
EA outreach is focused on "making people EAs" ... but it's more important to make the ideas acceptable without the polarizing EA brand.
Arising from discussion with Jaime Sevilla
Notable:
PREreview.org
commissions external reviews
NBER, CEPR etc -- very loosely filtered within member network
World Bank, Federal Reserve, etc. Internal review?
We designed and disseminated a survey taken by over 1,400 economists in order to (i) understand their experiences with peer review and (ii) collect opinions about potential proposals to improve the system.
...
We reviewed the existing literature about peer review, drawing on sources from inside and outside of economics. ... We then built a (non-comprehensive) themed bibliography,
... we took the additional step of preparing a list of over 160 proposals.
Other peer-review models Our current peer-review system relies on the feedback of a limited number of ad-hoc referees, given after a full manuscript was produced. We consider several changes that could be made to this model, including:
Post-publication peer review: Submissions could be published immediately and then subjected to peer review, or they could be subject to continued evaluation at the conclusion of the standard peer-review process.
Peer review of registered reports: Empirical papers could be conditionally accepted before the results are known, based on their research question and design. A limited number of journals have started to offer publication tracks for registered reports.
The official administrators are David Reinstein (working closely with the Operations Lead); David and Anirudh Tagat (Co-Director) have control and oversight of the budget.
Major decisions are made by majority vote by the Founding Committee (aka the ‘Management Committee’).
Members:
Advisory board members are kept informed and consulted on major decisions, and relied on for particular expertise.
Advisory Board Members:
Crowdsourced peer review and prediction markets: Rather than relying on a small number of referees, the wisdom of crowds could be leveraged to provide assessments of a manuscript's merits.
Non-economists and non-academics as referees: Besides enlarging the size of the pool of referees who assess a paper, the diversity of the pool could be increased by seeking the opinion of researchers from other disciplines or non-academics, such as policy makers.
Collaborative peer review platforms: Communication between authors, reviewers, and editors could be made more interactive, with the implementation of new channels for real-time discussion. Collaborative platforms could also be set up to solicit feedback before journal submission occurs.
(See sections below)
As part of The Unjournal's general approach, we keep track of and maintain contact with other initiatives in open science, open access, robustness/transparency, and encouraging impactful research. We want to be coordinated. We want to partner with other initiatives and tools where there is overlap, and clearly explain where (and why) we differentiate from other efforts.
See our Related Initiatives Comparison for a curated overview of initiatives in the same space, including:
- open review where Unjournal is an evaluation provider
- robustness replication of top papers
- diamond open access with recommender model
See also: and
This initiative and EA/gp Unjournal will interact with the EA forum and build on initiatives coming there.
Some of these links come from a conversation with Aaron Gertler
Here's where to suggest new Forum features.
Here's an example of a that led us to develop a new feature.
Note: Reinstein and Hamish Huggard have worked on tools to help transform R-markdown and bookdown files. Some work can be found on (but may need some explanation).
Jaime Sevilla has thoughts on creating a peer-review system for the Forum. (See embedded doc below, link .)
To create a quick and easy prototype to test, you fork the EA Forum and use that fork as a platform for the Unjournal project (maybe called something like "The Journal of Social Impact Improvement and Assessment").
People (ideally many from EA) would use the Forum-like interface to submit papers to this Unjournal.
These papers would look like EA Forum posts, but with an included OSF link to a PDF version. Any content (e.g., slides or video) could be embedded in the submission.
All submissions would be reviewed by a single admin (you?) for basic quality standards.
Most drafts would be accepted to The Unjournal.
Any accepted drafts would be publicly "peer reviewed." They would achieve peer-reviewed status when >x (3?) people from a predetermined or elected board of editors or experts had publicly or anonymously reviewed the paper by commenting publicly on the post. Reviews might also involve ratings the draft on relevant criteria (INT?). Public comment/review/rating would also be possible.
Draft revisions would be optional but could be requested. These would simply be new posts with version X/v X appended to the title.
All good comments or posts to the journal would receive upvotes, etc., so authors, editors and commentators would gain recognition, status and "points" from participation. This is sufficient for generating participation in most forums and notably lacking in most academic settings.
Good papers submitted to the journal would be distinguished by being more widely read, engaged with, and praised than others. If viable, they would also win prizes. As an example, there might be a call for papers on solving issue x with a reward pool of grant/unconditional funding of up to $x for winning submissions. The top x papers submitted to The Unjournal in response to that call would get grant funding for further research.
A change in rewards/incentives (from "I had a paper accepted/cited" to "I won a prize") seems to have various benefits.
It still works for traditional academic metrics—grant money is arguably even more prized than citations and publication in many settings
It works for non-academics who don't care about citations or prestigious journal publications.
As a metric, "funds received" would probably better track researchers' actual impact than their citations and acceptance in a top journal. People won't pay for more research that they don't value, but they will cite or accept that to a journal for other reasons.
Academics could of course still cite the DOIs and get citations tracked this way.
Reviewers could be paid per-review by research commissioners.
Here is a quick example of how it could work for the first run: Open Philanthropy calls for research on something they want to know about (e.g., interventions to reduce wild animal suffering). They commit to provide up $100,000 in research funding for good submissions and $10,000 for review support. Ten relevant experts apply and are elected to the expert editorial boards to review submissions. They will receive 300 USD per review and are expected to review at least x papers. People submit papers; these are reviewed; OP awards follow-up prizes to the winning papers. The cycle repeats with different funders, and so on.
I suppose I like the above because it seems pretty easy and actionable to do over as a test run for something to refine and scale. I estimate that I could probably do it myself if I had 6–12 months to focus on it. However, I imagine that I am missing a few key considerations as I am usually over-optimistic! Feel free to point those out and offer feedback.
9 Apr 2024: This section outlines our management structure and polices. More detailed content is being moved to our private (Coda.io) knowledge base.
Tech, tools and resources has been moved to it's own section
Update Feb. 2024: We are moving the discussion of the details of this process to an internal Coda link (here, accessible by team members only). We will present an overview in broad strokes below.
See also Mapping evaluation workflowfor an overview and flowchart of our full process (including the evaluation manager role).
Compensation: Evaluation managers are compensated a minimum of $300 per project, and up to $500 for detailed work. Further work on curating the evaluation, engaging with authors and evaluators, and writing detailed evaluation summary content can earn up to an additional $200.
If you are the evaluation manager please follow the process described in our private Coda space
Engage with our previous discussion of the papers; why we prioritized this work, what sort of evaluators would be appropriate, what to ask them to do.
Inform and engage with the paper's authors, asking them for updates and requests for feedback. The process varies depending on whether the work is part of our or whether we require authors' permission.
Find potential evaluators with relevant expertise, contact them. We generally seek two evaluators per paper.
See also:
See also:
We give the authors two weeks to respond before publishing the evaluation package (and they can always respond afterwards).
Once the evaluations are up on PubPub, reach out the evaluators again with the link, in case they want to view their evaluation and the others. The evaluators may be allowed to revise their evaluation, e.g., if the authors find an oversight in the evaluation. (We are working on a policy for this.)
At the moment (Nov. 2023) we don't have any explicit 'revise and resubmit' procedure, as part of the process. Authors are encouraged to share changes they plan to make, and a (perma)-link to where their revisions can be found. They are also welcome to independently (re)-submit an updated version of their work for a later Unjournal
Suggest research-specific issues for evaluators to consider. Guide evaluators on our process.
Read the evaluations as they come in, suggest additions or clarifications if necessary.
Rate the evaluations for awards and bonus incentives.
Share the evaluations with the authors, requesting their response.
Optionally, provide a brief "evaluation manager's report" (synthesis, discussion, implications, process) to accompany the evaluation package.
Mapping collaborator networks through Research Rabbit
We use a website called Research Rabbit (RR).
Our RR database contains papers we are considering evaluating. To check potential COI, we use the following steps:
After choosing a paper, we select the button "these authors." This presents all the authors for that paper.
After this, we choose "select all," and click "collaborators." This presents all the people that have collaborated on papers with the authors.
Finally, by using the "filter" function, we can determine whether the potential evaluator has ever collaborated with an author from the paper.
If a potential evaluator has no COI, we will add them to our list of possible evaluators for this paper.
Note: Coauthorship is not a disqualifier for a potential evaluator; however, we think it should be avoided where possible. If it cannot be avoided, we will note it publicly.
Resources and onboarding information for Unjournal team members.
Archived page: This page contains outdated information from our earlier processes.
Current team members: Please refer to our internal Coda workspace for up-to-date onboarding materials, process documentation, and team resources.
Contact if you need access.
Please ask for group access, as well as access to private channels, especially "management-policies". Each channel should have a description and some links at the top.
Slack is for quick conversations and coordination. Please don't consider Slack as a place to store and organize content. Longer, more structured content should go into Google Docs or similar.
See
Management team: You don't need to edit the GitBook if you don't want to, but we're trying to use it as our main place to 'explain everything' to ourselves and others. We will try to link all content here. Note you can use 'search' and 'lens' to look for things.
Access to the PubPub is mainly only needed for doing 'full-service evaluation manager work'.
Please ask for access to this drive. This drive contains meeting notes, discussion, grant applications and tech details.
Did the people who suggested the paper suggest any evaluators?
We prioritize our "evaluator pool" (people who signed up; see "")
Expertise in the aspects of the work that need evaluation
Interest in the topic/subject
Conflicts of interest (especially co-authorships)
Secondary concerns: Likely alignment and engagement with Unjournal's priorities. Good writing skills. Time and motivation to write the evaluation promptly and thoroughly.
15 Aug 2023: We are organizing some meetings and working groups, and building some private spaces ... where we are discussing 'which specified research themes and papers/projects we should prioritize for UJ evaluation.'
This is guided by concerns we discuss in other sections (e.g., 'what research to target', 'what is global priorities relevant research')
Research we prioritize, and short comments and ratings on its prioritization is currently maintained in our Airtable database (under 'crucial_research'). We consider 'who covers and monitors what' (in our core team) in the 'mapping_work' table). This exercise suggested some loose teams and projects. I link some (private) Gdocs for those project discussions below. We aim to make a useful discussion version/interface public when this is feasible.
Team members and field specialists: You should have access to a Google Doc called "Unjournal Field Specialists+: Proposed division (discussion), meeting notes", where we are dividing up the monitoring and prioritization work.
Some of the content in the sections below will overlap.
'Impactful, Neglected, Evaluation-Tractable' work in the global health & RCT-driven intervention-relevant part of development economics
Mental health and happiness; HLI suggestions
Syllabi (and ~agendas): Economics and global priorities (and adjacent work)
Microeconomic theory and its applications? When/what to consider?
The economics of animal welfare (market-focused; 'ag econ'), implications for policy
Attitudes towards animals/animal welfare; behavior change and 'go veg' campaigns
Impact of political and corporate campaigns
Environmental economics and policy
: How can UJ source and evaluate credible work in psychology? What to cover, when, who, with what standards...
Moral psychology/psychology of altruism and moral circles
Innovation, R&D, broad technological progress
Meta-science and scientific productivity
Social impact of AI (and other technology)
Techno-economic analysis of impactful products (e.g., cellular meat, geo-engineering)
Pandemics and other biological risks
Artificial intelligence; AI governance and strategy (is this in the UJ wheelhouse?)
International cooperation and conflict
See .\
Long term population, growth, macroeconomics
Normative/welfare economics and philosophy (should we cover this?)
Empirical methods (should we consider some highly-relevant subset, e.g., meta-analysis?)
Governance/political science
Global poverty: Macro, institutions, growth, market structure
Evidence-based policy organizations, their own assessments and syntheses (e.g., 3ie)
How to consider and incorporate adjacent work in epidemiology and medicine
To aim for consistency of style in all UJ documentation, a short style guide for the GitBook has been posted . Feel free to suggest changes or additions using the comments. Note this document, like so many, is under construction and likely to change without notice. The plan is to make use of it for any outward-facing communications.
This page is mainly for The Unjournal management, advisory board and staff, but outside opinions are also valuable.
Unjournal team members:
Priority 'ballot issues' are given in our 'Survey form', linked to the Airtable (ask for link)
Key discussion questions in the broad_issue_stuffview inquestions table, linking discussion Google docs
We are considering a second stream to evaluate non-traditional, less formal work, not written with academic standards in mind. This could include the strongest work published on the EA Forum, as well as a range of further applied research from EA/GP/LT linked organizations such as GPI, Rethink Priorities, Open Philanthropy, FLI, HLI, Faunalytics, etc., as well as EA-adjacent organizations and relevant government white papers. See comments ; see also Pete Slattery’s proposal , which namechecks the Unjournal.
E.g., for animal welfare...
We further discuss the case for this stream and sketch and consider some potential policies for this .
Internal discussion space:
DR: I suspect that signed reviews (cf blog posts) provide good feedback and evaluation. However, when it comes to rating (quantitative measures of a paper's value), my impression from existing initiatives and conversations is that people are reluctant to award anything less than 5/5 'full marks'.
Power dynamics: referees don't want to be 'punished', may want to flatter powerful authors
Connections and friendships may inhibit honesty
'Powerful referees signing critical reports' could hurt ECRs
Public reputation incentive for referees
(But note single-blind paid review has some private incentives.)
Fosters better public dialogue
Author and/or referee choose whether it should be single-blind or signed
Random trial: We can compare empirically (are signed reviews less informative?)
Use a mix (1 signed, 2 anonymous reviews) for each paper
We may revisit our "evaluators decide if they want to be anonymous" policy. Changes will, of course never apply retroactively: we will carefully keep our promises. However, we may consider requesting certain evaluators/evaluations to specifically be anonymous, or to publish their names. A mix of anonymous and signed reviews might be ideal, leveraging some of the benefits of each.
We are also researching other frameworks, templates, and past practices; we hope to draw from validated, theoretically grounded projects such as .
See the 'IDEAS protocol' and , 2022
Should we wait until all commissioned evaluations are in, as well as authors' responses, and release these as a group, or should we sometimes release a subset of these if we anticipate a long delay in others? (If we did this, we would still stick by our guarantee to give authors two weeks to respond before release.)
My memory is that when submitting a paper, I usually learn who the Senior Editor was but not the managing editor. But there are important differences in our case. For a traditional journal the editors make an ‘accept/reject/R&R’ decision. The referee’s role is technically an advisory one. In our case, there is no such decision to be made. For The Unjournal, ME’s are choosing evaluators, corresponding with them, explaining our processes, possibly suggesting what aspects to evaluate, and perhaps putting together a quick summary of the evaluations to be bundled into our output. But we don’t make any ‘accept/reject/R&R’ decisions … once the paper is in our system and on our track, there should be a fairly standardized approach. Because of this, my thinking is:
We don’t really need so many ‘layers of editor’ … a single Managing Editor (or co-ME’s) who consult other people on the UJ team informally … should be enough
ME anonymity is probably not necessary; there is less room for COI, bargaining, pleading, reputation issues etc.




