If one were to build an open platform to compete with news media rather than social or blogging media, they would need some sort of accountability/quality control mechanism to mirror the role of editorial review at prestigious news platforms.
The best tool for aggregating information that we have devised is markets. This is especially the case in open systems. The biggest risk to a well-functioning market is collusion. Any mechanism that was designed to employ markets in the place of editorial review would need to guard against this. With this in mind I propose this system:
A reporter with information on a climate event writes an article and publishes it to the platform. They must stake a certain TBD amount of money on this article.
The article is published without editorial review (unlike how news media currently works).
Prospective fact checkers (post editorial reviewers) also stake money and state their specialist topics. Some of those that listed climate as a topic are randomly chosen from the available pool of people. They are each given guidelines on how to judge an article and use these guidelines to give the article a trust score without colluding among themselves as they don’t who’s been asked to fact check. The article and the guidelines together represent a Schelling Point the fact checkers can converge on. A fact checking assignment is like jury duty. Some of your stake is slashed if you renege on giving a score for a given article. This requirement along with random selection should minimise any collusion.
The writer’s payout is determined by the score the fact checkers give him/her. The fact checkers’ payouts are determined by how close they are to the average score from the group. Everyone’s respective rep scores are also updated. Those that drop below a certain trust score will not have their articles listed on the platform or be chosen for post editorial review. Articles from the best performers will be amplified on the platform, minimising the chances of readers being supplied erroneous information.
Would be interested to hear people’s thoughts on this system? One worry I have is that after some time people will be armed with prior probabilty data and will simply strategically pick the high probabilty outcome without doing any fact checking. This could be mitigated by minimising the amount of tasks a fact checker gets, leading them to take the utmost care with the ones they are given. It’s not a perfect mitigation by any means though.