a1c3f03d7d
also: - refactor and significantly simplify database-wrapper code - rename manualSendToMongo to manualSendToDb |
||
---|---|---|
docs | ||
input | ||
public | ||
src | ||
.gitignore | ||
.nvmrc | ||
env.example | ||
next-env.d.ts | ||
package-lock.json | ||
package.json | ||
postcss.config.js | ||
Procfile | ||
README.md | ||
specification.json | ||
tailwind.config.js | ||
tsconfig.json |
What this is
https://metaforecast.org is a search engine for probabilities from various prediction markes and forecasting platforms (try searching "Trump", "China" or "Semiconductors").
This repository includes a source code for the website, as well as a set of libraries that fetches probabilities/forecasts.
We also provide a public database, which can be accessed with a script similar to this one.
I also created a search engine using Elicit's IDE, which uses GPT-3 to deliver vastly superior semantic search (as opposed to fuzzy word matching). If you have access to the Elicit IDE, you can use the action "Search Metaforecast database". However, I'm not currently updating it regularly.
How to run
1. Download this repository
$ git clone https://github.com/QURIresearch/metaforecast
$ cd metaforecasts
$ npm install
2. Set up a database and environment variables
You'll need a PostgreSQL instance, either local (see https://www.postgresql.org/download/) or in the cloud (for example, you can spin one up on https://www.digitalocean.com/products/managed-databases-postgresql or https://supabase.com/).
Environment can be set up with an .env
file. You'll need to configure at least DIGITALOCEAN_POSTGRES
for the fetching to work, and NEXT_PUBLIC_SITE_URL
for the frontend.
See ./docs/configuration.md for details.
3. Actually run
npm run cli
starts a local CLI which presents the user with choices; if you would like to skip each step, use the option number instead, e.g., npm run start 14
.
npm run next-dev
starts a Next.js dev server with the website on http://localhost:3000
.
4. Example: download the metaforecasts database
$ git clone https://github.com/QURIresearch/metaforecast
$ cd metaforecasts
$ npm install
$ node src/backend/manual/manualDownload.js
Code layout
- frontend code is in src/pages/, src/web/ and in a few other places which are required by Next.js (e.g. root-level configs in postcss.config.js and tailwind.config.js)
- various backend code is in src/backend/
- fetching libraries for various platforms is in src/backend/platforms/
- rudimentary documentation is in docs/
What are "stars" and how are they computed
Star ratings—e.g. ★★★☆☆—are an indicator of the quality of an aggregate forecast for a question. These ratings currently try to reflect my own best judgment and the best judgment of forecasting experts I've asked, based on our collective experience forecasting on these platforms. Thus, stars have a strong subjective component which could be formalized and refined in the future. You can see the code used to decide how many stars to assign here.
With regards the quality, I am most uncertain about Smarkets, Hypermind, Ladbrokes and WilliamHill, as I haven't used them as much. Also note that, whatever other redeeming features they might have, prediction markets rarely go above 95% or below 5%.
Tech stack
Overall, the services which we use are:
- Algolia for search
- Netlify for website deployment
- DigitalOcean for background jobs, e.g. fetching new forecasts
- Postgres on DigitalOcean for database
Various notes
- Commits follow conventional commits
- Right now, I'm fetching only a couple of common properties, such as the title, url, platform, whether a question is binary (yes/no), its percentage, and the number of forecasts.
- For elicit and metaculus, this library currently filters questions with <10 predictions.