Interview with the winner of the “Right!” said FRED Challenge

Interview with the winner of the “Right!” said FRED Challenge

In this interview, we sit down with the winner of the “Right!” said FRED Challenge for Q2 2024, Sigitas Keras. Known on GJ Open as sigis, Sigitas is an experienced quant and trader who decided to explore the world of forecasting after an impressive 25-year career in finance. With a PhD in mathematics and a natural curiosity about the world, he shares insights into the unique challenge he has taken on to forecast every question on GJO in 2024 and the strategies that helped him excel on the platform. Originally from Lithuania, Sigitas currently lives in Canada.

GJO: What is your background, and how did you first become interested in forecasting?

I was born in Lithuania, have a PhD in maths, but, as many others with a similar background, ended up in finance industry. After almost 25 years as a quant and a trader, I recently retired, which freed up a lot of time for other things. I tried forecasting on GJO for the first time a couple years ago. It seemed like an interesting challenge where I could combine analytical skills and general curiosity about the world.

GJO: How did you learn about GJ Open? How would you describe your experience on the platform so far?

I read Tetlock’s book Superforecasting, so likely that was an initial prompt, but to be honest I don’t remember full details anymore. Rightly or wrongly, I am one of the few forecasters who decided to forecast every question in 2024. It was very enjoyable, and I feel I learnt a lot both about forecasting and about various topics, but I have to admit this is getting too difficult to maintain. I don’t think I’ll continue doing all questions next year, and most likely will just focus on a few challenges, but I still like to maintain a good mixture of various topics.

GJO: What was your approach to the “Right!” said FRED Challenge? What do you think helped you come out on top?

I like questions that have good supporting data. In that sense, the FRED challenge is perfect for me. Whenever there is good data available, I try to use some mathematical model. Having a background in finance industry helps a bit with that, although I don’t think I use anything that requires more than FRED and other publicly available data and a Google spreadsheet. I also try to update my forecasts regularly, typically once a week. I think consistency is another important component of successful forecasting.

GJO: What topics would you consider of particular interest to forecast for 2025 and beyond?

I tend to forecast better when there is good data available for analysis. On the other hand, geopolitical questions are often much more challenging, so perhaps I will focus on improving there. My goal is to improve my score in the Superforecasting Workshops challenge!

GJO: Is there anything you would like to add that would be of particular interest to other forecasters on GJ Open?

I feel I am still very new to forecasting and to the community. One thing I hope is to learn more about other forecasters, their backgrounds, their approaches to forecasting. And if anyone has any questions for me, feel free to reach out.

See the latest forecasting challenges on GJ Open and try your hand at forecasting!

Keeping It Civil – Promoting an Open-Minded Dialog on Good Judgment Open

Keeping It Civil

A wise crowd encompasses diverse views.

Good Judgment Project research found that being an “actively open-minded thinker” is positively correlated with being an accurate forecaster. That’s no mystery. Exposure to views with which we disagree – even views that we find repugnant – can inform our understanding of the world in which we live. And the better we understand that world, the better we can project what will happen under various conditions.

On Good Judgment Open (GJO), our public forecasting platform, we strive to foster the candid exchange of opinions without personal attacks on others who express opposing views and without using profane language or offensive epithets. We do not censor comments simply because they express opinions with which we, individually or as a company, disagree.

We rely primarily on our forecasting community to let us know if any comments posted on GJO fall outside the reasonable bounds of civil discourse. Any forecaster can “flag” a comment, which triggers a review by our site administrators. We’re happy to report that flagged comments are rare; from time to time, however, we have deleted inflammatory remarks, especially ones that personally insult other forecasters, and have cautioned GJO forecasters to find ways to “disagree without being disagreeable.”

Can Forecasting Tournaments Reduce Political Polarization?

The relative rarity of nastiness on GJO may surprise those accustomed to the rough-and-tumble of the Twitterverse. But it’s little surprise to Good Judgment researchers. Our co-founder Barb Mellers launched a three-year, National-Science-Foundation-funded research program to investigate whether participation in forecasting tournaments could moderate political polarization. Specifically, Mellers and her co-authors Philip Tetlock and Hal Arkes wanted to explore:

How feasible is it to induce people to treat their political beliefs as testable probabilistic propositions open to revision in response to dissonant as well as consonant evidence and arguments?

Mellers, B., Cognition, https://doi.org/10.1016/j.cognition.2018.10.021.

They put this question to the test for two years in a forecasting tournament they dubbed the Foresight Project. This tournament went beyond prior Good Judgment Project research in that it included questions relating to controversies in US domestic politics and not just about geopolitical events elsewhere. That increased the likelihood that Foresight Project participants, who were predominantly US residents, would hold strong views related to the forecasting questions.

The Mellers et al. findings shed light on why the conversation on Good Judgment Open seems so civil compared to what we see elsewhere on social media.

Forecasting tournaments are not a panacea for what ails our political conversations. Even the most accurate forecasters – including some who earn the Superforecaster® credential – occasionally express their views in polarizing prose and endorse opinions that many consider to be immoral. In that respect, forecasting accuracy is like other forms of competence − in Phil Tetlock’s words, “[t]here is no divinely mandated link between morality and competence.”[1]

Nonetheless, we are optimistic that Good Judgment Open contributes to a more thoughtful public dialog and encourages our forecasters to listen carefully to points of view that they might otherwise dismiss. For example, we took great pleasure in hosting an “adversarial collaboration” challenge on the Iran nuclear deal, inspired by this New York Times op-ed co-written by Phil Tetlock and Peter Scoblic. We plan to engage GJO forecasters with more opportunities to test whether their views on controversial subjects lead to more or less accurate predictions.

As the 2020 election cycle intensifies, we forecast with near certainty (p>.99) that the public debate will grow even more heated and more personal. Together, let’s preserve Good Judgment Open as a place where facts and reasoned argument reign supreme. We have nothing to lose but our illusions.

[1] Tetlock, P., & D. Gardner (2015). Superforecasting: The Art and Science of Prediction. New York: Crown. P. 229.

How to Become a Superforecaster®

How to Become a Superforecaster®

A BBC listener asked the team behind their “CrowdScience” radio show and podcast whether she might, in fact, be a Superforecaster. She’s not the first to wonder if she would qualify – and not the first to be curious about how Good Judgment spots and recruits superior forecasting talent.

For all curious souls out there, here’s the inside scoop.

Superforecaster Origins: the Good Judgment Project

Superforecasters were a surprise discovery of the Good Judgment Project (GJP), the research-and-development project that preceded Good Judgment Inc. GJP was the winner of the massive US-government-sponsored four-year geopolitical forecasting tournament known as ACE.

As our co-founder Phil Tetlock explains in his bestseller Superforecasting: The Art and Science of Prediction, we set up GJP as a controlled experiment with random assignment of participants. Results from the first year of the tournament established that teaming and training could boost forecasting accuracy. Therefore, we selected GJP superforecasters from the top 2% of forecasters in each experimental condition to account for the advantages of having been on a team and/or received training. To minimize the chance that outstanding accuracy resulted from luck rather than skill, we limited eligibility for GJP superforecaster status to those forecasters who participated in at least 50 forecasting questions during a tournament “season.”

Professional Superforecaster Selection

When the ACE tournament ended in mid-2015, Good Judgment Inc invited the forecasters with the best-established track records to become the core of our professional Superforecaster contingent. Most of Good Judgment Inc’s 150+ professional Superforecasters qualified through their relative accuracy during GJP. They had not only earned GJP superforecaster status over one tournament season, but also confirmed their accuracy over 50+ questions in a second year of forecasting.

Good Judgment has continued to identify and recruit new professional Superforecasters since the ACE tournament ended via our public forecasting platform, Good Judgment Open. Many of these new recruits never had a chance to participate in ACE − and they have performed every bit as well as the “original” GJP superforecasters.

Do You Have What It Takes?

Each autumn, Good Judgment will identify and recruit potential Superforecasters from the ranks of GJ Open forecasters. If you think you have what it takes to be a Superforecaster, put that belief to an objective test: forecast on at least 100 GJ Open questions (cumulatively, not necessarily in one year). We look at many metrics, but the one we look at closest is average accuracy score per closed question. We also look at comment quality, because understanding the rationales for forecasts is important to our partners and clients. Collegiality is also important to ensure good teamwork with other Superforecasters. And it goes without saying, you’ll need to be in compliance with the GJ Open Terms of Service.

Those who consistently outperform the crowd are automatically eligible for our annual pro Super selection process and potentially a trial engagement. Those who successfully complete a three-month probation period will become full-fledged Superforecasters.

So, there you have it. Becoming a professional Superforecaster is the ultimate meritocracy. We don’t care where you live (as long as you have reliable Internet access!). When evaluating your qualifications for Superforecaster status, we ignore your gender, age, race, religion, and even education. (We are, however, keen to have an increasingly diverse pool of professional Superforecasters and encourage people of all backgrounds to test their skills on GJ Open!) We simply want to find the world’s most accurate forecasters – and to nurture their talents in a collaborative environment with other highly skilled professionals.

For those who are not invited, please continue on GJ Open and improve your forecast skills. We find the process often identifies GJ Open “veterans” who have recently elevated their performance into the top tier.

… If You’re Still Curious

While you’re pondering whether you have what it takes to be a Superforecaster, check out the BBC CrowdScience podcast episode that answers their listener’s question.

And, as always, you can visit our Superforecaster Analytics page to learn more about how Good Judgment’s professional Superforecasters provide early insights and well-calibrated probability estimates about key risks and opportunities to help governments, corporate clients, and NGOs make better decisions.