← Back

AI Investor Barometer

About the project

What is AI Investor Barometer?

AI Investor Barometer is an experimental comparison tool that uses five different AI models to generate valuation assumptions for listed stocks daily. Each model — GPT, Claude, Gemini, DeepSeek and Grok — independently produces its own estimate from the same company's public data. Results are presented side by side.

Currently tracking approximately 12 Finnish OMXH stocks and 12 US S&P 500 companies. The pipeline runs automatically on business days.

The project does not provide investment recommendations or advice. It is an experimental research tool: what do different AI models estimate about the same stocks, and do they differ from each other?

Platform in Numbers

5
AI Models
24
Companies Tracked
2.8k+
Estimates Generated
24
Trading Days
2.8k+
Model Runs
9
Weekly Reports
Dataset since: 2026-03-03Last updated: 2026-04-03

Why Compare AI Models?

AI is increasingly involved in investment decisions — either directly or indirectly through analysis tools, news aggregations and chatbots. The problem is that a single model appears reliable even when it is systematically over- or under-estimating certain stocks.

When five different models estimate the same company on the same day, you immediately see:

  • Do the models agree — or do estimates diverge dramatically?
  • Is any model consistently higher or lower than others?
  • Does any model change its estimates daily, while another remains stable?
  • Does any model stay closer to analyst consensus than another?

These questions are essential if you want to understand how different AI models differ. This tool makes model-specific comparison visible.

How AI Models Generate Estimates

Most AI outputs are black boxes: the model gives a number directly without you knowing how it arrived at it. In this project, the approach is different.

1
Model reads public data
Each model receives the same public financial data as input: revenue history, margins, debt structure and analyst consensus estimates.
2
Model produces valuation assumptions
The model does not give an estimate directly. Instead, it estimates the company's future growth rate, profitability and appropriate discount rate level — inputs, not the end result.
3
Deterministic calculator computes the estimate
The model's assumptions are fed into a fixed valuation model that calculates the estimate mathematically. Same calculator, same rules — for all models. This separates the model's "view" from mechanical calculation.

This approach makes model-specific differences transparent: if GPT and Claude arrive at different estimates, it's due to different growth assumptions — not because one "calculated incorrectly".

📐
Valuation Methodology
Full DCF documentation, formulas, and sector profiles
📋
Changelog
Version history and major platform updates
Questions & Answers
Common questions about AI Investor Barometer — how it works, what the numbers me

AI Model Comparison Metrics

Bias Index
Is the model generally higher or lower than consensus? Tracks each model's median valuation gap across the entire universe and visualizes its development over time.
Model Traits
The model's behavioral pattern: does it consistently produce higher or lower estimates, does it change its output daily or remain stable?
Calibration
How closely does the model stay within analyst consensus bounds without external corrections? Higher score means estimates are closer to analyst consensus.
Model Spread
At a glance: where each model places its estimate relative to the current price — per stock, all models side by side.
Accuracy
How well did models predict actual price direction? Tracks directional accuracy across different time horizons (7, 14, 30 days).
AI Market Indices
Three composite indices — AI Valuation Gap (ACDI), Model Disagreement (ADI), and Sentiment Shift (ANM) — track the collective AI view of the market over time.

Limitations and disclaimer

  • Not investment advice. All content consists of AI models' computational estimates, not investment advice. Draw your own conclusions.
  • AI models can be systematically wrong. Models learn from historical data that may be biased or incomplete. A high confidence score does not mean being correct.
  • Public data only. Models use only publicly available financial data and market information.
  • Experimental tool. This is an experimental AI comparison and measurement tool. Data may be incomplete, the scheduler may fail, results may be incorrect.

Pipeline Diagnostics

Last run: 2026-04-03Status: successDuration: 18m 13sCost: $2.0316
ModelAvg LatencyCost / RunValid %
Claude30.6s$0.954100%
Deepseek16.8s$0.055100%
Gemini16.6s$0.263100%
Gpt8.7s$0.391100%
Grok6.3s$0.369100%
Last 30 days: 34 runsTotal cost: $63.0619Avg: $1.8548/day

Contact & Feedback

Have feedback, a collaboration idea, or a question? We'd love to hear from you.

AI Investor Barometer · Experimental AI model comparison tool← Back to home page
All content is generated by AI models and may contain errors. This is an experimental tool — not investment advice, research, or recommendation. Terms of Use · Privacy Policy