1
Why Private Beats Public (at First)
The sceptic: “Private markets are just polls with extra steps. If public markets are hard, private ones will be irrelevant.”
The sceptic is right about one thing: a WhatsApp poll with a fancier interface is not a product. But a well-designed prediction market adds three things that no group chat can provide — a shared probability that moves as people commit to it, a scoreboard that persists beyond the conversation, and a resolution moment that everyone returns to. That is a structural difference.
In every Indian group chat with more than ten members, prediction is already happening. Before a cricket match, people state their views. After it, they argue about who called it correctly. The conversation evaporates. The person who called three matches correctly is indistinguishable from the one who called one and talked about it for a month. The signal is real. The architecture to capture it does not exist.
WePredict Private is that architecture. It is not a financial product. It is a game object — a shared scoreboard for groups that already argue about outcomes.

Private changes the cold start geometry
Public prediction markets suffer from the empty-room problem. You need density to create price discovery, movement, and social energy. Without it, every market looks dead. Building that density from scratch requires user acquisition, sustained engagement, and patience — and most public platforms have spent years on this problem.
Private prediction markets invert the geometry entirely. The room already exists. The WhatsApp group, the college alumni chat, the office cricket gang, the neighbourhood society — these are assembled communities, active daily, already predicting informally. You are not asking people to join something new. You are giving an existing room a game to play. The first market in a group of twenty friends who already argue about cricket does not need twenty strangers to make it meaningful. It needs one person to send a link.
Private also changes the content constraints
Public markets attract scrutiny around team names, brand identities, league rights, and financial instruments. In private groups, these conversations are already happening informally. A market on “Will Rohit score a fifty tonight?” inside a group of thirty cricket fans is not a public financial instrument — it is a structured version of something the group was already doing. The platform is not creating a new activity. It is giving an existing one a scoreboard.
The two surfaces — and why both are needed
The architectural framing that matters throughout this series is simple: NeoMails earns Mu. WePredict Private spends Mu inside groups. The inbox is the earn layer. The group is the burn layer. These are not competing surfaces — they are a loop. Without the earn layer, Mu has no credibility. Without the burn layer, Mu has no drama.

Does play money produce real behaviour?
The most common objection to this structure is that play money produces cheap talk. Real consequence requires real stakes. The evidence says otherwise. The Servan-Schreiber et al. study compared real-money and play-money prediction markets across 208 sports events and found no statistically significant systematic accuracy difference. Tetlock’s Good Judgment Project ran for years on pure reputation and scoring — no financial stake — and produced forecasters who beat intelligence analysts with access to classified information. At Manifold Markets today, the largest play-money platform, community predictions average within four percentage points of true probability.
The conclusion the evidence supports is not that money is irrelevant. It is that money is one mechanism for creating skin in the game — and social consequence is another. In a closed group of people you see regularly, social consequence may be the stronger force. Losing money in an anonymous public market is a private financial event. Losing Mu to your friend on the same market, in a group that watched both of you, is a social event. The social frame is what turns virtual currency into real consequence.
WePredict Private is group forecasting as a game — not public betting, not corporate analytics. A shared scoreboard for groups that already argue about outcomes.
Measurable commitment: We will optimise first for one metric — repeat use by the same group, not viral reach. If closed groups do not return for the next resolution moment, we have not built a product. We have built a gimmick.
2
The WhatsApp Mode: When Your Group Chat Gets a Scoreboard
The sceptic: “WhatsApp is for sharing and commenting. Nobody wants markets in family groups.”
This is true if you lead with the word “market”. WhatsApp groups do not want complexity. They want banter, speed, and status. The prediction is the occasion for the banter — not the other way around.
India already has a culture of informal social prediction that has no equivalent in most markets. The hostel senior who mapped the semester’s exam paper pattern before the syllabus was finalised. The market trader who reads a commodity’s direction in the quality of Tuesday morning enquiries. The old man at the temple who has predicted every local election in his ward for thirty years and keeps no record because he has never needed one. These are recognised social identities — people whose forecasting accuracy is tracked informally, remembered, and referenced for years. India is comfortable treating prediction as a form of expertise and social capital in a way that most cultures are not.
WePredict Private formalises what already exists, and adds the one thing informal prediction lacks: a persistent, compounding record that separates the genuinely calibrated from the merely confident. The old man at the temple knows his record. So does the ward. But the ward changes, and memory is not a ledger. What he has accumulated over thirty years lives only in the heads of people who were paying attention — and those people are not always the ones in the room when the next prediction is made. WePredict Private is the ledger he never had.
The unit of distribution is a forecast card, not a market
The instinct most product teams follow is wrong: build a market interface, then tell people to go visit it. This requires behaviour change. It asks people to add a new destination to their daily routine. Most people will not.
The right unit is a shareable forecast card — a visual object that travels into the group and brings the market to where the conversation already lives. The card shows the question, the current group probability, the time remaining, the top forecasters in the group, and one obvious action: Join. The market lives on a PWA; the card lives in the chat. The market comes to the group — the group does not come to the market.

Resolution follows the same logic. A results card arrives the next morning, shows who was right, updates the leaderboard, and gives the group something to react to. The NeoMail that arrives in each member’s inbox carries the resolution as a moment — pulling the inbox and the group into a shared ritual.
The rituals that fit WhatsApp naturally
The formats that work share three properties: they have natural close times that align with when the group is already active, they produce results the group cares about independently of the market, and they are light enough to run in mixed company.
Cricket is the anchor for India — matchday markets on match winner, top scorer, first wicket, first boundary. Weekend entertainment markets on box office bands and award winners work for film groups. Local life markets — will the wedding end before midnight, will the monsoon arrive before the meteorologists say it will, will the neighbourhood’s most eligible bachelor announce his engagement before the year is out — feel genuinely local in a way no public market can replicate. What all of these share is cadence: not an infinite menu of markets, but a small number of recurring rituals.
Why play money works better here than in public markets
In public markets, the primary stake is financial. In a WhatsApp group of people you see regularly, the stake is reputational — and reputational stakes bite harder when the audience is your actual peers. Mu earns its meaning here through three mechanisms: earned scarcity (a Mu balance represents weeks of NeoMails engagement, not a sign-up bonus), social comparison (your stake and result are visible to the group), and compounding record (you are not winning once — you are building something that persists).
Losing Mu alone is mildly annoying. Losing Mu to your friend, in a group that will reference it for the next fortnight, is genuinely felt. The social frame is the product.
Play money also enables mass participation that real-money platforms structurally cannot. In India, real-money prediction platforms face significant legal friction. WePredict Private has no cash barrier. Anyone with a Mu balance — earned through daily NeoMails engagement — can participate. The inclusivity is not a compromise. It is a structural advantage over any real-money competitor.
Guardrails to name honestly
Private reduces scrutiny. It does not remove responsibility. From day one: invite caps and rate limits (anti-spam), group admin controls over whether markets can be created, a clear list of what is not allowed (targeted harassment, political markets involving named candidates, anything that reproduces the information asymmetry of financial insider trading). These are not complex to implement. They are simple defaults that signal the platform takes its obligations seriously.
WePredict Private is working when a group creates a weekly ritual and sustains it for six to eight weeks without prompting. Not novelty. Habit.
3
The Slack Mode: Markets as a Thinking Tool
The sceptic: “In companies, prediction markets die. They’re fragile, politically sensitive, and they don’t survive champion churn. The history is clear.”
The sceptic is pointing at a real pattern — but drawing the wrong conclusion. The history of internal prediction markets does not show that they fail to produce useful intelligence. It shows that they fail to survive as side-project experiments. That is a design problem, not an evidence problem.
What the history actually shows
HP ran internal markets from 1996 to 1999 to forecast computer workstation sales — more accurate than official internal forecasts in six out of eight cases. Google’s Prophit launched in 2005; within three years, 20% of all employees had placed bets, and it became an HBS case study. Google ran a second market in 2020 with over 175,000 predictions from more than 10,000 employees, covering COVID-19 timelines, engineering milestones, and technology trends. Ford used prediction markets for car sales forecasting and achieved 25% lower mean squared error than its own expert forecasters.
The evidence that internal prediction markets can produce genuine intelligence is strong. The honest problem is durability. Most programmes faded when their internal champion left, or when the market was not embedded in operational workflow. HP’s market ended when the Caltech collaboration ended. Google’s Prophit ended when Bo Cowgill moved on. The lesson is not that markets do not work. It is that markets built as experiments — dependent on a single advocate — are fragile by design. The governance and the workflow integration must be built into the product itself.
Two jobs — and only two
To avoid overselling, keep the enterprise promise narrow. Internal prediction markets do two jobs well.
The first is forecasting: will we hit the quarterly number, will this sprint ship on time, will the partnership close by month-end, will the new feature reach 10,000 users by quarter-end. Questions with clear resolution criteria, meaningful consequences for being wrong, and dispersed information in the organisation that is not reaching decision-makers through normal reporting channels.
The second is alignment: surfacing what the organisation already suspects but cannot say cleanly because hierarchy distorts speech. Every company has a HiPPO problem — the Highest Paid Person’s Opinion dominates, not because it is most accurate but because the people with better information are not empowered to contradict it in a status meeting. A junior engineer who knows a project is going to be late cannot always say so in a stand-up. But they can stake Mu on a market asking whether the sprint will ship on time. The market aggregates the views of everyone willing to express a probability, and the result is visible to management without requiring any individual to go on record. That is not surveillance. It is psychological safety through structure.
Slack is not WhatsApp — and pretending otherwise kills both
This is the design principle that matters most for Slack-based private prediction markets. The WhatsApp mode and the Slack mode share an infrastructure — the Mu currency, the market engine, the Predictor Score. But they are not two modes of the same product. They are two products on a shared infrastructure. Treating them as the same, and building one interface to serve both, is how you end up serving neither well.

WhatsApp mode is entertainment-first. Banter is the feature. The prediction is the occasion for the banter. Friction should be minimal.
Slack mode is decision-support. The prediction is the product. Banter can be a bug. Some friction — a required “evidence link” when creating a market, a mandatory resolution date, an admin approval workflow — signals that this is a serious tool, not a game, and that matters for adoption in a professional context.
What Slack mode specifically requires that WhatsApp mode does not: templates for common market types (“Will Sprint 14 ship by Friday 6 pm?”, “Will Q4 sales land above ₹X crore?”), scheduled weekly rituals that run automatically without manual creation, an anonymity option for honest forecasting in hierarchical organisations, admin controls and topic restrictions (no markets on promotions, redundancies, HR matters, or public company financials), an audit trail, and a calibration dashboard that shows — over time — which individuals and teams are consistently well-calibrated on which types of questions.
That last element is the enterprise moat. A calibration record showing that a particular team consistently underestimates delivery time by two weeks is actionable management intelligence. It cannot be obtained through performance reviews, surveys, or observation — because all of those measure outcomes that individuals do not fully control. Calibration data measures the quality of probabilistic judgement over time, in conditions where there is a genuine incentive to be honest. That compounds with every market that runs.
The “no money needed” proof
For those who remain unconvinced that play money can drive serious enterprise forecasting: Metaculus runs entirely on points and public reputation, with no currency at all. It attracts policy analysts, researchers, and domain experts, and its aggregate predictions consistently outperform expert panels. The scoring system — a proper logarithmic rule that rewards honest probability estimates — does what financial incentives do in public markets: it creates skin in the game. The Predictor Score in WePredict Private is the same idea, applied to the contexts people actually inhabit.
Slack markets are not for everything. They are for decisions where being wrong is expensive — and learning fast is more valuable than protecting the plan. We will start with one team, one template, one monthly calibration report — and expand only if the forecasts are measurably better than existing status updates.
4
The Bridge: One Mu Wallet, Many Rooms
The sceptic: “Even if this works in groups, it won’t scale. Every group is its own island. There’s no compounding. You’ve built fragmentation by design.”
This is the most important challenge because it is actually a design question dressed as a sceptical one. The answer to it is the answer to why WePredict Private is not a standalone product — it is a critical layer in a larger architecture.
The identity problem that nobody has solved
Consider what the current state looks like for someone who predicts across multiple contexts. They have informal reputation in their WhatsApp cricket group as the person who always calls it right. They are a reliable forecaster in their office chat. They occasionally participate in public prediction markets. These identities are entirely disconnected. The calibration record from one context does not travel to another. The reputation earned in one room has no meaning anywhere else. Every new context starts from zero.
This is not a minor inconvenience. It is the structural reason prediction behaviour does not compound into a durable identity. Without portability, the forecaster is always a beginner somewhere, and the platform is always starting from scratch on every user.
One wallet, one score, many rooms
WePredict Private solves this through portable identity: one Mu wallet and one Predictor Score that follow the person across every context they inhabit.
The same person is the cricket pundit in their college alumni WhatsApp group, the delivery-timeline forecaster in their company Slack, and the NeoMails participant earning Mu through daily Magnets. WePredict Private should treat these as one identity — with a single Mu wallet earned in the inbox and spent across groups, a single Predictor Score that compounds across all resolved markets, and context-specific leaderboards that show their rank inside each particular group.
The group is the room. The Predictor Score is the passport.

Why this becomes defensible over time
Platforms can copy a market format. They can build an automated market maker, design a scoring system, create a social leaderboard. What they cannot easily copy is a Predictor Score that a user has been building for eight months across cricket markets, office prediction markets, and public WePredict questions. A calibration record of 74th percentile accuracy on delivery timelines, built over a full year, is not a feature that can be replicated overnight. Neither is the Mu balance that represents months of NeoMails engagement.
This is the moat that the broader WePredict architecture described in previous essays is designed to create. The record of attention — the compounding history of engagement, accuracy, and identity across contexts — cannot be shortcut. A late entrant who builds the same market format starts from zero on every user’s identity. They cannot give someone back the eight months of calibration history they built on WePredict.
How the surfaces strengthen each other
Public markets and private markets are not in competition for the same user behaviour. They are complementary rooms in the same economy.
Public WePredict gives Mu a discovery surface and a density of participants that private groups cannot replicate. A market on the Test series outcome has better price discovery at scale than in a group of twenty friends. It also provides the external calibration benchmark: if a user’s Predictor Score on public markets is strong, that credential travels into their private circles. The public market validates the score that the private market makes socially meaningful.
Private markets give Mu the social context that makes it worth earning in the first place. Staking Mu in an anonymous public market is an intellectual exercise. Staking it in front of the twenty people who will remember it for weeks is a social act. Private markets are where the Predictor Score becomes personal. Public markets are where it becomes credible. Each makes the other more valuable.
The sequencing — three rooms, built in order
The temptation is to build all three surfaces simultaneously. This is the complexity trap: multiple workstreams, each depending on the others, producing something too incomplete to prove and too complex to iterate.
The right order is staged and disciplined. Public WePredict launches first — seeded with cricket, building the Predictor Score infrastructure and establishing the platform as the system of record for forecasting identity. Without this foundation, the Predictor Score is a feature of a feature. With it, private markets are extending an existing identity into new contexts.
WhatsApp private markets launch second, as a feature for existing WePredict users. The cold start is solved because the user already has a Mu balance and a Predictor Score. They are not starting from scratch — they are extending something they have already built into a new social context. Every market card shared into a WhatsApp group is simultaneously a game invitation and a WePredict acquisition channel. The social distribution is organic.
Slack follows third, after the social mechanics are proven and calibration data exists to make the enterprise pitch credible. The claim that “our platform produces forecasters with meaningful calibration on delivery timelines after three months of participation” can only be made after three months of participation data exists. The enterprise case requires evidence, and the evidence comes from the public and social modes first.
Each stage provides what the next stage needs. None of this is simultaneous. All of it compounds.

The 90-day proof plan
The commitments for the first 90 days are intentionally minimal — not because the ambition is small, but because the discipline of proving one thing before adding the next is the entire lesson of the sequencing argument.
For WhatsApp: one weekly ritual, one category (cricket), group leaderboards only. No marketplace, no multi-category menu, no public sharing of group results. One question answered: do groups return after the first market?
For Slack: one team, one template market type, one monthly calibration report. One question answered: do the market forecasts tell us something the status updates did not?
One public learning metric across both: group repeat rate — the proportion of groups that create a second market after their first. If that number is above 50%, the social loop is forming. If it is below 20%, the problem is in the market design, not the currency, and the redesign is cheap.
The system-level proof that the whole architecture is working is a single observable pattern: Mu earned through NeoMails being spent in private group markets, generating crowd signals that flow back into the NeoMail as a teaser that earns more Mu. When that loop exists at scale — not as a feature demo, but as a measurable daily pattern — the attention economy has its social layer.
We will know WePredict Private is working when the Mu wallet earns in the inbox, spends in the group, and the resolution arrives back in the inbox as a ritual people return to. One loop. Many rooms. No shortcuts.
**
The argument about tonight’s match has always been a prediction market.
It just needed a scoreboard. And the scoreboard needs to follow you everywhere you go.
**
WePredict Private in the Wild
Concepts are cheap. Habits are not. Next up are four stories — two from WhatsApp, two from Slack — that show what WePredict Private looks like when it stops being a product spec and starts being something people live through on a Tuesday morning and a Thursday evening. All four are fictional. All four are assembled from patterns of behaviour that are entirely real.
5
WhatsApp Story 1: The Group That Finally Has Receipts
The WhatsApp group is called Hostel C Legends and it has twenty-three members.
It was created as an email list in 2009 by Vikram, who lived in Room 14 of Hostel C at NIT Trichy, on the night India won the T20 World Cup. The original purpose was to coordinate the celebration. A few years later, it transitioned to WhatsApp. Seventeen years later, the group is still active — somewhat improbably, given that its members are now scattered across Bengaluru, Mumbai, Singapore, New Jersey, and one persistent outlier in Coimbatore who nobody has visited but everyone likes — and its primary function is still, in some essential way, cricket.

The group has a mythology. It has recurring characters. There is Prashant, who works at a fintech in Bengaluru and is considered the group’s most reliable cricket analyst — calm, data-driven, occasionally insufferable about it. There is Deepak in New Jersey, who watches matches at 4am and compensates for the time zone with aggression. There is Meera, who joined in 2012 when she married Vikram and whose predictions everyone agrees are suspiciously accurate for someone who claims not to follow the game closely. There is Anand, who has predicted India to lose every pressure match for eight years on the grounds that “pressure is real,” and is technically correct often enough to remain credible. And there is Karthik — who confidently predicts whatever the group consensus appears to be, ten minutes after the consensus has formed, and presents it as independent analysis.
Through the years this group has argued about cricket the way families argue: with love, with memory, and with a running ledger of who was right and who was catastrophically wrong that exists nowhere except in individual recollections, and is therefore subject to endless, unresolvable dispute. Prashant believes his prediction record is excellent. Deepak believes his is better. Meera does not engage with this argument, and therefore wins it. Karthik has been wrong about nine consecutive finals and remembers none of them.
In late April 2026, in the middle of IPL, Vikram drops a card into the group.
**
It is a simple thing. A visual card, roughly the width of a phone screen, that sits in the chat the way a news article or a meme would sit — familiar, scrollable, immediately readable. It says:
WePredict Private — Hostel C Legends
Will Chennai beat Mumbai tonight?
Group probability: 54% Yes
Closes 7:30pm — 4 members have staked
[Join]
The first reaction is what first reactions always are:
“What is this?” “Are we gambling now?” “Who has time for this?”
And then, from Deepak in New Jersey at whatever ungodly hour it is there: “I’ll do it if Prashant does it.”
Prashant does it within four minutes. He stakes 200 Mu on Yes and explains his reasoning in three paragraphs. The group is used to this.
Deepak stakes 350 Mu on No and says: “CSK is finished. Dhoni is old. No debate.”
Anand stakes 150 Mu on No with the comment: “Pressure is real.”
Meera stakes 200 Mu on Yes. No comment. The group immediately begins speculating about whether she has inside information.
Karthik watches the probability move to 61% Yes, waits until 7:15pm, then stakes 300 Mu on Yes and says: “I’ve been thinking this for a while actually.”
Vikram, who set the whole thing up, stakes 100 Mu on No because he genuinely does not know and wants to participate more than he wants to win.
Chennai win by 6 runs. The results card arrives in everyone’s NeoMail the next morning. It shows the group probability at close — 63% Yes — the outcome — Yes — and the updated leaderboard. Prashant has climbed to first. Meera is second. Karthik, despite being right, has moved to fourth — because the scoring rewards early commitment to a correct position, not last-minute bandwagon-jumping. This single detail produces twenty minutes of the most animated conversation the group has had since the 2019 World Cup semi-final.
“This is rigged. I was right.” “You staked eight minutes before close.” “So? I was still right.” “The market was at 61% when you staked. You agreed with 61% of the group. That’s not a prediction, Karthik. That’s a headcount.”
This argument — which in previous years would have been impossible to have because there was no data to have it with — goes on for most of the following day and establishes a vocabulary that will persist for the entire season. Being early becomes honourable. Being late becomes known, formally, as The Karthik Move.
**
Three weeks in, something has changed in Hostel C Legends. Not the cricket discussion — that is exactly as it always was, which is to say loud, confident, and frequently wrong. What has changed is the scaffolding around it. After eleven markets, the group leaderboard looks like this:
- Meera — 847 points, 8/11 correct, top quartile on calibration
- Prashant — 791 points, 7/11 correct, strong early commitment
- Anand — 634 points, 5/11 correct, consistent early staking
- Deepak — 589 points, 5/11 correct, high stakes hurting him on losses
- Vikram — 423 points, 4/11 correct
- Karthik — 318 points, 5/11 correct, chronically late pattern
Three things have happened that the group did not predict.
The first: Meera, who has spent twelve years deflecting the group’s cricket analysis with mild amusement, is now first and cannot be argued with. The group has responded to this the way groups respond to uncomfortable data — by theorising about why the leaderboard is wrong. Prashant has suggested her edge is timing rather than cricket knowledge. Deepak has suggested she is googling things. Anand has said nothing, which is his version of agreement. Meera has said: “I just trust the batters who make it look easy.” Nobody knows what to do with this.
The second: Karthik’s late-staking pattern has been named and remembered. He is aware of this. He has started staking earlier. His calibration is not improving, but his commitment is, and the group finds this genuinely encouraging. Progress is progress.
The third is the one nobody predicted. Anand — the group’s permanent pessimist, the man who has predicted India to lose under pressure for eight years — is third on the leaderboard. His thesis, applied consistently and staked early, turns out to be calibrated at approximately the rate that India actually does struggle under pressure. The group is now in the uncomfortable position of having data that partially vindicates Anand’s worldview, and this is producing a level of collective cognitive dissonance that may take the rest of the season to work through.
**
By June, Hostel C Legends has run thirty-one markets. Nobody has been prompted to create any of them since Week 3. Vikram set up a Friday reminder that a new match market is available, and the group now creates its own markets without being asked — including, in the ninth week, a market on whether Deepak will visit India before the year ends. (He will not. He staked Mu on Yes. The group found this poetic.)
The NeoMail each member receives on match mornings carries a WePredict card — Your group market closes tonight, 61% say Yes, 9 members have staked — and this card has become, for several members, the primary reason they open the NeoMail at all. The inbox has acquired gravity it did not previously have. It is no longer a place you go reluctantly to process things. It is a place you go because something is happening there that involves people you care about.
The social texture of the group has shifted in a way that is hard to describe precisely but easy to recognise. The arguments still happen. The confidence is unchanged. What is different is that the arguments now happen in reference to a record — a real, unambiguous, publicly visible record of who has been right about what over thirty-one resolved questions. The punditry has not diminished. It has been grounded.
And Meera, who has led the leaderboard for eleven consecutive weeks, receives a message from Deepak on a Thursday evening that says only: “I accept it.”
This is, in its small way, a resolution that seventeen years of argument could not produce.
6
WhatsApp Story 2: The Family Group Discovers Mu
The Sharma family group has twenty-five members and a name that nobody remembers choosing: Sharma Parivar ❤️🙏. It was created for a cousin’s wedding in 2018 and never disbanded because nobody wanted to be the person who disbanded it. It is active in the way all large family groups are active — in bursts, around events, with a long undercurrent of unread messages that everyone has muted but nobody has left.
During IPL season, the group comes alive. It comes alive the way a chai shop comes alive before a big match — with opinions that arrived fully formed, delivered with certainty, attributed to no particular evidence. Riya’s father-in-law, Uncle Sameer, is the group’s most prolific predictor. He has strong views about every team, player, and decision, delivered in capital letters with a cheerful disregard for whether his previous predictions turned out to be correct. He is, in the precise sense of the term, unaccountable. There is no record. There never has been.

One Friday afternoon, right before an RCB vs CSK match, Riya — who is twenty-seven, works at a startup, and has been using NeoMails for three months — drops a forecast card into the group.
She does not introduce it. She does not explain it. She simply drops it into the chat the way you drop any link, without ceremony, and waits to see what happens.
WePredict Private — Sharma Parivar
Will RCB beat CSK tonight?
Group probability: 57% Yes
Closes 7:25pm
Top forecasters this week: 1) Riya 2) Uncle Sameer 3) Neha
[Join — 1 tap, no app needed]
The first responses arrive within ninety seconds:
“What is this?” “Riya beta, are we gambling now?” “Is this legal?”
And then, from Uncle Sameer, in capital letters: “I WILL JOIN. RCB WILL WIN. TELL EVERYONE.”
This is, it turns out, the real distribution mechanic. Not a notification. Not a product feature. Status dynamics. Once Uncle Sameer joins, three cousins who would not otherwise have clicked join immediately — partly to play, mostly to have grounds to argue with him later.
The link opens a lightweight page. No app to install. No form to fill. Two buttons: Yes and No. Under them, fixed stake sizes: 10 Mu, 50 Mu, 200 Mu. No custom amounts. The product is, deliberately, anti-clever. It is designed to be used in thirty seconds by someone who has never heard of a prediction market and does not want to learn.
Two family members do not have enough Mu to stake. This produces the moment Riya has been waiting for:
“How do I get Mu?”
“Open the NeoMail with the quiz in it. The one with the subject line that shows your balance. Takes two minutes.”
“Oh those. I’ve been ignoring those.”
“Don’t. That’s where you earn.”
Three family members who have been deleting NeoMails for weeks open them that evening and engage for the first time. They earn enough Mu to stake. They join the market. The loop, which was invisible to them until this moment, suddenly makes sense: the inbox is where you earn the currency that lets you play.
**
By 7:20pm, the group probability has moved from 57% to 63%. The banter has reached a pitch that the group has not seen since Kohli’s 89 not out against West Indies in the 2016 World T20 semi-final.
“Stop inflating it. You’ll jinx it.” “You’re just scared you’ll be wrong again.” “I’m not scared. I’m calibrated.”
That last word — calibrated — is new in this context. It does not belong to the usual vocabulary of family cricket arguments. It has arrived because the leaderboard has created a new social identity: the person whose predictions have a track record. Uncle Sameer, who has been the group’s loudest voice for eight years, is second on the leaderboard. Riya is first. This fact is visible to all twenty-five members.
Uncle Sameer handles this with more grace than anyone expected. “Next week,” he says. “I am warming up.”
RCB win. The results card arrives in everyone’s NeoMail the next morning — a clean visual showing the group probability at close, the actual outcome, the updated leaderboard, and a single line that will carry more weight than any full sentence could: Next market drops tomorrow at 10am.
The group explodes. Not because anyone won money. Nobody won anything except Mu, and most of them still have only a vague understanding of what Mu is. They explode because the card has done something that twenty-five people in a family group have never experienced: it has created a public record inside a private space. Uncle Sameer’s ranking is now social reality. Riya’s first place is documented. Neha, who has been quiet in the group for months, is third and has started typing again.
Identities are emerging. And identities, once they exist, are sticky.
**
The following week, Riya notices that only twelve of twenty-five family members participated. She sends a message — not from the platform, just a regular WhatsApp message — that says: “If you don’t have enough Mu, open the NeoMail today. There’s a quiz. Five minutes, you’re in for tonight’s market.”
Four more family members open their NeoMails. Three of them have been subscribers for months but have never clicked anything. The prediction market is the reason they finally do.
This is how an ecosystem grows without advertising. Not through a campaign. Through a cousin saying: “You’re missing out, and it only takes five minutes to fix that.”
By the fifth week, twenty of twenty-five Sharma family members are participating in at least one market per week. Uncle Sameer has climbed to first place. He has announced this in the group seventeen times. The group has pointed out each time that he announced it while it was still happening, which the scoring system does not reward. He remains unmoved. First place is first place.
The group is not what it was. It is louder, more specific, more willing to commit to positions before the outcome is known. It has a leaderboard. It has a vocabulary. It has receipts. For a family that has been arguing about cricket since before some of its younger members were born, this is not a small thing.
“We were already arguing every match,” Riya’s mother says, on a Sunday evening in Week 6, after a market she staked correctly and Uncle Sameer staked wrong. “Now we have receipts.”
7
Slack Story 1: The Sprint the Market Knew Would Slip
GrowthStack is a mid-sized SaaS company in Bengaluru with about 340 employees. Its product is a B2B analytics platform for retail chains. Its engineering organisation is split into six squads. The squad relevant to this story is called Polaris — seven engineers, a product manager named Shreya, and a squad lead named Rohan. Standard configuration. Standard pressures.
In January 2026, GrowthStack’s head of engineering, Arjun, decides to pilot WePredict Private in Slack. He has read about internal prediction markets, he has looked at what Google and HP did with them, and he believes the company has a specific problem: sprint commitments are consistently overconfident, and management’s view of delivery timelines is consistently more optimistic than what the engineering team believes in private. He has tried asking engineers directly about this. The answers are carefully hedged. Nobody wants to be the person who tells the VP of Product that the quarter’s roadmap is aspirational rather than achievable.
He sets up WePredict Private in a single Slack channel — #polaris-forecasts — with the intention of running it for one quarter before deciding whether to expand. He explains the mechanics to the team in a fifteen-minute session: anonymous staking, fixed Mu amounts to reduce signalling games, explicit resolution criteria tied to Jira, a calibration dashboard that will show accuracy over time. He emphasises one thing above all others: the point is not to find out who was pessimistic. The point is to surface what the team collectively knows before it becomes a problem.
The team listens carefully. They are engineers. They appreciate precision. Several of them are privately sceptical. None of them say so.
**
The first market goes up on a Monday morning in the second week of January.
WePredict Private — Polaris
Will Polaris complete the Retailer Dashboard v2.1 feature by end of Sprint 23 — Friday 31 January?
Closes Wednesday 5pm
Resolution: automatic — Jira “Released” status + deployment timestamp
Anonymity: enabled
[Join]
By Tuesday morning, nine people have staked. The probability has settled at 34% Yes.
This is significant. The official sprint plan says this feature will be complete by Friday. The commitment communicated to the VP of Product in the Monday stand-up says yes. The Jira board says in progress. The team’s public posture is confident. The market says 34%.

Rohan, the squad lead, sees this and feels something that does not have a clean name but is recognisable to anyone who has ever been responsible for delivering something on time while privately suspecting it will not arrive. It is the discomfort of someone who knows a thing is true but has been communicating a more optimistic version of it upwards, not out of dishonesty but out of the reasonable hope that effort and goodwill will close the gap.
The market has said, in aggregate and anonymously, what the team has been thinking but not saying. Nobody said it. The crowd said it. And somehow that makes it easier to act on.
Rohan sends a message in #polaris-forecasts: the team is behind on two blocking items, and could the resolution criteria be amended to cover a working subset of the feature rather than the full scope? Shreya agrees within an hour. The criteria are updated. The market is amended.
By Wednesday 5pm close, the probability has risen to 61% Yes on the narrowed scope.
The feature ships on Thursday — one day early on the narrowed criteria, and a conversation about the remaining scope moved cleanly into the next sprint planning session. The VP of Product is told that the team delivered ahead of schedule. The broader scope question is surfaced as a planning discussion rather than a missed commitment.
Nobody in this story has been dishonest. But without the market, the most likely outcome was a Friday miss, an explanation, and the specific kind of post-mortem that assigns blame to everyone and changes nothing. With the market, the miss was anticipated on Tuesday, the scope was renegotiated on Tuesday, and the team delivered on Thursday. The difference is not in capability or effort. It is in the speed at which private knowledge became collective information that someone could act on.
**
By the end of the first quarter, #polaris-forecasts has run fourteen markets across three sprints. The calibration dashboard has produced several things that Arjun finds genuinely, specifically useful — not in the vague sense that dashboards are often called useful, but in the sense of things that change decisions.
The first: Polaris systematically overestimates sprint completion for features that involve the data layer. The market probability for data-layer-dependent features closes below 50% three times out of four, and the feature has slipped three times out of four. This is not news to anyone who has been paying attention — the data layer’s unpredictability has been mentioned in retrospectives for months. But it has never appeared as a number before. It has existed as a vague collective concern that surfaces and evaporates. Now it is a number: 27% average completion probability for data-layer-dependent features at market close. The team uses this in the next sprint planning to explicitly flag any feature with a data-layer dependency. The VP of Product asks why. Shreya shows him the calibration data. He does not argue with a number.
The second finding is more personal. Of the eleven people who have staked in at least ten markets, the three most accurate forecasters — ranked by calibration score — are Nisha, a data engineer formally assigned to a different squad but spending most of her time on Polaris work; Rohan; and a junior engineer named Siddharth who joined GrowthStack six months ago. The three least accurate are the two most senior engineers on the squad, and — somewhat awkwardly — Arjun himself, who has staked in every market from the beginning.
Arjun looks at this information for a long time. It is visible to everyone in the channel.
The senior engineers’ poor calibration follows a specific pattern: they consistently overestimate how quickly refactoring work will complete. They are optimistic about their own estimates. This is not a character flaw; it is a systematic bias that has now been made legible. In the next sprint planning, Arjun asks both senior engineers to add a 20% buffer to any refactoring estimate. They do not push back. The data is the data, and arguing with a calibration score in front of the whole team is not a position anyone wants to occupy.
Siddharth’s high calibration score produces a different kind of movement. He is six months in and has been hesitant to express strong views in planning meetings. The Predictor Score is not a formal credential — it does not appear on his employment record or his performance review. But it is a real credential within the team, visible to everyone in the channel, and it is difficult to ignore. Rohan begins copying him into planning discussions that would previously not have included a junior engineer. His estimates begin carrying weight in conversations that were previously shaped entirely by the senior engineers’ views. This is not a promotion. It is something smaller and in some ways more significant: the quiet expansion of whose knowledge gets counted.
**
The market that matters most to this story runs in the third week of March.
GrowthStack is bidding on a large enterprise contract with a regional retail chain. The bid includes a commitment to deliver a custom integration feature by the end of April. The VP of Sales wants this commitment in the proposal. The VP of Product is supportive. Arjun is uncertain, in the specific way that heads of engineering are uncertain when they have calibration data and the people above them do not.
He creates a private channel with seven people — Rohan, Shreya, the three most calibrated forecasters from the dashboard, and one senior engineer — and runs a single market: Can Polaris deliver the RetailChain integration feature to production-ready status by April 30?
He gives it 24 hours. Seven people stake. The market closes at 29% Yes.
There is no ambiguity in this number. The seven people who staked are the seven people who know the codebase, the team’s current capacity, and the feature’s complexity most precisely. They have been forecasting together for a quarter. Their calibration scores are real and documented. The market says 29%.
Arjun takes this number to the VP of Sales. He explains how the market works and what the calibration data behind it means. He suggests that the proposal commit to May 31 instead of April 30.
The VP of Sales pushes back. “This is just Arjun being cautious. We’ve had this conversation before.”
Arjun says: “It’s not me being cautious. It’s seven people being asked to stake something anonymously, with three months of calibration data behind them, and 71% of them saying April 30 is not realistic.”
The proposal goes out with a May 31 delivery commitment. GrowthStack wins the contract. The feature ships on May 19 — twelve days ahead of the committed date, three weeks after the original impossible ask.
The VP of Sales does not say anything to Arjun directly. But she is the one who forwards his internal note about the Polaris experiment to the CEO, with a single line of her own above it: “Worth reading.”
8
Slack Story 2: The Market That Said What Nobody Would
The Slack channel is called #release-ops, and it is the kind of channel that exists in every product company — useful, necessary, and quietly dysfunctional in a way that everyone understands and nobody fixes.
The dysfunction is not dramatic. It is mundane. It is the drama of optimism. Every Monday, the stand-up notes land in #release-ops: features in progress, timelines green, confidence expressed. By Wednesday, the features are still in progress. By Thursday, private messages begin circulating — between engineers who trust each other, between PMs who have done this before — in which the actual status of things is discussed honestly and usefully. By Friday, something ships, or something does not, and either way the public account of why is shaped more by what is comfortable to say than by what actually happened.
This is not dishonesty. It is a rational response to the social environment of status meetings. People communicate the version of the truth that preserves relationships, avoids blame, and keeps the energy positive. The problem is that this version of the truth, communicated upwards, reaches the people who make resourcing and prioritisation decisions too late to change outcomes. The surprise slip — the feature that was green on Monday and missed on Friday — is not a technical failure. It is an information failure. The team knew. The information did not travel.
Priya, the Head of Product, has been thinking about this for a year. She does not think the team is being dishonest. She thinks the environment makes honesty expensive in a way that a different mechanism might change. She sets up WePredict Private in #release-ops on a Tuesday afternoon in February with a short message to the team that says: Trying something. No grades, no blame. Just signal.
The first market goes up the following Monday morning, automated, via a bot that Priya has configured to run every week without manual input:
WePredict Private — Release Ops
Will Sprint 14 ship by Friday 6pm?
Resolution: Jira “Released” + deployment confirmed Closes Thursday 5pm
Stake: 20 Mu fixed — anonymity enabled
[Join]
By Monday afternoon, the market has opened at 70% Yes. This is roughly the mood of the room, which is roughly the mood of every Monday.
By Wednesday morning, it is at 58% Yes.

Nothing has been said publicly. The stand-up notes for Wednesday still read: features in progress, timeline on track. But the market has moved twelve points in two days, and that movement represents the private accumulation of signals — a dependency that has not resolved, a review cycle that is taking longer than expected, an estimate that was always slightly optimistic — none of which would survive a status meeting on their own but which together produce a probability that a crowd of informed people has honestly expressed.
Priya does not treat 58% as a verdict. She treats it as a signal to ask better questions. Not “is there a problem?” — which creates defensive responses — but: “What would need to happen for this to land above 70%? Which dependency is driving the uncertainty? If we do slip, what is the smallest scope adjustment that preserves the value?”
The conversation that follows is different from a normal status discussion. Instead of debating opinions — the engineer who believes it will ship, the PM who is less sure, the designer who knows a review is late — the team debates conditions. The question is not whether someone is right or wrong. The question is what the market is reflecting and whether it can be changed. This is a calmer and more productive conversation than the one that usually happens in status meetings, and the reason it is calmer is that nobody’s personal credibility is on the line. The market said it. Everyone is just responding to the market.
By Thursday close, the probability is 43% Yes.
Priya does not need courage at this point. The market has provided it. She can say: “The crowd is telling us we are unlikely to ship as scoped. Let’s act accordingly” — and what follows is a scoping conversation rather than a blame conversation. One non-critical feature is moved to the following sprint. A QA cycle is brought forward by a day. An external dependency is escalated.
On Friday at 4:30pm, they ship.
The celebration in #release-ops is real, and it is also slightly unusual, because the team knows that what they are celebrating is not just a delivery. They are celebrating a system that told them the truth early enough for them to change the outcome. The ship happened in part because of the slip that the market predicted and the team prevented. Both things are true simultaneously.
**
Over six weeks, #release-ops runs six markets. The calibration picture that emerges is specific enough to be actionable.
The market is systematically too optimistic on Monday mornings. By Wednesday, it corrects. The gap between Monday sentiment and Wednesday sentiment is the gap between how the team feels at the start of a sprint and what they collectively know by the middle of it. Priya uses this to change the timing of her escalation conversations: she stops asking about status on Mondays, when the answer is always optimistic, and starts asking on Wednesdays, when the market has had time to incorporate the week’s actual signals.
One sub-team — a pair of engineers who joined the company eight months ago and have been largely quiet in planning meetings — is consistently better calibrated than the rest. Their market predictions are accurate at a rate that is notably higher than the team average. Priya does not share this observation in a meeting. She starts copying them into sprint planning discussions. Their estimates begin influencing scope decisions in ways that would not have been possible six months ago, when their tenure and seniority would have made their views easy to overlook. The calibration data has given them a credential that their job title had not yet provided.
The market that Priya considers most important runs in Week 5. A major feature — the biggest deliverable of the quarter — opens on Monday at 65% Yes. It ends Thursday at 39% Yes. The feature does not ship that Friday.
This is, by one measure, a failure. By another measure, it is the product working exactly as intended. The market predicted the slip on Monday and confirmed it by Thursday. The team adjusted scope early enough to deliver a meaningful subset rather than nothing. The miss was not a surprise to anyone who had been watching the channel. It was a managed, anticipated, documented event — documented not in a post-mortem but in a probability curve that moved from optimism to realism over four days.
Priya writes a short note in the channel after the week ends. It says: “The market told us on Monday. We listened by Wednesday. We shipped something real on Friday. That’s the whole point.”
Twenty-three people react to this message with a thumbs-up. One person — the junior engineer who was among the most accurate forecasters in the channel — reacts with a small, specific emoji that Priya will think about for a while afterwards: not a thumbs-up, not a celebration, but a simple green check mark.
It means: yes, that is what happened. And it can happen again.
**
What Four Stories Prove That Two Cannot
Read across all four, and a pattern emerges that no single story contains on its own.
The Hostel C Legends story is about what happens when a long-standing mythology — of who knows cricket, whose predictions count — meets an impartial record. The mythology does not disappear. It gets grounded. The arguments continue; they just happen in reference to something real now.
The Sharma Family story is about discovery and the ecosystem loop. The prediction market is the reason people open the NeoMail. The NeoMail is the reason they have Mu to stake. The stake is the reason they care about the outcome. None of these things works without the others, and none of them is visible until a cousin drops a card into a family chat on a Friday afternoon.
The GrowthStack story is about the accumulation of calibration intelligence and what it makes possible — not in a single dramatic moment, but over a quarter, through a series of small revelations that reshape how decisions are made and whose knowledge gets counted.
The #release-ops story is about the thing that prediction markets do that no other management tool can: they give a crowd a mechanism to say what no individual will say, early enough to change the outcome rather than explain it.
Four different contexts. Four different emotional registers. The same infrastructure, the same currency, the same portable identity layer underneath.
What they prove together is the claim the earlier parts of this series made theoretically: social consequence is real consequence. Closed groups do not need cash to create stakes. They need a scoreboard, a record, and the knowledge that the people who will see the result are the people who matter to them.
WePredict Private is the scoreboard.
The rest — the arguments, the revelations, the junior engineer’s green check mark, Uncle Sameer announcing his first place ranking seventeen times — is what happens when groups finally have receipts.