Posts by Valeriy M., PhD, MBA, CQF
Modern gradient-boosted trees arrive pre-calibrated. Measuring first is not optional — it is the whole job.
Which model did you calibrate last that you now suspect you broke?
Full write-up in the comments.
MachineLearning #Calibration #ConformalPrediction
The model is calibrated. Any rewrite is destructive.
If |Z| > 3.0, use Venn-Abers — distribution-free, mathematically guaranteed. The 1.96–3.0 band is where you actually test both.
The standard advice comes from an era of SVMs and shallow nets.
Distort a calibrated distribution through a parametric function it never needed, you lose sharpness faster than you gain anything.
The Spiegelhalter decision rule — three steps:
Compute the Spiegelhalter Z on a held-out set before any calibration.
If |Z| < 1.96, do nothing.
Platt fit a two-parameter sigmoid anyway, compressed the tails, pulled confident predictions toward the center, and erased meaningful variation the model had already earned.
Log-loss does not forgive that. It measures calibration and sharpness together.
Platt hurt every one of them.
The reason is mechanical. CatBoost’s Spiegelhalter |Z| was 1.88 — below the 1.96 significance threshold. The model was not miscalibrated. There was nothing to fix.
The damage on well-behaved models was not subtle.
— CatBoost: log-loss worsened on 93% of folds. Platt added 5.3%.
— TabICL: worse on 91% of folds. 6.0% penalty.
— EBM: worse on 90%. 4.4% penalty.
— TabPFN: worse on 87%. 5.0% penalty.
Four of the five best classifiers in the study.
Stop using Platt scaling by default. It’s the bug, not the fix.
If you’re serious about shipping calibrated models, that one-line reflex is quietly making your best classifier worse.
We ran Platt scaling across 21 classifiers, 30 binary datasets, 150 cross-validation folds (arXiv 2601.19944).
Knowing is cheap. Performing is what counts.
Video based on my English translation of Kiselev’s iconic Arithmetic:
www.youtube.com/watc...
My new article explains "The Great Translation: How 1950s China Cloned the Soviet Mathematical Machine"
The real foundation of China’s 🇨🇳 STEM and AI rise was not the Western liberal arts model. It was the opposite: the systematic dismantling of that model and the large-scale import of Soviet curricula, textbooks, and teaching methods.
The faithful edition of Markov is meant to read like a classical mathematics book — and part of what classical mathematics looks like is Latin and Greek sharing a page without pretending they belong to the same family.
If you wanted a century of mathematical writing to stay typographically continuous with the tradition that came before it, Knuth got it exactly right.
I'm leaving the font alone in my translation.
You've been reading it your whole career and probably never noticed — because somewhere along the way, it became the look of math.
Was it a mistake?
If you wanted every symbol to look cut from one block of wood, yes. Later fonts harmonise Greek and Latin better.
Knuth drew them from different typographic traditions. Neoclassical Latin italic for a. Handwritten Greek italic for α. Different centuries, different hands.
Nearly every mathematical paper typeset in LaTeX carries this mismatch. Textbooks. Monographs. Fields Medal lectures.
In Computer Modern — the math font nearly every LaTeX paper on earth defaults to, designed by Donald Knuth in the late 1970s — italic a and italic α don't harmonize. The Greek alpha is visibly taller and meaningfully wider than the Latin a sitting right next to it.
Knuth is a genius. He still got some things wrong.
A reader of my Markov translation sent me a photo of a fraction on page 20:
(a − α) / (a + b − α − β)
"The 'a' and the 'α' look like different sizes. Feels off."
He was right.
And that is why no one should impressed by the cult of Kaggle medals.
#AI #MachineLearning #DataScience #DeepLearning #Kaggle #DeepMind #Mathematics #ArtificialIntelligence
But it does not produce DeepMind.
Kaggle trained people to win competitions.
Rigorous mathematics trained people to move civilization forward.
Major scientific breakthroughs.
Nobel Prize-level impact.
That is what happens when you optimize for first principles instead of public scoreboards.
The uncomfortable truth is simple:
In AI, rigorous mathematics beats contrived hacks.
Every time.
Kaggle can produce clever competitors.
And it is definitely not the same thing as changing the world.
Now compare that with DeepMind.
Founded by elite mathematicians, neuroscientists, and top-tier researchers.
Built on rigorous theory, not leaderboard theatre.
And what did it produce?
AlphaGo.
AlphaFold.
Kaggle rewards people for squeezing the last few basis points out of contrived benchmark problems with endless feature hacks, ensembling tricks, and competition-specific gymnastics.
That is not the same thing as building enduring technology.
That is not the same thing as advancing science.
For years, the industry was told that Kaggle Grandmasters were the "ultimate proof of machine learning excellence."
They were not.
Kaggle was founded in 2010.
More than 16 years later, what exactly is its legacy?
❌ A generation of leaderboard chasers.
❌ Very little foundational AI.
❌ Very few serious companies.
❌ No scientific breakthroughs.
Proof that great early education + relentless curiosity can take you from action hero to AI innovator.
Who else credits their weird or rigorous childhood schooling for skills they use today? Drop your stories 👇
#MillaJovovich #MemPalace #AI
From Soviet kindergarten math foundations → Hollywood superstar → building cutting-edge AI architecture.
The USSR preschool program didn’t mess around with early STEM foundations… and apparently neither does Milla.
After dominating Hollywood as the Resident Evil queen and Fifth Element warrior, Milla didn’t stop learning.
She just architected MemPalace — a free, open-source AI memory system on GitHub that’s currently posting some of the highest benchmark scores ever seen for memory/retrieval.
In the USSR, little Milla Jovovich was getting a serious head start in mathematics — counting, patterns, shapes, basic logic, and spatial thinking that the Soviet preschool system was famous for teaching systematically from age 3–6.
Fast-forward decades:
Most kids leave kindergarten knowing colors and how to share toys.