•1 min read•from Machine Learning
Formalizing statistical learning theory in Lean 4 [R]
![Formalizing statistical learning theory in Lean 4 [R]](/_next/image?url=https%3A%2F%2Fexternal-preview.redd.it%2FvcCPMFQJ_Ba1RGnmGbl8xPO9fb1vZ01Ht3Hg4xF7i0Q.png%3Fwidth%3D640%26crop%3Dsmart%26auto%3Dwebp%26s%3D66c23bc95d8b7a4cf205d906d8b6fed9fda955ab&w=3840&q=75)
| I’ve been working on a Lean 4 project focused on formalizing parts of statistical learning theory: Current results include:
The main idea is to build a readable and pedagogically structured “theorem ladder” for ML theory rather than just isolated declarations. I’m trying to keep:
Compared to some existing Lean SLT efforts that focus more heavily on empirical-process infrastructure and abstract probability machinery, this project is currently more focused on explicit finite-sample PAC/Rademacher/stability routes and readable end-to-end theorem chains. I’d especially appreciate feedback on:
Thank you, [link] [comments] |
Want to read more?
Check out the full article on the original site
Tagged with
#rows.com
#machine learning in spreadsheet applications
#natural language processing for spreadsheets
#generative AI for data analysis
#row zero
#Excel alternatives for data analysis
#financial modeling with spreadsheets
#no-code spreadsheet solutions
#spreadsheet API integration
#statistical learning theory
#Lean 4
#finite-class ERM bounds
#Rademacher symmetrization
#high-probability Rademacher bounds
#Sauer-Shelah VC-dimension bridge
#finite scalar contraction
#linear predictor bounds
#finite PAC-Bayes bounds
#algorithmic stability
#theorem ladder