sam_ezeh 8 days ago

The repository is extremely disappointing

>This repository provides a framework for specifying and proving properties—such as robustness, fairness, and interpretability—of machine learning models using Lean 4

FV of AI systems is hard and this repo massively oversells itself. The Lean code itself is weird (AI generated?) and the "fairness" example proves that if a (linear) classifier always returns "yes" then the percentage of people classified as "yes" is the same across all demographics which isn't meaningful since this percentage is 100% for every possible group of people.

I was watching a video [1] the other day about the challenges FV faces for AI if it's something you're be interested in. I'm not really familiar with FV for neural networks and not sure what progress looks like in the field but I know ANSR [2] and TIAMAT [3] are doing related things

[1]: https://www.youtube.com/watch?v=bs5snugP1VA

[2]: https://www.darpa.mil/research/programs/assured-neuro-symbol...

[3]: https://www.darpa.mil/research/programs/transfer-from-imprec...

2
almostgotcaught 5 days ago

> this repo massively oversells itself

that's literally every single paper in academic CS (except theory). every claim is an exaggeration or lie by omission ("this only works on a small subset of ..."). every title is clickbait. every impl is worse than junior-level code. you become so jaded to it that you start to evaluate papers/projects based a premonition of what it could be rather than what it is.

mentalgear 3 days ago

At least the author[0] seems to have some clout behind him. However, given that his code doesn't even compile and the premise seems massively over-stated, I wonder how his credentials (Stanford, etc) can even be genuine.

[0] https://mateopetel.xyz/