This is decent for what it is. Some of the problems are pretty open ended which has pros and cons, but that is very different from leetcode, which has very specific data and test cases.
For example, implement linear regression but the example solution uses a random number generator without a fixed seed. It’s fine, reproducibility isn’t the point, but leetcode problems are more structured.
In leetcode they usually don’t tell you exactly what data structure you must use, only that it must pass certain test cases. By analogy this might not tell you which architecture to use but require that it passes certain eval metrics.
Most ML problems in real life don’t constrain you to use linear regression or a CNN either. But there will be some metric you need to optimize.
What would take this repo to the next level is to have a reproducible data generation function for each exercise as well as a reasonable metric which must be passed. I don’t see anything that requires my classification auc to be over 0.5 which would be a basic criteria of bug-free code.
One time my teacher used a computer to make a math test for me, but then told me I couldn't use my computer during the exam. I dropped out of school immediately.
Accounts with high karma have a "flag" link in the header of comments and submissions for indicating that it may not be the right kind of content for the site. Human moderators actively monitor the site, so your comment about this user's behavior will be seen by one of them soon.
For example, implement linear regression but the example solution uses a random number generator without a fixed seed. It’s fine, reproducibility isn’t the point, but leetcode problems are more structured.
In leetcode they usually don’t tell you exactly what data structure you must use, only that it must pass certain test cases. By analogy this might not tell you which architecture to use but require that it passes certain eval metrics.
What would take this repo to the next level is to have a reproducible data generation function for each exercise as well as a reasonable metric which must be passed. I don’t see anything that requires my classification auc to be over 0.5 which would be a basic criteria of bug-free code.
I was reverse engineering the ML interview pipeline for myself and that's how I stumbled upon all this.
I think the data aspect does make sense tho. I might add that as the next thing to do
I mean...this entire project appears to be mostly GPT-generated?
[0]: https://news.ycombinator.com/submitted?id=NetRunnerSu