GenAi, Deep Learning and Computer Vision
3.3K subscribers
31 photos
5 videos
5 files
168 links
Deep LearningπŸ’‘,
Computer Vision πŸ“½ &
#Ai 🧠

Get #free_books,
#Online_courses,
#Research_papers,
#Codes, and #Projects,
Tricks and Hacks, coding, training Stuff

Suggestion @AIindian
Download Telegram
GenAi, Deep Learning and Computer Vision
Photo
Google engineers offered 28 actionable tests for #machinelearning systems. πŸ‘‡

Introducing πŸ‘‰ The ML Test Score: A Rubric for ML Production Readiness and Technical Debt Reduction (2017). πŸ‘ˆ

If #ml #training is like compilation, then ML testing shall be applied to both #data and code.

7 model tests

1⃣ πŸ‘‰ Review model specs and version-control it. It makes training auditable and improve reproducibility.

2⃣ πŸ‘‰ Ensure model loss is correlated with user engagement.

3⃣ πŸ‘‰ Tune all hyperparameters. Grid search, Bayesian method whatever you use, tune all of them.

4⃣ πŸ‘‰ Measure the impact of model staleness. The age-versus-quality curve shows what amount of staleness is tolerable.

5⃣ πŸ‘‰ Test against a simpler model regularly to confirm the benefit more sophisticated techniques.

6⃣ πŸ‘‰ Check the model quality is good across different data segment, e.g. user countries, movie genre etc.

7⃣ πŸ‘‰ Test model inclusion by checking against the protected dimensions or enrich under-represented categories.

7 data tests

1⃣ πŸ‘‰ Capture feature expectations in schema using statistics from data + domain knowledge + expectations.

2⃣ πŸ‘‰ Use beneficial features only, e.g. training a set of models each with one feature removed.

3⃣ πŸ‘‰ Avoid costly features. Cost includes running time, RAM as well as upstream work and instability.

4⃣ πŸ‘‰ Adhere to feature requirements. If certain features can’t be used, enforce it programmatically.

5⃣ πŸ‘‰ Set privacy controls. Budget enough time for new feature that depends on sensitive data.

6⃣ πŸ‘‰ Add new features quickly. If conflicting with 5⃣ , privacy goes first.

7⃣ πŸ‘‰ Test code for all input features. Bugs do exist in feature creation code.

See 7 Infrastructure & 7 monitoring tests in paper. πŸ‘‡

They interviewed 36 teams across Google and found

πŸ‘‰ Using a checklist helps avoid mistakes (like a surgeon would do).

πŸ‘‰ Data dependencies leads to outsourcing responsibility. Other teams’ validation may not validate your use case.

πŸ‘‰ A good framework promotes integration test which is not well adopted.

πŸ‘‰ Assess the assessment to better assess your system.
https://research.google.com/pubs/archive/aad9f93b86b7addfea4c419b9100c6cdd26cacea.pdf
πŸ‘3