在线时间:8:00-16:00
迪恩网络APP
随时随地掌握行业动态
扫描二维码
关注迪恩网络微信公众号
开源软件名称(OpenSource Name):amusi/machine-learning-yearning-cn开源软件地址(OpenSource Url):https://github.com/amusi/machine-learning-yearning-cn开源编程语言(OpenSource Language):开源软件介绍(OpenSource Introduction):machine-learning-yearning-cn吴恩达《Machine Learning Yearning》的英文版完结:[第1~第58章](Machine Learning Yearning 1-58(by Andrew NG).pdf) 官网:
原作者:Andrew NG 申明:本文旨在传播知识,并无商业行为之意 TODO
目录Chapter 1. Why Machine Learning Strategy Chapter 2. How to use this book to help your team Chapter 3. Prerequisites and Notation Chapter 4. Scale drives machine learning progress Chapter 5. Your development and test sets Chapter 6. Your dev and test sets should come from the same distribution Chapter 7. How large do the dev/test sets need to be? Chapter 8. Establish a single-number evaluation metric for your team to optimize Chapter 9. Optimizingandsatisficingmetrics Chapter 10. Having a dev set and metric speeds up iterations Chapter 11. When to change dev/test sets and metrics Chapter 12. Takeaways: Setting up development and test sets Chapter 13. Build your first system quickly, then iterate Chapter 14. Error analysis: Look at dev set examples to evaluate ideas Chapter 15. Evaluate multiple ideas in parallel during error analysis Chapter 16. Cleaning up mislabeled dev and test set examples Chapter 17. If you have a large dev set, split it into two subsets, only one of which you look at Chapter 18. How big should the Eyeball and Blackbox dev sets be? Chapter 19. Takeaways: Basic error analysis Chapter 20. Bias and Variance: The two big sources of error Chapter 21. Examples of Bias and Variance Chapter 22. Comparing to the optimal error rate Chapter 23. Addressing Bias and Variance Chapter 24. Bias vs. Variance tradeoff Chapter 25. Techniques for reducing avoidable bias Chapter 26. Techniques for reducing Variance Chapter 27. Error analysis on the training set Chapter 28. Diagnosing bias and variance: Learning curves Chapter 29. Plotting training error Chapter 30. Interpreting learning curves: High bias Chapter 31. Interpreting learning curves: Other cases Chapter 32. Plotting learning curves Chapter 33. Why we compare to human-level performance Chapter 34. How to define human-level performance Chapter 35. Surpassing human-level performance Chapter 36. Why train and test on different distributions Chapter 37. Whether to use all your data Chapter 38. Whether to include inconsistent data Chapter 39. Weighting data Chapter 40. Generalizing from the training set to the dev set Chapete 41. Identifying Bias, Variance, and Data Mismatch Errors Chapter 42. Addressing data mismatch Chapter 43. Artificial data synthesis Chapter 44. The Optimization Verification test Chapter 45. General form of Optimization Verification test Chapter 46. Reinforcement learning example Chapter 47. The rise of end-to-end learning Chapter 48. More end-to-end learning examples Chapter 49. Pros and cons of end-to-end learning Chapter 50. Choosing pipeline components: Data availability Chapter 51. Choosing pipeline components: Task simplicity Chapter 52. Directly learning rich outputs Chapter 53. Error Analysis by Parts Chapter 54. Attributing error to one part Chapter 55: General case of error attribution Chapter 56. Error analysis by parts and comparison to human-level performance Chapter 57. Spotting a flawed ML pipeline Chapter 58. Building a superhero team - Get your teammates to read this |
2023-10-27
2022-08-15
2022-08-17
2022-09-23
2022-08-13
请发表评论