دسترسی نامحدود
برای کاربرانی که ثبت نام کرده اند
برای ارتباط با ما می توانید از طریق شماره موبایل زیر از طریق تماس و پیامک با ما در ارتباط باشید
در صورت عدم پاسخ گویی از طریق پیامک با پشتیبان در ارتباط باشید
برای کاربرانی که ثبت نام کرده اند
درصورت عدم همخوانی توضیحات با کتاب
از ساعت 7 صبح تا 10 شب
ویرایش:
نویسندگان: Jeremy Howard. Sylvain Gugger
سری:
ISBN (شابک) : 1492045527, 9781492045526
ناشر: Oreilly & Associates Inc
سال نشر: 2020
تعداد صفحات: 0
زبان: English
فرمت فایل : EPUB (درصورت درخواست کاربر به PDF، EPUB یا AZW3 تبدیل می شود)
حجم فایل: 30 مگابایت
در صورت تبدیل فایل کتاب Deep Learning for Coders With Fastai and Pytorch: Ai Applications Without a Phd به فرمت های PDF، EPUB، AZW3، MOBI و یا DJVU می توانید به پشتیبان اطلاع دهید تا فایل مورد نظر را تبدیل نمایند.
توجه داشته باشید کتاب یادگیری عمیق برای برنامه نویسان با Fastai و Pytorch: برنامه های کاربردی Ai بدون مدرک دکترا نسخه زبان اصلی می باشد و کتاب ترجمه شده به فارسی نمی باشد. وبسایت اینترنشنال لایبرری ارائه دهنده کتاب های زبان اصلی می باشد و هیچ گونه کتاب ترجمه شده یا نوشته شده به فارسی را ارائه نمی دهد.
یادگیری عمیق اغلب به عنوان حوزه انحصاری دکترای ریاضی و شرکت های بزرگ فناوری در نظر گرفته می شود. اما همانطور که این راهنمای عملی نشان میدهد، برنامهنویسانی که با پایتون راحت هستند، میتوانند با پیشزمینه ریاضی کم، مقادیر کم داده و حداقل کد به نتایج چشمگیری در یادگیری عمیق دست یابند. چگونه؟ با fastai، اولین کتابخانه ای که یک رابط ثابت برای برنامه های کاربردی یادگیری عمیق ارائه می دهد.
نویسندگان جرمی هوارد و سیلوین گوگر به شما نشان می دهند که چگونه با استفاده از fastai و PyTorch یک مدل را در طیف وسیعی از وظایف آموزش دهید. . همچنین برای به دست آوردن درک کاملی از الگوریتم های پشت صحنه، به تدریج در تئوری یادگیری عمیق فرو خواهید رفت.
Deep learning is often viewed as the exclusive domain of math PhDs and big tech companies. But as this hands-on guide demonstrates, programmers comfortable with Python can achieve impressive results in deep learning with little math background, small amounts of data, and minimal code. How? With fastai, the first library to provide a consistent interface to the most frequently used deep learning applications.
Authors Jeremy Howard and Sylvain Gugger show you how to train a model on a wide range of tasks using fastai and PyTorch. You&;ll also dive progressively further into deep learning theory to gain a complete understanding of the algorithms behind the scenes.
Copyright Table of Contents Preface Who This Book Is For What You Need to Know What You Will Learn O’Reilly Online Learning How to Contact Us Foreword Part I. Deep Learning in Practice Chapter 1. Your Deep Learning Journey Deep Learning Is for Everyone Neural Networks: A Brief History Who We Are How to Learn Deep Learning Your Projects and Your Mindset The Software: PyTorch, fastai, and Jupyter (And Why It Doesn’t Matter) Your First Model Getting a GPU Deep Learning Server Running Your First Notebook What Is Machine Learning? What Is a Neural Network? A Bit of Deep Learning Jargon Limitations Inherent to Machine Learning How Our Image Recognizer Works What Our Image Recognizer Learned Image Recognizers Can Tackle Non-Image Tasks Jargon Recap Deep Learning Is Not Just for Image Classification Validation Sets and Test Sets Use Judgment in Defining Test Sets A Choose Your Own Adventure Moment Questionnaire Further Research Chapter 2. From Model to Production The Practice of Deep Learning Starting Your Project The State of Deep Learning The Drivetrain Approach Gathering Data From Data to DataLoaders Data Augmentation Training Your Model, and Using It to Clean Your Data Turning Your Model into an Online Application Using the Model for Inference Creating a Notebook App from the Model Turning Your Notebook into a Real App Deploying Your App How to Avoid Disaster Unforeseen Consequences and Feedback Loops Get Writing! Questionnaire Further Research Chapter 3. Data Ethics Key Examples for Data Ethics Bugs and Recourse: Buggy Algorithm Used for Healthcare Benefits Feedback Loops: YouTube’s Recommendation System Bias: Professor Latanya Sweeney “Arrested” Why Does This Matter? Integrating Machine Learning with Product Design Topics in Data Ethics Recourse and Accountability Feedback Loops Bias Disinformation Identifying and Addressing Ethical Issues Analyze a Project You Are Working On Processes to Implement The Power of Diversity Fairness, Accountability, and Transparency Role of Policy The Effectiveness of Regulation Rights and Policy Cars: A Historical Precedent Conclusion Questionnaire Further Research Deep Learning in Practice: That’s a Wrap! Part II. Understanding fastai’s Applications Chapter 4. Under the Hood: Training a Digit Classifier Pixels: The Foundations of Computer Vision First Try: Pixel Similarity NumPy Arrays and PyTorch Tensors Computing Metrics Using Broadcasting Stochastic Gradient Descent Calculating Gradients Stepping with a Learning Rate An End-to-End SGD Example Summarizing Gradient Descent The MNIST Loss Function Sigmoid SGD and Mini-Batches Putting It All Together Creating an Optimizer Adding a Nonlinearity Going Deeper Jargon Recap Questionnaire Further Research Chapter 5. Image Classification From Dogs and Cats to Pet Breeds Presizing Checking and Debugging a DataBlock Cross-Entropy Loss Viewing Activations and Labels Softmax Log Likelihood Taking the log Model Interpretation Improving Our Model The Learning Rate Finder Unfreezing and Transfer Learning Discriminative Learning Rates Selecting the Number of Epochs Deeper Architectures Conclusion Questionnaire Further Research Chapter 6. Other Computer Vision Problems Multi-Label Classification The Data Constructing a DataBlock Binary Cross Entropy Regression Assembling the Data Training a Model Conclusion Questionnaire Further Research Chapter 7. Training a State-of-the-Art Model Imagenette Normalization Progressive Resizing Test Time Augmentation Mixup Label Smoothing Conclusion Questionnaire Further Research Chapter 8. Collaborative Filtering Deep Dive A First Look at the Data Learning the Latent Factors Creating the DataLoaders Collaborative Filtering from Scratch Weight Decay Creating Our Own Embedding Module Interpreting Embeddings and Biases Using fastai.collab Embedding Distance Bootstrapping a Collaborative Filtering Model Deep Learning for Collaborative Filtering Conclusion Questionnaire Further Research Chapter 9. Tabular Modeling Deep Dive Categorical Embeddings Beyond Deep Learning The Dataset Kaggle Competitions Look at the Data Decision Trees Handling Dates Using TabularPandas and TabularProc Creating the Decision Tree Categorical Variables Random Forests Creating a Random Forest Out-of-Bag Error Model Interpretation Tree Variance for Prediction Confidence Feature Importance Removing Low-Importance Variables Removing Redundant Features Partial Dependence Data Leakage Tree Interpreter Extrapolation and Neural Networks The Extrapolation Problem Finding Out-of-Domain Data Using a Neural Network Ensembling Boosting Combining Embeddings with Other Methods Conclusion Questionnaire Further Research Chapter 10. NLP Deep Dive: RNNs Text Preprocessing Tokenization Word Tokenization with fastai Subword Tokenization Numericalization with fastai Putting Our Texts into Batches for a Language Model Training a Text Classifier Language Model Using DataBlock Fine-Tuning the Language Model Saving and Loading Models Text Generation Creating the Classifier DataLoaders Fine-Tuning the Classifier Disinformation and Language Models Conclusion Questionnaire Further Research Chapter 11. Data Munging with fastai’s Mid-Level API Going Deeper into fastai’s Layered API Transforms Writing Your Own Transform Pipeline TfmdLists and Datasets: Transformed Collections TfmdLists Datasets Applying the Mid-Level Data API: SiamesePair Conclusion Questionnaire Further Research Understanding fastai’s Applications: Wrap Up Part III. Foundations of Deep Learning Chapter 12. A Language Model from Scratch The Data Our First Language Model from Scratch Our Language Model in PyTorch Our First Recurrent Neural Network Improving the RNN Maintaining the State of an RNN Creating More Signal Multilayer RNNs The Model Exploding or Disappearing Activations LSTM Building an LSTM from Scratch Training a Language Model Using LSTMs Regularizing an LSTM Dropout Activation Regularization and Temporal Activation Regularization Training a Weight-Tied Regularized LSTM Conclusion Questionnaire Further Research Chapter 13. Convolutional Neural Networks The Magic of Convolutions Mapping a Convolutional Kernel Convolutions in PyTorch Strides and Padding Understanding the Convolution Equations Our First Convolutional Neural Network Creating the CNN Understanding Convolution Arithmetic Receptive Fields A Note About Twitter Color Images Improving Training Stability A Simple Baseline Increase Batch Size 1cycle Training Batch Normalization Conclusion Questionnaire Further Research Chapter 14. ResNets Going Back to Imagenette Building a Modern CNN: ResNet Skip Connections A State-of-the-Art ResNet Bottleneck Layers Conclusion Questionnaire Further Research Chapter 15. Application Architectures Deep Dive Computer Vision cnn_learner unet_learner A Siamese Network Natural Language Processing Tabular Conclusion Questionnaire Further Research Chapter 16. The Training Process Establishing a Baseline A Generic Optimizer Momentum RMSProp Adam Decoupled Weight Decay Callbacks Creating a Callback Callback Ordering and Exceptions Conclusion Questionnaire Further Research Foundations of Deep Learning: Wrap Up Part IV. Deep Learning from Scratch Chapter 17. A Neural Net from the Foundations Building a Neural Net Layer from Scratch Modeling a Neuron Matrix Multiplication from Scratch Elementwise Arithmetic Broadcasting Einstein Summation The Forward and Backward Passes Defining and Initializing a Layer Gradients and the Backward Pass Refactoring the Model Going to PyTorch Conclusion Questionnaire Further Research Chapter 18. CNN Interpretation with CAM CAM and Hooks Gradient CAM Conclusion Questionnaire Further Research Chapter 19. A fastai Learner from Scratch Data Dataset Module and Parameter Simple CNN Loss Learner Callbacks Scheduling the Learning Rate Conclusion Questionnaire Further Research Chapter 20. Concluding Thoughts Appendix A. Creating a Blog Blogging with GitHub Pages Creating the Repository Setting Up Your Home Page Creating Posts Synchronizing GitHub and Your Computer Jupyter for Blogging Appendix B. Data Project Checklist Data Scientists Strategy Data Analytics Implementation Maintenance Constraints Index About the Authors Acknowledgments Colophon