دسترسی نامحدود
برای کاربرانی که ثبت نام کرده اند
برای ارتباط با ما می توانید از طریق شماره موبایل زیر از طریق تماس و پیامک با ما در ارتباط باشید
در صورت عدم پاسخ گویی از طریق پیامک با پشتیبان در ارتباط باشید
برای کاربرانی که ثبت نام کرده اند
درصورت عدم همخوانی توضیحات با کتاب
از ساعت 7 صبح تا 10 شب
ویرایش:
نویسندگان: Piegorsch
سری:
ISBN (شابک) : 9781119561071
ناشر:
سال نشر: 2021
تعداد صفحات: [674]
زبان: English
فرمت فایل : PDF (درصورت درخواست کاربر به PDF، EPUB یا AZW3 تبدیل می شود)
حجم فایل: 27 Mb
در صورت تبدیل فایل کتاب Handbook of Computational Statistics and Data Science (2021) [et al] [] به فرمت های PDF، EPUB، AZW3، MOBI و یا DJVU می توانید به پشتیبان اطلاع دهید تا فایل مورد نظر را تبدیل نمایند.
توجه داشته باشید کتاب کتابچه راهنمای آمار محاسباتی و علوم داده (2021) [و همکاران] [] نسخه زبان اصلی می باشد و کتاب ترجمه شده به فارسی نمی باشد. وبسایت اینترنشنال لایبرری ارائه دهنده کتاب های زبان اصلی می باشد و هیچ گونه کتاب ترجمه شده یا نوشته شده به فارسی را ارائه نمی دهد.
یک نقشه راه ضروری برای کاربرد آمار محاسباتی در علوم داده معاصر در آمار محاسباتی در علم داده، تیمی از ریاضیدانان و آماردانان برجسته مجموعه ای تخصصی از مفاهیم، نظریه ها، تکنیک ها و شیوه های آمار محاسباتی را برای خوانندگانی که به دنبال یک آمار محاسباتی هستند، ارائه می کنند. منبع مستقل آمار در علم داده معاصر. این کتاب شامل بخشهای متعددی است که به حوزههای کلیدی و خاص در آمار محاسباتی اختصاص دارد و ارائههای مدرن و قابل دسترس از تکنیکهای بهروز را ارائه میکند. آمار محاسباتی در علم داده دسترسی رایگان به مدخل های نهایی شده در مجموعه Wiley StatsRef: Statistics Reference Online را فراهم می کند. خوانندگان همچنین خواهند یافت: مقدمه ای کامل بر آمار محاسباتی مرتبط و قابل دسترسی برای پزشکان و محققان در زمینه های مختلف داده فشرده کاوش های جامع از موضوعات فعال در آمار، از جمله داده های بزرگ، پردازش جریان داده، تجسم کمی، و یادگیری عمیق ایده آل برای محققان و محققانی که در هر زمینه ای که به تکنیک های آمار محاسباتی متوسط و پیشرفته نیاز دارد، کار می کنند، آمار محاسباتی در علم داده نیز جایگاهی را در کتابخانه های محققینی که در حال تحقیق و توسعه فناوری های داده های محاسباتی-علمی و گرافیک های آماری هستند، به دست خواهد آورد.
An essential roadmap to the application of computational statistics in contemporary data science In Computational Statistics in Data Science, a team of distinguished mathematicians and statisticians delivers an expert compilation of concepts, theories, techniques, and practices in computational statistics for readers who seek a single, standalone sourcebook on statistics in contemporary data science. The book contains multiple sections devoted to key, specific areas in computational statistics, offering modern and accessible presentations of up-to-date techniques. Computational Statistics in Data Science provides complimentary access to finalized entries in the Wiley StatsRef: Statistics Reference Online compendium. Readers will also find: A thorough introduction to computational statistics relevant and accessible to practitioners and researchers in a variety of data-intensive areas Comprehensive explorations of active topics in statistics, including big data, data stream processing, quantitative visualization, and deep learning Perfect for researchers and scholars working in any field requiring intermediate and advanced computational statistics techniques, Computational Statistics in Data Science will also earn a place in the libraries of scholars researching and developing computational data-scientific technologies and statistical graphics.
Cover Title Page Copyright Contents List of Contributors Preface Part I Computational Statistics and Data Science 1 Computational Statistics and Data Science in the Twenty‐First Century 1 Introduction 2 Core Challenges 1–3 2.1 Big N 2.2 Big P 2.3 Big M 3 Model‐Specific Advances 3.1 Bayesian Sparse Regression in the Age of Big N and Big P 3.1.1 Continuous shrinkage: alleviating big M 3.1.2 Conjugate gradient sampler for structured high‐dimensional Gaussians 3.2 Phylogenetic Reconstruction 4 Core Challenges 4 and 5 4.1 Fast, Flexible, and Friendly Statistical Algo‐Ware 4.2 Hardware‐Optimized Inference 5 Rise of Data Science Acknowledgments References 2 Statistical Software 1 User Development Environments 1.1 Extensible Text Editors: Emacs and Vim 1.2 Jupyter Notebooks 1.3 RStudio and Rmarkdown 2 Popular Statistical Software 2.1 R 2.1.1 Why use R over Python or Minitab? 2.1.2 Where can users find R support? 2.1.3 How easy is R to develop? 2.1.4 What is the downside of R? 2.1.5 Summary of R 2.2 Python 2.3 SAS® 2.4 SPSS® 3 Noteworthy Statistical Software and Related Tools 3.1 BUGS/JAGS 3.2 C++ 3.3 Microsoft Excel/Spreadsheets 3.4 Git 3.5 Java 3.6 JavaScript, Typescript 3.7 Maple 3.8 MATLAB, GNU Octave 3.9 Minitab® 3.10 Workload Managers: SLURM/LSF 3.11 SQL 3.12 Stata® 3.13 Tableau® 4 Promising and Emerging Statistical Software 4.1 Edward, Pyro, NumPyro, and PyMC3 4.2 Julia 4.3 NIMBLE 4.4 Scala 4.5 Stan 5 The Future of Statistical Computing 6 Concluding Remarks Acknowledgments References Further Reading 3 An Introduction to Deep Learning Methods 1 Introduction 2 Machine Learning: An Overview 2.1 Introduction 2.2 Supervised Learning 2.3 Gradient Descent 3 Feedforward Neural Networks 3.1 Introduction 3.2 Model Description 3.3 Training an MLP 4 Convolutional Neural Networks 4.1 Introduction 4.2 Convolutional Layer 4.3 LeNet‐5 5 Autoencoders 5.1 Introduction 5.2 Objective Function 5.3 Variational Autoencoder 6 Recurrent Neural Networks 6.1 Introduction 6.2 Architecture 6.3 Long Short‐Term Memory Networks 7 Conclusion References 4 Streaming Data and Data Streams 1 Introduction 2 Data Stream Computing 3 Issues in Data Stream Mining 3.1 Scalability 3.2 Integration 3.3 Fault‐Tolerance 3.4 Timeliness 3.5 Consistency 3.6 Heterogeneity and Incompleteness 3.7 Load Balancing 3.8 High Throughput 3.9 Privacy 3.10 Accuracy 4 Streaming Data Tools and Technologies 5 Streaming Data Pre‐Processing: Concept and Implementation 6 Streaming Data Algorithms 6.1 Unsupervised Learning 6.2 Semi‐Supervised Learning 6.3 Supervised Learning 6.4 Ontology‐Based Methods 7 Strategies for Processing Data Streams 8 Best Practices for Managing Data Streams 9 Conclusion and the Way Forward References Part II Simulation‐Based Methods 5 Monte Carlo Simulation: Are We There Yet? 1 Introduction 2 Estimation 2.1 Expectations 2.2 Quantiles 2.3 Other Estimators 3 Sampling Distribution 3.1 Means 3.2 Quantiles 3.3 Other Estimators 3.4 Confidence Regions for Means 4 Estimating ???? 5 Stopping Rules 5.1 IID Monte Carlo 5.2 MCMC 6 Workflow 7 Examples 7.1 Action Figure Collector Problem 7.2 Estimating Risk for Empirical Bayes 7.3 Bayesian Nonlinear Regression References 6 Sequential Monte Carlo: Particle Filters and Beyond 1 Introduction 2 Sequential Importance Sampling and Resampling 2.1 Extended State Spaces and SMC Samplers 2.2 Particle MCMC and Related Methods 3 SMC in Statistical Contexts 3.1 SMC for Hidden Markov Models 3.1.1 Filtering 3.1.2 Smoothing 3.1.3 Parameter estimation 3.2 SMC for Bayesian Inference 3.2.1 SMC for model comparison 3.2.2 SMC for ABC 3.3 SMC for Maximum‐Likelihood Estimation 3.4 SMC for Rare Event Estimation 4 Selected Recent Developments Acknowledgments References 7 Markov Chain Monte Carlo Methods, A Survey with Some Frequent Misunderstandings 1 Introduction 2 Monte Carlo Methods 3 Markov Chain Monte Carlo Methods 3.1 Metropolis–Hastings Algorithms 3.2 Gibbs Sampling 3.3 Hamiltonian Monte Carlo 4 Approximate Bayesian Computation 5 Further Reading Abbreviations and Acronyms References 8 Bayesian Inference with Adaptive Markov Chain Monte Carlo 1 Introduction 2 Random‐Walk Metropolis Algorithm 3 Adaptation of Random‐Walk Metropolis 3.1 Adaptive Metropolis (AM) 3.2 Adaptive Scaling Metropolis (ASM) 3.3 Robust Adaptive Metropolis (RAM) 3.4 Rationale behind the Adaptations 3.5 Summary and Discussion on the Methods 4 Multimodal Targets with Parallel Tempering 5 Dynamic Models with Particle Filters 6 Discussion Acknowledgments References 9 Advances in Importance Sampling 1 Introduction and Problem Statement 1.1 Standard Monte Carlo Integration 2 Importance Sampling 2.1 Origins 2.2 Basics 2.3 Theoretical Analysis 2.4 Diagnostics 2.5 Other IS Schemes 2.5.1 Transformation of the importance weights 2.5.2 Particle filtering (sequential Monte Carlo) 3 Multiple Importance Sampling (MIS) 3.1 Generalized MIS 3.1.1 MIS with different number of samples per proposal 3.2 Rare Event Estimation 3.3 Compressed and Distributed IS 4 Adaptive Importance Sampling (AIS) Acknowledgments References Part III Statistical Learning 10 Supervised Learning 1 Introduction 2 Penalized Empirical Risk Minimization 2.1 Bias–Variance Trade‐Off 2.2 First‐Order Optimization Methods 3 Linear Regression 3.1 Linear Regression and Ridge Regression 3.2 LASSO 4 Classification 4.1 Model‐Based Methods 4.2 Support Vector Machine (SVM) 4.3 Convex Surrogate Loss 4.3.1 Surrogate risk minimization 4.3.2 Large‐margin unified machines (LUMs) 4.4 Nonconvex Surrogate Loss 4.5 Multicategory Classification Problem 5 Extensions for Complex Data 5.1 Reproducing Kernel Hilbert Space (RKHS) 5.2 Large‐Scale Optimization 6 Discussion References 11 Unsupervised and Semisupervised Learning 1 Introduction 2 Unsupervised Learning 2.1 Mixture‐Model‐Based Clustering 2.1.1 Gaussian mixture model 2.1.2 Clustering by mode association 2.1.3 Hidden Markov model on variable blocks 2.1.4 Variable selection 2.2 Clustering of Distributional Data 2.3 Uncertainty Analysis 3 Semisupervised Learning 3.1 Setting 3.2 Self‐Training 3.3 Generative Models 3.4 Graphical Models 3.5 Entropy Minimization 3.6 Consistency Regularization 3.7 Mixup 3.8 MixMatch 4 Conclusions Acknowledgment References 12 Random Forests 1 Introduction 2 Random Forest (RF) 2.1 RF Algorithm 2.2 RF Advantages and Limitations 3 Random Forest Extensions 3.1 Extremely Randomized Trees (ERT) 3.2 Acceptance‐Rejection Trees (ART) 3.3 Conditional Random Forest (CRF) 3.4 Miscellaneous 4 Random Forests of Interaction Trees (RFIT) 4.1 Modified Splitting Statistic 4.2 Standard Errors 4.3 Concomitant Outputs 4.4 Illustration of RFIT 5 Random Forest of Interaction Trees for Observational Studies 5.1 Propensity Score 5.2 Random Forest Adjusting for Propensity Score 5.3 Variable Importance 5.4 Simulation Study 6 Discussion References 13 Network Analysis 1 Introduction 2 Gaussian Graphical Models for Mixed Partial Compositional Data 2.1 A Statistical Framework for Mixed Partial Compositional Data 2.2 Estimation of Gaussian Graphical Models of Mixed Partial Compositional Data 3 Theoretical Properties 3.1 Assumptions 3.2 Rates of Convergence 4 Graphical Model Selection 5 Analysis of a Microbiome–Metabolomics Data 6 Discussion References 14 Tensors in Modern Statistical Learning 1 Introduction 2 Background 2.1 Definitions and Notation 2.2 Tensor Operations 2.3 Tensor Decompositions 3 Tensor Supervised Learning 3.1 Tensor Predictor Regression 3.1.1 Motivating examples 3.1.2 Low‐rank linear and generalized linear model 3.1.3 Large‐scale tensor regression via sketching 3.1.4 Nonparametric tensor regression 3.1.5 Future directions 3.2 Tensor Response Regression 3.2.1 Motivating examples 3.2.2 Sparse low‐rank tensor response model 3.2.3 Additional tensor response regression models 3.2.4 Future directions 4 Tensor Unsupervised Learning 4.1 Tensor Clustering 4.1.1 Motivating examples 4.1.2 Convex tensor co‐clustering 4.1.3 Tensor clustering via low‐rank decomposition 4.1.4 Additional tensor clustering approaches 4.1.5 Future directions 4.2 Tensor Graphical Model 4.2.1 Motivating examples 4.2.2 Gaussian graphical model 4.2.3 Variation in the Kronecker structure 4.2.4 Future directions 5 Tensor Reinforcement Learning 5.1 Stochastic Low‐Rank Tensor Bandit 5.1.1 Motivating examples 5.1.2 Low‐rank tensor bandit problem formulation 5.1.3 Rank‐1 bandit 5.1.4 General‐rank bandit 5.1.5 Future directions 5.2 Learning Markov Decision Process via Tensor Decomposition 5.2.1 Motivating examples 5.2.2 Dimension reduction of Markov decision process 5.2.3 Maximum‐likelihood estimation and Tucker decomposition 5.2.4 Future directions 6 Tensor Deep Learning 6.1 Tensor‐Based Deep Neural Network Compression 6.1.1 Motivating examples 6.1.2 Compression of convolutional layers of CNN 6.1.3 Compression of fully‐connected layers of CNN 6.1.4 Compression of all layers of CNN 6.1.5 Compression of RNN 6.1.6 Future directions 6.2 Deep Learning Theory through Tensor Methods 6.2.1 Motivating examples 6.2.2 Expressive power, compressibility and generalizability 6.2.3 Additional connections 6.2.4 Future directions Acknowledgments References 15 Computational Approaches to Bayesian Additive Regression Trees 1 Introduction 2 Bayesian CART 2.1 A Single‐Tree Model 2.2 Tree Model Likelihood 2.3 Tree Model Prior 2.3.1 p(????) 2.3.2 p(????|????) 3 Tree MCMC 3.1 The BIRTH/DEATH Move 3.2 CHANGE Rule 3.3 SWAP Rule 3.4 Improved Tree Space Moves 3.4.1 Rotate 3.4.2 Perturb 3.4.3 The complex mixtures that are tree proposals 4 The BART Model 4.1 Specification of the BART Regularization Prior 5 BART Example: Boston Housing Values and Air Pollution 6 BART MCMC 7 BART Extentions 7.1 The DART Sparsity Prior 7.1.1 Grouped variables and the DART prior 7.2 XBART 7.2.1 The XBART algorithm and GrowFromRoot 7.2.2 Warm‐start XBART 8 Conclusion References Part IV High‐Dimensional Data Analysis 16 Penalized Regression 1 Introduction 2 Penalization for Smoothness 3 Penalization for Sparsity 4 Tuning Parameter Selection References 17 Model Selection in High‐Dimensional Regression 1 Model Selection Problem 2 Model Selection in High‐Dimensional Linear Regression 2.1 Shrinkage Methods 2.2 Sure Screening Methods 2.3 Model Selection Theory 2.4 Tuning Parameter Selection 2.5 Numerical Computation 3 Interaction‐Effect Selection for High‐Dimensional Data 3.1 Problem Setup 3.2 Joint Selection of Main Effects and Interactions 3.3 Two‐Stage Approach 3.4 Regularization Path Algorithm under Marginality Principle (RAMP) 4 Model Selection in High‐Dimensional Nonparametric Models 4.1 Model Selection Problem Penalty on basis coefficients Function soft‐thresholding methods 4.2 Penalty on Basis Coefficients Polynomial basis example Basis pursuit Adaptive group LASSO 4.3 Component Selection and Smoothing Operator (COSSO) 4.4 Adaptive COSSO Nonparametric oracle property 4.5 Sparse Additive Models (SpAM) Persistent property 4.6 Sparsity‐Smoothness Penalty 4.7 Nonparametric Independence Screening (NIS) 5 Concluding Remarks References 18 Sampling Local Scale Parameters in High‐Dimensional Regression Models 1 Introduction 2 A Blocked Gibbs Sampler for the Horseshoe 2.1 Some Highlights for the Blocked Algorithm 3 Sampling (ξ,σ2,β) 3.1 Sampling ξ 3.2 Sampling σ2 3.3 Sampling β 4 Sampling η 4.1 The Slice Sampling Strategy 4.2 Direct Sampling 4.2.1 Inverse‐cdf sampler 5 Appendix: A. Newton–Raphson Steps for the Inverse‐cdf Sampler for η Acknowledgment References 19 Factor Modeling for High‐Dimensional Time Series 1 Introduction 2 Identifiability 3 Estimation of High‐Dimensional Factor Model 3.1 Least‐Squares or Principal Component Estimation 3.2 Factor Loading Space Estimation 3.2.1 Improved Estimation of Factor Process 3.3 Frequency‐Domain Approach 3.4 Likelihood‐Based Estimation 3.4.1 Exact likelihood via Kalman filtering 3.4.2 Exact likelihood via matrix decomposition 3.4.3 Bai and Li's Quasi‐likelihood estimation 3.4.4 Breitung and Tenhofen's Quasi‐likelihood estimation 3.4.5 Frequency‐domain (Whittle) likelihood 4 Determining the Number of Factors 4.1 Information Criterion 4.2 Eigenvalues Difference/Ratio Estimators 4.3 Testing Approaches 4.4 Estimation of Dynamic Factors Acknowledgment References Part V Quantitative Visualization 20 Visual Communication of Data: It Is Not a Programming Problem, It Is Viewer Perception 1 Introduction 1.1 Observation 1.2 Available Guidance 1.3 Our Message 2 Case Studies Part 1 2.1 Imogene: A Senior Data Analyst Who Becomes Too Interested in the Program 2.2 Regis: An Intern Who Wants to Get the Job Done Quickly 3 Let StAR Be Your Guide 4 Case Studies Part 2: Using StAR Principles to Develop Better Graphics 4.1 StAR Method: Imogene Thinks through and Investigates Changing Scales 4.2 StAR Method: Regis Thinks through and Discovers an Interesting Way to Depict Uncertainty 5 Ask Colleagues Their Opinion 6 Case Studies: Part 3 6.1 Imogene Gets Advice on Using Dot Plots 6.2 Regis Gets Advice on Visualizing in the Presence of Multiple Tests 7 Iterate 8 Final Thoughts References 21 Uncertainty Visualization 1 Introduction 1.1 Uncertainty Visualization Design Space 2 Uncertainty Visualization Theories 2.1 Frequency Framing 2.1.1 Icon arrays 2.1.2 Quantile dotplots 2.2 Attribute Substitution 2.2.1 Hypothetical outcome plots 2.3 Visual Boundaries = Cognitive Categories 2.3.1 Ensemble displays 2.3.2 Error bars 2.4 Visual Semiotics of Uncertainty 3 General Discussion References 22 Big Data Visualization 1 Introduction 2 Architecture for Big Data Analytics 3 Filtering 3.1 Sampling 4 Aggregating 4.1 1D Continuous Aggregation 4.2 1D Categorical Aggregation 4.3 2D Aggregation 4.3.1 2D binning on the surface of a sphere 4.3.2 2D categorical versus continuous aggregation 4.3.3 2D categorical versus categorical aggregation 4.4 nD Aggregation 4.5 Two‐Way Aggregation 5 Analyzing 6 Big Data Graphics 6.1 Box Plots 6.2 Histograms 6.3 Scatterplot Matrices 6.4 Parallel Coordinates 7 Conclusion References 23 Visualization‐Assisted Statistical Learning 1 Introduction 2 Better Visualizations with Seriation 3 Visualizing Machine Learning Fits 3.1 Partial Dependence 3.2 FEV Dataset 3.3 Interactive Conditional Visualization 4 Condvis2 Case Studies 4.1 Interactive Exploration of FEV Regression Models 4.2 Interactive Exploration of Pima Classification Models 4.3 Interactive Exploration of Models for Wages Repeated Measures Data 5 Discussion References 24 Functional Data Visualization 1 Introduction 2 Univariate Functional Data Visualization 2.1 Functional Boxplots 2.2 Surface Boxplots 3 Multivariate Functional Data Visualization 3.1 Magnitude–Shape Plots 3.2 Two‐Stage Functional Boxplots 3.3 Trajectory Functional Boxplots 4 Conclusions Acknowledgment References Part VI Numerical Approximation and Optimization 25 Gradient‐Based Optimizers for Statistics and Machine Learning 1 Introduction 2 Convex Versus Nonconvex Optimization 3 Gradient Descent 3.1 Basic Formulation 3.2 How to Find the Step Size? 3.3 Examples 4 Proximal Gradient Descent: Handling Nondifferentiable Regularization 5 Stochastic Gradient Descent 5.1 Basic Formulation 5.2 Challenges References 26 Alternating Minimization Algorithms 1 Introduction 2 Coordinate Descent 3 EM as Alternating Minimization 3.1 Finite Mixture Models 3.2 Variational EM 4 Matrix Approximation Algorithms 4.1 k‐Means Clustering 4.2 Low‐Rank Matrix Factorization 4.3 Reduced Rank Regression 5 Conclusion References 27 A Gentle Introduction to Alternating Direction Method of Multipliers (ADMM) for Statistical Problems 1 Introduction 2 Two Perfect Examples of ADMM 3 Variable Splitting and Linearized ADMM 4 Multiblock ADMM 5 Nonconvex Problems 6 Stopping Criteria 7 Convergence Results of ADMM 7.1 Convex Problems 7.1.1 Convex case 7.1.2 Strongly convex case 7.1.3 Linearized ADMM 7.2 Nonconvex Problems Acknowledgments References 28 Nonconvex Optimization via MM Algorithms: Convergence Theory 1 Background 2 Convergence Theorems 2.1 Classical Convergence Theorem 2.2 Smooth Objective Functions 2.3 Nonsmooth Objective Functions 2.3.1 MM convergence for semialgebraic functions 2.4 A Proximal Trick to Prevent Cycling 3 Paracontraction 4 Bregman Majorization 4.1 Convergence Analysis via SUMMA 4.2 Examples 4.2.1 Proximal gradient method 4.2.2 Mirror descent method References Part VII High‐Performance Computing 29 Massive Parallelization 1 Introduction 2 Gaussian Process Regression and Surrogate Modeling 2.1 GP Basics 2.2 Pushing the Envelope 3 Divide‐and‐Conquer GP Regression 3.1 Local Approximate Gaussian Processes 3.2 Massively Parallelized Global GP Approximation 3.3 Off‐Loading Subroutines to GPUs 4 Empirical Results 4.1 SARCOS 4.2 Supercomputer Cascade 5 Conclusion Acknowledgments References 30 Divide‐and‐Conquer Methods for Big Data Analysis 1 Introduction 2 Linear Regression Model 3 Parametric Models 3.1 Sparse High‐Dimensional Models 3.2 Marginal Proportional Hazards Model 3.3 One‐Step Estimator and Multiround Divide‐and‐Conquer 3.4 Performance in Nonstandard Problems 4 Nonparametric and Semiparametric Models 5 Online Sequential Updating 6 Splitting the Number of Covariates 7 Bayesian Divide‐and‐Conquer and Median‐Based Combining 8 Real‐World Applications 9 Discussion Acknowledgment References 31 Bayesian Aggregation 1 From Model Selection to Model Combination 1.1 The Bayesian Decision Framework for Model Assessment 1.2 Remodeling: ℳ‐Closed, ℳ‐Complete, and ℳ‐Open Views 2 From Bayesian Model Averaging to Bayesian Stacking 2.1 ℳ‐Closed: Bayesian Model Averaging 2.2 ℳ‐Open: Stacking 2.2.1 Choice of utility 2.3 ℳ‐Complete: Reference‐Model Stacking 2.4 The Connection between BMA and Stacking 2.5 Hierarchical Stacking 2.6 Other Related Methods and Generalizations 3 Asymptotic Theories of Stacking 3.1 Model Aggregation Is No Worse than Model Selection 3.2 Stacking Viewed as Pointwise Model Selection 3.3 Selection or Averaging? 4 Stacking in Practice 4.1 Practical Implementation Using Pareto Smoothed Importance Sampling 4.2 Stacking for Multilevel Data 4.3 Stacking for Time Series Data 4.4 The Choice of Model List 5 Discussion References 32 Asynchronous Parallel Computing 1 Introduction 1.1 Synchronous and Asynchronous Parallel Computing 1.2 Not All Algorithms Can Benefit from Parallelization 1.3 Outline 1.4 Notation 2 Asynchronous Parallel Coordinate Update 2.1 Least Absolute Shrinkage and Selection Operator (LASSO) 2.2 Nonnegative Matrix Factorization 2.3 Kernel Support Vector Machine 2.4 Decentralized Algorithms 3 Asynchronous Parallel Stochastic Approaches 3.1 Hogwild! 3.2 Federated Learning 4 Doubly Stochastic Coordinate Optimization with Variance Reduction 5 Concluding Remarks References Index Abbreviations and Acronyms EULA