您好,欢迎光临有路网!
统计学习基础
QQ咨询:
有路璐璐:

统计学习基础

  • 作者:(美)哈斯蒂(Hastie T)
  • 出版社:世界图书出版社
  • ISBN:9787506292313
  • 出版日期:2009年01月01日
  • 页数:533
  • 定价:¥88.00
  • 分享领佣金
    手机购买
    城市
    店铺名称
    店主联系方式
    店铺售价
    库存
    店铺得分/总交易量
    发布时间
    操作

    新书比价

    网站名称
    书名
    售价
    优惠
    操作

    图书详情

    内容提要
    The learning problems that we consider can be roughly categorized as either supervised or unsupervised. In supervised learning, the goal is to predict the value of an outcome measure based on a number of input measures; in unsupervised learning, there is no outcome measure, and the goal is to describe the associations and patterns among a set of input measures.
    目录
    Preface
    1 Introduction
    2 Overview of Supervised Learning
    2.1 Introduction
    2.2 Variable Types and Terminology
    2.3 Two Simple Approaches to Prediction: Least Squares and Nearest Neighbors
    2.3.1 Linear Models and Least Squares
    2.3.2 Nearest-Neighbor Methods
    2.3.3 From Least Squares to Nearest Neighbors
    2.4 Statistical Decision Theory
    2.5 Local Methods in High Dimensions
    2.6 Statistical Models, Supervised Learning and Function Approximation
    2.6.1 A Statistical Model for the Joint Distribution Pr(X,Y)
    2.6.2 Supervised Learning
    2.6.3 Function Approximation
    2.7 Structured Regression Models
    2.7.1 Difficulty of the Problem
    2.8 Classes of Restricted Estimators
    2.8.1 Roughness Penalty and Bayesian Methods
    2.8.2 Kernel Methods and Local Regression
    2.8.3 Basis Functions and Dictionary Methods
    2.9 Model Selection and the Bias-Variance Tradeoff
    Bibliographic Notes
    Exercises
    3 Linear Methods for Regression
    3.1 Introduction
    3.2 Linear Regression Models and Least Squares
    3.2.1 Example:Prostate Cancer
    3.2.2 The Ganss-Markov Theorem
    3.3 Multiple Regression from Simple Univariate Regression
    3.3.1 Multiple Outputs
    3.4 Subset Selection and Coefficient Shrinkage
    3.4.1 Subset Selection
    3.4.2 Prostate Cancer Data Example fContinued)
    3.4.3 Shrinkage Methods
    3.4.4 Methods Using Derived Input Directions
    3.4.5 Discussion:A Comparison of the Selection and Shrinkage Methods
    3.4.6 Multiple Outcome Shrinkage and Selection
    3.5 Compntational Considerations
    Bibliographic Notes
    Exercises
    4 Linear Methods for Classification
    4.1 Introduction
    4.2 Linear Regression of an Indicator Matrix
    4.3 Linear Discriminant Analysis
    4.3.1 Regularized Discriminant Analysis
    4.3.2 Computations for LDA
    4.3.3 Reduced-Rank Linear Discriminant Analysis
    4.4 Logistic Regression
    4.4.1 Fitting Logistic Regression Models
    4.4.2 Example:South African Heart Disease
    4.4.3 Quadratic Approximations and Inference
    4.4.4 Logistic Regression or LDA7
    4.5 Separating Hyper planes
    4.5.1 Rosenblatt's Perceptron Learning Algorithm
    4.5.2 Optimal Separating Hyper planes
    Bibliographic Notes
    Exercises
    5 Basis Expansions and Regularizatlon
    5.1 Introduction
    5.2 Piecewise Polynomials and Splines
    5.2.1 Natural Cubic Splines
    5.2.2 Example: South African Heart Disease (Continued)
    5.2.3 Example: Phoneme Recognition
    5.3 Filtering and Feature Extraction
    5.4 Smoothing Splines
    5.4.1 Degrees of Freedom and Smoother Matrices
    5.5 Automatic Selection of the Smoothing Parameters
    5.5.1 Fixing the Degrees of Freedom
    5.5.2 The Bias-Variance Tradeoff
    5.6 Nonparametric Logistic Regression
    5.7 Multidimensional Splines
    5.8 Regularization and Reproducing Kernel Hilbert Spaces . .
    5.8.1 Spaces of Phnctions Generated by Kernels
    5.8.2 Examples of RKHS
    5.9 Wavelet Smoothing
    5.9.1 Wavelet Bases and the Wavelet Transform
    5.9.2 Adaptive Wavelet Filtering
    Bibliographic Notes
    Exercises
    Appendix: Computational Considerations for Splines
    Appendix: B-splines
    Appendix: Computations for Smoothing Splines
    6 Kernel Methods
    7 Model Assessment and Selection
    8 Model Inference and Averaging
    9 Additive Models, Trees, and Related Methods
    10 Boosting and Additive Trees
    11 Neural Networks
    12 Support Vector Machines and Flexible Discriminants
    13 Prototype Methods and Nearest-Neighbors
    14 Unsupervised Learning
    References
    Author Index
    Index

    与描述相符

    100

    北京 天津 河北 山西 内蒙古 辽宁 吉林 黑龙江 上海 江苏 浙江 安徽 福建 江西 山东 河南 湖北 湖南 广东 广西 海南 重庆 四川 贵州 云南 西藏 陕西 甘肃 青海 宁夏 新疆 台湾 香港 澳门 海外