reduced rank regression model

    • The Fixed-X Case
    • Classical Multivariate Regression Model
    • The Random-X Case
  • Multivariate Reduced-Rank Regression
    • RRR的应用

关于reduced rank regression model的内容可以直接跳到Multivariate Reduced-Rank Regression的部分

Multivariate linear regression is a natural extension of multiple linear regression in that both techniques try to interpret possible linear relationships between certain input and output variables. Multiple regression is concerned with studying to what extent the behavior of a single output variable Y is influenced by a set of r input variables X = (X1, ··· ,Xr) T ^T T .
Multivariate regression has s output variables Y = (Y1, ··· ,Ys) T ^T T , each of whose behavior may be influenced by exactly the same set of inputs
X = (X1, ··· ,Xr) T ^T T .
So, not only are the components of X correlated with each other, but in multivariate regression, the components of Y are also correlated with each other (and with the components of X). In this chapter, we are interested in estimating the regression relationship between Y and X, taking into account the various dependencies between the r-vector X and the s-vector Y and the dependencies within X and within Y.
we describe the multivariate reduced-rank regression model (RRR) (Izenman, 1975), which is an enhancement of the classical multivariate regression model and has recently received research attention in the statistics and econometrics literature. The following reasons explain the popularity of this model: RRR provides a unified approach to many of the diverse classical multivariate statistical techniques; it lends itself quite naturally to analyzing a wide variety of statistical problems involving reduction of dimensionality and the search for structure in multivariate data; and it is relatively simple to program because the regression estimates depend only upon the sample covariance matrices of X and Y and the eigendecomposition of a certain symmetric matrix that generalizes the multiple squared correlation coefficient R2 from multiple regression.

我们需要考虑X是随机和非随机两种情况

The Fixed-X Case

Let Y = (Y1, ··· ,Ys) T ^T T be a random s-vector-valued output variate with
mean vector µ Y µ_Y µY​ and covariance matrix Σ Y Y Σ_{Y Y} ΣYY​ , and let X = (X1, ··· ,Xr) T ^T T
be a fixed (nonstochastic) r-vector-valued input variate. The components
of the output vector Y will typically be continuous responses, and the
components of the input vector X may be indicator or “dummy” variables
that are set up by the researcher to identify known groupings of the data
associated with distinct subpopulations or experimental conditions.

Suppose we observe n replications,
( X j T , Y j T ) τ (X_j^T , Y_j^T )^τ (XjT​,YjT​)τ, j = 1, 2,… ,n,
on the (r + s)-vector ( X τ , Y τ ) τ (X^τ , Y^τ )^τ (Xτ,Yτ)τ . We define an (r × n)-matrix X and an
(s × n)-matrix Y by X = ( X 1 , ⋅ ⋅ ⋅ , X n ) , Y = ( Y 1 , ⋅ ⋅ ⋅ , Y n ) \mathcal X = (X1, ··· , Xn), \mathcal Y = (Y1, ··· , Yn) X=(X1,⋅⋅⋅,Xn),Y=(Y1,⋅⋅⋅,Yn).

Form the mean vectors,
X ˉ = n − 1 ∑ j = 1 n X j \bar{X} = n^{−1}\sum_{j=1}^n Xj Xˉ=n−1∑j=1n​Xj , Y ˉ = n − 1 ∑ j = 1 n Y j \bar{Y} = n^{−1}\sum_{j=1}^n Yj Yˉ=n−1∑j=1n​Yj
and let
X = ( X ˉ , ⋅ ⋅ ⋅ , X ˉ ) , Y = ( Y ˉ , ⋅ ⋅ ⋅ , Y ˉ ) \mathcal{X} = (\bar{X} , ··· , \bar{X} ), \mathcal{Y} = (\bar{Y} , ··· , \bar{Y} ) X=(Xˉ,⋅⋅⋅,Xˉ),Y=(Yˉ,⋅⋅⋅,Yˉ)
The centered versions of X and Y are defined by
X c = X − X ˉ = ( X 1 − X ˉ , ⋅ ⋅ ⋅ , X n − X ˉ ) \mathcal X_c=\mathcal X − \bar{\mathcal X}=(X1 − \bar{X} , ··· , Xn − \bar{X} ) Xc​=X−Xˉ=(X1−Xˉ,⋅⋅⋅,Xn−Xˉ),
Y c = Y − Y ˉ = ( Y 1 − Y ˉ , ⋅ ⋅ ⋅ , Y n − Y ˉ ) \mathcal Y_c = \mathcal Y − \bar{\mathcal Y} = (Y1 − \bar{Y} , ··· , Yn − \bar{Y} ) Yc​=Y−Yˉ​=(Y1−Yˉ,⋅⋅⋅,Yn−Yˉ),
respectively.

Classical Multivariate Regression Model

Consider the multivariate linear regression model
Y s × n = µ s × n + Θ s × r X r × n + ϵ s × n Y_{s×n} =µ_{s×n} + Θ_{s×r} X_ {r×n} +\epsilon_ {s×n} Ys×n​=µs×n​+Θs×r​Xr×n​+ϵs×n​,
( θ 11 θ 12 . . . θ 1 n θ 21 θ 22 . . . θ 2 n . . . . . . . . . X s 1 θ s 2 . . . θ s r ) ∗ ( X 11 X 12 . . . X 1 n X 21 X 22 . . . X 2 n . . . . . . . . . X r 1 X r 2 . . . X r n ) \begin{pmatrix} \theta_{11} & \theta_{12} &...& \theta_{1n} \\\theta_{21} & \theta_{22} &...& \theta_{2n}\\ ...&...& &...&\\X_{s1} & \theta_{s2} &...& \theta_{sr} \end{pmatrix}*\begin{pmatrix} X_{11} & X_{12} &...& X_{1n} \\ X_{21} & X_{22} &...& X_{2n}\\ ...&...& &...&\\X_{r1} & X_{r2} &...& X_{rn} \end{pmatrix} ⎝⎜⎜⎛​θ11​θ21​...Xs1​​θ12​θ22​...θs2​​.........​θ1n​θ2n​...θsr​​​⎠⎟⎟⎞​∗⎝⎜⎜⎛​X11​X21​...Xr1​​X12​X22​...Xr2​​.........​X1n​X2n​...Xrn​​​⎠⎟⎟⎞​

ϵ = ( ϵ 1 , ϵ 2 , ⋅ ⋅ ⋅ , ϵ n ) \epsilon= (\epsilon_1, \epsilon_2, ··· , \epsilon_n) ϵ=(ϵ1​,ϵ2​,⋅⋅⋅,ϵn​) is the (s × n) error matrix whose columns are each random s-vectors with mean 0 and the same unknown nonsingular (s × s) error covariance matrix Σ ϵ ϵ Σ_{ \epsilon \epsilon} Σϵϵ​,and pairs of column vectors, ( ϵ j , ϵ k ) ( \epsilon_j , \epsilon_k) (ϵj​,ϵk​), j = k, are uncorrelated with each other

When the Xs are considered to be fixed in repeated sampling (e.g., in designed experiments), the so-called design matrix X consists of known constants and possibly also observed values of covariates, Θ is a full-rank matrix of unknown fixed effects, and µ = µ 0 1 n τ µ = µ_01^τ_n µ=µ0​1nτ​, where µ 0 µ_0 µ0​ is an unknown s-vector of constants.

LS Estimation
If we set µ = Y ˉ − Θ X ˉ \bar{\mathcal Y} - Θ\bar{\mathcal X} Yˉ​−ΘXˉ, the model (6.7) reduces to Y c = Θ X c + ϵ \mathcal Y_c = Θ \mathcal X_c + \epsilon Yc​=ΘXc​+ϵ .



依据

Kronecker products

从上面的分析可知,multivariate regression没什么用
In its basic classical formulation, therefore, we see that multivariate regression is a procedure that has no true multivariate content. That is, there
is no reason to create specialized software to carry out a multivariate regression of Y on X when the same result can more easily be obtained by using existing multiple regression routines. This is one reason why many books on multivariate analysis do not contain a separate chapter on multivariate regression and also why the topics of multiple regression and multivariate regression are so often confused with each other.

The Random-X Case



Multivariate Reduced-Rank Regression

Most applications of reduced-rank regression have been directed toward problems in time series (time domain and frequency domain) and econometrics. This development has led to the introduction of the related topic of cointegration into the econometric literature.

下面给出Reduced-Rank Regression的模型


于是我们将矩阵C分解

为了求得最优解我们需要一个定理

简单来说


上面第二个等式好像中间两个是减号

RRR的应用

摘自:《Modern Multivariate Statistical Techniques》

reduced rank regression model相关推荐

  1. R语言caret包构建机器学习回归模型(regression model)、使用DALEX包进行模型解释分析、特征重要度、偏依赖分析等

    R语言caret包构建机器学习回归模型(regression model).使用DALEX包进行模型解释分析.特征重要度.偏依赖分析等 目录

  2. 回归模型(regression model)有哪些指标?如何计算回归指标(regression metrics)?如何可视化他们?

    回归模型(regression model)有哪些指标?如何计算回归指标(regression metrics)?如何可视化他们? metrics.explained_variance_score(y ...

  3. Logistic Regression Model

    Logistic Regression Model Cost Function Logistic Function will not be a convex function, it will cau ...

  4. 二项逻辑回归模型(logistic regression model)

    Binary logistic regression model 是分类模型,由概率分布P(Y∣X)P(Y|X)P(Y∣X)计算,是参数化的Logistic分布 先概述一下这个模型的条件概率分布 P( ...

  5. 线性回归模型(Linear regression model)

    绪论 在开始之前,先大致说明下机器学习中的一些大的概念及分类. 机器学习中的学习算法很多,包括线性回归.逻辑回归.随机森林.决策树.聚类.贝叶斯等等.但是这些学习算法可以归为两类,即监督学习(supe ...

  6. 高斯过程分类 matlab,matlab 高斯过程回归模型 matlab Gaussian process regression model

    [实例简介] 高斯过程回归及分类的代码,内容全,有实例,注释清晰.包括分类系列和预测回归系列,值得感兴趣的同学学习借鉴.里面有对应的数据和demo程序,程序可运行,MATLAB2014a下测试通过,其 ...

  7. r语言的MASS包干什么的_R语言常用包汇总

    转载于:https://blog.csdn.net/sinat_26917383/article/details/50651464?locationNum=2&fps=1 一.一些函数包大汇总 ...

  8. CCF A类会议或期刊----回归相关论文

    工具代码位置 期待您的PR 会议/期刊 论文 sigmod2020 PrIU: A Provenance-Based Approach for Incrementally Updating Regre ...

  9. r语言 tunerf函数_R语言︱常用统计方法包 机器学习包(名称、简介)

    一.一些函数包大汇总 转载于:http://www.dataguru.cn/thread-116761-1-1.html 时间上有点过期,下面的资料供大家参考 基本的R包已经实现了传统多元统计的很多功 ...

最新文章

  1. [Leetcode] Max Area of Island 最大岛屿面积
  2. 调试机械臂一体化控制电路:STM32F103控制器初步调试
  3. 用户在购买机器之后不知如何正确保养笔记本电脑
  4. python数组随机打乱_对Python random模块打乱数组顺序的实例讲解
  5. leetcode LCP 19. 秋叶收藏集(dp)
  6. 机器学习中的数学--数学知识复习
  7. cuSPARSE库:(十二)cusparseDestroyMatDescr()
  8. 对于tf.nn.sparse_softmax_cross_entropy_with_logits(logits=y,labels=tf.argmax(y_,1))的研究
  9. java opencv 计算滑块横移距离
  10. 【计算机毕业设计】疫情社区管理系统的设计与实现
  11. troublemaker中文谐音_trouble maker的中文音译歌词 希望准确的遵循原来的发音。。。不是中文歌词和韩语歌词!要的是音译啊 注意...
  12. 学校固定资产计算机台账,(一)固定资产台账.doc
  13. minio服务报403错误
  14. 黑鲨会升级鸿蒙吗,黑鲨4首批用户评价已出炉,不吹不黑,优缺点都很明显!...
  15. 2022-2027年中国人力资源外包服务行业市场全景评估及发展战略规划报告
  16. 增强学习和OpeAI Gym的介绍:基础增强学习问题的演示
  17. 安装golang出现 Hey! The program ‘go‘ is currently not installed. You can install it by typing...
  18. Spring Security(12)——Remember-Me功能
  19. 如何学习自动化测试?一文4从个方面带你入门自动化测试【建议收藏】
  20. 看互联网将如何颠覆这17个传统行业

热门文章

  1. 创业者是否值得投入做微信小程序(应用号)
  2. 初学安卓:安卓小游戏之2048
  3. 金在中:心情不好的时候张根硕会陪我玩
  4. 微信小程序实现搜索功能
  5. GEE学习笔记【5】GEE的数据类型之Image Collection,卫星数据,矩阵
  6. 中国铁路警方严打“盗抢骗” 去年12月至今打掉团伙85个
  7. ps的简单用法和网页F12
  8. Java去除字符串中的空格以及特殊符号
  9. cmd运行Java jar包
  10. 阿里百川即时通讯的介绍