已知兩個高斯分布及他們的關係,如何求條件期望?
下面h和x是兩個符合高斯分布的變數,分布函數已給出,並且x=Ph,請問最終的條件期望E(h|x)是如何得到的?
我試著推理如下,但沒法繼續了
謝邀。
Again I am passionate about this kind of technicalities.
I do not redefine the variables in the question. But let me define something more. Note that has one more dimension than , where except the first dimension of , . For later convenience, I define , where is the first dimension and contains the rest.
On the other hand, given calculated above, one can easily show that its inverse is given by
.(Note: I found this out by clumsily going through the calculation below, but verification is easy. And this is actually the main obstacle of this exercise.)First we define the joint probability distribution function . We know that obeys the Gaussian distribution, and is a linear superposition of the elements in . It is the most appropriate to use Dirac delta function to describe this:
.As we know, the probability distribution for is in the form
.However, we can also get by integrating out :,where you can see the determinant of is .Then the conditional probability can be calculated:
In the above, we simply exploit the definition of conditional probability, and integrate over all with the Dirac delta function. There are a lot of algebra involved which the readers can diligently verify on their own. Continuing the calculation givesA careful operation of algebra can show that this is exactly equal to .I have skipped many details, but readers can verify this on their own.
P.S.: For me, the most difficult part is to find the inverse of , which I found by calculating the integrals above and got the matrix elements, and then verify it by multiplying it by itself to see if it gives an identity matrix. The difference in the number of dimensions of and does impose some inconvenience. However, it is more about menial algebra instead of an intellectually challenging problems. If you want to have a taste without much algebraic operation, take the fewer dimensions and do the calculation first using Mathematica to get a feeling.
首先先安利下 The Matrix Cookbook,從此媽媽再也不用擔心我的矩陣推導了!
http://www2.imm.dtu.dk/pubdb/views/edoc_download.php/3274/pdf/imm3274.pdf將向量h與x合在一起看做一個符合多元高斯分布的隨機向量(h,x),依據cookbook中公式(353)
可得 (1)再由cookbook中公式(314):
則由可得代入式(1)既得結果 (2)
Remark:注意這裡的P是一般的實矩陣,其維度由h與x的維度決定,所以P不一定可逆,甚至不是方陣,所以式(2)小括弧不能再展開繼續推導了,但如果P可逆會如何?那繼續推導唄
繞了半天。。。再從貝葉斯角度來看的話:P不是方陣但行滿秩,求M-P廣義逆,有
若列滿秩,有這個是線性方程組中的極小範數最小二乘解,其實也是隨機向量符合多元高斯分布假設下的極大似然估計,而對應題主問題P應為行滿秩,觀察行滿秩的形式與式(2)頗為相似,這是由於加入了h的先驗分布,而對極大似然估計進行了對先驗分布的修正,求得的對應的最大後驗估計。推薦閱讀:
※深度學習乃至機器學習和凸論有什麼本質聯繫?
※怎麼看待現在一些論文提出的新點子就只在mnist或者cifar上跑跑實驗?
※如何評價CVPR best paper simGAN?
※如何解讀論文《Generalization in Deep Learning》?
※如何開發一個特定領域的自動問答機器人(Chat Bot)?