畢業(yè)設(shè)計(jì)論文 外文文獻(xiàn)翻譯 數(shù)學(xué)專(zhuān)業(yè)

上傳人:仙*** 文檔編號(hào):28873084 上傳時(shí)間:2021-09-17 格式:DOC 頁(yè)數(shù):14 大?。?71.50KB
收藏 版權(quán)申訴 舉報(bào) 下載
畢業(yè)設(shè)計(jì)論文 外文文獻(xiàn)翻譯 數(shù)學(xué)專(zhuān)業(yè)_第1頁(yè)
第1頁(yè) / 共14頁(yè)
畢業(yè)設(shè)計(jì)論文 外文文獻(xiàn)翻譯 數(shù)學(xué)專(zhuān)業(yè)_第2頁(yè)
第2頁(yè) / 共14頁(yè)
畢業(yè)設(shè)計(jì)論文 外文文獻(xiàn)翻譯 數(shù)學(xué)專(zhuān)業(yè)_第3頁(yè)
第3頁(yè) / 共14頁(yè)

下載文檔到電腦,查找使用更方便

15 積分

下載資源

還剩頁(yè)未讀,繼續(xù)閱讀

資源描述:

《畢業(yè)設(shè)計(jì)論文 外文文獻(xiàn)翻譯 數(shù)學(xué)專(zhuān)業(yè)》由會(huì)員分享,可在線閱讀,更多相關(guān)《畢業(yè)設(shè)計(jì)論文 外文文獻(xiàn)翻譯 數(shù)學(xué)專(zhuān)業(yè)(14頁(yè)珍藏版)》請(qǐng)?jiān)谘b配圖網(wǎng)上搜索。

1、 重 慶 理 工 大 學(xué) 文 獻(xiàn) 翻 譯 二級(jí)學(xué)院 數(shù)學(xué)與統(tǒng)計(jì)學(xué)院 譯文: 摘自:The Newton Raphson Algorithm for FunctionOptimization Kevin Quinn Assistant Professor Department of Political Science and The C

2、enter for Statistics and the Social Sciences Box 354322, Padelford Hall University of Washington Seattle, WA 98195-4322 October 25, 2001 一、引言 通過(guò)這個(gè)課程的學(xué)習(xí)我們將有興趣計(jì)算最大似然估計(jì)(極大似然估計(jì))。例如我們常常觀察到的復(fù)雜的非線性函數(shù)的數(shù)據(jù)。因此,通過(guò)我們的計(jì)算封閉形式的去表達(dá)這極大似然估計(jì)的形式一般是不會(huì)存在模型的 牛頓拉夫森算法是一個(gè)迭代的過(guò)程,可用于計(jì)算出極大似然估計(jì)。其背后的算法的基本思想的內(nèi)容。首先,圍繞一些初步的參數(shù)

3、值構(gòu)造一個(gè)二次近似逼近的有利函數(shù)(希望能接近最大似然估計(jì))。其次是,調(diào)整參數(shù)值讓其最大限度地提高二次近似。此過(guò)程再不斷的重復(fù)進(jìn)行,直到參數(shù)值穩(wěn)定。 這就說(shuō)明開(kāi)始容易想象出一個(gè)函數(shù)遇到最大化的一個(gè)變量。在這種情況下開(kāi)發(fā),我們轉(zhuǎn)而更為一般的情況下最大化的一個(gè)變量k的函數(shù) 。 二、牛頓拉夫森算法求變量1的函數(shù)的最大值 2、1泰勒系列的逼近問(wèn)題 牛頓拉夫遜算法的第一部分的發(fā)展是設(shè)計(jì)一個(gè)近似函數(shù)表示似然函數(shù)就可以很容易的最大化的分析。要做到這一點(diǎn),我們需要利用泰勒定理。 定理1(泰勒定理(1維))假設(shè)函數(shù)f次可微的開(kāi)區(qū)間I上的,對(duì)于任意的一點(diǎn)到在I區(qū)間上存在的一點(diǎn)在到上例如:

4、 . (1) 他可以表示成為從到的方程的高階項(xiàng)從到更快于從到。這就意味著(最小值) 這被稱(chēng)作一階泰勒的近似函數(shù)f在x上的,小的值可以構(gòu)建一個(gè)更準(zhǔn)確的逼近函數(shù): 請(qǐng)注意第一階泰勒的近似可以重寫(xiě)為被稱(chēng)為一個(gè)二階泰勒的近似函數(shù)f在上的值如: 從到.這凸顯一個(gè)事實(shí),即一階泰勒的近似的線性函數(shù)在上的。同樣的,二階泰勒的近似值可以被改寫(xiě)成為: 當(dāng),,且。這凸顯出的一個(gè)事實(shí),即是二階泰勒近似值是在上的第二階多項(xiàng)式。 2、2查找到的其最大值的二階多項(xiàng)式 假設(shè)出我們想要找出的值能最大化的 首先,我們計(jì)算出的一階導(dǎo)數(shù)的函數(shù)為: 我們了

5、解到這,當(dāng)?shù)闹凳菚r(shí),其中函數(shù)的值達(dá)到最大,換句話說(shuō),我們都知道 求解我們發(fā)現(xiàn)。第二階的條件就是。這意味著 的值將是最大無(wú)論什么時(shí)候當(dāng). 2、3牛頓拉夫森的算法 假設(shè)我們想要找到的值當(dāng)最大化的二次連續(xù)可微的函數(shù)的值。 記得 當(dāng),且。這就意味著: 一階條件的(記為)值能最大化就是是: 這就意味著。換而言之就是, 在的函數(shù)值能最大化的二階泰勒近似值為函數(shù) 考慮到這一點(diǎn),我們可以指定用于一維的函數(shù)優(yōu)化問(wèn)題的牛頓拉夫森算法。 算法2、1:牛頓拉夫森一維的(,,公差) 發(fā)表評(píng)論:找出求的值能最大化的函數(shù): 當(dāng) Do

6、 回到 注意事項(xiàng):注意牛頓拉夫森算法,不檢查的二階的必要條件為是最大化。這就意味著,如果你給一個(gè)錯(cuò)的開(kāi)始x的值的算法,你可能最終是最小的,而不是一個(gè)最大的。 2、4例如:計(jì)算二項(xiàng)式抽樣模型的極大似然估計(jì) 看到牛頓拉夫森算法的工程實(shí)踐中如何讓看一個(gè)簡(jiǎn)單的示例,二項(xiàng)式抽樣與分析解的簡(jiǎn)單的模型。 我們的對(duì)數(shù)似然函數(shù)是: 當(dāng)為樣本容量時(shí),就是成功的次數(shù),是一個(gè)取得成功的概率。一階導(dǎo)數(shù)對(duì)數(shù)似然函數(shù)是: 二階導(dǎo)數(shù)對(duì)數(shù)似然函數(shù)就是: 解析,我們知道的最大似然估計(jì)是:。 舉一個(gè)例子,假設(shè)且。解析,我們知道的最大似然估計(jì)是。讓我們來(lái)看看如何在這種情況下解出牛

7、頓拉夫森算法。 我們首先設(shè)置公差級(jí)別。在這種情況下的,讓將它設(shè)置為0.01(在實(shí)踐中你可能想要的東西更接近0.00001)。下一步,我們初始猜測(cè)的最大似然估計(jì)(記為)。假設(shè)。的這是在絕對(duì)值大于0.01的公差。 因此我們?cè)O(shè)置為: 。 現(xiàn)在我們計(jì)算出,它仍然是在絕對(duì)值大于的公差。因此我們?cè)O(shè)置為: 是約等于是絕對(duì)值小于的公差,這樣我們就可以停止了。牛頓拉夫森算法返回pi的的值等于到接近0.3994,這是合理的分析值0.40。請(qǐng)注意,我們可以設(shè)置的容忍水平接近的牛頓拉夫森過(guò)程更準(zhǔn)確(機(jī)器精密度范圍內(nèi))。 三、牛頓拉夫森算法求最大的變量的函數(shù) 3、1泰勒級(jí)數(shù)逼近問(wèn)題維度

8、 考慮函數(shù)至少有兩次的連續(xù)可微。假設(shè)且。然后給出一階泰勒近似值在函數(shù)上的一個(gè)被寫(xiě)為: 給出二階泰勒近似值在函數(shù)上的一個(gè)被寫(xiě)為: 當(dāng)是梯度(一階導(dǎo)數(shù)的向量)的函數(shù)在上時(shí),且是 Hessian矩陣(第二衍生矩陣)在函數(shù)屬于上時(shí)。 3、2找到最大值的變量的二階多項(xiàng)式 考慮 當(dāng)是一個(gè)標(biāo)量,和是關(guān)于K-向量,且是一個(gè)的對(duì)稱(chēng)矩陣,負(fù)正定矩陣。這的梯度在上表示為: 我們知道,由于最大化的值滿(mǎn)足能最大化的梯度將是一個(gè)零矢量, 求解 我們找出結(jié)果如: 由于被認(rèn)為是負(fù)定的,而且我們知道這就是最大的。 3、

9、3在維度的牛頓拉夫森算法 假設(shè)我們要找出的最大限度地提高二次連續(xù)可微函數(shù)。 記得 當(dāng)且。請(qǐng)注意矩陣將是對(duì)稱(chēng)的,這就意味著是: 再一次,最大值的一階條件就是: 這就意味著: 換句話說(shuō)就是,向量能最大化的在的二階泰勒近似值為函數(shù): 考慮到這一點(diǎn),我們就可以指定的k維函數(shù)優(yōu)化問(wèn)題的牛頓拉夫森算法。 算法研究3、1:牛頓拉夫森算法的KD(,,公差) 發(fā)表評(píng)論:求關(guān)于的值的的最大函數(shù)。 當(dāng) Do 回到()。 譯文: 摘自: The Ne

10、wton Raphson Algorithm for FunctionOptimization Kevin Quinn Assistant Professor Department of Political Science and The Center for Statistics and the Social Sciences Box 354322, Padelford Hall University of Washington Seattle, WA 98195-4322 October 25, 2001 1 Introduction T

11、hroughout this course we will be interested in calculating maximum likelihood estimate(MLEs). Such estimates are often extremely complicated nonlinear functions of the observed data. As a result, closed form expressions for the MLEs will generally not exist for the models we are working with. Th

12、e Newton Raphson algorithm is an iterative procedure that can be used to calculate MLEs. The basic idea behind the algorithm is the following. First, construct a quadratic approximation to the function of interest around some initial parameter value (hopefully close to the MLE). Next, adjust the par

13、ameter value to that which maximizes the quadratic approximation. This procedure is iterated until the parameter values stabilize. These notes begin with the easy to visualize case of maximizing a function of one variable. After this case is developed, we turn to the more general case of maximiz

14、ing a function of variables. 2 The Newton Raphson Algorithm for Finding the Maximum of a Function of 1 Variable 2.1 Taylor Series Approximations The first part of developing the Newton Raphson algorithm is to devise a way to approximate the likelihood function with a function that can b

15、e easily maximized analytically. To do this we need to make use of Taylor’s Theorem. Theorem 1 (Taylor’s Theorem (1 Dimension)). Suppose the function f is times differentiable on an open interval I. For any points and in I there exists a point between and such that . (1) It can be sho

16、wn that as goes to the higher order terms in equation go to 0 much faster than goes to . This means that (for small values of ) This is referred to as a first order Taylor approximation of f at . A more accurate approximation to can be constructed for small values of as: This is known

17、as a second order Taylor approximation of f at Note that the first order Taylor approximation can be rewritten as: where and . This highlights the fact that the first order Taylor approximation is a linear function in . Similarly, the second order Taylor approximation can be rewritten as: W

18、here , and . This highlights the fact that the second order Taylor approximation is a second order polynomial in 2.2 Finding the Maximum of a Second Order Polynomial Suppose we want to find the value of that maximizes First, we calculate the first derivative of : We know that , where

19、 is the value of at which f attains its maximum. In other words, we know that Solving for we find that . The second order condition is . This implies that will be a maximum whenever . 2.3 The Newton Raphson Algorithm Suppose we want to find the value of that maximizes some twice continu

20、ously differentiable function . Recall where , , and . This implies . The first order condition for the value of (denoted ) that maximizes is Which implies . In other words, the value that maximizes the second order Taylor approximation to at is With this in mind

21、we can specify the Newton Raphson algorithm for dimensional function optimization. Algorithm 2.1: NewtonRaphson1D(,,tolerance) comment: Find the value of that maximizes While do return Caution: Note that the Newton Raphson Algorithm doesn’t check the second order conditions nece

22、ssary for to be a maximizer. This means that if you give the algorithm a bad starting value for you may end up with a min rather than a max. 2.4 Example: Calculating the MLE of a Binomial Sampling Model To see how the Newton Raphson algorithm works in practice lets look at a simple example with

23、an analytical solution– a simple model of binomial sampling. Our log-likelihood function is: where is the sample size, is the number of successes, and is the probability of a success.The first derivative of the log-likelihood function is and the second derivative of the log-likelihood func

24、tion is Analytically, we know that the MLE is . For the sake of example, suppose and . Analytically, we know that the MLE is Let’s see how the Newton Raphson algorithm works in this situation. We begin by setting a tolerance level. In this case, let’s set it to (In practice you probably want

25、 something closer to ). Next we make an initial guess (denoted ) as to the MLE. Suppose . which is larger in absolute value than our tolerance of . Thus we set . Now we calculate which is still larger in absolute value than our tolerance of . Thus we set is approximately equal to which is s

26、maller in absolute value than our tolerance of so we can stop. The Newton Raphson algorithm here returns a value of pi equal to which is reasonably close to the analytical value of . Note we can make the Newton Raphson procedure more accurate (within machine precision) by setting the tolerance lev

27、el closer to . 3 The Newton Raphson Algorithm for Finding the Maximum of a Function of Variables 3.1 Taylor Series Approximations in Dimensions Consider a function that is at least twice continuously differentiable. Suppose and . Then the first order Taylor approximation to at is given

28、by and the second order Taylor approximation to f at is given by where is the gradient (vector of first derivatives) at , and is the Hessian (matrix of second derivatives) of at . 3.2 Finding the Maximum of a Second Order Polynomial in Variables Consider where is a scalar, and

29、 are k-vectors, and is a symmetric, negative definite matrix. The gradient of at is Since the gradient at the value that maximizes will be a vector of zeros we know that the maximizer satisfies Solving for we find that Since is assumed to be negative definite we know that this is a

30、maximum. 3.3 The Newton Raphson Algorithm in k Dimensions Suppose we want to find the that maximizes the twice continuously differentiable function Recall where and . Note that will be symmetric. This implies Once again, the first order condition for a maximum is which implies that

31、 In other words, the vector that maximizes the second order Taylor approximation to at is With this in mind we can specify the Newton Raphson algorithm for k-dimensional function optimization. Algorithm 3.1: NewtonRaphsonKD comment: Find the value of that maximizes While do Return()

展開(kāi)閱讀全文
溫馨提示:
1: 本站所有資源如無(wú)特殊說(shuō)明,都需要本地電腦安裝OFFICE2007和PDF閱讀器。圖紙軟件為CAD,CAXA,PROE,UG,SolidWorks等.壓縮文件請(qǐng)下載最新的WinRAR軟件解壓。
2: 本站的文檔不包含任何第三方提供的附件圖紙等,如果需要附件,請(qǐng)聯(lián)系上傳者。文件的所有權(quán)益歸上傳用戶(hù)所有。
3.本站RAR壓縮包中若帶圖紙,網(wǎng)頁(yè)內(nèi)容里面會(huì)有圖紙預(yù)覽,若沒(méi)有圖紙預(yù)覽就沒(méi)有圖紙。
4. 未經(jīng)權(quán)益所有人同意不得將文件中的內(nèi)容挪作商業(yè)或盈利用途。
5. 裝配圖網(wǎng)僅提供信息存儲(chǔ)空間,僅對(duì)用戶(hù)上傳內(nèi)容的表現(xiàn)方式做保護(hù)處理,對(duì)用戶(hù)上傳分享的文檔內(nèi)容本身不做任何修改或編輯,并不能對(duì)任何下載內(nèi)容負(fù)責(zé)。
6. 下載文件中如有侵權(quán)或不適當(dāng)內(nèi)容,請(qǐng)與我們聯(lián)系,我們立即糾正。
7. 本站不保證下載資源的準(zhǔn)確性、安全性和完整性, 同時(shí)也不承擔(dān)用戶(hù)因使用這些下載資源對(duì)自己和他人造成任何形式的傷害或損失。

相關(guān)資源

更多
正為您匹配相似的精品文檔
關(guān)于我們 - 網(wǎng)站聲明 - 網(wǎng)站地圖 - 資源地圖 - 友情鏈接 - 網(wǎng)站客服 - 聯(lián)系我們

copyright@ 2023-2025  zhuangpeitu.com 裝配圖網(wǎng)版權(quán)所有   聯(lián)系電話:18123376007

備案號(hào):ICP2024067431號(hào)-1 川公網(wǎng)安備51140202000466號(hào)


本站為文檔C2C交易模式,即用戶(hù)上傳的文檔直接被用戶(hù)下載,本站只是中間服務(wù)平臺(tái),本站所有文檔下載所得的收益歸上傳人(含作者)所有。裝配圖網(wǎng)僅提供信息存儲(chǔ)空間,僅對(duì)用戶(hù)上傳內(nèi)容的表現(xiàn)方式做保護(hù)處理,對(duì)上載內(nèi)容本身不做任何修改或編輯。若文檔所含內(nèi)容侵犯了您的版權(quán)或隱私,請(qǐng)立即通知裝配圖網(wǎng),我們立即給予刪除!