如何在python中使用scipy.optimize中的lesssq函数来将直线和二次线拟合到数据集x和y



如何使用 scipy.optimize 的 minimumsq 函数将直线和二次线拟合到下面的数据集中?我知道如何使用polyfit来做到这一点。但是我需要使用 minimumsq 函数。

以下是 x 和 y 数据集:

x: 1.0,2.5,3.5,4.0,1.1,1.8,2.2,3.7
y: 6.008,15.722,27.130,33.772,5.257,9.549,11.098,28.828

有人可以帮我吗?

minimumsq() 方法查找最小化误差函数的参数集(yExperimental 和 yFit 之间的差异)。我使用元组传递线性和二次拟合的参数和 lambda 函数。

Minimumsq 从第一次猜测(参数的初始元组)开始,并尝试最小化误差函数。最后,如果 minimumsq 成功,它将返回最适合数据的参数列表。(我打印了看)。我希望它有效此致敬意

from scipy.optimize import leastsq
import numpy as np
import matplotlib.pyplot as plt

def main():
   # data provided
   x=np.array([1.0,2.5,3.5,4.0,1.1,1.8,2.2,3.7])
   y=np.array([6.008,15.722,27.130,33.772,5.257,9.549,11.098,28.828])
   # here, create lambda functions for Line, Quadratic fit
   # tpl is a tuple that contains the parameters of the fit
   funcLine=lambda tpl,x : tpl[0]*x+tpl[1]
   funcQuad=lambda tpl,x : tpl[0]*x**2+tpl[1]*x+tpl[2]
   # func is going to be a placeholder for funcLine,funcQuad or whatever 
   # function we would like to fit
   func=funcLine
   # ErrorFunc is the diference between the func and the y "experimental" data
   ErrorFunc=lambda tpl,x,y: func(tpl,x)-y
   #tplInitial contains the "first guess" of the parameters 
   tplInitial1=(1.0,2.0)
   # leastsq finds the set of parameters in the tuple tpl that minimizes
   # ErrorFunc=yfit-yExperimental
   tplFinal1,success=leastsq(ErrorFunc,tplInitial1[:],args=(x,y))
   print " linear fit ",tplFinal1
   xx1=np.linspace(x.min(),x.max(),50)
   yy1=func(tplFinal1,xx1)
   #------------------------------------------------
   # now the quadratic fit
   #-------------------------------------------------
   func=funcQuad
   tplInitial2=(1.0,2.0,3.0)
   tplFinal2,success=leastsq(ErrorFunc,tplInitial2[:],args=(x,y))
   print "quadratic fit" ,tplFinal2
   xx2=xx1
   yy2=func(tplFinal2,xx2)
   plt.plot(xx1,yy1,'r-',x,y,'bo',xx2,yy2,'g-')
   plt.show()
if __name__=="__main__":
   main()
from scipy.optimize import leastsq
import scipy as sc
import numpy as np
import matplotlib.pyplot as plt

optimize.curve_fit代码更简单,无需定义残差(错误)函数。

fig, ax = plt.subplots ()
# data
x=np.array([1.0,2.5,3.5,4.0,1.1,1.8,2.2,3.7])
y=np.array([6.008,15.722,27.130,33.772,5.257,9.549,11.098,28.828])
# modeling functions
def funcLine(x, a,b):
    return a*x+b
def funcQuad(x, a, b, c):
    return a*x**2+b*x+c
# optimize constants for the linear function
constantsLine, _ = sc.optimize.curve_fit (funcLine, x, y)
X=np.linspace(x.min(),x.max(),50)
Y1=funcLine(X, *constantsLine)
# optimize constants for the quadratic function
constantsQuad, _ = sc.optimize.curve_fit (funcQuad, x, y)
Y2=funcQuad(X,*constantsQuad)
plt.plot(X,Y1,'r-',label='linear approximation')
plt.plot(x,y,'bo',label='data points')
plt.plot(X,Y2,'g-', label='quadratic approximation')
matplotlib.pylab.legend ()
ax.set_title("Nonlinear Least Square Problems", fontsize=18)
plt.show()

这是一个超级简单的例子。想象一个抛物面,就像一个碗,边长得像抛物线。如果我们将底部放在坐标 (x, y) = (a, b) 处,然后在 x 和 y 的所有值上最小化抛物面的高度 - 我们期望最小值为 x=a 和 y=b。下面是可以执行此操作的代码。

import random
from scipy.optimize import least_squares

a, b = random.randint(1, 1000), random.randint(1, 1000)
print("Expect", [a, b])
def f(args):
    x, y = args
    return (x-a)**2 + (y-b)**2
x0 = [-1, -3]
result = least_squares(fun=f, x0=x0)
print(result.x)

最新更新