Skip to main content

modeling - Model fitting to noisy data with a custom minimization function


I'm looking into fitting some data with Mathematica. I've got my head around how NonlinearModelFit works (I've been using the Levenberg-Marquardt algorithm for some other work).


But my data this time is Poisson distributed, and I want to see if using the appropriate MLE for Poisson data is better for my scenario than nonlinear least squares fitting.


According to the paper Efficient Levenberg-Marquardt minimization of the maximum likelihood estimator for Poisson deviates, then the minimization for least-squares fitting, for data yi and the model fi is


χ2=2N∑i=1(fi−yi)2σ2i


whereas for Poisson distributed data according to the paper, the minimization is


χ2=2N∑i=1fi−yi−2N∑i=1,y≠0yiln(fiyi)


Is it possible to run a model-fitting in Mathematica using this minimization? And can a (modified?) Levenberg-Marquardt algorithm still be used?



Edit


There's an associated Nature Methods letter at http://dx.doi.org/10.1038/nmeth0510-338, with a revised version of the above link: http://www.nature.com/nmeth/journal/v7/n5/extref/nmeth0510-338-S1.pdf (courtesy of @belisarius)


Update #1


So here's the sort of data/model I'm looking to fit: the sum of two (or more) Gaussians, which may sometimes overlap as shown in the example below.


The amount of Poisson noise is deliberately significant as I'm dealing with very low counts. I've only posted a one-dimensional example here, but the data is in 2D, so there are more variables (x,y,means,heights,sigma...). I'm happy with using NonlinearModelFit to solve the problem, but I'm curious about dealing with the Poisson noise "more appropriately".


twoGaussianFunction[x_, A1_, sigma1_, mean1_, A2_, sigma2_, mean2_] := 
A1 Exp[-((x - mean1)^2/(2 sigma1^2))] +
A2 Exp[-((x - mean2)^2/(2 sigma2^2))];

cleandata = Table[twoGaussianFunction[i, 10, 10, 30, 10, 10, 60], {i, 0, 100}];


noisydata = RandomVariate[PoissonDistribution[0.5 #]] & /@ cleandata;
ListLinePlot[{cleandata, noisydata}, PlotRange -> Full]

enter image description here



Answer



To answer my own question, I went with the suggestion by Oleksandr R:



Here I'm suggesting that you just write it directly as a minimization problem. In 4700, Ajasja and I did exactly this to perform a least squares fit using a custom minimizer, but you can of course minimize anything you want. By the way, you may also like to see GeneralizedLinearModelFit, which can be used for fitting Poisson-distributed data directly.




I ended up using NMinimize for my problem, rather than FindMinimum, but writing it as a minimisation problem was the solution. I used my Gaussian model as fi to solve for the data yi this:


χ2=2N∑i=1fi−yi−2N∑i=1,y≠0yiln(fiyi)


as intended, with decent results.


First, the data:


twoGaussianFunction[x_, A1_, sigma1_, mean1_, A2_, sigma2_, mean2_] :=
A1 Exp[-((x - mean1)^2/(2 sigma1^2))] +
A2 Exp[-((x - mean2)^2/(2 sigma2^2))];

cleandata =
Table[twoGaussianFunction[i, 2, 3, 20, 2, 3, 30], {i, 0, 50}];


noisydata = RandomVariate[PoissonDistribution[2 #]]/2 & /@ cleandata;

ListLinePlot[{cleandata, noisydata}, PlotRange -> Full, PlotLegends -> {"Original", "Noisy"}]

enter image description here


Then the minimisation function:


minimizeFunction[A1guess_, sigma1guess_, mean1guess_, A2guess_, 
sigma2guess_, mean2guess_] :=
2 Sum[twoGaussianFunction[i, A1guess, sigma1guess, mean1guess,

A2guess, sigma2guess, mean2guess] - noisydata[[i]], {i, 50}] -
2 Sum[If[noisydata[[i]] == 0., 0,noisydata[[i]]*
Log[twoGaussianFunction[i, A1guess, sigma1guess, mean1guess,
A2guess, sigma2guess, mean2guess]/noisydata[[i]]]], {i,50}];

Followed by NMinimize:


bestfit = 
NMinimize[{minimizeFunction[a, b, c, d, e, f],
a > 0 && b > 0 && c > 0 && d > 0 && e > 0 && f > 0},
{{a, 1, 3},

{b, 2, 4},
{c, 15, 25},
{d, 1, 3},
{e, 2, 4},
{f, 25, 35}},
Method -> "NelderMead",
MaxIterations -> 100
];

cleaneddata =

Table[twoGaussianFunction[i, a /. Last[bestfit], b /. Last[bestfit],
c /. Last[bestfit], d /. Last[bestfit], e /. Last[bestfit],
f /. Last[bestfit]], {i, 1, 50}];

ListLinePlot[{cleandata, cleaneddata}, PlotRange -> Full,
PlotLegends -> {"Original", "Fitted"}]

enter image description here


There's probably room for improvement in the way I've implemented it - certainly for speed, perhaps using Parallelize in the minimisation function? (That was my first thought). I've still got to test it fully against the standard least-squares method though...


Comments

Popular posts from this blog

functions - Get leading series expansion term?

Given a function f[x] , I would like to have a function leadingSeries that returns just the leading term in the series around x=0 . For example: leadingSeries[(1/x + 2)/(4 + 1/x^2 + x)] x and leadingSeries[(1/x + 2 + (1 - 1/x^3)/4)/(4 + x)] -(1/(16 x^3)) Is there such a function in Mathematica? Or maybe one can implement it efficiently? EDIT I finally went with the following implementation, based on Carl Woll 's answer: lds[ex_,x_]:=( (ex/.x->(x+O[x]^2))/.SeriesData[U_,Z_,L_List,Mi_,Ma_,De_]:>SeriesData[U,Z,{L[[1]]},Mi,Mi+1,De]//Quiet//Normal) The advantage is, that this one also properly works with functions whose leading term is a constant: lds[Exp[x],x] 1 Answer Update 1 Updated to eliminate SeriesData and to not return additional terms Perhaps you could use: leadingSeries[expr_, x_] := Normal[expr /. x->(x+O[x]^2) /. a_List :> Take[a, 1]] Then for your examples: leadingSeries[(1/x + 2)/(4 + 1/x^2 + x), x] leadingSeries[Exp[x], x] leadingSeries[(1/x + 2 + (1 - 1/x...

mathematical optimization - Minimizing using indices, error: Part::pkspec1: The expression cannot be used as a part specification

I want to use Minimize where the variables to minimize are indices pointing into an array. Here a MWE that hopefully shows what my problem is. vars = u@# & /@ Range[3]; cons = Flatten@ { Table[(u[j] != #) & /@ vars[[j + 1 ;; -1]], {j, 1, 3 - 1}], 1 vec1 = {1, 2, 3}; vec2 = {1, 2, 3}; Minimize[{Total@((vec1[[#]] - vec2[[u[#]]])^2 & /@ Range[1, 3]), cons}, vars, Integers] The error I get: Part::pkspec1: The expression u[1] cannot be used as a part specification. >> Answer Ok, it seems that one can get around Mathematica trying to evaluate vec2[[u[1]]] too early by using the function Indexed[vec2,u[1]] . The working MWE would then look like the following: vars = u@# & /@ Range[3]; cons = Flatten@{ Table[(u[j] != #) & /@ vars[[j + 1 ;; -1]], {j, 1, 3 - 1}], 1 vec1 = {1, 2, 3}; vec2 = {1, 2, 3}; NMinimize[ {Total@((vec1[[#]] - Indexed[vec2, u[#]])^2 & /@ R...

How to remap graph properties?

Graph objects support both custom properties, which do not have special meanings, and standard properties, which may be used by some functions. When importing from formats such as GraphML, we usually get a result with custom properties. What is the simplest way to remap one property to another, e.g. to remap a custom property to a standard one so it can be used with various functions? Example: Let's get Zachary's karate club network with edge weights and vertex names from here: http://nexus.igraph.org/api/dataset_info?id=1&format=html g = Import[ "http://nexus.igraph.org/api/dataset?id=1&format=GraphML", {"ZIP", "karate.GraphML"}] I can remap "name" to VertexLabels and "weights" to EdgeWeight like this: sp[prop_][g_] := SetProperty[g, prop] g2 = g // sp[EdgeWeight -> (PropertyValue[{g, #}, "weight"] & /@ EdgeList[g])] // sp[VertexLabels -> (# -> PropertyValue[{g, #}, "name"]...