diff --git a/paper.md b/paper.md index 567e18c8c8f5f9ca4309ed638ada1e0219bfe15c..d3b18bf7f61f8397154ca22c287dca66e0b898ca 100755 --- a/paper.md +++ b/paper.md @@ -52,7 +52,7 @@ Each candidate point has a loss $L$ indicated by the size of the red dots. The candidate point with the largest loss will be chosen, which in this case is the one with $L_{1,2}$. ](figures/loss_1D.pdf){#fig:loss_1D} -{#fig:adaptive_vs_grid} @@ -114,7 +114,10 @@ The local loss function values then serve as a criterion for choosing the next p This means that upon adding new data points only the intervals near the new point needs to have their loss value updated. #### As an example the interpoint distance is a good loss function in one dimension. -<!-- Plot here --> +An example of such a loss function for a one-dimensional function is the interpoint distance, such as in Fig. @fig:loss_1D. +This loss will suggest to sample a point in the middle of an interval with the largest Euclidean distance and thereby ensure the continuity of the function. +A more complex loss function that also takes the first neighboring intervals into account, is one that adds more points where the second derivative (or curvature) is the highest. +Figure @fig:adaptive_vs_grid shows a comparison between this loss and a function that is sampled on a grid. #### In general local loss functions only have a logarithmic overhead.