From a15762112e352b6a209dba5a977c8090fd1dbd6c Mon Sep 17 00:00:00 2001 From: Bas Nijholt <basnijholt@gmail.com> Date: Thu, 12 Sep 2019 17:21:47 +0200 Subject: [PATCH] updates --- paper.md | 7 +++++-- 1 file changed, 5 insertions(+), 2 deletions(-) diff --git a/paper.md b/paper.md index 567e18c..d3b18bf 100755 --- a/paper.md +++ b/paper.md @@ -52,7 +52,7 @@ Each candidate point has a loss $L$ indicated by the size of the red dots. The candidate point with the largest loss will be chosen, which in this case is the one with $L_{1,2}$. ](figures/loss_1D.pdf){#fig:loss_1D} -{#fig:adaptive_vs_grid} @@ -114,7 +114,10 @@ The local loss function values then serve as a criterion for choosing the next p This means that upon adding new data points only the intervals near the new point needs to have their loss value updated. #### As an example the interpoint distance is a good loss function in one dimension. -<!-- Plot here --> +An example of such a loss function for a one-dimensional function is the interpoint distance, such as in Fig. @fig:loss_1D. +This loss will suggest to sample a point in the middle of an interval with the largest Euclidean distance and thereby ensure the continuity of the function. +A more complex loss function that also takes the first neighboring intervals into account, is one that adds more points where the second derivative (or curvature) is the highest. +Figure @fig:adaptive_vs_grid shows a comparison between this loss and a function that is sampled on a grid. #### In general local loss functions only have a logarithmic overhead. -- GitLab