diff --git a/paper.md b/paper.md
index 35b83c6dba148cce2ad138706a4456fdee5b5250..567e18c8c8f5f9ca4309ed638ada1e0219bfe15c 100755
--- a/paper.md
+++ b/paper.md
@@ -102,14 +102,18 @@ An example of such a polygonal remeshing method is one where the polygons align
 
 #### We aim to sample low dimensional low to intermediate cost functions in parallel.
 The general algorithm that we describe in this paper works best for low to intermediary cost functions.
-The point suggestion step happens in a single sequential process while the function execution can happen in parallel.
-This means that $t_\textrm{function} / N_\textrm{workers} \gg t_\textrm{suggest}$, in order to benefit from our adaptive sampling algorithm.
-Very fast functions can be calculated on a dense grid and extremely slow functions might benefit from full scale Bayesian optimization, nonetheless a large class of functions is inside the right regime for Adaptive to be beneficial.
+The point suggestion step happens in a single sequential process while the function executions can be in parallel.
+This means that to benefit from an adaptive sampling algorithm $t_\textrm{function} / N_\textrm{workers} \gg t_\textrm{suggest}$ must hold, here $t_\textrm{function}$ is the average function execution time, $N_\textrm{workers}$ the number of parallel processes, and $t_\textrm{suggest}$ the time it takes to suggest a new point.
+Very fast functions can be calculated on a dense grid and extremely slow functions might benefit from full-scale Bayesian optimization where $t_\textrm{suggest}$ is large, nonetheless, a large class of functions is inside the right regime for Adaptive to be beneficial.
 Further, because of the curse of dimensionality---the sparsity of space in higher dimensions---our local algorithm works best in low dimensional space; typically calculations that can reasonably be plotted, so with 1, 2, or 3 degrees of freedom.
 
 #### We propose to use a local loss function as a criterion for choosing the next point.
+To minimize $t_\textrm{suggest}$ and equivalently make the point suggestion algorithm as fast as possible, we propose to assign a loss to each interval.
+This loss is determined only by the function values of the points inside that interval and optionally of its neighboring intervals too.
+The local loss function values then serve as a criterion for choosing the next point by the virtue of choosing a new candidate point inside the interval with the maximum loss.
+This means that upon adding new data points only the intervals near the new point needs to have their loss value updated.
 
-#### As an example interpoint distance is a good loss function in one dimension.
+#### As an example the interpoint distance is a good loss function in one dimension.
 <!-- Plot here -->
 
 #### In general local loss functions only have a logarithmic overhead.