From 94cedc0fb20c4507968f4344c734830ed2ab34ce Mon Sep 17 00:00:00 2001
From: Bas Nijholt <basnijholt@gmail.com>
Date: Mon, 30 Sep 2019 17:09:59 +0200
Subject: [PATCH] changes

---
 paper.md | 4 ++--
 1 file changed, 2 insertions(+), 2 deletions(-)

diff --git a/paper.md b/paper.md
index 3d0e79c..2ba7c2a 100755
--- a/paper.md
+++ b/paper.md
@@ -30,7 +30,7 @@ Even though it is suboptimal, one usually resorts to sampling $X$ on a homogeneo
 
 #### Choosing new points based on existing data improves the simulation efficiency.
 <!-- This should convey the point that it is advantageous to do this. -->
-An alternative, that improves the simulation efficiency is to choose new, potentially interesting points in $X$ based on existing data [@Gramacy2004; @Figueiredo1995; @Castro2008; @Chen2017]. <!-- cite i.e., hydrodynamics-->
+An alternative, which improves the simulation efficiency, is to choose new potentially interesting points in $X$, based on existing data [@Gramacy2004; @Figueiredo1995; @Castro2008; @Chen2017]. <!-- cite i.e., hydrodynamics-->
 Bayesian optimization works well for high-cost simulations where one needs to find a minimum (or maximum) [@Takhtaganov2018].
 However, if the goal of the simulation is to approximate a continuous function using the fewest points, the continuity of the approximation is achieved by a greedy algorithm that samples mid-points of intervals with the largest distance or curvature [@Wolfram2011].
 Such a sampling strategy (i.e., in Fig. @fig:algo) would trivially speedup many simulations.
@@ -55,7 +55,7 @@ For a one-dimensional function with three points known (its boundary points and
 (4) calculate $f(x_\textrm{new})$,
 (5) repeat the previous steps, without redoing calculations for unchanged intervals.
 
-In this paper, we describe a class of algorithms that rely on local criteria for sampling, such as in the former example.
+In this paper, we present a class of algorithms that rely on local criteria for sampling, such as in the former example.
 Here we associate a *local loss* to each interval and pick a *candidate point* inside the interval with the largest loss.
 For example, in the case of the integration algorithm, the loss is the error estimate.
 The advantage of these *local* algorithms is that they allow for easy parallelization and have a low computational overhead.
-- 
GitLab