Patchwork Improve math benchmark infrastructure

login
register
mail settings
Submitter Wilco Dijkstra
Date June 19, 2017, 4:52 p.m.
Message ID <AM5PR0802MB26106CCEAB7753D6C06D901C83C40@AM5PR0802MB2610.eurprd08.prod.outlook.com>
Download mbox | patch
Permalink /patch/21110/
State Committed
Headers show

Comments

Wilco Dijkstra - June 19, 2017, 4:52 p.m.
Siddhesh Poyarekar wrote:
>
> There is no easy way to actually replicate the cache effects of a real
> workload without replicating the program in some form, but this is a
> good start.  Long term we should consider doing things like invalidating
> cache at specific points in the program in an attempt to emulate the
> workload a little more closely.  This may be more relevant for string
> benchmarks than for math though.

Yes it's hard to replicate the full cache effects in a micro benchmark.
I'm not sure flushing is the best approach - at least for the string functions
it is possible to avoid using the same pointers again and again and do 
something similar to memcpy-random by precomputing a set of operations
and then do them in one go. For math functions it's likely more about
I-cache pressure.

> Use the name convention workload-* to better reflect what the dataset
> is.  So for your wrf powf trace, call it workload-spec2006.wrf.  Also,
> please document this feature in benchtests/README.

Done, see below:

Improve support for math function benchmarking.  This patch adds
a feature that allows accurate benchmarking of traces extracted
from real workloads.  This is done by iterating over all samples
rather than repeating each sample many times (which completely 
ignores branch prediction and cache effects).  A trace can be
added to existing math function inputs via 
"## name: workload-<name>", followed by the trace.

OK for commit?

ChangeLog:
2017-06-19  Wilco Dijkstra  <wdijkstr@arm.com>

        * benchtests/README: Describe workload feature.
        * benchtests/bench-skeleton.c (main): Add support for
        benchmarking traces from workloads.
--
Carlos O'Donell - June 19, 2017, 8:16 p.m.
On 06/19/2017 12:52 PM, Wilco Dijkstra wrote:
> Improve support for math function benchmarking.  This patch adds
> a feature that allows accurate benchmarking of traces extracted
> from real workloads.  This is done by iterating over all samples
> rather than repeating each sample many times (which completely 
> ignores branch prediction and cache effects).  A trace can be
> added to existing math function inputs via 
> "## name: workload-<name>", followed by the trace.

There are similarities between this and what DJ and the Red Hat glibc
team did for malloc tracing.

We hope to have this merged soon, but were blocked by the recent
CVE work.

One thing that was suggested at LPC2016 was that we could do better
at trace capture time, perhaps getting a histogram of cache behaviour
and then trying to emulate that same density function in the
whole-system benchmark.

Our page touch algorithm in the malloc simulator was too naive and
real access patterns might have drastic consequences on behaviour.

So I think maybe we'll be able to leverage the simulator we have a
bit more in the coming year to do more detailed modelling of a
workload.
Siddhesh Poyarekar - June 20, 2017, 5:22 a.m.
On Monday 19 June 2017 10:22 PM, Wilco Dijkstra wrote:
> Done, see below:
> 
> Improve support for math function benchmarking.  This patch adds
> a feature that allows accurate benchmarking of traces extracted
> from real workloads.  This is done by iterating over all samples
> rather than repeating each sample many times (which completely 
> ignores branch prediction and cache effects).  A trace can be
> added to existing math function inputs via 
> "## name: workload-<name>", followed by the trace.
> 
> OK for commit?

OK with a tiny nit fix below.

> ChangeLog:
> 2017-06-19  Wilco Dijkstra  <wdijkstr@arm.com>
> 
>         * benchtests/README: Describe workload feature.
>         * benchtests/bench-skeleton.c (main): Add support for
>         benchmarking traces from workloads.
> --
> diff --git a/benchtests/README b/benchtests/README
> index 2c5f38113593ea7da90895266c8fd523fa21c5a1..67333707d5bbc2c6cf5a4de5698c18dfdf086076 100644
> --- a/benchtests/README
> +++ b/benchtests/README
> @@ -102,6 +102,12 @@ the same file by using the `name' directive that looks something like this:
>  See the pow-inputs file for an example of what such a partitioned input file
>  would look like.
>  
> +It is also possible to measure throughput of a (partial) trace extracted from
> +a real workload.  In this case the whole trace is iterated over 'iter' times

  "the whole trace is iterated over multiple times"

since 'iter' is not defined here.

> +rather than repeating every input multiple times.  This can be done via:
> +
> +  ##name: workload-<name>
> +
>  Benchmark Sets:
>  ==============
>  
> diff --git a/benchtests/bench-skeleton.c b/benchtests/bench-skeleton.c
> index 09eb78df1bce2d9f5e410e3e82821eb9b271e70d..8c98ed673c055a5cf4d774604eb7bf0a383cecb2 100644
> --- a/benchtests/bench-skeleton.c
> +++ b/benchtests/bench-skeleton.c
> @@ -68,34 +68,50 @@ main (int argc, char **argv)
>        clock_gettime (CLOCK_MONOTONIC_RAW, &runtime);
>        runtime.tv_sec += DURATION;
>  
> +      bool is_bench = strncmp (VARIANT (v), "workload-", 9) == 0;
>        double d_total_i = 0;
>        timing_t total = 0, max = 0, min = 0x7fffffffffffffff;
>        int64_t c = 0;
> +      uint64_t cur;
>        while (1)
>  	{
> -	  for (i = 0; i < NUM_SAMPLES (v); i++)
> +	  if (is_bench)
>  	    {
> -	      uint64_t cur;
> +	      /* Benchmark a real trace of calls - all samples are iterated
> +		 over once before repeating.  This models actual use more
> +		 accurately than repeating the same sample many times.  */
>  	      TIMING_NOW (start);
>  	      for (k = 0; k < iters; k++)
> -		BENCH_FUNC (v, i);
> +		for (i = 0; i < NUM_SAMPLES (v); i++)
> +		  BENCH_FUNC (v, i);
>  	      TIMING_NOW (end);
> -
>  	      TIMING_DIFF (cur, start, end);
> +	      TIMING_ACCUM (total, cur);
> +	      d_total_i += iters * NUM_SAMPLES (v);
> +	    }
> +	  else
> +	    for (i = 0; i < NUM_SAMPLES (v); i++)
> +	      {
> +		TIMING_NOW (start);
> +		for (k = 0; k < iters; k++)
> +		  BENCH_FUNC (v, i);
> +		TIMING_NOW (end);
>  
> -	      if (cur > max)
> -		max = cur;
> +		TIMING_DIFF (cur, start, end);
>  
> -	      if (cur < min)
> -		min = cur;
> +		if (cur > max)
> +		  max = cur;
>  
> -	      TIMING_ACCUM (total, cur);
> -	      /* Accumulate timings for the value.  In the end we will divide
> -	         by the total iterations.  */
> -	      RESULT_ACCUM (cur, v, i, c * iters, (c + 1) * iters);
> +		if (cur < min)
> +		  min = cur;
>  
> -	      d_total_i += iters;
> -	    }
> +		TIMING_ACCUM (total, cur);
> +		/* Accumulate timings for the value.  In the end we will divide
> +		   by the total iterations.  */
> +		RESULT_ACCUM (cur, v, i, c * iters, (c + 1) * iters);
> +
> +		d_total_i += iters;
> +	      }
>  	  c++;
>  	  struct timespec curtime;
>  
> @@ -117,11 +133,18 @@ main (int argc, char **argv)
>  
>        json_attr_double (&json_ctx, "duration", d_total_s);
>        json_attr_double (&json_ctx, "iterations", d_total_i);
> -      json_attr_double (&json_ctx, "max", max / d_iters);
> -      json_attr_double (&json_ctx, "min", min / d_iters);
> -      json_attr_double (&json_ctx, "mean", d_total_s / d_total_i);
> +      if (is_bench)
> +	{
> +	  json_attr_double (&json_ctx, "throughput", d_total_s / d_total_i);
> +	}

Redundant braces.

> +      else
> +	{
> +	  json_attr_double (&json_ctx, "max", max / d_iters);
> +	  json_attr_double (&json_ctx, "min", min / d_iters);
> +	  json_attr_double (&json_ctx, "mean", d_total_s / d_total_i);
> +	}
>  
> -      if (detailed)
> +      if (detailed && !is_bench)
>  	{
>  	  json_array_begin (&json_ctx, "timings");
>  
>
Siddhesh Poyarekar - June 20, 2017, 5:29 a.m.
On Tuesday 20 June 2017 01:46 AM, Carlos O'Donell wrote:
> There are similarities between this and what DJ and the Red Hat glibc
> team did for malloc tracing.
> 
> We hope to have this merged soon, but were blocked by the recent
> CVE work.
> 
> One thing that was suggested at LPC2016 was that we could do better
> at trace capture time, perhaps getting a histogram of cache behaviour
> and then trying to emulate that same density function in the
> whole-system benchmark.

That sounds interesting.  What's the plan for the implementation?  Is it
planned to be part of benchtests or is it going to be a separate project
like we had discussed a couple of years ago?

> Our page touch algorithm in the malloc simulator was too naive and
> real access patterns might have drastic consequences on behaviour.
> 
> So I think maybe we'll be able to leverage the simulator we have a
> bit more in the coming year to do more detailed modelling of a
> workload.

Sorry I've lost track of this again, but what's up with DJ's malloc patch?

Siddhesh

Patch

diff --git a/benchtests/README b/benchtests/README
index 2c5f38113593ea7da90895266c8fd523fa21c5a1..67333707d5bbc2c6cf5a4de5698c18dfdf086076 100644
--- a/benchtests/README
+++ b/benchtests/README
@@ -102,6 +102,12 @@  the same file by using the `name' directive that looks something like this:
 See the pow-inputs file for an example of what such a partitioned input file
 would look like.
 
+It is also possible to measure throughput of a (partial) trace extracted from
+a real workload.  In this case the whole trace is iterated over 'iter' times
+rather than repeating every input multiple times.  This can be done via:
+
+  ##name: workload-<name>
+
 Benchmark Sets:
 ==============
 
diff --git a/benchtests/bench-skeleton.c b/benchtests/bench-skeleton.c
index 09eb78df1bce2d9f5e410e3e82821eb9b271e70d..8c98ed673c055a5cf4d774604eb7bf0a383cecb2 100644
--- a/benchtests/bench-skeleton.c
+++ b/benchtests/bench-skeleton.c
@@ -68,34 +68,50 @@  main (int argc, char **argv)
       clock_gettime (CLOCK_MONOTONIC_RAW, &runtime);
       runtime.tv_sec += DURATION;
 
+      bool is_bench = strncmp (VARIANT (v), "workload-", 9) == 0;
       double d_total_i = 0;
       timing_t total = 0, max = 0, min = 0x7fffffffffffffff;
       int64_t c = 0;
+      uint64_t cur;
       while (1)
 	{
-	  for (i = 0; i < NUM_SAMPLES (v); i++)
+	  if (is_bench)
 	    {
-	      uint64_t cur;
+	      /* Benchmark a real trace of calls - all samples are iterated
+		 over once before repeating.  This models actual use more
+		 accurately than repeating the same sample many times.  */
 	      TIMING_NOW (start);
 	      for (k = 0; k < iters; k++)
-		BENCH_FUNC (v, i);
+		for (i = 0; i < NUM_SAMPLES (v); i++)
+		  BENCH_FUNC (v, i);
 	      TIMING_NOW (end);
-
 	      TIMING_DIFF (cur, start, end);
+	      TIMING_ACCUM (total, cur);
+	      d_total_i += iters * NUM_SAMPLES (v);
+	    }
+	  else
+	    for (i = 0; i < NUM_SAMPLES (v); i++)
+	      {
+		TIMING_NOW (start);
+		for (k = 0; k < iters; k++)
+		  BENCH_FUNC (v, i);
+		TIMING_NOW (end);
 
-	      if (cur > max)
-		max = cur;
+		TIMING_DIFF (cur, start, end);
 
-	      if (cur < min)
-		min = cur;
+		if (cur > max)
+		  max = cur;
 
-	      TIMING_ACCUM (total, cur);
-	      /* Accumulate timings for the value.  In the end we will divide
-	         by the total iterations.  */
-	      RESULT_ACCUM (cur, v, i, c * iters, (c + 1) * iters);
+		if (cur < min)
+		  min = cur;
 
-	      d_total_i += iters;
-	    }
+		TIMING_ACCUM (total, cur);
+		/* Accumulate timings for the value.  In the end we will divide
+		   by the total iterations.  */
+		RESULT_ACCUM (cur, v, i, c * iters, (c + 1) * iters);
+
+		d_total_i += iters;
+	      }
 	  c++;
 	  struct timespec curtime;
 
@@ -117,11 +133,18 @@  main (int argc, char **argv)
 
       json_attr_double (&json_ctx, "duration", d_total_s);
       json_attr_double (&json_ctx, "iterations", d_total_i);
-      json_attr_double (&json_ctx, "max", max / d_iters);
-      json_attr_double (&json_ctx, "min", min / d_iters);
-      json_attr_double (&json_ctx, "mean", d_total_s / d_total_i);
+      if (is_bench)
+	{
+	  json_attr_double (&json_ctx, "throughput", d_total_s / d_total_i);
+	}
+      else
+	{
+	  json_attr_double (&json_ctx, "max", max / d_iters);
+	  json_attr_double (&json_ctx, "min", min / d_iters);
+	  json_attr_double (&json_ctx, "mean", d_total_s / d_total_i);
+	}
 
-      if (detailed)
+      if (detailed && !is_bench)
 	{
 	  json_array_begin (&json_ctx, "timings");