Abstract class that provides helpers for TensorFlow benchmarks.
tf.test.Benchmark()
Methods
evaluate
View source
evaluate(
    tensors
)
Evaluates tensors and returns numpy values.
| Args | 
tensors
 | 
A Tensor or a nested list/tuple of Tensors.
 | 
| Returns | 
| 
tensors numpy values.
 | 
is_abstract
View source
@classmethod
is_abstract()
report_benchmark
View source
report_benchmark(
    iters=None, cpu_time=None, wall_time=None, throughput=None, extras=None,
    name=None, metrics=None
)
Report a benchmark.
| Args | 
iters
 | 
(optional) How many iterations were run
 | 
cpu_time
 | 
(optional) Median or mean cpu time in seconds.
 | 
wall_time
 | 
(optional) Median or mean wall time in seconds.
 | 
throughput
 | 
(optional) Throughput (in MB/s)
 | 
extras
 | 
(optional) Dict mapping string keys to additional benchmark info.
Values may be either floats or values that are convertible to strings.
 | 
name
 | 
(optional) Override the BenchmarkEntry name with name.
Otherwise it is inferred from the top-level method name.
 | 
metrics
 | 
(optional) A list of dict, where each dict has the keys below
name (required), string, metric name
value (required), double, metric value
min_value (optional), double, minimum acceptable metric value
max_value (optional), double, maximum acceptable metric value
 | 
run_op_benchmark
View source
run_op_benchmark(
    sess, op_or_tensor, feed_dict=None, burn_iters=2, min_iters=10,
    store_trace=False, store_memory_usage=True, name=None, extras=None, mbs=0
)
Run an op or tensor in the given session.  Report the results.
| Args | 
sess
 | 
Session object to use for timing.
 | 
op_or_tensor
 | 
Operation or Tensor to benchmark.
 | 
feed_dict
 | 
A dict of values to feed for each op iteration (see the
feed_dict parameter of Session.run).
 | 
burn_iters
 | 
Number of burn-in iterations to run.
 | 
min_iters
 | 
Minimum number of iterations to use for timing.
 | 
store_trace
 | 
Boolean, whether to run an extra untimed iteration and
store the trace of iteration in returned extras.
The trace will be stored as a string in Google Chrome trace format
in the extras field "full_trace_chrome_format". Note that trace
will not be stored in test_log_pb2.TestResults proto.
 | 
store_memory_usage
 | 
Boolean, whether to run an extra untimed iteration,
calculate memory usage, and store that in extras fields.
 | 
name
 | 
(optional) Override the BenchmarkEntry name with name.
Otherwise it is inferred from the top-level method name.
 | 
extras
 | 
(optional) Dict mapping string keys to additional benchmark info.
Values may be either floats or values that are convertible to strings.
 | 
mbs
 | 
(optional) The number of megabytes moved by this op, used to
calculate the ops throughput.
 | 
| Returns | 
A dict containing the key-value pairs that were passed to
report_benchmark. If store_trace option is used, then
full_chrome_trace_format will be included in return dictionary even
though it is not passed to report_benchmark with extras.
 |