Abstract class that provides helpers for TensorFlow benchmarks.
tf.test.Benchmark()
Methods
evaluate
View source
evaluate(
    tensors
)
Evaluates tensors and returns numpy values.
| Args | 
|---|
| tensors | A Tensor or a nested list/tuple of Tensors. | 
| Returns | 
|---|
| tensors numpy values. | 
is_abstract
View source
@classmethod
is_abstract()
report_benchmark
View source
report_benchmark(
    iters=None,
    cpu_time=None,
    wall_time=None,
    throughput=None,
    extras=None,
    name=None,
    metrics=None
)
Report a benchmark.
| Args | 
|---|
| iters | (optional) How many iterations were run | 
| cpu_time | (optional) Median or mean cpu time in seconds. | 
| wall_time | (optional) Median or mean wall time in seconds. | 
| throughput | (optional) Throughput (in MB/s) | 
| extras | (optional) Dict mapping string keys to additional benchmark info.
Values may be either floats or values that are convertible to strings. | 
| name | (optional) Override the BenchmarkEntry name with name.
Otherwise it is inferred from the top-level method name. | 
| metrics | (optional) A list of dict, where each dict has the keys below
name (required), string, metric name
value (required), double, metric value
min_value (optional), double, minimum acceptable metric value
max_value (optional), double, maximum acceptable metric value | 
run_op_benchmark
View source
run_op_benchmark(
    sess,
    op_or_tensor,
    feed_dict=None,
    burn_iters=2,
    min_iters=10,
    store_trace=False,
    store_memory_usage=True,
    name=None,
    extras=None,
    mbs=0
)
Run an op or tensor in the given session.  Report the results.
| Args | 
|---|
| sess | Sessionobject to use for timing. | 
| op_or_tensor | OperationorTensorto benchmark. | 
| feed_dict | A dictof values to feed for each op iteration (see thefeed_dictparameter ofSession.run). | 
| burn_iters | Number of burn-in iterations to run. | 
| min_iters | Minimum number of iterations to use for timing. | 
| store_trace | Boolean, whether to run an extra untimed iteration and
store the trace of iteration in returned extras.
The trace will be stored as a string in Google Chrome trace format
in the extras field "full_trace_chrome_format". Note that trace
will not be stored in test_log_pb2.TestResults proto. | 
| store_memory_usage | Boolean, whether to run an extra untimed iteration,
calculate memory usage, and store that in extras fields. | 
| name | (optional) Override the BenchmarkEntry name with name.
Otherwise it is inferred from the top-level method name. | 
| extras | (optional) Dict mapping string keys to additional benchmark info.
Values may be either floats or values that are convertible to strings. | 
| mbs | (optional) The number of megabytes moved by this op, used to
calculate the ops throughput. | 
| Returns | 
|---|
| A dictcontaining the key-value pairs that were passed toreport_benchmark. Ifstore_traceoption is used, thenfull_chrome_trace_formatwill be included in return dictionary even
though it is not passed toreport_benchmarkwithextras. |