tf.distribute.NcclAllReduce
Stay organized with collections
Save and categorize content based on your preferences.
Reduction using NCCL all-reduce.
Inherits From: AllReduceCrossDeviceOps
tf.distribute.NcclAllReduce(
num_packs=1
)
Args |
num_packs
|
values will be packed in this many splits. num_packs should
be greater than or equals 0. When it is zero, no packing will be done.
|
Raises |
ValueError if num_packs is negative.
|
Methods
batch_reduce
View source
batch_reduce(
reduce_op, value_destination_pairs
)
Reduce PerReplica objects in a batch.
Reduce each first element in value_destination_pairs
to each second
element which indicates the destinations.
Args |
reduce_op
|
Indicates how per_replica_value will be reduced. Accepted
values are tf.distribute.ReduceOp.SUM , tf.distribute.ReduceOp.MEAN .
|
value_destination_pairs
|
a list or a tuple of tuples of PerReplica objects
(or tensors with device set if there is one device) and destinations.
|
Returns |
a list of Mirrored objects.
|
Raises |
ValueError
|
if value_destination_pairs is not a list or a tuple of
tuples of PerReplica objects and destinations
|
broadcast
View source
broadcast(
tensor, destinations
)
Broadcast the tensor
to destinations.
Args |
tensor
|
the tensor to broadcast.
|
destinations
|
the broadcast destinations.
|
Returns |
a Mirrored object.
|
reduce
View source
reduce(
reduce_op, per_replica_value, destinations
)
Reduce per_replica_value
to destinations
.
It runs the reduction operation defined by reduce_op
and put the
result on destinations
.
Returns |
a Mirrored object.
|
Raises |
ValueError
|
if per_replica_value can't be converted to a PerReplica
object.
|
Except as otherwise noted, the content of this page is licensed under the Creative Commons Attribution 4.0 License, and code samples are licensed under the Apache 2.0 License. For details, see the Google Developers Site Policies. Java is a registered trademark of Oracle and/or its affiliates.
Last updated 2020-10-01 UTC.
[null,null,["Last updated 2020-10-01 UTC."],[],[],null,["# tf.distribute.NcclAllReduce\n\n\u003cbr /\u003e\n\n|----------------------------------------------------------------------|-------------------------------------------------------------------------------------------------------------------------------------------|\n| [TensorFlow 2 version](/api_docs/python/tf/distribute/NcclAllReduce) | [View source on GitHub](https://github.com/tensorflow/tensorflow/blob/v1.15.0/tensorflow/python/distribute/cross_device_ops.py#L795-L817) |\n\nReduction using NCCL all-reduce.\n\nInherits From: [`AllReduceCrossDeviceOps`](../../tf/contrib/distribute/AllReduceCrossDeviceOps)\n\n#### View aliases\n\n\n**Compat aliases for migration**\n\nSee\n[Migration guide](https://www.tensorflow.org/guide/migrate) for\nmore details.\n\n[`tf.compat.v1.distribute.NcclAllReduce`](/api_docs/python/tf/distribute/NcclAllReduce), \\`tf.compat.v2.distribute.NcclAllReduce\\`\n\n\u003cbr /\u003e\n\n tf.distribute.NcclAllReduce(\n num_packs=1\n )\n\n\u003cbr /\u003e\n\n\u003cbr /\u003e\n\n\u003cbr /\u003e\n\n| Args ---- ||\n|-------------|--------------------------------------------------------------------------------------------------------------------------------------|\n| `num_packs` | values will be packed in this many splits. `num_packs` should be greater than or equals 0. When it is zero, no packing will be done. |\n\n\u003cbr /\u003e\n\n\u003cbr /\u003e\n\n\u003cbr /\u003e\n\n\u003cbr /\u003e\n\n| Raises ------ ||\n|---|---|\n| ValueError if `num_packs` is negative. ||\n\n\u003cbr /\u003e\n\nMethods\n-------\n\n### `batch_reduce`\n\n[View source](https://github.com/tensorflow/tensorflow/blob/v1.15.0/tensorflow/python/distribute/cross_device_ops.py#L284-L324) \n\n batch_reduce(\n reduce_op, value_destination_pairs\n )\n\nReduce PerReplica objects in a batch.\n\nReduce each first element in `value_destination_pairs` to each second\nelement which indicates the destinations.\n\n\u003cbr /\u003e\n\n\u003cbr /\u003e\n\n\u003cbr /\u003e\n\n| Args ||\n|---------------------------|------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|\n| `reduce_op` | Indicates how per_replica_value will be reduced. Accepted values are [`tf.distribute.ReduceOp.SUM`](../../tf/distribute/ReduceOp#SUM), [`tf.distribute.ReduceOp.MEAN`](../../tf/distribute/ReduceOp#MEAN). |\n| `value_destination_pairs` | a list or a tuple of tuples of PerReplica objects (or tensors with device set if there is one device) and destinations. |\n\n\u003cbr /\u003e\n\n\u003cbr /\u003e\n\n\u003cbr /\u003e\n\n\u003cbr /\u003e\n\n| Returns ||\n|---|---|\n| a list of Mirrored objects. ||\n\n\u003cbr /\u003e\n\n\u003cbr /\u003e\n\n\u003cbr /\u003e\n\n\u003cbr /\u003e\n\n| Raises ||\n|--------------|--------------------------------------------------------------------------------------------------------|\n| `ValueError` | if `value_destination_pairs` is not a list or a tuple of tuples of PerReplica objects and destinations |\n\n\u003cbr /\u003e\n\n### `broadcast`\n\n[View source](https://github.com/tensorflow/tensorflow/blob/v1.15.0/tensorflow/python/distribute/cross_device_ops.py#L326-L337) \n\n broadcast(\n tensor, destinations\n )\n\nBroadcast the `tensor` to destinations.\n\n\u003cbr /\u003e\n\n\u003cbr /\u003e\n\n\u003cbr /\u003e\n\n| Args ||\n|----------------|-----------------------------|\n| `tensor` | the tensor to broadcast. |\n| `destinations` | the broadcast destinations. |\n\n\u003cbr /\u003e\n\n\u003cbr /\u003e\n\n\u003cbr /\u003e\n\n\u003cbr /\u003e\n\n| Returns ||\n|---|---|\n| a Mirrored object. ||\n\n\u003cbr /\u003e\n\n### `reduce`\n\n[View source](https://github.com/tensorflow/tensorflow/blob/v1.15.0/tensorflow/python/distribute/cross_device_ops.py#L248-L282) \n\n reduce(\n reduce_op, per_replica_value, destinations\n )\n\nReduce `per_replica_value` to `destinations`.\n\nIt runs the reduction operation defined by `reduce_op` and put the\nresult on `destinations`.\n\n\u003cbr /\u003e\n\n\u003cbr /\u003e\n\n\u003cbr /\u003e\n\n| Args ||\n|---------------------|------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|\n| `reduce_op` | Indicates how per_replica_value will be reduced. Accepted values are [`tf.distribute.ReduceOp.SUM`](../../tf/distribute/ReduceOp#SUM), [`tf.distribute.ReduceOp.MEAN`](../../tf/distribute/ReduceOp#MEAN). |\n| `per_replica_value` | a PerReplica object or a tensor with device set. |\n| `destinations` | the reduction destinations. |\n\n\u003cbr /\u003e\n\n\u003cbr /\u003e\n\n\u003cbr /\u003e\n\n\u003cbr /\u003e\n\n| Returns ||\n|---|---|\n| a Mirrored object. ||\n\n\u003cbr /\u003e\n\n\u003cbr /\u003e\n\n\u003cbr /\u003e\n\n\u003cbr /\u003e\n\n| Raises ||\n|--------------|-----------------------------------------------------------------|\n| `ValueError` | if per_replica_value can't be converted to a PerReplica object. |\n\n\u003cbr /\u003e"]]