使用集合让一切井井有条
根据您的偏好保存内容并对其进行分类。
工具
探索可支持和加速 TensorFlow 工作流程的工具。
Colab
Colaboratory 是一个免费的 Jupyter 笔记本环境,不需要进行任何设置就可以使用,并且完全在云端运行。借助 Colaboratory,您只需点击一下鼠标,即可在浏览器中执行 TensorFlow 代码。
Visual Blocks
一种可视化编码 Web 框架,可将 I/O 设备、模型、数据增强甚至 Colab 代码作为可重复使用的基本构件,设计机器学习工作流原型。
What-If Tool
一种以无代码的方式探究机器学习模型的工具,对模型的理解、调试和公平性很有用。可在 TensorBoard 和 Jupyter 或 Colab 笔记本中使用。
ML Perf
全面的机器学习基准测试套件,用于衡量机器学习软件框架、机器学习硬件加速器和机器学习云端平台的性能。
XLA
XLA(加速线性代数)是一种针对特定领域的线性代数编译器,能够优化 TensorFlow 计算,它可以提高服务器和移动平台的运行速度,并改进内存使用情况和可移植性。
TPU Research Cloud
加入 TPU Research Cloud (TRC) 计划后,研究人员可以申请访问 Cloud TPU 来加快实现下一波研究突破;我们免费提供了 1000 多个 Cloud TPU。
[null,null,[],[],[],null,["# Tools\n=====\n\nExplore tools to support and accelerate TensorFlow workflows. \n[Colab](https://colab.sandbox.google.com/notebooks/welcome.ipynb) \nColaboratory is a free Jupyter notebook environment that requires no setup and runs entirely in the cloud, allowing you to execute TensorFlow code in your browser with a single click. \n[Learn more](https://colab.sandbox.google.com/notebooks/welcome.ipynb) \n[Visual Blocks](https://visualblocks.withgoogle.com/) \nA visual coding web framework to prototype ML workflows using I/O devices, models, data augmentation, and even Colab code as reusable building blocks. \n[Learn more](https://visualblocks.withgoogle.com/) \n[TensorBoard](/tensorboard) \nA suite of visualization tools to understand, debug, and optimize TensorFlow programs. \n[Learn more](/tensorboard) [View code](https://github.com/tensorflow/tensorboard) \n[What-If Tool](https://pair-code.github.io/what-if-tool/) \nA tool for code-free probing of machine learning models, useful for model understanding, debugging, and fairness. Available in TensorBoard and jupyter or colab notebooks. \n[Learn more](https://pair-code.github.io/what-if-tool/) [Get started](https://colab.research.google.com/github/PAIR-code/what-if-tool/blob/master/What_If_Tool_Notebook_Usage.ipynb) \n[ML Perf](https://mlperf.org/) \nA broad ML benchmark suite for measuring performance of ML software frameworks, ML hardware accelerators, and ML cloud platforms. \n[Learn more](https://mlperf.org/) \n[XLA](/xla) \nXLA (Accelerated Linear Algebra) is a domain-specific compiler for linear algebra that optimizes TensorFlow computations. The results are improvements in speed, memory usage, and portability on server and mobile platforms. \n[Learn more](/xla) \n[TensorFlow Playground](https://playground.tensorflow.org/#activation=tanh&batchSize=10&dataset=circle®Dataset=reg-plane&learningRate=0.03®ularizationRate=0&noise=0&networkShape=4,2&seed=0.04620&showTestData=false&discretize=false&percTrainData=50&x=true&y=true&xTimesY=false&xSquared=false&ySquared=false&cosX=false&sinX=false&cosY=false&sinY=false&collectStats=false&problem=classification&initZero=false&hideText=false) \nTinker with a neural network in your browser. Don't worry, you can't break it. \n[Learn more](https://playground.tensorflow.org/#activation=tanh&batchSize=10&dataset=circle®Dataset=reg-plane&learningRate=0.03®ularizationRate=0&noise=0&networkShape=4,2&seed=0.04620&showTestData=false&discretize=false&percTrainData=50&x=true&y=true&xTimesY=false&xSquared=false&ySquared=false&cosX=false&sinX=false&cosY=false&sinY=false&collectStats=false&problem=classification&initZero=false&hideText=false) \n[TPU Research Cloud](https://sites.research.google/trc/) \nThe TPU Research Cloud (TRC) program enables researchers to apply for access to a cluster of more than 1,000 Cloud TPUs at no charge to help them accelerate the next wave of research breakthroughs. \n[Learn more](https://sites.research.google/trc/) \n[MLIR](/mlir) \nA new intermediate representation and compiler framework. \n[Learn more](/mlir) \n\nExplore libraries that build advanced models, methods, and extensions using TensorFlow\n--------------------------------------------------------------------------------------\n\n[See libraries](/resources/libraries-extensions)"]]