aboutsummaryrefslogtreecommitdiffhomepage
path: root/tensorflow/docs_src/performance/quantization.md
diff options
context:
space:
mode:
Diffstat (limited to 'tensorflow/docs_src/performance/quantization.md')
-rw-r--r--tensorflow/docs_src/performance/quantization.md2
1 files changed, 1 insertions, 1 deletions
diff --git a/tensorflow/docs_src/performance/quantization.md b/tensorflow/docs_src/performance/quantization.md
index 4499f5715c..3326d82964 100644
--- a/tensorflow/docs_src/performance/quantization.md
+++ b/tensorflow/docs_src/performance/quantization.md
@@ -80,7 +80,7 @@ need for a separate calibration step.
TensorFlow can train models with quantization in the loop. Because training
requires small gradient adjustments, floating point values are still used. To
keep models as floating point while adding the quantization error in the training
-loop, @{$array_ops#Fake_quantization$fake quantization} nodes simulate the
+loop, [fake quantization](../api_guides/python/array_ops.md#Fake_quantization) nodes simulate the
effect of quantization in the forward and backward passes.
Since it's difficult to add these fake quantization operations to all the