fix markdown syntax mistakes

This commit is contained in:
meijun 2017-08-29 07:42:09 +08:00
parent 86d9171b00
commit 51eb71cd3b
6 changed files with 20 additions and 20 deletions

View File

@ -86,7 +86,7 @@ For example:
* Checks RecvTensor RPC latency and bandwidth. * Checks RecvTensor RPC latency and bandwidth.
* Checks CPU/Memory utilization of the job. * Checks CPU/Memory utilization of the job.
####AcceleratorUtilization Checker #### AcceleratorUtilization Checker
* Checks what percentage of time the accelerator spends on computation. * Checks what percentage of time the accelerator spends on computation.
#### OperationChecker #### OperationChecker
@ -100,7 +100,7 @@ For example:
* Checks the most expensive graph nodes. * Checks the most expensive graph nodes.
* Checks the most expensive graph-building Python codes. * Checks the most expensive graph-building Python codes.
####Contribute Your Checker #### Contribute Your Checker
Follow examples of accelerator_utilization_checker.h Follow examples of accelerator_utilization_checker.h

View File

@ -51,7 +51,7 @@ It defines _checkpoint_variable op type. It also provides checkpointed tensors'
Note: this feature is not well maintained now. Note: this feature is not well maintained now.
###Start `tfprof` ### Start `tfprof`
#### Build `tfprof` #### Build `tfprof`
@ -140,9 +140,9 @@ tfprof>
-output -output
``` ```
###Examples ### Examples
####Profile Python Time #### Profile Python Time
```shell ```shell
# Requires --graph_path --op_log_path # Requires --graph_path --op_log_path
tfprof> code -max_depth 1000 -show_name_regexes .*model_analyzer.*py.* -select micros -account_type_regexes .* -order_by micros tfprof> code -max_depth 1000 -show_name_regexes .*model_analyzer.*py.* -select micros -account_type_regexes .* -order_by micros

View File

@ -1,6 +1,6 @@
##Options ## Options
###Overview ### Overview
For all tfprof views, the profiles are processed with the following procedures For all tfprof views, the profiles are processed with the following procedures
@ -35,7 +35,7 @@ For all tfprof views, the profiles are processed with the following procedures
4) Finally, the filtered data structure is output in a format depending 4) Finally, the filtered data structure is output in a format depending
on the `-output` option. on the `-output` option.
####Option Semantics In Different View #### Option Semantics In Different View
options usually have the same semantics in different views. However, some options usually have the same semantics in different views. However, some
can vary. For example `-max_depth` in scope view means the depth of can vary. For example `-max_depth` in scope view means the depth of
name scope <b>tree</b>. In op view, it means the length of operation <b>list</b>. name scope <b>tree</b>. In op view, it means the length of operation <b>list</b>.
@ -68,7 +68,7 @@ output_bytes: The memory output by the operation. It's not necessarily requested
by the current operation. For example, it can be a tensor by the current operation. For example, it can be a tensor
forwarded from input to output, with in-place mutation. forwarded from input to output, with in-place mutation.
###Docs ### Docs
`-max_depth`: Show nodes that are at most this number of hops from starting node in the data structure. `-max_depth`: Show nodes that are at most this number of hops from starting node in the data structure.

View File

@ -1,4 +1,4 @@
##Profile Memory ## Profile Memory
It is generally a good idea to visualize the memory usage in timeline. It is generally a good idea to visualize the memory usage in timeline.
It allows you to see the memory consumption of each GPU over time. It allows you to see the memory consumption of each GPU over time.

View File

@ -1,9 +1,9 @@
##Profile Model Architecture ## Profile Model Architecture
* [Profile Model Parameters](#profile-model-parameters) * [Profile Model Parameters](#profile-model-parameters)
* [Profile Model Float Operations](#profile-model-float-operations) * [Profile Model Float Operations](#profile-model-float-operations)
###Profile Model Parameters ### Profile Model Parameters
<b>Notes:</b> <b>Notes:</b>
`VariableV2` operation type might contain variables created by TensorFlow `VariableV2` operation type might contain variables created by TensorFlow
@ -39,9 +39,9 @@ param_stats = tf.profiler.profile(
sys.stdout.write('total_params: %d\n' % param_stats.total_parameters) sys.stdout.write('total_params: %d\n' % param_stats.total_parameters)
``` ```
###Profile Model Float Operations ### Profile Model Float Operations
####Caveats #### Caveats
For an operation to have float operation statistics: For an operation to have float operation statistics:

View File

@ -1,4 +1,4 @@
##Profile Time ## Profile Time
* [Times in TensorFlow and tfprof](#times-in-tensorflow-and-tfprof) * [Times in TensorFlow and tfprof](#times-in-tensorflow-and-tfprof)
* [Profile by Python Code](#profile-by-python-code) * [Profile by Python Code](#profile-by-python-code)
@ -7,7 +7,7 @@
* [Profile by Name Scope](#profile-by-name-scope) * [Profile by Name Scope](#profile-by-name-scope)
###Times in TensorFlow and tfprof ### Times in TensorFlow and tfprof
When we run a model, Tensorflow schedules and runs the nodes (operations) When we run a model, Tensorflow schedules and runs the nodes (operations)
in the graph. An operation can be placed on an accelerator or on CPU. in the graph. An operation can be placed on an accelerator or on CPU.
@ -37,7 +37,7 @@ When an operation is placed on CPU, it will completely run on CPU. Hence,
should be 0. should be 0.
###Profile by Python Code ### Profile by Python Code
```python ```python
# In code view, the time of each line of Python code is the aggregated # In code view, the time of each line of Python code is the aggregated
# times of all operations created by that line. # times of all operations created by that line.
@ -112,7 +112,7 @@ Set ```-output timeline:outfile=<filename>``` to generate timeline instead of st
</left> </left>
###Profile by Operation Type ### Profile by Operation Type
```python ```python
# In op view, you can view the aggregated time of each operation type. # In op view, you can view the aggregated time of each operation type.
tfprof> op -select micros,occurrence -order_by micros tfprof> op -select micros,occurrence -order_by micros
@ -138,7 +138,7 @@ MatMul 618.97ms (63.56%, 16.51%), |/job:worker/replica:0/
``` ```
###Profile by Graph ### Profile by Graph
Usually, use graph view to generate a timeline to visualize the result. Usually, use graph view to generate a timeline to visualize the result.
@ -163,7 +163,7 @@ Open a Chrome browser, enter URL chrome://tracing and load the timeline file.
****************************************************** ******************************************************
``` ```
###Profile by Name Scope ### Profile by Name Scope
Usually scope view allows you to pin point the problematic places if you Usually scope view allows you to pin point the problematic places if you
have properly named your operations with tf.name_scope or tf.variable_scope. have properly named your operations with tf.name_scope or tf.variable_scope.