correct lint warnings
This commit is contained in:
parent
6f0cb4f11c
commit
4abe7b53ba
@ -439,7 +439,8 @@ class Timeline(object):
|
|||||||
tid = nodestats.thread_id
|
tid = nodestats.thread_id
|
||||||
inputs = []
|
inputs = []
|
||||||
if is_gputrace:
|
if is_gputrace:
|
||||||
node_name, op = self._parse_kernel_label(nodestats.timeline_label, node_name)
|
node_name, op = self._parse_kernel_label(nodestats.timeline_label,
|
||||||
|
node_name)
|
||||||
elif node_name == 'RecvTensor':
|
elif node_name == 'RecvTensor':
|
||||||
# RPC tracing does not use the standard timeline_label format.
|
# RPC tracing does not use the standard timeline_label format.
|
||||||
op = 'RecvTensor'
|
op = 'RecvTensor'
|
||||||
@ -623,11 +624,13 @@ class Timeline(object):
|
|||||||
Args:
|
Args:
|
||||||
op_time: How the execution time of op is shown in timeline.
|
op_time: How the execution time of op is shown in timeline.
|
||||||
Possible values are "schedule", "gpu" and "all".
|
Possible values are "schedule", "gpu" and "all".
|
||||||
"schedule" will show op from the time it is scheduled to the end of the scheduling.
|
"schedule" will show op from the time it is scheduled to the end of
|
||||||
|
the scheduling.
|
||||||
Notice by the end of its scheduling its async kernels may not start yet.
|
Notice by the end of its scheduling its async kernels may not start yet.
|
||||||
It is shown using the default value from step_stats.
|
It is shown using the default value from step_stats.
|
||||||
"gpu" will show op with the execution time of its kernels on GPU.
|
"gpu" will show op with the execution time of its kernels on GPU.
|
||||||
"all" will show op from the start of its scheduling to the end of its last kernel.
|
"all" will show op from the start of its scheduling to the end of
|
||||||
|
its last kernel.
|
||||||
"""
|
"""
|
||||||
if op_time == "schedule":
|
if op_time == "schedule":
|
||||||
self._step_stats = self._origin_step_stats
|
self._step_stats = self._origin_step_stats
|
||||||
@ -648,7 +651,8 @@ class Timeline(object):
|
|||||||
op_gpu_end = {}
|
op_gpu_end = {}
|
||||||
for stats in stream_all_stats:
|
for stats in stream_all_stats:
|
||||||
for kernel in stats.node_stats:
|
for kernel in stats.node_stats:
|
||||||
name, _ = self._parse_kernel_label(kernel.timeline_label, kernel.node_name)
|
name, _ = self._parse_kernel_label(kernel.timeline_label,
|
||||||
|
kernel.node_name)
|
||||||
start = kernel.all_start_micros
|
start = kernel.all_start_micros
|
||||||
end = kernel.all_start_micros + kernel.all_end_rel_micros
|
end = kernel.all_start_micros + kernel.all_end_rel_micros
|
||||||
if name in op_gpu_start:
|
if name in op_gpu_start:
|
||||||
@ -662,13 +666,17 @@ class Timeline(object):
|
|||||||
for stats in job_stats:
|
for stats in job_stats:
|
||||||
for op in stats.node_stats:
|
for op in stats.node_stats:
|
||||||
if op.node_name in op_gpu_start:
|
if op.node_name in op_gpu_start:
|
||||||
end = max(op_gpu_end[op.node_name], op.all_start_micros + op.all_end_rel_micros)
|
end = max(op_gpu_end[op.node_name],
|
||||||
|
op.all_start_micros + op.all_end_rel_micros)
|
||||||
if op_time == "gpu":
|
if op_time == "gpu":
|
||||||
op.all_start_micros = op_gpu_start[op.node_name]
|
op.all_start_micros = op_gpu_start[op.node_name]
|
||||||
op.all_end_rel_micros = end - op.all_start_micros
|
op.all_end_rel_micros = end - op.all_start_micros
|
||||||
|
|
||||||
def analyze_step_stats(self, show_dataflow=True, show_memory=True, op_time="schedule"):
|
def analyze_step_stats(self,
|
||||||
"""Analyze the step stats and format it into Chrome Trace Format
|
show_dataflow=True,
|
||||||
|
show_memory=True,
|
||||||
|
op_time="schedule"):
|
||||||
|
"""Analyze the step stats and format it into Chrome Trace Format.
|
||||||
|
|
||||||
Args:
|
Args:
|
||||||
show_dataflow: (Optional.) If True, add flow events to the trace
|
show_dataflow: (Optional.) If True, add flow events to the trace
|
||||||
@ -677,11 +685,13 @@ class Timeline(object):
|
|||||||
showing the sizes and lifetimes of tensors.
|
showing the sizes and lifetimes of tensors.
|
||||||
op_time: (Optional.) How the execution time of op is shown in timeline.
|
op_time: (Optional.) How the execution time of op is shown in timeline.
|
||||||
Possible values are "schedule", "gpu" and "all".
|
Possible values are "schedule", "gpu" and "all".
|
||||||
"schedule" will show op from the time it is scheduled to the end of the scheduling.
|
"schedule" will show op from the time it is scheduled to the end of
|
||||||
Notice by the end of its scheduling its async kernels may not start yet.
|
the scheduling.
|
||||||
It is shown using the default value from step_stats.
|
Notice by the end of its scheduling its async kernels may not start
|
||||||
|
yet. It is shown using the default value from step_stats.
|
||||||
"gpu" will show op with the execution time of its kernels on GPU.
|
"gpu" will show op with the execution time of its kernels on GPU.
|
||||||
"all" will show op from the start of its scheduling to the end of its last kernel.
|
"all" will show op from the start of its scheduling to the end of
|
||||||
|
its last kernel.
|
||||||
|
|
||||||
Returns:
|
Returns:
|
||||||
A 'StepStatsAnalysis' object.
|
A 'StepStatsAnalysis' object.
|
||||||
@ -697,7 +707,10 @@ class Timeline(object):
|
|||||||
chrome_trace=self._chrome_trace,
|
chrome_trace=self._chrome_trace,
|
||||||
allocator_maximums=self._allocator_maximums)
|
allocator_maximums=self._allocator_maximums)
|
||||||
|
|
||||||
def generate_chrome_trace_format(self, show_dataflow=True, show_memory=False, op_time="schedule"):
|
def generate_chrome_trace_format(self,
|
||||||
|
show_dataflow=True,
|
||||||
|
show_memory=False,
|
||||||
|
op_time="schedule"):
|
||||||
"""Produces a trace in Chrome Trace Format.
|
"""Produces a trace in Chrome Trace Format.
|
||||||
|
|
||||||
Args:
|
Args:
|
||||||
@ -707,11 +720,13 @@ class Timeline(object):
|
|||||||
showing the sizes and lifetimes of tensors.
|
showing the sizes and lifetimes of tensors.
|
||||||
op_time: (Optional.) How the execution time of op is shown in timeline.
|
op_time: (Optional.) How the execution time of op is shown in timeline.
|
||||||
Possible values are "schedule", "gpu" and "all".
|
Possible values are "schedule", "gpu" and "all".
|
||||||
"schedule" will show op from the time it is scheduled to the end of the scheduling.
|
"schedule" will show op from the time it is scheduled to the end of
|
||||||
Notice by the end of its scheduling its async kernels may not start yet.
|
the scheduling.
|
||||||
It is shown using the default value from step_stats.
|
Notice by the end of its scheduling its async kernels may not start
|
||||||
|
yet. It is shown using the default value from step_stats.
|
||||||
"gpu" will show op with the execution time of its kernels on GPU.
|
"gpu" will show op with the execution time of its kernels on GPU.
|
||||||
"all" will show op from the start of its scheduling to the end of its last kernel.
|
"all" will show op from the start of its scheduling to the end of
|
||||||
|
its last kernel.
|
||||||
|
|
||||||
Returns:
|
Returns:
|
||||||
A JSON formatted string in Chrome Trace format.
|
A JSON formatted string in Chrome Trace format.
|
||||||
|
Loading…
Reference in New Issue
Block a user