correct lint warnings

This commit is contained in:
zilinzhu 2020-03-20 10:35:08 +08:00
parent 6f0cb4f11c
commit 4abe7b53ba

View File

@ -439,7 +439,8 @@ class Timeline(object):
tid = nodestats.thread_id tid = nodestats.thread_id
inputs = [] inputs = []
if is_gputrace: if is_gputrace:
node_name, op = self._parse_kernel_label(nodestats.timeline_label, node_name) node_name, op = self._parse_kernel_label(nodestats.timeline_label,
node_name)
elif node_name == 'RecvTensor': elif node_name == 'RecvTensor':
# RPC tracing does not use the standard timeline_label format. # RPC tracing does not use the standard timeline_label format.
op = 'RecvTensor' op = 'RecvTensor'
@ -623,11 +624,13 @@ class Timeline(object):
Args: Args:
op_time: How the execution time of op is shown in timeline. op_time: How the execution time of op is shown in timeline.
Possible values are "schedule", "gpu" and "all". Possible values are "schedule", "gpu" and "all".
"schedule" will show op from the time it is scheduled to the end of the scheduling. "schedule" will show op from the time it is scheduled to the end of
the scheduling.
Notice by the end of its scheduling its async kernels may not start yet. Notice by the end of its scheduling its async kernels may not start yet.
It is shown using the default value from step_stats. It is shown using the default value from step_stats.
"gpu" will show op with the execution time of its kernels on GPU. "gpu" will show op with the execution time of its kernels on GPU.
"all" will show op from the start of its scheduling to the end of its last kernel. "all" will show op from the start of its scheduling to the end of
its last kernel.
""" """
if op_time == "schedule": if op_time == "schedule":
self._step_stats = self._origin_step_stats self._step_stats = self._origin_step_stats
@ -648,7 +651,8 @@ class Timeline(object):
op_gpu_end = {} op_gpu_end = {}
for stats in stream_all_stats: for stats in stream_all_stats:
for kernel in stats.node_stats: for kernel in stats.node_stats:
name, _ = self._parse_kernel_label(kernel.timeline_label, kernel.node_name) name, _ = self._parse_kernel_label(kernel.timeline_label,
kernel.node_name)
start = kernel.all_start_micros start = kernel.all_start_micros
end = kernel.all_start_micros + kernel.all_end_rel_micros end = kernel.all_start_micros + kernel.all_end_rel_micros
if name in op_gpu_start: if name in op_gpu_start:
@ -662,13 +666,17 @@ class Timeline(object):
for stats in job_stats: for stats in job_stats:
for op in stats.node_stats: for op in stats.node_stats:
if op.node_name in op_gpu_start: if op.node_name in op_gpu_start:
end = max(op_gpu_end[op.node_name], op.all_start_micros + op.all_end_rel_micros) end = max(op_gpu_end[op.node_name],
op.all_start_micros + op.all_end_rel_micros)
if op_time == "gpu": if op_time == "gpu":
op.all_start_micros = op_gpu_start[op.node_name] op.all_start_micros = op_gpu_start[op.node_name]
op.all_end_rel_micros = end - op.all_start_micros op.all_end_rel_micros = end - op.all_start_micros
def analyze_step_stats(self, show_dataflow=True, show_memory=True, op_time="schedule"): def analyze_step_stats(self,
"""Analyze the step stats and format it into Chrome Trace Format show_dataflow=True,
show_memory=True,
op_time="schedule"):
"""Analyze the step stats and format it into Chrome Trace Format.
Args: Args:
show_dataflow: (Optional.) If True, add flow events to the trace show_dataflow: (Optional.) If True, add flow events to the trace
@ -677,11 +685,13 @@ class Timeline(object):
showing the sizes and lifetimes of tensors. showing the sizes and lifetimes of tensors.
op_time: (Optional.) How the execution time of op is shown in timeline. op_time: (Optional.) How the execution time of op is shown in timeline.
Possible values are "schedule", "gpu" and "all". Possible values are "schedule", "gpu" and "all".
"schedule" will show op from the time it is scheduled to the end of the scheduling. "schedule" will show op from the time it is scheduled to the end of
Notice by the end of its scheduling its async kernels may not start yet. the scheduling.
It is shown using the default value from step_stats. Notice by the end of its scheduling its async kernels may not start
yet. It is shown using the default value from step_stats.
"gpu" will show op with the execution time of its kernels on GPU. "gpu" will show op with the execution time of its kernels on GPU.
"all" will show op from the start of its scheduling to the end of its last kernel. "all" will show op from the start of its scheduling to the end of
its last kernel.
Returns: Returns:
A 'StepStatsAnalysis' object. A 'StepStatsAnalysis' object.
@ -697,7 +707,10 @@ class Timeline(object):
chrome_trace=self._chrome_trace, chrome_trace=self._chrome_trace,
allocator_maximums=self._allocator_maximums) allocator_maximums=self._allocator_maximums)
def generate_chrome_trace_format(self, show_dataflow=True, show_memory=False, op_time="schedule"): def generate_chrome_trace_format(self,
show_dataflow=True,
show_memory=False,
op_time="schedule"):
"""Produces a trace in Chrome Trace Format. """Produces a trace in Chrome Trace Format.
Args: Args:
@ -707,11 +720,13 @@ class Timeline(object):
showing the sizes and lifetimes of tensors. showing the sizes and lifetimes of tensors.
op_time: (Optional.) How the execution time of op is shown in timeline. op_time: (Optional.) How the execution time of op is shown in timeline.
Possible values are "schedule", "gpu" and "all". Possible values are "schedule", "gpu" and "all".
"schedule" will show op from the time it is scheduled to the end of the scheduling. "schedule" will show op from the time it is scheduled to the end of
Notice by the end of its scheduling its async kernels may not start yet. the scheduling.
It is shown using the default value from step_stats. Notice by the end of its scheduling its async kernels may not start
yet. It is shown using the default value from step_stats.
"gpu" will show op with the execution time of its kernels on GPU. "gpu" will show op with the execution time of its kernels on GPU.
"all" will show op from the start of its scheduling to the end of its last kernel. "all" will show op from the start of its scheduling to the end of
its last kernel.
Returns: Returns:
A JSON formatted string in Chrome Trace format. A JSON formatted string in Chrome Trace format.