Last active
December 18, 2017 04:11
-
-
Save WuXianglong/880de9ef637214d62976 to your computer and use it in GitHub Desktop.
Generate a dot graph from the output of several profilers.
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
#!/usr/bin/env python | |
# | |
# Copyright 2008-2014 Jose Fonseca | |
# | |
# This program is free software: you can redistribute it and/or modify it | |
# under the terms of the GNU Lesser General Public License as published | |
# by the Free Software Foundation, either version 3 of the License, or | |
# (at your option) any later version. | |
# | |
# This program is distributed in the hope that it will be useful, | |
# but WITHOUT ANY WARRANTY; without even the implied warranty of | |
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the | |
# GNU Lesser General Public License for more details. | |
# | |
# You should have received a copy of the GNU Lesser General Public License | |
# along with this program. If not, see <http://www.gnu.org/licenses/>. | |
# | |
"""Generate a dot graph from the output of several profilers.""" | |
__author__ = "Jose Fonseca et al" | |
import sys | |
import math | |
import os.path | |
import re | |
import textwrap | |
import optparse | |
import xml.parsers.expat | |
import collections | |
import locale | |
import json | |
# Python 2.x/3.x compatibility | |
if sys.version_info[0] >= 3: | |
PYTHON_3 = True | |
def compat_iteritems(x): return x.items() # No iteritems() in Python 3 | |
def compat_itervalues(x): return x.values() # No itervalues() in Python 3 | |
def compat_keys(x): return list(x.keys()) # keys() is a generator in Python 3 | |
basestring = str # No class basestring in Python 3 | |
unichr = chr # No unichr in Python 3 | |
xrange = range # No xrange in Python 3 | |
else: | |
PYTHON_3 = False | |
def compat_iteritems(x): return x.iteritems() | |
def compat_itervalues(x): return x.itervalues() | |
def compat_keys(x): return x.keys() | |
try: | |
# Debugging helper module | |
import debug | |
except ImportError: | |
pass | |
######################################################################## | |
# Model | |
MULTIPLICATION_SIGN = unichr(0xd7) | |
def times(x): | |
return "%u%s" % (x, MULTIPLICATION_SIGN) | |
def percentage(p): | |
return "%.02f%%" % (p*100.0,) | |
def add(a, b): | |
return a + b | |
def fail(a, b): | |
assert False | |
tol = 2 ** -23 | |
def ratio(numerator, denominator): | |
try: | |
ratio = float(numerator)/float(denominator) | |
except ZeroDivisionError: | |
# 0/0 is undefined, but 1.0 yields more useful results | |
return 1.0 | |
if ratio < 0.0: | |
if ratio < -tol: | |
sys.stderr.write('warning: negative ratio (%s/%s)\n' % (numerator, denominator)) | |
return 0.0 | |
if ratio > 1.0: | |
if ratio > 1.0 + tol: | |
sys.stderr.write('warning: ratio greater than one (%s/%s)\n' % (numerator, denominator)) | |
return 1.0 | |
return ratio | |
class UndefinedEvent(Exception): | |
"""Raised when attempting to get an event which is undefined.""" | |
def __init__(self, event): | |
Exception.__init__(self) | |
self.event = event | |
def __str__(self): | |
return 'unspecified event %s' % self.event.name | |
class Event(object): | |
"""Describe a kind of event, and its basic operations.""" | |
def __init__(self, name, null, aggregator, formatter = str): | |
self.name = name | |
self._null = null | |
self._aggregator = aggregator | |
self._formatter = formatter | |
def __eq__(self, other): | |
return self is other | |
def __hash__(self): | |
return id(self) | |
def null(self): | |
return self._null | |
def aggregate(self, val1, val2): | |
"""Aggregate two event values.""" | |
assert val1 is not None | |
assert val2 is not None | |
return self._aggregator(val1, val2) | |
def format(self, val): | |
"""Format an event value.""" | |
assert val is not None | |
return self._formatter(val) | |
CALLS = Event("Calls", 0, add, times) | |
SAMPLES = Event("Samples", 0, add, times) | |
SAMPLES2 = Event("Samples", 0, add, times) | |
# Count of samples where a given function was either executing or on the stack. | |
# This is used to calculate the total time ratio according to the | |
# straightforward method described in Mike Dunlavey's answer to | |
# stackoverflow.com/questions/1777556/alternatives-to-gprof, item 4 (the myth | |
# "that recursion is a tricky confusing issue"), last edited 2012-08-30: it's | |
# just the ratio of TOTAL_SAMPLES over the number of samples in the profile. | |
# | |
# Used only when totalMethod == callstacks | |
TOTAL_SAMPLES = Event("Samples", 0, add, times) | |
TIME = Event("Time", 0.0, add, lambda x: '(' + str(x) + ')') | |
TIME_RATIO = Event("Time ratio", 0.0, add, lambda x: '(' + percentage(x) + ')') | |
TOTAL_TIME = Event("Total time", 0.0, fail) | |
TOTAL_TIME_RATIO = Event("Total time ratio", 0.0, fail, percentage) | |
totalMethod = 'callratios' | |
class Object(object): | |
"""Base class for all objects in profile which can store events.""" | |
def __init__(self, events=None): | |
if events is None: | |
self.events = {} | |
else: | |
self.events = events | |
def __hash__(self): | |
return id(self) | |
def __eq__(self, other): | |
return self is other | |
def __contains__(self, event): | |
return event in self.events | |
def __getitem__(self, event): | |
try: | |
return self.events[event] | |
except KeyError: | |
raise UndefinedEvent(event) | |
def __setitem__(self, event, value): | |
if value is None: | |
if event in self.events: | |
del self.events[event] | |
else: | |
self.events[event] = value | |
class Call(Object): | |
"""A call between functions. | |
There should be at most one call object for every pair of functions. | |
""" | |
def __init__(self, callee_id): | |
Object.__init__(self) | |
self.callee_id = callee_id | |
self.ratio = None | |
self.weight = None | |
class Function(Object): | |
"""A function.""" | |
def __init__(self, id, name): | |
Object.__init__(self) | |
self.id = id | |
self.name = name | |
self.module = None | |
self.process = None | |
self.calls = {} | |
self.called = None | |
self.weight = None | |
self.cycle = None | |
def add_call(self, call): | |
if call.callee_id in self.calls: | |
sys.stderr.write('warning: overwriting call from function %s to %s\n' % (str(self.id), str(call.callee_id))) | |
self.calls[call.callee_id] = call | |
def get_call(self, callee_id): | |
if not callee_id in self.calls: | |
call = Call(callee_id) | |
call[SAMPLES] = 0 | |
call[SAMPLES2] = 0 | |
call[CALLS] = 0 | |
self.calls[callee_id] = call | |
return self.calls[callee_id] | |
_parenthesis_re = re.compile(r'\([^()]*\)') | |
_angles_re = re.compile(r'<[^<>]*>') | |
_const_re = re.compile(r'\s+const$') | |
def stripped_name(self): | |
"""Remove extraneous information from C++ demangled function names.""" | |
name = self.name | |
# Strip function parameters from name by recursively removing paired parenthesis | |
while True: | |
name, n = self._parenthesis_re.subn('', name) | |
if not n: | |
break | |
# Strip const qualifier | |
name = self._const_re.sub('', name) | |
# Strip template parameters from name by recursively removing paired angles | |
while True: | |
name, n = self._angles_re.subn('', name) | |
if not n: | |
break | |
return name | |
# TODO: write utility functions | |
def __repr__(self): | |
return self.name | |
class Cycle(Object): | |
"""A cycle made from recursive function calls.""" | |
def __init__(self): | |
Object.__init__(self) | |
self.functions = set() | |
def add_function(self, function): | |
assert function not in self.functions | |
self.functions.add(function) | |
if function.cycle is not None: | |
for other in function.cycle.functions: | |
if function not in self.functions: | |
self.add_function(other) | |
function.cycle = self | |
class Profile(Object): | |
"""The whole profile.""" | |
def __init__(self): | |
Object.__init__(self) | |
self.functions = {} | |
self.cycles = [] | |
def add_function(self, function): | |
if function.id in self.functions: | |
sys.stderr.write('warning: overwriting function %s (id %s)\n' % (function.name, str(function.id))) | |
self.functions[function.id] = function | |
def add_cycle(self, cycle): | |
self.cycles.append(cycle) | |
def validate(self): | |
"""Validate the edges.""" | |
for function in compat_itervalues(self.functions): | |
for callee_id in compat_keys(function.calls): | |
assert function.calls[callee_id].callee_id == callee_id | |
if callee_id not in self.functions: | |
sys.stderr.write('warning: call to undefined function %s from function %s\n' % (str(callee_id), function.name)) | |
del function.calls[callee_id] | |
def find_cycles(self): | |
"""Find cycles using Tarjan's strongly connected components algorithm.""" | |
# Apply the Tarjan's algorithm successively until all functions are visited | |
visited = set() | |
for function in compat_itervalues(self.functions): | |
if function not in visited: | |
self._tarjan(function, 0, [], {}, {}, visited) | |
cycles = [] | |
for function in compat_itervalues(self.functions): | |
if function.cycle is not None and function.cycle not in cycles: | |
cycles.append(function.cycle) | |
self.cycles = cycles | |
if 0: | |
for cycle in cycles: | |
sys.stderr.write("Cycle:\n") | |
for member in cycle.functions: | |
sys.stderr.write("\tFunction %s\n" % member.name) | |
def prune_root(self, root): | |
visited = set() | |
frontier = set([root]) | |
while len(frontier) > 0: | |
node = frontier.pop() | |
visited.add(node) | |
f = self.functions[node] | |
newNodes = f.calls.keys() | |
frontier = frontier.union(set(newNodes) - visited) | |
subtreeFunctions = {} | |
for n in visited: | |
subtreeFunctions[n] = self.functions[n] | |
self.functions = subtreeFunctions | |
def prune_leaf(self, leaf): | |
edgesUp = collections.defaultdict(set) | |
for f in self.functions.keys(): | |
for n in self.functions[f].calls.keys(): | |
edgesUp[n].add(f) | |
# build the tree up | |
visited = set() | |
frontier = set([leaf]) | |
while len(frontier) > 0: | |
node = frontier.pop() | |
visited.add(node) | |
frontier = frontier.union(edgesUp[node] - visited) | |
downTree = set(self.functions.keys()) | |
upTree = visited | |
path = downTree.intersection(upTree) | |
pathFunctions = {} | |
for n in path: | |
f = self.functions[n] | |
newCalls = {} | |
for c in f.calls.keys(): | |
if c in path: | |
newCalls[c] = f.calls[c] | |
f.calls = newCalls | |
pathFunctions[n] = f | |
self.functions = pathFunctions | |
def getFunctionId(self, funcName): | |
for f in self.functions: | |
if self.functions[f].name == funcName: | |
return f | |
return False | |
def _tarjan(self, function, order, stack, orders, lowlinks, visited): | |
"""Tarjan's strongly connected components algorithm. | |
See also: | |
- http://en.wikipedia.org/wiki/Tarjan's_strongly_connected_components_algorithm | |
""" | |
visited.add(function) | |
orders[function] = order | |
lowlinks[function] = order | |
order += 1 | |
pos = len(stack) | |
stack.append(function) | |
for call in compat_itervalues(function.calls): | |
callee = self.functions[call.callee_id] | |
# TODO: use a set to optimize lookup | |
if callee not in orders: | |
order = self._tarjan(callee, order, stack, orders, lowlinks, visited) | |
lowlinks[function] = min(lowlinks[function], lowlinks[callee]) | |
elif callee in stack: | |
lowlinks[function] = min(lowlinks[function], orders[callee]) | |
if lowlinks[function] == orders[function]: | |
# Strongly connected component found | |
members = stack[pos:] | |
del stack[pos:] | |
if len(members) > 1: | |
cycle = Cycle() | |
for member in members: | |
cycle.add_function(member) | |
return order | |
def call_ratios(self, event): | |
# Aggregate for incoming calls | |
cycle_totals = {} | |
for cycle in self.cycles: | |
cycle_totals[cycle] = 0.0 | |
function_totals = {} | |
for function in compat_itervalues(self.functions): | |
function_totals[function] = 0.0 | |
# Pass 1: function_total gets the sum of call[event] for all | |
# incoming arrows. Same for cycle_total for all arrows | |
# that are coming into the *cycle* but are not part of it. | |
for function in compat_itervalues(self.functions): | |
for call in compat_itervalues(function.calls): | |
if call.callee_id != function.id: | |
callee = self.functions[call.callee_id] | |
if event in call.events: | |
function_totals[callee] += call[event] | |
if callee.cycle is not None and callee.cycle is not function.cycle: | |
cycle_totals[callee.cycle] += call[event] | |
else: | |
sys.stderr.write("call_ratios: No data for " + function.name + " call to " + callee.name + "\n") | |
# Pass 2: Compute the ratios. Each call[event] is scaled by the | |
# function_total of the callee. Calls into cycles use the | |
# cycle_total, but not calls within cycles. | |
for function in compat_itervalues(self.functions): | |
for call in compat_itervalues(function.calls): | |
assert call.ratio is None | |
if call.callee_id != function.id: | |
callee = self.functions[call.callee_id] | |
if event in call.events: | |
if callee.cycle is not None and callee.cycle is not function.cycle: | |
total = cycle_totals[callee.cycle] | |
else: | |
total = function_totals[callee] | |
call.ratio = ratio(call[event], total) | |
else: | |
# Warnings here would only repeat those issued above. | |
call.ratio = 0.0 | |
def integrate(self, outevent, inevent): | |
"""Propagate function time ratio along the function calls. | |
Must be called after finding the cycles. | |
See also: | |
- http://citeseer.ist.psu.edu/graham82gprof.html | |
""" | |
# Sanity checking | |
assert outevent not in self | |
for function in compat_itervalues(self.functions): | |
assert outevent not in function | |
assert inevent in function | |
for call in compat_itervalues(function.calls): | |
assert outevent not in call | |
if call.callee_id != function.id: | |
assert call.ratio is not None | |
# Aggregate the input for each cycle | |
for cycle in self.cycles: | |
total = inevent.null() | |
for function in compat_itervalues(self.functions): | |
total = inevent.aggregate(total, function[inevent]) | |
self[inevent] = total | |
# Integrate along the edges | |
total = inevent.null() | |
for function in compat_itervalues(self.functions): | |
total = inevent.aggregate(total, function[inevent]) | |
self._integrate_function(function, outevent, inevent) | |
self[outevent] = total | |
def _integrate_function(self, function, outevent, inevent): | |
if function.cycle is not None: | |
return self._integrate_cycle(function.cycle, outevent, inevent) | |
else: | |
if outevent not in function: | |
total = function[inevent] | |
for call in compat_itervalues(function.calls): | |
if call.callee_id != function.id: | |
total += self._integrate_call(call, outevent, inevent) | |
function[outevent] = total | |
return function[outevent] | |
def _integrate_call(self, call, outevent, inevent): | |
assert outevent not in call | |
assert call.ratio is not None | |
callee = self.functions[call.callee_id] | |
subtotal = call.ratio *self._integrate_function(callee, outevent, inevent) | |
call[outevent] = subtotal | |
return subtotal | |
def _integrate_cycle(self, cycle, outevent, inevent): | |
if outevent not in cycle: | |
# Compute the outevent for the whole cycle | |
total = inevent.null() | |
for member in cycle.functions: | |
subtotal = member[inevent] | |
for call in compat_itervalues(member.calls): | |
callee = self.functions[call.callee_id] | |
if callee.cycle is not cycle: | |
subtotal += self._integrate_call(call, outevent, inevent) | |
total += subtotal | |
cycle[outevent] = total | |
# Compute the time propagated to callers of this cycle | |
callees = {} | |
for function in compat_itervalues(self.functions): | |
if function.cycle is not cycle: | |
for call in compat_itervalues(function.calls): | |
callee = self.functions[call.callee_id] | |
if callee.cycle is cycle: | |
try: | |
callees[callee] += call.ratio | |
except KeyError: | |
callees[callee] = call.ratio | |
for member in cycle.functions: | |
member[outevent] = outevent.null() | |
for callee, call_ratio in compat_iteritems(callees): | |
ranks = {} | |
call_ratios = {} | |
partials = {} | |
self._rank_cycle_function(cycle, callee, 0, ranks) | |
self._call_ratios_cycle(cycle, callee, ranks, call_ratios, set()) | |
partial = self._integrate_cycle_function(cycle, callee, call_ratio, partials, ranks, call_ratios, outevent, inevent) | |
assert partial == max(partials.values()) | |
assert abs(call_ratio*total - partial) <= 0.001*call_ratio*total | |
return cycle[outevent] | |
def _rank_cycle_function(self, cycle, function, rank, ranks): | |
if function not in ranks or ranks[function] > rank: | |
ranks[function] = rank | |
for call in compat_itervalues(function.calls): | |
if call.callee_id != function.id: | |
callee = self.functions[call.callee_id] | |
if callee.cycle is cycle: | |
self._rank_cycle_function(cycle, callee, rank + 1, ranks) | |
def _call_ratios_cycle(self, cycle, function, ranks, call_ratios, visited): | |
if function not in visited: | |
visited.add(function) | |
for call in compat_itervalues(function.calls): | |
if call.callee_id != function.id: | |
callee = self.functions[call.callee_id] | |
if callee.cycle is cycle: | |
if ranks[callee] > ranks[function]: | |
call_ratios[callee] = call_ratios.get(callee, 0.0) + call.ratio | |
self._call_ratios_cycle(cycle, callee, ranks, call_ratios, visited) | |
def _integrate_cycle_function(self, cycle, function, partial_ratio, partials, ranks, call_ratios, outevent, inevent): | |
if function not in partials: | |
partial = partial_ratio*function[inevent] | |
for call in compat_itervalues(function.calls): | |
if call.callee_id != function.id: | |
callee = self.functions[call.callee_id] | |
if callee.cycle is not cycle: | |
assert outevent in call | |
partial += partial_ratio*call[outevent] | |
else: | |
if ranks[callee] > ranks[function]: | |
callee_partial = self._integrate_cycle_function(cycle, callee, partial_ratio, partials, ranks, call_ratios, outevent, inevent) | |
call_ratio = ratio(call.ratio, call_ratios[callee]) | |
call_partial = call_ratio*callee_partial | |
try: | |
call[outevent] += call_partial | |
except UndefinedEvent: | |
call[outevent] = call_partial | |
partial += call_partial | |
partials[function] = partial | |
try: | |
function[outevent] += partial | |
except UndefinedEvent: | |
function[outevent] = partial | |
return partials[function] | |
def aggregate(self, event): | |
"""Aggregate an event for the whole profile.""" | |
total = event.null() | |
for function in compat_itervalues(self.functions): | |
try: | |
total = event.aggregate(total, function[event]) | |
except UndefinedEvent: | |
return | |
self[event] = total | |
def ratio(self, outevent, inevent): | |
assert outevent not in self | |
assert inevent in self | |
for function in compat_itervalues(self.functions): | |
assert outevent not in function | |
assert inevent in function | |
function[outevent] = ratio(function[inevent], self[inevent]) | |
for call in compat_itervalues(function.calls): | |
assert outevent not in call | |
if inevent in call: | |
call[outevent] = ratio(call[inevent], self[inevent]) | |
self[outevent] = 1.0 | |
def prune(self, node_thres, edge_thres): | |
"""Prune the profile""" | |
# compute the prune ratios | |
for function in compat_itervalues(self.functions): | |
try: | |
function.weight = function[TOTAL_TIME_RATIO] | |
except UndefinedEvent: | |
pass | |
for call in compat_itervalues(function.calls): | |
callee = self.functions[call.callee_id] | |
if TOTAL_TIME_RATIO in call: | |
# handle exact cases first | |
call.weight = call[TOTAL_TIME_RATIO] | |
else: | |
try: | |
# make a safe estimate | |
call.weight = min(function[TOTAL_TIME_RATIO], callee[TOTAL_TIME_RATIO]) | |
except UndefinedEvent: | |
pass | |
# prune the nodes | |
for function_id in compat_keys(self.functions): | |
function = self.functions[function_id] | |
if function.weight is not None: | |
if function.weight < node_thres: | |
del self.functions[function_id] | |
# prune the egdes | |
for function in compat_itervalues(self.functions): | |
for callee_id in compat_keys(function.calls): | |
call = function.calls[callee_id] | |
if callee_id not in self.functions or call.weight is not None and call.weight < edge_thres: | |
del function.calls[callee_id] | |
def dump(self): | |
for function in compat_itervalues(self.functions): | |
sys.stderr.write('Function %s:\n' % (function.name,)) | |
self._dump_events(function.events) | |
for call in compat_itervalues(function.calls): | |
callee = self.functions[call.callee_id] | |
sys.stderr.write(' Call %s:\n' % (callee.name,)) | |
self._dump_events(call.events) | |
for cycle in self.cycles: | |
sys.stderr.write('Cycle:\n') | |
self._dump_events(cycle.events) | |
for function in cycle.functions: | |
sys.stderr.write(' Function %s\n' % (function.name,)) | |
def _dump_events(self, events): | |
for event, value in compat_iteritems(events): | |
sys.stderr.write(' %s: %s\n' % (event.name, event.format(value))) | |
######################################################################## | |
# Parsers | |
class Struct: | |
"""Masquerade a dictionary with a structure-like behavior.""" | |
def __init__(self, attrs = None): | |
if attrs is None: | |
attrs = {} | |
self.__dict__['_attrs'] = attrs | |
def __getattr__(self, name): | |
try: | |
return self._attrs[name] | |
except KeyError: | |
raise AttributeError(name) | |
def __setattr__(self, name, value): | |
self._attrs[name] = value | |
def __str__(self): | |
return str(self._attrs) | |
def __repr__(self): | |
return repr(self._attrs) | |
class ParseError(Exception): | |
"""Raised when parsing to signal mismatches.""" | |
def __init__(self, msg, line): | |
Exception.__init__(self) | |
self.msg = msg | |
# TODO: store more source line information | |
self.line = line | |
def __str__(self): | |
return '%s: %r' % (self.msg, self.line) | |
class Parser: | |
"""Parser interface.""" | |
stdinInput = True | |
multipleInput = False | |
def __init__(self): | |
pass | |
def parse(self): | |
raise NotImplementedError | |
class JsonParser(Parser): | |
"""Parser for a custom JSON representation of profile data. | |
See schema.json for details. | |
""" | |
def __init__(self, stream): | |
Parser.__init__(self) | |
self.stream = stream | |
def parse(self): | |
obj = json.load(self.stream) | |
assert obj['version'] == 0 | |
profile = Profile() | |
profile[SAMPLES] = 0 | |
fns = obj['functions'] | |
for functionIndex in range(len(fns)): | |
fn = fns[functionIndex] | |
function = Function(functionIndex, fn['name']) | |
try: | |
function.module = fn['module'] | |
except KeyError: | |
pass | |
try: | |
function.process = fn['process'] | |
except KeyError: | |
pass | |
function[SAMPLES] = 0 | |
profile.add_function(function) | |
for event in obj['events']: | |
callchain = [] | |
for functionIndex in event['callchain']: | |
function = profile.functions[functionIndex] | |
callchain.append(function) | |
cost = event['cost'][0] | |
callee = callchain[0] | |
callee[SAMPLES] += cost | |
profile[SAMPLES] += cost | |
for caller in callchain[1:]: | |
try: | |
call = caller.calls[callee.id] | |
except KeyError: | |
call = Call(callee.id) | |
call[SAMPLES2] = cost | |
caller.add_call(call) | |
else: | |
call[SAMPLES2] += cost | |
callee = caller | |
if False: | |
profile.dump() | |
# compute derived data | |
profile.validate() | |
profile.find_cycles() | |
profile.ratio(TIME_RATIO, SAMPLES) | |
profile.call_ratios(SAMPLES2) | |
profile.integrate(TOTAL_TIME_RATIO, TIME_RATIO) | |
return profile | |
class LineParser(Parser): | |
"""Base class for parsers that read line-based formats.""" | |
def __init__(self, stream): | |
Parser.__init__(self) | |
self._stream = stream | |
self.__line = None | |
self.__eof = False | |
self.line_no = 0 | |
def readline(self): | |
line = self._stream.readline() | |
if not line: | |
self.__line = '' | |
self.__eof = True | |
else: | |
self.line_no += 1 | |
line = line.rstrip('\r\n') | |
if not PYTHON_3: | |
encoding = self._stream.encoding | |
if encoding is None: | |
encoding = locale.getpreferredencoding() | |
line = line.decode(encoding) | |
self.__line = line | |
def lookahead(self): | |
assert self.__line is not None | |
return self.__line | |
def consume(self): | |
assert self.__line is not None | |
line = self.__line | |
self.readline() | |
return line | |
def eof(self): | |
assert self.__line is not None | |
return self.__eof | |
XML_ELEMENT_START, XML_ELEMENT_END, XML_CHARACTER_DATA, XML_EOF = range(4) | |
class XmlToken: | |
def __init__(self, type, name_or_data, attrs = None, line = None, column = None): | |
assert type in (XML_ELEMENT_START, XML_ELEMENT_END, XML_CHARACTER_DATA, XML_EOF) | |
self.type = type | |
self.name_or_data = name_or_data | |
self.attrs = attrs | |
self.line = line | |
self.column = column | |
def __str__(self): | |
if self.type == XML_ELEMENT_START: | |
return '<' + self.name_or_data + ' ...>' | |
if self.type == XML_ELEMENT_END: | |
return '</' + self.name_or_data + '>' | |
if self.type == XML_CHARACTER_DATA: | |
return self.name_or_data | |
if self.type == XML_EOF: | |
return 'end of file' | |
assert 0 | |
class XmlTokenizer: | |
"""Expat based XML tokenizer.""" | |
def __init__(self, fp, skip_ws = True): | |
self.fp = fp | |
self.tokens = [] | |
self.index = 0 | |
self.final = False | |
self.skip_ws = skip_ws | |
self.character_pos = 0, 0 | |
self.character_data = '' | |
self.parser = xml.parsers.expat.ParserCreate() | |
self.parser.StartElementHandler = self.handle_element_start | |
self.parser.EndElementHandler = self.handle_element_end | |
self.parser.CharacterDataHandler = self.handle_character_data | |
def handle_element_start(self, name, attributes): | |
self.finish_character_data() | |
line, column = self.pos() | |
token = XmlToken(XML_ELEMENT_START, name, attributes, line, column) | |
self.tokens.append(token) | |
def handle_element_end(self, name): | |
self.finish_character_data() | |
line, column = self.pos() | |
token = XmlToken(XML_ELEMENT_END, name, None, line, column) | |
self.tokens.append(token) | |
def handle_character_data(self, data): | |
if not self.character_data: | |
self.character_pos = self.pos() | |
self.character_data += data | |
def finish_character_data(self): | |
if self.character_data: | |
if not self.skip_ws or not self.character_data.isspace(): | |
line, column = self.character_pos | |
token = XmlToken(XML_CHARACTER_DATA, self.character_data, None, line, column) | |
self.tokens.append(token) | |
self.character_data = '' | |
def next(self): | |
size = 16*1024 | |
while self.index >= len(self.tokens) and not self.final: | |
self.tokens = [] | |
self.index = 0 | |
data = self.fp.read(size) | |
self.final = len(data) < size | |
self.parser.Parse(data, self.final) | |
if self.index >= len(self.tokens): | |
line, column = self.pos() | |
token = XmlToken(XML_EOF, None, None, line, column) | |
else: | |
token = self.tokens[self.index] | |
self.index += 1 | |
return token | |
def pos(self): | |
return self.parser.CurrentLineNumber, self.parser.CurrentColumnNumber | |
class XmlTokenMismatch(Exception): | |
def __init__(self, expected, found): | |
Exception.__init__(self) | |
self.expected = expected | |
self.found = found | |
def __str__(self): | |
return '%u:%u: %s expected, %s found' % (self.found.line, self.found.column, str(self.expected), str(self.found)) | |
class XmlParser(Parser): | |
"""Base XML document parser.""" | |
def __init__(self, fp): | |
Parser.__init__(self) | |
self.tokenizer = XmlTokenizer(fp) | |
self.consume() | |
def consume(self): | |
self.token = self.tokenizer.next() | |
def match_element_start(self, name): | |
return self.token.type == XML_ELEMENT_START and self.token.name_or_data == name | |
def match_element_end(self, name): | |
return self.token.type == XML_ELEMENT_END and self.token.name_or_data == name | |
def element_start(self, name): | |
while self.token.type == XML_CHARACTER_DATA: | |
self.consume() | |
if self.token.type != XML_ELEMENT_START: | |
raise XmlTokenMismatch(XmlToken(XML_ELEMENT_START, name), self.token) | |
if self.token.name_or_data != name: | |
raise XmlTokenMismatch(XmlToken(XML_ELEMENT_START, name), self.token) | |
attrs = self.token.attrs | |
self.consume() | |
return attrs | |
def element_end(self, name): | |
while self.token.type == XML_CHARACTER_DATA: | |
self.consume() | |
if self.token.type != XML_ELEMENT_END: | |
raise XmlTokenMismatch(XmlToken(XML_ELEMENT_END, name), self.token) | |
if self.token.name_or_data != name: | |
raise XmlTokenMismatch(XmlToken(XML_ELEMENT_END, name), self.token) | |
self.consume() | |
def character_data(self, strip = True): | |
data = '' | |
while self.token.type == XML_CHARACTER_DATA: | |
data += self.token.name_or_data | |
self.consume() | |
if strip: | |
data = data.strip() | |
return data | |
class GprofParser(Parser): | |
"""Parser for GNU gprof output. | |
See also: | |
- Chapter "Interpreting gprof's Output" from the GNU gprof manual | |
http://sourceware.org/binutils/docs-2.18/gprof/Call-Graph.html#Call-Graph | |
- File "cg_print.c" from the GNU gprof source code | |
http://sourceware.org/cgi-bin/cvsweb.cgi/~checkout~/src/gprof/cg_print.c?rev=1.12&cvsroot=src | |
""" | |
def __init__(self, fp): | |
Parser.__init__(self) | |
self.fp = fp | |
self.functions = {} | |
self.cycles = {} | |
def readline(self): | |
line = self.fp.readline() | |
if not line: | |
sys.stderr.write('error: unexpected end of file\n') | |
sys.exit(1) | |
line = line.rstrip('\r\n') | |
return line | |
_int_re = re.compile(r'^\d+$') | |
_float_re = re.compile(r'^\d+\.\d+$') | |
def translate(self, mo): | |
"""Extract a structure from a match object, while translating the types in the process.""" | |
attrs = {} | |
groupdict = mo.groupdict() | |
for name, value in compat_iteritems(groupdict): | |
if value is None: | |
value = None | |
elif self._int_re.match(value): | |
value = int(value) | |
elif self._float_re.match(value): | |
value = float(value) | |
attrs[name] = (value) | |
return Struct(attrs) | |
_cg_header_re = re.compile( | |
# original gprof header | |
r'^\s+called/total\s+parents\s*$|' + | |
r'^index\s+%time\s+self\s+descendents\s+called\+self\s+name\s+index\s*$|' + | |
r'^\s+called/total\s+children\s*$|' + | |
# GNU gprof header | |
r'^index\s+%\s+time\s+self\s+children\s+called\s+name\s*$' | |
) | |
_cg_ignore_re = re.compile( | |
# spontaneous | |
r'^\s+<spontaneous>\s*$|' | |
# internal calls (such as "mcount") | |
r'^.*\((\d+)\)$' | |
) | |
_cg_primary_re = re.compile( | |
r'^\[(?P<index>\d+)\]?' + | |
r'\s+(?P<percentage_time>\d+\.\d+)' + | |
r'\s+(?P<self>\d+\.\d+)' + | |
r'\s+(?P<descendants>\d+\.\d+)' + | |
r'\s+(?:(?P<called>\d+)(?:\+(?P<called_self>\d+))?)?' + | |
r'\s+(?P<name>\S.*?)' + | |
r'(?:\s+<cycle\s(?P<cycle>\d+)>)?' + | |
r'\s\[(\d+)\]$' | |
) | |
_cg_parent_re = re.compile( | |
r'^\s+(?P<self>\d+\.\d+)?' + | |
r'\s+(?P<descendants>\d+\.\d+)?' + | |
r'\s+(?P<called>\d+)(?:/(?P<called_total>\d+))?' + | |
r'\s+(?P<name>\S.*?)' + | |
r'(?:\s+<cycle\s(?P<cycle>\d+)>)?' + | |
r'\s\[(?P<index>\d+)\]$' | |
) | |
_cg_child_re = _cg_parent_re | |
_cg_cycle_header_re = re.compile( | |
r'^\[(?P<index>\d+)\]?' + | |
r'\s+(?P<percentage_time>\d+\.\d+)' + | |
r'\s+(?P<self>\d+\.\d+)' + | |
r'\s+(?P<descendants>\d+\.\d+)' + | |
r'\s+(?:(?P<called>\d+)(?:\+(?P<called_self>\d+))?)?' + | |
r'\s+<cycle\s(?P<cycle>\d+)\sas\sa\swhole>' + | |
r'\s\[(\d+)\]$' | |
) | |
_cg_cycle_member_re = re.compile( | |
r'^\s+(?P<self>\d+\.\d+)?' + | |
r'\s+(?P<descendants>\d+\.\d+)?' + | |
r'\s+(?P<called>\d+)(?:\+(?P<called_self>\d+))?' + | |
r'\s+(?P<name>\S.*?)' + | |
r'(?:\s+<cycle\s(?P<cycle>\d+)>)?' + | |
r'\s\[(?P<index>\d+)\]$' | |
) | |
_cg_sep_re = re.compile(r'^--+$') | |
def parse_function_entry(self, lines): | |
parents = [] | |
children = [] | |
while True: | |
if not lines: | |
sys.stderr.write('warning: unexpected end of entry\n') | |
line = lines.pop(0) | |
if line.startswith('['): | |
break | |
# read function parent line | |
mo = self._cg_parent_re.match(line) | |
if not mo: | |
if self._cg_ignore_re.match(line): | |
continue | |
sys.stderr.write('warning: unrecognized call graph entry: %r\n' % line) | |
else: | |
parent = self.translate(mo) | |
parents.append(parent) | |
# read primary line | |
mo = self._cg_primary_re.match(line) | |
if not mo: | |
sys.stderr.write('warning: unrecognized call graph entry: %r\n' % line) | |
return | |
else: | |
function = self.translate(mo) | |
while lines: | |
line = lines.pop(0) | |
# read function subroutine line | |
mo = self._cg_child_re.match(line) | |
if not mo: | |
if self._cg_ignore_re.match(line): | |
continue | |
sys.stderr.write('warning: unrecognized call graph entry: %r\n' % line) | |
else: | |
child = self.translate(mo) | |
children.append(child) | |
function.parents = parents | |
function.children = children | |
self.functions[function.index] = function | |
def parse_cycle_entry(self, lines): | |
# read cycle header line | |
line = lines[0] | |
mo = self._cg_cycle_header_re.match(line) | |
if not mo: | |
sys.stderr.write('warning: unrecognized call graph entry: %r\n' % line) | |
return | |
cycle = self.translate(mo) | |
# read cycle member lines | |
cycle.functions = [] | |
for line in lines[1:]: | |
mo = self._cg_cycle_member_re.match(line) | |
if not mo: | |
sys.stderr.write('warning: unrecognized call graph entry: %r\n' % line) | |
continue | |
call = self.translate(mo) | |
cycle.functions.append(call) | |
self.cycles[cycle.cycle] = cycle | |
def parse_cg_entry(self, lines): | |
if lines[0].startswith("["): | |
self.parse_cycle_entry(lines) | |
else: | |
self.parse_function_entry(lines) | |
def parse_cg(self): | |
"""Parse the call graph.""" | |
# skip call graph header | |
while not self._cg_header_re.match(self.readline()): | |
pass | |
line = self.readline() | |
while self._cg_header_re.match(line): | |
line = self.readline() | |
# process call graph entries | |
entry_lines = [] | |
while line != '\014': # form feed | |
if line and not line.isspace(): | |
if self._cg_sep_re.match(line): | |
self.parse_cg_entry(entry_lines) | |
entry_lines = [] | |
else: | |
entry_lines.append(line) | |
line = self.readline() | |
def parse(self): | |
self.parse_cg() | |
self.fp.close() | |
profile = Profile() | |
profile[TIME] = 0.0 | |
cycles = {} | |
for index in self.cycles: | |
cycles[index] = Cycle() | |
for entry in compat_itervalues(self.functions): | |
# populate the function | |
function = Function(entry.index, entry.name) | |
function[TIME] = entry.self | |
if entry.called is not None: | |
function.called = entry.called | |
if entry.called_self is not None: | |
call = Call(entry.index) | |
call[CALLS] = entry.called_self | |
function.called += entry.called_self | |
# populate the function calls | |
for child in entry.children: | |
call = Call(child.index) | |
assert child.called is not None | |
call[CALLS] = child.called | |
if child.index not in self.functions: | |
# NOTE: functions that were never called but were discovered by gprof's | |
# static call graph analysis dont have a call graph entry so we need | |
# to add them here | |
missing = Function(child.index, child.name) | |
function[TIME] = 0.0 | |
function.called = 0 | |
profile.add_function(missing) | |
function.add_call(call) | |
profile.add_function(function) | |
if entry.cycle is not None: | |
try: | |
cycle = cycles[entry.cycle] | |
except KeyError: | |
sys.stderr.write('warning: <cycle %u as a whole> entry missing\n' % entry.cycle) | |
cycle = Cycle() | |
cycles[entry.cycle] = cycle | |
cycle.add_function(function) | |
profile[TIME] = profile[TIME] + function[TIME] | |
for cycle in compat_itervalues(cycles): | |
profile.add_cycle(cycle) | |
# Compute derived events | |
profile.validate() | |
profile.ratio(TIME_RATIO, TIME) | |
profile.call_ratios(CALLS) | |
profile.integrate(TOTAL_TIME, TIME) | |
profile.ratio(TOTAL_TIME_RATIO, TOTAL_TIME) | |
return profile | |
# Clone&hack of GprofParser for VTune Amplifier XE 2013 gprof-cc output. | |
# Tested only with AXE 2013 for Windows. | |
# - Use total times as reported by AXE. | |
# - In the absence of call counts, call ratios are faked from the relative | |
# proportions of total time. This affects only the weighting of the calls. | |
# - Different header, separator, and end marker. | |
# - Extra whitespace after function names. | |
# - You get a full entry for <spontaneous>, which does not have parents. | |
# - Cycles do have parents. These are saved but unused (as they are | |
# for functions). | |
# - Disambiguated "unrecognized call graph entry" error messages. | |
# Notes: | |
# - Total time of functions as reported by AXE passes the val3 test. | |
# - CPU Time:Children in the input is sometimes a negative number. This | |
# value goes to the variable descendants, which is unused. | |
# - The format of gprof-cc reports is unaffected by the use of | |
# -knob enable-call-counts=true (no call counts, ever), or | |
# -show-as=samples (results are quoted in seconds regardless). | |
class AXEParser(Parser): | |
"Parser for VTune Amplifier XE 2013 gprof-cc report output." | |
def __init__(self, fp): | |
Parser.__init__(self) | |
self.fp = fp | |
self.functions = {} | |
self.cycles = {} | |
def readline(self): | |
line = self.fp.readline() | |
if not line: | |
sys.stderr.write('error: unexpected end of file\n') | |
sys.exit(1) | |
line = line.rstrip('\r\n') | |
return line | |
_int_re = re.compile(r'^\d+$') | |
_float_re = re.compile(r'^\d+\.\d+$') | |
def translate(self, mo): | |
"""Extract a structure from a match object, while translating the types in the process.""" | |
attrs = {} | |
groupdict = mo.groupdict() | |
for name, value in compat_iteritems(groupdict): | |
if value is None: | |
value = None | |
elif self._int_re.match(value): | |
value = int(value) | |
elif self._float_re.match(value): | |
value = float(value) | |
attrs[name] = (value) | |
return Struct(attrs) | |
_cg_header_re = re.compile( | |
'^Index |' | |
'^-----+ ' | |
) | |
_cg_footer_re = re.compile(r'^Index\s+Function\s*$') | |
_cg_primary_re = re.compile( | |
r'^\[(?P<index>\d+)\]?' + | |
r'\s+(?P<percentage_time>\d+\.\d+)' + | |
r'\s+(?P<self>\d+\.\d+)' + | |
r'\s+(?P<descendants>\d+\.\d+)' + | |
r'\s+(?P<name>\S.*?)' + | |
r'(?:\s+<cycle\s(?P<cycle>\d+)>)?' + | |
r'\s+\[(\d+)\]$' | |
) | |
_cg_parent_re = re.compile( | |
r'^\s+(?P<self>\d+\.\d+)?' + | |
r'\s+(?P<descendants>\d+\.\d+)?' + | |
r'\s+(?P<name>\S.*?)' + | |
r'(?:\s+<cycle\s(?P<cycle>\d+)>)?' + | |
r'\s+\[(?P<index>\d+)\]$' | |
) | |
_cg_child_re = _cg_parent_re | |
_cg_cycle_header_re = re.compile( | |
r'^\[(?P<index>\d+)\]?' + | |
r'\s+(?P<percentage_time>\d+\.\d+)' + | |
r'\s+(?P<self>\d+\.\d+)' + | |
r'\s+(?P<descendants>\d+\.\d+)' + | |
r'\s+<cycle\s(?P<cycle>\d+)\sas\sa\swhole>' + | |
r'\s+\[(\d+)\]$' | |
) | |
_cg_cycle_member_re = re.compile( | |
r'^\s+(?P<self>\d+\.\d+)?' + | |
r'\s+(?P<descendants>\d+\.\d+)?' + | |
r'\s+(?P<name>\S.*?)' + | |
r'(?:\s+<cycle\s(?P<cycle>\d+)>)?' + | |
r'\s+\[(?P<index>\d+)\]$' | |
) | |
def parse_function_entry(self, lines): | |
parents = [] | |
children = [] | |
while True: | |
if not lines: | |
sys.stderr.write('warning: unexpected end of entry\n') | |
return | |
line = lines.pop(0) | |
if line.startswith('['): | |
break | |
# read function parent line | |
mo = self._cg_parent_re.match(line) | |
if not mo: | |
sys.stderr.write('warning: unrecognized call graph entry (1): %r\n' % line) | |
else: | |
parent = self.translate(mo) | |
if parent.name != '<spontaneous>': | |
parents.append(parent) | |
# read primary line | |
mo = self._cg_primary_re.match(line) | |
if not mo: | |
sys.stderr.write('warning: unrecognized call graph entry (2): %r\n' % line) | |
return | |
else: | |
function = self.translate(mo) | |
while lines: | |
line = lines.pop(0) | |
# read function subroutine line | |
mo = self._cg_child_re.match(line) | |
if not mo: | |
sys.stderr.write('warning: unrecognized call graph entry (3): %r\n' % line) | |
else: | |
child = self.translate(mo) | |
if child.name != '<spontaneous>': | |
children.append(child) | |
if function.name != '<spontaneous>': | |
function.parents = parents | |
function.children = children | |
self.functions[function.index] = function | |
def parse_cycle_entry(self, lines): | |
# Process the parents that were not there in gprof format. | |
parents = [] | |
while True: | |
if not lines: | |
sys.stderr.write('warning: unexpected end of cycle entry\n') | |
return | |
line = lines.pop(0) | |
if line.startswith('['): | |
break | |
mo = self._cg_parent_re.match(line) | |
if not mo: | |
sys.stderr.write('warning: unrecognized call graph entry (6): %r\n' % line) | |
else: | |
parent = self.translate(mo) | |
if parent.name != '<spontaneous>': | |
parents.append(parent) | |
# read cycle header line | |
mo = self._cg_cycle_header_re.match(line) | |
if not mo: | |
sys.stderr.write('warning: unrecognized call graph entry (4): %r\n' % line) | |
return | |
cycle = self.translate(mo) | |
# read cycle member lines | |
cycle.functions = [] | |
for line in lines[1:]: | |
mo = self._cg_cycle_member_re.match(line) | |
if not mo: | |
sys.stderr.write('warning: unrecognized call graph entry (5): %r\n' % line) | |
continue | |
call = self.translate(mo) | |
cycle.functions.append(call) | |
cycle.parents = parents | |
self.cycles[cycle.cycle] = cycle | |
def parse_cg_entry(self, lines): | |
if any("as a whole" in linelooper for linelooper in lines): | |
self.parse_cycle_entry(lines) | |
else: | |
self.parse_function_entry(lines) | |
def parse_cg(self): | |
"""Parse the call graph.""" | |
# skip call graph header | |
line = self.readline() | |
while self._cg_header_re.match(line): | |
line = self.readline() | |
# process call graph entries | |
entry_lines = [] | |
# An EOF in readline terminates the program without returning. | |
while not self._cg_footer_re.match(line): | |
if line.isspace(): | |
self.parse_cg_entry(entry_lines) | |
entry_lines = [] | |
else: | |
entry_lines.append(line) | |
line = self.readline() | |
def parse(self): | |
sys.stderr.write('warning: for axe format, edge weights are unreliable estimates derived from\nfunction total times.\n') | |
self.parse_cg() | |
self.fp.close() | |
profile = Profile() | |
profile[TIME] = 0.0 | |
cycles = {} | |
for index in self.cycles: | |
cycles[index] = Cycle() | |
for entry in compat_itervalues(self.functions): | |
# populate the function | |
function = Function(entry.index, entry.name) | |
function[TIME] = entry.self | |
function[TOTAL_TIME_RATIO] = entry.percentage_time / 100.0 | |
# populate the function calls | |
for child in entry.children: | |
call = Call(child.index) | |
# The following bogus value affects only the weighting of | |
# the calls. | |
call[TOTAL_TIME_RATIO] = function[TOTAL_TIME_RATIO] | |
if child.index not in self.functions: | |
# NOTE: functions that were never called but were discovered by gprof's | |
# static call graph analysis dont have a call graph entry so we need | |
# to add them here | |
# FIXME: Is this applicable? | |
missing = Function(child.index, child.name) | |
function[TIME] = 0.0 | |
profile.add_function(missing) | |
function.add_call(call) | |
profile.add_function(function) | |
if entry.cycle is not None: | |
try: | |
cycle = cycles[entry.cycle] | |
except KeyError: | |
sys.stderr.write('warning: <cycle %u as a whole> entry missing\n' % entry.cycle) | |
cycle = Cycle() | |
cycles[entry.cycle] = cycle | |
cycle.add_function(function) | |
profile[TIME] = profile[TIME] + function[TIME] | |
for cycle in compat_itervalues(cycles): | |
profile.add_cycle(cycle) | |
# Compute derived events. | |
profile.validate() | |
profile.ratio(TIME_RATIO, TIME) | |
# Lacking call counts, fake call ratios based on total times. | |
profile.call_ratios(TOTAL_TIME_RATIO) | |
# The TOTAL_TIME_RATIO of functions is already set. Propagate that | |
# total time to the calls. (TOTAL_TIME is neither set nor used.) | |
for function in compat_itervalues(profile.functions): | |
for call in compat_itervalues(function.calls): | |
if call.ratio is not None: | |
callee = profile.functions[call.callee_id] | |
call[TOTAL_TIME_RATIO] = call.ratio * callee[TOTAL_TIME_RATIO] | |
return profile | |
class CallgrindParser(LineParser): | |
"""Parser for valgrind's callgrind tool. | |
See also: | |
- http://valgrind.org/docs/manual/cl-format.html | |
""" | |
_call_re = re.compile(r'^calls=\s*(\d+)\s+((\d+|\+\d+|-\d+|\*)\s+)+$') | |
def __init__(self, infile): | |
LineParser.__init__(self, infile) | |
# Textual positions | |
self.position_ids = {} | |
self.positions = {} | |
# Numeric positions | |
self.num_positions = 1 | |
self.cost_positions = ['line'] | |
self.last_positions = [0] | |
# Events | |
self.num_events = 0 | |
self.cost_events = [] | |
self.profile = Profile() | |
self.profile[SAMPLES] = 0 | |
def parse(self): | |
# read lookahead | |
self.readline() | |
self.parse_key('version') | |
self.parse_key('creator') | |
while self.parse_part(): | |
pass | |
if not self.eof(): | |
sys.stderr.write('warning: line %u: unexpected line\n' % self.line_no) | |
sys.stderr.write('%s\n' % self.lookahead()) | |
# compute derived data | |
self.profile.validate() | |
self.profile.find_cycles() | |
self.profile.ratio(TIME_RATIO, SAMPLES) | |
self.profile.call_ratios(SAMPLES2) | |
self.profile.integrate(TOTAL_TIME_RATIO, TIME_RATIO) | |
return self.profile | |
def parse_part(self): | |
if not self.parse_header_line(): | |
return False | |
while self.parse_header_line(): | |
pass | |
if not self.parse_body_line(): | |
return False | |
while self.parse_body_line(): | |
pass | |
return True | |
def parse_header_line(self): | |
return \ | |
self.parse_empty() or \ | |
self.parse_comment() or \ | |
self.parse_part_detail() or \ | |
self.parse_description() or \ | |
self.parse_event_specification() or \ | |
self.parse_cost_line_def() or \ | |
self.parse_cost_summary() | |
_detail_keys = set(('cmd', 'pid', 'thread', 'part')) | |
def parse_part_detail(self): | |
return self.parse_keys(self._detail_keys) | |
def parse_description(self): | |
return self.parse_key('desc') is not None | |
def parse_event_specification(self): | |
event = self.parse_key('event') | |
if event is None: | |
return False | |
return True | |
def parse_cost_line_def(self): | |
pair = self.parse_keys(('events', 'positions')) | |
if pair is None: | |
return False | |
key, value = pair | |
items = value.split() | |
if key == 'events': | |
self.num_events = len(items) | |
self.cost_events = items | |
if key == 'positions': | |
self.num_positions = len(items) | |
self.cost_positions = items | |
self.last_positions = [0]*self.num_positions | |
return True | |
def parse_cost_summary(self): | |
pair = self.parse_keys(('summary', 'totals')) | |
if pair is None: | |
return False | |
return True | |
def parse_body_line(self): | |
return \ | |
self.parse_empty() or \ | |
self.parse_comment() or \ | |
self.parse_cost_line() or \ | |
self.parse_position_spec() or \ | |
self.parse_association_spec() | |
__subpos_re = r'(0x[0-9a-fA-F]+|\d+|\+\d+|-\d+|\*)' | |
_cost_re = re.compile(r'^' + | |
__subpos_re + r'( +' + __subpos_re + r')*' + | |
r'( +\d+)*' + | |
'$') | |
def parse_cost_line(self, calls=None): | |
line = self.lookahead().rstrip() | |
mo = self._cost_re.match(line) | |
if not mo: | |
return False | |
function = self.get_function() | |
if calls is None: | |
# Unlike other aspects, call object (cob) is relative not to the | |
# last call object, but to the caller's object (ob), so try to | |
# update it when processing a functions cost line | |
try: | |
self.positions['cob'] = self.positions['ob'] | |
except KeyError: | |
pass | |
values = line.split() | |
assert len(values) <= self.num_positions + self.num_events | |
positions = values[0 : self.num_positions] | |
events = values[self.num_positions : ] | |
events += ['0']*(self.num_events - len(events)) | |
for i in range(self.num_positions): | |
position = positions[i] | |
if position == '*': | |
position = self.last_positions[i] | |
elif position[0] in '-+': | |
position = self.last_positions[i] + int(position) | |
elif position.startswith('0x'): | |
position = int(position, 16) | |
else: | |
position = int(position) | |
self.last_positions[i] = position | |
events = [float(event) for event in events] | |
if calls is None: | |
function[SAMPLES] += events[0] | |
self.profile[SAMPLES] += events[0] | |
else: | |
callee = self.get_callee() | |
callee.called += calls | |
try: | |
call = function.calls[callee.id] | |
except KeyError: | |
call = Call(callee.id) | |
call[CALLS] = calls | |
call[SAMPLES2] = events[0] | |
function.add_call(call) | |
else: | |
call[CALLS] += calls | |
call[SAMPLES2] += events[0] | |
self.consume() | |
return True | |
def parse_association_spec(self): | |
line = self.lookahead() | |
if not line.startswith('calls='): | |
return False | |
_, values = line.split('=', 1) | |
values = values.strip().split() | |
calls = int(values[0]) | |
call_position = values[1:] | |
self.consume() | |
self.parse_cost_line(calls) | |
return True | |
_position_re = re.compile(r'^(?P<position>[cj]?(?:ob|fl|fi|fe|fn))=\s*(?:\((?P<id>\d+)\))?(?:\s*(?P<name>.+))?') | |
_position_table_map = { | |
'ob': 'ob', | |
'fl': 'fl', | |
'fi': 'fl', | |
'fe': 'fl', | |
'fn': 'fn', | |
'cob': 'ob', | |
'cfl': 'fl', | |
'cfi': 'fl', | |
'cfe': 'fl', | |
'cfn': 'fn', | |
'jfi': 'fl', | |
} | |
_position_map = { | |
'ob': 'ob', | |
'fl': 'fl', | |
'fi': 'fl', | |
'fe': 'fl', | |
'fn': 'fn', | |
'cob': 'cob', | |
'cfl': 'cfl', | |
'cfi': 'cfl', | |
'cfe': 'cfl', | |
'cfn': 'cfn', | |
'jfi': 'jfi', | |
} | |
def parse_position_spec(self): | |
line = self.lookahead() | |
if line.startswith('jump=') or line.startswith('jcnd='): | |
self.consume() | |
return True | |
mo = self._position_re.match(line) | |
if not mo: | |
return False | |
position, id, name = mo.groups() | |
if id: | |
table = self._position_table_map[position] | |
if name: | |
self.position_ids[(table, id)] = name | |
else: | |
name = self.position_ids.get((table, id), '') | |
self.positions[self._position_map[position]] = name | |
self.consume() | |
return True | |
def parse_empty(self): | |
if self.eof(): | |
return False | |
line = self.lookahead() | |
if line.strip(): | |
return False | |
self.consume() | |
return True | |
def parse_comment(self): | |
line = self.lookahead() | |
if not line.startswith('#'): | |
return False | |
self.consume() | |
return True | |
_key_re = re.compile(r'^(\w+):') | |
def parse_key(self, key): | |
pair = self.parse_keys((key,)) | |
if not pair: | |
return None | |
key, value = pair | |
return value | |
def parse_keys(self, keys): | |
line = self.lookahead() | |
mo = self._key_re.match(line) | |
if not mo: | |
return None | |
key, value = line.split(':', 1) | |
if key not in keys: | |
return None | |
value = value.strip() | |
self.consume() | |
return key, value | |
def make_function(self, module, filename, name): | |
# FIXME: module and filename are not being tracked reliably | |
#id = '|'.join((module, filename, name)) | |
id = name | |
try: | |
function = self.profile.functions[id] | |
except KeyError: | |
function = Function(id, name) | |
if module: | |
function.module = os.path.basename(module) | |
function[SAMPLES] = 0 | |
function.called = 0 | |
self.profile.add_function(function) | |
return function | |
def get_function(self): | |
module = self.positions.get('ob', '') | |
filename = self.positions.get('fl', '') | |
function = self.positions.get('fn', '') | |
return self.make_function(module, filename, function) | |
def get_callee(self): | |
module = self.positions.get('cob', '') | |
filename = self.positions.get('cfi', '') | |
function = self.positions.get('cfn', '') | |
return self.make_function(module, filename, function) | |
class PerfParser(LineParser): | |
"""Parser for linux perf callgraph output. | |
It expects output generated with | |
perf record -g | |
perf script | gprof2dot.py --format=perf | |
""" | |
def __init__(self, infile): | |
LineParser.__init__(self, infile) | |
self.profile = Profile() | |
def readline(self): | |
# Override LineParser.readline to ignore comment lines | |
while True: | |
LineParser.readline(self) | |
if self.eof() or not self.lookahead().startswith('#'): | |
break | |
def parse(self): | |
# read lookahead | |
self.readline() | |
profile = self.profile | |
profile[SAMPLES] = 0 | |
while not self.eof(): | |
self.parse_event() | |
# compute derived data | |
profile.validate() | |
profile.find_cycles() | |
profile.ratio(TIME_RATIO, SAMPLES) | |
profile.call_ratios(SAMPLES2) | |
if totalMethod == "callratios": | |
# Heuristic approach. TOTAL_SAMPLES is unused. | |
profile.integrate(TOTAL_TIME_RATIO, TIME_RATIO) | |
elif totalMethod == "callstacks": | |
# Use the actual call chains for functions. | |
profile[TOTAL_SAMPLES] = profile[SAMPLES] | |
profile.ratio(TOTAL_TIME_RATIO, TOTAL_SAMPLES) | |
# Then propagate that total time to the calls. | |
for function in compat_itervalues(profile.functions): | |
for call in compat_itervalues(function.calls): | |
if call.ratio is not None: | |
callee = profile.functions[call.callee_id] | |
call[TOTAL_TIME_RATIO] = call.ratio * callee[TOTAL_TIME_RATIO] | |
else: | |
assert False | |
return profile | |
def parse_event(self): | |
if self.eof(): | |
return | |
line = self.consume() | |
assert line | |
callchain = self.parse_callchain() | |
if not callchain: | |
return | |
callee = callchain[0] | |
callee[SAMPLES] += 1 | |
self.profile[SAMPLES] += 1 | |
for caller in callchain[1:]: | |
try: | |
call = caller.calls[callee.id] | |
except KeyError: | |
call = Call(callee.id) | |
call[SAMPLES2] = 1 | |
caller.add_call(call) | |
else: | |
call[SAMPLES2] += 1 | |
callee = caller | |
# Increment TOTAL_SAMPLES only once on each function. | |
stack = set(callchain) | |
for function in stack: | |
function[TOTAL_SAMPLES] += 1 | |
def parse_callchain(self): | |
callchain = [] | |
while self.lookahead(): | |
function = self.parse_call() | |
if function is None: | |
break | |
callchain.append(function) | |
if self.lookahead() == '': | |
self.consume() | |
return callchain | |
call_re = re.compile(r'^\s+(?P<address>[0-9a-fA-F]+)\s+(?P<symbol>.*)\s+\((?P<module>[^)]*)\)$') | |
def parse_call(self): | |
line = self.consume() | |
mo = self.call_re.match(line) | |
assert mo | |
if not mo: | |
return None | |
function_name = mo.group('symbol') | |
if not function_name: | |
function_name = mo.group('address') | |
module = mo.group('module') | |
function_id = function_name + ':' + module | |
try: | |
function = self.profile.functions[function_id] | |
except KeyError: | |
function = Function(function_id, function_name) | |
function.module = os.path.basename(module) | |
function[SAMPLES] = 0 | |
function[TOTAL_SAMPLES] = 0 | |
self.profile.add_function(function) | |
return function | |
class OprofileParser(LineParser): | |
"""Parser for oprofile callgraph output. | |
See also: | |
- http://oprofile.sourceforge.net/doc/opreport.html#opreport-callgraph | |
""" | |
_fields_re = { | |
'samples': r'(\d+)', | |
'%': r'(\S+)', | |
'linenr info': r'(?P<source>\(no location information\)|\S+:\d+)', | |
'image name': r'(?P<image>\S+(?:\s\(tgid:[^)]*\))?)', | |
'app name': r'(?P<application>\S+)', | |
'symbol name': r'(?P<symbol>\(no symbols\)|.+?)', | |
} | |
def __init__(self, infile): | |
LineParser.__init__(self, infile) | |
self.entries = {} | |
self.entry_re = None | |
def add_entry(self, callers, function, callees): | |
try: | |
entry = self.entries[function.id] | |
except KeyError: | |
self.entries[function.id] = (callers, function, callees) | |
else: | |
callers_total, function_total, callees_total = entry | |
self.update_subentries_dict(callers_total, callers) | |
function_total.samples += function.samples | |
self.update_subentries_dict(callees_total, callees) | |
def update_subentries_dict(self, totals, partials): | |
for partial in compat_itervalues(partials): | |
try: | |
total = totals[partial.id] | |
except KeyError: | |
totals[partial.id] = partial | |
else: | |
total.samples += partial.samples | |
def parse(self): | |
# read lookahead | |
self.readline() | |
self.parse_header() | |
while self.lookahead(): | |
self.parse_entry() | |
profile = Profile() | |
reverse_call_samples = {} | |
# populate the profile | |
profile[SAMPLES] = 0 | |
for _callers, _function, _callees in compat_itervalues(self.entries): | |
function = Function(_function.id, _function.name) | |
function[SAMPLES] = _function.samples | |
profile.add_function(function) | |
profile[SAMPLES] += _function.samples | |
if _function.application: | |
function.process = os.path.basename(_function.application) | |
if _function.image: | |
function.module = os.path.basename(_function.image) | |
total_callee_samples = 0 | |
for _callee in compat_itervalues(_callees): | |
total_callee_samples += _callee.samples | |
for _callee in compat_itervalues(_callees): | |
if not _callee.self: | |
call = Call(_callee.id) | |
call[SAMPLES2] = _callee.samples | |
function.add_call(call) | |
# compute derived data | |
profile.validate() | |
profile.find_cycles() | |
profile.ratio(TIME_RATIO, SAMPLES) | |
profile.call_ratios(SAMPLES2) | |
profile.integrate(TOTAL_TIME_RATIO, TIME_RATIO) | |
return profile | |
def parse_header(self): | |
while not self.match_header(): | |
self.consume() | |
line = self.lookahead() | |
fields = re.split(r'\s\s+', line) | |
entry_re = r'^\s*' + r'\s+'.join([self._fields_re[field] for field in fields]) + r'(?P<self>\s+\[self\])?$' | |
self.entry_re = re.compile(entry_re) | |
self.skip_separator() | |
def parse_entry(self): | |
callers = self.parse_subentries() | |
if self.match_primary(): | |
function = self.parse_subentry() | |
if function is not None: | |
callees = self.parse_subentries() | |
self.add_entry(callers, function, callees) | |
self.skip_separator() | |
def parse_subentries(self): | |
subentries = {} | |
while self.match_secondary(): | |
subentry = self.parse_subentry() | |
subentries[subentry.id] = subentry | |
return subentries | |
def parse_subentry(self): | |
entry = Struct() | |
line = self.consume() | |
mo = self.entry_re.match(line) | |
if not mo: | |
raise ParseError('failed to parse', line) | |
fields = mo.groupdict() | |
entry.samples = int(mo.group(1)) | |
if 'source' in fields and fields['source'] != '(no location information)': | |
source = fields['source'] | |
filename, lineno = source.split(':') | |
entry.filename = filename | |
entry.lineno = int(lineno) | |
else: | |
source = '' | |
entry.filename = None | |
entry.lineno = None | |
entry.image = fields.get('image', '') | |
entry.application = fields.get('application', '') | |
if 'symbol' in fields and fields['symbol'] != '(no symbols)': | |
entry.symbol = fields['symbol'] | |
else: | |
entry.symbol = '' | |
if entry.symbol.startswith('"') and entry.symbol.endswith('"'): | |
entry.symbol = entry.symbol[1:-1] | |
entry.id = ':'.join((entry.application, entry.image, source, entry.symbol)) | |
entry.self = fields.get('self', None) != None | |
if entry.self: | |
entry.id += ':self' | |
if entry.symbol: | |
entry.name = entry.symbol | |
else: | |
entry.name = entry.image | |
return entry | |
def skip_separator(self): | |
while not self.match_separator(): | |
self.consume() | |
self.consume() | |
def match_header(self): | |
line = self.lookahead() | |
return line.startswith('samples') | |
def match_separator(self): | |
line = self.lookahead() | |
return line == '-'*len(line) | |
def match_primary(self): | |
line = self.lookahead() | |
return not line[:1].isspace() | |
def match_secondary(self): | |
line = self.lookahead() | |
return line[:1].isspace() | |
class HProfParser(LineParser): | |
"""Parser for java hprof output | |
See also: | |
- http://java.sun.com/developer/technicalArticles/Programming/HPROF.html | |
""" | |
trace_re = re.compile(r'\t(.*)\((.*):(.*)\)') | |
trace_id_re = re.compile(r'^TRACE (\d+):$') | |
def __init__(self, infile): | |
LineParser.__init__(self, infile) | |
self.traces = {} | |
self.samples = {} | |
def parse(self): | |
# read lookahead | |
self.readline() | |
while not self.lookahead().startswith('------'): self.consume() | |
while not self.lookahead().startswith('TRACE '): self.consume() | |
self.parse_traces() | |
while not self.lookahead().startswith('CPU'): | |
self.consume() | |
self.parse_samples() | |
# populate the profile | |
profile = Profile() | |
profile[SAMPLES] = 0 | |
functions = {} | |
# build up callgraph | |
for id, trace in compat_iteritems(self.traces): | |
if not id in self.samples: continue | |
mtime = self.samples[id][0] | |
last = None | |
for func, file, line in trace: | |
if not func in functions: | |
function = Function(func, func) | |
function[SAMPLES] = 0 | |
profile.add_function(function) | |
functions[func] = function | |
function = functions[func] | |
# allocate time to the deepest method in the trace | |
if not last: | |
function[SAMPLES] += mtime | |
profile[SAMPLES] += mtime | |
else: | |
c = function.get_call(last) | |
c[SAMPLES2] += mtime | |
last = func | |
# compute derived data | |
profile.validate() | |
profile.find_cycles() | |
profile.ratio(TIME_RATIO, SAMPLES) | |
profile.call_ratios(SAMPLES2) | |
profile.integrate(TOTAL_TIME_RATIO, TIME_RATIO) | |
return profile | |
def parse_traces(self): | |
while self.lookahead().startswith('TRACE '): | |
self.parse_trace() | |
def parse_trace(self): | |
l = self.consume() | |
mo = self.trace_id_re.match(l) | |
tid = mo.group(1) | |
last = None | |
trace = [] | |
while self.lookahead().startswith('\t'): | |
l = self.consume() | |
match = self.trace_re.search(l) | |
if not match: | |
#sys.stderr.write('Invalid line: %s\n' % l) | |
break | |
else: | |
function_name, file, line = match.groups() | |
trace += [(function_name, file, line)] | |
self.traces[int(tid)] = trace | |
def parse_samples(self): | |
self.consume() | |
self.consume() | |
while not self.lookahead().startswith('CPU'): | |
rank, percent_self, percent_accum, count, traceid, method = self.lookahead().split() | |
self.samples[int(traceid)] = (int(count), method) | |
self.consume() | |
class SysprofParser(XmlParser): | |
def __init__(self, stream): | |
XmlParser.__init__(self, stream) | |
def parse(self): | |
objects = {} | |
nodes = {} | |
self.element_start('profile') | |
while self.token.type == XML_ELEMENT_START: | |
if self.token.name_or_data == 'objects': | |
assert not objects | |
objects = self.parse_items('objects') | |
elif self.token.name_or_data == 'nodes': | |
assert not nodes | |
nodes = self.parse_items('nodes') | |
else: | |
self.parse_value(self.token.name_or_data) | |
self.element_end('profile') | |
return self.build_profile(objects, nodes) | |
def parse_items(self, name): | |
assert name[-1] == 's' | |
items = {} | |
self.element_start(name) | |
while self.token.type == XML_ELEMENT_START: | |
id, values = self.parse_item(name[:-1]) | |
assert id not in items | |
items[id] = values | |
self.element_end(name) | |
return items | |
def parse_item(self, name): | |
attrs = self.element_start(name) | |
id = int(attrs['id']) | |
values = self.parse_values() | |
self.element_end(name) | |
return id, values | |
def parse_values(self): | |
values = {} | |
while self.token.type == XML_ELEMENT_START: | |
name = self.token.name_or_data | |
value = self.parse_value(name) | |
assert name not in values | |
values[name] = value | |
return values | |
def parse_value(self, tag): | |
self.element_start(tag) | |
value = self.character_data() | |
self.element_end(tag) | |
if value.isdigit(): | |
return int(value) | |
if value.startswith('"') and value.endswith('"'): | |
return value[1:-1] | |
return value | |
def build_profile(self, objects, nodes): | |
profile = Profile() | |
profile[SAMPLES] = 0 | |
for id, object in compat_iteritems(objects): | |
# Ignore fake objects (process names, modules, "Everything", "kernel", etc.) | |
if object['self'] == 0: | |
continue | |
function = Function(id, object['name']) | |
function[SAMPLES] = object['self'] | |
profile.add_function(function) | |
profile[SAMPLES] += function[SAMPLES] | |
for id, node in compat_iteritems(nodes): | |
# Ignore fake calls | |
if node['self'] == 0: | |
continue | |
# Find a non-ignored parent | |
parent_id = node['parent'] | |
while parent_id != 0: | |
parent = nodes[parent_id] | |
caller_id = parent['object'] | |
if objects[caller_id]['self'] != 0: | |
break | |
parent_id = parent['parent'] | |
if parent_id == 0: | |
continue | |
callee_id = node['object'] | |
assert objects[caller_id]['self'] | |
assert objects[callee_id]['self'] | |
function = profile.functions[caller_id] | |
samples = node['self'] | |
try: | |
call = function.calls[callee_id] | |
except KeyError: | |
call = Call(callee_id) | |
call[SAMPLES2] = samples | |
function.add_call(call) | |
else: | |
call[SAMPLES2] += samples | |
# Compute derived events | |
profile.validate() | |
profile.find_cycles() | |
profile.ratio(TIME_RATIO, SAMPLES) | |
profile.call_ratios(SAMPLES2) | |
profile.integrate(TOTAL_TIME_RATIO, TIME_RATIO) | |
return profile | |
class XPerfParser(Parser): | |
"""Parser for CSVs generted by XPerf, from Microsoft Windows Performance Tools. | |
""" | |
def __init__(self, stream): | |
Parser.__init__(self) | |
self.stream = stream | |
self.profile = Profile() | |
self.profile[SAMPLES] = 0 | |
self.column = {} | |
def parse(self): | |
import csv | |
reader = csv.reader( | |
self.stream, | |
delimiter = ',', | |
quotechar = None, | |
escapechar = None, | |
doublequote = False, | |
skipinitialspace = True, | |
lineterminator = '\r\n', | |
quoting = csv.QUOTE_NONE) | |
header = True | |
for row in reader: | |
if header: | |
self.parse_header(row) | |
header = False | |
else: | |
self.parse_row(row) | |
# compute derived data | |
self.profile.validate() | |
self.profile.find_cycles() | |
self.profile.ratio(TIME_RATIO, SAMPLES) | |
self.profile.call_ratios(SAMPLES2) | |
self.profile.integrate(TOTAL_TIME_RATIO, TIME_RATIO) | |
return self.profile | |
def parse_header(self, row): | |
for column in range(len(row)): | |
name = row[column] | |
assert name not in self.column | |
self.column[name] = column | |
def parse_row(self, row): | |
fields = {} | |
for name, column in compat_iteritems(self.column): | |
value = row[column] | |
for factory in int, float: | |
try: | |
value = factory(value) | |
except ValueError: | |
pass | |
else: | |
break | |
fields[name] = value | |
process = fields['Process Name'] | |
symbol = fields['Module'] + '!' + fields['Function'] | |
weight = fields['Weight'] | |
count = fields['Count'] | |
if process == 'Idle': | |
return | |
function = self.get_function(process, symbol) | |
function[SAMPLES] += weight * count | |
self.profile[SAMPLES] += weight * count | |
stack = fields['Stack'] | |
if stack != '?': | |
stack = stack.split('/') | |
assert stack[0] == '[Root]' | |
if stack[-1] != symbol: | |
# XXX: some cases the sampled function does not appear in the stack | |
stack.append(symbol) | |
caller = None | |
for symbol in stack[1:]: | |
callee = self.get_function(process, symbol) | |
if caller is not None: | |
try: | |
call = caller.calls[callee.id] | |
except KeyError: | |
call = Call(callee.id) | |
call[SAMPLES2] = count | |
caller.add_call(call) | |
else: | |
call[SAMPLES2] += count | |
caller = callee | |
def get_function(self, process, symbol): | |
function_id = process + '!' + symbol | |
try: | |
function = self.profile.functions[function_id] | |
except KeyError: | |
module, name = symbol.split('!', 1) | |
function = Function(function_id, name) | |
function.process = process | |
function.module = module | |
function[SAMPLES] = 0 | |
self.profile.add_function(function) | |
return function | |
class SleepyParser(Parser): | |
"""Parser for GNU gprof output. | |
See also: | |
- http://www.codersnotes.com/sleepy/ | |
- http://sleepygraph.sourceforge.net/ | |
""" | |
stdinInput = False | |
def __init__(self, filename): | |
Parser.__init__(self) | |
from zipfile import ZipFile | |
self.database = ZipFile(filename) | |
self.symbols = {} | |
self.calls = {} | |
self.profile = Profile() | |
_symbol_re = re.compile( | |
r'^(?P<id>\w+)' + | |
r'\s+"(?P<module>[^"]*)"' + | |
r'\s+"(?P<procname>[^"]*)"' + | |
r'\s+"(?P<sourcefile>[^"]*)"' + | |
r'\s+(?P<sourceline>\d+)$' | |
) | |
def openEntry(self, name): | |
# Some versions of verysleepy use lowercase filenames | |
for database_name in self.database.namelist(): | |
if name.lower() == database_name.lower(): | |
name = database_name | |
break | |
return self.database.open(name, 'rU') | |
def parse_symbols(self): | |
for line in self.openEntry('Symbols.txt'): | |
line = line.decode('UTF-8') | |
mo = self._symbol_re.match(line) | |
if mo: | |
symbol_id, module, procname, sourcefile, sourceline = mo.groups() | |
function_id = ':'.join([module, procname]) | |
try: | |
function = self.profile.functions[function_id] | |
except KeyError: | |
function = Function(function_id, procname) | |
function.module = module | |
function[SAMPLES] = 0 | |
self.profile.add_function(function) | |
self.symbols[symbol_id] = function | |
def parse_callstacks(self): | |
for line in self.openEntry('Callstacks.txt'): | |
line = line.decode('UTF-8') | |
fields = line.split() | |
samples = float(fields[0]) | |
callstack = fields[1:] | |
callstack = [self.symbols[symbol_id] for symbol_id in callstack] | |
callee = callstack[0] | |
callee[SAMPLES] += samples | |
self.profile[SAMPLES] += samples | |
for caller in callstack[1:]: | |
try: | |
call = caller.calls[callee.id] | |
except KeyError: | |
call = Call(callee.id) | |
call[SAMPLES2] = samples | |
caller.add_call(call) | |
else: | |
call[SAMPLES2] += samples | |
callee = caller | |
def parse(self): | |
profile = self.profile | |
profile[SAMPLES] = 0 | |
self.parse_symbols() | |
self.parse_callstacks() | |
# Compute derived events | |
profile.validate() | |
profile.find_cycles() | |
profile.ratio(TIME_RATIO, SAMPLES) | |
profile.call_ratios(SAMPLES2) | |
profile.integrate(TOTAL_TIME_RATIO, TIME_RATIO) | |
return profile | |
class PstatsParser: | |
"""Parser python profiling statistics saved with te pstats module.""" | |
stdinInput = False | |
multipleInput = True | |
def __init__(self, *filename): | |
import pstats | |
try: | |
self.stats = pstats.Stats(*filename) | |
except ValueError: | |
if PYTHON_3: | |
sys.stderr.write('error: failed to load %s\n' % ', '.join(filename)) | |
sys.exit(1) | |
import hotshot.stats | |
self.stats = hotshot.stats.load(filename[0]) | |
self.profile = Profile() | |
self.function_ids = {} | |
def get_function_name(self, key): | |
filename, line, name = key | |
module = os.path.splitext(filename)[0] | |
module = os.path.basename(module) | |
return "%s:%d:%s" % (module, line, name) | |
def get_function(self, key): | |
try: | |
id = self.function_ids[key] | |
except KeyError: | |
id = len(self.function_ids) | |
name = self.get_function_name(key) | |
function = Function(id, name) | |
self.profile.functions[id] = function | |
self.function_ids[key] = id | |
else: | |
function = self.profile.functions[id] | |
return function | |
def parse(self): | |
self.profile[TIME] = 0.0 | |
self.profile[TOTAL_TIME] = self.stats.total_tt | |
for fn, (cc, nc, tt, ct, callers) in compat_iteritems(self.stats.stats): | |
callee = self.get_function(fn) | |
callee.called = nc | |
callee[TOTAL_TIME] = ct | |
callee[TIME] = tt | |
self.profile[TIME] += tt | |
self.profile[TOTAL_TIME] = max(self.profile[TOTAL_TIME], ct) | |
for fn, value in compat_iteritems(callers): | |
caller = self.get_function(fn) | |
call = Call(callee.id) | |
if isinstance(value, tuple): | |
for i in xrange(0, len(value), 4): | |
nc, cc, tt, ct = value[i:i+4] | |
if CALLS in call: | |
call[CALLS] += cc | |
else: | |
call[CALLS] = cc | |
if TOTAL_TIME in call: | |
call[TOTAL_TIME] += ct | |
else: | |
call[TOTAL_TIME] = ct | |
else: | |
call[CALLS] = value | |
call[TOTAL_TIME] = ratio(value, nc)*ct | |
caller.add_call(call) | |
if False: | |
self.stats.print_stats() | |
self.stats.print_callees() | |
# Compute derived events | |
self.profile.validate() | |
self.profile.ratio(TIME_RATIO, TIME) | |
self.profile.ratio(TOTAL_TIME_RATIO, TOTAL_TIME) | |
return self.profile | |
formats = { | |
"axe": AXEParser, | |
"callgrind": CallgrindParser, | |
"hprof": HProfParser, | |
"json": JsonParser, | |
"oprofile": OprofileParser, | |
"perf": PerfParser, | |
"prof": GprofParser, | |
"pstats": PstatsParser, | |
"sleepy": SleepyParser, | |
"sysprof": SysprofParser, | |
"xperf": XPerfParser, | |
} | |
######################################################################## | |
# Output | |
class Theme: | |
def __init__(self, | |
bgcolor = (0.0, 0.0, 1.0), | |
mincolor = (0.0, 0.0, 0.0), | |
maxcolor = (0.0, 0.0, 1.0), | |
fontname = "Arial", | |
fontcolor = "white", | |
nodestyle = "filled", | |
minfontsize = 10.0, | |
maxfontsize = 10.0, | |
minpenwidth = 0.5, | |
maxpenwidth = 4.0, | |
gamma = 2.2, | |
skew = 1.0): | |
self.bgcolor = bgcolor | |
self.mincolor = mincolor | |
self.maxcolor = maxcolor | |
self.fontname = fontname | |
self.fontcolor = fontcolor | |
self.nodestyle = nodestyle | |
self.minfontsize = minfontsize | |
self.maxfontsize = maxfontsize | |
self.minpenwidth = minpenwidth | |
self.maxpenwidth = maxpenwidth | |
self.gamma = gamma | |
self.skew = skew | |
def graph_bgcolor(self): | |
return self.hsl_to_rgb(*self.bgcolor) | |
def graph_fontname(self): | |
return self.fontname | |
def graph_fontcolor(self): | |
return self.fontcolor | |
def graph_fontsize(self): | |
return self.minfontsize | |
def node_bgcolor(self, weight): | |
return self.color(weight) | |
def node_fgcolor(self, weight): | |
if self.nodestyle == "filled": | |
return self.graph_bgcolor() | |
else: | |
return self.color(weight) | |
def node_fontsize(self, weight): | |
return self.fontsize(weight) | |
def node_style(self): | |
return self.nodestyle | |
def edge_color(self, weight): | |
return self.color(weight) | |
def edge_fontsize(self, weight): | |
return self.fontsize(weight) | |
def edge_penwidth(self, weight): | |
weight = min(weight, 1.0) | |
return max(weight*self.maxpenwidth, self.minpenwidth) | |
def edge_arrowsize(self, weight): | |
return 0.5 * math.sqrt(self.edge_penwidth(weight)) | |
def fontsize(self, weight): | |
weight = min(weight, 1.0) | |
return max(weight**2 * self.maxfontsize, self.minfontsize) | |
def color(self, weight): | |
weight = min(max(weight, 0.0), 1.0) | |
hmin, smin, lmin = self.mincolor | |
hmax, smax, lmax = self.maxcolor | |
if self.skew < 0: | |
raise ValueError("Skew must be greater than 0") | |
elif self.skew == 1.0: | |
h = hmin + weight*(hmax - hmin) | |
s = smin + weight*(smax - smin) | |
l = lmin + weight*(lmax - lmin) | |
else: | |
base = self.skew | |
h = hmin + ((hmax-hmin)*(-1.0 + (base ** weight)) / (base - 1.0)) | |
s = smin + ((smax-smin)*(-1.0 + (base ** weight)) / (base - 1.0)) | |
l = lmin + ((lmax-lmin)*(-1.0 + (base ** weight)) / (base - 1.0)) | |
return self.hsl_to_rgb(h, s, l) | |
def hsl_to_rgb(self, h, s, l): | |
"""Convert a color from HSL color-model to RGB. | |
See also: | |
- http://www.w3.org/TR/css3-color/#hsl-color | |
""" | |
h = h % 1.0 | |
s = min(max(s, 0.0), 1.0) | |
l = min(max(l, 0.0), 1.0) | |
if l <= 0.5: | |
m2 = l*(s + 1.0) | |
else: | |
m2 = l + s - l*s | |
m1 = l*2.0 - m2 | |
r = self._hue_to_rgb(m1, m2, h + 1.0/3.0) | |
g = self._hue_to_rgb(m1, m2, h) | |
b = self._hue_to_rgb(m1, m2, h - 1.0/3.0) | |
# Apply gamma correction | |
r **= self.gamma | |
g **= self.gamma | |
b **= self.gamma | |
return (r, g, b) | |
def _hue_to_rgb(self, m1, m2, h): | |
if h < 0.0: | |
h += 1.0 | |
elif h > 1.0: | |
h -= 1.0 | |
if h*6 < 1.0: | |
return m1 + (m2 - m1)*h*6.0 | |
elif h*2 < 1.0: | |
return m2 | |
elif h*3 < 2.0: | |
return m1 + (m2 - m1)*(2.0/3.0 - h)*6.0 | |
else: | |
return m1 | |
TEMPERATURE_COLORMAP = Theme( | |
mincolor = (2.0/3.0, 0.80, 0.25), # dark blue | |
maxcolor = (0.0, 1.0, 0.5), # satured red | |
gamma = 1.0 | |
) | |
PINK_COLORMAP = Theme( | |
mincolor = (0.0, 1.0, 0.90), # pink | |
maxcolor = (0.0, 1.0, 0.5), # satured red | |
) | |
GRAY_COLORMAP = Theme( | |
mincolor = (0.0, 0.0, 0.85), # light gray | |
maxcolor = (0.0, 0.0, 0.0), # black | |
) | |
BW_COLORMAP = Theme( | |
minfontsize = 8.0, | |
maxfontsize = 24.0, | |
mincolor = (0.0, 0.0, 0.0), # black | |
maxcolor = (0.0, 0.0, 0.0), # black | |
minpenwidth = 0.1, | |
maxpenwidth = 8.0, | |
) | |
PRINT_COLORMAP = Theme( | |
minfontsize = 18.0, | |
maxfontsize = 30.0, | |
fontcolor = "black", | |
nodestyle = "solid", | |
mincolor = (0.0, 0.0, 0.0), # black | |
maxcolor = (0.0, 0.0, 0.0), # black | |
minpenwidth = 0.1, | |
maxpenwidth = 8.0, | |
) | |
themes = { | |
"color": TEMPERATURE_COLORMAP, | |
"pink": PINK_COLORMAP, | |
"gray": GRAY_COLORMAP, | |
"bw": BW_COLORMAP, | |
"print": PRINT_COLORMAP, | |
} | |
def sorted_iteritems(d): | |
# Used mostly for result reproducibility (while testing.) | |
keys = compat_keys(d) | |
keys.sort() | |
for key in keys: | |
value = d[key] | |
yield key, value | |
class DotWriter: | |
"""Writer for the DOT language. | |
See also: | |
- "The DOT Language" specification | |
http://www.graphviz.org/doc/info/lang.html | |
""" | |
strip = False | |
wrap = False | |
def __init__(self, fp): | |
self.fp = fp | |
def wrap_function_name(self, name): | |
"""Split the function name on multiple lines.""" | |
if len(name) > 32: | |
ratio = 2.0/3.0 | |
height = max(int(len(name)/(1.0 - ratio) + 0.5), 1) | |
width = max(len(name)/height, 32) | |
# TODO: break lines in symbols | |
name = textwrap.fill(name, width, break_long_words=False) | |
# Take away spaces | |
name = name.replace(", ", ",") | |
name = name.replace("> >", ">>") | |
name = name.replace("> >", ">>") # catch consecutive | |
return name | |
show_function_events = [TOTAL_TIME_RATIO, TIME_RATIO] | |
show_edge_events = [TOTAL_TIME_RATIO, CALLS] | |
def graph(self, profile, theme): | |
self.begin_graph() | |
fontname = theme.graph_fontname() | |
fontcolor = theme.graph_fontcolor() | |
nodestyle = theme.node_style() | |
self.attr('graph', fontname=fontname, ranksep=0.25, nodesep=0.125) | |
self.attr('node', fontname=fontname, shape="box", style=nodestyle, fontcolor=fontcolor, width=0, height=0) | |
self.attr('edge', fontname=fontname) | |
for _, function in sorted_iteritems(profile.functions): | |
labels = [] | |
if function.process is not None: | |
labels.append(function.process) | |
if function.module is not None: | |
labels.append(function.module) | |
if self.strip: | |
function_name = function.stripped_name() | |
else: | |
function_name = function.name | |
if self.wrap: | |
function_name = self.wrap_function_name(function_name) | |
labels.append(function_name) | |
for event in self.show_function_events: | |
if event in function.events: | |
label = event.format(function[event]) | |
labels.append(label) | |
if function.called is not None: | |
labels.append("%u%s" % (function.called, MULTIPLICATION_SIGN)) | |
if function.weight is not None: | |
weight = function.weight | |
else: | |
weight = 0.0 | |
label = '\n'.join(labels) | |
self.node(function.id, | |
label = label, | |
color = self.color(theme.node_bgcolor(weight)), | |
fontcolor = self.color(theme.node_fgcolor(weight)), | |
fontsize = "%.2f" % theme.node_fontsize(weight), | |
) | |
for _, call in sorted_iteritems(function.calls): | |
callee = profile.functions[call.callee_id] | |
labels = [] | |
for event in self.show_edge_events: | |
if event in call.events: | |
label = event.format(call[event]) | |
labels.append(label) | |
if call.weight is not None: | |
weight = call.weight | |
elif callee.weight is not None: | |
weight = callee.weight | |
else: | |
weight = 0.0 | |
label = '\n'.join(labels) | |
self.edge(function.id, call.callee_id, | |
label = label, | |
color = self.color(theme.edge_color(weight)), | |
fontcolor = self.color(theme.edge_color(weight)), | |
fontsize = "%.2f" % theme.edge_fontsize(weight), | |
penwidth = "%.2f" % theme.edge_penwidth(weight), | |
labeldistance = "%.2f" % theme.edge_penwidth(weight), | |
arrowsize = "%.2f" % theme.edge_arrowsize(weight), | |
) | |
self.end_graph() | |
def begin_graph(self): | |
self.write('digraph {\n') | |
def end_graph(self): | |
self.write('}\n') | |
def attr(self, what, **attrs): | |
self.write("\t") | |
self.write(what) | |
self.attr_list(attrs) | |
self.write(";\n") | |
def node(self, node, **attrs): | |
self.write("\t") | |
self.id(node) | |
self.attr_list(attrs) | |
self.write(";\n") | |
def edge(self, src, dst, **attrs): | |
self.write("\t") | |
self.id(src) | |
self.write(" -> ") | |
self.id(dst) | |
self.attr_list(attrs) | |
self.write(";\n") | |
def attr_list(self, attrs): | |
if not attrs: | |
return | |
self.write(' [') | |
first = True | |
for name, value in sorted_iteritems(attrs): | |
if first: | |
first = False | |
else: | |
self.write(", ") | |
self.id(name) | |
self.write('=') | |
self.id(value) | |
self.write(']') | |
def id(self, id): | |
if isinstance(id, (int, float)): | |
s = str(id) | |
elif isinstance(id, basestring): | |
if id.isalnum() and not id.startswith('0x'): | |
s = id | |
else: | |
s = self.escape(id) | |
else: | |
raise TypeError | |
self.write(s) | |
def color(self, rgb): | |
r, g, b = rgb | |
def float2int(f): | |
if f <= 0.0: | |
return 0 | |
if f >= 1.0: | |
return 255 | |
return int(255.0*f + 0.5) | |
return "#" + "".join(["%02x" % float2int(c) for c in (r, g, b)]) | |
def escape(self, s): | |
if not PYTHON_3: | |
s = s.encode('utf-8') | |
s = s.replace('\\', r'\\') | |
s = s.replace('\n', r'\n') | |
s = s.replace('\t', r'\t') | |
s = s.replace('"', r'\"') | |
return '"' + s + '"' | |
def write(self, s): | |
self.fp.write(s) | |
######################################################################## | |
# Main program | |
def naturalJoin(values): | |
if len(values) >= 2: | |
return ', '.join(values[:-1]) + ' or ' + values[-1] | |
else: | |
return ''.join(values) | |
def main(): | |
"""Main program.""" | |
global totalMethod | |
formatNames = list(formats.keys()) | |
formatNames.sort() | |
optparser = optparse.OptionParser( | |
usage="\n\t%prog [options] [file] ...") | |
optparser.add_option( | |
'-o', '--output', metavar='FILE', | |
type="string", dest="output", | |
help="output filename [stdout]") | |
optparser.add_option( | |
'-n', '--node-thres', metavar='PERCENTAGE', | |
type="float", dest="node_thres", default=0.5, | |
help="eliminate nodes below this threshold [default: %default]") | |
optparser.add_option( | |
'-e', '--edge-thres', metavar='PERCENTAGE', | |
type="float", dest="edge_thres", default=0.1, | |
help="eliminate edges below this threshold [default: %default]") | |
optparser.add_option( | |
'-f', '--format', | |
type="choice", choices=formatNames, | |
dest="format", default="prof", | |
help="profile format: %s [default: %%default]" % naturalJoin(formatNames)) | |
optparser.add_option( | |
'--total', | |
type="choice", choices=('callratios', 'callstacks'), | |
dest="totalMethod", default=totalMethod, | |
help="preferred method of calculating total time: callratios or callstacks (currently affects only perf format) [default: %default]") | |
optparser.add_option( | |
'-c', '--colormap', | |
type="choice", choices=('color', 'pink', 'gray', 'bw', 'print'), | |
dest="theme", default="color", | |
help="color map: color, pink, gray, bw, or print [default: %default]") | |
optparser.add_option( | |
'-s', '--strip', | |
action="store_true", | |
dest="strip", default=False, | |
help="strip function parameters, template parameters, and const modifiers from demangled C++ function names") | |
optparser.add_option( | |
'-w', '--wrap', | |
action="store_true", | |
dest="wrap", default=False, | |
help="wrap function names") | |
optparser.add_option( | |
'--show-samples', | |
action="store_true", | |
dest="show_samples", default=False, | |
help="show function samples") | |
# add option to create subtree or show paths | |
optparser.add_option( | |
'-z', '--root', | |
type="string", | |
dest="root", default="", | |
help="prune call graph to show only descendants of specified root function") | |
optparser.add_option( | |
'-l', '--leaf', | |
type="string", | |
dest="leaf", default="", | |
help="prune call graph to show only ancestors of specified leaf function") | |
# add a new option to control skew of the colorization curve | |
optparser.add_option( | |
'--skew', | |
type="float", dest="theme_skew", default=1.0, | |
help="skew the colorization curve. Values < 1.0 give more variety to lower percentages. Values > 1.0 give less variety to lower percentages") | |
(options, args) = optparser.parse_args(sys.argv[1:]) | |
if len(args) > 1 and options.format != 'pstats': | |
optparser.error('incorrect number of arguments') | |
try: | |
theme = themes[options.theme] | |
except KeyError: | |
optparser.error('invalid colormap \'%s\'' % options.theme) | |
# set skew on the theme now that it has been picked. | |
if options.theme_skew: | |
theme.skew = options.theme_skew | |
totalMethod = options.totalMethod | |
try: | |
Format = formats[options.format] | |
except KeyError: | |
optparser.error('invalid format \'%s\'' % options.format) | |
if Format.stdinInput: | |
if not args: | |
fp = sys.stdin | |
else: | |
fp = open(args[0], 'rt') | |
parser = Format(fp) | |
elif Format.multipleInput: | |
if not args: | |
optparser.error('at least a file must be specified for %s input' % options.format) | |
parser = Format(*args) | |
else: | |
if len(args) != 1: | |
optparser.error('exactly one file must be specified for %s input' % options.format) | |
parser = Format(args[0]) | |
profile = parser.parse() | |
if options.output is None: | |
output = sys.stdout | |
else: | |
if PYTHON_3: | |
output = open(options.output, 'wt', encoding='UTF-8') | |
else: | |
output = open(options.output, 'wt') | |
dot = DotWriter(output) | |
dot.strip = options.strip | |
dot.wrap = options.wrap | |
if options.show_samples: | |
dot.show_function_events.append(SAMPLES) | |
profile = profile | |
profile.prune(options.node_thres/100.0, options.edge_thres/100.0) | |
if options.root: | |
rootId = profile.getFunctionId(options.root) | |
if not rootId: | |
sys.stderr.write('root node ' + options.root + ' not found (might already be pruned : try -e0 -n0 flags)\n') | |
sys.exit(1) | |
profile.prune_root(rootId) | |
if options.leaf: | |
leafId = profile.getFunctionId(options.leaf) | |
if not leafId: | |
sys.stderr.write('leaf node ' + options.leaf + ' not found (maybe already pruned : try -e0 -n0 flags)\n') | |
sys.exit(1) | |
profile.prune_leaf(leafId) | |
dot.graph(profile, theme) | |
if __name__ == '__main__': | |
main() |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment