[opt-viewer] Put critical items in parallel
Summary:
Put opt-viewer critical items in parallel
Patch by Brian Cain!
Requires features from Python 2.7
**Performance**
Below are performance results across various configurations. These were taken on an i5-5200U (dual core + HT). They were taken with a small subset of the YAML output of building Python 3.6.0b3 with LTO+PGO. 60 YAML files.
"multiprocessing" is the current submission contents. "baseline" is as of 544f14c6b2a07a94168df31833dba9dc35fd8289 (I think this is aka r287505).
"ImportError" vs "class<...CLoader>" below are just confirming the expected configuration (with/without CLoader).
The below was measured on AMD A8-5500B (4 cores) with 224 input YAML files, showing a ~1.75x speed increase over the baseline with libYAML. I suspect it would scale well on high-end servers.
```
**************************************** MULTIPROCESSING ****************************************
PyYAML:
Traceback (most recent call last):
File "<string>", line 1, in <module>
ImportError: cannot import name CLoader
Python 2.7.10
489.42user 5.53system 2:38.03elapsed 313%CPU (0avgtext+0avgdata 400308maxresident)k
0inputs+31392outputs (0major+473540minor)pagefaults 0swaps
PyYAML+libYAML:
<class 'yaml.cyaml.CLoader'>
Python 2.7.10
78.69user 5.45system 0:32.63elapsed 257%CPU (0avgtext+0avgdata 398560maxresident)k
0inputs+31392outputs (0major+542022minor)pagefaults 0swaps
PyPy/PyYAML:
Traceback (most recent call last):
File "<builtin>/app_main.py", line 75, in run_toplevel
File "<builtin>/app_main.py", line 601, in run_it
File "<string>", line 1, in <module>
ImportError: cannot import name 'CLoader'
Python 2.7.9 (2.6.0+dfsg-3, Jul 04 2015, 05:43:17)
[PyPy 2.6.0 with GCC 4.9.3]
154.27user 8.12system 0:53.83elapsed 301%CPU (0avgtext+0avgdata 627960maxresident)k
808inputs+30376outputs (0major+727994minor)pagefaults 0swaps
**************************************** BASELINE ****************************************
PyYAML:
Traceback (most recent call last):
File "<string>", line 1, in <module>
ImportError: cannot import name CLoader
Python 2.7.10
358.08user 4.05system 6:08.37elapsed 98%CPU (0avgtext+0avgdata 315004maxresident)k
0inputs+31392outputs (0major+85252minor)pagefaults 0swaps
PyYAML+libYAML:
<class 'yaml.cyaml.CLoader'>
Python 2.7.10
50.32user 3.30system 0:56.59elapsed 94%CPU (0avgtext+0avgdata 307296maxresident)k
0inputs+31392outputs (0major+79335minor)pagefaults 0swaps
PyPy/PyYAML:
Traceback (most recent call last):
File "<builtin>/app_main.py", line 75, in run_toplevel
File "<builtin>/app_main.py", line 601, in run_it
File "<string>", line 1, in <module>
ImportError: cannot import name 'CLoader'
Python 2.7.9 (2.6.0+dfsg-3, Jul 04 2015, 05:43:17)
[PyPy 2.6.0 with GCC 4.9.3]
72.94user 5.18system 1:23.41elapsed 93%CPU (0avgtext+0avgdata 455312maxresident)k
0inputs+30392outputs (0major+110280minor)pagefaults 0swaps
```
Reviewers: fhahn, anemet
Reviewed By: anemet
Subscribers: llvm-commits, mehdi_amini
Differential Revision: https://reviews.llvm.org/D26967
llvm-svn: 293261
This commit is contained in:
parent
0b79aa3373
commit
55bfb497d2
|
|
@ -15,6 +15,13 @@ try:
|
||||||
from yaml import CLoader as Loader
|
from yaml import CLoader as Loader
|
||||||
except ImportError:
|
except ImportError:
|
||||||
from yaml import Loader
|
from yaml import Loader
|
||||||
|
|
||||||
|
import functools
|
||||||
|
from collections import defaultdict
|
||||||
|
import itertools
|
||||||
|
from multiprocessing import Pool
|
||||||
|
from multiprocessing import Lock, cpu_count
|
||||||
|
import errno
|
||||||
import argparse
|
import argparse
|
||||||
import os.path
|
import os.path
|
||||||
import re
|
import re
|
||||||
|
|
@ -24,16 +31,12 @@ from pygments import highlight
|
||||||
from pygments.lexers.c_cpp import CppLexer
|
from pygments.lexers.c_cpp import CppLexer
|
||||||
from pygments.formatters import HtmlFormatter
|
from pygments.formatters import HtmlFormatter
|
||||||
|
|
||||||
parser = argparse.ArgumentParser(description=desc)
|
|
||||||
parser.add_argument('yaml_files', nargs='+')
|
|
||||||
parser.add_argument('output_dir')
|
|
||||||
parser.add_argument('-source-dir', '-s', default='', help='set source directory')
|
|
||||||
args = parser.parse_args()
|
|
||||||
|
|
||||||
p = subprocess.Popen(['c++filt', '-n'], stdin=subprocess.PIPE, stdout=subprocess.PIPE)
|
p = subprocess.Popen(['c++filt', '-n'], stdin=subprocess.PIPE, stdout=subprocess.PIPE)
|
||||||
|
p_lock = Lock()
|
||||||
|
|
||||||
|
|
||||||
def demangle(name):
|
def demangle(name):
|
||||||
|
with p_lock:
|
||||||
p.stdin.write(name + '\n')
|
p.stdin.write(name + '\n')
|
||||||
return p.stdout.readline().rstrip()
|
return p.stdout.readline().rstrip()
|
||||||
|
|
||||||
|
|
@ -156,16 +159,16 @@ class Missed(Remark):
|
||||||
|
|
||||||
|
|
||||||
class SourceFileRenderer:
|
class SourceFileRenderer:
|
||||||
def __init__(self, filename):
|
def __init__(self, source_dir, output_dir, filename):
|
||||||
existing_filename = None
|
existing_filename = None
|
||||||
if os.path.exists(filename):
|
if os.path.exists(filename):
|
||||||
existing_filename = filename
|
existing_filename = filename
|
||||||
else:
|
else:
|
||||||
fn = os.path.join(args.source_dir, filename)
|
fn = os.path.join(source_dir, filename)
|
||||||
if os.path.exists(fn):
|
if os.path.exists(fn):
|
||||||
existing_filename = fn
|
existing_filename = fn
|
||||||
|
|
||||||
self.stream = open(os.path.join(args.output_dir, SourceFileRenderer.html_file_name(filename)), 'w')
|
self.stream = open(os.path.join(output_dir, SourceFileRenderer.html_file_name(filename)), 'w')
|
||||||
if existing_filename:
|
if existing_filename:
|
||||||
self.source_stream = open(existing_filename)
|
self.source_stream = open(existing_filename)
|
||||||
else:
|
else:
|
||||||
|
|
@ -243,8 +246,8 @@ class SourceFileRenderer:
|
||||||
|
|
||||||
|
|
||||||
class IndexRenderer:
|
class IndexRenderer:
|
||||||
def __init__(self):
|
def __init__(self, output_dir):
|
||||||
self.stream = open(os.path.join(args.output_dir, 'index.html'), 'w')
|
self.stream = open(os.path.join(output_dir, 'index.html'), 'w')
|
||||||
|
|
||||||
def render_entry(self, r):
|
def render_entry(self, r):
|
||||||
print('''
|
print('''
|
||||||
|
|
@ -278,41 +281,123 @@ class IndexRenderer:
|
||||||
</html>''', file=self.stream)
|
</html>''', file=self.stream)
|
||||||
|
|
||||||
|
|
||||||
all_remarks = dict()
|
def get_remarks(input_file):
|
||||||
file_remarks = dict()
|
max_hotness = 0
|
||||||
|
all_remarks = dict()
|
||||||
|
file_remarks = defaultdict(functools.partial(defaultdict, list))
|
||||||
|
|
||||||
for input_file in args.yaml_files:
|
with open(input_file) as f:
|
||||||
f = open(input_file)
|
|
||||||
docs = yaml.load_all(f, Loader=Loader)
|
docs = yaml.load_all(f, Loader=Loader)
|
||||||
|
|
||||||
for remark in docs:
|
for remark in docs:
|
||||||
# Avoid remarks withoug debug location or if they are duplicated
|
# Avoid remarks withoug debug location or if they are duplicated
|
||||||
if not hasattr(remark, 'DebugLoc') or remark.key in all_remarks:
|
if not hasattr(remark, 'DebugLoc') or remark.key in all_remarks:
|
||||||
continue
|
continue
|
||||||
all_remarks[remark.key] = remark
|
all_remarks[remark.key] = remark
|
||||||
|
|
||||||
file_remarks.setdefault(remark.File, dict()).setdefault(remark.Line, []).append(remark)
|
file_remarks[remark.File][remark.Line].append(remark)
|
||||||
|
|
||||||
Remark.max_hotness = max(Remark.max_hotness, remark.Hotness)
|
max_hotness = max(max_hotness, remark.Hotness)
|
||||||
|
|
||||||
# Set up a map between function names and their source location for function where inlining happened
|
return max_hotness, all_remarks, file_remarks
|
||||||
for remark in all_remarks.itervalues():
|
|
||||||
if type(remark) == Passed and remark.Pass == "inline" and remark.Name == "Inlined":
|
|
||||||
|
def _render_file(source_dir, output_dir, entry):
|
||||||
|
filename, remarks = entry
|
||||||
|
SourceFileRenderer(source_dir, output_dir, filename).render(remarks)
|
||||||
|
|
||||||
|
|
||||||
|
def gather_results(pool, filenames):
|
||||||
|
all_remarks = dict()
|
||||||
|
remarks = pool.map(get_remarks, filenames)
|
||||||
|
|
||||||
|
def merge_dicts(dicts):
|
||||||
|
''' Takes an iterable of dicts and merges them into
|
||||||
|
a single dict. Nested dicts are merged as well.
|
||||||
|
>>> merge_dicts([ {'a': [3], }, {'a': [4], }, {'b': [6] }])
|
||||||
|
{'a': [3,4,], 'b': [6]}
|
||||||
|
>>> merge_dicts([ {'a': {'q': [6,3], 'f': [30],}, }, {'a': {'f': [4,10]}, }, {'b': [6] }])
|
||||||
|
{'a': [{'q': [6,3]}, {'f': [4,10,30]}], 'b': [6]}
|
||||||
|
|
||||||
|
'''
|
||||||
|
merged = defaultdict(functools.partial(defaultdict, list))
|
||||||
|
|
||||||
|
for k, v in itertools.chain(*[d.iteritems() for d in dicts]):
|
||||||
|
for k_, v_ in v.items():
|
||||||
|
merged[k][k_] += v_
|
||||||
|
|
||||||
|
return merged
|
||||||
|
|
||||||
|
file_remark_dicts = [entry[2] for entry in remarks]
|
||||||
|
# merge the list of remarks at each line of each file
|
||||||
|
file_remarks = merge_dicts(file_remark_dicts)
|
||||||
|
|
||||||
|
# merge individual 'all_remark' results:
|
||||||
|
for _, all_rem, _ in remarks:
|
||||||
|
all_remarks.update(all_rem)
|
||||||
|
|
||||||
|
Remark.max_hotness = max(entry[0] for entry in remarks)
|
||||||
|
|
||||||
|
return all_remarks, file_remarks
|
||||||
|
|
||||||
|
|
||||||
|
def map_remarks(all_remarks):
|
||||||
|
# Set up a map between function names and their source location for
|
||||||
|
# function where inlining happened
|
||||||
|
for remark in all_remarks.itervalues():
|
||||||
|
if isinstance(remark, Passed) and remark.Pass == "inline" and remark.Name == "Inlined":
|
||||||
for arg in remark.Args:
|
for arg in remark.Args:
|
||||||
caller = arg.get('Caller')
|
caller = arg.get('Caller')
|
||||||
if caller:
|
if caller:
|
||||||
Remark.caller_loc[caller] = arg['DebugLoc']
|
Remark.caller_loc[caller] = arg['DebugLoc']
|
||||||
|
|
||||||
if Remark.should_display_hotness():
|
|
||||||
|
def generate_report(pool, all_remarks, file_remarks, source_dir, output_dir):
|
||||||
|
try:
|
||||||
|
os.makedirs(output_dir)
|
||||||
|
except OSError as e:
|
||||||
|
if e.errno == errno.EEXIST and os.path.isdir(output_dir):
|
||||||
|
pass
|
||||||
|
else:
|
||||||
|
raise
|
||||||
|
|
||||||
|
_render_file_bound = functools.partial(_render_file, source_dir, output_dir)
|
||||||
|
pool.map(_render_file_bound, file_remarks.items())
|
||||||
|
|
||||||
|
if Remark.should_display_hotness():
|
||||||
sorted_remarks = sorted(all_remarks.itervalues(), key=lambda r: r.Hotness, reverse=True)
|
sorted_remarks = sorted(all_remarks.itervalues(), key=lambda r: r.Hotness, reverse=True)
|
||||||
else:
|
else:
|
||||||
sorted_remarks = sorted(all_remarks.itervalues(), key=lambda r: (r.File, r.Line, r.Column))
|
sorted_remarks = sorted(all_remarks.itervalues(), key=lambda r: (r.File, r.Line, r.Column))
|
||||||
|
IndexRenderer(args.output_dir).render(sorted_remarks)
|
||||||
|
|
||||||
if not os.path.exists(args.output_dir):
|
shutil.copy(os.path.join(os.path.dirname(os.path.realpath(__file__)),
|
||||||
os.mkdir(args.output_dir)
|
"style.css"), output_dir)
|
||||||
|
|
||||||
for (filename, remarks) in file_remarks.iteritems():
|
|
||||||
SourceFileRenderer(filename).render(remarks)
|
|
||||||
|
|
||||||
IndexRenderer().render(sorted_remarks)
|
if __name__ == '__main__':
|
||||||
|
parser = argparse.ArgumentParser(description=desc)
|
||||||
|
parser.add_argument('yaml_files', nargs='+')
|
||||||
|
parser.add_argument('output_dir')
|
||||||
|
parser.add_argument(
|
||||||
|
'--jobs',
|
||||||
|
'-j',
|
||||||
|
default=cpu_count(),
|
||||||
|
type=int,
|
||||||
|
help='Max job count (defaults to current CPU count)')
|
||||||
|
parser.add_argument(
|
||||||
|
'-source-dir',
|
||||||
|
'-s',
|
||||||
|
default='',
|
||||||
|
help='set source directory')
|
||||||
|
args = parser.parse_args()
|
||||||
|
|
||||||
shutil.copy(os.path.join(os.path.dirname(os.path.realpath(__file__)), "style.css"), args.output_dir)
|
if len(args.yaml_files) == 0:
|
||||||
|
parser.print_help()
|
||||||
|
sys.exit(1)
|
||||||
|
|
||||||
|
pool = Pool(processes=args.jobs)
|
||||||
|
all_remarks, file_remarks = gather_results(pool, args.yaml_files)
|
||||||
|
|
||||||
|
map_remarks(all_remarks)
|
||||||
|
|
||||||
|
generate_report(pool, all_remarks, file_remarks, args.source_dir, args.output_dir)
|
||||||
|
|
|
||||||
Loading…
Reference in New Issue