perf hists: Fix indent for multiple hierarchy sort key
[deliverable/linux.git] / tools / perf / ui / stdio / hist.c
1 #include <stdio.h>
2
3 #include "../../util/util.h"
4 #include "../../util/hist.h"
5 #include "../../util/sort.h"
6 #include "../../util/evsel.h"
7
8
9 static size_t callchain__fprintf_left_margin(FILE *fp, int left_margin)
10 {
11 int i;
12 int ret = fprintf(fp, " ");
13
14 for (i = 0; i < left_margin; i++)
15 ret += fprintf(fp, " ");
16
17 return ret;
18 }
19
20 static size_t ipchain__fprintf_graph_line(FILE *fp, int depth, int depth_mask,
21 int left_margin)
22 {
23 int i;
24 size_t ret = callchain__fprintf_left_margin(fp, left_margin);
25
26 for (i = 0; i < depth; i++)
27 if (depth_mask & (1 << i))
28 ret += fprintf(fp, "| ");
29 else
30 ret += fprintf(fp, " ");
31
32 ret += fprintf(fp, "\n");
33
34 return ret;
35 }
36
37 static size_t ipchain__fprintf_graph(FILE *fp, struct callchain_node *node,
38 struct callchain_list *chain,
39 int depth, int depth_mask, int period,
40 u64 total_samples, int left_margin)
41 {
42 int i;
43 size_t ret = 0;
44 char bf[1024];
45
46 ret += callchain__fprintf_left_margin(fp, left_margin);
47 for (i = 0; i < depth; i++) {
48 if (depth_mask & (1 << i))
49 ret += fprintf(fp, "|");
50 else
51 ret += fprintf(fp, " ");
52 if (!period && i == depth - 1) {
53 ret += fprintf(fp, "--");
54 ret += callchain_node__fprintf_value(node, fp, total_samples);
55 ret += fprintf(fp, "--");
56 } else
57 ret += fprintf(fp, "%s", " ");
58 }
59 fputs(callchain_list__sym_name(chain, bf, sizeof(bf), false), fp);
60 fputc('\n', fp);
61 return ret;
62 }
63
64 static struct symbol *rem_sq_bracket;
65 static struct callchain_list rem_hits;
66
67 static void init_rem_hits(void)
68 {
69 rem_sq_bracket = malloc(sizeof(*rem_sq_bracket) + 6);
70 if (!rem_sq_bracket) {
71 fprintf(stderr, "Not enough memory to display remaining hits\n");
72 return;
73 }
74
75 strcpy(rem_sq_bracket->name, "[...]");
76 rem_hits.ms.sym = rem_sq_bracket;
77 }
78
79 static size_t __callchain__fprintf_graph(FILE *fp, struct rb_root *root,
80 u64 total_samples, int depth,
81 int depth_mask, int left_margin)
82 {
83 struct rb_node *node, *next;
84 struct callchain_node *child = NULL;
85 struct callchain_list *chain;
86 int new_depth_mask = depth_mask;
87 u64 remaining;
88 size_t ret = 0;
89 int i;
90 uint entries_printed = 0;
91 int cumul_count = 0;
92
93 remaining = total_samples;
94
95 node = rb_first(root);
96 while (node) {
97 u64 new_total;
98 u64 cumul;
99
100 child = rb_entry(node, struct callchain_node, rb_node);
101 cumul = callchain_cumul_hits(child);
102 remaining -= cumul;
103 cumul_count += callchain_cumul_counts(child);
104
105 /*
106 * The depth mask manages the output of pipes that show
107 * the depth. We don't want to keep the pipes of the current
108 * level for the last child of this depth.
109 * Except if we have remaining filtered hits. They will
110 * supersede the last child
111 */
112 next = rb_next(node);
113 if (!next && (callchain_param.mode != CHAIN_GRAPH_REL || !remaining))
114 new_depth_mask &= ~(1 << (depth - 1));
115
116 /*
117 * But we keep the older depth mask for the line separator
118 * to keep the level link until we reach the last child
119 */
120 ret += ipchain__fprintf_graph_line(fp, depth, depth_mask,
121 left_margin);
122 i = 0;
123 list_for_each_entry(chain, &child->val, list) {
124 ret += ipchain__fprintf_graph(fp, child, chain, depth,
125 new_depth_mask, i++,
126 total_samples,
127 left_margin);
128 }
129
130 if (callchain_param.mode == CHAIN_GRAPH_REL)
131 new_total = child->children_hit;
132 else
133 new_total = total_samples;
134
135 ret += __callchain__fprintf_graph(fp, &child->rb_root, new_total,
136 depth + 1,
137 new_depth_mask | (1 << depth),
138 left_margin);
139 node = next;
140 if (++entries_printed == callchain_param.print_limit)
141 break;
142 }
143
144 if (callchain_param.mode == CHAIN_GRAPH_REL &&
145 remaining && remaining != total_samples) {
146 struct callchain_node rem_node = {
147 .hit = remaining,
148 };
149
150 if (!rem_sq_bracket)
151 return ret;
152
153 if (callchain_param.value == CCVAL_COUNT && child && child->parent) {
154 rem_node.count = child->parent->children_count - cumul_count;
155 if (rem_node.count <= 0)
156 return ret;
157 }
158
159 new_depth_mask &= ~(1 << (depth - 1));
160 ret += ipchain__fprintf_graph(fp, &rem_node, &rem_hits, depth,
161 new_depth_mask, 0, total_samples,
162 left_margin);
163 }
164
165 return ret;
166 }
167
168 /*
169 * If have one single callchain root, don't bother printing
170 * its percentage (100 % in fractal mode and the same percentage
171 * than the hist in graph mode). This also avoid one level of column.
172 *
173 * However when percent-limit applied, it's possible that single callchain
174 * node have different (non-100% in fractal mode) percentage.
175 */
176 static bool need_percent_display(struct rb_node *node, u64 parent_samples)
177 {
178 struct callchain_node *cnode;
179
180 if (rb_next(node))
181 return true;
182
183 cnode = rb_entry(node, struct callchain_node, rb_node);
184 return callchain_cumul_hits(cnode) != parent_samples;
185 }
186
187 static size_t callchain__fprintf_graph(FILE *fp, struct rb_root *root,
188 u64 total_samples, u64 parent_samples,
189 int left_margin)
190 {
191 struct callchain_node *cnode;
192 struct callchain_list *chain;
193 u32 entries_printed = 0;
194 bool printed = false;
195 struct rb_node *node;
196 int i = 0;
197 int ret = 0;
198 char bf[1024];
199
200 node = rb_first(root);
201 if (node && !need_percent_display(node, parent_samples)) {
202 cnode = rb_entry(node, struct callchain_node, rb_node);
203 list_for_each_entry(chain, &cnode->val, list) {
204 /*
205 * If we sort by symbol, the first entry is the same than
206 * the symbol. No need to print it otherwise it appears as
207 * displayed twice.
208 */
209 if (!i++ && field_order == NULL &&
210 sort_order && !prefixcmp(sort_order, "sym"))
211 continue;
212 if (!printed) {
213 ret += callchain__fprintf_left_margin(fp, left_margin);
214 ret += fprintf(fp, "|\n");
215 ret += callchain__fprintf_left_margin(fp, left_margin);
216 ret += fprintf(fp, "---");
217 left_margin += 3;
218 printed = true;
219 } else
220 ret += callchain__fprintf_left_margin(fp, left_margin);
221
222 ret += fprintf(fp, "%s\n", callchain_list__sym_name(chain, bf, sizeof(bf),
223 false));
224
225 if (++entries_printed == callchain_param.print_limit)
226 break;
227 }
228 root = &cnode->rb_root;
229 }
230
231 if (callchain_param.mode == CHAIN_GRAPH_REL)
232 total_samples = parent_samples;
233
234 ret += __callchain__fprintf_graph(fp, root, total_samples,
235 1, 1, left_margin);
236 if (ret) {
237 /* do not add a blank line if it printed nothing */
238 ret += fprintf(fp, "\n");
239 }
240
241 return ret;
242 }
243
244 static size_t __callchain__fprintf_flat(FILE *fp, struct callchain_node *node,
245 u64 total_samples)
246 {
247 struct callchain_list *chain;
248 size_t ret = 0;
249 char bf[1024];
250
251 if (!node)
252 return 0;
253
254 ret += __callchain__fprintf_flat(fp, node->parent, total_samples);
255
256
257 list_for_each_entry(chain, &node->val, list) {
258 if (chain->ip >= PERF_CONTEXT_MAX)
259 continue;
260 ret += fprintf(fp, " %s\n", callchain_list__sym_name(chain,
261 bf, sizeof(bf), false));
262 }
263
264 return ret;
265 }
266
267 static size_t callchain__fprintf_flat(FILE *fp, struct rb_root *tree,
268 u64 total_samples)
269 {
270 size_t ret = 0;
271 u32 entries_printed = 0;
272 struct callchain_node *chain;
273 struct rb_node *rb_node = rb_first(tree);
274
275 while (rb_node) {
276 chain = rb_entry(rb_node, struct callchain_node, rb_node);
277
278 ret += fprintf(fp, " ");
279 ret += callchain_node__fprintf_value(chain, fp, total_samples);
280 ret += fprintf(fp, "\n");
281 ret += __callchain__fprintf_flat(fp, chain, total_samples);
282 ret += fprintf(fp, "\n");
283 if (++entries_printed == callchain_param.print_limit)
284 break;
285
286 rb_node = rb_next(rb_node);
287 }
288
289 return ret;
290 }
291
292 static size_t __callchain__fprintf_folded(FILE *fp, struct callchain_node *node)
293 {
294 const char *sep = symbol_conf.field_sep ?: ";";
295 struct callchain_list *chain;
296 size_t ret = 0;
297 char bf[1024];
298 bool first;
299
300 if (!node)
301 return 0;
302
303 ret += __callchain__fprintf_folded(fp, node->parent);
304
305 first = (ret == 0);
306 list_for_each_entry(chain, &node->val, list) {
307 if (chain->ip >= PERF_CONTEXT_MAX)
308 continue;
309 ret += fprintf(fp, "%s%s", first ? "" : sep,
310 callchain_list__sym_name(chain,
311 bf, sizeof(bf), false));
312 first = false;
313 }
314
315 return ret;
316 }
317
318 static size_t callchain__fprintf_folded(FILE *fp, struct rb_root *tree,
319 u64 total_samples)
320 {
321 size_t ret = 0;
322 u32 entries_printed = 0;
323 struct callchain_node *chain;
324 struct rb_node *rb_node = rb_first(tree);
325
326 while (rb_node) {
327
328 chain = rb_entry(rb_node, struct callchain_node, rb_node);
329
330 ret += callchain_node__fprintf_value(chain, fp, total_samples);
331 ret += fprintf(fp, " ");
332 ret += __callchain__fprintf_folded(fp, chain);
333 ret += fprintf(fp, "\n");
334 if (++entries_printed == callchain_param.print_limit)
335 break;
336
337 rb_node = rb_next(rb_node);
338 }
339
340 return ret;
341 }
342
343 static size_t hist_entry_callchain__fprintf(struct hist_entry *he,
344 u64 total_samples, int left_margin,
345 FILE *fp)
346 {
347 u64 parent_samples = he->stat.period;
348
349 if (symbol_conf.cumulate_callchain)
350 parent_samples = he->stat_acc->period;
351
352 switch (callchain_param.mode) {
353 case CHAIN_GRAPH_REL:
354 return callchain__fprintf_graph(fp, &he->sorted_chain, total_samples,
355 parent_samples, left_margin);
356 break;
357 case CHAIN_GRAPH_ABS:
358 return callchain__fprintf_graph(fp, &he->sorted_chain, total_samples,
359 parent_samples, left_margin);
360 break;
361 case CHAIN_FLAT:
362 return callchain__fprintf_flat(fp, &he->sorted_chain, total_samples);
363 break;
364 case CHAIN_FOLDED:
365 return callchain__fprintf_folded(fp, &he->sorted_chain, total_samples);
366 break;
367 case CHAIN_NONE:
368 break;
369 default:
370 pr_err("Bad callchain mode\n");
371 }
372
373 return 0;
374 }
375
376 static int hist_entry__snprintf(struct hist_entry *he, struct perf_hpp *hpp)
377 {
378 const char *sep = symbol_conf.field_sep;
379 struct perf_hpp_fmt *fmt;
380 char *start = hpp->buf;
381 int ret;
382 bool first = true;
383
384 if (symbol_conf.exclude_other && !he->parent)
385 return 0;
386
387 hists__for_each_format(he->hists, fmt) {
388 if (perf_hpp__should_skip(fmt, he->hists))
389 continue;
390
391 /*
392 * If there's no field_sep, we still need
393 * to display initial ' '.
394 */
395 if (!sep || !first) {
396 ret = scnprintf(hpp->buf, hpp->size, "%s", sep ?: " ");
397 advance_hpp(hpp, ret);
398 } else
399 first = false;
400
401 if (perf_hpp__use_color() && fmt->color)
402 ret = fmt->color(fmt, hpp, he);
403 else
404 ret = fmt->entry(fmt, hpp, he);
405
406 ret = hist_entry__snprintf_alignment(he, hpp, fmt, ret);
407 advance_hpp(hpp, ret);
408 }
409
410 return hpp->buf - start;
411 }
412
413 static int hist_entry__hierarchy_fprintf(struct hist_entry *he,
414 struct perf_hpp *hpp,
415 struct hists *hists,
416 FILE *fp)
417 {
418 const char *sep = symbol_conf.field_sep;
419 struct perf_hpp_fmt *fmt;
420 char *buf = hpp->buf;
421 size_t size = hpp->size;
422 int ret, printed = 0;
423 bool first = true;
424
425 if (symbol_conf.exclude_other && !he->parent)
426 return 0;
427
428 ret = scnprintf(hpp->buf, hpp->size, "%*s", he->depth * HIERARCHY_INDENT, "");
429 advance_hpp(hpp, ret);
430
431 hists__for_each_format(he->hists, fmt) {
432 if (perf_hpp__is_sort_entry(fmt) || perf_hpp__is_dynamic_entry(fmt))
433 break;
434
435 /*
436 * If there's no field_sep, we still need
437 * to display initial ' '.
438 */
439 if (!sep || !first) {
440 ret = scnprintf(hpp->buf, hpp->size, "%s", sep ?: " ");
441 advance_hpp(hpp, ret);
442 } else
443 first = false;
444
445 if (perf_hpp__use_color() && fmt->color)
446 ret = fmt->color(fmt, hpp, he);
447 else
448 ret = fmt->entry(fmt, hpp, he);
449
450 ret = hist_entry__snprintf_alignment(he, hpp, fmt, ret);
451 advance_hpp(hpp, ret);
452 }
453
454 if (!sep)
455 ret = scnprintf(hpp->buf, hpp->size, "%*s",
456 (hists->nr_hpp_node - 2) * HIERARCHY_INDENT, "");
457 advance_hpp(hpp, ret);
458
459 printed += fprintf(fp, "%s", buf);
460
461 perf_hpp_list__for_each_format(he->hpp_list, fmt) {
462 hpp->buf = buf;
463 hpp->size = size;
464
465 /*
466 * No need to call hist_entry__snprintf_alignment() since this
467 * fmt is always the last column in the hierarchy mode.
468 */
469 if (perf_hpp__use_color() && fmt->color)
470 fmt->color(fmt, hpp, he);
471 else
472 fmt->entry(fmt, hpp, he);
473
474 /*
475 * dynamic entries are right-aligned but we want left-aligned
476 * in the hierarchy mode
477 */
478 printed += fprintf(fp, "%s%s", sep ?: " ", ltrim(buf));
479 }
480 printed += putc('\n', fp);
481
482 if (symbol_conf.use_callchain && he->leaf) {
483 u64 total = hists__total_period(hists);
484
485 printed += hist_entry_callchain__fprintf(he, total, 0, fp);
486 goto out;
487 }
488
489 out:
490 return printed;
491 }
492
493 static int hist_entry__fprintf(struct hist_entry *he, size_t size,
494 struct hists *hists,
495 char *bf, size_t bfsz, FILE *fp)
496 {
497 int ret;
498 struct perf_hpp hpp = {
499 .buf = bf,
500 .size = size,
501 };
502 u64 total_period = hists->stats.total_period;
503
504 if (size == 0 || size > bfsz)
505 size = hpp.size = bfsz;
506
507 if (symbol_conf.report_hierarchy)
508 return hist_entry__hierarchy_fprintf(he, &hpp, hists, fp);
509
510 hist_entry__snprintf(he, &hpp);
511
512 ret = fprintf(fp, "%s\n", bf);
513
514 if (symbol_conf.use_callchain)
515 ret += hist_entry_callchain__fprintf(he, total_period, 0, fp);
516
517 return ret;
518 }
519
520 static int print_hierarchy_indent(const char *sep, int indent,
521 const char *line, FILE *fp)
522 {
523 if (sep != NULL || indent < 2)
524 return 0;
525
526 return fprintf(fp, "%-.*s", (indent - 2) * HIERARCHY_INDENT, line);
527 }
528
529 static int print_hierarchy_header(struct hists *hists, struct perf_hpp *hpp,
530 const char *sep, FILE *fp)
531 {
532 bool first = true;
533 int indent;
534 int depth;
535 unsigned width = 0;
536 unsigned header_width = 0;
537 struct perf_hpp_fmt *fmt;
538
539 indent = hists->nr_hpp_node;
540
541 /* preserve max indent depth for column headers */
542 print_hierarchy_indent(sep, indent, spaces, fp);
543
544 hists__for_each_format(hists, fmt) {
545 if (perf_hpp__is_sort_entry(fmt) || perf_hpp__is_dynamic_entry(fmt))
546 break;
547
548 if (!first)
549 fprintf(fp, "%s", sep ?: " ");
550 else
551 first = false;
552
553 fmt->header(fmt, hpp, hists_to_evsel(hists));
554 fprintf(fp, "%s", hpp->buf);
555 }
556
557 /* combine sort headers with ' / ' */
558 first = true;
559 hists__for_each_format(hists, fmt) {
560 if (!perf_hpp__is_sort_entry(fmt) && !perf_hpp__is_dynamic_entry(fmt))
561 continue;
562 if (perf_hpp__should_skip(fmt, hists))
563 continue;
564
565 if (!first)
566 header_width += fprintf(fp, " / ");
567 else {
568 fprintf(fp, "%s", sep ?: " ");
569 first = false;
570 }
571
572 fmt->header(fmt, hpp, hists_to_evsel(hists));
573 rtrim(hpp->buf);
574
575 header_width += fprintf(fp, "%s", ltrim(hpp->buf));
576 }
577
578 fprintf(fp, "\n# ");
579
580 /* preserve max indent depth for initial dots */
581 print_hierarchy_indent(sep, indent, dots, fp);
582
583 first = true;
584 hists__for_each_format(hists, fmt) {
585 if (perf_hpp__is_sort_entry(fmt) || perf_hpp__is_dynamic_entry(fmt))
586 break;
587
588 if (!first)
589 fprintf(fp, "%s", sep ?: " ");
590 else
591 first = false;
592
593 width = fmt->width(fmt, hpp, hists_to_evsel(hists));
594 fprintf(fp, "%.*s", width, dots);
595 }
596
597 depth = 0;
598 hists__for_each_format(hists, fmt) {
599 if (!perf_hpp__is_sort_entry(fmt) && !perf_hpp__is_dynamic_entry(fmt))
600 continue;
601 if (perf_hpp__should_skip(fmt, hists))
602 continue;
603
604 width = fmt->width(fmt, hpp, hists_to_evsel(hists));
605 width += depth * HIERARCHY_INDENT;
606
607 if (width > header_width)
608 header_width = width;
609
610 depth++;
611 }
612
613 fprintf(fp, "%s%-.*s", sep ?: " ", header_width, dots);
614
615 fprintf(fp, "\n#\n");
616
617 return 2;
618 }
619
620 size_t hists__fprintf(struct hists *hists, bool show_header, int max_rows,
621 int max_cols, float min_pcnt, FILE *fp)
622 {
623 struct perf_hpp_fmt *fmt;
624 struct rb_node *nd;
625 size_t ret = 0;
626 unsigned int width;
627 const char *sep = symbol_conf.field_sep;
628 int nr_rows = 0;
629 char bf[96];
630 struct perf_hpp dummy_hpp = {
631 .buf = bf,
632 .size = sizeof(bf),
633 };
634 bool first = true;
635 size_t linesz;
636 char *line = NULL;
637 unsigned indent;
638
639 init_rem_hits();
640
641 hists__for_each_format(hists, fmt)
642 perf_hpp__reset_width(fmt, hists);
643
644 if (symbol_conf.col_width_list_str)
645 perf_hpp__set_user_width(symbol_conf.col_width_list_str);
646
647 if (!show_header)
648 goto print_entries;
649
650 fprintf(fp, "# ");
651
652 if (symbol_conf.report_hierarchy) {
653 nr_rows += print_hierarchy_header(hists, &dummy_hpp, sep, fp);
654 goto print_entries;
655 }
656
657 hists__for_each_format(hists, fmt) {
658 if (perf_hpp__should_skip(fmt, hists))
659 continue;
660
661 if (!first)
662 fprintf(fp, "%s", sep ?: " ");
663 else
664 first = false;
665
666 fmt->header(fmt, &dummy_hpp, hists_to_evsel(hists));
667 fprintf(fp, "%s", bf);
668 }
669
670 fprintf(fp, "\n");
671 if (max_rows && ++nr_rows >= max_rows)
672 goto out;
673
674 if (sep)
675 goto print_entries;
676
677 first = true;
678
679 fprintf(fp, "# ");
680
681 hists__for_each_format(hists, fmt) {
682 unsigned int i;
683
684 if (perf_hpp__should_skip(fmt, hists))
685 continue;
686
687 if (!first)
688 fprintf(fp, "%s", sep ?: " ");
689 else
690 first = false;
691
692 width = fmt->width(fmt, &dummy_hpp, hists_to_evsel(hists));
693 for (i = 0; i < width; i++)
694 fprintf(fp, ".");
695 }
696
697 fprintf(fp, "\n");
698 if (max_rows && ++nr_rows >= max_rows)
699 goto out;
700
701 fprintf(fp, "#\n");
702 if (max_rows && ++nr_rows >= max_rows)
703 goto out;
704
705 print_entries:
706 linesz = hists__sort_list_width(hists) + 3 + 1;
707 linesz += perf_hpp__color_overhead();
708 line = malloc(linesz);
709 if (line == NULL) {
710 ret = -1;
711 goto out;
712 }
713
714 indent = hists__overhead_width(hists) + 4;
715
716 for (nd = rb_first(&hists->entries); nd; nd = __rb_hierarchy_next(nd, HMD_FORCE_CHILD)) {
717 struct hist_entry *h = rb_entry(nd, struct hist_entry, rb_node);
718 float percent;
719
720 if (h->filtered)
721 continue;
722
723 percent = hist_entry__get_percent_limit(h);
724 if (percent < min_pcnt)
725 continue;
726
727 ret += hist_entry__fprintf(h, max_cols, hists, line, linesz, fp);
728
729 if (max_rows && ++nr_rows >= max_rows)
730 break;
731
732 /*
733 * If all children are filtered out or percent-limited,
734 * display "no entry >= x.xx%" message.
735 */
736 if (!h->leaf && !hist_entry__has_hierarchy_children(h, min_pcnt)) {
737 int nr_sort = hists->nr_sort_keys;
738
739 print_hierarchy_indent(sep, nr_sort + h->depth + 1, spaces, fp);
740 fprintf(fp, "%*sno entry >= %.2f%%\n", indent, "", min_pcnt);
741
742 if (max_rows && ++nr_rows >= max_rows)
743 break;
744 }
745
746 if (h->ms.map == NULL && verbose > 1) {
747 __map_groups__fprintf_maps(h->thread->mg,
748 MAP__FUNCTION, fp);
749 fprintf(fp, "%.10s end\n", graph_dotted_line);
750 }
751 }
752
753 free(line);
754 out:
755 zfree(&rem_sq_bracket);
756
757 return ret;
758 }
759
760 size_t events_stats__fprintf(struct events_stats *stats, FILE *fp)
761 {
762 int i;
763 size_t ret = 0;
764
765 for (i = 0; i < PERF_RECORD_HEADER_MAX; ++i) {
766 const char *name;
767
768 if (stats->nr_events[i] == 0)
769 continue;
770
771 name = perf_event__name(i);
772 if (!strcmp(name, "UNKNOWN"))
773 continue;
774
775 ret += fprintf(fp, "%16s events: %10d\n", name,
776 stats->nr_events[i]);
777 }
778
779 return ret;
780 }
This page took 0.059276 seconds and 5 git commands to generate.