]> git.kernelconcepts.de Git - karo-tx-linux.git/blobdiff - tools/perf/util/hist.c
Merge branch 'master' of git://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux...
[karo-tx-linux.git] / tools / perf / util / hist.c
index b262b44b7a656d8dca4fcab7488e43e3c49ebb8c..30df6187ee026acf0fb9f12013fbde891d3bdd70 100644 (file)
@@ -4,6 +4,7 @@
 #include "session.h"
 #include "sort.h"
 #include "evsel.h"
+#include "annotate.h"
 #include <math.h>
 
 static bool hists__filter_entry_by_dso(struct hists *hists,
@@ -127,6 +128,8 @@ void hists__calc_col_len(struct hists *hists, struct hist_entry *h)
                               + unresolved_col_width + 2;
                        hists__new_col_len(hists, HISTC_MEM_DADDR_SYMBOL,
                                           symlen);
+                       hists__new_col_len(hists, HISTC_MEM_DCACHELINE,
+                                          symlen + 1);
                } else {
                        symlen = unresolved_col_width + 4 + 2;
                        hists__new_col_len(hists, HISTC_MEM_DADDR_SYMBOL,
@@ -231,6 +234,8 @@ static bool hists__decay_entry(struct hists *hists, struct hist_entry *he)
                return true;
 
        he_stat__decay(&he->stat);
+       if (symbol_conf.cumulate_callchain)
+               he_stat__decay(he->stat_acc);
 
        diff = prev_period - he->stat.period;
 
@@ -276,14 +281,31 @@ void hists__decay_entries(struct hists *hists, bool zap_user, bool zap_kernel)
  * histogram, sorted on item, collects periods
  */
 
-static struct hist_entry *hist_entry__new(struct hist_entry *template)
+static struct hist_entry *hist_entry__new(struct hist_entry *template,
+                                         bool sample_self)
 {
-       size_t callchain_size = symbol_conf.use_callchain ? sizeof(struct callchain_root) : 0;
-       struct hist_entry *he = zalloc(sizeof(*he) + callchain_size);
+       size_t callchain_size = 0;
+       struct hist_entry *he;
+
+       if (symbol_conf.use_callchain || symbol_conf.cumulate_callchain)
+               callchain_size = sizeof(struct callchain_root);
+
+       he = zalloc(sizeof(*he) + callchain_size);
 
        if (he != NULL) {
                *he = *template;
 
+               if (symbol_conf.cumulate_callchain) {
+                       he->stat_acc = malloc(sizeof(he->stat));
+                       if (he->stat_acc == NULL) {
+                               free(he);
+                               return NULL;
+                       }
+                       memcpy(he->stat_acc, &he->stat, sizeof(he->stat));
+                       if (!sample_self)
+                               memset(&he->stat, 0, sizeof(he->stat));
+               }
+
                if (he->ms.map)
                        he->ms.map->referenced = true;
 
@@ -295,6 +317,7 @@ static struct hist_entry *hist_entry__new(struct hist_entry *template)
                         */
                        he->branch_info = malloc(sizeof(*he->branch_info));
                        if (he->branch_info == NULL) {
+                               free(he->stat_acc);
                                free(he);
                                return NULL;
                        }
@@ -333,7 +356,8 @@ static u8 symbol__parent_filter(const struct symbol *parent)
 
 static struct hist_entry *add_hist_entry(struct hists *hists,
                                         struct hist_entry *entry,
-                                        struct addr_location *al)
+                                        struct addr_location *al,
+                                        bool sample_self)
 {
        struct rb_node **p;
        struct rb_node *parent = NULL;
@@ -357,7 +381,10 @@ static struct hist_entry *add_hist_entry(struct hists *hists,
                cmp = hist_entry__cmp(he, entry);
 
                if (!cmp) {
-                       he_stat__add_period(&he->stat, period, weight);
+                       if (sample_self)
+                               he_stat__add_period(&he->stat, period, weight);
+                       if (symbol_conf.cumulate_callchain)
+                               he_stat__add_period(he->stat_acc, period, weight);
 
                        /*
                         * This mem info was allocated from sample__resolve_mem
@@ -385,14 +412,17 @@ static struct hist_entry *add_hist_entry(struct hists *hists,
                        p = &(*p)->rb_right;
        }
 
-       he = hist_entry__new(entry);
+       he = hist_entry__new(entry, sample_self);
        if (!he)
                return NULL;
 
        rb_link_node(&he->rb_node_in, parent, p);
        rb_insert_color(&he->rb_node_in, hists->entries_in);
 out:
-       he_stat__add_cpumode_period(&he->stat, al->cpumode, period);
+       if (sample_self)
+               he_stat__add_cpumode_period(&he->stat, al->cpumode, period);
+       if (symbol_conf.cumulate_callchain)
+               he_stat__add_cpumode_period(he->stat_acc, al->cpumode, period);
        return he;
 }
 
@@ -401,7 +431,8 @@ struct hist_entry *__hists__add_entry(struct hists *hists,
                                      struct symbol *sym_parent,
                                      struct branch_info *bi,
                                      struct mem_info *mi,
-                                     u64 period, u64 weight, u64 transaction)
+                                     u64 period, u64 weight, u64 transaction,
+                                     bool sample_self)
 {
        struct hist_entry entry = {
                .thread = al->thread,
@@ -410,9 +441,10 @@ struct hist_entry *__hists__add_entry(struct hists *hists,
                        .map    = al->map,
                        .sym    = al->sym,
                },
-               .cpu    = al->cpu,
-               .ip     = al->addr,
-               .level  = al->level,
+               .cpu     = al->cpu,
+               .cpumode = al->cpumode,
+               .ip      = al->addr,
+               .level   = al->level,
                .stat = {
                        .nr_events = 1,
                        .period = period,
@@ -426,7 +458,429 @@ struct hist_entry *__hists__add_entry(struct hists *hists,
                .transaction = transaction,
        };
 
-       return add_hist_entry(hists, &entry, al);
+       return add_hist_entry(hists, &entry, al, sample_self);
+}
+
+static int
+iter_next_nop_entry(struct hist_entry_iter *iter __maybe_unused,
+                   struct addr_location *al __maybe_unused)
+{
+       return 0;
+}
+
+static int
+iter_add_next_nop_entry(struct hist_entry_iter *iter __maybe_unused,
+                       struct addr_location *al __maybe_unused)
+{
+       return 0;
+}
+
+static int
+iter_prepare_mem_entry(struct hist_entry_iter *iter, struct addr_location *al)
+{
+       struct perf_sample *sample = iter->sample;
+       struct mem_info *mi;
+
+       mi = sample__resolve_mem(sample, al);
+       if (mi == NULL)
+               return -ENOMEM;
+
+       iter->priv = mi;
+       return 0;
+}
+
+static int
+iter_add_single_mem_entry(struct hist_entry_iter *iter, struct addr_location *al)
+{
+       u64 cost;
+       struct mem_info *mi = iter->priv;
+       struct hist_entry *he;
+
+       if (mi == NULL)
+               return -EINVAL;
+
+       cost = iter->sample->weight;
+       if (!cost)
+               cost = 1;
+
+       /*
+        * must pass period=weight in order to get the correct
+        * sorting from hists__collapse_resort() which is solely
+        * based on periods. We want sorting be done on nr_events * weight
+        * and this is indirectly achieved by passing period=weight here
+        * and the he_stat__add_period() function.
+        */
+       he = __hists__add_entry(&iter->evsel->hists, al, iter->parent, NULL, mi,
+                               cost, cost, 0, true);
+       if (!he)
+               return -ENOMEM;
+
+       iter->he = he;
+       return 0;
+}
+
+static int
+iter_finish_mem_entry(struct hist_entry_iter *iter,
+                     struct addr_location *al __maybe_unused)
+{
+       struct perf_evsel *evsel = iter->evsel;
+       struct hist_entry *he = iter->he;
+       int err = -EINVAL;
+
+       if (he == NULL)
+               goto out;
+
+       hists__inc_nr_samples(&evsel->hists, he->filtered);
+
+       err = hist_entry__append_callchain(he, iter->sample);
+
+out:
+       /*
+        * We don't need to free iter->priv (mem_info) here since
+        * the mem info was either already freed in add_hist_entry() or
+        * passed to a new hist entry by hist_entry__new().
+        */
+       iter->priv = NULL;
+
+       iter->he = NULL;
+       return err;
+}
+
+static int
+iter_prepare_branch_entry(struct hist_entry_iter *iter, struct addr_location *al)
+{
+       struct branch_info *bi;
+       struct perf_sample *sample = iter->sample;
+
+       bi = sample__resolve_bstack(sample, al);
+       if (!bi)
+               return -ENOMEM;
+
+       iter->curr = 0;
+       iter->total = sample->branch_stack->nr;
+
+       iter->priv = bi;
+       return 0;
+}
+
+static int
+iter_add_single_branch_entry(struct hist_entry_iter *iter __maybe_unused,
+                            struct addr_location *al __maybe_unused)
+{
+       /* to avoid calling callback function */
+       iter->he = NULL;
+
+       return 0;
+}
+
+static int
+iter_next_branch_entry(struct hist_entry_iter *iter, struct addr_location *al)
+{
+       struct branch_info *bi = iter->priv;
+       int i = iter->curr;
+
+       if (bi == NULL)
+               return 0;
+
+       if (iter->curr >= iter->total)
+               return 0;
+
+       al->map = bi[i].to.map;
+       al->sym = bi[i].to.sym;
+       al->addr = bi[i].to.addr;
+       return 1;
+}
+
+static int
+iter_add_next_branch_entry(struct hist_entry_iter *iter, struct addr_location *al)
+{
+       struct branch_info *bi;
+       struct perf_evsel *evsel = iter->evsel;
+       struct hist_entry *he = NULL;
+       int i = iter->curr;
+       int err = 0;
+
+       bi = iter->priv;
+
+       if (iter->hide_unresolved && !(bi[i].from.sym && bi[i].to.sym))
+               goto out;
+
+       /*
+        * The report shows the percentage of total branches captured
+        * and not events sampled. Thus we use a pseudo period of 1.
+        */
+       he = __hists__add_entry(&evsel->hists, al, iter->parent, &bi[i], NULL,
+                               1, 1, 0, true);
+       if (he == NULL)
+               return -ENOMEM;
+
+       hists__inc_nr_samples(&evsel->hists, he->filtered);
+
+out:
+       iter->he = he;
+       iter->curr++;
+       return err;
+}
+
+static int
+iter_finish_branch_entry(struct hist_entry_iter *iter,
+                        struct addr_location *al __maybe_unused)
+{
+       zfree(&iter->priv);
+       iter->he = NULL;
+
+       return iter->curr >= iter->total ? 0 : -1;
+}
+
+static int
+iter_prepare_normal_entry(struct hist_entry_iter *iter __maybe_unused,
+                         struct addr_location *al __maybe_unused)
+{
+       return 0;
+}
+
+static int
+iter_add_single_normal_entry(struct hist_entry_iter *iter, struct addr_location *al)
+{
+       struct perf_evsel *evsel = iter->evsel;
+       struct perf_sample *sample = iter->sample;
+       struct hist_entry *he;
+
+       he = __hists__add_entry(&evsel->hists, al, iter->parent, NULL, NULL,
+                               sample->period, sample->weight,
+                               sample->transaction, true);
+       if (he == NULL)
+               return -ENOMEM;
+
+       iter->he = he;
+       return 0;
+}
+
+static int
+iter_finish_normal_entry(struct hist_entry_iter *iter,
+                        struct addr_location *al __maybe_unused)
+{
+       struct hist_entry *he = iter->he;
+       struct perf_evsel *evsel = iter->evsel;
+       struct perf_sample *sample = iter->sample;
+
+       if (he == NULL)
+               return 0;
+
+       iter->he = NULL;
+
+       hists__inc_nr_samples(&evsel->hists, he->filtered);
+
+       return hist_entry__append_callchain(he, sample);
+}
+
+static int
+iter_prepare_cumulative_entry(struct hist_entry_iter *iter __maybe_unused,
+                             struct addr_location *al __maybe_unused)
+{
+       struct hist_entry **he_cache;
+
+       callchain_cursor_commit(&callchain_cursor);
+
+       /*
+        * This is for detecting cycles or recursions so that they're
+        * cumulated only one time to prevent entries more than 100%
+        * overhead.
+        */
+       he_cache = malloc(sizeof(*he_cache) * (PERF_MAX_STACK_DEPTH + 1));
+       if (he_cache == NULL)
+               return -ENOMEM;
+
+       iter->priv = he_cache;
+       iter->curr = 0;
+
+       return 0;
+}
+
+static int
+iter_add_single_cumulative_entry(struct hist_entry_iter *iter,
+                                struct addr_location *al)
+{
+       struct perf_evsel *evsel = iter->evsel;
+       struct perf_sample *sample = iter->sample;
+       struct hist_entry **he_cache = iter->priv;
+       struct hist_entry *he;
+       int err = 0;
+
+       he = __hists__add_entry(&evsel->hists, al, iter->parent, NULL, NULL,
+                               sample->period, sample->weight,
+                               sample->transaction, true);
+       if (he == NULL)
+               return -ENOMEM;
+
+       iter->he = he;
+       he_cache[iter->curr++] = he;
+
+       callchain_append(he->callchain, &callchain_cursor, sample->period);
+
+       /*
+        * We need to re-initialize the cursor since callchain_append()
+        * advanced the cursor to the end.
+        */
+       callchain_cursor_commit(&callchain_cursor);
+
+       hists__inc_nr_samples(&evsel->hists, he->filtered);
+
+       return err;
+}
+
+static int
+iter_next_cumulative_entry(struct hist_entry_iter *iter,
+                          struct addr_location *al)
+{
+       struct callchain_cursor_node *node;
+
+       node = callchain_cursor_current(&callchain_cursor);
+       if (node == NULL)
+               return 0;
+
+       return fill_callchain_info(al, node, iter->hide_unresolved);
+}
+
+static int
+iter_add_next_cumulative_entry(struct hist_entry_iter *iter,
+                              struct addr_location *al)
+{
+       struct perf_evsel *evsel = iter->evsel;
+       struct perf_sample *sample = iter->sample;
+       struct hist_entry **he_cache = iter->priv;
+       struct hist_entry *he;
+       struct hist_entry he_tmp = {
+               .cpu = al->cpu,
+               .thread = al->thread,
+               .comm = thread__comm(al->thread),
+               .ip = al->addr,
+               .ms = {
+                       .map = al->map,
+                       .sym = al->sym,
+               },
+               .parent = iter->parent,
+       };
+       int i;
+       struct callchain_cursor cursor;
+
+       callchain_cursor_snapshot(&cursor, &callchain_cursor);
+
+       callchain_cursor_advance(&callchain_cursor);
+
+       /*
+        * Check if there's duplicate entries in the callchain.
+        * It's possible that it has cycles or recursive calls.
+        */
+       for (i = 0; i < iter->curr; i++) {
+               if (hist_entry__cmp(he_cache[i], &he_tmp) == 0) {
+                       /* to avoid calling callback function */
+                       iter->he = NULL;
+                       return 0;
+               }
+       }
+
+       he = __hists__add_entry(&evsel->hists, al, iter->parent, NULL, NULL,
+                               sample->period, sample->weight,
+                               sample->transaction, false);
+       if (he == NULL)
+               return -ENOMEM;
+
+       iter->he = he;
+       he_cache[iter->curr++] = he;
+
+       callchain_append(he->callchain, &cursor, sample->period);
+       return 0;
+}
+
+static int
+iter_finish_cumulative_entry(struct hist_entry_iter *iter,
+                            struct addr_location *al __maybe_unused)
+{
+       zfree(&iter->priv);
+       iter->he = NULL;
+
+       return 0;
+}
+
+const struct hist_iter_ops hist_iter_mem = {
+       .prepare_entry          = iter_prepare_mem_entry,
+       .add_single_entry       = iter_add_single_mem_entry,
+       .next_entry             = iter_next_nop_entry,
+       .add_next_entry         = iter_add_next_nop_entry,
+       .finish_entry           = iter_finish_mem_entry,
+};
+
+const struct hist_iter_ops hist_iter_branch = {
+       .prepare_entry          = iter_prepare_branch_entry,
+       .add_single_entry       = iter_add_single_branch_entry,
+       .next_entry             = iter_next_branch_entry,
+       .add_next_entry         = iter_add_next_branch_entry,
+       .finish_entry           = iter_finish_branch_entry,
+};
+
+const struct hist_iter_ops hist_iter_normal = {
+       .prepare_entry          = iter_prepare_normal_entry,
+       .add_single_entry       = iter_add_single_normal_entry,
+       .next_entry             = iter_next_nop_entry,
+       .add_next_entry         = iter_add_next_nop_entry,
+       .finish_entry           = iter_finish_normal_entry,
+};
+
+const struct hist_iter_ops hist_iter_cumulative = {
+       .prepare_entry          = iter_prepare_cumulative_entry,
+       .add_single_entry       = iter_add_single_cumulative_entry,
+       .next_entry             = iter_next_cumulative_entry,
+       .add_next_entry         = iter_add_next_cumulative_entry,
+       .finish_entry           = iter_finish_cumulative_entry,
+};
+
+int hist_entry_iter__add(struct hist_entry_iter *iter, struct addr_location *al,
+                        struct perf_evsel *evsel, struct perf_sample *sample,
+                        int max_stack_depth, void *arg)
+{
+       int err, err2;
+
+       err = sample__resolve_callchain(sample, &iter->parent, evsel, al,
+                                       max_stack_depth);
+       if (err)
+               return err;
+
+       iter->evsel = evsel;
+       iter->sample = sample;
+
+       err = iter->ops->prepare_entry(iter, al);
+       if (err)
+               goto out;
+
+       err = iter->ops->add_single_entry(iter, al);
+       if (err)
+               goto out;
+
+       if (iter->he && iter->add_entry_cb) {
+               err = iter->add_entry_cb(iter, al, true, arg);
+               if (err)
+                       goto out;
+       }
+
+       while (iter->ops->next_entry(iter, al)) {
+               err = iter->ops->add_next_entry(iter, al);
+               if (err)
+                       break;
+
+               if (iter->he && iter->add_entry_cb) {
+                       err = iter->add_entry_cb(iter, al, false, arg);
+                       if (err)
+                               goto out;
+               }
+       }
+
+out:
+       err2 = iter->ops->finish_entry(iter, al);
+       if (!err)
+               err = err2;
+
+       return err;
 }
 
 int64_t
@@ -469,6 +923,7 @@ void hist_entry__free(struct hist_entry *he)
 {
        zfree(&he->branch_info);
        zfree(&he->mem_info);
+       zfree(&he->stat_acc);
        free_srcline(he->srcline);
        free(he);
 }
@@ -494,6 +949,8 @@ static bool hists__collapse_insert_entry(struct hists *hists __maybe_unused,
 
                if (!cmp) {
                        he_stat__add_stat(&iter->stat, &he->stat);
+                       if (symbol_conf.cumulate_callchain)
+                               he_stat__add_stat(iter->stat_acc, he->stat_acc);
 
                        if (symbol_conf.use_callchain) {
                                callchain_cursor_reset(&callchain_cursor);
@@ -800,6 +1257,13 @@ void hists__inc_nr_events(struct hists *hists, u32 type)
        events_stats__inc(&hists->stats, type);
 }
 
+void hists__inc_nr_samples(struct hists *hists, bool filtered)
+{
+       events_stats__inc(&hists->stats, PERF_RECORD_SAMPLE);
+       if (!filtered)
+               hists->stats.nr_non_filtered_samples++;
+}
+
 static struct hist_entry *hists__add_dummy_entry(struct hists *hists,
                                                 struct hist_entry *pair)
 {
@@ -831,7 +1295,7 @@ static struct hist_entry *hists__add_dummy_entry(struct hists *hists,
                        p = &(*p)->rb_right;
        }
 
-       he = hist_entry__new(pair);
+       he = hist_entry__new(pair, true);
        if (he) {
                memset(&he->stat, 0, sizeof(he->stat));
                he->hists = hists;