libata-acpi: missing _SDD is not an error
[safe/jmp/linux-2.6] / tools / perf / builtin-sched.c
index f856a02..ce2d5be 100644 (file)
@@ -50,7 +50,7 @@ static u64                    sleep_measurement_overhead;
 
 static unsigned long           nr_tasks;
 
-struct sched_event;
+struct sched_atom;
 
 struct task_desc {
        unsigned long           nr;
@@ -59,7 +59,7 @@ struct task_desc {
 
        unsigned long           nr_events;
        unsigned long           curr_event;
-       struct sched_event      **events;
+       struct sched_atom       **atoms;
 
        pthread_t               thread;
        sem_t                   sleep_sem;
@@ -76,7 +76,7 @@ enum sched_event_type {
        SCHED_EVENT_WAKEUP,
 };
 
-struct sched_event {
+struct sched_atom {
        enum sched_event_type   type;
        u64                     timestamp;
        u64                     duration;
@@ -117,7 +117,12 @@ static u64                 run_avg;
 
 static unsigned long           replay_repeat = 10;
 static unsigned long           nr_timestamps;
-static unsigned long           unordered_timestamps;
+static unsigned long           nr_unordered_timestamps;
+static unsigned long           nr_state_machine_bugs;
+static unsigned long           nr_context_switch_bugs;
+static unsigned long           nr_events;
+static unsigned long           nr_lost_chunks;
+static unsigned long           nr_lost_events;
 
 #define TASK_STATE_TO_CHAR_STR "RSDTtZX"
 
@@ -137,8 +142,8 @@ struct work_atom {
        u64                     runtime;
 };
 
-struct task_atoms {
-       struct list_head        atom_list;
+struct work_atoms {
+       struct list_head        work_list;
        struct thread           *thread;
        struct rb_node          node;
        u64                     max_lat;
@@ -147,15 +152,13 @@ struct task_atoms {
        u64                     total_runtime;
 };
 
-typedef int (*sort_fn_t)(struct task_atoms *, struct task_atoms *);
+typedef int (*sort_fn_t)(struct work_atoms *, struct work_atoms *);
 
 static struct rb_root          atom_root, sorted_atom_root;
 
 static u64                     all_runtime;
 static u64                     all_count;
 
-static int read_events(void);
-
 
 static u64 get_nsecs(void)
 {
@@ -220,10 +223,10 @@ static void calibrate_sleep_measurement_overhead(void)
        printf("sleep measurement overhead: %Ld nsecs\n", min_delta);
 }
 
-static struct sched_event *
+static struct sched_atom *
 get_new_event(struct task_desc *task, u64 timestamp)
 {
-       struct sched_event *event = calloc(1, sizeof(*event));
+       struct sched_atom *event = calloc(1, sizeof(*event));
        unsigned long idx = task->nr_events;
        size_t size;
 
@@ -231,27 +234,27 @@ get_new_event(struct task_desc *task, u64 timestamp)
        event->nr = idx;
 
        task->nr_events++;
-       size = sizeof(struct sched_event *) * task->nr_events;
-       task->events = realloc(task->events, size);
-       BUG_ON(!task->events);
+       size = sizeof(struct sched_atom *) * task->nr_events;
+       task->atoms = realloc(task->atoms, size);
+       BUG_ON(!task->atoms);
 
-       task->events[idx] = event;
+       task->atoms[idx] = event;
 
        return event;
 }
 
-static struct sched_event *last_event(struct task_desc *task)
+static struct sched_atom *last_event(struct task_desc *task)
 {
        if (!task->nr_events)
                return NULL;
 
-       return task->events[task->nr_events - 1];
+       return task->atoms[task->nr_events - 1];
 }
 
 static void
 add_sched_event_run(struct task_desc *task, u64 timestamp, u64 duration)
 {
-       struct sched_event *event, *curr_event = last_event(task);
+       struct sched_atom *event, *curr_event = last_event(task);
 
        /*
         * optimize an existing RUN event by merging this one
@@ -275,7 +278,7 @@ static void
 add_sched_event_wakeup(struct task_desc *task, u64 timestamp,
                       struct task_desc *wakee)
 {
-       struct sched_event *event, *wakee_event;
+       struct sched_atom *event, *wakee_event;
 
        event = get_new_event(task, timestamp);
        event->type = SCHED_EVENT_WAKEUP;
@@ -303,7 +306,7 @@ static void
 add_sched_event_sleep(struct task_desc *task, u64 timestamp,
                      u64 task_state __used)
 {
-       struct sched_event *event = get_new_event(task, timestamp);
+       struct sched_atom *event = get_new_event(task, timestamp);
 
        event->type = SCHED_EVENT_SLEEP;
 
@@ -372,27 +375,27 @@ static void add_cross_task_wakeups(void)
 }
 
 static void
-process_sched_event(struct task_desc *this_task __used, struct sched_event *event)
+process_sched_event(struct task_desc *this_task __used, struct sched_atom *atom)
 {
        int ret = 0;
        u64 now;
        long long delta;
 
        now = get_nsecs();
-       delta = start_time + event->timestamp - now;
+       delta = start_time + atom->timestamp - now;
 
-       switch (event->type) {
+       switch (atom->type) {
                case SCHED_EVENT_RUN:
-                       burn_nsecs(event->duration);
+                       burn_nsecs(atom->duration);
                        break;
                case SCHED_EVENT_SLEEP:
-                       if (event->wait_sem)
-                               ret = sem_wait(event->wait_sem);
+                       if (atom->wait_sem)
+                               ret = sem_wait(atom->wait_sem);
                        BUG_ON(ret);
                        break;
                case SCHED_EVENT_WAKEUP:
-                       if (event->wait_sem)
-                               ret = sem_post(event->wait_sem);
+                       if (atom->wait_sem)
+                               ret = sem_post(atom->wait_sem);
                        BUG_ON(ret);
                        break;
                default:
@@ -467,7 +470,7 @@ again:
 
        for (i = 0; i < this_task->nr_events; i++) {
                this_task->curr_event = i;
-               process_sched_event(this_task, this_task->events[i]);
+               process_sched_event(this_task, this_task->atoms[i]);
        }
 
        cpu_usage_1 = get_cpu_usage_nsec_self();
@@ -629,38 +632,6 @@ static void test_calibrations(void)
        printf("the sleep test took %Ld nsecs\n", T1-T0);
 }
 
-static void __cmd_replay(void)
-{
-       unsigned long i;
-
-       calibrate_run_measurement_overhead();
-       calibrate_sleep_measurement_overhead();
-
-       test_calibrations();
-
-       read_events();
-
-       printf("nr_run_events:        %ld\n", nr_run_events);
-       printf("nr_sleep_events:      %ld\n", nr_sleep_events);
-       printf("nr_wakeup_events:     %ld\n", nr_wakeup_events);
-
-       if (targetless_wakeups)
-               printf("target-less wakeups:  %ld\n", targetless_wakeups);
-       if (multitarget_wakeups)
-               printf("multi-target wakeups: %ld\n", multitarget_wakeups);
-       if (nr_run_events_optimized)
-               printf("run events optimized: %ld\n",
-                       nr_run_events_optimized);
-
-       print_task_traces();
-       add_cross_task_wakeups();
-
-       create_tasks();
-       printf("------------------------------------------------------------\n");
-       for (i = 0; i < replay_repeat; i++)
-               run_one_test();
-}
-
 static int
 process_comm_event(event_t *event, unsigned long offset, unsigned long head)
 {
@@ -668,14 +639,14 @@ process_comm_event(event_t *event, unsigned long offset, unsigned long head)
 
        thread = threads__findnew(event->comm.pid, &threads, &last_match);
 
-       dump_printf("%p [%p]: PERF_EVENT_COMM: %s:%d\n",
+       dump_printf("%p [%p]: perf_event_comm: %s:%d\n",
                (void *)(offset + head),
                (void *)(long)(event->header.size),
                event->comm.comm, event->comm.pid);
 
        if (thread == NULL ||
            thread__set_comm(thread, event->comm.comm)) {
-               dump_printf("problem processing PERF_EVENT_COMM, skipping event.\n");
+               dump_printf("problem processing perf_event_comm, skipping event.\n");
                return -1;
        }
        total_comm++;
@@ -727,6 +698,20 @@ struct trace_switch_event {
        u32 next_prio;
 };
 
+struct trace_runtime_event {
+       u32 size;
+
+       u16 common_type;
+       u8 common_flags;
+       u8 common_preempt_count;
+       u32 common_pid;
+       u32 common_tgid;
+
+       char comm[16];
+       u32 pid;
+       u64 runtime;
+       u64 vruntime;
+};
 
 struct trace_wakeup_event {
        u32 size;
@@ -767,6 +752,12 @@ struct trace_sched_handler {
                             u64 timestamp,
                             struct thread *thread);
 
+       void (*runtime_event)(struct trace_runtime_event *,
+                             struct event *,
+                             int cpu,
+                             u64 timestamp,
+                             struct thread *thread);
+
        void (*wakeup_event)(struct trace_wakeup_event *,
                             struct event *,
                             int cpu,
@@ -881,7 +872,7 @@ struct sort_dimension {
 static LIST_HEAD(cmp_pid);
 
 static int
-thread_lat_cmp(struct list_head *list, struct task_atoms *l, struct task_atoms *r)
+thread_lat_cmp(struct list_head *list, struct work_atoms *l, struct work_atoms *r)
 {
        struct sort_dimension *sort;
        int ret = 0;
@@ -897,18 +888,18 @@ thread_lat_cmp(struct list_head *list, struct task_atoms *l, struct task_atoms *
        return ret;
 }
 
-static struct task_atoms *
+static struct work_atoms *
 thread_atoms_search(struct rb_root *root, struct thread *thread,
                         struct list_head *sort_list)
 {
        struct rb_node *node = root->rb_node;
-       struct task_atoms key = { .thread = thread };
+       struct work_atoms key = { .thread = thread };
 
        while (node) {
-               struct task_atoms *atoms;
+               struct work_atoms *atoms;
                int cmp;
 
-               atoms = container_of(node, struct task_atoms, node);
+               atoms = container_of(node, struct work_atoms, node);
 
                cmp = thread_lat_cmp(sort_list, &key, atoms);
                if (cmp > 0)
@@ -924,16 +915,16 @@ thread_atoms_search(struct rb_root *root, struct thread *thread,
 }
 
 static void
-__thread_latency_insert(struct rb_root *root, struct task_atoms *data,
+__thread_latency_insert(struct rb_root *root, struct work_atoms *data,
                         struct list_head *sort_list)
 {
        struct rb_node **new = &(root->rb_node), *parent = NULL;
 
        while (*new) {
-               struct task_atoms *this;
+               struct work_atoms *this;
                int cmp;
 
-               this = container_of(*new, struct task_atoms, node);
+               this = container_of(*new, struct work_atoms, node);
                parent = *new;
 
                cmp = thread_lat_cmp(sort_list, data, this);
@@ -950,14 +941,14 @@ __thread_latency_insert(struct rb_root *root, struct task_atoms *data,
 
 static void thread_atoms_insert(struct thread *thread)
 {
-       struct task_atoms *atoms;
+       struct work_atoms *atoms;
 
        atoms = calloc(sizeof(*atoms), 1);
        if (!atoms)
                die("No memory");
 
        atoms->thread = thread;
-       INIT_LIST_HEAD(&atoms->atom_list);
+       INIT_LIST_HEAD(&atoms->work_list);
        __thread_latency_insert(&atom_root, atoms, &cmp_pid);
 }
 
@@ -980,10 +971,9 @@ static char sched_out_state(struct trace_switch_event *switch_event)
 }
 
 static void
-lat_sched_out(struct task_atoms *atoms,
-             struct trace_switch_event *switch_event __used,
-             u64 delta,
-             u64 timestamp)
+add_sched_out_event(struct work_atoms *atoms,
+                   char run_state,
+                   u64 timestamp)
 {
        struct work_atom *atom;
 
@@ -993,25 +983,37 @@ lat_sched_out(struct task_atoms *atoms,
 
        atom->sched_out_time = timestamp;
 
-       if (sched_out_state(switch_event) == 'R') {
+       if (run_state == 'R') {
                atom->state = THREAD_WAIT_CPU;
                atom->wake_up_time = atom->sched_out_time;
        }
 
-       atom->runtime = delta;
-       list_add_tail(&atom->list, &atoms->atom_list);
+       list_add_tail(&atom->list, &atoms->work_list);
+}
+
+static void
+add_runtime_event(struct work_atoms *atoms, u64 delta, u64 timestamp __used)
+{
+       struct work_atom *atom;
+
+       BUG_ON(list_empty(&atoms->work_list));
+
+       atom = list_entry(atoms->work_list.prev, struct work_atom, list);
+
+       atom->runtime += delta;
+       atoms->total_runtime += delta;
 }
 
 static void
-lat_sched_in(struct task_atoms *atoms, u64 timestamp)
+add_sched_in_event(struct work_atoms *atoms, u64 timestamp)
 {
        struct work_atom *atom;
        u64 delta;
 
-       if (list_empty(&atoms->atom_list))
+       if (list_empty(&atoms->work_list))
                return;
 
-       atom = list_entry(atoms->atom_list.prev, struct work_atom, list);
+       atom = list_entry(atoms->work_list.prev, struct work_atom, list);
 
        if (atom->state != THREAD_WAIT_CPU)
                return;
@@ -1029,7 +1031,6 @@ lat_sched_in(struct task_atoms *atoms, u64 timestamp)
        if (delta > atoms->max_lat)
                atoms->max_lat = delta;
        atoms->nb_atoms++;
-       atoms->total_runtime += atom->runtime;
 }
 
 static void
@@ -1039,13 +1040,12 @@ latency_switch_event(struct trace_switch_event *switch_event,
                     u64 timestamp,
                     struct thread *thread __used)
 {
-       struct task_atoms *out_atoms, *in_atoms;
+       struct work_atoms *out_events, *in_events;
        struct thread *sched_out, *sched_in;
        u64 timestamp0;
        s64 delta;
 
-       if (cpu >= MAX_CPUS || cpu < 0)
-               return;
+       BUG_ON(cpu >= MAX_CPUS || cpu < 0);
 
        timestamp0 = cpu_last_switched[cpu];
        cpu_last_switched[cpu] = timestamp;
@@ -1061,34 +1061,63 @@ latency_switch_event(struct trace_switch_event *switch_event,
        sched_out = threads__findnew(switch_event->prev_pid, &threads, &last_match);
        sched_in = threads__findnew(switch_event->next_pid, &threads, &last_match);
 
-       in_atoms = thread_atoms_search(&atom_root, sched_in, &cmp_pid);
-       if (!in_atoms) {
+       out_events = thread_atoms_search(&atom_root, sched_out, &cmp_pid);
+       if (!out_events) {
+               thread_atoms_insert(sched_out);
+               out_events = thread_atoms_search(&atom_root, sched_out, &cmp_pid);
+               if (!out_events)
+                       die("out-event: Internal tree error");
+       }
+       add_sched_out_event(out_events, sched_out_state(switch_event), timestamp);
+
+       in_events = thread_atoms_search(&atom_root, sched_in, &cmp_pid);
+       if (!in_events) {
                thread_atoms_insert(sched_in);
-               in_atoms = thread_atoms_search(&atom_root, sched_in, &cmp_pid);
-               if (!in_atoms)
-                       die("in-atom: Internal tree error");
+               in_events = thread_atoms_search(&atom_root, sched_in, &cmp_pid);
+               if (!in_events)
+                       die("in-event: Internal tree error");
+               /*
+                * Take came in we have not heard about yet,
+                * add in an initial atom in runnable state:
+                */
+               add_sched_out_event(in_events, 'R', timestamp);
        }
+       add_sched_in_event(in_events, timestamp);
+}
 
-       out_atoms = thread_atoms_search(&atom_root, sched_out, &cmp_pid);
-       if (!out_atoms) {
-               thread_atoms_insert(sched_out);
-               out_atoms = thread_atoms_search(&atom_root, sched_out, &cmp_pid);
-               if (!out_atoms)
-                       die("out-atom: Internal tree error");
+static void
+latency_runtime_event(struct trace_runtime_event *runtime_event,
+                    struct event *event __used,
+                    int cpu,
+                    u64 timestamp,
+                    struct thread *this_thread __used)
+{
+       struct work_atoms *atoms;
+       struct thread *thread;
+
+       BUG_ON(cpu >= MAX_CPUS || cpu < 0);
+
+       thread = threads__findnew(runtime_event->pid, &threads, &last_match);
+       atoms = thread_atoms_search(&atom_root, thread, &cmp_pid);
+       if (!atoms) {
+               thread_atoms_insert(thread);
+               atoms = thread_atoms_search(&atom_root, thread, &cmp_pid);
+               if (!atoms)
+                       die("in-event: Internal tree error");
+               add_sched_out_event(atoms, 'R', timestamp);
        }
 
-       lat_sched_in(in_atoms, timestamp);
-       lat_sched_out(out_atoms, switch_event, delta, timestamp);
+       add_runtime_event(atoms, runtime_event->runtime, timestamp);
 }
 
 static void
 latency_wakeup_event(struct trace_wakeup_event *wakeup_event,
-                    struct event *event __used,
+                    struct event *__event __used,
                     int cpu __used,
                     u64 timestamp,
                     struct thread *thread __used)
 {
-       struct task_atoms *atoms;
+       struct work_atoms *atoms;
        struct work_atom *atom;
        struct thread *wakee;
 
@@ -1100,20 +1129,22 @@ latency_wakeup_event(struct trace_wakeup_event *wakeup_event,
        atoms = thread_atoms_search(&atom_root, wakee, &cmp_pid);
        if (!atoms) {
                thread_atoms_insert(wakee);
-               return;
+               atoms = thread_atoms_search(&atom_root, wakee, &cmp_pid);
+               if (!atoms)
+                       die("wakeup-event: Internal tree error");
+               add_sched_out_event(atoms, 'S', timestamp);
        }
 
-       if (list_empty(&atoms->atom_list))
-               return;
+       BUG_ON(list_empty(&atoms->work_list));
 
-       atom = list_entry(atoms->atom_list.prev, struct work_atom, list);
+       atom = list_entry(atoms->work_list.prev, struct work_atom, list);
 
        if (atom->state != THREAD_SLEEPING)
-               return;
+               nr_state_machine_bugs++;
 
        nr_timestamps++;
        if (atom->sched_out_time > timestamp) {
-               unordered_timestamps++;
+               nr_unordered_timestamps++;
                return;
        }
 
@@ -1124,40 +1155,41 @@ latency_wakeup_event(struct trace_wakeup_event *wakeup_event,
 static struct trace_sched_handler lat_ops  = {
        .wakeup_event           = latency_wakeup_event,
        .switch_event           = latency_switch_event,
+       .runtime_event          = latency_runtime_event,
        .fork_event             = latency_fork_event,
 };
 
-static void output_lat_thread(struct task_atoms *atom_list)
+static void output_lat_thread(struct work_atoms *work_list)
 {
        int i;
        int ret;
        u64 avg;
 
-       if (!atom_list->nb_atoms)
+       if (!work_list->nb_atoms)
                return;
        /*
         * Ignore idle threads:
         */
-       if (!atom_list->thread->pid)
+       if (!strcmp(work_list->thread->comm, "swapper"))
                return;
 
-       all_runtime += atom_list->total_runtime;
-       all_count += atom_list->nb_atoms;
+       all_runtime += work_list->total_runtime;
+       all_count += work_list->nb_atoms;
 
-       ret = printf(" %s ", atom_list->thread->comm);
+       ret = printf("  %s:%d ", work_list->thread->comm, work_list->thread->pid);
 
-       for (i = 0; i < 19 - ret; i++)
+       for (i = 0; i < 24 - ret; i++)
                printf(" ");
 
-       avg = atom_list->total_lat / atom_list->nb_atoms;
+       avg = work_list->total_lat / work_list->nb_atoms;
 
-       printf("|%9.3f ms |%9llu | avg:%9.3f ms | max:%9.3f ms |\n",
-             (double)atom_list->total_runtime / 1e6,
-                atom_list->nb_atoms, (double)avg / 1e6,
-                (double)atom_list->max_lat / 1e6);
+       printf("|%11.3f ms |%9llu | avg:%9.3f ms | max:%9.3f ms |\n",
+             (double)work_list->total_runtime / 1e6,
+                work_list->nb_atoms, (double)avg / 1e6,
+                (double)work_list->max_lat / 1e6);
 }
 
-static int pid_cmp(struct task_atoms *l, struct task_atoms *r)
+static int pid_cmp(struct work_atoms *l, struct work_atoms *r)
 {
        if (l->thread->pid < r->thread->pid)
                return -1;
@@ -1172,7 +1204,7 @@ static struct sort_dimension pid_sort_dimension = {
        .cmp                    = pid_cmp,
 };
 
-static int avg_cmp(struct task_atoms *l, struct task_atoms *r)
+static int avg_cmp(struct work_atoms *l, struct work_atoms *r)
 {
        u64 avgl, avgr;
 
@@ -1198,7 +1230,7 @@ static struct sort_dimension avg_sort_dimension = {
        .cmp                    = avg_cmp,
 };
 
-static int max_cmp(struct task_atoms *l, struct task_atoms *r)
+static int max_cmp(struct work_atoms *l, struct work_atoms *r)
 {
        if (l->max_lat < r->max_lat)
                return -1;
@@ -1213,7 +1245,7 @@ static struct sort_dimension max_sort_dimension = {
        .cmp                    = max_cmp,
 };
 
-static int switch_cmp(struct task_atoms *l, struct task_atoms *r)
+static int switch_cmp(struct work_atoms *l, struct work_atoms *r)
 {
        if (l->nb_atoms < r->nb_atoms)
                return -1;
@@ -1228,7 +1260,7 @@ static struct sort_dimension switch_sort_dimension = {
        .cmp                    = switch_cmp,
 };
 
-static int runtime_cmp(struct task_atoms *l, struct task_atoms *r)
+static int runtime_cmp(struct work_atoms *l, struct work_atoms *r)
 {
        if (l->total_runtime < r->total_runtime)
                return -1;
@@ -1255,7 +1287,7 @@ static struct sort_dimension *available_sorts[] = {
 
 static LIST_HEAD(sort_list);
 
-static int sort_dimension__add(char *tok, struct list_head *list)
+static int sort_dimension__add(const char *tok, struct list_head *list)
 {
        int i;
 
@@ -1277,53 +1309,17 @@ static void sort_lat(void)
        struct rb_node *node;
 
        for (;;) {
-               struct task_atoms *data;
+               struct work_atoms *data;
                node = rb_first(&atom_root);
                if (!node)
                        break;
 
                rb_erase(node, &atom_root);
-               data = rb_entry(node, struct task_atoms, node);
+               data = rb_entry(node, struct work_atoms, node);
                __thread_latency_insert(&sorted_atom_root, data, &sort_list);
        }
 }
 
-static void __cmd_lat(void)
-{
-       struct rb_node *next;
-
-       setup_pager();
-       read_events();
-       sort_lat();
-
-       printf("-----------------------------------------------------------------------------------\n");
-       printf(" Task              |  Runtime ms | Switches | Average delay ms | Maximum delay ms |\n");
-       printf("-----------------------------------------------------------------------------------\n");
-
-       next = rb_first(&sorted_atom_root);
-
-       while (next) {
-               struct task_atoms *atom_list;
-
-               atom_list = rb_entry(next, struct task_atoms, node);
-               output_lat_thread(atom_list);
-               next = rb_next(next);
-       }
-
-       printf("-----------------------------------------------------------------------------------\n");
-       printf(" TOTAL:            |%9.3f ms |%9Ld |",
-               (double)all_runtime/1e6, all_count);
-
-       if (unordered_timestamps && nr_timestamps) {
-               printf(" INFO: %.2f%% unordered events.\n",
-                       (double)unordered_timestamps/(double)nr_timestamps*100.0);
-       } else {
-               printf("\n");
-       }
-
-       printf("---------------------------------------------\n");
-}
-
 static struct trace_sched_handler *trace_handler;
 
 static void
@@ -1343,13 +1339,106 @@ process_sched_wakeup_event(struct raw_event_sample *raw,
        FILL_FIELD(wakeup_event, success, event, raw->data);
        FILL_FIELD(wakeup_event, cpu, event, raw->data);
 
-       trace_handler->wakeup_event(&wakeup_event, event, cpu, timestamp, thread);
+       if (trace_handler->wakeup_event)
+               trace_handler->wakeup_event(&wakeup_event, event, cpu, timestamp, thread);
+}
+
+/*
+ * Track the current task - that way we can know whether there's any
+ * weird events, such as a task being switched away that is not current.
+ */
+static int max_cpu;
+
+static u32 curr_pid[MAX_CPUS] = { [0 ... MAX_CPUS-1] = -1 };
+
+static struct thread *curr_thread[MAX_CPUS];
+
+static char next_shortname1 = 'A';
+static char next_shortname2 = '0';
+
+static void
+map_switch_event(struct trace_switch_event *switch_event,
+                struct event *event __used,
+                int this_cpu,
+                u64 timestamp,
+                struct thread *thread __used)
+{
+       struct thread *sched_out, *sched_in;
+       int new_shortname;
+       u64 timestamp0;
+       s64 delta;
+       int cpu;
+
+       BUG_ON(this_cpu >= MAX_CPUS || this_cpu < 0);
+
+       if (this_cpu > max_cpu)
+               max_cpu = this_cpu;
+
+       timestamp0 = cpu_last_switched[this_cpu];
+       cpu_last_switched[this_cpu] = timestamp;
+       if (timestamp0)
+               delta = timestamp - timestamp0;
+       else
+               delta = 0;
+
+       if (delta < 0)
+               die("hm, delta: %Ld < 0 ?\n", delta);
+
+
+       sched_out = threads__findnew(switch_event->prev_pid, &threads, &last_match);
+       sched_in = threads__findnew(switch_event->next_pid, &threads, &last_match);
+
+       curr_thread[this_cpu] = sched_in;
+
+       printf("  ");
+
+       new_shortname = 0;
+       if (!sched_in->shortname[0]) {
+               sched_in->shortname[0] = next_shortname1;
+               sched_in->shortname[1] = next_shortname2;
+
+               if (next_shortname1 < 'Z') {
+                       next_shortname1++;
+               } else {
+                       next_shortname1='A';
+                       if (next_shortname2 < '9') {
+                               next_shortname2++;
+                       } else {
+                               next_shortname2='0';
+                       }
+               }
+               new_shortname = 1;
+       }
+
+       for (cpu = 0; cpu <= max_cpu; cpu++) {
+               if (cpu != this_cpu)
+                       printf(" ");
+               else
+                       printf("*");
+
+               if (curr_thread[cpu]) {
+                       if (curr_thread[cpu]->pid)
+                               printf("%2s ", curr_thread[cpu]->shortname);
+                       else
+                               printf(".  ");
+               } else
+                       printf("   ");
+       }
+
+       printf("  %12.6f secs ", (double)timestamp/1e9);
+       if (new_shortname) {
+               printf("%s => %s:%d\n",
+                       sched_in->shortname, sched_in->comm, sched_in->pid);
+       } else {
+               printf("\n");
+       }
 }
 
+
 static void
 process_sched_switch_event(struct raw_event_sample *raw,
                           struct event *event,
-                          int cpu __used,
+                          int this_cpu,
                           u64 timestamp __used,
                           struct thread *thread __used)
 {
@@ -1365,7 +1454,36 @@ process_sched_switch_event(struct raw_event_sample *raw,
        FILL_FIELD(switch_event, next_pid, event, raw->data);
        FILL_FIELD(switch_event, next_prio, event, raw->data);
 
-       trace_handler->switch_event(&switch_event, event, cpu, timestamp, thread);
+       if (curr_pid[this_cpu] != (u32)-1) {
+               /*
+                * Are we trying to switch away a PID that is
+                * not current?
+                */
+               if (curr_pid[this_cpu] != switch_event.prev_pid)
+                       nr_context_switch_bugs++;
+       }
+       if (trace_handler->switch_event)
+               trace_handler->switch_event(&switch_event, event, this_cpu, timestamp, thread);
+
+       curr_pid[this_cpu] = switch_event.next_pid;
+}
+
+static void
+process_sched_runtime_event(struct raw_event_sample *raw,
+                          struct event *event,
+                          int cpu __used,
+                          u64 timestamp __used,
+                          struct thread *thread __used)
+{
+       struct trace_runtime_event runtime_event;
+
+       FILL_ARRAY(runtime_event, comm, event, raw->data);
+       FILL_FIELD(runtime_event, pid, event, raw->data);
+       FILL_FIELD(runtime_event, runtime, event, raw->data);
+       FILL_FIELD(runtime_event, vruntime, event, raw->data);
+
+       if (trace_handler->runtime_event)
+               trace_handler->runtime_event(&runtime_event, event, cpu, timestamp, thread);
 }
 
 static void
@@ -1384,7 +1502,8 @@ process_sched_fork_event(struct raw_event_sample *raw,
        FILL_ARRAY(fork_event, child_comm, event, raw->data);
        FILL_FIELD(fork_event, child_pid, event, raw->data);
 
-       trace_handler->fork_event(&fork_event, event, cpu, timestamp, thread);
+       if (trace_handler->fork_event)
+               trace_handler->fork_event(&fork_event, event, cpu, timestamp, thread);
 }
 
 static void
@@ -1410,6 +1529,8 @@ process_raw_event(event_t *raw_event __used, void *more_data,
 
        if (!strcmp(event->name, "sched_switch"))
                process_sched_switch_event(raw, event, cpu, timestamp, thread);
+       if (!strcmp(event->name, "sched_stat_runtime"))
+               process_sched_runtime_event(raw, event, cpu, timestamp, thread);
        if (!strcmp(event->name, "sched_wakeup"))
                process_sched_wakeup_event(raw, event, cpu, timestamp, thread);
        if (!strcmp(event->name, "sched_wakeup_new"))
@@ -1452,7 +1573,7 @@ process_sample_event(event_t *event, unsigned long offset, unsigned long head)
                more_data += sizeof(u64);
        }
 
-       dump_printf("%p [%p]: PERF_EVENT_SAMPLE (IP, %d): %d/%d: %p period: %Ld\n",
+       dump_printf("%p [%p]: PERF_RECORD_SAMPLE (IP, %d): %d/%d: %p period: %Ld\n",
                (void *)(offset + head),
                (void *)(long)(event->header.size),
                event->header.misc,
@@ -1468,9 +1589,9 @@ process_sample_event(event_t *event, unsigned long offset, unsigned long head)
                return -1;
        }
 
-       cpumode = event->header.misc & PERF_EVENT_MISC_CPUMODE_MASK;
+       cpumode = event->header.misc & PERF_RECORD_MISC_CPUMODE_MASK;
 
-       if (cpumode == PERF_EVENT_MISC_KERNEL) {
+       if (cpumode == PERF_RECORD_MISC_KERNEL) {
                show = SHOW_KERNEL;
                level = 'k';
 
@@ -1478,7 +1599,7 @@ process_sample_event(event_t *event, unsigned long offset, unsigned long head)
 
                dump_printf(" ...... dso: %s\n", dso->name);
 
-       } else if (cpumode == PERF_EVENT_MISC_USER) {
+       } else if (cpumode == PERF_RECORD_MISC_USER) {
 
                show = SHOW_USER;
                level = '.';
@@ -1503,20 +1624,25 @@ process_event(event_t *event, unsigned long offset, unsigned long head)
 {
        trace_event(event);
 
+       nr_events++;
        switch (event->header.type) {
-       case PERF_EVENT_MMAP ... PERF_EVENT_LOST:
+       case PERF_RECORD_MMAP:
+               return 0;
+       case PERF_RECORD_LOST:
+               nr_lost_chunks++;
+               nr_lost_events += event->lost.lost;
                return 0;
 
-       case PERF_EVENT_COMM:
+       case PERF_RECORD_COMM:
                return process_comm_event(event, offset, head);
 
-       case PERF_EVENT_EXIT ... PERF_EVENT_READ:
+       case PERF_RECORD_EXIT ... PERF_RECORD_READ:
                return 0;
 
-       case PERF_EVENT_SAMPLE:
+       case PERF_RECORD_SAMPLE:
                return process_sample_event(event, offset, head);
 
-       case PERF_EVENT_MAX:
+       case PERF_RECORD_MAX:
        default:
                return -1;
        }
@@ -1620,12 +1746,126 @@ more:
        return rc;
 }
 
+static void print_bad_events(void)
+{
+       if (nr_unordered_timestamps && nr_timestamps) {
+               printf("  INFO: %.3f%% unordered timestamps (%ld out of %ld)\n",
+                       (double)nr_unordered_timestamps/(double)nr_timestamps*100.0,
+                       nr_unordered_timestamps, nr_timestamps);
+       }
+       if (nr_lost_events && nr_events) {
+               printf("  INFO: %.3f%% lost events (%ld out of %ld, in %ld chunks)\n",
+                       (double)nr_lost_events/(double)nr_events*100.0,
+                       nr_lost_events, nr_events, nr_lost_chunks);
+       }
+       if (nr_state_machine_bugs && nr_timestamps) {
+               printf("  INFO: %.3f%% state machine bugs (%ld out of %ld)",
+                       (double)nr_state_machine_bugs/(double)nr_timestamps*100.0,
+                       nr_state_machine_bugs, nr_timestamps);
+               if (nr_lost_events)
+                       printf(" (due to lost events?)");
+               printf("\n");
+       }
+       if (nr_context_switch_bugs && nr_timestamps) {
+               printf("  INFO: %.3f%% context switch bugs (%ld out of %ld)",
+                       (double)nr_context_switch_bugs/(double)nr_timestamps*100.0,
+                       nr_context_switch_bugs, nr_timestamps);
+               if (nr_lost_events)
+                       printf(" (due to lost events?)");
+               printf("\n");
+       }
+}
+
+static void __cmd_lat(void)
+{
+       struct rb_node *next;
+
+       setup_pager();
+       read_events();
+       sort_lat();
+
+       printf("\n -----------------------------------------------------------------------------------------\n");
+       printf("  Task                  |   Runtime ms  | Switches | Average delay ms | Maximum delay ms |\n");
+       printf(" -----------------------------------------------------------------------------------------\n");
+
+       next = rb_first(&sorted_atom_root);
+
+       while (next) {
+               struct work_atoms *work_list;
+
+               work_list = rb_entry(next, struct work_atoms, node);
+               output_lat_thread(work_list);
+               next = rb_next(next);
+       }
+
+       printf(" -----------------------------------------------------------------------------------------\n");
+       printf("  TOTAL:                |%11.3f ms |%9Ld |\n",
+               (double)all_runtime/1e6, all_count);
+
+       printf(" ---------------------------------------------------\n");
+
+       print_bad_events();
+       printf("\n");
+
+}
+
+static struct trace_sched_handler map_ops  = {
+       .wakeup_event           = NULL,
+       .switch_event           = map_switch_event,
+       .runtime_event          = NULL,
+       .fork_event             = NULL,
+};
+
+static void __cmd_map(void)
+{
+       max_cpu = sysconf(_SC_NPROCESSORS_CONF);
+
+       setup_pager();
+       read_events();
+       print_bad_events();
+}
+
+static void __cmd_replay(void)
+{
+       unsigned long i;
+
+       calibrate_run_measurement_overhead();
+       calibrate_sleep_measurement_overhead();
+
+       test_calibrations();
+
+       read_events();
+
+       printf("nr_run_events:        %ld\n", nr_run_events);
+       printf("nr_sleep_events:      %ld\n", nr_sleep_events);
+       printf("nr_wakeup_events:     %ld\n", nr_wakeup_events);
+
+       if (targetless_wakeups)
+               printf("target-less wakeups:  %ld\n", targetless_wakeups);
+       if (multitarget_wakeups)
+               printf("multi-target wakeups: %ld\n", multitarget_wakeups);
+       if (nr_run_events_optimized)
+               printf("run atoms optimized: %ld\n",
+                       nr_run_events_optimized);
+
+       print_task_traces();
+       add_cross_task_wakeups();
+
+       create_tasks();
+       printf("------------------------------------------------------------\n");
+       for (i = 0; i < replay_repeat; i++)
+               run_one_test();
+}
+
+
 static const char * const sched_usage[] = {
-       "perf sched [<options>] {record|latency|replay|trace}",
+       "perf sched [<options>] {record|latency|map|replay|trace}",
        NULL
 };
 
 static const struct option sched_options[] = {
+       OPT_STRING('i', "input", &input_name, "file",
+                   "input file name"),
        OPT_BOOLEAN('v', "verbose", &verbose,
                    "be more verbose (show symbol address, etc)"),
        OPT_BOOLEAN('D', "dump-raw-trace", &dump_trace,
@@ -1677,7 +1917,7 @@ static void setup_sorting(void)
 
        free(str);
 
-       sort_dimension__add((char *)"pid", &cmp_pid);
+       sort_dimension__add("pid", &cmp_pid);
 }
 
 static const char *record_args[] = {
@@ -1686,6 +1926,7 @@ static const char *record_args[] = {
        "-R",
        "-M",
        "-f",
+       "-m", "1024",
        "-c", "1",
        "-e", "sched:sched_switch:r",
        "-e", "sched:sched_stat_wait:r",
@@ -1738,6 +1979,10 @@ int cmd_sched(int argc, const char **argv, const char *prefix __used)
                }
                setup_sorting();
                __cmd_lat();
+       } else if (!strcmp(argv[0], "map")) {
+               trace_handler = &map_ops;
+               setup_sorting();
+               __cmd_map();
        } else if (!strncmp(argv[0], "rep", 3)) {
                trace_handler = &replay_ops;
                if (argc) {