perf_counter: Remove munmap stuff
[safe/jmp/linux-2.6] / kernel / perf_counter.c
index 78c5862..195712e 100644 (file)
@@ -41,7 +41,6 @@ static int perf_overcommit __read_mostly = 1;
 
 static atomic_t nr_counters __read_mostly;
 static atomic_t nr_mmap_counters __read_mostly;
-static atomic_t nr_munmap_counters __read_mostly;
 static atomic_t nr_comm_counters __read_mostly;
 
 int sysctl_perf_counter_priv __read_mostly; /* do we need to be privileged */
@@ -1448,8 +1447,6 @@ static void free_counter(struct perf_counter *counter)
        atomic_dec(&nr_counters);
        if (counter->attr.mmap)
                atomic_dec(&nr_mmap_counters);
-       if (counter->attr.munmap)
-               atomic_dec(&nr_munmap_counters);
        if (counter->attr.comm)
                atomic_dec(&nr_comm_counters);
 
@@ -2510,7 +2507,7 @@ static void perf_counter_fork_output(struct perf_counter *counter,
 
 static int perf_counter_fork_match(struct perf_counter *counter)
 {
-       if (counter->attr.comm || counter->attr.mmap || counter->attr.munmap)
+       if (counter->attr.comm || counter->attr.mmap)
                return 1;
 
        return 0;
@@ -2557,8 +2554,7 @@ void perf_counter_fork(struct task_struct *task)
        struct perf_fork_event fork_event;
 
        if (!atomic_read(&nr_comm_counters) &&
-           !atomic_read(&nr_mmap_counters) &&
-           !atomic_read(&nr_munmap_counters))
+           !atomic_read(&nr_mmap_counters))
                return;
 
        fork_event = (struct perf_fork_event){
@@ -2722,12 +2718,7 @@ static void perf_counter_mmap_output(struct perf_counter *counter,
 static int perf_counter_mmap_match(struct perf_counter *counter,
                                   struct perf_mmap_event *mmap_event)
 {
-       if (counter->attr.mmap &&
-           mmap_event->event.header.type == PERF_EVENT_MMAP)
-               return 1;
-
-       if (counter->attr.munmap &&
-           mmap_event->event.header.type == PERF_EVENT_MUNMAP)
+       if (counter->attr.mmap)
                return 1;
 
        return 0;
@@ -2821,27 +2812,6 @@ void perf_counter_mmap(unsigned long addr, unsigned long len,
        perf_counter_mmap_event(&mmap_event);
 }
 
-void perf_counter_munmap(unsigned long addr, unsigned long len,
-                        unsigned long pgoff, struct file *file)
-{
-       struct perf_mmap_event mmap_event;
-
-       if (!atomic_read(&nr_munmap_counters))
-               return;
-
-       mmap_event = (struct perf_mmap_event){
-               .file   = file,
-               .event  = {
-                       .header = { .type = PERF_EVENT_MUNMAP, },
-                       .start  = addr,
-                       .len    = len,
-                       .pgoff  = pgoff,
-               },
-       };
-
-       perf_counter_mmap_event(&mmap_event);
-}
-
 /*
  * Log sample_period changes so that analyzing tools can re-normalize the
  * event flow.
@@ -3525,8 +3495,6 @@ done:
        atomic_inc(&nr_counters);
        if (counter->attr.mmap)
                atomic_inc(&nr_mmap_counters);
-       if (counter->attr.munmap)
-               atomic_inc(&nr_munmap_counters);
        if (counter->attr.comm)
                atomic_inc(&nr_comm_counters);