📄 timer_stats.c
字号:
/* * kernel/time/timer_stats.c * * Collect timer usage statistics. * * Copyright(C) 2006, Red Hat, Inc., Ingo Molnar * Copyright(C) 2006 Timesys Corp., Thomas Gleixner <tglx@timesys.com> * * timer_stats is based on timer_top, a similar functionality which was part of * Con Kolivas dyntick patch set. It was developed by Daniel Petrini at the * Instituto Nokia de Tecnologia - INdT - Manaus. timer_top's design was based * on dynamic allocation of the statistics entries and linear search based * lookup combined with a global lock, rather than the static array, hash * and per-CPU locking which is used by timer_stats. It was written for the * pre hrtimer kernel code and therefore did not take hrtimers into account. * Nevertheless it provided the base for the timer_stats implementation and * was a helpful source of inspiration. Kudos to Daniel and the Nokia folks * for this effort. * * timer_top.c is * Copyright (C) 2005 Instituto Nokia de Tecnologia - INdT - Manaus * Written by Daniel Petrini <d.pensator@gmail.com> * timer_top.c was released under the GNU General Public License version 2 * * We export the addresses and counting of timer functions being called, * the pid and cmdline from the owner process if applicable. * * Start/stop data collection: * # echo 1[0] >/proc/timer_stats * * Display the information collected so far: * # cat /proc/timer_stats * * This program is free software; you can redistribute it and/or modify * it under the terms of the GNU General Public License version 2 as * published by the Free Software Foundation. */#include <linux/proc_fs.h>#include <linux/module.h>#include <linux/spinlock.h>#include <linux/sched.h>#include <linux/seq_file.h>#include <linux/kallsyms.h>#include <asm/uaccess.h>/* * This is our basic unit of interest: a timer expiry event identified * by the timer, its start/expire functions and the PID of the task that * started the timer. We count the number of times an event happens: */struct entry { /* * Hash list: */ struct entry *next; /* * Hash keys: */ void *timer; void *start_func; void *expire_func; pid_t pid; /* * Number of timeout events: */ unsigned long count; /* * We save the command-line string to preserve * this information past task exit: */ char comm[TASK_COMM_LEN + 1];} ____cacheline_aligned_in_smp;/* * Spinlock protecting the tables - not taken during lookup: */static DEFINE_SPINLOCK(table_lock);/* * Per-CPU lookup locks for fast hash lookup: */static DEFINE_PER_CPU(spinlock_t, lookup_lock);/* * Mutex to serialize state changes with show-stats activities: */static DEFINE_MUTEX(show_mutex);/* * Collection status, active/inactive: */static int __read_mostly active;/* * Beginning/end timestamps of measurement: */static ktime_t time_start, time_stop;/* * tstat entry structs only get allocated while collection is * active and never freed during that time - this simplifies * things quite a bit. * * They get freed when a new collection period is started. */#define MAX_ENTRIES_BITS 10#define MAX_ENTRIES (1UL << MAX_ENTRIES_BITS)static unsigned long nr_entries;static struct entry entries[MAX_ENTRIES];static atomic_t overflow_count;/* * The entries are in a hash-table, for fast lookup: */#define TSTAT_HASH_BITS (MAX_ENTRIES_BITS - 1)#define TSTAT_HASH_SIZE (1UL << TSTAT_HASH_BITS)#define TSTAT_HASH_MASK (TSTAT_HASH_SIZE - 1)#define __tstat_hashfn(entry) \ (((unsigned long)(entry)->timer ^ \ (unsigned long)(entry)->start_func ^ \ (unsigned long)(entry)->expire_func ^ \ (unsigned long)(entry)->pid ) & TSTAT_HASH_MASK)#define tstat_hashentry(entry) (tstat_hash_table + __tstat_hashfn(entry))static struct entry *tstat_hash_table[TSTAT_HASH_SIZE] __read_mostly;static void reset_entries(void){ nr_entries = 0; memset(entries, 0, sizeof(entries)); memset(tstat_hash_table, 0, sizeof(tstat_hash_table)); atomic_set(&overflow_count, 0);}static struct entry *alloc_entry(void){ if (nr_entries >= MAX_ENTRIES) return NULL; return entries + nr_entries++;}static int match_entries(struct entry *entry1, struct entry *entry2){ return entry1->timer == entry2->timer && entry1->start_func == entry2->start_func && entry1->expire_func == entry2->expire_func && entry1->pid == entry2->pid;}/* * Look up whether an entry matching this item is present * in the hash already. Must be called with irqs off and the * lookup lock held: */static struct entry *tstat_lookup(struct entry *entry, char *comm){ struct entry **head, *curr, *prev; head = tstat_hashentry(entry); curr = *head; /* * The fastpath is when the entry is already hashed, * we do this with the lookup lock held, but with the * table lock not held: */ while (curr) { if (match_entries(curr, entry)) return curr; curr = curr->next; } /* * Slowpath: allocate, set up and link a new hash entry: */ prev = NULL; curr = *head; spin_lock(&table_lock); /* * Make sure we have not raced with another CPU: */ while (curr) { if (match_entries(curr, entry)) goto out_unlock; prev = curr; curr = curr->next; } curr = alloc_entry(); if (curr) { *curr = *entry; curr->count = 0; curr->next = NULL; memcpy(curr->comm, comm, TASK_COMM_LEN); smp_mb(); /* Ensure that curr is initialized before insert */ if (prev) prev->next = curr; else *head = curr; } out_unlock: spin_unlock(&table_lock); return curr;}/** * timer_stats_update_stats - Update the statistics for a timer. * @timer: pointer to either a timer_list or a hrtimer * @pid: the pid of the task which set up the timer * @startf: pointer to the function which did the timer setup * @timerf: pointer to the timer callback function of the timer * @comm: name of the process which set up the timer * * When the timer is already registered, then the event counter is * incremented. Otherwise the timer is registered in a free slot. */void timer_stats_update_stats(void *timer, pid_t pid, void *startf, void *timerf, char * comm){ /* * It doesnt matter which lock we take: */ spinlock_t *lock; struct entry *entry, input; unsigned long flags; if (likely(!active)) return; lock = &per_cpu(lookup_lock, raw_smp_processor_id()); input.timer = timer; input.start_func = startf; input.expire_func = timerf; input.pid = pid; spin_lock_irqsave(lock, flags); if (!active) goto out_unlock; entry = tstat_lookup(&input, comm); if (likely(entry)) entry->count++; else atomic_inc(&overflow_count); out_unlock: spin_unlock_irqrestore(lock, flags);}static void print_name_offset(struct seq_file *m, unsigned long addr){ char symname[KSYM_NAME_LEN+1]; if (lookup_symbol_name(addr, symname) < 0) seq_printf(m, "<%p>", (void *)addr); else seq_printf(m, "%s", symname);}static int tstats_show(struct seq_file *m, void *v){ struct timespec period; struct entry *entry; unsigned long ms; long events = 0; ktime_t time; int i; mutex_lock(&show_mutex); /* * If still active then calculate up to now: */ if (active) time_stop = ktime_get(); time = ktime_sub(time_stop, time_start); period = ktime_to_timespec(time); ms = period.tv_nsec / 1000000; seq_puts(m, "Timer Stats Version: v0.1\n"); seq_printf(m, "Sample period: %ld.%03ld s\n", period.tv_sec, ms); if (atomic_read(&overflow_count)) seq_printf(m, "Overflow: %d entries\n", atomic_read(&overflow_count)); for (i = 0; i < nr_entries; i++) { entry = entries + i; seq_printf(m, "%4lu, %5d %-16s ", entry->count, entry->pid, entry->comm); print_name_offset(m, (unsigned long)entry->start_func); seq_puts(m, " ("); print_name_offset(m, (unsigned long)entry->expire_func); seq_puts(m, ")\n"); events += entry->count; } ms += period.tv_sec * 1000; if (!ms) ms = 1; if (events && period.tv_sec) seq_printf(m, "%ld total events, %ld.%ld events/sec\n", events, events / period.tv_sec, events * 1000 / ms); else seq_printf(m, "%ld total events\n", events); mutex_unlock(&show_mutex); return 0;}/* * After a state change, make sure all concurrent lookup/update * activities have stopped: */static void sync_access(void){ unsigned long flags; int cpu; for_each_online_cpu(cpu) { spin_lock_irqsave(&per_cpu(lookup_lock, cpu), flags); /* nothing */ spin_unlock_irqrestore(&per_cpu(lookup_lock, cpu), flags); }}static ssize_t tstats_write(struct file *file, const char __user *buf, size_t count, loff_t *offs){ char ctl[2]; if (count != 2 || *offs) return -EINVAL; if (copy_from_user(ctl, buf, count)) return -EFAULT; mutex_lock(&show_mutex); switch (ctl[0]) { case '0': if (active) { active = 0; time_stop = ktime_get(); sync_access(); } break; case '1': if (!active) { reset_entries(); time_start = ktime_get(); smp_mb(); active = 1; } break; default: count = -EINVAL; } mutex_unlock(&show_mutex); return count;}static int tstats_open(struct inode *inode, struct file *filp){ return single_open(filp, tstats_show, NULL);}static struct file_operations tstats_fops = { .open = tstats_open, .read = seq_read, .write = tstats_write, .llseek = seq_lseek, .release = seq_release,};void __init init_timer_stats(void){ int cpu; for_each_possible_cpu(cpu) spin_lock_init(&per_cpu(lookup_lock, cpu));}static int __init init_tstats_procfs(void){ struct proc_dir_entry *pe; pe = create_proc_entry("timer_stats", 0644, NULL); if (!pe) return -ENOMEM; pe->proc_fops = &tstats_fops; return 0;}__initcall(init_tstats_procfs);
⌨️ 快捷键说明
复制代码
Ctrl + C
搜索代码
Ctrl + F
全屏模式
F11
切换主题
Ctrl + Shift + D
显示快捷键
?
增大字号
Ctrl + =
减小字号
Ctrl + -