Re: [kernel-hardening] [PATCH v2] time: Remove CONFIG_TIMER_STATS

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi,

Don't forget to send to linux-api@xxxxxxxxxxxxxxx 

Le mercredi 08 février 2017 à 11:26 -0800, Kees Cook a écrit :
> Currently CONFIG_TIMER_STATS exposes process information across
> namespaces:
> 
> kernel/time/timer_list.c print_timer():
> 
>         SEQ_printf(m, ", %s/%d", tmp, timer->start_pid);
> 
> /proc/timer_list:
> 
>  #11: <0000000000000000>, hrtimer_wakeup, S:01, do_nanosleep,
> cron/2570
> 
> Given that the tracer can give the same information, this patch
> entirely
> removes CONFIG_TIMER_STATS.
> 
> Suggested-by: Thomas Gleixner <tglx@xxxxxxxxxxxxx>
> Signed-off-by: Kees Cook <keescook@xxxxxxxxxxxx>
> Acked-by: John Stultz <john.stultz@xxxxxxxxxx>
> ---
> v2:
> - dropped doc comments for removed structure elements; thx 0-day
> builder.
> ---
>  Documentation/timers/timer_stats.txt |  73 ------
>  include/linux/hrtimer.h              |  11 -
>  include/linux/timer.h                |  45 ----
>  kernel/kthread.c                     |   1 -
>  kernel/time/Makefile                 |   1 -
>  kernel/time/hrtimer.c                |  38 ----
>  kernel/time/timer.c                  |  48 +---
>  kernel/time/timer_list.c             |  10 -
>  kernel/time/timer_stats.c            | 425 -----------------------
> ------------
>  kernel/workqueue.c                   |   2 -
>  lib/Kconfig.debug                    |  14 --
>  11 files changed, 2 insertions(+), 666 deletions(-)
>  delete mode 100644 Documentation/timers/timer_stats.txt
>  delete mode 100644 kernel/time/timer_stats.c
> 
> diff --git a/Documentation/timers/timer_stats.txt
> b/Documentation/timers/timer_stats.txt
> deleted file mode 100644
> index de835ee97455..000000000000
> --- a/Documentation/timers/timer_stats.txt
> +++ /dev/null
> @@ -1,73 +0,0 @@
> -timer_stats - timer usage statistics
> -------------------------------------
> -
> -timer_stats is a debugging facility to make the timer (ab)usage in a
> Linux
> -system visible to kernel and userspace developers. If enabled in the
> config
> -but not used it has almost zero runtime overhead, and a relatively
> small
> -data structure overhead. Even if collection is enabled runtime all
> the
> -locking is per-CPU and lookup is hashed.
> -
> -timer_stats should be used by kernel and userspace developers to
> verify that
> -their code does not make unduly use of timers. This helps to avoid
> unnecessary
> -wakeups, which should be avoided to optimize power consumption.
> -
> -It can be enabled by CONFIG_TIMER_STATS in the "Kernel hacking"
> configuration
> -section.
> -
> -timer_stats collects information about the timer events which are
> fired in a
> -Linux system over a sample period:
> -
> -- the pid of the task(process) which initialized the timer
> -- the name of the process which initialized the timer
> -- the function where the timer was initialized
> -- the callback function which is associated to the timer
> -- the number of events (callbacks)
> -
> -timer_stats adds an entry to /proc: /proc/timer_stats
> -
> -This entry is used to control the statistics functionality and to
> read out the
> -sampled information.
> -
> -The timer_stats functionality is inactive on bootup.
> -
> -To activate a sample period issue:
> -# echo 1 >/proc/timer_stats
> -
> -To stop a sample period issue:
> -# echo 0 >/proc/timer_stats
> -
> -The statistics can be retrieved by:
> -# cat /proc/timer_stats
> -
> -While sampling is enabled, each readout from /proc/timer_stats will
> see
> -newly updated statistics. Once sampling is disabled, the sampled
> information
> -is kept until a new sample period is started. This allows multiple
> readouts.
> -
> -Sample output of /proc/timer_stats:
> -
> -Timerstats sample period: 3.888770 s
> -  12,     0 swapper          hrtimer_stop_sched_tick
> (hrtimer_sched_tick)
> -  15,     1 swapper          hcd_submit_urb (rh_timer_func)
> -   4,   959 kedac            schedule_timeout (process_timeout)
> -   1,     0 swapper          page_writeback_init (wb_timer_fn)
> -  28,     0 swapper          hrtimer_stop_sched_tick
> (hrtimer_sched_tick)
> -  22,  2948 IRQ 4            tty_flip_buffer_push
> (delayed_work_timer_fn)
> -   3,  3100 bash             schedule_timeout (process_timeout)
> -   1,     1 swapper          queue_delayed_work_on
> (delayed_work_timer_fn)
> -   1,     1 swapper          queue_delayed_work_on
> (delayed_work_timer_fn)
> -   1,     1 swapper          neigh_table_init_no_netlink
> (neigh_periodic_timer)
> -   1,  2292 ip               __netdev_watchdog_up (dev_watchdog)
> -   1,    23 events/1         do_cache_clean (delayed_work_timer_fn)
> -90 total events, 30.0 events/sec
> -
> -The first column is the number of events, the second column the pid,
> the third
> -column is the name of the process. The forth column shows the
> function which
> -initialized the timer and in parenthesis the callback function which
> was
> -executed on expiry.
> -
> -    Thomas, Ingo
> -
> -Added flag to indicate 'deferrable timer' in /proc/timer_stats. A
> deferrable
> -timer will appear as follows
> -  10D,     1 swapper          queue_delayed_work_on
> (delayed_work_timer_fn)
> -
> diff --git a/include/linux/hrtimer.h b/include/linux/hrtimer.h
> index cdab81ba29f8..e52b427223ba 100644
> --- a/include/linux/hrtimer.h
> +++ b/include/linux/hrtimer.h
> @@ -88,12 +88,6 @@ enum hrtimer_restart {
>   * @base:	pointer to the timer base (per cpu and per clock)
>   * @state:	state information (See bit values above)
>   * @is_rel:	Set if the timer was armed relative
> - * @start_pid:  timer statistics field to store the pid of the task
> which
> - *		started the timer
> - * @start_site:	timer statistics field to store the site
> where the timer
> - *		was started
> - * @start_comm: timer statistics field to store the name of the
> process which
> - *		started the timer
>   *
>   * The hrtimer structure must be initialized by hrtimer_init()
>   */
> @@ -104,11 +98,6 @@ struct hrtimer {
>  	struct hrtimer_clock_base	*base;
>  	u8				state;
>  	u8				is_rel;
> -#ifdef CONFIG_TIMER_STATS
> -	int				start_pid;
> -	void				*start_site;
> -	char				start_comm[16];
> -#endif
>  };
>  
>  /**
> diff --git a/include/linux/timer.h b/include/linux/timer.h
> index 51d601f192d4..5a209b84fd9e 100644
> --- a/include/linux/timer.h
> +++ b/include/linux/timer.h
> @@ -20,11 +20,6 @@ struct timer_list {
>  	unsigned long		data;
>  	u32			flags;
>  
> -#ifdef CONFIG_TIMER_STATS
> -	int			start_pid;
> -	void			*start_site;
> -	char			start_comm[16];
> -#endif
>  #ifdef CONFIG_LOCKDEP
>  	struct lockdep_map	lockdep_map;
>  #endif
> @@ -197,46 +192,6 @@ extern int mod_timer_pending(struct timer_list
> *timer, unsigned long expires);
>   */
>  #define NEXT_TIMER_MAX_DELTA	((1UL << 30) - 1)
>  
> -/*
> - * Timer-statistics info:
> - */
> -#ifdef CONFIG_TIMER_STATS
> -
> -extern int timer_stats_active;
> -
> -extern void init_timer_stats(void);
> -
> -extern void timer_stats_update_stats(void *timer, pid_t pid, void
> *startf,
> -				     void *timerf, char *comm, u32
> flags);
> -
> -extern void __timer_stats_timer_set_start_info(struct timer_list
> *timer,
> -					       void *addr);
> -
> -static inline void timer_stats_timer_set_start_info(struct
> timer_list *timer)
> -{
> -	if (likely(!timer_stats_active))
> -		return;
> -	__timer_stats_timer_set_start_info(timer,
> __builtin_return_address(0));
> -}
> -
> -static inline void timer_stats_timer_clear_start_info(struct
> timer_list *timer)
> -{
> -	timer->start_site = NULL;
> -}
> -#else
> -static inline void init_timer_stats(void)
> -{
> -}
> -
> -static inline void timer_stats_timer_set_start_info(struct
> timer_list *timer)
> -{
> -}
> -
> -static inline void timer_stats_timer_clear_start_info(struct
> timer_list *timer)
> -{
> -}
> -#endif
> -
>  extern void add_timer(struct timer_list *timer);
>  
>  extern int try_to_del_timer_sync(struct timer_list *timer);
> diff --git a/kernel/kthread.c b/kernel/kthread.c
> index 2318fba86277..8461a4372e8a 100644
> --- a/kernel/kthread.c
> +++ b/kernel/kthread.c
> @@ -850,7 +850,6 @@ void __kthread_queue_delayed_work(struct
> kthread_worker *worker,
>  
>  	list_add(&work->node, &worker->delayed_work_list);
>  	work->worker = worker;
> -	timer_stats_timer_set_start_info(&dwork->timer);
>  	timer->expires = jiffies + delay;
>  	add_timer(timer);
>  }
> diff --git a/kernel/time/Makefile b/kernel/time/Makefile
> index 976840d29a71..938dbf33ef49 100644
> --- a/kernel/time/Makefile
> +++ b/kernel/time/Makefile
> @@ -15,6 +15,5 @@ ifeq ($(CONFIG_GENERIC_CLOCKEVENTS_BROADCAST),y)
>  endif
>  obj-$(CONFIG_GENERIC_SCHED_CLOCK)		+= sched_clock.o
>  obj-$(CONFIG_TICK_ONESHOT)			+= tick-oneshot.o
> tick-sched.o
> -obj-$(CONFIG_TIMER_STATS)			+= timer_stats.o
>  obj-$(CONFIG_DEBUG_FS)				+=
> timekeeping_debug.o
>  obj-$(CONFIG_TEST_UDELAY)			+= test_udelay.o
> diff --git a/kernel/time/hrtimer.c b/kernel/time/hrtimer.c
> index c6ecedd3b839..edabde646e58 100644
> --- a/kernel/time/hrtimer.c
> +++ b/kernel/time/hrtimer.c
> @@ -766,34 +766,6 @@ void hrtimers_resume(void)
>  	clock_was_set_delayed();
>  }
>  
> -static inline void timer_stats_hrtimer_set_start_info(struct hrtimer
> *timer)
> -{
> -#ifdef CONFIG_TIMER_STATS
> -	if (timer->start_site)
> -		return;
> -	timer->start_site = __builtin_return_address(0);
> -	memcpy(timer->start_comm, current->comm, TASK_COMM_LEN);
> -	timer->start_pid = current->pid;
> -#endif
> -}
> -
> -static inline void timer_stats_hrtimer_clear_start_info(struct
> hrtimer *timer)
> -{
> -#ifdef CONFIG_TIMER_STATS
> -	timer->start_site = NULL;
> -#endif
> -}
> -
> -static inline void timer_stats_account_hrtimer(struct hrtimer
> *timer)
> -{
> -#ifdef CONFIG_TIMER_STATS
> -	if (likely(!timer_stats_active))
> -		return;
> -	timer_stats_update_stats(timer, timer->start_pid, timer-
> >start_site,
> -				 timer->function, timer->start_comm, 
> 0);
> -#endif
> -}
> -
>  /*
>   * Counterpart to lock_hrtimer_base above:
>   */
> @@ -932,7 +904,6 @@ remove_hrtimer(struct hrtimer *timer, struct
> hrtimer_clock_base *base, bool rest
>  		 * rare case and less expensive than a smp call.
>  		 */
>  		debug_deactivate(timer);
> -		timer_stats_hrtimer_clear_start_info(timer);
>  		reprogram = base->cpu_base ==
> this_cpu_ptr(&hrtimer_bases);
>  
>  		if (!restart)
> @@ -990,8 +961,6 @@ void hrtimer_start_range_ns(struct hrtimer
> *timer, ktime_t tim,
>  	/* Switch the timer base, if necessary: */
>  	new_base = switch_hrtimer_base(timer, base, mode &
> HRTIMER_MODE_PINNED);
>  
> -	timer_stats_hrtimer_set_start_info(timer);
> -
>  	leftmost = enqueue_hrtimer(timer, new_base);
>  	if (!leftmost)
>  		goto unlock;
> @@ -1128,12 +1097,6 @@ static void __hrtimer_init(struct hrtimer
> *timer, clockid_t clock_id,
>  	base = hrtimer_clockid_to_base(clock_id);
>  	timer->base = &cpu_base->clock_base[base];
>  	timerqueue_init(&timer->node);
> -
> -#ifdef CONFIG_TIMER_STATS
> -	timer->start_site = NULL;
> -	timer->start_pid = -1;
> -	memset(timer->start_comm, 0, TASK_COMM_LEN);
> -#endif
>  }
>  
>  /**
> @@ -1217,7 +1180,6 @@ static void __run_hrtimer(struct
> hrtimer_cpu_base *cpu_base,
>  	raw_write_seqcount_barrier(&cpu_base->seq);
>  
>  	__remove_hrtimer(timer, base, HRTIMER_STATE_INACTIVE, 0);
> -	timer_stats_account_hrtimer(timer);
>  	fn = timer->function;
>  
>  	/*
> diff --git a/kernel/time/timer.c b/kernel/time/timer.c
> index ec33a6933eae..82a6bfa0c307 100644
> --- a/kernel/time/timer.c
> +++ b/kernel/time/timer.c
> @@ -571,38 +571,6 @@ internal_add_timer(struct timer_base *base,
> struct timer_list *timer)
>  	trigger_dyntick_cpu(base, timer);
>  }
>  
> -#ifdef CONFIG_TIMER_STATS
> -void __timer_stats_timer_set_start_info(struct timer_list *timer,
> void *addr)
> -{
> -	if (timer->start_site)
> -		return;
> -
> -	timer->start_site = addr;
> -	memcpy(timer->start_comm, current->comm, TASK_COMM_LEN);
> -	timer->start_pid = current->pid;
> -}
> -
> -static void timer_stats_account_timer(struct timer_list *timer)
> -{
> -	void *site;
> -
> -	/*
> -	 * start_site can be concurrently reset by
> -	 * timer_stats_timer_clear_start_info()
> -	 */
> -	site = READ_ONCE(timer->start_site);
> -	if (likely(!site))
> -		return;
> -
> -	timer_stats_update_stats(timer, timer->start_pid, site,
> -				 timer->function, timer->start_comm,
> -				 timer->flags);
> -}
> -
> -#else
> -static void timer_stats_account_timer(struct timer_list *timer) {}
> -#endif
> -
>  #ifdef CONFIG_DEBUG_OBJECTS_TIMERS
>  
>  static struct debug_obj_descr timer_debug_descr;
> @@ -789,11 +757,6 @@ static void do_init_timer(struct timer_list
> *timer, unsigned int flags,
>  {
>  	timer->entry.pprev = NULL;
>  	timer->flags = flags | raw_smp_processor_id();
> -#ifdef CONFIG_TIMER_STATS
> -	timer->start_site = NULL;
> -	timer->start_pid = -1;
> -	memset(timer->start_comm, 0, TASK_COMM_LEN);
> -#endif
>  	lockdep_init_map(&timer->lockdep_map, name, key, 0);
>  }
>  
> @@ -1001,8 +964,6 @@ __mod_timer(struct timer_list *timer, unsigned
> long expires, bool pending_only)
>  		base = lock_timer_base(timer, &flags);
>  	}
>  
> -	timer_stats_timer_set_start_info(timer);
> -
>  	ret = detach_if_pending(timer, base, false);
>  	if (!ret && pending_only)
>  		goto out_unlock;
> @@ -1130,7 +1091,6 @@ void add_timer_on(struct timer_list *timer, int
> cpu)
>  	struct timer_base *new_base, *base;
>  	unsigned long flags;
>  
> -	timer_stats_timer_set_start_info(timer);
>  	BUG_ON(timer_pending(timer) || !timer->function);
>  
>  	new_base = get_timer_cpu_base(timer->flags, cpu);
> @@ -1176,7 +1136,6 @@ int del_timer(struct timer_list *timer)
>  
>  	debug_assert_init(timer);
>  
> -	timer_stats_timer_clear_start_info(timer);
>  	if (timer_pending(timer)) {
>  		base = lock_timer_base(timer, &flags);
>  		ret = detach_if_pending(timer, base, true);
> @@ -1204,10 +1163,9 @@ int try_to_del_timer_sync(struct timer_list
> *timer)
>  
>  	base = lock_timer_base(timer, &flags);
>  
> -	if (base->running_timer != timer) {
> -		timer_stats_timer_clear_start_info(timer);
> +	if (base->running_timer != timer)
>  		ret = detach_if_pending(timer, base, true);
> -	}
> +
>  	spin_unlock_irqrestore(&base->lock, flags);
>  
>  	return ret;
> @@ -1331,7 +1289,6 @@ static void expire_timers(struct timer_base
> *base, struct hlist_head *head)
>  		unsigned long data;
>  
>  		timer = hlist_entry(head->first, struct timer_list,
> entry);
> -		timer_stats_account_timer(timer);
>  
>  		base->running_timer = timer;
>  		detach_timer(timer, true);
> @@ -1868,7 +1825,6 @@ static void __init init_timer_cpus(void)
>  void __init init_timers(void)
>  {
>  	init_timer_cpus();
> -	init_timer_stats();
>  	open_softirq(TIMER_SOFTIRQ, run_timer_softirq);
>  }
>  
> diff --git a/kernel/time/timer_list.c b/kernel/time/timer_list.c
> index afe6cd1944fc..387a3a5aa388 100644
> --- a/kernel/time/timer_list.c
> +++ b/kernel/time/timer_list.c
> @@ -62,21 +62,11 @@ static void
>  print_timer(struct seq_file *m, struct hrtimer *taddr, struct
> hrtimer *timer,
>  	    int idx, u64 now)
>  {
> -#ifdef CONFIG_TIMER_STATS
> -	char tmp[TASK_COMM_LEN + 1];
> -#endif
>  	SEQ_printf(m, " #%d: ", idx);
>  	print_name_offset(m, taddr);
>  	SEQ_printf(m, ", ");
>  	print_name_offset(m, timer->function);
>  	SEQ_printf(m, ", S:%02x", timer->state);
> -#ifdef CONFIG_TIMER_STATS
> -	SEQ_printf(m, ", ");
> -	print_name_offset(m, timer->start_site);
> -	memcpy(tmp, timer->start_comm, TASK_COMM_LEN);
> -	tmp[TASK_COMM_LEN] = 0;
> -	SEQ_printf(m, ", %s/%d", tmp, timer->start_pid);
> -#endif
>  	SEQ_printf(m, "\n");
>  	SEQ_printf(m, " # expires at %Lu-%Lu nsecs [in %Ld to %Ld
> nsecs]\n",
>  		(unsigned long
> long)ktime_to_ns(hrtimer_get_softexpires(timer)),
> diff --git a/kernel/time/timer_stats.c b/kernel/time/timer_stats.c
> deleted file mode 100644
> index afddded947df..000000000000
> --- a/kernel/time/timer_stats.c
> +++ /dev/null
> @@ -1,425 +0,0 @@
> -/*
> - * kernel/time/timer_stats.c
> - *
> - * Collect timer usage statistics.
> - *
> - * Copyright(C) 2006, Red Hat, Inc., Ingo Molnar
> - * Copyright(C) 2006 Timesys Corp., Thomas Gleixner <tglx@xxxxxxxxxx
> m>
> - *
> - * timer_stats is based on timer_top, a similar functionality which
> was part of
> - * Con Kolivas dyntick patch set. It was developed by Daniel Petrini
> at the
> - * Instituto Nokia de Tecnologia - INdT - Manaus. timer_top's design
> was based
> - * on dynamic allocation of the statistics entries and linear search
> based
> - * lookup combined with a global lock, rather than the static array,
> hash
> - * and per-CPU locking which is used by timer_stats. It was written
> for the
> - * pre hrtimer kernel code and therefore did not take hrtimers into
> account.
> - * Nevertheless it provided the base for the timer_stats
> implementation and
> - * was a helpful source of inspiration. Kudos to Daniel and the
> Nokia folks
> - * for this effort.
> - *
> - * timer_top.c is
> - *	Copyright (C) 2005 Instituto Nokia de Tecnologia - INdT -
> Manaus
> - *	Written by Daniel Petrini <d.pensator@xxxxxxxxx>
> - *	timer_top.c was released under the GNU General Public
> License version 2
> - *
> - * We export the addresses and counting of timer functions being
> called,
> - * the pid and cmdline from the owner process if applicable.
> - *
> - * Start/stop data collection:
> - * # echo [1|0] >/proc/timer_stats
> - *
> - * Display the information collected so far:
> - * # cat /proc/timer_stats
> - *
> - * This program is free software; you can redistribute it and/or
> modify
> - * it under the terms of the GNU General Public License version 2 as
> - * published by the Free Software Foundation.
> - */
> -
> -#include <linux/proc_fs.h>
> -#include <linux/module.h>
> -#include <linux/spinlock.h>
> -#include <linux/sched.h>
> -#include <linux/seq_file.h>
> -#include <linux/kallsyms.h>
> -
> -#include <linux/uaccess.h>
> -
> -/*
> - * This is our basic unit of interest: a timer expiry event
> identified
> - * by the timer, its start/expire functions and the PID of the task
> that
> - * started the timer. We count the number of times an event happens:
> - */
> -struct entry {
> -	/*
> -	 * Hash list:
> -	 */
> -	struct entry		*next;
> -
> -	/*
> -	 * Hash keys:
> -	 */
> -	void			*timer;
> -	void			*start_func;
> -	void			*expire_func;
> -	pid_t			pid;
> -
> -	/*
> -	 * Number of timeout events:
> -	 */
> -	unsigned long		count;
> -	u32			flags;
> -
> -	/*
> -	 * We save the command-line string to preserve
> -	 * this information past task exit:
> -	 */
> -	char			comm[TASK_COMM_LEN + 1];
> -
> -} ____cacheline_aligned_in_smp;
> -
> -/*
> - * Spinlock protecting the tables - not taken during lookup:
> - */
> -static DEFINE_RAW_SPINLOCK(table_lock);
> -
> -/*
> - * Per-CPU lookup locks for fast hash lookup:
> - */
> -static DEFINE_PER_CPU(raw_spinlock_t, tstats_lookup_lock);
> -
> -/*
> - * Mutex to serialize state changes with show-stats activities:
> - */
> -static DEFINE_MUTEX(show_mutex);
> -
> -/*
> - * Collection status, active/inactive:
> - */
> -int __read_mostly timer_stats_active;
> -
> -/*
> - * Beginning/end timestamps of measurement:
> - */
> -static ktime_t time_start, time_stop;
> -
> -/*
> - * tstat entry structs only get allocated while collection is
> - * active and never freed during that time - this simplifies
> - * things quite a bit.
> - *
> - * They get freed when a new collection period is started.
> - */
> -#define MAX_ENTRIES_BITS	10
> -#define MAX_ENTRIES		(1UL << MAX_ENTRIES_BITS)
> -
> -static unsigned long nr_entries;
> -static struct entry entries[MAX_ENTRIES];
> -
> -static atomic_t overflow_count;
> -
> -/*
> - * The entries are in a hash-table, for fast lookup:
> - */
> -#define TSTAT_HASH_BITS		(MAX_ENTRIES_BITS - 1)
> -#define TSTAT_HASH_SIZE		(1UL << TSTAT_HASH_BITS)
> -#define TSTAT_HASH_MASK		(TSTAT_HASH_SIZE - 1)
> -
> -#define __tstat_hashfn(entry)					
> 	\
> -	(((unsigned long)(entry)->timer       ^			
> 	\
> -	  (unsigned long)(entry)->start_func  ^			
> 	\
> -	  (unsigned long)(entry)->expire_func ^			
> 	\
> -	  (unsigned long)(entry)->pid		) &
> TSTAT_HASH_MASK)
> -
> -#define tstat_hashentry(entry)	(tstat_hash_table +
> __tstat_hashfn(entry))
> -
> -static struct entry *tstat_hash_table[TSTAT_HASH_SIZE]
> __read_mostly;
> -
> -static void reset_entries(void)
> -{
> -	nr_entries = 0;
> -	memset(entries, 0, sizeof(entries));
> -	memset(tstat_hash_table, 0, sizeof(tstat_hash_table));
> -	atomic_set(&overflow_count, 0);
> -}
> -
> -static struct entry *alloc_entry(void)
> -{
> -	if (nr_entries >= MAX_ENTRIES)
> -		return NULL;
> -
> -	return entries + nr_entries++;
> -}
> -
> -static int match_entries(struct entry *entry1, struct entry *entry2)
> -{
> -	return entry1->timer       == entry2->timer	  &&
> -	       entry1->start_func  == entry2->start_func  &&
> -	       entry1->expire_func == entry2->expire_func &&
> -	       entry1->pid	   == entry2->pid;
> -}
> -
> -/*
> - * Look up whether an entry matching this item is present
> - * in the hash already. Must be called with irqs off and the
> - * lookup lock held:
> - */
> -static struct entry *tstat_lookup(struct entry *entry, char *comm)
> -{
> -	struct entry **head, *curr, *prev;
> -
> -	head = tstat_hashentry(entry);
> -	curr = *head;
> -
> -	/*
> -	 * The fastpath is when the entry is already hashed,
> -	 * we do this with the lookup lock held, but with the
> -	 * table lock not held:
> -	 */
> -	while (curr) {
> -		if (match_entries(curr, entry))
> -			return curr;
> -
> -		curr = curr->next;
> -	}
> -	/*
> -	 * Slowpath: allocate, set up and link a new hash entry:
> -	 */
> -	prev = NULL;
> -	curr = *head;
> -
> -	raw_spin_lock(&table_lock);
> -	/*
> -	 * Make sure we have not raced with another CPU:
> -	 */
> -	while (curr) {
> -		if (match_entries(curr, entry))
> -			goto out_unlock;
> -
> -		prev = curr;
> -		curr = curr->next;
> -	}
> -
> -	curr = alloc_entry();
> -	if (curr) {
> -		*curr = *entry;
> -		curr->count = 0;
> -		curr->next = NULL;
> -		memcpy(curr->comm, comm, TASK_COMM_LEN);
> -
> -		smp_mb(); /* Ensure that curr is initialized before
> insert */
> -
> -		if (prev)
> -			prev->next = curr;
> -		else
> -			*head = curr;
> -	}
> - out_unlock:
> -	raw_spin_unlock(&table_lock);
> -
> -	return curr;
> -}
> -
> -/**
> - * timer_stats_update_stats - Update the statistics for a timer.
> - * @timer:	pointer to either a timer_list or a hrtimer
> - * @pid:	the pid of the task which set up the timer
> - * @startf:	pointer to the function which did the timer setup
> - * @timerf:	pointer to the timer callback function of the
> timer
> - * @comm:	name of the process which set up the timer
> - * @tflags:	The flags field of the timer
> - *
> - * When the timer is already registered, then the event counter is
> - * incremented. Otherwise the timer is registered in a free slot.
> - */
> -void timer_stats_update_stats(void *timer, pid_t pid, void *startf,
> -			      void *timerf, char *comm, u32 tflags)
> -{
> -	/*
> -	 * It doesn't matter which lock we take:
> -	 */
> -	raw_spinlock_t *lock;
> -	struct entry *entry, input;
> -	unsigned long flags;
> -
> -	if (likely(!timer_stats_active))
> -		return;
> -
> -	lock = &per_cpu(tstats_lookup_lock, raw_smp_processor_id());
> -
> -	input.timer = timer;
> -	input.start_func = startf;
> -	input.expire_func = timerf;
> -	input.pid = pid;
> -	input.flags = tflags;
> -
> -	raw_spin_lock_irqsave(lock, flags);
> -	if (!timer_stats_active)
> -		goto out_unlock;
> -
> -	entry = tstat_lookup(&input, comm);
> -	if (likely(entry))
> -		entry->count++;
> -	else
> -		atomic_inc(&overflow_count);
> -
> - out_unlock:
> -	raw_spin_unlock_irqrestore(lock, flags);
> -}
> -
> -static void print_name_offset(struct seq_file *m, unsigned long
> addr)
> -{
> -	char symname[KSYM_NAME_LEN];
> -
> -	if (lookup_symbol_name(addr, symname) < 0)
> -		seq_printf(m, "<%p>", (void *)addr);
> -	else
> -		seq_printf(m, "%s", symname);
> -}
> -
> -static int tstats_show(struct seq_file *m, void *v)
> -{
> -	struct timespec64 period;
> -	struct entry *entry;
> -	unsigned long ms;
> -	long events = 0;
> -	ktime_t time;
> -	int i;
> -
> -	mutex_lock(&show_mutex);
> -	/*
> -	 * If still active then calculate up to now:
> -	 */
> -	if (timer_stats_active)
> -		time_stop = ktime_get();
> -
> -	time = ktime_sub(time_stop, time_start);
> -
> -	period = ktime_to_timespec64(time);
> -	ms = period.tv_nsec / 1000000;
> -
> -	seq_puts(m, "Timer Stats Version: v0.3\n");
> -	seq_printf(m, "Sample period: %ld.%03ld s\n",
> (long)period.tv_sec, ms);
> -	if (atomic_read(&overflow_count))
> -		seq_printf(m, "Overflow: %d entries\n",
> atomic_read(&overflow_count));
> -	seq_printf(m, "Collection: %s\n", timer_stats_active ?
> "active" : "inactive");
> -
> -	for (i = 0; i < nr_entries; i++) {
> -		entry = entries + i;
> -		if (entry->flags & TIMER_DEFERRABLE) {
> -			seq_printf(m, "%4luD, %5d %-16s ",
> -				entry->count, entry->pid, entry-
> >comm);
> -		} else {
> -			seq_printf(m, " %4lu, %5d %-16s ",
> -				entry->count, entry->pid, entry-
> >comm);
> -		}
> -
> -		print_name_offset(m, (unsigned long)entry-
> >start_func);
> -		seq_puts(m, " (");
> -		print_name_offset(m, (unsigned long)entry-
> >expire_func);
> -		seq_puts(m, ")\n");
> -
> -		events += entry->count;
> -	}
> -
> -	ms += period.tv_sec * 1000;
> -	if (!ms)
> -		ms = 1;
> -
> -	if (events && period.tv_sec)
> -		seq_printf(m, "%ld total events, %ld.%03ld
> events/sec\n",
> -			   events, events * 1000 / ms,
> -			   (events * 1000000 / ms) % 1000);
> -	else
> -		seq_printf(m, "%ld total events\n", events);
> -
> -	mutex_unlock(&show_mutex);
> -
> -	return 0;
> -}
> -
> -/*
> - * After a state change, make sure all concurrent lookup/update
> - * activities have stopped:
> - */
> -static void sync_access(void)
> -{
> -	unsigned long flags;
> -	int cpu;
> -
> -	for_each_online_cpu(cpu) {
> -		raw_spinlock_t *lock = &per_cpu(tstats_lookup_lock,
> cpu);
> -
> -		raw_spin_lock_irqsave(lock, flags);
> -		/* nothing */
> -		raw_spin_unlock_irqrestore(lock, flags);
> -	}
> -}
> -
> -static ssize_t tstats_write(struct file *file, const char __user
> *buf,
> -			    size_t count, loff_t *offs)
> -{
> -	char ctl[2];
> -
> -	if (count != 2 || *offs)
> -		return -EINVAL;
> -
> -	if (copy_from_user(ctl, buf, count))
> -		return -EFAULT;
> -
> -	mutex_lock(&show_mutex);
> -	switch (ctl[0]) {
> -	case '0':
> -		if (timer_stats_active) {
> -			timer_stats_active = 0;
> -			time_stop = ktime_get();
> -			sync_access();
> -		}
> -		break;
> -	case '1':
> -		if (!timer_stats_active) {
> -			reset_entries();
> -			time_start = ktime_get();
> -			smp_mb();
> -			timer_stats_active = 1;
> -		}
> -		break;
> -	default:
> -		count = -EINVAL;
> -	}
> -	mutex_unlock(&show_mutex);
> -
> -	return count;
> -}
> -
> -static int tstats_open(struct inode *inode, struct file *filp)
> -{
> -	return single_open(filp, tstats_show, NULL);
> -}
> -
> -static const struct file_operations tstats_fops = {
> -	.open		= tstats_open,
> -	.read		= seq_read,
> -	.write		= tstats_write,
> -	.llseek		= seq_lseek,
> -	.release	= single_release,
> -};
> -
> -void __init init_timer_stats(void)
> -{
> -	int cpu;
> -
> -	for_each_possible_cpu(cpu)
> -		raw_spin_lock_init(&per_cpu(tstats_lookup_lock,
> cpu));
> -}
> -
> -static int __init init_tstats_procfs(void)
> -{
> -	struct proc_dir_entry *pe;
> -
> -	pe = proc_create("timer_stats", 0644, NULL, &tstats_fops);
> -	if (!pe)
> -		return -ENOMEM;
> -	return 0;
> -}
> -__initcall(init_tstats_procfs);
> diff --git a/kernel/workqueue.c b/kernel/workqueue.c
> index 1d9fb6543a66..072cbc9b175d 100644
> --- a/kernel/workqueue.c
> +++ b/kernel/workqueue.c
> @@ -1523,8 +1523,6 @@ static void __queue_delayed_work(int cpu,
> struct workqueue_struct *wq,
>  		return;
>  	}
>  
> -	timer_stats_timer_set_start_info(&dwork->timer);
> -
>  	dwork->wq = wq;
>  	dwork->cpu = cpu;
>  	timer->expires = jiffies + delay;
> diff --git a/lib/Kconfig.debug b/lib/Kconfig.debug
> index eb9e9a7870fa..132af338d6dd 100644
> --- a/lib/Kconfig.debug
> +++ b/lib/Kconfig.debug
> @@ -980,20 +980,6 @@ config DEBUG_TIMEKEEPING
>  
>  	  If unsure, say N.
>  
> -config TIMER_STATS
> -	bool "Collect kernel timers statistics"
> -	depends on DEBUG_KERNEL && PROC_FS
> -	help
> -	  If you say Y here, additional code will be inserted into
> the
> -	  timer routines to collect statistics about kernel timers
> being
> -	  reprogrammed. The statistics can be read from
> /proc/timer_stats.
> -	  The statistics collection is started by writing 1 to
> /proc/timer_stats,
> -	  writing 0 stops it. This feature is useful to collect
> information
> -	  about timer usage patterns in kernel and userspace. This
> feature
> -	  is lightweight if enabled in the kernel config but not
> activated
> -	  (it defaults to deactivated on bootup and will only be
> activated
> -	  if some application like powertop activates it
> explicitly).
> -
>  config DEBUG_PREEMPT
>  	bool "Debug preemptible kernel"
>  	depends on DEBUG_KERNEL && PREEMPT && TRACE_IRQFLAGS_SUPPORT
> -- 
> 2.7.4
> 
> 

--
To unsubscribe from this list: send the line "unsubscribe linux-doc" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html



[Index of Archives]     [Kernel Newbies]     [Security]     [Netfilter]     [Bugtraq]     [Linux FS]     [Yosemite Forum]     [MIPS Linux]     [ARM Linux]     [Linux Security]     [Linux RAID]     [Samba]     [Video 4 Linux]     [Device Mapper]     [Linux Resources]

  Powered by Linux