/* * linux/fs/nfs/iostat.h * * Declarations for NFS client per-mount statistics * * Copyright (C) 2005, 2006 Chuck Lever * */ #ifndef _NFS_IOSTAT #define _NFS_IOSTAT #include #include #include struct nfs_iostats { unsigned long long bytes[__NFSIOS_BYTESMAX]; #ifdef CONFIG_NFS_FSCACHE unsigned long long fscache[__NFSIOS_FSCACHEMAX]; #endif unsigned long events[__NFSIOS_COUNTSMAX]; } ____cacheline_aligned; static inline void nfs_inc_server_stats(const struct nfs_server *server, enum nfs_stat_eventcounters stat) { this_cpu_inc(server->io_stats->events[stat]); } static inline void nfs_inc_stats(const struct inode *inode, enum nfs_stat_eventcounters stat) { nfs_inc_server_stats(NFS_SERVER(inode), stat); } static inline void nfs_add_server_stats(const struct nfs_server *server, enum nfs_stat_bytecounters stat, long addend) { this_cpu_add(server->io_stats->bytes[stat], addend); } static inline void nfs_add_stats(const struct inode *inode, enum nfs_stat_bytecounters stat, long addend) { nfs_add_server_stats(NFS_SERVER(inode), stat, addend); } #ifdef CONFIG_NFS_FSCACHE static inline void nfs_add_fscache_stats(struct inode *inode, enum nfs_stat_fscachecounters stat, long addend) { this_cpu_add(NFS_SERVER(inode)->io_stats->fscache[stat], addend); } static inline void nfs_inc_fscache_stats(struct inode *inode, enum nfs_stat_fscachecounters stat) { this_cpu_inc(NFS_SERVER(inode)->io_stats->fscache[stat]); } #endif static inline struct nfs_iostats __percpu *nfs_alloc_iostats(void) { return alloc_percpu(struct nfs_iostats); } static inline void nfs_free_iostats(struct nfs_iostats __percpu *stats) { if (stats != NULL) free_percpu(stats); } #endif /* _NFS_IOSTAT */ ae5b42828785b589f686725ff5bc3b9b25'>treecommitdiff
diff options
context:
space:
mode:
authorThomas Gleixner <tglx@linutronix.de>2017-01-31 09:37:34 +0100
committerThomas Gleixner <tglx@linutronix.de>2017-01-31 21:47:58 +0100
commit0becc0ae5b42828785b589f686725ff5bc3b9b25 (patch)
treebe6d0e1f37c38ed0a7dd5da2d4b1e93f0fb43101 /sound/soc/codecs/wm8996.c
parent24c2503255d35c269b67162c397a1a1c1e02f6ce (diff)
x86/mce: Make timer handling more robust
Erik reported that on a preproduction hardware a CMCI storm triggers the BUG_ON in add_timer_on(). The reason is that the per CPU MCE timer is started by the CMCI logic before the MCE CPU hotplug callback starts the timer with add_timer_on(). So the timer is already queued which triggers the BUG. Using add_timer_on() is pretty pointless in this code because the timer is strictlty per CPU, initialized as pinned and all operations which arm the timer happen on the CPU to which the timer belongs. Simplify the whole machinery by using mod_timer() instead of add_timer_on() which avoids the problem because mod_timer() can handle already queued timers. Use __start_timer() everywhere so the earliest armed expiry time is preserved. Reported-by: Erik Veijola <erik.veijola@intel.com> Tested-by: Borislav Petkov <bp@alien8.de> Signed-off-by: Thomas Gleixner <tglx@linutronix.de> Reviewed-by: Borislav Petkov <bp@alien8.de> Cc: Tony Luck <tony.luck@intel.com> Link: http://lkml.kernel.org/r/alpine.DEB.2.20.1701310936080.3457@nanos Signed-off-by: Thomas Gleixner <tglx@linutronix.de>
Diffstat (limited to 'sound/soc/codecs/wm8996.c')