summaryrefslogtreecommitdiff
path: root/proc.h
blob: 0a31f36a1fca264575a906a523a8e2a5c644a16d (plain)
1
2
3
4
5
6
7
8
#ifndef PROC_H
#define PROC_H

extern void cpu_affinity(int cpu);
extern int set_proc_prio(int prio);
extern int set_sched_status(int policy, int priority);

#endif /* PROC_H */
l'>
authorWill Deacon <will.deacon@arm.com>2016-09-05 11:56:05 +0100
committerCatalin Marinas <catalin.marinas@arm.com>2016-09-09 12:33:48 +0100
commit872c63fbf9e153146b07f0cece4da0d70b283eeb (patch)
tree86880c47da4f3557f5673fdb1cb3e00a660fff7d
parentc6935931c1894ff857616ff8549b61236a19148f (diff)
arm64: spinlocks: implement smp_mb__before_spinlock() as smp_mb()
smp_mb__before_spinlock() is intended to upgrade a spin_lock() operation to a full barrier, such that prior stores are ordered with respect to loads and stores occuring inside the critical section. Unfortunately, the core code defines the barrier as smp_wmb(), which is insufficient to provide the required ordering guarantees when used in conjunction with our load-acquire-based spinlock implementation. This patch overrides the arm64 definition of smp_mb__before_spinlock() to map to a full smp_mb(). Cc: <stable@vger.kernel.org> Cc: Peter Zijlstra <peterz@infradead.org> Reported-by: Alan Stern <stern@rowland.harvard.edu> Signed-off-by: Will Deacon <will.deacon@arm.com> Signed-off-by: Catalin Marinas <catalin.marinas@arm.com>
Diffstat