/* * Xen Event Channels (internal header) * * Copyright (C) 2013 Citrix Systems R&D Ltd. * * This source code is licensed under the GNU General Public License, * Version 2 or later. See the file COPYING for more details. */ #ifndef __EVENTS_INTERNAL_H__ #define __EVENTS_INTERNAL_H__ /* Interrupt types. */ enum xen_irq_type { IRQT_UNBOUND = 0, IRQT_PIRQ, IRQT_VIRQ, IRQT_IPI, IRQT_EVTCHN }; /* * Packed IRQ information: * type - enum xen_irq_type * event channel - irq->event channel mapping * cpu - cpu this event channel is bound to * index - type-specific information: * PIRQ - vector, with MSB being "needs EIO", or physical IRQ of the HVM * guest, or GSI (real passthrough IRQ) of the device. * VIRQ - virq number * IPI - IPI vector * EVTCHN - */ struct irq_info { struct list_head list; int refcnt; enum xen_irq_type type; /* type */ unsigned irq; unsigned int evtchn; /* event channel */ unsigned short cpu; /* cpu bound */ union { unsigned short virq; enum ipi_vector ipi; struct { unsigned short pirq; unsigned short gsi; unsigned char vector; unsigned char flags; uint16_t domid; } pirq; } u; }; #define PIRQ_NEEDS_EOI (1 << 0) #define PIRQ_SHAREABLE (1 << 1) #define PIRQ_MSI_GROUP (1 << 2) struct evtchn_ops { unsigned (*max_channels)(void); unsigned (*nr_channels)(void); int (*setup)(struct irq_info *info); void (*bind_to_cpu)(struct irq_info *info, unsigned cpu); void (*clear_pending)(unsigned port); void (*set_pending)(unsigned port); bool (*is_pending)(unsigned port); bool (*test_and_set_mask)(unsigned port); void (*mask)(unsigned port); void (*unmask)(unsigned port); void (*handle_events)(unsigned cpu); void (*resume)(void); }; extern const struct evtchn_ops *evtchn_ops; extern int **evtchn_to_irq; int get_evtchn_to_irq(unsigned int evtchn); struct irq_info *info_for_irq(unsigned irq); unsigned cpu_from_irq(unsigned irq); unsigned cpu_from_evtchn(unsigned int evtchn); static inline unsigned xen_evtchn_max_channels(void) { return evtchn_ops->max_channels(); } /* * Do any ABI specific setup for a bound event channel before it can * be unmasked and used. */ static inline int xen_evtchn_port_setup(struct irq_info *info) { if (evtchn_ops->setup) return evtchn_ops->setup(info); return 0; } static inline void xen_evtchn_port_bind_to_cpu(struct irq_info *info, unsigned cpu) { evtchn_ops->bind_to_cpu(info, cpu); } static inline void clear_evtchn(unsigned port) { evtchn_ops->clear_pending(port); } static inline void set_evtchn(unsigned port) { evtchn_ops->set_pending(port); } static inline bool test_evtchn(unsigned port) { return evtchn_ops->is_pending(port); } static inline bool test_and_set_mask(unsigned port) { return evtchn_ops->test_and_set_mask(port); } static inline void mask_evtchn(unsigned port) { return evtchn_ops->mask(port); } static inline void unmask_evtchn(unsigned port) { return evtchn_ops->unmask(port); } static inline void xen_evtchn_handle_events(unsigned cpu) { return evtchn_ops->handle_events(cpu); } static inline void xen_evtchn_resume(void) { if (evtchn_ops->resume) evtchn_ops->resume(); } void xen_evtchn_2l_init(void); int xen_evtchn_fifo_init(void); #endif /* #ifndef __EVENTS_INTERNAL_H__ */ '/>context:space:mode:
authorAlexander Stein <alexander.stein@systec-electronic.com>2017-01-30 12:35:28 +0100
committerLinus Walleij <linus.walleij@linaro.org>2017-01-30 15:53:57 +0100
commitcdca06e4e85974d8a3503ab15709dbbaf90d3dd1 (patch)
tree268b0c01070a957006db97c88dccb5fb061440ff /net/9p/trans_common.h
parent1b89970d81bbd52720fc64a3fe9572ee33588363 (diff)
pinctrl: baytrail: Add missing spinlock usage in byt_gpio_irq_handler
According to VLI64 Intel Atom E3800 Specification Update (#329901) concurrent read accesses may result in returning 0xffffffff and write accesses may be dropped silently. To workaround all accesses must be protected by locks. Cc: stable@vger.kernel.org Signed-off-by: Alexander Stein <alexander.stein@systec-electronic.com> Acked-by: Mika Westerberg <mika.westerberg@linux.intel.com> Signed-off-by: Linus Walleij <linus.walleij@linaro.org>
Diffstat (limited to 'net/9p/trans_common.h')