/* * Copyright (c) 2016, Mellanox Technologies. All rights reserved. * * This software is available to you under a choice of one of two * licenses. You may choose to be licensed under the terms of the GNU * General Public License (GPL) Version 2, available from the file * COPYING in the main directory of this source tree, or the * OpenIB.org BSD license below: * * Redistribution and use in source and binary forms, with or * without modification, are permitted provided that the following * conditions are met: * * - Redistributions of source code must retain the above * copyright notice, this list of conditions and the following * disclaimer. * * - Redistributions in binary form must reproduce the above * copyright notice, this list of conditions and the following * disclaimer in the documentation and/or other materials * provided with the distribution. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE * SOFTWARE. */ #ifndef SOC_NPS_COMMON_H #define SOC_NPS_COMMON_H #ifdef CONFIG_SMP #define NPS_IPI_IRQ 5 #endif #define NPS_HOST_REG_BASE 0xF6000000 #define NPS_MSU_BLKID 0x018 #define CTOP_INST_RSPI_GIC_0_R12 0x3C56117E #define CTOP_INST_MOV2B_FLIP_R3_B1_B2_INST 0x5B60 #define CTOP_INST_MOV2B_FLIP_R3_B1_B2_LIMM 0x00010422 #ifndef __ASSEMBLY__ /* In order to increase compilation test coverage */ #ifdef CONFIG_ARC static inline void nps_ack_gic(void) { __asm__ __volatile__ ( " .word %0\n" : : "i"(CTOP_INST_RSPI_GIC_0_R12) : "memory"); } #else static inline void nps_ack_gic(void) { } #define write_aux_reg(r, v) #define read_aux_reg(r) 0 #endif /* CPU global ID */ struct global_id { union { struct { #ifdef CONFIG_EZNPS_MTM_EXT u32 __reserved:20, cluster:4, core:4, thread:4; #else u32 __reserved:24, cluster:4, core:4; #endif }; u32 value; }; }; /* * Convert logical to physical CPU IDs * * The conversion swap bits 1 and 2 of cluster id (out of 4 bits) * Now quad of logical clusters id's are adjacent physically, * and not like the id's physically came with each cluster. * Below table is 4x4 mesh of core clusters as it layout on chip. * Cluster ids are in format: logical (physical) * * ----------------- ------------------ * 3 | 5 (3) 7 (7) | | 13 (11) 15 (15)| * * 2 | 4 (2) 6 (6) | | 12 (10) 14 (14)| * ----------------- ------------------ * 1 | 1 (1) 3 (5) | | 9 (9) 11 (13)| * * 0 | 0 (0) 2 (4) | | 8 (8) 10 (12)| * ----------------- ------------------ * 0 1 2 3 */ static inline int nps_cluster_logic_to_phys(int cluster) { #ifdef __arc__ __asm__ __volatile__( " mov r3,%0\n" " .short %1\n" " .word %2\n" " mov %0,r3\n" : "+r"(cluster) : "i"(CTOP_INST_MOV2B_FLIP_R3_B1_B2_INST), "i"(CTOP_INST_MOV2B_FLIP_R3_B1_B2_LIMM) : "r3"); #endif return cluster; } #define NPS_CPU_TO_CLUSTER_NUM(cpu) \ ({ struct global_id gid; gid.value = cpu; \ nps_cluster_logic_to_phys(gid.cluster); }) struct nps_host_reg_address { union { struct { u32 base:8, cl_x:4, cl_y:4, blkid:6, reg:8, __reserved:2; }; u32 value; }; }; struct nps_host_reg_address_non_cl { union { struct { u32 base:7, blkid:11, reg:12, __reserved:2; }; u32 value; }; }; static inline void *nps_host_reg_non_cl(u32 blkid, u32 reg) { struct nps_host_reg_address_non_cl reg_address; reg_address.value = NPS_HOST_REG_BASE; reg_address.blkid = blkid; reg_address.reg = reg; return (void *)reg_address.value; } static inline void *nps_host_reg(u32 cpu, u32 blkid, u32 reg) { struct nps_host_reg_address reg_address; u32 cl = NPS_CPU_TO_CLUSTER_NUM(cpu); reg_address.value = NPS_HOST_REG_BASE; reg_address.cl_x = (cl >> 2) & 0x3; reg_address.cl_y = cl & 0x3; reg_address.blkid = blkid; reg_address.reg = reg; return (void *)reg_address.value; } #endif /* __ASSEMBLY__ */ #endif /* SOC_NPS_COMMON_H */ lue='reload'/>
authorzhong jiang <zhongjiang@huawei.com>2017-01-24 15:18:52 -0800
committerLinus Torvalds <torvalds@linux-foundation.org>2017-01-24 16:26:14 -0800
commit3277953de2f31dd03c6375e9a9f680ac37fc9d27 (patch)
tree6ba22924faa1b82ef2f0277db10a3f5abdc44157 /include/net/sctp
parent3705ccfdd1e8b539225ce20e3925a945cc788d67 (diff)
mm: do not export ioremap_page_range symbol for external module
Recently, I've found cases in which ioremap_page_range was used incorrectly, in external modules, leading to crashes. This can be partly attributed to the fact that ioremap_page_range is lower-level, with fewer protections, as compared to the other functions that an external module would typically call. Those include: ioremap_cache ioremap_nocache ioremap_prot ioremap_uc ioremap_wc ioremap_wt ...each of which wraps __ioremap_caller, which in turn provides a safer way to achieve the mapping. Therefore, stop EXPORT-ing ioremap_page_range. Link: http://lkml.kernel.org/r/1485173220-29010-1-git-send-email-zhongjiang@huawei.com Signed-off-by: zhong jiang <zhongjiang@huawei.com> Reviewed-by: John Hubbard <jhubbard@nvidia.com> Suggested-by: John Hubbard <jhubbard@nvidia.com> Acked-by: Michal Hocko <mhocko@suse.com> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
Diffstat (limited to 'include/net/sctp')