/* * CMA DebugFS Interface * * Copyright (c) 2015 Sasha Levin */ #include #include #include #include #include #include #include "cma.h" struct cma_mem { struct hlist_node node; struct page *p; unsigned long n; }; static struct dentry *cma_debugfs_root; static int cma_debugfs_get(void *data, u64 *val) { unsigned long *p = data; *val = *p; return 0; } DEFINE_SIMPLE_ATTRIBUTE(cma_debugfs_fops, cma_debugfs_get, NULL, "%llu\n"); static int cma_used_get(void *data, u64 *val) { struct cma *cma = data; unsigned long used; mutex_lock(&cma->lock); /* pages counter is smaller than sizeof(int) */ used = bitmap_weight(cma->bitmap, (int)cma_bitmap_maxno(cma)); mutex_unlock(&cma->lock); *val = (u64)used << cma->order_per_bit; return 0; } DEFINE_SIMPLE_ATTRIBUTE(cma_used_fops, cma_used_get, NULL, "%llu\n"); static int cma_maxchunk_get(void *data, u64 *val) { struct cma *cma = data; unsigned long maxchunk = 0; unsigned long start, end = 0; unsigned long bitmap_maxno = cma_bitmap_maxno(cma); mutex_lock(&cma->lock); for (;;) { start = find_next_zero_bit(cma->bitmap, bitmap_maxno, end); if (start >= cma->count) break; end = find_next_bit(cma->bitmap, bitmap_maxno, start); maxchunk = max(end - start, maxchunk); } mutex_unlock(&cma->lock); *val = (u64)maxchunk << cma->order_per_bit; return 0; } DEFINE_SIMPLE_ATTRIBUTE(cma_maxchunk_fops, cma_maxchunk_get, NULL, "%llu\n"); static void cma_add_to_cma_mem_list(struct cma *cma, struct cma_mem *mem) { spin_lock(&cma->mem_head_lock); hlist_add_head(&mem->node, &cma->mem_head); spin_unlock(&cma->mem_head_lock); } static struct cma_mem *cma_get_entry_from_list(struct cma *cma) { struct cma_mem *mem = NULL; spin_lock(&cma->mem_head_lock); if (!hlist_empty(&cma->mem_head)) { mem = hlist_entry(cma->mem_head.first, struct cma_mem, node); hlist_del_init(&mem->node); } spin_unlock(&cma->mem_head_lock); return mem; } static int cma_free_mem(struct cma *cma, int count) { struct cma_mem *mem = NULL; while (count) { mem = cma_get_entry_from_list(cma); if (mem == NULL) return 0; if (mem->n <= count) { cma_release(cma, mem->p, mem->n); count -= mem->n; kfree(mem); } else if (cma->order_per_bit == 0) { cma_release(cma, mem->p, count); mem->p += count; mem->n -= count; count = 0; cma_add_to_cma_mem_list(cma, mem); } else { pr_debug("cma: cannot release partial block when order_per_bit != 0\n"); cma_add_to_cma_mem_list(cma, mem); break; } } return 0; } static int cma_free_write(void *data, u64 val) { int pages = val; struct cma *cma = data; return cma_free_mem(cma, pages); } DEFINE_SIMPLE_ATTRIBUTE(cma_free_fops, NULL, cma_free_write, "%llu\n"); static int cma_alloc_mem(struct cma *cma, int count) { struct cma_mem *mem; struct page *p; mem = kzalloc(sizeof(*mem), GFP_KERNEL); if (!mem) return -ENOMEM; p = cma_alloc(cma, count, 0); if (!p) { kfree(mem); return -ENOMEM; } mem->p = p; mem->n = count; cma_add_to_cma_mem_list(cma, mem); return 0; } static int cma_alloc_write(void *data, u64 val) { int pages = val; struct cma *cma = data; return cma_alloc_mem(cma, pages); } DEFINE_SIMPLE_ATTRIBUTE(cma_alloc_fops, NULL, cma_alloc_write, "%llu\n"); static void cma_debugfs_add_one(struct cma *cma, int idx) { struct dentry *tmp; char name[16]; int u32s; sprintf(name, "cma-%d", idx); tmp = debugfs_create_dir(name, cma_debugfs_root); debugfs_create_file("alloc", S_IWUSR, tmp, cma, &cma_alloc_fops); debugfs_create_file("free", S_IWUSR, tmp, cma, &cma_free_fops); debugfs_create_file("base_pfn", S_IRUGO, tmp, &cma->base_pfn, &cma_debugfs_fops); debugfs_create_file("count", S_IRUGO, tmp, &cma->count, &cma_debugfs_fops); debugfs_create_file("order_per_bit", S_IRUGO, tmp, &cma->order_per_bit, &cma_debugfs_fops); debugfs_create_file("used", S_IRUGO, tmp, cma, &cma_used_fops); debugfs_create_file("maxchunk", S_IRUGO, tmp, cma, &cma_maxchunk_fops); u32s = DIV_ROUND_UP(cma_bitmap_maxno(cma), BITS_PER_BYTE * sizeof(u32)); debugfs_create_u32_array("bitmap", S_IRUGO, tmp, (u32*)cma->bitmap, u32s); } static int __init cma_debugfs_init(void) { int i; cma_debugfs_root = debugfs_create_dir("cma", NULL); if (!cma_debugfs_root) return -ENOMEM; for (i = 0; i < cma_area_count; i++) cma_debugfs_add_one(&cma_areas[i], i); return 0; } late_initcall(cma_debugfs_init); >space:mode:
authorJack Morgenstein <jackm@dev.mellanox.co.il>2017-01-15 20:15:00 +0200
committerDoug Ledford <dledford@redhat.com>2017-01-27 14:29:04 -0500
commitb4cfe3971f6eab542dd7ecc398bfa1aeec889934 (patch)
treec7ad49d05da0535170c8e7710cd44ae1cecc271f /tools/include/asm/barrier.h
parent2d4b21e0a2913612274a69a3ba1bfee4cffc6e77 (diff)
RDMA/cma: Fix unknown symbol when CONFIG_IPV6 is not enabled
If IPV6 has not been enabled in the underlying kernel, we must avoid calling IPV6 procedures in rdma_cm.ko. This requires using "IS_ENABLED(CONFIG_IPV6)" in "if" statements surrounding any code which calls external IPV6 procedures. In the instance fixed here, procedure cma_bind_addr() called ipv6_addr_type() -- which resulted in calling external procedure __ipv6_addr_type(). Fixes: 6c26a77124ff ("RDMA/cma: fix IPv6 address resolution") Cc: <stable@vger.kernel.org> # v4.2+ Cc: Spencer Baugh <sbaugh@catern.com> Signed-off-by: Jack Morgenstein <jackm@dev.mellanox.co.il> Reviewed-by: Moni Shoua <monis@mellanox.com> Signed-off-by: Leon Romanovsky <leon@kernel.org> Signed-off-by: Doug Ledford <dledford@redhat.com>
Diffstat (limited to 'tools/include/asm/barrier.h')