/* * CMA DebugFS Interface * * Copyright (c) 2015 Sasha Levin */ #include #include #include #include #include #include #include "cma.h" struct cma_mem { struct hlist_node node; struct page *p; unsigned long n; }; static struct dentry *cma_debugfs_root; static int cma_debugfs_get(void *data, u64 *val) { unsigned long *p = data; *val = *p; return 0; } DEFINE_SIMPLE_ATTRIBUTE(cma_debugfs_fops, cma_debugfs_get, NULL, "%llu\n"); static int cma_used_get(void *data, u64 *val) { struct cma *cma = data; unsigned long used; mutex_lock(&cma->lock); /* pages counter is smaller than sizeof(int) */ used = bitmap_weight(cma->bitmap, (int)cma_bitmap_maxno(cma)); mutex_unlock(&cma->lock); *val = (u64)used << cma->order_per_bit; return 0; } DEFINE_SIMPLE_ATTRIBUTE(cma_used_fops, cma_used_get, NULL, "%llu\n"); static int cma_maxchunk_get(void *data, u64 *val) { struct cma *cma = data; unsigned long maxchunk = 0; unsigned long start, end = 0; unsigned long bitmap_maxno = cma_bitmap_maxno(cma); mutex_lock(&cma->lock); for (;;) { start = find_next_zero_bit(cma->bitmap, bitmap_maxno, end); if (start >= cma->count) break; end = find_next_bit(cma->bitmap, bitmap_maxno, start); maxchunk = max(end - start, maxchunk); } mutex_unlock(&cma->lock); *val = (u64)maxchunk << cma->order_per_bit; return 0; } DEFINE_SIMPLE_ATTRIBUTE(cma_maxchunk_fops, cma_maxchunk_get, NULL, "%llu\n"); static void cma_add_to_cma_mem_list(struct cma *cma, struct cma_mem *mem) { spin_lock(&cma->mem_head_lock); hlist_add_head(&mem->node, &cma->mem_head); spin_unlock(&cma->mem_head_lock); } static struct cma_mem *cma_get_entry_from_list(struct cma *cma) { struct cma_mem *mem = NULL; spin_lock(&cma->mem_head_lock); if (!hlist_empty(&cma->mem_head)) { mem = hlist_entry(cma->mem_head.first, struct cma_mem, node); hlist_del_init(&mem->node); } spin_unlock(&cma->mem_head_lock); return mem; } static int cma_free_mem(struct cma *cma, int count) { struct cma_mem *mem = NULL; while (count) { mem = cma_get_entry_from_list(cma); if (mem == NULL) return 0; if (mem->n <= count) { cma_release(cma, mem->p, mem->n); count -= mem->n; kfree(mem); } else if (cma->order_per_bit == 0) { cma_release(cma, mem->p, count); mem->p += count; mem->n -= count; count = 0; cma_add_to_cma_mem_list(cma, mem); } else { pr_debug("cma: cannot release partial block when order_per_bit != 0\n"); cma_add_to_cma_mem_list(cma, mem); break; } } return 0; } static int cma_free_write(void *data, u64 val) { int pages = val; struct cma *cma = data; return cma_free_mem(cma, pages); } DEFINE_SIMPLE_ATTRIBUTE(cma_free_fops, NULL, cma_free_write, "%llu\n"); static int cma_alloc_mem(struct cma *cma, int count) { struct cma_mem *mem; struct page *p; mem = kzalloc(sizeof(*mem), GFP_KERNEL); if (!mem) return -ENOMEM; p = cma_alloc(cma, count, 0); if (!p) { kfree(mem); return -ENOMEM; } mem->p = p; mem->n = count; cma_add_to_cma_mem_list(cma, mem); return 0; } static int cma_alloc_write(void *data, u64 val) { int pages = val; struct cma *cma = data; return cma_alloc_mem(cma, pages); } DEFINE_SIMPLE_ATTRIBUTE(cma_alloc_fops, NULL, cma_alloc_write, "%llu\n"); static void cma_debugfs_add_one(struct cma *cma, int idx) { struct dentry *tmp; char name[16]; int u32s; sprintf(name, "cma-%d", idx); tmp = debugfs_create_dir(name, cma_debugfs_root); debugfs_create_file("alloc", S_IWUSR, tmp, cma, &cma_alloc_fops); debugfs_create_file("free", S_IWUSR, tmp, cma, &cma_free_fops); debugfs_create_file("base_pfn", S_IRUGO, tmp, &cma->base_pfn, &cma_debugfs_fops); debugfs_create_file("count", S_IRUGO, tmp, &cma->count, &cma_debugfs_fops); debugfs_create_file("order_per_bit", S_IRUGO, tmp, &cma->order_per_bit, &cma_debugfs_fops); debugfs_create_file("used", S_IRUGO, tmp, cma, &cma_used_fops); debugfs_create_file("maxchunk", S_IRUGO, tmp, cma, &cma_maxchunk_fops); u32s = DIV_ROUND_UP(cma_bitmap_maxno(cma), BITS_PER_BYTE * sizeof(u32)); debugfs_create_u32_array("bitmap", S_IRUGO, tmp, (u32*)cma->bitmap, u32s); } static int __init cma_debugfs_init(void) { int i; cma_debugfs_root = debugfs_create_dir("cma", NULL); if (!cma_debugfs_root) return -ENOMEM; for (i = 0; i < cma_area_count; i++) cma_debugfs_add_one(&cma_areas[i], i); return 0; } late_initcall(cma_debugfs_init); >30space:mode:
authorFeng <fgao@ikuai8.com>2017-01-20 21:40:43 +0800
committerPablo Neira Ayuso <pablo@netfilter.org>2017-02-02 14:30:19 +0100
commit10435c1192d06bdb0bac7666452d8219d7e7c477 (patch)
tree93b76419142fe17b1d162d062c663297a3e8a965 /include/dt-bindings/clock/bcm281xx.h
parent1a28ad74ebd8f9d3c7eae0d781f72a6d30545e17 (diff)
netfilter: nf_tables: Eliminate duplicated code in nf_tables_table_enable()
If something fails in nf_tables_table_enable(), it unregisters the chains. But the rollback code is the same as nf_tables_table_disable() almostly, except there is one counter check. Now create one wrapper function to eliminate the duplicated codes. Signed-off-by: Feng <fgao@ikuai8.com> Signed-off-by: Pablo Neira Ayuso <pablo@netfilter.org>
Diffstat (limited to 'include/dt-bindings/clock/bcm281xx.h')