#define _GNU_SOURCE #include #include #include #include #include "locking.h" #include "cpusched.h" #include "xmalloc.h" #include "hash.h" struct map_entry { int fd; unsigned int cpu; struct map_entry *next; }; static struct hash_table mapper; static unsigned int *cpu_work_map = NULL; static unsigned int cpu_len = 0; static struct rwlock map_lock; static unsigned int get_appropriate_cpu(void) { unsigned int i, cpu = 0; unsigned int work = UINT_MAX; for (i = 0; i < cpu_len; ++i) { if (cpu_work_map[i] < work) { work = cpu_work_map[i]; cpu = i; } } return cpu; } unsigned int socket_to_cpu(int fd) { int cpu = 0; struct map_entry *entry; rwlock_rd_lock(&map_lock); entry = lookup_hash(fd, &mapper); while (entry && fd != entry->fd) entry = entry->next; if (entry && fd == entry->fd) cpu = entry->cpu; rwlock_unlock(&map_lock); return cpu; } unsigned int register_socket(int fd) { void **pos; struct map_entry *entry; rwlock_wr_lock(&map_lock); entry = xzmalloc(sizeof(*entry)); entry->fd = fd; entry->cpu = get_appropriate_cpu(); cpu_work_map[entry->cpu]++; pos = insert_hash(entry->fd, entry, &mapper); if (pos) { entry->next = (*pos); (*pos) = entry; } rwlock_unlock(&map_lock); return entry->cpu; } static struct map_entry *socket_to_map_entry(int fd) { struct map_entry *entry, *ret = NULL; rwlock_rd_lock(&map_lock); entry = lookup_hash(fd, &mapper); while (entry && fd != entry->fd) entry = entry->next; if (entry && fd == entry->fd) ret = entry; rwlock_unlock(&map_lock); return ret; } void unregister_socket(int fd) { struct map_entry *pos; struct map_entry *entry = socket_to_map_entry(fd); if (entry == NULL) return; rwlock_wr_lock(&map_lock); cpu_work_map[entry->cpu]--; pos = remove_hash(entry->fd, entry, entry->next, &mapper); while (pos && pos->next && pos->next != entry) pos = pos->next; if (pos && pos->next && pos->next == entry) pos->next = entry->next; entry->next = NULL; xfree(entry); rwlock_unlock(&map_lock); } static int cleanup_cpusched_batch(void *ptr) { struct map_entry *next; struct map_entry *entry = ptr; if (!entry) return 0; while ((next = entry->next)) { entry->next = NULL; xfree(entry); entry = next; } xfree(entry); return 0; } void init_cpusched(unsigned int cpus) { rwlock_init(&map_lock); cpu_len = cpus; cpu_work_map = xcalloc(cpu_len, sizeof(*cpu_work_map)); init_hash(&mapper); } void destroy_cpusched(void) { xfree(cpu_work_map); for_each_hash(&mapper, cleanup_cpusched_batch); free_hash(&mapper); rwlock_destroy(&map_lock); } >
diff options
context:
space:
mode:
authorMatias Bjørling <m@bjorling.me>2016-05-06 20:03:07 +0200
committerJens Axboe <axboe@fb.com>2016-05-06 12:51:10 -0600
commit5ebc7d9fe13ff9bd3622d0be3cd39c8751459be6 (patch)
treedfca4e0b1029af8fcf8fe52ce6de69a4eb7bcdb0 /include/linux/lightnvm.h
parent6659d4d80c6fc6b4b5bc2e99988f32b1b3865f1d (diff)
lightnvm: make nvm_set_rqd_ppalist() aware of vblks
A virtual block enables a block to identify multiple physical blocks. This is useful for metadata where a device media supports multiple planes. In that case, a block, with multiple planes can be managed as a single vblk. Reducing the metadata required by one forth. nvm_set_rqd_ppalist() takes care of expanding a ppa_list with vblks automatically. However, for some use-cases, where only a single physical block is required, the ppa_list should not be expanded. Therefore, add a vblk parameter to nvm_set_rqd_ppalist(), and only expand the ppa_list if vblk is set. Signed-off-by: Matias Bjørling <m@bjorling.me> Signed-off-by: Jens Axboe <axboe@fb.com>
Diffstat (limited to 'include/linux/lightnvm.h')
-rw-r--r--include/linux/lightnvm.h2
1 files changed, 1 insertions, 1 deletions
diff --git a/include/linux/lightnvm.h b/include/linux/lightnvm.h