/* * Basic general purpose allocator for managing special purpose memory * not managed by the regular kmalloc/kfree interface. * Uses for this includes on-device special memory, uncached memory * etc. * * This code is based on the buddy allocator found in the sym53c8xx_2 * driver Copyright (C) 1999-2001 Gerard Roudier , * and adapted for general purpose use. * * Copyright 2005 (C) Jes Sorensen * * This source code is licensed under the GNU General Public License, * Version 2. See the file COPYING for more details. */ #include #include #include #include #include #include #include #include #include #include struct gen_pool *gen_pool_create(int nr_chunks, int max_chunk_shift, unsigned long (*fp)(struct gen_pool *), unsigned long data) { struct gen_pool *poolp; unsigned long tmp; int i; /* * This is really an arbitrary limit, +10 is enough for * IA64_GRANULE_SHIFT, aka 16MB. If anyone needs a large limit * this can be increased without problems. */ if ((max_chunk_shift > (PAGE_SHIFT + 10)) || ((max_chunk_shift < ALLOC_MIN_SHIFT) && max_chunk_shift)) return NULL; if (!max_chunk_shift) max_chunk_shift = PAGE_SHIFT; poolp = kmalloc(sizeof(struct gen_pool), GFP_KERNEL); if (!poolp) return NULL; memset(poolp, 0, sizeof(struct gen_pool)); poolp->h = kmalloc(sizeof(struct gen_pool_link) * (max_chunk_shift - ALLOC_MIN_SHIFT + 1), GFP_KERNEL); if (!poolp->h) { printk(KERN_WARNING "gen_pool_alloc() failed to allocate\n"); kfree(poolp); return NULL; } memset(poolp->h, 0, sizeof(struct gen_pool_link) * (max_chunk_shift - ALLOC_MIN_SHIFT + 1)); spin_lock_init(&poolp->lock); poolp->get_new_chunk = fp; poolp->max_chunk_shift = max_chunk_shift; poolp->private = data; for (i = 0; i < nr_chunks; i++) { tmp = poolp->get_new_chunk(poolp); printk(KERN_INFO "allocated %lx\n", tmp); if (!tmp) break; gen_pool_free(poolp, tmp, (1 << poolp->max_chunk_shift)); } return poolp; } EXPORT_SYMBOL(gen_pool_create); /* * Simple power of two buddy-like generic allocator. * Provides naturally aligned memory chunks. */ unsigned long gen_pool_alloc(struct gen_pool *poolp, int size) { int j, i, s, max_chunk_size; unsigned long a, flags; struct gen_pool_link *h = poolp->h; max_chunk_size = 1 << poolp->max_chunk_shift; if (size > max_chunk_size) return 0; size = max(size, 1 << ALLOC_MIN_SHIFT); i = fls(size - 1); s = 1 << i; j = i -= ALLOC_MIN_SHIFT; spin_lock_irqsave(&poolp->lock, flags); while (!h[j].next) { if (s == max_chunk_size) { struct gen_pool_link *ptr; spin_unlock_irqrestore(&poolp->lock, flags); ptr = (struct gen_pool_link *)poolp->get_new_chunk(poolp); spin_lock_irqsave(&poolp->lock, flags); h[j].next = ptr; if (h[j].next) h[j].next->next = NULL; break; } j++; s <<= 1; } a = (unsigned long) h[j].next; if (a) { h[j].next = h[j].next->next; /* * This should be split into a seperate function doing * the chunk split in order to support custom * handling memory not physically accessible by host */ while (j > i) { j -= 1; s >>= 1; h[j].next = (struct gen_pool_link *) (a + s); h[j].next->next = NULL; } } spin_unlock_irqrestore(&poolp->lock, flags); return a; } EXPORT_SYMBOL(gen_pool_alloc); /* * Counter-part of the generic allocator. */ void gen_pool_free(struct gen_pool *poolp, unsigned long ptr, int size) { struct gen_pool_link *q; struct gen_pool_link *h = poolp->h; unsigned long a, b, flags; int i, s, max_chunk_size; max_chunk_size = 1 << poolp->max_chunk_shift; if (size > max_chunk_size) return; size = max(size, 1 << ALLOC_MIN_SHIFT); i = fls(size - 1); s = 1 << i; i -= ALLOC_MIN_SHIFT; a = ptr; spin_lock_irqsave(&poolp->lock, flags); while (1) { if (s == max_chunk_size) { ((struct gen_pool_link *)a)->next = h[i].next; h[i].next = (struct gen_pool_link *)a; break; } b = a ^ s; q = &h[i]; while (q->next && q->next != (struct gen_pool_link *)b) q = q->next; if (!q->next) { ((struct gen_pool_link *)a)->next = h[i].next; h[i].next = (struct gen_pool_link *)a; break; } q->next = q->next->next; a = a & b; s <<= 1; i++; } spin_unlock_irqrestore(&poolp->lock, flags); } EXPORT_SYMBOL(gen_pool_free);