/* * soc-cache.c -- ASoC register cache helpers * * Copyright 2009 Wolfson Microelectronics PLC. * * Author: Mark Brown * * This program is free software; you can redistribute it and/or modify it * under the terms of the GNU General Public License as published by the * Free Software Foundation; either version 2 of the License, or (at your * option) any later version. */ #include #include #include #include #include #include #include static bool snd_soc_set_cache_val(void *base, unsigned int idx, unsigned int val, unsigned int word_size) { switch (word_size) { case 1: { u8 *cache = base; if (cache[idx] == val) return true; cache[idx] = val; break; } case 2: { u16 *cache = base; if (cache[idx] == val) return true; cache[idx] = val; break; } default: BUG(); } return false; } static unsigned int snd_soc_get_cache_val(const void *base, unsigned int idx, unsigned int word_size) { if (!base) return -1; switch (word_size) { case 1: { const u8 *cache = base; return cache[idx]; } case 2: { const u16 *cache = base; return cache[idx]; } default: BUG(); } /* unreachable */ return -1; } struct snd_soc_rbtree_node { struct rb_node node; /* the actual rbtree node holding this block */ unsigned int base_reg; /* base register handled by this block */ unsigned int word_size; /* number of bytes needed to represent the register index */ void *block; /* block of adjacent registers */ unsigned int blklen; /* number of registers available in the block */ } __attribute__ ((packed)); struct snd_soc_rbtree_ctx { struct rb_root root; struct snd_soc_rbtree_node *cached_rbnode; }; static inline void snd_soc_rbtree_get_base_top_reg( struct snd_soc_rbtree_node *rbnode, unsigned int *base, unsigned int *top) { *base = rbnode->base_reg; *top = rbnode->base_reg + rbnode->blklen - 1; } static unsigned int snd_soc_rbtree_get_register( struct snd_soc_rbtree_node *rbnode, unsigned int idx) { unsigned int val; switch (rbnode->word_size) { case 1: { u8 *p = rbnode->block; val = p[idx]; return val; } case 2: { u16 *p = rbnode->block; val = p[idx]; return val; } default: BUG(); break; } return -1; } static void snd_soc_rbtree_set_register(struct snd_soc_rbtree_node *rbnode, unsigned int idx, unsigned int val) { switch (rbnode->word_size) { case 1: { u8 *p = rbnode->block; p[idx] = val; break; } case 2: { u16 *p = rbnode->block; p[idx] = val; break; } default: BUG(); break; } } static struct snd_soc_rbtree_node *snd_soc_rbtree_lookup( struct rb_root *root, unsigned int reg) { struct rb_node *node; struct snd_soc_rbtree_node *rbnode; unsigned int base_reg, top_reg; node = root->rb_node; while (node) { rbnode = container_of(node, struct snd_soc_rbtree_node, node); snd_soc_rbtree_get_base_top_reg(rbnode, &base_reg, &top_reg); if (reg >= base_reg && reg <= top_reg) return rbnode; else if (reg > top_reg) node = node->rb_right; else if (reg < base_reg) node = node->rb_left; } return NULL; } static int snd_soc_rbtree_insert(struct rb_root *root, struct snd_soc_rbtree_node *rbnode) { struct rb_node **new, *parent; struct snd_soc_rbtree_node *rbnode_tmp; unsigned int base_reg_tmp, top_reg_tmp; unsigned int base_reg; parent = NULL; new = &root->rb_node; while (*new) { rbnode_tmp = container_of(*new, struct snd_soc_rbtree_node, node); /* base and top registers of the current rbnode */ snd_soc_rbtree_get_base_top_reg(rbnode_tmp, &base_reg_tmp, &top_reg_tmp); /* base register of the rbnode to be added */ base_reg = rbnode->base_reg; parent = *new; /* if this register has already been inserted, just return */ if (base_reg >= base_reg_tmp && base_reg <= top_reg_tmp) return 0; else if (base_reg > top_reg_tmp) new = &((*new)->rb_right); else if (base_reg < base_reg_tmp) new = &((*new)->rb_left); } /* insert the node into the rbtree */ rb_link_node(&rbnode->node, parent, new); rb_insert_color(&rbnode->node, root); return 1; } static int snd_soc_rbtree_cache_sync(struct snd_soc_codec *codec) { struct snd_soc_rbtree_ctx *rbtree_ctx; struct rb_node *node; struct snd_soc_rbtree_node *rbnode; unsigned int regtmp; unsigned int val, def; int ret; int i; rbtree_ctx = codec->reg_cache; for (node = rb_first(&rbtree_ctx->root); node; node = rb_next(node)) { rbnode = rb_entry(node, struct snd_soc_rbtree_node, node); for (i = 0; i < rbnode->blklen; ++i) { regtmp = rbnode->base_reg + i; val = snd_soc_rbtree_get_register(rbnode, i); def = snd_soc_get_cache_val(codec->reg_def_copy, i, rbnode->word_size); if (val == def) continue; WARN_ON(!snd_soc_codec_writable_register(codec, regtmp)); codec->cache_bypass = 1; ret = snd_soc_write(codec, regtmp, val); codec->cache_bypass = 0; if (ret) return ret; dev_dbg(codec->dev, "Synced register %#x, value = %#x\n", regtmp, val); } } return 0; } static int snd_soc_rbtree_insert_to_block(struct snd_soc_rbtree_node *rbnode, unsigned int pos, unsigned int reg, unsigned int value) { u8 *blk; blk = krealloc(rbnode->block, (rbnode->blklen + 1) * rbnode->word_size, GFP_KERNEL); if (!blk) return -ENOMEM; /* insert the register value in the correct place in the rbnode block */ memmove(blk + (pos + 1) * rbnode->word_size, blk + pos * rbnode->word_size, (rbnode->blklen - pos) * rbnode->word_size); /* update the rbnode block, its size and the base register */ rbnode->block = blk; rbnode->blklen++; if (!pos) rbnode->base_reg = reg; snd_soc_rbtree_set_register(rbnode, pos, value); return 0; } static int snd_soc_rbtree_cache_write(struct snd_soc_codec *codec, unsigned int reg, unsigned int value) { struct snd_soc_rbtree_ctx *rbtree_ctx; struct snd_soc_rbtree_node *rbnode, *rbnode_tmp; struct rb_node *node; unsigned int val; unsigned int reg_tmp; unsigned int base_reg, top_reg; unsigned int pos; int i; int ret; rbtree_ctx = codec->reg_cache; /* look up the required register in the cached rbnode */ rbnode = rbtree_ctx->cached_rbnode; if (rbnode) { snd_soc_rbtree_get_base_top_reg(rbnode, &base_reg, &top_reg); if (reg >= base_reg && reg <= top_reg) { reg_tmp = reg - base_reg; val = snd_soc_rbtree_get_register(rbnode, reg_tmp); if (val == value) return 0; snd_soc_rbtree_set_register(rbnode, reg_tmp, value); return 0; } } /* if we can't locate it in the cached rbnode we'll have * to traverse the rbtree looking for it. */ rbnode = snd_soc_rbtree_lookup(&rbtree_ctx->root, reg); if (rbnode) { reg_tmp = reg - rbnode->base_reg; val = snd_soc_rbtree_get_register(rbnode, reg_tmp); if (val == value) return 0; snd_soc_rbtree_set_register(rbnode, reg_tmp, value); rbtree_ctx->cached_rbnode = rbnode; } else { /* bail out early, no need to create the rbnode yet */ if (!value) return 0; /* look for an adjacent register to the one we are about to add */ for (node = rb_first(&rbtree_ctx->root); node; node = rb_next(node)) { rbnode_tmp = rb_entry(node, struct snd_soc_rbtree_node, node); for (i = 0; i < rbnode_tmp->blklen; ++i) { reg_tmp = rbnode_tmp->base_reg + i; if (abs(reg_tmp - reg) != 1) continue; /* decide where in the block to place our register */ if (reg_tmp + 1 == reg) pos = i + 1; else pos = i; ret = snd_soc_rbtree_insert_to_block(rbnode_tmp, pos, reg, value); if (ret) return ret; rbtree_ctx->cached_rbnode = rbnode_tmp; return 0; } } /* we did not manage to find a place to insert it in an existing * block so create a new rbnode with a single register in its block. * This block will get populated further if any other adjacent * registers get modified in the future. */ rbnode = kzalloc(sizeof *rbnode, GFP_KERNEL); if (!rbnode) return -ENOMEM; rbnode->blklen = 1; rbnode->base_reg = reg; rbnode->word_size = codec->driver->reg_word_size; rbnode->block = kmalloc(rbnode->blklen * rbnode->word_size, GFP_KERNEL); if (!rbnode->block) { kfree(rbnode); return -ENOMEM; } snd_soc_rbtree_set_register(rbnode, 0, value); snd_soc_rbtree_insert(&rbtree_ctx->root, rbnode); rbtree_ctx->cached_rbnode = rbnode; } return 0; } static int snd_soc_rbtree_cache_read(struct snd_soc_codec *codec, unsigned int reg, unsigned int *value) { struct snd_soc_rbtree_ctx *rbtree_ctx; struct snd_soc_rbtree_node *rbnode; unsigned int base_reg, top_reg; unsigned int reg_tmp; rbtree_ctx = codec->reg_cache; /* look up the required register in the cached rbnode */ rbnode = rbtree_ctx->cached_rbnode; if (rbnode) { snd_soc_rbtree_get_base_top_reg(rbnode, &base_reg, &top_reg); if (reg >= base_reg && reg <= top_reg) { reg_tmp = reg - base_reg; *value = snd_soc_rbtree_get_register(rbnode, reg_tmp); return 0; } } /* if we can't locate it in the cached rbnode we'll have * to traverse the rbtree looking for it. */ rbnode = snd_soc_rbtree_lookup(&rbtree_ctx->root, reg); if (rbnode) { reg_tmp = reg - rbnode->base_reg; *value = snd_soc_rbtree_get_register(rbnode, reg_tmp); rbtree_ctx->cached_rbnode = rbnode; } else { /* uninitialized registers default to 0 */ *value = 0; } return 0; } static int snd_soc_rbtree_cache_exit(struct snd_soc_codec *codec) { struct rb_node *next; struct snd_soc_rbtree_ctx *rbtree_ctx; struct snd_soc_rbtree_node *rbtree_node; /* if we've already been called then just return */ rbtree_ctx = codec->reg_cache; if (!rbtree_ctx) return 0; /* free up the rbtree */ next = rb_first(&rbtree_ctx->root); while (next) { rbtree_node = rb_entry(next, struct snd_soc_rbtree_node, node); next = rb_next(&rbtree_node->node); rb_erase(&rbtree_node->node, &rbtree_ctx->root); kfree(rbtree_node->block); kfree(rbtree_node); } /* release the resources */ kfree(codec->reg_cache); codec->reg_cache = NULL; return 0; } static int snd_soc_rbtree_cache_init(struct snd_soc_codec *codec) { struct snd_soc_rbtree_ctx *rbtree_ctx; unsigned int word_size; unsigned int val; int i; int ret; codec->reg_cache = kmalloc(sizeof *rbtree_ctx, GFP_KERNEL); if (!codec->reg_cache) return -ENOMEM; rbtree_ctx = codec->reg_cache; rbtree_ctx->root = RB_ROOT; rbtree_ctx->cached_rbnode = NULL; if (!codec->reg_def_copy) return 0; word_size = codec->driver->reg_word_size; for (i = 0; i < codec->driver->reg_cache_size; ++i) { val = snd_soc_get_cache_val(codec->reg_def_copy, i, word_size); if (!val) continue; ret = snd_soc_rbtree_cache_write(codec, i, val); if (ret) goto err; } return 0; err: snd_soc_cache_exit(codec); return ret; } #ifdef CONFIG_SND_SOC_CACHE_LZO struct snd_soc_lzo_ctx { void *wmem; void *dst; const void *src; size_t src_len; size_t dst_len; size_t decompressed_size; unsigned long *sync_bmp; int sync_bmp_nbits; }; #define LZO_BLOCK_NUM 8 static int snd_soc_lzo_block_count(void) { return LZO_BLOCK_NUM; } static int snd_soc_lzo_prepare(struct snd_soc_lzo_ctx *lzo_ctx) { lzo_ctx->wmem = kmalloc(LZO1X_MEM_COMPRESS, GFP_KERNEL); if (!lzo_ctx->wmem) return -ENOMEM; return 0; } static int snd_soc_lzo_compress(struct snd_soc_lzo_ctx *lzo_ctx) { size_t compress_size; int ret; ret = lzo1x_1_compress(lzo_ctx->src, lzo_ctx->src_len, lzo_ctx->dst, &compress_size, lzo_ctx->wmem); if (ret != LZO_E_OK || compress_size > lzo_ctx->dst_len) return -EINVAL; lzo_ctx->dst_len = compress_size; return 0; } static int snd_soc_lzo_decompress(struct snd_soc_lzo_ctx *lzo_ctx) { size_t dst_len; int ret; dst_len = lzo_ctx->dst_len; ret = lzo1x_decompress_safe(lzo_ctx->src, lzo_ctx->src_len, lzo_ctx->dst, &dst_len); if (ret != LZO_E_OK || dst_len != lzo_ctx->dst_len) return -EINVAL; return 0; } static int snd_soc_lzo_compress_cache_block(struct snd_soc_codec *codec, struct snd_soc_lzo_ctx *lzo_ctx) { int ret; lzo_ctx->dst_len = lzo1x_worst_compress(PAGE_SIZE); lzo_ctx->dst = kmalloc(lzo_ctx->dst_len, GFP_KERNEL); if (!lzo_ctx->dst) { lzo_ctx->dst_len = 0; return -ENOMEM; } ret = snd_soc_lzo_compress(lzo_ctx); if (ret < 0) return ret; return 0; } static int snd_soc_lzo_decompress_cache_block(struct snd_soc_codec *codec, struct snd_soc_lzo_ctx *lzo_ctx) { int ret; lzo_ctx->dst_len = lzo_ctx->decompressed_size; lzo_ctx->dst = kmalloc(lzo_ctx->dst_len, GFP_KERNEL); if (!lzo_ctx->dst) { lzo_ctx->dst_len = 0; return -ENOMEM; } ret = snd_soc_lzo_decompress(lzo_ctx); if (ret < 0) return ret; return 0; } static inline int snd_soc_lzo_get_blkindex(struct snd_soc_codec *codec, unsigned int reg) { const struct snd_soc_codec_driver *codec_drv; codec_drv = codec->driver; return (reg * codec_drv->reg_word_size) / DIV_ROUND_UP(codec->reg_size, snd_soc_lzo_block_count()); } static inline int snd_soc_lzo_get_blkpos(struct snd_soc_codec *codec, unsigned int reg) { const struct snd_soc_codec_driver *codec_drv; codec_drv = codec->driver; return reg % (DIV_ROUND_UP(codec->reg_size, snd_soc_lzo_block_count()) / codec_drv->reg_word_size); } static inline int snd_soc_lzo_get_blksize(struct snd_soc_codec *codec) { const struct snd_soc_codec_driver *codec_drv; codec_drv = codec->driver; return DIV_ROUND_UP(codec->reg_size, snd_soc_lzo_block_count()); } static int snd_soc_lzo_cache_sync(struct snd_soc_codec *codec) { struct snd_soc_lzo_ctx **lzo_blocks; unsigned int val; int i; int ret; lzo_blocks = codec->reg_cache; for_each_set_bit(i, lzo_blocks[0]->sync_bmp, lzo_blocks[0]->sync_bmp_nbits) { WARN_ON(!snd_soc_codec_writable_register(codec, i)); ret = snd_soc_cache_read(codec, i, &val); if (ret) return ret; codec->cache_bypass = 1; ret = snd_soc_write(codec, i, val); codec->cache_bypass = 0; if (ret) return ret; dev_dbg(codec->dev, "Synced register %#x, value = %#x\n", i, val); } return 0; } static int snd_soc_lzo_cache_write(struct snd_soc_codec *codec, unsigned int reg, unsigned int value) { struct snd_soc_lzo_ctx *lzo_block, **lzo_blocks; int ret, blkindex, blkpos; size_t blksize, tmp_dst_len; void *tmp_dst; /* index of the compressed lzo block */ blkindex = snd_soc_lzo_get_blkindex(codec, reg); /* register index within the decompressed block */ blkpos = snd_soc_lzo_get_blkpos(codec, reg); /* size of the compressed block */ blksize = snd_soc_lzo_get_blksize(codec); lzo_blocks = codec->reg_cache; lzo_block = lzo_blocks[blkindex]; /* save the pointer and length of the compressed block */ tmp_dst = lzo_block->dst; tmp_dst_len = lzo_block->dst_len; /* prepare the source to be the compressed block */ lzo_block->src = lzo_block->dst; lzo_block->src_len = lzo_block->dst_len; /* decompress the block */ ret = snd_soc_lzo_decompress_cache_block(codec, lzo_block); if (ret < 0) { kfree(lzo_block->dst); goto out; } /* write the new value to the cache */ if (snd_soc_set_cache_val(lzo_block->dst, blkpos, value, codec->driver->reg_word_size)) { kfree(lzo_block->dst); goto out; } /* prepare the source to be the decompressed block */ lzo_block->src = lzo_block->dst; lzo_block->src_len = lzo_block->dst_len; /* compress the block */ ret = snd_soc_lzo_compress_cache_block(codec, lzo_block); if (ret < 0) { kfree(lzo_block->dst); kfree(lzo_block->src); goto out; } /* set the bit so we know we have to sync this register */ set_bit(reg, lzo_block->sync_bmp); kfree(tmp_dst); kfree(lzo_block->src); return 0; out: lzo_block->dst = tmp_dst; lzo_block->dst_len = tmp_dst_len; return ret; } static int snd_soc_lzo_cache_read(struct snd_soc_codec *codec, unsigned int reg, unsigned int *value) { struct snd_soc_lzo_ctx *lzo_block, **lzo_blocks; int ret, blkindex, blkpos; size_t blksize, tmp_dst_len; void *tmp_dst; *value = 0; /* index of the compressed lzo block */ blkindex = snd_soc_lzo_get_blkindex(codec, reg); /* register index within the decompressed block */ blkpos = snd_soc_lzo_get_blkpos(codec, reg); /* size of the compressed block */ blksize = snd_soc_lzo_get_blksize(codec); lzo_blocks = codec->reg_cache; lzo_block = lzo_blocks[blkindex]; /* save the pointer and length of the compressed block */ tmp_dst = lzo_block->dst; tmp_dst_len = lzo_block->dst_len; /* prepare the source to be the compressed block */ lzo_block->src = lzo_block->dst; lzo_block->src_len = lzo_block->dst_len; /* decompress the block */ ret = snd_soc_lzo_decompress_cache_block(codec, lzo_block); if (ret >= 0) /* fetch the value from the cache */ *value = snd_soc_get_cache_val(lzo_block->dst, blkpos, codec->driver->reg_word_size); kfree(lzo_block->dst); /* restore the pointer and length of the compressed block */ lzo_block->dst = tmp_dst; lzo_block->dst_len = tmp_dst_len; return 0; } static int snd_soc_lzo_cache_exit(struct snd_soc_codec *codec) { struct snd_soc_lzo_ctx **lzo_blocks; int i, blkcount; lzo_blocks = codec->reg_cache; if (!lzo_blocks) return 0; blkcount = snd_soc_lzo_block_count(); /* * the pointer to the bitmap used for syncing the cache * is shared amongst all lzo_blocks. Ensure it is freed * only once. */ if (lzo_blocks[0]) kfree(lzo_blocks[0]->sync_bmp); for (i = 0; i < blkcount; ++i) { if (lzo_blocks[i]) { kfree(lzo_blocks[i]->wmem); kfree(lzo_blocks[i]->dst); } /* each lzo_block is a pointer returned by kmalloc or NULL */ kfree(lzo_blocks[i]); } kfree(lzo_blocks); codec->reg_cache = NULL; return 0; } static int snd_soc_lzo_cache_init(struct snd_soc_codec *codec) { struct snd_soc_lzo_ctx **lzo_blocks; size_t bmp_size; const struct snd_soc_codec_driver *codec_drv; int ret, tofree, i, blksize, blkcount; const char *p, *end; unsigned long *sync_bmp; ret = 0; codec_drv = codec->driver; /* * If we have not been given a default register cache * then allocate a dummy zero-ed out region, compress it * and remember to free it afterwards. */ tofree = 0; if (!codec->reg_def_copy) tofree = 1; if (!codec->reg_def_copy) { codec->reg_def_copy = kzalloc(codec->reg_size, GFP_KERNEL); if (!codec->reg_def_copy) return -ENOMEM; } blkcount = snd_soc_lzo_block_count(); codec->reg_cache = kzalloc(blkcount * sizeof *lzo_blocks, GFP_KERNEL); if (!codec->reg_cache) { ret = -ENOMEM; goto err_tofree; } lzo_blocks = codec->reg_cache; /* * allocate a bitmap to be used when syncing the cache with * the hardware. Each time a register is modified, the corresponding * bit is set in the bitmap, so we know that we have to sync * that register. */ bmp_size = codec_drv->reg_cache_size; sync_bmp = kmalloc(BITS_TO_LONGS(bmp_size) * sizeof(long), GFP_KERNEL); if (!sync_bmp) { ret = -ENOMEM; goto err; } bitmap_zero(sync_bmp, bmp_size); /* allocate the lzo blocks and initialize them */ for (i = 0; i < blkcount; ++i) { lzo_blocks[i] = kzalloc(sizeof **lzo_blocks, GFP_KERNEL); if (!lzo_blocks[i]) { kfree(sync_bmp); ret = -ENOMEM; goto err; } lzo_blocks[i]->sync_bmp = sync_bmp; lzo_blocks[i]->sync_bmp_nbits = bmp_size; /* alloc the working space for the compressed block */ ret = snd_soc_lzo_prepare(lzo_blocks[i]); if (ret < 0) goto err; } blksize = snd_soc_lzo_get_blksize(codec); p = codec->reg_def_copy; end = codec->reg_def_copy + codec->reg_size; /* compress the register map and fill the lzo blocks */ for (i = 0; i < blkcount; ++i, p += blksize) { lzo_blocks[i]->src = p; if (p + blksize > end) lzo_blocks[i]->src_len = end - p; else lzo_blocks[i]->src_len = blksize; ret = snd_soc_lzo_compress_cache_block(codec, lzo_blocks[i]); if (ret < 0) goto err; lzo_blocks[i]->decompressed_size = lzo_blocks[i]->src_len; } if (tofree) { kfree(codec->reg_def_copy); codec->reg_def_copy = NULL; } return 0; err: snd_soc_cache_exit(codec); err_tofree: if (tofree) { kfree(codec->reg_def_copy); codec->reg_def_copy = NULL; } return ret; } #endif static int snd_soc_flat_cache_sync(struct snd_soc_codec *codec) { int i; int ret; const struct snd_soc_codec_driver *codec_drv; unsigned int val; codec_drv = codec->driver; for (i = 0; i < codec_drv->reg_cache_size; ++i) { ret = snd_soc_cache_read(codec, i, &val); if (ret) return ret; if (codec->reg_def_copy) if (snd_soc_get_cache_val(codec->reg_def_copy, i, codec_drv->reg_word_size) == val) continue; WARN_ON(!snd_soc_codec_writable_register(codec, i)); ret = snd_soc_write(codec, i, val); if (ret) return ret; dev_dbg(codec->dev, "Synced register %#x, value = %#x\n", i, val); } return 0; } static int snd_soc_flat_cache_write(struct snd_soc_codec *codec, unsigned int reg, unsigned int value) { snd_soc_set_cache_val(codec->reg_cache, reg, value, codec->driver->reg_word_size); return 0; } static int snd_soc_flat_cache_read(struct snd_soc_codec *codec, unsigned int reg, unsigned int *value) { *value = snd_soc_get_cache_val(codec->reg_cache, reg, codec->driver->reg_word_size); return 0; } static int snd_soc_flat_cache_exit(struct snd_soc_codec *codec) { if (!codec->reg_cache) return 0; kfree(codec->reg_cache); codec->reg_cache = NULL; return 0; } static int snd_soc_flat_cache_init(struct snd_soc_codec *codec) { const struct snd_soc_codec_driver *codec_drv; codec_drv = codec->driver; if (codec->reg_def_copy) codec->reg_cache = kmemdup(codec->reg_def_copy, codec->reg_size, GFP_KERNEL); else codec->reg_cache = kzalloc(codec->reg_size, GFP_KERNEL); if (!codec->reg_cache) return -ENOMEM; return 0; } /* an array of all supported compression types */ static const struct snd_soc_cache_ops cache_types[] = { /* Flat *must* be the first entry for fallback */ { .id = SND_SOC_FLAT_COMPRESSION, .name = "flat", .init = snd_soc_flat_cache_init, .exit = snd_soc_flat_cache_exit, .read = snd_soc_flat_cache_read, .write = snd_soc_flat_cache_write, .sync = snd_soc_flat_cache_sync }, #ifdef CONFIG_SND_SOC_CACHE_LZO { .id = SND_SOC_LZO_COMPRESSION, .name = "LZO", .init = snd_soc_lzo_cache_init, .exit = snd_soc_lzo_cache_exit, .read = snd_soc_lzo_cache_read, .write = snd_soc_lzo_cache_write, .sync = snd_soc_lzo_cache_sync }, #endif { .id = SND_SOC_RBTREE_COMPRESSION, .name = "rbtree", .init = snd_soc_rbtree_cache_init, .exit = snd_soc_rbtree_cache_exit, .read = snd_soc_rbtree_cache_read, .write = snd_soc_rbtree_cache_write, .sync = snd_soc_rbtree_cache_sync } }; int snd_soc_cache_init(struct snd_soc_codec *codec) { int i; for (i = 0; i < ARRAY_SIZE(cache_types); ++i) if (cache_types[i].id == codec->compress_type) break; /* Fall back to flat compression */ if (i == ARRAY_SIZE(cache_types)) { dev_warn(codec->dev, "Could not match compress type: %d\n", codec->compress_type); i = 0; } mutex_init(&codec->cache_rw_mutex); codec->cache_ops = &cache_types[i]; if (codec->cache_ops->init) { if (codec->cache_ops->name) dev_dbg(codec->dev, "Initializing %s cache for %s codec\n", codec->cache_ops->name, codec->name); return codec->cache_ops->init(codec); } return -ENOSYS; } /* * NOTE: keep in mind that this function might be called * multiple times. */ int snd_soc_cache_exit(struct snd_soc_codec *codec) { if (codec->cache_ops && codec->cache_ops->exit) { if (codec->cache_ops->name) dev_dbg(codec->dev, "Destroying %s cache for %s codec\n", codec->cache_ops->name, codec->name); return codec->cache_ops->exit(codec); } return -ENOSYS; } /** * snd_soc_cache_read: Fetch the value of a given register from the cache. * * @codec: CODEC to configure. * @reg: The register index. * @value: The value to be returned. */ int snd_soc_cache_read(struct snd_soc_codec *codec, unsigned int reg, unsigned int *value) { int ret; mutex_lock(&codec->cache_rw_mutex); if (value && codec->cache_ops && codec->cache_ops->read) { ret = codec->cache_ops->read(codec, reg, value); mutex_unlock(&codec->cache_rw_mutex); return ret; } mutex_unlock(&codec->cache_rw_mutex); return -ENOSYS; } EXPORT_SYMBOL_GPL(snd_soc_cache_read); /** * snd_soc_cache_write: Set the value of a given register in the cache. * * @codec: CODEC to configure. * @reg: The register index. * @value: The new register value. */ int snd_soc_cache_write(struct snd_soc_codec *codec, unsigned int reg, unsigned int value) { int ret; mutex_lock(&codec->cache_rw_mutex); if (codec->cache_ops && codec->cache_ops->write) { ret = codec->cache_ops->write(codec, reg, value); mutex_unlock(&codec->cache_rw_mutex); return ret; } mutex_unlock(&codec->cache_rw_mutex); return -ENOSYS; } EXPORT_SYMBOL_GPL(snd_soc_cache_write); /** * snd_soc_cache_sync: Sync the register cache with the hardware. * * @codec: CODEC to configure. * * Any registers that should not be synced should be marked as * volatile. In general drivers can choose not to use the provided * syncing functionality if they so require. */ int snd_soc_cache_sync(struct snd_soc_codec *codec) { int ret; const char *name; if (!codec->cache_sync) { return 0; } if (!codec->cache_ops || !codec->cache_ops->sync) return -ENOSYS; if (codec->cache_ops->name) name = codec->cache_ops->name; else name = "unknown"; if (codec->cache_ops->name) dev_dbg(codec->dev, "Syncing %s cache for %s codec\n", codec->cache_ops->name, codec->name); trace_snd_soc_cache_sync(codec, name, "start"); ret = codec->cache_ops->sync(codec); if (!ret) codec->cache_sync = 0; trace_snd_soc_cache_sync(codec, name, "end"); return ret; } EXPORT_SYMBOL_GPL(snd_soc_cache_sync); static int snd_soc_get_reg_access_index(struct snd_soc_codec *codec, unsigned int reg) { const struct snd_soc_codec_driver *codec_drv; unsigned int min, max, index; codec_drv = codec->driver; min = 0; max = codec_drv->reg_access_size - 1; do { index = (min + max) / 2; if (codec_drv->reg_access_default[index].reg == reg) return index; if (codec_drv->reg_access_default[index].reg < reg) min = index + 1; else max = index; } while (min <= max); return -1; } int snd_soc_default_volatile_register(struct snd_soc_codec *codec, unsigned int reg) { int index; if (reg >= codec->driver->reg_cache_size) return 1; index = snd_soc_get_reg_access_index(codec, reg); if (index < 0) return 0; return codec->driver->reg_access_default[index].vol; } EXPORT_SYMBOL_GPL(snd_soc_default_volatile_register); int snd_soc_default_readable_register(struct snd_soc_codec *codec, unsigned int reg) { int index; if (reg >= codec->driver->reg_cache_size) return 1; index = snd_soc_get_reg_access_index(codec, reg); if (index < 0) return 0; return codec->driver->reg_access_default[index].read; } EXPORT_SYMBOL_GPL(snd_soc_default_readable_register); int snd_soc_default_writable_register(struct snd_soc_codec *codec, unsigned int reg) { int index; if (reg >= codec->driver->reg_cache_size) return 1; index = snd_soc_get_reg_access_index(codec, reg); if (index < 0) return 0; return codec->driver->reg_access_default[index].write; } EXPORT_SYMBOL_GPL(snd_soc_default_writable_register);