regmap: Cut down on the average # of nodes in the rbtree cache

This patch aims to bring down the average number of nodes
in the rbtree cache and increase the average number of registers
per node.  This should improve general lookup and traversal times.
This is achieved by setting the minimum size of a block within the
rbnode to the size of the rbnode itself.  This will essentially
cache possibly non-existent registers so to combat this scenario,
we keep a separate bitmap in memory which keeps track of which register
exists.  The memory overhead of this change is likely in the order of
~5-10%, possibly less depending on the register file layout.  On my test
system with a bitmap of ~4300 bits and a relatively sparse register
layout, the memory requirements for the entire cache did not increase
(the cutting down of nodes which was about 50% of the original number
compensated the situation).

A second patch that can be built on top of this can look at the
ratio `sizeof(*rbnode) / map->cache_word_size' in order to suitably
adjust the block length of each block.

Signed-off-by: Dimitris Papastamos <dp@opensource.wolfsonmicro.com>
Signed-off-by: Mark Brown <broonie@opensource.wolfsonmicro.com>
This commit is contained in:
Dimitris Papastamos 2013-03-15 14:54:35 +00:00 committed by Mark Brown
parent 584de329ca
commit 0c7ed8563a

View file

@ -36,6 +36,8 @@ struct regcache_rbtree_node {
struct regcache_rbtree_ctx {
struct rb_root root;
struct regcache_rbtree_node *cached_rbnode;
unsigned long *reg_present;
unsigned int reg_present_nbits;
};
static inline void regcache_rbtree_get_base_top_reg(
@ -146,6 +148,7 @@ static int rbtree_show(struct seq_file *s, void *ignored)
map->lock(map);
mem_size = sizeof(*rbtree_ctx);
mem_size += BITS_TO_LONGS(rbtree_ctx->reg_present_nbits) * sizeof(long);
for (node = rb_first(&rbtree_ctx->root); node != NULL;
node = rb_next(node)) {
@ -196,6 +199,44 @@ static void rbtree_debugfs_init(struct regmap *map)
}
#endif
static int enlarge_reg_present_bitmap(struct regmap *map, unsigned int reg)
{
struct regcache_rbtree_ctx *rbtree_ctx;
unsigned long *reg_present;
unsigned int reg_present_size;
unsigned int nregs;
int i;
rbtree_ctx = map->cache;
nregs = reg + 1;
reg_present_size = BITS_TO_LONGS(nregs);
reg_present_size *= sizeof(long);
if (!rbtree_ctx->reg_present) {
reg_present = kmalloc(reg_present_size, GFP_KERNEL);
if (!reg_present)
return -ENOMEM;
bitmap_zero(reg_present, nregs);
rbtree_ctx->reg_present = reg_present;
rbtree_ctx->reg_present_nbits = nregs;
return 0;
}
if (nregs > rbtree_ctx->reg_present_nbits) {
reg_present = krealloc(rbtree_ctx->reg_present,
reg_present_size, GFP_KERNEL);
if (!reg_present)
return -ENOMEM;
for (i = 0; i < nregs; i++)
if (i >= rbtree_ctx->reg_present_nbits)
clear_bit(i, reg_present);
rbtree_ctx->reg_present = reg_present;
rbtree_ctx->reg_present_nbits = nregs;
}
return 0;
}
static int regcache_rbtree_init(struct regmap *map)
{
struct regcache_rbtree_ctx *rbtree_ctx;
@ -209,6 +250,8 @@ static int regcache_rbtree_init(struct regmap *map)
rbtree_ctx = map->cache;
rbtree_ctx->root = RB_ROOT;
rbtree_ctx->cached_rbnode = NULL;
rbtree_ctx->reg_present = NULL;
rbtree_ctx->reg_present_nbits = 0;
for (i = 0; i < map->num_reg_defaults; i++) {
ret = regcache_rbtree_write(map,
@ -238,6 +281,8 @@ static int regcache_rbtree_exit(struct regmap *map)
if (!rbtree_ctx)
return 0;
kfree(rbtree_ctx->reg_present);
/* free up the rbtree */
next = rb_first(&rbtree_ctx->root);
while (next) {
@ -255,6 +300,17 @@ static int regcache_rbtree_exit(struct regmap *map)
return 0;
}
static int regcache_reg_present(struct regmap *map, unsigned int reg)
{
struct regcache_rbtree_ctx *rbtree_ctx;
rbtree_ctx = map->cache;
if (!(rbtree_ctx->reg_present[BIT_WORD(reg)] & BIT_MASK(reg)))
return 0;
return 1;
}
static int regcache_rbtree_read(struct regmap *map,
unsigned int reg, unsigned int *value)
{
@ -264,6 +320,8 @@ static int regcache_rbtree_read(struct regmap *map,
rbnode = regcache_rbtree_lookup(map, reg);
if (rbnode) {
reg_tmp = (reg - rbnode->base_reg) / map->reg_stride;
if (!regcache_reg_present(map, reg))
return -ENOENT;
*value = regcache_rbtree_get_register(map, rbnode, reg_tmp);
} else {
return -ENOENT;
@ -313,6 +371,12 @@ static int regcache_rbtree_write(struct regmap *map, unsigned int reg,
int ret;
rbtree_ctx = map->cache;
/* update the reg_present bitmap, make space if necessary */
ret = enlarge_reg_present_bitmap(map, reg);
if (ret < 0)
return ret;
set_bit(reg, rbtree_ctx->reg_present);
/* if we can't locate it in the cached rbnode we'll have
* to traverse the rbtree looking for it.
*/
@ -354,7 +418,7 @@ static int regcache_rbtree_write(struct regmap *map, unsigned int reg,
rbnode = kzalloc(sizeof *rbnode, GFP_KERNEL);
if (!rbnode)
return -ENOMEM;
rbnode->blklen = 1;
rbnode->blklen = sizeof(*rbnode);
rbnode->base_reg = reg;
rbnode->block = kmalloc(rbnode->blklen * map->cache_word_size,
GFP_KERNEL);
@ -404,6 +468,10 @@ static int regcache_rbtree_sync(struct regmap *map, unsigned int min,
for (i = base; i < end; i++) {
regtmp = rbnode->base_reg + (i * map->reg_stride);
if (!regcache_reg_present(map, regtmp))
continue;
val = regcache_rbtree_get_register(map, rbnode, i);
/* Is this the hardware default? If so skip. */