I have my hash table set to size = 16
modparam("htable", "htable", "a=>size=16;autoexpire=0;dbtable=htable;")
I had around 3000 entries in that table, some entries with more than 10 duplicate keys (key_type = 1) and we found that when reading the entries of the duplicated keys it could not go pass index 9 even though in the physical table there were 17 entries.
I deleted some old entries and now the table is about 2000 items and it seems to work properly again ,(loads all value in memory). I say seems because I can not dump all values:
./kamcmd htable.dump a
ERROR: reply too big
But I check manually all the ones that were not being loaded and they are now
Questions:
1- the size parameter is for the buckets or entries size ?
2- is 16 a good value for a table with 3000 entries? ( 2^16 is 65k so I figured it would be ok)
3- can the the way I define my key impose a limitation?
4- what is the max size for an htable to be dumped with the above command?
—
You are receiving this because you are subscribed to this thread.
Reply to this email directly, view it on GitHub, or unsubscribe.