I have my hash table set to size = 16
modparam("htable", "htable", "a=>size=16;autoexpire=0;dbtable=htable;")

I had around 3000 entries in that table, some entries with more than 10 duplicate keys (key_type = 1) and we found that when reading the entries of the duplicated keys it could not go pass index 9 even though in the physical table there were 17 entries.
I deleted some old entries and now the table is about 2000 items and it seems to work properly again ,(loads all value in memory). I say seems because I can not dump all values:

./kamcmd htable.dump a
ERROR: reply too big

But I check manually all the ones that were not being loaded and they are now
Questions:
1- the size parameter is for the buckets or entries size ?
2- is 16 a good value for a table with 3000 entries? ( 2^16 is 65k so I figured it would be ok)
3- can the the way I define my key impose a limitation?
4- what is the max size for an htable to be dumped with the above command?


You are receiving this because you are subscribed to this thread.
Reply to this email directly, view it on GitHub, or unsubscribe.