ketama - a consistent hashing algo for memcache clients
tbone
tbone at nexopia.com
Wed Apr 11 04:47:02 UTC 2007
Which could happen if the issue was intermittent (network hiccup,
semi-bad network cable, or even capped bandwidth on the memcached
machine [which I have seen happen])
Ian Holsman wrote:
>
> the old entry would only exist as long as your expiry time.
>
> so steps 3-5 would have to occur in your expiry window for this to
> take effect.
>
>
> regards
> Ian
>
> On 11/04/2007, at 2:18 PM, Timo Ewalds wrote:
>
>> I'd like to point out that the problem of stale cache entries exists
>> if you add AND remove servers without flushing all the data. Here's
>> the sequence that would cause it:
>>
>> 1. Start a bunch of servers
>> 2. Run for a while
>> 3. Add 1 server
>> 4. Run for a while
>> 5. Remove the server added in step 3.
>>
>> Now a key that mapped to one of the original servers, but moved to
>> the new server, maps to the old server again. If the key hadn't
>> expired on the old server, it'll be returned as if it was valid,
>> though it may have been changed/invalidated on the new server. To
>> avoid this, you can never remove servers after adding servers. Adding
>> is fine, removing is fine, adding fresh/flushed ones after removing
>> is fine, but removing after adding is not.
>>
>> Note that this problem is very similar to the one of rehashing key
>> locations when a server goes down and comes back up without being
>> flushed (ie intermittent network issues). I believe the consensus to
>> fix it is just to not rehash and deal with the extra db load from
>> having a downed memcache server.
>>
>> As a separate issue, presumably if you're adding servers often,
>> you've got many servers. If you're adding 100 buckets per server (to
>> make the distribution even), what's the cpu usage like to figure out
>> which server to choose, compared to the simple hash method. I'd guess
>> in php that could be quite a bit heavier where the bucket list needs
>> to be rebuilt each page.
>>
>> Has anyone tried using consistent hashing not to make server
>> management easier, but to increase locality of keys to make
>> get_multi's more likely to use fewer servers? Something like have the
>> top 20 bits be a hash of the prefix, and the lower 12 bits be a hash
>> of the suffix (id?). Using consistent hashing, it means keys with the
>> same prefix will be nearby in the number line, so likely to map to
>> the same server.
>> Another solution which we already use for a specific case would be to
>> have different server pools for different data groups (I believe
>> Steven Grimm mentioned doing this), but it is annoying to do in the
>> general case.
>>
>> Timo
>>
>>
>> Richard Jones wrote:
>>> Every time we added memcached servers to the pool, the rehashing of
>>> all keys would kill our database while the cache filled up again. We
>>> implemented a consistent hashing algorithm that our memcached
>>> clients are now using to map keys to servers. Now we can add new
>>> memcached servers without causing chaos :)
>>>
>>> Here is the blurb and the code:
>>> http://www.last.fm/user/RJ/journal/2007/04/10/392555/
>>>
>>> Source download includes:
>>> libketama, a c library
>>> php_ketama, a php extension that uses the library
>>> a pure java implementation
>>>
>>> I'm interested in other approaches (or even if this is a concern for
>>> most people?). I read about the lots-of-virtual-buckets concept, but
>>> didn't find an implementation. Anyone already using consistent
>>> hashing for memcache?
>>>
>>> RJ
>>>
>>>
>
> --
> Ian Holsman
> Ian at Holsman.net
> http://garden-gossip.com/ -- what's in your garden?
>
>
More information about the memcached
mailing list