Available since: 6.2.0
Time complexity: O(1)
Atomically returns and removes the first/last element (head/tail depending on
wherefrom argument) of the list stored at
source, and pushes the
element at the first/last element (head/tail depending on the
argument) of the list stored at
For example: consider
source holding the list
holding the list
LMOVE source destination RIGHT LEFT results in
source does not exist, the value
nil is returned and no operation is
destination are the same, the operation is equivalent to
removing the first/last element from the list and pushing it as first/last
element of the list, so it can be considered as a list rotation command (or a
wherefrom is the same as
This command comes in place of the now deprecated
LMOVE RIGHT LEFT is equivalent.
Bulk string reply: the element being popped and pushed.
Redis is often used as a messaging server to implement processing of background
jobs or other kinds of messaging tasks.
A simple form of queue is often obtained pushing values into a list in the
producer side, and waiting for this values in the consumer side using
(using polling), or
BRPOP if the client is better served by a blocking
However in this context the obtained queue is not reliable as messages can be lost, for example in the case there is a network problem or if the consumer crashes just after the message is received but it is still to process.
BLMOVE for the blocking variant) offers a way to avoid
this problem: the consumer fetches the message and at the same time pushes it
into a processing list.
It will use the
LREM command in order to remove the message from the
processing list once the message has been processed.
An additional client may monitor the processing list for items that remain there for too much time, and will push those timed out items into the queue again if needed.
LMOVE with the same source and destination key, a client can visit
all the elements of an N-elements list, one after the other, in O(N) without
transferring the full list from the server to the client using a single
The above pattern works even if the following two conditions:
The above makes it very simple to implement a system where a set of items must be processed by N workers continuously as fast as possible. An example is a monitoring system that must check that a set of web sites are reachable, with the smallest delay possible, using a number of parallel workers.
Note that this implementation of workers is trivially scalable and reliable, because even if a message is lost the item is still in the queue and will be processed at the next iteration.