On Mon, Mar 18, 2013 at 6:09 PM, Pedro Ruivo <pedro@infinispan.org> wrote:
Hi all,

To solve ISPN-2808 (avoid blocking JGroups threads in order to allow to
deliver the request responses), I've created another thread pool to move
the possible blocking commands (i.e. the commands that may block until
some state is achieved).

Problem description:

With this solution, the new thread pool should be large in order to be
able to handle the remote commands without deadlocks. The problem is
that all the threads can be block to process the command that may
unblock other commands.

Example: a bunch of commands are blocked waiting for a new topology ID
and the command that will increment the topology ID is in the thread
pool queue.

Solution:

Use a smart command dispatcher, i.e., keep the command in the queue
until we are sure that it will not wait for other commands. I've already
implemented some kind of executor service (ConditionalExecutorService,
in ISPN-2635 and ISPN-2636 branches, Total Order stuff) that only put
the Runnable (more precisely a new interface called ConditionalRunnable)
in the thread pool when it is ready to be processed. Creative guys, it
may need a better name :)

The ConditionalRunnable has a new method (boolean isReady()) that should
return true when the runnable should not block.

Example how to apply this to ISPN-2808:

Most of the commands awaits for a particular topology ID and/or for lock
acquisition.

Well, the original problem description was about the DistributionInterceptor forwarding the command from the primary owner to the backup owners and waiting for a response from them :)
The forwarding done by StateTransferInterceptor is also synchronous and can block.

It's true that you can't check how long a remote call will take beforehand...
 
In this way, the isReady() implementation can be something
like:

isReady()
  return commandTopologyId <= currentTopologyId && (for all keys; do if
!lock(key).tryLock(); return false; done)


Shouldn't you release the locks you managed to lock already if one of the lock acquisitions failed?

Actually you may want to release the locks even if lock acquisition did succeed... In non-tx mode, the locks are owned by the current thread, so you can't lock a key on one thread and unlock it on another (though you could skip the check completely in non-tx mode). And in transactional mode, you could have a deadlock because 2 txs lock the same keys in a different order.

Which leads me to a different point, how would you handle deadlocks? With pessimistic mode, if tx1 holds lock k1 and wants to acquire k2, but tx2 holds k2 and wants to acquire k1, will the LockCommands tx1:k2 and tx2:k1 ever be scheduled? In general, can we make the time that a Lock/PrepareCommand spends in the ConditionalExecutorService queue count against lockAcquisitionTimeout?

With this, I believe we can keep the number of thread low and avoid the
thread deadlocks.

Now, I have two possible implementations:

1) put a reference for StateTransferManager and/or LockManager in the
commands, and invoke the methods directly (a little dirty)

2) added new method in the CommandInterceptor like: boolean
preProcess<command>(Command, InvocationContext). each interceptor will
check if the command will block on it (returning false) or not (invoke
the next interceptor). For example, the StateTransferInterceptor returns
immediately false if the commandToplogyId is higher than the
currentTopologyId and the *LockingIntercerptor will return false if it
cannot acquire some lock.

Any other suggestions? If I was not clear let me know.

TBH I would only check for the topology id before scheduling the commands to the Infinispan thread pool, because checking the locks is very complicated with all the configuration options that we have.

This is how I was imagining solving the lock problem: The OOB threads would execute directly the commands that do not acquire locks (CommitCommand, TxCompletionNotificationCommand) directly and submit the others to the ISPN thread pool; if the command's topology id was higher than the current topology id, the OOB thread would just stick it in a queue. A separate thread would read only the commands with the current topology id from the queue, and process them just like an OOB thread.

However... the CommitCommands may very well have a *lower* topology id by the time they are executed, so the OOB/executor thread may well block while waiting for the results of the forwarding RPC. Maybe we could work around it by having StateTransferInterceptor submit a task with the command forwarding only to the thread pool, but again it gets quite complicated. So for the first phase I recommend handling only the topology id problem.