I think that Spark's streaming API, particularly the window operations, could be an effective way to do computations in real time as data as ingested
+1
not only for processing the streams, but also for any kind of post processing needed. plus it would supply the abstractions to run computations across large number of nodes.