[infinispan-issues] [JBoss JIRA] Issue Comment Edited: (ISPN-78) Large object support

Olaf Bergner (JIRA) jira-events at lists.jboss.org
Tue Mar 29 19:47:39 EDT 2011


    [ https://issues.jboss.org/browse/ISPN-78?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12592546#comment-12592546 ] 

Olaf Bergner edited comment on ISPN-78 at 3/29/11 7:46 PM:
-----------------------------------------------------------

Wouldn't a non streaming API somehow defeat this feature's sole purpose? It is my understanding that it is driven by the desire to store objects the size of which exceeds that of every single JVM in the cluster, correct?

If that is indeed the case, it would be *impossible* to store such an object without first fragmenting it into pieces each of which fits into a single node's heap. So we need a streaming API to read a large object from some place - think file - outside the current JVM's address space in order not to blow up that JVM's heap. It is out of the question that this approach considerably complicates matters, but I don't see a way around it.

      was (Author: O.Bergner):
    Wouldn't a non streaming API somehow defeat this feature's sole purpose? It is my understanding that it is driven by the desire to store objects the size of which exceeds that of every single JVM in the cluster, correct?

If that is indeed the case, it would be *impossible* to store such an object without first fragmenting into it pieces each of which fits into a single node's heap. So we need a streaming API to read a large object from some place - think file - outside the current JVM's address space in order not to blow up that JVM's heap. It is out of the question that this approach considerably complicates matters, but I don't see a way around it.
  
> Large object support
> --------------------
>
>                 Key: ISPN-78
>                 URL: https://issues.jboss.org/browse/ISPN-78
>             Project: Infinispan
>          Issue Type: Feature Request
>          Components: Core API
>            Reporter: Manik Surtani
>            Assignee: Manik Surtani
>             Fix For: 5.1.0.BETA1, 5.1.0.Final
>
>
> if each VM is allocated a 2GB heap and you have a 100 nodes in a grid with 1 redundant copy for each key, you have a theoretical addressable heap of 100GB.  But you are limited by (half) the heap of a single VM per entry, since entries are stored whole.
> E.g., cache.put(k, my2GBObject) will fail since you need at least 2GB for the object + another 2GB for its serialized form.
> This gets worse when you try cache.put(k, my10GBObject).  This *should* be possible if we have a theoretical 100GB heap.
> Potential solutions here are to fragment large objects, and store each fragment under separate keys.  Another approach would be to directly stream objects to disk. etc.  Needs thought and design, possibly a separate API to prevent 'pollution" of the more simplistic API.  (JumboCache?)
> Re: fragmenting, issues to overcome:
> How many chunks to fragment into?  Max size of each key could be configured, but how do we determine the size of an Object?  VM instrumentation?  Or perhaps the JumboCache only stores byte[]'s?  

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


More information about the infinispan-issues mailing list