Couldn't extract the data from large partition in cassandra -


i realized after sometime partition size in cassandra data model incorrect. , had designed primary key created 1 partition. so, getting read timeout in queries. have fixed issue creating new column family problem how recover data partition appreciated. have around 40gb of partition given below :

38g     ./eventdata-7b5004e02bdc11e69ebb19dbbcf6d4b0 55m     ./eventdata-8f8317a01e5711e69ebb19dbbcf6d4b0 2.4m    ./eventdata-3c20dd401cf911e69ebb19dbbcf6d4b0 88k     ./eventdata-1866015012a911e69ebb19dbbcf6d4b0 

so, read data using sstable2json convert in json gave error :

exception in thread "main" java.lang.outofmemoryerror: gc overhead limit exceeded         @ java.nio.heapbytebuffer.duplicate(heapbytebuffer.java:107)         @ org.apache.cassandra.db.composites.abstractctype.slicebytes(abstractctype.java:369)         @ org.apache.cassandra.db.composites.abstractcompoundcellnametype.frombytebuffer(abstractcompoundcellnametype.java:101)         @ org.apache.cassandra.db.composites.abstractctype$serializer.deserialize(abstractctype.java:398)         @ org.apache.cassandra.db.composites.abstractctype$serializer.deserialize(abstractctype.java:382)         @ org.apache.cassandra.io.sstable.indexhelper$indexinfo$serializer.deserialize(indexhelper.java:194)         @ org.apache.cassandra.io.sstable.indexhelper$indexinfo$serializer.deserialize(indexhelper.java:174)         @ org.apache.cassandra.db.rowindexentry$serializer.deserialize(rowindexentry.java:129)         @ org.apache.cassandra.io.sstable.sstablescanner$keyscanningiterator.computenext(sstablescanner.java:257)         @ org.apache.cassandra.io.sstable.sstablescanner$keyscanningiterator.computenext(sstablescanner.java:231)         @ com.google.common.collect.abstractiterator.trytocomputenext(abstractiterator.java:143)         @ com.google.common.collect.abstractiterator.hasnext(abstractiterator.java:138)         @ org.apache.cassandra.io.sstable.sstablescanner.hasnext(sstablescanner.java:211)         @ org.apache.cassandra.tools.sstableexport.export(sstableexport.java:332)         @ org.apache.cassandra.tools.sstableexport.export(sstableexport.java:369)         @ org.apache.cassandra.tools.sstableexport.export(sstableexport.java:382)         @ org.apache.cassandra.tools.sstableexport.main(sstableexport.java:467) error 14:02:13 leak detected: reference (org.apache.cassandra.utils.concurrent.ref$state@2835df6c) class org.apache.cassandra.io.sstable.s stablereader$instancetidier@1560150690:/mnt/eventdata/eventdata-7b5004e02bdc11e69ebb19dbbcf6d4b0/./cw-eventdata-ka-2839 not released before reference garbage collected 

then trying split sstable using sstablesplit utility again got error. have cassandra 2.1.12 running on node having 8gb of ram , 8gb of swap space :

exception in thread "main" java.lang.outofmemoryerror: gc overhead limit exceeded         @ org.apache.cassandra.io.util.randomaccessreader.readbytes(randomaccessreader.java:375)         @ org.apache.cassandra.utils.bytebufferutil.read(bytebufferutil.java:352)         @ org.apache.cassandra.utils.bytebufferutil.readwithshortlength(bytebufferutil.java:331)         @ org.apache.cassandra.db.composites.abstractctype$serializer.deserialize(abstractctype.java:398)         @ org.apache.cassandra.db.composites.abstractctype$serializer.deserialize(abstractctype.java:382)         @ org.apache.cassandra.io.sstable.indexhelper$indexinfo$serializer.deserialize(indexhelper.java:193)         @ org.apache.cassandra.io.sstable.indexhelper$indexinfo$serializer.deserialize(indexhelper.java:174)         @ org.apache.cassandra.db.rowindexentry$serializer.deserialize(rowindexentry.java:129)         @ org.apache.cassandra.io.sstable.sstablescanner$keyscanningiterator.computenext(sstablescanner.java:257)         @ org.apache.cassandra.io.sstable.sstablescanner$keyscanningiterator.computenext(sstablescanner.java:231)         @ com.google.common.collect.abstractiterator.trytocomputenext(abstractiterator.java:143)         @ com.google.common.collect.abstractiterator.hasnext(abstractiterator.java:138)         @ org.apache.cassandra.io.sstable.sstablescanner.hasnext(sstablescanner.java:211)         @ org.apache.cassandra.utils.mergeiterator$onetoone.computenext(mergeiterator.java:202)         @ com.google.common.collect.abstractiterator.trytocomputenext(abstractiterator.java:143)         @ com.google.common.collect.abstractiterator.hasnext(abstractiterator.java:138)         @ org.apache.cassandra.db.compaction.compactiontask.runmaythrow(compactiontask.java:181)         @ org.apache.cassandra.utils.wrappedrunnable.run(wrappedrunnable.java:28)         @ org.apache.cassandra.db.compaction.compactiontask.executeinternal(compactiontask.java:73)         @ org.apache.cassandra.db.compaction.abstractcompactiontask.execute(abstractcompactiontask.java:59)         @ org.apache.cassandra.db.compaction.sstablesplitter.split(sstablesplitter.java:38)         @ org.apache.cassandra.tools.standalonesplitter.main(standalonesplitter.java:153) 


Comments

Popular posts from this blog

jOOQ update returning clause with Oracle -

java - Warning equals/hashCode on @Data annotation lombok with inheritance -

java - BasicPathUsageException: Cannot join to attribute of basic type -