i did an export (Instance running v5.2.4, ubuntu/docker) and trying to import it into another instance v5.3.2 (windows, non docker). The import file is 31gb. the import machine is my local potent gaming machine with 54GB of ram and really fast nvme disks.
unfortunately i run into the following exception after ~ 5mio documents (75mio in total).
Fatal error. System.AccessViolationException: Attempted to read or write protected memory. This is often an indication that other memory is corrupt.
at Sparrow.Compression.LZ4.LZ4_decompress_generic[[Sparrow.Compression.LZ4+EndOnInputSize, Sparrow, Version=5.3.2.0, Culture=neutral, PublicKeyToken=37f41c7f99471593],[Sparrow.Co
mpression.LZ4+Full, Sparrow, Version=5.3.2.0, Culture=neutral, PublicKeyToken=37f41c7f99471593],[Sparrow.Compression.LZ4+NoDict, Sparrow, Version=5.3.2.0, Culture=neutral, PublicKey
Token=37f41c7f99471593]](Byte*, Byte*, Int32, Int32, Int32, Byte*, Byte*, Int32)
at Sparrow.Compression.LZ4.Decode64(Byte*, Int32, Byte*, Int32, Boolean)
at Sparrow.Compression.LZ4.Decode64LongBuffers(Byte*, Int64, Byte*, Int64, Boolean)
at Voron.Data.BTrees.Tree.DecompressFromBuffer(Voron.Data.Compression.DecompressionUsage, DecompressionInput ByRef)
at Voron.Data.BTrees.Tree.DecompressPage(Voron.Data.BTrees.TreePage, Voron.Data.Compression.DecompressionUsage, Boolean)
at Voron.Data.BTrees.Tree.ReadDecompressed(Voron.Slice)
at Raven.Server.Documents.Indexes.MapReduce.MapReduceResultsStore.Get(Int64)
at Raven.Server.Documents.Indexes.MapReduce.MapReduceIndexBase`2[[System.__Canon, System.Private.CoreLib, Version=5.0.0.0, Culture=neutral, PublicKeyToken=7cec85d7bea7798e],[Syst
em.__Canon, System.Private.CoreLib, Version=5.0.0.0, Culture=neutral, PublicKeyToken=7cec85d7bea7798e]].PutMapResults(Sparrow.Json.LazyStringValue, Sparrow.Json.LazyStringValue, Sys
tem.Collections.Generic.IEnumerable`1<Raven.Server.Documents.Indexes.MapReduce.MapResult>, Raven.Server.ServerWide.Context.TransactionOperationContext, Raven.Server.Documents.Indexe
s.IndexingStatsScope)
at Raven.Server.Documents.Indexes.MapReduce.Static.MapReduceIndex.HandleMap(Raven.Server.Documents.Indexes.IndexItem, System.Collections.IEnumerable, System.Lazy`1<Raven.Server.D
ocuments.Indexes.Persistence.Lucene.IndexWriteOperation>, Raven.Server.ServerWide.Context.TransactionOperationContext, Raven.Server.Documents.Indexes.IndexingStatsScope)
at Raven.Server.Documents.Indexes.Workers.MapItems.Execute(Raven.Server.Documents.Indexes.QueryOperationContext, Raven.Server.ServerWide.Context.TransactionOperationContext, Syst
em.Lazy`1<Raven.Server.Documents.Indexes.Persistence.Lucene.IndexWriteOperation>, Raven.Server.Documents.Indexes.IndexingStatsScope, System.Threading.CancellationToken)
at Raven.Server.Documents.Indexes.Index.DoIndexingWork(Raven.Server.Documents.Indexes.IndexingStatsScope, System.Threading.CancellationToken)
at Raven.Server.Documents.Indexes.Index.ExecuteIndexing()
at Raven.Server.Documents.Indexes.Index.<StartIndexingThread>b__142_0(System.Object)
at Raven.Server.Utils.PoolOfThreads+PooledThread.DoWork()
at Raven.Server.Utils.PoolOfThreads+PooledThread.Run()
at System.Threading.ThreadHelper.ThreadStart_Context(System.Object)
at System.Threading.ExecutionContext.RunInternal(System.Threading.ExecutionContext, System.Threading.ContextCallback, System.Object)
at System.Threading.ThreadHelper.ThreadStart()
I did uncheck "import index".
i see the exception in the console window and the server is also dead -> needs to be closed and started again. after restarting the server, the db can't be loaded anymore.
next try would be to delete the database and import it again. but just in case if there is some debug info around, i didn't yet do it.