Over a million developers have joined DZone.
{{announcement.body}}
{{announcement.title}}

Bulk Insert in MongoDB With a C# Driver [Code Snippet]

DZone's Guide to

Bulk Insert in MongoDB With a C# Driver [Code Snippet]

In this post we take a look at how to efficiently save a large collection of documents in Mongo. Read on for details.

· Database Zone ·
Free Resource

Running out of memory? Learn how Redis Enterprise enables large dataset analysis with the highest throughput and lowest latency while reducing costs over 75%! 

There are situations where you need to save a lot of documents inside a collection in MongoDB. My scenario is a migration of documents from a collection to another database, with in-memory manipulation of the documents.

The most common error in these situations is to read the documents from the original collection, then execute a function that modify the document in-memory, and finally issuing an insert in destination collection. This is wrong because you have a roundtrip against MongoDB for each document you are saving.

Whenever you are calling Insert or Save function, you are paying the penality of a call to MongoDb process, network latency, etc, whenever possible you should reduce the number of calls to database engine.

In such a scenario, the MongoDB driver has a function called InsertBatch that allows you to insert documents in batches — and the fun part is that it simply accepts an IEnumerable. As an example, I have a function that manipulates a BsonDocument stored in a variable called Action. I have source and test database where I need to copy documents with manipulation and this is the code that does everything.

var sourceQueue = source.GetCollection(queue);
var destQueue = dest.GetCollection(queue);

if (sourceQueue.Count() == 0) return;

//migrate counterCollection
Console.WriteLine("Migrating Queue " + queue);
var allElement = sourceQueue
.FindAll()
.AsEnumerable()
.Select(document => {
    Action(document);
    return document;
});
destQueue.InsertBatch(allElement);

The name of the collection is contained in queue variable (actually I’m transforming a software that manages jobs), and as you can verify I can simply enumerate all source documents with FindAll (this code uses the old 1.10 driver), for each object I’m calling the Action function that manipulates the document, and finally I can simply use the InsertBatch to insert documents in batches.

This function runs faster than saving each document with a separate call, even if the MongoDB instance runs on the very same machine, so you do not pay the network latency.

If you use the latest version of the drivers, you have the InsertMany method that offers even more options and basically does the very same operation as InsertBatch.

Running out of memory? Never run out of memory with Redis Enterprise databaseStart your free trial today.

Topics:
database ,driver ,mongodb ,c# ,insert

Published at DZone with permission of

Opinions expressed by DZone contributors are their own.

{{ parent.title || parent.header.title}}

{{ parent.tldr }}

{{ parent.urlSource.name }}