diff --git a/doc/parse_many.md b/doc/parse_many.md index 1e2524a0..b7002f97 100644 --- a/doc/parse_many.md +++ b/doc/parse_many.md @@ -1,8 +1,14 @@ parse_many ========== -An interface providing features to work with files or streams containing multiple small JSON documents. -As fast and convenient as possible. +An interface providing features to work with files or streams containing multiple small JSON documents. Given an input such as +```JSON +{"text":"a"} +{"text":"b"} +{"text":"c"} +... +``` +... you want to read the entries (individual JSON documents) as quickly and as conveniently as possible. Importantly, the input might span several gigabytes, but you want to use a small (fixed) amount of memory. Ideally, you'd also like the parallelize the processing (using more than one core) to speed up the process. Contents -------- @@ -227,4 +233,4 @@ Consider the following example where a truncated document (`{"key":"intentionall ``` -Importantly, you should only call `truncated_bytes()` after iterating through all of the documents since the stream cannot tell whether there are truncated documents at the very end when it may not have accessed that part of the data yet. \ No newline at end of file +Importantly, you should only call `truncated_bytes()` after iterating through all of the documents since the stream cannot tell whether there are truncated documents at the very end when it may not have accessed that part of the data yet.