search
  • Sign In
  • Sign Up
Password reset successful

Follow the proiects vou are interested in andi aet the latestnews about them taster

Table of Contents
Understand the differences between NDJSON format and traditional JSON parsing
Solution: Use JsonReader for iterative parsing
core steps
Sample code
Home Java javaTutorial Parse and map multiple records in NDJSON file using Gson

Parse and map multiple records in NDJSON file using Gson

Dec 03, 2025 pm 01:00 PM

Parse and map multiple records in NDJSON file using Gson

This article details how to use the Gson library to efficiently parse and map multiple JSON records in NDJSON (Newline Delimited JSON) files in Java. To solve the problem that the traditional Gson `fromJson` method can only read the first record, the tutorial proposes an iterative parsing solution based on the `JsonReader` and `peek()` methods, and provides a complete Java code example to ensure that all independent JSON objects in the file can be successfully read and deserialized.

NDJSON (Newline Delimited JSON) is a common format when dealing with text files containing multiple independent JSON objects, where each JSON object occupies one line. However, when trying to use the Gson library to parse such files, developers often encounter the problem that only the first JSON record can be read successfully. This is because Gson's fromJson method treats the entire input stream as a complete JSON structure by default, and the input is considered to end once the first top-level JSON object is parsed.

In order to solve this problem, we need to use the JsonReader class provided by Gson to perform iterative parsing and read the JSON objects in the file one by one.

Understand the differences between NDJSON format and traditional JSON parsing

Each record in the NDJSON file is an independent, complete JSON object, separated by newlines. For example:

 {"id": 1, "name": "Alice"}
{"id": 2, "name": "Bob"}
{"id": 3, "name": "Charlie"}

If you use gson.fromJson(reader, MyObject.class) directly, Gson will read the first {"id": 1, "name": "Alice"} and map it successfully, but the following content will be ignored because Gson thinks that the file has been parsed.

Solution: Use JsonReader for iterative parsing

To correctly parse all records in an NDJSON file, a combination of JsonReader and a loop structure is required. JsonReader provides finer-grained control over JSON streaming.

core steps

  1. Initialize JsonReader : Create a JsonReader instance from a file or input stream.
  2. Set relaxed mode (setLenient(true)) : There is usually no comma separation between each JSON object in the NDJSON file. JsonReader by default will expect a single JSON document or a comma-separated JSON array. In order for JsonReader to tolerate this non-strict JSON format (i.e. multiple top-level JSON objects), it needs to be set to relaxed mode.
  3. Iterative reading : Use a loop structure, combined with the reader.peek() method to determine whether the end of the file is reached. In each loop, call gson.fromJson(reader, YourDTO.class) to read a JSON object and add it to the list.

Sample code

Let's say we have a CustomerFeedDTO class that maps each customer record in an NDJSON file:

 import java.util.ArrayList;
import java.util.List;
import java.util.Map;

//Customer Data Transfer Object (DTO)
class CustomerFeedDTO {
    private Map<string> profile;
    private Map<string> phone;
    private ArrayList&gt; addresses;
    private Map<string> orders;
    private ArrayList&gt; customs;

    //Constructor, Getter and Setter methods (omitted here, but needed in actual applications)
    public Map<string> getProfile() { return profile; }
    public void setProfile(Map<string> profile) { this.profile = profile; }

    public Map<string> getPhone() { return phone; }
    public void setPhone(Map<string> phone) { this.phone = phone; }

    public ArrayList&gt; getAddresses() { return addresses; }
    public void setAddresses(ArrayList&gt; addresses) { this.addresses = addresses; }

    public Map<string> getOrders() { return orders; }
    public void setOrders(Map<string> orders) { this.orders = orders; }

    public ArrayList&gt; getCustoms() { return customs; }
    public void setCustoms(ArrayList&gt; customs) { this.customs = customs; }

    @Override
    public String toString() {
        return "CustomerFeedDTO{"  
               "profile=" profile  
               ", phone=" phone  
               ", addresses=" addresses  
               ", orders=" orders  
               ", customs=" customs  
               '}';
    }
}</string></string></string></string></string></string></string></string></string>

Using the above CustomerFeedDTO and NDJSON files, the following code demonstrates how to parse all records:

 import com.google.gson.Gson;
import com.google.gson.stream.JsonReader;
import com.google.gson.stream.JsonToken;

import java.io.FileReader;
import java.io.IOException;
import java.util.ArrayList;
import java.util.List;

public class NdJsonParser {

    public static void main(String[] args) {
        List<customerfeeddto> customerFeedDTOs = new ArrayList();
        Gson gson = new Gson();

        // Assume customer.json is your NDJSON file try (FileReader fileReader = new FileReader("customer.json");
             JsonReader reader = new JsonReader(fileReader)) {

            // Must be set to relaxed mode to handle multiple top-level JSON objects reader.setLenient(true);

            // Loop reading until the end of the document while (reader.peek() != JsonToken.END_DOCUMENT) {
                CustomerFeedDTO customerFeedDTO = gson.fromJson(reader, CustomerFeedDTO.class);
                customerFeedDTOs.add(customerFeedDTO);
            }

            // Print all parsed customer records for (int i = 0; i <h4> Choice between reader.peek() and reader.hasNext()</h4>
<p> In the loop condition, we used reader.peek() != JsonToken.END_DOCUMENT. The peek() method returns the type of the next token without actually consuming it. When peek() returns JsonToken.END_DOCUMENT, it means that the end of the JSON document has been reached.</p>
<p> Although JsonReader also has a hasNext() method, it may cause exceptions in NDJSON scenarios. hasNext() is mainly used to check if there are more elements in a JSON array or object, and may throw an IllegalStateException when it tries to read an unexpected token (for example, when there are no more structured JSON elements at the end of the document). Therefore, for a stream like NDJSON, which is composed of multiple independent JSON objects, it is more robust and recommended to use peek() == JsonToken.END_DOCUMENT to determine the loop termination condition.</p>
<h3> Things to note and best practices</h3>
<ol>
<li> <strong>Resource management</strong> : Always use the try-with-resources statement to ensure that FileReader and JsonReader can be closed correctly after use to avoid resource leaks.</li>
<li> <strong>Error handling</strong> : In actual applications, more detailed exception handling logic should be added, such as catching JsonSyntaxException to handle malformed JSON records.</li>
<li> <strong>DTO design</strong> : The CustomerFeedDTO in the example uses Map<string> and ArrayList&gt; to handle unknown or dynamic JSON structures. When the JSON structure is known, it is recommended to use specific Java types (such as String, Integer, custom objects, etc.) instead of Map and ArrayList&gt; to improve type safety and code readability. For example, profile can be defined as a specific ProfileDTO class.</string>
</li>
<li> <strong>Performance considerations</strong> : For very large NDJSON files, loading all records into a List at once can consume a lot of memory. In this case, you can consider processing records one by one, or using a streaming API (such as Java 8 Stream API) combined with a custom iterator to process the data to avoid loading all the data at once.</li>
</ol>
<h3> Summarize</h3>
<p> Through the iterative parsing mechanism of JsonReader, combined with reader.setLenient(true) and reader.peek() != JsonToken.END_DOCUMENT as a loop condition, all JSON records in the NDJSON file can be effectively parsed and mapped. This approach provides fine-grained control over the JSON flow and is the standard and recommended practice when processing multiple independent rows of JSON data. Proper understanding and application of these techniques ensures reliable processing of NDJSON data in Java applications.</p></customerfeeddto>

The above is the detailed content of Parse and map multiple records in NDJSON file using Gson. For more information, please follow other related articles on the PHP Chinese website!

Statement of this Website
The content of this article is voluntarily contributed by netizens, and the copyright belongs to the original author. This site does not assume corresponding legal responsibility. If you find any content suspected of plagiarism or infringement, please contact admin@php.cn

Hot AI Tools

Undress AI Tool

Undress AI Tool

Undress images for free

AI Clothes Remover

AI Clothes Remover

Online AI tool for removing clothes from photos.

Undresser.AI Undress

Undresser.AI Undress

AI-powered app for creating realistic nude photos

ArtGPT

ArtGPT

AI image generator for creative art from text prompts.

Stock Market GPT

Stock Market GPT

AI powered investment research for smarter decisions

Popular tool

Notepad++7.3.1

Notepad++7.3.1

Easy-to-use and free code editor

SublimeText3 Chinese version

SublimeText3 Chinese version

Chinese version, very easy to use

Zend Studio 13.0.1

Zend Studio 13.0.1

Powerful PHP integrated development environment

Dreamweaver CS6

Dreamweaver CS6

Visual web development tools

SublimeText3 Mac version

SublimeText3 Mac version

God-level code editing software (SublimeText3)

How to configure Spark distributed computing environment in Java_Java big data processing How to configure Spark distributed computing environment in Java_Java big data processing Mar 09, 2026 pm 08:45 PM

Spark cannot run in local mode, ClassNotFoundException: org.apache.spark.sql.SparkSession. This is the most common first step of getting stuck: even the dependencies are not correct. Only spark-core_2.12 is written in Maven, but spark-sql_2.12 is not added. SparkSession crashes as soon as it is built. The Scala version must strictly match the official Spark compiled version - Spark3.4.x uses Scala2.12 by default. If you use spark-sqljar of 2.13, the class loader cannot directly find the main class. Practical advice: Go to mvnre

How to safely map user-entered weekday string to integer value and implement date offset operation in Java How to safely map user-entered weekday string to integer value and implement date offset operation in Java Mar 09, 2026 pm 09:43 PM

This article introduces a concise and maintainable way to map the weekday string (such as "Monday") to the corresponding serial number (1-7), and use the modulo operation to realize the forward and backward offset of any number of days (such as Monday plus 4 days to get Friday), avoiding lengthy if chains and hard-coded logic.

How to generate a list of duplicate elements using Java's Collections.nCopies_Initialization tips How to generate a list of duplicate elements using Java's Collections.nCopies_Initialization tips Mar 06, 2026 am 06:24 AM

Collections.nCopies returns an immutable view. Calling add/remove will throw UnsupportedOperationException; it needs to be wrapped with newArrayList() to modify it, and it is disabled for mutable objects.

What is exception masking (Suppressed Exceptions) in Java_Multiple resource shutdown exception handling What is exception masking (Suppressed Exceptions) in Java_Multiple resource shutdown exception handling Mar 10, 2026 pm 06:57 PM

What is SuppressedException: It is not "swallowed", but actively archived by the JVM. SuppressedException is not an exception loss, but the JVM quietly attaches the secondary exception to the main exception under the premise that "only one exception must be thrown" for you to verify afterwards. It is automatically triggered by the JVM in only two scenarios: one is that the resource closure in try-with-resources fails, and the other is that you manually call addSuppressed() in finally. The key difference is: the former is fully automatic and safe; the latter requires you to keep it to yourself, and it can be written as shadowing if you are not careful. try-

How to use Homebrew to install Java on Mac_A must-have Java tool chain for developers How to use Homebrew to install Java on Mac_A must-have Java tool chain for developers Mar 09, 2026 pm 09:48 PM

Homebrew installs the latest stable version of openjdk (such as JDK22) by default, not the LTS version; you need to explicitly execute brewinstallopenjdk@17 or brewinstallopenjdk@21 to install the LTS version, and manually configure PATH and JAVA_HOME to be correctly recognized by the system and IDE.

How to correctly implement runtime file writing in Java applications (avoiding JAR internal write failures) How to correctly implement runtime file writing in Java applications (avoiding JAR internal write failures) Mar 09, 2026 pm 07:57 PM

After a Java application is packaged as a JAR, data cannot be written directly to the resources in the JAR package (such as test.txt) because the JAR is essentially a read-only ZIP archive; the correct approach is to write variable data to an external path (such as a user directory, a temporary directory, or a configuration-specified path).

What is the underlying principle of array expansion in Java_Java memory dynamic adjustment analysis What is the underlying principle of array expansion in Java_Java memory dynamic adjustment analysis Mar 09, 2026 pm 09:45 PM

ArrayList.add() triggers expansion because grow() is called when size is equal to elementData.length. The first add allocates 10 capacity, and subsequent expansion is 1.5 times and not less than the minimum requirement, relying on delayed initialization and System.arraycopy optimization.

How to safely read a line of integer input in Java and avoid Scanner blocking How to safely read a line of integer input in Java and avoid Scanner blocking Mar 06, 2026 am 06:21 AM

This article introduces typical blocking problems when using Scanner to read multiple integers in a single line. It points out that hasNextInt() will wait indefinitely when there is no subsequent input, and recommends a safe alternative with nextLine() string splitting as the core.

Related articles