MockTwitterStreamRecorded.java

package kafka_tweetoscope.tweetsProducer;

import com.google.gson.Gson;
import com.google.gson.GsonBuilder;
import com.google.gson.JsonArray;
import com.google.gson.JsonDeserializer;
import com.google.gson.JsonElement;
import com.google.gson.JsonObject;
import com.twitter.clientlib.model.Tweet;
import org.apache.kafka.clients.producer.ProducerRecord;

import java.io.BufferedReader;
import java.io.IOException;
import java.nio.charset.StandardCharsets;
import java.io.InputStream;
import java.io.InputStreamReader;
import java.time.Instant;
import java.time.OffsetDateTime;
import java.time.ZoneOffset;

public class MockTwitterStreamRecorded extends AbstractTweetsProducer {
    protected String fileName;

    public MockTwitterStreamRecorded(String bootstrapServer, String topicName, String fileName) {
        super(bootstrapServer, topicName);
        switch (fileName) {
            case "large":
                this.fileName = "/largeTestBase.txt";
                break;
            case "scenario":
                this.fileName = "/scenarioTestBase.txt";
                break;
            case "mini":
                this.fileName = "/miniTestBase.txt";
                break;
            default:
                throw new IllegalArgumentException(
                    "heoooo c'est pas le bon arg"
                );
        }
    }
    
    private String readResourceFile() throws IOException {
        try (InputStream inputStream = getClass().getResourceAsStream(this.fileName);
             BufferedReader reader = new BufferedReader(new InputStreamReader(inputStream, StandardCharsets.UTF_8))) {
            if (inputStream == null) {
                throw new IOException("Resource not found: " + this.fileName);
            }
            
            StringBuilder content = new StringBuilder();
            String line;
            while ((line = reader.readLine()) != null) {
                content.append(line);
            }
            return content.toString();
        }
    }

    public static void main(String[] args) {
        MockTwitterStreamRecorded tweetsProducer = new MockTwitterStreamRecorded("localhost:9092", "raw-tweets", "large");
        tweetsProducer.run();
    }

    @Override
    public void run() {
        try {
            // Use readResourceFile() instead of Files.readAllBytes()
            String jsonString = readResourceFile();
            Gson gson = new GsonBuilder()
                .registerTypeAdapter(OffsetDateTime.class,
                    (JsonDeserializer<OffsetDateTime>) (json, type, context) ->
                        Instant.parse(json.getAsString()).atOffset(ZoneOffset.UTC))
                .create();
            
            JsonObject jsonObject = gson.fromJson(jsonString, JsonObject.class);
            JsonArray prerecordedtweets = jsonObject.getAsJsonArray("tweets");

            int processed = 0;
            int total = prerecordedtweets.size();

            for (JsonElement je : prerecordedtweets) {
                try {
                    Tweet tweet = gson.fromJson(je, Tweet.class);
                    producer.send(new ProducerRecord<Void, Tweet>(topicName, null, tweet));
                    
                    processed++;
                    if (processed % 100 == 0) {
                        System.out.println("Processed " + processed + " of " + total + " tweets");
                    }
                } catch (Exception e) {
                    System.err.println("Error processing tweet: Skipping");
                    System.err.println(e.getMessage());
                }
            }
            
            System.out.println("Finished processing " + processed + " out of " + total + " tweets");

        } catch (IOException e) {
            System.err.println("Error reading file: " + fileName);
            e.printStackTrace();
        } finally {
            producer.close();
        }
    }
}