MockTwitterStreamRecorded.java
package kafka_tweetoscope.tweetsProducer;
import com.google.gson.Gson;
import com.google.gson.GsonBuilder;
import com.google.gson.JsonArray;
import com.google.gson.JsonDeserializer;
import com.google.gson.JsonElement;
import com.google.gson.JsonObject;
import com.twitter.clientlib.model.Tweet;
import org.apache.kafka.clients.producer.ProducerRecord;
import java.io.BufferedReader;
import java.io.IOException;
import java.nio.charset.StandardCharsets;
import java.io.InputStream;
import java.io.InputStreamReader;
import java.time.Instant;
import java.time.OffsetDateTime;
import java.time.ZoneOffset;
public class MockTwitterStreamRecorded extends AbstractTweetsProducer {
protected String fileName;
public MockTwitterStreamRecorded(String bootstrapServer, String topicName, String fileName) {
super(bootstrapServer, topicName);
switch (fileName) {
case "large":
this.fileName = "/largeTestBase.txt";
break;
case "scenario":
this.fileName = "/scenarioTestBase.txt";
break;
case "mini":
this.fileName = "/miniTestBase.txt";
break;
default:
throw new IllegalArgumentException(
"heoooo c'est pas le bon arg"
);
}
}
private String readResourceFile() throws IOException {
try (InputStream inputStream = getClass().getResourceAsStream(this.fileName);
BufferedReader reader = new BufferedReader(new InputStreamReader(inputStream, StandardCharsets.UTF_8))) {
if (inputStream == null) {
throw new IOException("Resource not found: " + this.fileName);
}
StringBuilder content = new StringBuilder();
String line;
while ((line = reader.readLine()) != null) {
content.append(line);
}
return content.toString();
}
}
public static void main(String[] args) {
MockTwitterStreamRecorded tweetsProducer = new MockTwitterStreamRecorded("localhost:9092", "raw-tweets", "large");
tweetsProducer.run();
}
@Override
public void run() {
try {
// Use readResourceFile() instead of Files.readAllBytes()
String jsonString = readResourceFile();
Gson gson = new GsonBuilder()
.registerTypeAdapter(OffsetDateTime.class,
(JsonDeserializer<OffsetDateTime>) (json, type, context) ->
Instant.parse(json.getAsString()).atOffset(ZoneOffset.UTC))
.create();
JsonObject jsonObject = gson.fromJson(jsonString, JsonObject.class);
JsonArray prerecordedtweets = jsonObject.getAsJsonArray("tweets");
int processed = 0;
int total = prerecordedtweets.size();
for (JsonElement je : prerecordedtweets) {
try {
Tweet tweet = gson.fromJson(je, Tweet.class);
producer.send(new ProducerRecord<Void, Tweet>(topicName, null, tweet));
processed++;
if (processed % 100 == 0) {
System.out.println("Processed " + processed + " of " + total + " tweets");
}
} catch (Exception e) {
System.err.println("Error processing tweet: Skipping");
System.err.println(e.getMessage());
}
}
System.out.println("Finished processing " + processed + " out of " + total + " tweets");
} catch (IOException e) {
System.err.println("Error reading file: " + fileName);
e.printStackTrace();
} finally {
producer.close();
}
}
}