X-Git-Url: https://gerrit.o-ran-sc.org/r/gitweb?a=blobdiff_plain;f=dmaap-adaptor-java%2Fsrc%2Fmain%2Fjava%2Forg%2Foran%2Fdmaapadapter%2Ftasks%2FKafkaTopicConsumers.java;h=0ed85c6a1b1262d76fca69ec325d65077436ddfc;hb=46a0fd717e5f49ebae6cb2c4fbcf54f0e329dc86;hp=23d9da2c9294c60e9bf33e7be4c50bbf578b89e4;hpb=968b89731a192c2ee3f3393d00519879ad89ce56;p=nonrtric.git diff --git a/dmaap-adaptor-java/src/main/java/org/oran/dmaapadapter/tasks/KafkaTopicConsumers.java b/dmaap-adaptor-java/src/main/java/org/oran/dmaapadapter/tasks/KafkaTopicConsumers.java index 23d9da2c..0ed85c6a 100644 --- a/dmaap-adaptor-java/src/main/java/org/oran/dmaapadapter/tasks/KafkaTopicConsumers.java +++ b/dmaap-adaptor-java/src/main/java/org/oran/dmaapadapter/tasks/KafkaTopicConsumers.java @@ -23,57 +23,97 @@ package org.oran.dmaapadapter.tasks; import java.util.HashMap; import java.util.Map; +import lombok.Getter; + import org.oran.dmaapadapter.configuration.ApplicationConfig; import org.oran.dmaapadapter.repository.InfoType; import org.oran.dmaapadapter.repository.InfoTypes; import org.oran.dmaapadapter.repository.Job; +import org.oran.dmaapadapter.repository.Jobs; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.springframework.beans.factory.annotation.Autowired; +import org.springframework.scheduling.annotation.EnableScheduling; +import org.springframework.scheduling.annotation.Scheduled; import org.springframework.stereotype.Component; -import reactor.core.Disposable; -/** - * The class fetches incoming requests from DMAAP and sends them further to the - * consumers that has a job for this InformationType. - */ @SuppressWarnings("squid:S2629") // Invoke method(s) only conditionally @Component +@EnableScheduling public class KafkaTopicConsumers { private static final Logger logger = LoggerFactory.getLogger(KafkaTopicConsumers.class); - private final Map topicConsumers = new HashMap<>(); - private final Map activeSubscriptions = new HashMap<>(); - private final ApplicationConfig appConfig; + private final Map topicListeners = new HashMap<>(); // Key is typeId - public KafkaTopicConsumers(@Autowired ApplicationConfig appConfig) { - this.appConfig = appConfig; - } + @Getter + private final Map consumers = new HashMap<>(); // Key is jobId + + private static final int CONSUMER_SUPERVISION_INTERVAL_MS = 1000 * 60 * 3; + + public KafkaTopicConsumers(@Autowired ApplicationConfig appConfig, @Autowired InfoTypes types, + @Autowired Jobs jobs) { - public void start(InfoTypes types) { for (InfoType type : types.getAll()) { if (type.isKafkaTopicDefined()) { - KafkaTopicConsumer topicConsumer = new KafkaTopicConsumer(appConfig, type); - topicConsumers.put(type.getId(), topicConsumer); + KafkaTopicListener topicConsumer = new KafkaTopicListener(appConfig, type); + topicListeners.put(type.getId(), topicConsumer); } } + + jobs.addObserver(new Jobs.Observer() { + @Override + public void onJobbAdded(Job job) { + addJob(job); + } + + @Override + public void onJobRemoved(Job job) { + removeJob(job); + } + + }); } public synchronized void addJob(Job job) { - if (this.activeSubscriptions.get(job.getId()) == null && job.getType().isKafkaTopicDefined()) { + if (this.consumers.get(job.getId()) == null && job.getType().isKafkaTopicDefined()) { logger.debug("Kafka job added {}", job.getId()); - KafkaTopicConsumer topicConsumer = topicConsumers.get(job.getType().getId()); - Disposable subscription = topicConsumer.startDistributeToConsumer(job); - activeSubscriptions.put(job.getId(), subscription); + KafkaTopicListener topicConsumer = topicListeners.get(job.getType().getId()); + KafkaJobDataConsumer subscription = new KafkaJobDataConsumer(job); + subscription.start(topicConsumer.getOutput()); + consumers.put(job.getId(), subscription); } } public synchronized void removeJob(Job job) { - Disposable d = activeSubscriptions.remove(job.getId()); + KafkaJobDataConsumer d = consumers.remove(job.getId()); if (d != null) { logger.debug("Kafka job removed {}", job.getId()); - d.dispose(); + d.stop(); } } + @Scheduled(fixedRate = CONSUMER_SUPERVISION_INTERVAL_MS) + public synchronized void restartNonRunningTasks() { + + for (KafkaJobDataConsumer consumer : consumers.values()) { + if (!consumer.isRunning()) { + restartTopic(consumer); + } + } + } + + private void restartTopic(KafkaJobDataConsumer consumer) { + InfoType type = consumer.getJob().getType(); + KafkaTopicListener topic = this.topicListeners.get(type.getId()); + topic.start(); + restartConsumersOfType(topic, type); + } + + private void restartConsumersOfType(KafkaTopicListener topic, InfoType type) { + this.consumers.forEach((jobId, consumer) -> { + if (consumer.getJob().getType().getId().equals(type.getId())) { + consumer.start(topic.getOutput()); + } + }); + } }