I have the following scenario:
- 4 wearable sensors attached on individuals.
- Potentially infinite individuals.
- A Kafka cluster.
I have to perform real-time processing on data streams on a cluster with a running instance of apache flink. Kafka is the data hub between flink cluster and sensors. Moreover, subject's streams are totally independent and also different streams belonging to same subject are independent each other.
I imagine this setup in my mind: I set a specific topic for each subject and each topic is partitioned in 4 partition, each one for each sensor on specific person. In this way I though to establish a consumer group for every topic.
Actually, my data amount is not so much big but mine interest is to build an easily scalable system. A day maybe I can have hundreds of individuals for instance...
My questions are:
- Is this setup good? What do you think about it?
- In this way I will have 4 kafka broker and each one handles a partition, right (without consider potential backups)?
Destroy me guys, and thanks in advance