Job description
Job Title : Kafka Developer
Location : Noida/ Chennai/ Pune
Education: B.E. / B.Tech.
Year of Experience: 8-10 Years
About us
At R Systems we are shaping the future of technology by designing cutting-edge software #products, platforms, and #digital experiences that drive business growth for our clients.
Our product mindset and advanced engineering capabilities in #Cloud, Data, #AI, and Customer Experience empower us to deliver innovative solutions to key players across the high-tech industry.
This includes ISVs, SaaS, and Internet companies, as well as leading organizations in #telecom, #media, #healthcare, #finance, and #manufacturing.
We are Great Place To Work Certified in 10 countries where we have a full-time workforce - India, the USA, Canada, Poland, Romania, Moldova, Indonesia, Singapore, Malaysia, and Thailand.
Mandatory Requirements:
- A degree in Computer Science, IT, or a related field.
- Kafka Expertise – Strong knowledge of Kafka architecture, brokers, producers, consumers, and stream processing.
- Programming Skills – Proficiency in Java, Scala, or Python for developing Kafka-based applications.
- Big Data & Streaming Technologies – Experience with Spark, Flink, or Apache Storm is a plus.
- Database Knowledge – Familiarity with SQL and NoSQL databases like Cassandra, MongoDB, or PostgreSQL.
- Cloud & DevOps – Experience with cloud platforms (AWS, Azure, GCP) and Kubernetes/Docker.
- Event-Driven Architecture – Understanding of event-driven and microservices architectures.
- Monitoring & Debugging – Experience with Kafka monitoring tools like Confluent Control Center, Kafka Manager, or ELK stack.
- Security & Scalability – Knowledge of Kafka security, access control, and scaling strategies.
- Problem-Solving & Communication – Strong analytical skills and ability to work in cross-functional teams
Roles & Responsibilities:
- Kafka Application Development – Design, develop, and maintain real-time data streaming applications using Apache Kafka.
- Topic Management – Create and manage Kafka topics, partitions, and replication factors.
- Data Pipeline Development – Build and optimize data pipelines for real-time data processing.
- Producer & Consumer Implementation – Develop Kafka producers and consumers for seamless data exchange between systems.
- Integration & Connectivity – Integrate Kafka with databases, cloud services, and other messaging platforms.
- Performance Optimization – Monitor and fine-tune Kafka clusters for low latency and high throughput.
- Security & Compliance – Implement security best practices, including SSL, SASL, and authentication mechanisms.
- Cluster Administration – Manage Kafka clusters, brokers, ZooKeeper, and ensure high availability.
- Monitoring & Logging – Use tools like Prometheus, Grafana, or Kafka Manager for monitoring and troubleshooting.
- Documentation & Best Practices – Maintain documentation for Kafka configurations, data flow, and architectural decisions.
Mandatory Skills
Java, Apache Kafka, Kafka Streams, Kafka, Apache Zookeeper, Kafka Cluster, Kafka Administrator, Microservices
Required Skill Profession
Other General