KubeCraftJobs

DevOps & Cloud Job Board

Kafka developer

Tata Consultancy Services

Bengaluru, Karnataka

On-site
Mid Level
Full Time
Posted January 05, 2026

Tech Stack

kafka kafka-streams java scala python confluent amazon-web-services avro apache-avro protobuf docker kubernetes hadoop apache-hadoop spark prometheus grafana kafka-manager

Please log in or register to view job application links.

Job Description

Dear Candidates, Greetings from TCS!!! TCS is looking for Kafka developer Experience: 6-10 years Location: Hyderabad/Pune/Bangalore/Chennai/Kolkata Required Technical Skill Set: Kafka developer Job description: We are seeking a skilled and motivated Mid-Level Kafka Developer to join our team. The ideal candidate will have a solid background in building scalable, high-throughput data pipelines and real-time streaming solutions using Apache Kafka. You will work closely with architects, engineers, and product managers to implement event-driven architectures and integrate Kafka with various systems. Key Responsibilities: - Design, develop, and maintain real-time streaming solutions using Apache Kafka. - Implement Kafka producers, consumers, topics, and stream processing logic. - Develop and maintain Kafka Connectors and Kafka Streams applications. - Optimize Kafka clusters for performance, scalability, and reliability. - Collaborate with DevOps teams for Kafka deployment, monitoring, and alerting. - Integrate Kafka with data sources like databases, microservices, and cloud services. - Participate in code reviews, design discussions, and sprint planning. - Troubleshoot and resolve issues related to message delivery, data loss, and system failures. Primary skills: - 6+ years of software development experience. - Hands-on experience with Apache Kafka. - Strong knowledge of Kafka architecture (brokers, zookeepers, partitions, offsets, replication). - Experience with Kafka Streams, Kafka Connect, or ksqlDB. - Proficiency in Java, Scala, or Python. - Understanding of distributed systems and real-time data processing. - Experience with RESTful APIs and microservices architecture Good to have skill: - Experience with Kafka on Confluent Platform or AWS MSK. - Familiarity with Schema Registry, Avro/Protobuf, and serialization formats. - Knowledge of containerization (Docker, Kubernetes). - Experience integrating Kafka with Big Data technologies (e.g., Hadoop, Spark, Flink). - Familiarity with monitoring tools like Prometheus, Grafana, Kafka Manager, or Confluent Control Center