About the Role    
We are looking for a highly skilled Backend Engineer   with a strong background in Python, system design, and infrastructure   to join our team.
You will be responsible for designing, building, and maintaining scalable backend systems, while collaborating with cross-functional teams to deliver robust and efficient solutions.
 
This role requires someone who can think end-to-end  , from designing high-level architecture, implementing core services, to ensuring production-grade reliability and performance.
 
  
Key Responsibilities    
- Develop and maintain backend services and APIs using Python   & NodeJs     
- Design scalable, resilient, and maintainable systems, focusing on system architecture and distributed systems  .
 
 
- Integrate AI and large language models (LLMs)   into applications, ensuring performance, scalability, and cost-efficiency.
 
 
- Collaborate with AI/ML teams to deploy models into production pipelines.
 
 
- Optimize infrastructure for AI workloads (GPU usage, caching, batch processing)  
- Build and maintain monitoring, logging, and observability   for AI-powered systems.
 
 
- Troubleshoot and resolve issues in production systems while maintaining high reliability.
 
 
- Participate in design and code reviews, and drive engineering best practices across the team.
 
 
- Automate deployment pipelines for backend and AI services (CI/CD, IaC).
 
 
  
Required Skills & Qualifications    
- Strong experience in Python   (FastAPI (most-preferred), Flask, Django, or similar) or NodeJS   (Express (most-preferred), Fastify or similar)  
- Solid understanding of system design principles  : scalability, fault tolerance, distributed systems.
 
 
- Experience with infrastructure and DevOps  : Docker, Kubernetes, Terraform, CI/CD pipelines.
 
 
- Hands-on experience with cloud platforms   (AWS, Azure, GCP), especially for AI workloads.
 
 
- Knowledge of databases   (SQL & NoSQL) and caching systems (Redis, Memcached).
 
 
- Experience integrating LLMs   or AI APIs into production systems (OpenAI, HuggingFace, LangChain, etc.).
 
 
- Familiarity with messaging/streaming systems   (Kafka, RabbitMQ).
 
 
- Monitoring and observability experience (Prometheus, Grafana, ELK).
 
 
- Strong problem-solving, debugging, and analytical skills.
 
 
- Excellent communication and collaboration skills.
 
 
  
  
Nice to Have    
- Experience with generative AI pipelines  , vector databases, and embeddings.
 
 
- Familiarity with ML Ops tools   (MLflow, BentoML, Ray Serve, etc.).
 
 
- Knowledge of event-driven architectures   and microservices.
 
 
- Prior experience in AI/LLM-focused startups or high-scale AI systems  .
 
 
  
What We Offer    
- Opportunity to work on challenging, large-scale systems   with real-world impact.
 
 
- Collaborative team culture with focus on learning and innovation  .
 
 
- Competitive compensation and growth opportunities.