50+ AWS (Amazon Web Services) Jobs in Bangalore (Bengaluru) | AWS (Amazon Web Services) Job openings in Bangalore (Bengaluru)
Apply to 50+ AWS (Amazon Web Services) Jobs in Bangalore (Bengaluru) on CutShort.io. Explore the latest AWS (Amazon Web Services) Job opportunities across top companies like Google, Amazon & Adobe.

🚀 Hiring: Automation Tester
⭐ Experience: 6+ Years
📍 Location: Bangalore
⭐ Work Mode:- Hybrid
⏱️ Notice Period: Immediate Joiners
(Only immediate joiners & candidates serving notice period)
We’re looking for a skilled Automation Tester with a strong background in Java, Selenium, and AWS to join our growing QA team. If you're passionate about automation, love solving problems, and thrive in a dynamic environment, we’d love to hear from you!
Must-Have Skills:
✅ Strong hands-on experience in Automation Testing
✅ Proficiency in Java for writing test scripts
✅ Expertise in Selenium WebDriver
✅ Working knowledge of AWS services relevant to test environments
✅ Solid understanding of the SDLC and Agile methodologies
The candidate should have extensive experience in designing and developing scalable data pipelines and real-time data processing solutions. As a key member of the team, the Senior Data Engineer will play a critical role in building end-to-end data workflows, supporting machine learning model deployment, and driving MLOps practices in a fast-paced, agile environment. Strong expertise in Apache Kafka, Apache Flink, AWS SageMaker, and Terraform is essential. Additional experience with infrastructure automation and CI/CD for ML models is a significant advantage.
Key Responsibilities
- Design, develop, and maintain high-performance ETL and real-time data pipelines using Apache Kafka and Apache Flink.
- Build scalable and automated MLOps pipelines for training, validation, and deployment of models using AWS SageMaker and associated services.
- Implement and manage Infrastructure as Code (IaC) using Terraform to provision and manage AWS environments.
- Collaborate with data scientists, ML engineers, and DevOps teams to streamline model deployment workflows and ensure reliable production delivery.
- Optimize data storage and retrieval strategies for large-scale structured and unstructured datasets.
- Develop data transformation logic and integrate data from various internal and external sources into data lakes and warehouses.
- Monitor, troubleshoot, and enhance performance of data systems in a cloud-native, fast-evolving production setup.
- Ensure adherence to data governance, privacy, and security standards across all data handling activities.
- Document data engineering solutions and workflows to facilitate cross-functional understanding and ongoing maintenance.
Proficient with Node.js, Javascript: server-side applications;
framework: such as Express.js;
others: Restful API, DB(relational), Git;
strong in communication
[Prefer to have]
cloud services: such as AWS, Azure, or GCP;
container: like Docker;
microservice; CICD; mentor/leader experience.

Key Responsibilities
- Design and implement ETL/ELT pipelines using Databricks, PySpark, and AWS Glue
- Develop and maintain scalable data architectures on AWS (S3, EMR, Lambda, Redshift, RDS)
- Perform data wrangling, cleansing, and transformation using Python and SQL
- Collaborate with data scientists to integrate Generative AI models into analytics workflows
- Build dashboards and reports to visualize insights using tools like Power BI or Tableau
- Ensure data quality, governance, and security across all data assets
- Optimize performance of data pipelines and troubleshoot bottlenecks
- Work closely with stakeholders to understand data requirements and deliver actionable insights
🧪 Required Skills
Skill AreaTools & TechnologiesCloud PlatformsAWS (S3, Lambda, Glue, EMR, Redshift)Big DataDatabricks, Apache Spark, PySparkProgrammingPython, SQLData EngineeringETL/ELT, Data Lakes, Data WarehousingAnalyticsData Modeling, Visualization, BI ReportingGen AI IntegrationOpenAI, Hugging Face, LangChain (preferred)DevOps (Bonus)Git, Jenkins, Terraform, Docker
📚 Qualifications
- Bachelor's or Master’s degree in Computer Science, Data Science, or related field
- 3+ years of experience in data engineering or data analytics
- Hands-on experience with Databricks, PySpark, and AWS
- Familiarity with Generative AI tools and frameworks is a strong plus
- Strong problem-solving and communication skills
🌟 Preferred Traits
- Analytical mindset with attention to detail
- Passion for data and emerging technologies
- Ability to work independently and in cross-functional teams
- Eagerness to learn and adapt in a fast-paced environment

It's a Product based company, they are in to network domain

Your Role
- Develop and implement test strategies for Software-as-a-Service offering in hybrid cloud environments
- Create automated tests and tools to ensure release qualification
- Simulate and test large-scale customer scenarios
- Find and track bugs, providing RCA and assisting with closure
- Collaborate with top-notch engineers to influence release quality
- Stay up-to-date with latest datacenter and cloud technologies
- Work with computing/storage/networking services of Azure and AWS
- Strong knowledge of operating systems and TCP-IP protocols
What You Will Bring
- Strong automation skills using Python
- Experience with computing/storage/networking services of Azure and AWS
- Software development or test automation experience in Python,
- Strong understanding of operating systems
- Knowledge of TCP-IP Suite of protocols
- Prior experience with large-scale system testing
- Understanding of virtualisation concepts
- Ability to debug complex distributed applications

5+ years of hands-on strong experience in design, develop and implementation of complex software solutions using Python fast API
•LLM experience with closed source LLMs like GPT 3.5, 4, 4 – Turbo, CoHere etc
. •Excellent hands on expertise in Python coding, Pep 8 standards & best practices/design patterns, Object Oriented Design methodologies
•Hands on experience of deploying as AI/ML solutions as a service/Rest API endpoints on Cloud or Kubernetes
•Experience with development methodologies, experience in writing unit tests in Python
•Some experience with Open Source LLMs like LLAMA2, Falcon and fine tuning will be preferred.
•Experience with using PostGreSQL or any other RDBMS and SQL query writing. •Familiar with GIT Lab, CI/CD and Swagger
•Strong problem-solving and communication skills, ability to work independently and as part of Agile teams

Key Responsibilities:
· Design, develop, and maintain high-quality, scalable software solutions
· Work extensively with AWS Serverless services such as Lambda, API Gateway, and DynamoDB
· Implement and optimize Elasticsearch for search and analytics
· Build backend services using NodeJS and develop front-end applications using Angular
· Design and develop robust REST APIs with a solid understanding of web services
· Write clean, efficient, and maintainable code in JavaScript and TypeScript
· Troubleshoot, debug, and resolve performance issues in complex systems
· Collaborate with cross-functional teams to deliver robust, end-to-end software solutions
Required Skills & Qualifications:
· Bachelor’s degree in Computer Science or a related field
· 8+ years of experience in software development
· Strong hands-on expertise in AWS Serverless (Lambda, API Gateway, DynamoDB) or .NET with SQL
· Experience with Elasticsearch for efficient data retrieval and indexing
· Proficiency in NodeJS, Angular, JavaScript, and TypeScript
· Solid understanding of REST APIs and web services
· Strong UI development skills with CSS
· Proven experience designing solutions for complex technical requirements
· Excellent debugging, analytical, and problem-solving abilities

Job Title: Senior Data Engineer
Location: Bangalore | Hybrid
Company: krtrimaIQ Cognitive Solutions
Role Overview:
As a Senior Data Engineer, you will design, build, and optimize robust data foundations and end-to-end solutions to unlock maximum value from data across the organization. You will play a key role in fostering data-driven thinking, not only within the IT function but also among broader business stakeholders. You will serve as a technology and subject matter expert, providing mentorship to junior engineers and translating the company’s vision and Data Strategy into actionable, high-impact IT solutions.
Key Responsibilities:
- Design, develop, and implement scalable data solutions to support business objectives and drive digital transformation.
- Serve as a subject matter expert in data engineering, providing guidance and mentorship to junior team members.
- Enable and promote data-driven culture throughout the organization, engaging both technical and business stakeholders.
- Lead the design and delivery of Data Foundation initiatives, ensuring adoption and value realization across business units.
- Collaborate with business and IT teams to capture requirements, design optimal data models, and deliver high-value insights.
- Manage and drive change management, incident management, and problem management processes related to data platforms.
- Present technical reports and actionable insights to stakeholders and leadership teams, acting as the expert in Data Analysis and Design.
- Continuously improve efficiency and effectiveness of solution delivery, driving down costs and reducing implementation times.
- Contribute to organizational knowledge-sharing and capability building (e.g., Centers of Excellence, Communities of Practice).
- Champion best practices in code quality, DevOps, CI/CD, and data governance throughout the solution lifecycle.
Key Characteristics:
- Technology expert with a passion for continuous learning and exploring multiple perspectives.
- Deep expertise in the data engineering/technology domain, with hands-on experience across the full data stack.
- Excellent communicator, able to bridge the gap between technical teams and business stakeholders.
- Trusted leader, respected across levels for subject matter expertise and collaborative approach.
Mandatory Skills & Experience:
- Mastery in public cloud platforms: AWS, Azure, SAP
- Mastery in ELT (Extract, Load, Transform) operations
- Advanced data modeling expertise for enterprise data platforms
Hands-on skills:
- Data Integration & Ingestion
- Data Manipulation and Processing
- Source/version control and DevOps tools: GITHUB, Actions, Azure DevOps
- Data engineering/data platform tools: Azure Data Factory, Databricks, SQL Database, Synapse Analytics, Stream Analytics, AWS Glue, Apache Airflow, AWS Kinesis, Amazon Redshift, SonarQube, PyTest
- Experience building scalable and reliable data pipelines for analytics and other business applications
Optional/Preferred Skills:
- Project management experience, especially running or contributing to Scrum teams
- Experience working with BPC (Business Planning and Consolidation), Planning tools
- Exposure to working with external partners in the technology ecosystem and vendor management
What We Offer:
- Opportunity to leverage cutting-edge technologies in a high-impact, global business environment
- Collaborative, growth-oriented culture with strong community and knowledge-sharing
- Chance to influence and drive key data initiatives across the organization
About HelloRamp.ai
HelloRamp is on a mission to revolutionize media creation for automotive and retail using AI. Our platform powers 3D/AR experiences for leading brands like Cars24, Spinny, and Samsung. We’re now building the next generation of Computer Vision + AI products, including cutting-edge NeRF pipelines and AI-driven video generation.
What You’ll Work On
- Develop and optimize Computer Vision pipelines for large-scale media creation.
- Implement NeRF-based systems for high-quality 3D reconstruction.
- Build and fine-tune AI video generation models using state-of-the-art techniques.
- Optimize AI inference for production (CUDA, TensorRT, ONNX).
- Collaborate with the engineering team to integrate AI features into scalable cloud systems.
- Research latest AI/CV advancements and bring them into production.
Skills & Experience
- Strong Python programming skills.
- Deep expertise in Computer Vision and Machine Learning.
- Hands-on with PyTorch/TensorFlow.
- Experience with NeRF frameworks (Instant-NGP, Nerfstudio, Plenoxels) and/or video synthesis models.
- Familiarity with 3D graphics concepts (meshes, point clouds, depth maps).
- GPU programming and optimization skills.
Nice to Have
- Knowledge of Three.js or WebGL for rendering AI outputs on the web.
- Familiarity with FFmpeg and video processing pipelines.
- Experience in cloud-based GPU environments (AWS/GCP).
Why Join Us?
- Work on cutting-edge AI and Computer Vision projects with global impact.
- Join a small, high-ownership team where your work matters.
- Opportunity to experiment, publish, and contribute to open-source.
- Competitive pay and flexible work setup.

Job Type : Contract
Location : Bangalore
Experience : 5+yrs
The role focuses on cloud security engineering with a strong emphasis on GCP, while also covering AWS and Azure.
Required Skills:
- 5+ years of experience in software and/or cloud platform engineering, particularly focused on GCP environment.
- Knowledge of the Shared Responsibility Model; keen understanding of the security risks inherent in hosting cloud-based applications and data.
- Experience developing across the security assurance lifecycle (including prevent, detect, respond, and remediate controls)?Experience in configuring Public Cloud native security tooling and capabilities with a focus on Google Cloud Organizational policies/constraints, VPC SC, IAM policies and GCP APIs.
- Experience with Cloud Security Posture Management (CSPM) 3rd Party tools such as Wiz, Prisma, Check Point CloudGuard, etc.
- Experience in Policy-as-code (Rego) and OPA platform.
- Experience solutioning and configuring event-driven serverless-based security controls in Azure, including but not limited to technologies such as Azure Function, Automation Runbook, AWS Lambda and Google Cloud Functions.
- Deep understanding of DevOps processes and workflows.
- Working knowledge of the Secure SDLC process
- Experience with Infrastructure as Code (IaC) tooling, preferably Terraform.
- Familiarity with Logging and data pipeline concepts and architectures in cloud.
- Strong in scripting languages such as PowerShell or Python or Bash or Go.
- Knowledge of Agile best practices and methodologies
- Experience creating technical architecture documentation.? Excellent communication, written and interpersonal skills.
- Practical experience in designing and configuring CICD pipelines. Practical experience in GitHub Actions and Jenkins.
- Experience in ITSM.
- Ability to articulate complex technical concepts to non-technical stakeholders.
- Experience with risk control frameworks and engagements with risk and regulatory functions
- Experience in the financial industry would be a plus.

Participate and contribute in platform requirements/story development.
Contribute to the design and design alternatives to the requirements/stories and also participate in design reviews.
Involved in Platform Sprint activities.
Development of assigned stories in appropriate languages defined for each module.
Develop Unit test cases and execute them part of continuous integration pipeline.
Participate in peer code reviews
Years of experience needed –
4+ years or more on Golang development.
Technical Skills:
4+ years of development experience on Golang based development projects
Mandatory Skills – Golang, AWS , JavaScript, CSS, HTML
Must have experience in design and development of front end and back-end services for various business processes
Good experience on Java, Spring boot & Microservices
Experience using versioning controls such as GitHub
Experience in working with SQL & NOSQL databases
Understanding of containerization technologies such as docker, Kubernetes
Good knowledge and hands-on on Unit Testing and available Test Frameworks
Role Overview
We are seeking a DevOps Engineer with 2 years of experience to join our innovative team. The ideal
candidate will bridge the gap between development and operations, implementing and maintaining our
cloud infrastructure while ensuring secure deployment pipelines and robust security practices for our
client projects.
Responsibilities:
- Design, implement, and maintain CI/CD pipelines.
- Containerize applications using Docker and orchestrate deployments
- Manage and optimize cloud infrastructure on AWS and Azure platforms
- Monitor system performance and implement automation for operational tasks to ensure optimal
- performance, security, and scalability.
- Troubleshoot and resolve infrastructure and deployment issues
- Create and maintain documentation for processes and configurations
- Collaborate with cross-functional teams to gather requirements, prioritise tasks, and contribute to project completion.
- Stay informed about emerging technologies and best practices within the fields of DevOps and cloud computing.
Requirements:
- 2+ years of hands-on experience with AWS cloud services
- Strong proficiency in CI/CD pipeline configuration
- Expertise in Docker containerisation and container management
- Proficiency in shell scripting (Bash/Power-Shell)
- Working knowledge of monitoring and logging tools
- Knowledge of network security and firewall configuration
- Strong communication and collaboration skills, with the ability to work effectively within a team
- environment
- Understanding of networking concepts and protocols in AWS and/or Azure

Quidcash is seeking a highly skilled and passionate Engineering Manager to lead, mentor, and grow a talented team of software engineers. You will be instrumental in shaping our technical direction, driving the development of our core products, and championing engineering excellence. This is a hands-on leadership role where you'll contribute to architectural decisions, foster a culture of innovation, and ensure your team is equipped to build scalable, robust, and intelligent systems.
If you're a leader who thrives on technical challenges, loves building high-performing teams, and is excited by the potential of AI/ML in fintech, we want to hear from you!
What You'll Do:
· Lead & Mentor: Manage, coach, and develop a team of software engineers, data scientists fostering an inclusive, collaborative, and high-performance culture. Conduct regular 1:1s, performance reviews, and support career growth.
· Technical Leadership: Provide strong technical guidance on architecture, design, and implementation of complex systems, particularly in microservices, OOPS principles, and cloud-native applications.
· AI/ML Integration: Drive the strategy and execution for integrating AI/ML models and techniques into our products and platforms, working closely with data scientists and engineers.
· Engineering Best Practices: Establish, evangelize, and enforce best practices for software development, including code quality, testing (unit, integration, E2E), CI/CD, security, and documentation.
· DevOps Culture: Champion and implement DevOps principles to improve deployment frequency, system reliability, and operational efficiency. Oversee CI/CD pipelines and infrastructure-as-code practices.
· Roadmap & Execution: Collaborate with Product Management, Design, and other stakeholders to define the technical roadmap, translate product requirements into actionable engineering tasks, and ensure timely delivery of high-quality software.
· Architectural Vision: Contribute to and influence the long-term architectural vision for Quidcash platforms, ensuring scalability, resilience, and maintainability.
· Problem Solving: Dive deep into complex technical challenges, lead troubleshooting efforts, and make critical technical decisions.
· Recruitment & Team Building: Actively participate in recruiting, interviewing, and onboarding new engineering talent.
What You'll Bring (Must-Haves):
· Experience:
o Proven experience (6+ years) in software engineering, with a strong foundation in Object-Oriented Programming (OOP/S) using languages like Java, Python, C#, Go, or similar.
o Demonstrable experience (2+ years) in an engineering leadership or management role, directly managing a team of engineers.
· Technical Acumen:
o Deep understanding and practical experience with microservice architecture, including design patterns, deployment strategies (e.g., Kubernetes, Docker), and inter-service communication.
o Solid experience with cloud platforms (AWS, GCP, or Azure).
o Familiarity and practical experience with AI/ML concepts, tools, and their application in real-world products (e.g., machine learning pipelines, model deployment, MLOps).
o Proficiency in establishing and driving DevOps practices (CI/CD, monitoring, alerting, infrastructure automation).
· Leadership & Management:
o Excellent leadership, communication, and interpersonal skills with a proven ability to mentor and grow engineering talent.
o Experience in setting up and enforcing engineering best practices (code reviews, testing methodologies, agile processes).
o Strong project management skills, with experience in Agile/Scrum methodologies.
· Mindset:
o A proactive, problem-solving attitude with a passion for continuous improvement.
o Ability to thrive in a fast-paced, dynamic startup environment.
o Strong business acumen and ability to align technical strategy with company goals.
Nice-to-Haves:
· Experience in the FinTech or financial services, lending industry.
· Hands-on experience with specific AI/ML frameworks (e.g., TensorFlow, PyTorch, scikit-learn).
· Experience with event-driven architectures (e.g., Kafka, RabbitMQ).
· Contributions to open-source projects or a strong public technical presence.
· Advanced degree (M.S. or Ph.D.) in Computer Science, Engineering, or a related field.
Why Join Quidcash?
· Impact: Play a pivotal role in shaping a product that directly impacts Indian SMEs business growth.
· Innovation: Work with cutting-edge technologies, including AI/ML, in a forward-thinking environment.
· Growth: Opportunities for professional development and career advancement in a growing company.
· Culture: Be part of a collaborative, supportive, and brilliant team that values every contribution.
· Benefits: Competitive salary, comprehensive benefits package, be a part of the next fin-tech evolution.

Role Overview
We're looking for experienced Data Engineers who can independently design, build, and manage scalable data platforms. You'll work directly with clients and internal teams to develop robust data pipelines that support analytics, AI/ML, and operational systems.
You’ll also play a mentorship role and help establish strong engineering practices across our data projects.
Key Responsibilities
- Design and develop large-scale, distributed data pipelines (batch and streaming)
- Implement scalable data models, warehouses/lakehouses, and data lakes
- Translate business requirements into technical data solutions
- Optimize data pipelines for performance and reliability
- Ensure code is clean, modular, tested, and documented
- Contribute to architecture, tooling decisions, and platform setup
- Review code/design and mentor junior engineers
Must-Have Skills
- Strong programming skills in Python and advanced SQL
- Solid grasp of ETL/ELT, data modeling (OLTP & OLAP), and stream processing
- Hands-on experience with frameworks like Apache Spark, Flink, etc.
- Experience with orchestration tools like Airflow
- Familiarity with CI/CD pipelines and Git
- Ability to debug and scale data pipelines in production
Preferred Skills
- Experience with cloud platforms (AWS preferred, GCP or Azure also fine)
- Exposure to Databricks, dbt, or similar tools
- Understanding of data governance, quality frameworks, and observability
- Certifications (e.g., AWS Data Analytics, Solutions Architect, Databricks) are a bonus
What We’re Looking For
- Problem-solver with strong analytical skills and attention to detail
- Fast learner who can adapt across tools, tech stacks, and domains
- Comfortable working in fast-paced, client-facing environments
- Willingness to travel within India when required

Job Title: Frontend Engineer- Reactjs, Nextjs, MUI
Location: Hybrid weekly ⅔ days WFO (Bengaluru- India)
About the Role:
We're looking for a passionate and skilled Frontend Engineer with 1–3 years of experience to join our growing development team. This role is front-end-heavy, focused on building clean, scalable, and high-performance user interfaces using the latest technologies in the MERN stack—particularly Next.js, React, TypeScript, and Material UI (MUI).
You’ll work alongside a collaborative and talented team to design and build seamless web experiences that delight users. If you're excited about modern frontend architecture and want to grow in a fast-moving, remote-first environment, we'd love to hear from you.
Key Responsibilities:
- Develop responsive, high-performance web applications using Next.js, React, and TypeScript.
- Translate UI/UX designs into functional frontend components using MUI.
- Collaborate with backend developers, designers, and product managers to deliver new features and improvements.
- Ensure code quality through best practices, code reviews, and testing.
- Optimize applications for maximum speed and scalability.
Must-Have Skills:
- 1–3 years of professional experience in frontend development.
- Strong proficiency in React, Next.js, and TypeScript.
- Experience with Material UI (MUI) or similar component libraries.
- Understanding of responsive design, modern frontend tooling, and web performance best practices.
- Familiarity with Git and collaborative workflows.
Nice-to-Have (Bonus) Skills:
- Familiarity with testing libraries (Jest, React Testing Library, Cypress).
- Experience working with design tools like Figma or Adobe XD.
- Basic knowledge of accessibility (a11y) standards and performance optimization.
- Basic experience with Node.js, MongoDB, or working in a MERN stack environment.
- Familiarity with AWS services or cloud deployment practices.
- Experience with RESTful APIs or integrating with backend services.

Job Title: MERN STACK Developer
Location: Hybrid weekly ⅔ days WFO (Bengaluru- India)
About the Role:
We're looking for a passionate and skilled Frontend Engineer with 1–3 years of experience to join our growing development team. This role is front-end-heavy, focused on building clean, scalable, and high-performance user interfaces using the latest technologies in the MERN stack—particularly Next.js, React, TypeScript, and Material UI (MUI).
You’ll work alongside a collaborative and talented team to design and build seamless web experiences that delight users. If you're excited about modern frontend architecture and want to grow in a fast-moving, remote-first environment, we'd love to hear from you.
Key Responsibilities:
- Develop responsive, high-performance web applications using Next.js, React, and TypeScript.
- Translate UI/UX designs into functional frontend components using MUI.
- Collaborate with backend developers, designers, and product managers to deliver new features and improvements.
- Ensure code quality through best practices, code reviews, and testing.
- Optimize applications for maximum speed and scalability.
Must-Have Skills:
- 1–3 years of professional experience in frontend development.
- Strong proficiency in React, Next.js, and TypeScript.
- Experience with Material UI (MUI) or similar component libraries.
- Understanding of responsive design, modern frontend tooling, and web performance best practices.
- Familiarity with Git and collaborative workflows.
Nice-to-Have (Bonus) Skills:
- Familiarity with testing libraries (Jest, React Testing Library, Cypress).
- Experience working with design tools like Figma or Adobe XD.
- Basic knowledge of accessibility (a11y) standards and performance optimization.
- Basic experience with Node.js, MongoDB, or working in a MERN stack environment.
- Familiarity with AWS services or cloud deployment practices.
- Experience with RESTful APIs or integrating with backend services.
Job Title: Backend Engineer - NodeJS, NestJS, and Python
Location: Hybrid weekly ⅔ days WFO (Bengaluru- India)
About the role:
We are looking for a skilled and passionate Senior Backend Developer to join our dynamic team. The ideal candidate should have strong experience in Node.js and NestJS, along with a solid understanding of database management, query optimization, and microservices architecture. As a backend developer, you will be responsible for developing and maintaining scalable backend systems, building robust APIs, integrating databases, and working closely with frontend and DevOps teams to deliver high-quality software solutions.
What You'll Do 🛠️
- Design, develop, and maintain server-side logic using Node.js, NestJS, and Python.
- Develop and integrate RESTful APIs and microservices to support scalable systems.
- Work with NoSQL and SQL databases (e.g., MongoDB, PostgreSQL, MySQL) to create and manage schemas, write complex queries, and optimize performance.
- Collaborate with cross-functional teams including frontend, DevOps, and QA.
- Ensure code quality, maintainability, and scalability through code reviews, testing, and documentation.
- Monitor and troubleshoot production systems, ensuring high availability and performance.
- Implement security and data protection best practices.
What You'll Bring 💼
- 4 to 6 years of professional experience as a backend developer.
- Strong proficiency in Node.js and NestJS framework.
- Good hands-on experience with Python (Django/Flask experience is a plus).
- Solid understanding of relational and non-relational databases.
- Proficient in writing complex NoSQL queries and SQL queries
- Experience with microservices architecture and distributed systems.
- Familiarity with version control systems like Git.
- Basic understanding of containerization (e.g., Docker) and cloud services is a plus.
- Excellent problem-solving skills and a collaborative mindset.
Bonus Points ➕
- Experience with CI/CD pipelines.
- Exposure to cloud platforms like AWS, GCP or Azure.
- Familiarity with event-driven architecture or message brokers (MQTT, Kafka, RabbitMQ)
Why this role matters
You will help build the company from the ground up—shaping our culture and having an impact from Day 1 as part of the foundational team.

Job Title: Sr Dev Ops Engineer
Location: Bengaluru- India (Hybrid work type)
Reports to: Sr Engineer manager
About Our Client :
We are a solution-based, fast-paced tech company with a team that thrives on collaboration and innovative thinking. Our Client's IoT solutions provide real-time visibility and actionable insights for logistics and supply chain management. Cloud-based, AI-enhanced metrics coupled with patented hardware optimize processes, inform strategic decision making and enable intelligent supply chains without the costly infrastructure
About the role : We're looking for a passionate DevOps Engineer to optimize our software delivery and infrastructure. You'll build and maintain CI/CD pipelines for our microservices, automate infrastructure, and ensure our systems are reliable, scalable, and secure. If you thrive on enhancing performance and fostering operational excellence, this role is for you.
What You'll Do 🛠️
- Cloud Platform Management: Administer and optimize AWS resources, ensuring efficient billing and cost management.
- Billing & Cost Optimization: Monitor and optimize cloud spending.
- Containerization & Orchestration: Deploy and manage applications and orchestrate them.
- Database Management: Deploy, manage, and optimize database instances and their lifecycles.
- Authentication Solutions: Implement and manage authentication systems.
- Backup & Recovery: Implement robust backup and disaster recovery strategies, for Kubernetes cluster and database backups.
- Monitoring & Alerting: Set up and maintain robust systems using tools for application and infrastructure health and integrate with billing dashboards.
- Automation & Scripting: Automate repetitive tasks and infrastructure provisioning.
- Security & Reliability: Implement best practices and ensure system performance and security across all deployments.
- Collaboration & Support: Work closely with development teams, providing DevOps expertise and support for their various application stacks.
What You'll Bring 💼
- Minimum of 4 years of experience in a DevOps or SRE role.
- Strong proficiency in AWS Cloud, including services like Lambda, IoT Core, ElastiCache, CloudFront, and S3.
- Solid understanding of Linux fundamentals and command-line tools.
- Extensive experience with CI/CD tools, GitLab CI.
- Hands-on experience with Docker and Kubernetes, specifically AWS EKS.
- Proven experience deploying and managing microservices.
- Expertise in database deployment, optimization, and lifecycle management (MongoDB, PostgreSQL, and Redis).
- Experience with Identity and Access management solutions like Keycloak.
- Experience implementing backup and recovery solutions.
- Familiarity with optimizing scaling, ideally with Karpenter.
- Proficiency in scripting (Python, Bash).
- Experience with monitoring tools such as Prometheus, Grafana, AWS CloudWatch, Elastic Stack.
- Excellent problem-solving and communication skills.
Bonus Points ➕
- Basic understanding of MQTT or general IoT concepts and protocols.
- Direct experience optimizing React.js (Next.js), Node.js (Express.js, Nest.js) or Python (Flask) deployments in a containerized environment.
- Knowledge of specific AWS services relevant to application stacks.
- Contributions to open-source projects related to Kubernetes, MongoDB, or any of the mentioned frameworks.
- AWS Certifications (AWS Certified DevOps Engineer, AWS Certified Solutions Architect, AWS Certified SysOps Administrator, AWS Certified Advanced Networking).
Why this role:
•You will help build the company from the ground up—shaping our culture and having an impact from Day 1 as part of the foundational team.
Job Title: Sr Dev Ops Engineer
Location: Bengaluru- India (Hybrid work type)
Reports to: Sr Engineer manager
About Our Client :
We are a solution-based, fast-paced tech company with a team that thrives on collaboration and innovative thinking. Our Client's IoT solutions provide real-time visibility and actionable insights for logistics and supply chain management. Cloud-based, AI-enhanced metrics coupled with patented hardware optimize processes, inform strategic decision making and enable intelligent supply chains without the costly infrastructure
About the role : We're looking for a passionate DevOps Engineer to optimize our software delivery and infrastructure. You'll build and maintain CI/CD pipelines for our microservices, automate infrastructure, and ensure our systems are reliable, scalable, and secure. If you thrive on enhancing performance and fostering operational excellence, this role is for you.
What You'll Do 🛠️
- Cloud Platform Management: Administer and optimize AWS resources, ensuring efficient billing and cost management.
- Billing & Cost Optimization: Monitor and optimize cloud spending.
- Containerization & Orchestration: Deploy and manage applications and orchestrate them.
- Database Management: Deploy, manage, and optimize database instances and their lifecycles.
- Authentication Solutions: Implement and manage authentication systems.
- Backup & Recovery: Implement robust backup and disaster recovery strategies, for Kubernetes cluster and database backups.
- Monitoring & Alerting: Set up and maintain robust systems using tools for application and infrastructure health and integrate with billing dashboards.
- Automation & Scripting: Automate repetitive tasks and infrastructure provisioning.
- Security & Reliability: Implement best practices and ensure system performance and security across all deployments.
- Collaboration & Support: Work closely with development teams, providing DevOps expertise and support for their various application stacks.
What You'll Bring 💼
- Minimum of 4 years of experience in a DevOps or SRE role.
- Strong proficiency in AWS Cloud, including services like Lambda, IoT Core, ElastiCache, CloudFront, and S3.
- Solid understanding of Linux fundamentals and command-line tools.
- Extensive experience with CI/CD tools, GitLab CI.
- Hands-on experience with Docker and Kubernetes, specifically AWS EKS.
- Proven experience deploying and managing microservices.
- Expertise in database deployment, optimization, and lifecycle management (MongoDB, PostgreSQL, and Redis).
- Experience with Identity and Access management solutions like Keycloak.
- Experience implementing backup and recovery solutions.
- Familiarity with optimizing scaling, ideally with Karpenter.
- Proficiency in scripting (Python, Bash).
- Experience with monitoring tools such as Prometheus, Grafana, AWS CloudWatch, Elastic Stack.
- Excellent problem-solving and communication skills.
Bonus Points ➕
- Basic understanding of MQTT or general IoT concepts and protocols.
- Direct experience optimizing React.js (Next.js), Node.js (Express.js, Nest.js) or Python (Flask) deployments in a containerized environment.
- Knowledge of specific AWS services relevant to application stacks.
- Contributions to open-source projects related to Kubernetes, MongoDB, or any of the mentioned frameworks.
- AWS Certifications (AWS Certified DevOps Engineer, AWS Certified Solutions Architect, AWS Certified SysOps Administrator, AWS Certified Advanced Networking).
Why this role:
•You will help build the company from the ground up—shaping our culture and having an impact from Day 1 as part of the foundational team.
We are seeking an experienced and highly skilled Technical Lead with a strong background in Java, SaaS architectures, firewalls, and cybersecurity products, including SIEM and SOAR platforms. The ideal candidate will lead technical initiatives, design and implement scalable systems, and drive best practices across the engineering team. This role requires deep technical expertise, leadership abilities, and a passion for building secure and high-performing security solutions.
Key Roles & Responsibilities:
- Lead the design and development of scalable and secure software solutions using Java.
- Architect and build SaaS-based cybersecurity applications, ensuring high availability, performance, and reliability.
- Provide technical leadership, mentoring, and guidance to the development team.
- Ensure best practices in secure coding, threat modeling, and compliance with industry standards.
- Collaborate with cross-functional teams, including Product Management, Security, and DevOps to deliver high-quality security solutions.
- Design and implement security analytics, automation workflows and ITSM integrations.
- Drive continuous improvements in engineering processes, tools, and technologies.
- Troubleshoot complex technical issues and lead incident response for critical production systems.
Basic Qualifications:
- A bachelor’s or master’s degree in computer science, electronics engineering or a related field
- 3-6 years of software development experience, with expertise in Java.
- Strong background in building SaaS applications with cloud-native architectures (AWS, GCP, or Azure).
- In-depth understanding of microservices architecture, APIs, and distributed systems.
- Experience with containerization and orchestration tools like Docker and Kubernetes.
- Knowledge of DevSecOps principles, CI/CD pipelines, and infrastructure as code (Terraform, Ansible, etc.).
- Strong problem-solving skills and ability to work in an agile, fast-paced environment.
- Excellent communication and leadership skills, with a track record of mentoring engineers.
Preferred Qualifications:
- Experience with cybersecurity solutions, including SIEM (e.g., Splunk, ELK, IBM QRadar) and SOAR (e.g., Palo Alto XSOAR, Swimlane).
- Knowledge of zero-trust security models and secure API development.
- Hands-on experience with machine learning or AI-driven security analytics.


Company Description
Appiness Interactive Pvt. Ltd. is a Bangalore-based product development and UX firm that
specializes in digital services for startups to fortune-500s. We work closely with our clients to
create a comprehensive soul for their brand in the online world, engaged through multiple
platforms of digital media. Our team is young, passionate, and aggressive, not afraid to think
out of the box or tread the un-trodden path in order to deliver the best results for our clients.
We pride ourselves on Practical Creativity where the idea is only as good as the returns it
fetches for our clients.
Key Responsibilities:
- Design and implement advanced AI/ML models and algorithms to address real-world challenges.
- Analyze large and complex datasets to derive actionable insights and train predictive models.
- Build and deploy scalable, production-ready AI solutions on cloud platforms such as AWS, Azure, or GCP.
- Collaborate closely with cross-functional teams, including data engineers, product managers, and software developers, to integrate AI solutions into business workflows.
- Continuously monitor and optimize model performance, ensuring scalability, robustness, and reliability.
- Stay abreast of the latest advancements in AI, ML, and Generative AI technologies, and proactively apply them where applicable.
- Implement MLOps best practices using tools such as MLflow, Docker, and CI/CD pipelines.
- Work with Large Language Models (LLMs) like GPT and LLaMA, and develop Retrieval-Augmented Generation (RAG) pipelines when needed.
Required Skills:
- Strong programming skills in Python (preferred); experience with R or Java is also valuable.
- Proficiency with machine learning libraries and frameworks such as TensorFlow, PyTorch, and Scikit-learn.
- Hands-on experience with cloud platforms like AWS, Azure, or GCP.
- Solid foundation in data structures, algorithms, statistics, and machine learning principles.
- Familiarity with MLOps tools and practices, including MLflow, Docker, and Kubernetes.
- Proven experience in deploying and maintaining AI/ML models in production environments.
- Exposure to Large Language Models (LLMs), Generative AI, and vector databases is a strong plus.

Backend Engineer - Python
Location
Bangalore, India
Experience Required
2-3 years minimum
Job Overview
We are seeking a skilled Backend Engineer with expertise in Python to join our engineering team. The ideal candidate will have hands-on experience building and maintaining enterprise-level, scalable backend systems.
Key Requirements
Technical Skills
CS fundamentals are must (CN, DBMS, OS, System Design, OOPS) • Python Expertise: Advanced proficiency in Python with deep understanding of frameworks like Django, FastAPI, or Flask
• Database Management: Experience with PostgreSQL, MySQL, MongoDB, and database optimization
• API Development: Strong experience in designing and implementing RESTful APIs and GraphQL
• Cloud Platforms: Hands-on experience with AWS, GCP, or Azure services
• Containerization: Proficiency with Docker and Kubernetes
• Message Queues: Experience with Redis, RabbitMQ, or Apache Kafka
• Version Control: Advanced Git workflows and collaboration
Experience Requirements
• Minimum 2-3 years of backend development experience
• Proven track record of working on enterprise-level applications
• Experience building scalable systems handling high traffic loads
• Background in microservices architecture and distributed systems
• Experience with CI/CD pipelines and DevOps practices
Responsibilities
• Design, develop, and maintain robust backend services and APIs
• Optimize application performance and scalability
• Collaborate with frontend teams and product managers
• Implement security best practices and data protection measures
• Write comprehensive tests and maintain code quality
• Participate in code reviews and architectural discussions
• Monitor system performance and troubleshoot production issues
Preferred Qualifications
• Knowledge of caching strategies (Redis, Memcached)
• Understanding of software architecture patterns
• Experience with Agile/Scrum methodologies
• Open source contributions or personal projects

Location: Hybrid/ Remote
Type: Contract / Full‑Time
Experience: 5+ Years
Qualification: Bachelor’s or Master’s in Computer Science or a related technical field
Responsibilities:
- Architect & implement the RAG pipeline: embeddings ingestion, vector search (MongoDB Atlas or similar), and context-aware chat generation.
- Design and build Python‑based services (FastAPI) for generating and updating embeddings.
- Host and apply LoRA/QLoRA adapters for per‑user fine‑tuning.
- Automate data pipelines to ingest daily user logs, chunk text, and upsert embeddings into the vector store.
- Develop Node.js/Express APIs that orchestrate embedding, retrieval, and LLM inference for real‑time chat.
- Manage vector index lifecycle and similarity metrics (cosine/dot‑product).
- Deploy and optimize on AWS (Lambda, EC2, SageMaker), containerization (Docker), and monitoring for latency, costs, and error rates.
- Collaborate with frontend engineers to define API contracts and demo endpoints.
- Document architecture diagrams, API specifications, and runbooks for future team onboarding.
Required Skills
- Strong Python expertise (FastAPI, async programming).
- Proficiency with Node.js and Express for API development.
- Experience with vector databases (MongoDB Atlas Vector Search, Pinecone, Weaviate) and similarity search.
- Familiarity with OpenAI’s APIs (embeddings, chat completions).
- Hands‑on with parameters‑efficient fine‑tuning (LoRA, QLoRA, PEFT/Hugging Face).
- Knowledge of LLM hosting best practices on AWS (EC2, Lambda, SageMaker).
Containerization skills (Docker):
- Good understanding of RAG architectures, prompt design, and memory management.
- Strong Git workflow and collaborative development practices (GitHub, CI/CD).
Nice‑to‑Have:
- Experience with Llama family models or other open‑source LLMs.
- Familiarity with MongoDB Atlas free tier and cluster management.
- Background in data engineering for streaming or batch processing.
- Knowledge of monitoring & observability tools (Prometheus, Grafana, CloudWatch).
- Frontend skills in React to prototype demo UIs.

Location: Hybrid/ Remote
Openings: 2
Experience: 5–12 Years
Qualification: Bachelor’s or Master’s in Computer Science or a related technical field
Key Responsibilities
Architect & Design:
- Provide technical and architectural direction for complex frontend solutions,ensuring alignment with enterprise standards and best practices.
- Conduct design and code reviews to maintain high-quality, reusable, and scalable frontend interfaces for enterprise applications.
- Collaborate with cross-functional teams to define and enforce UI/UX design guidelines, accessibility standards, and performance benchmarks.
- Identify and address potential security vulnerabilities in frontend implementations, ensuring compliance with security and data privacy requirements.
Development & Debugging:
- Write clean, maintainable, and efficient frontend code.
- Debug and troubleshoot code to ensure robust, high-performing applications.
- Develop reusable frontend libraries that can be leveraged across multiple projects.
AI Awareness (Preferred):
- Understand AI/ML fundamentals and how they can enhance frontend applications.
- Collaborate with teams integrating AI-based features into chat applications.
Collaboration & Reporting:
- Work closely with cross-functional teams to align on architecture and deliverables.
- Regularly report progress, identify risks, and propose mitigation strategies.
Quality Assurance:
- Implement unit tests and end-to-end tests to ensure code quality.
- Participate in code reviews and enforce best practices.
Required Skills
- 5-10 years of experience architecting and developing cloud-based global applications in a public cloud environment (AWS, Azure, or GCP).
- Strong hands-on expertise in frontend technologies: JavaScript, HTML5, CSS3
- Proficiency with Modern frameworks like React, Angular, or Node.js
- Backend familiarity with Java, Spring Boot (or similar technologies).
- Experience developing real-world, at-scale products.
- General knowledge of cloud platforms (AWS, Azure, or GCP) and their structure, use, and capabilities.
- Strong problem-solving, debugging, and performance optimization skills.

Location: Hybrid/ Remote
Openings: 2
Experience: 5+ Years
Qualification: Bachelor’s or Master’s in Computer Science or related field
Job Responsibilities
Problem Solving & Optimization:
- Analyze and resolve complex technical and application issues.
- Optimize application performance, scalability, and reliability.
Design & Develop:
- Build, test, and deploy scalable full-stack applications with high performance and security.
- Develop clean, reusable, and maintainable code for both frontend and backend.
AI Integration (Preferred):
- Collaborate with the team to integrate AI/ML models into applications where applicable.
- Explore Generative AI, NLP, or machine learning solutions that enhance product capabilities.
Technical Leadership & Mentorship:
- Provide guidance, mentorship, and code reviews for junior developers.
- Foster a culture of technical excellence and knowledge sharing.
Agile & Delivery Management:
- Participate in Agile ceremonies (sprint planning, stand-ups, retrospectives).
- Define and scope backlog items, track progress, and ensure timely delivery.
Collaboration:
- Work closely with cross-functional teams (product managers, designers, QA) to deliver high-quality solutions.
- Coordinate with geographically distributed teams.
Quality Assurance & Security:
- Conduct peer reviews of designs and code to ensure best practices.
- Implement security measures and ensure compliance with industry standards.
Innovation & Continuous Improvement:
- Identify areas for improvement in the software development lifecycle.
- Stay updated with the latest tech trends, especially in AI and cloud technologies, and recommend new tools or frameworks.
Required Skills
- Strong proficiency in JavaScript, HTML5, CSS3
- Hands-on expertise with frontend frameworks like React, Angular, or Vue.js
- Backend development experience with Java, Spring Boot (Node.js is a plus)
- Knowledge of REST APIs, microservices, and scalable architectures
- Familiarity with cloud platforms (AWS, Azure, or GCP)
- Experience with Agile/Scrum methodologies and JIRA for project tracking
- Proficiency in Git and version control best practices
- Strong debugging, performance optimization, and problem-solving skills
- Ability to analyze customer requirements and translate them into technical specifications

Location: Hybrid/ Remote
Openings: 5
Experience: 0 - 2Years
Qualification: Bachelor’s or Master’s in Computer Science or a related technical field
Key Responsibilities:
Backend Development & APIs
- Build microservices that provide REST APIs to power web frontends.
- Design clean, reusable, and scalable backend code meeting enterprise security standards.
- Conceptualize and implement optimized data storage solutions for high-performance systems.
Deployment & Cloud
- Deploy microservices using a common deployment framework on AWS and GCP.
- Inspect and optimize server code for speed, security, and scalability.
Frontend Integration
- Work on modern front-end frameworks to ensure seamless integration with back-end services.
- Develop reusable libraries for both frontend and backend codebases.
AI Awareness (Preferred)
- Understand how AI/ML or Generative AI can enhance enterprise software workflows.
- Collaborate with AI specialists to integrate AI-driven features where applicable.
Quality & Collaboration
- Participate in code reviews to maintain high code quality.
- Collaborate with teams using Agile/Scrum methodologies for rapid and structured delivery.
Required Skills:
- Proficiency in JavaScript (ES6+), Webpack, Mocha, Jest
- Experience with recent frontend frameworks – React.js, Redux.js, Node.js (or similar)
- Deep understanding of HTML5, CSS3, SASS/LESS, and Content Management Systems
- Ability to design and implement RESTful APIs and understand their impact on client-side applications
- Familiarity with cloud platforms (AWS, Azure, or GCP) – deployment, storage, and scalability
- Experience working with Agile and Scrum methodologies
- Strong backend expertise in Java, J2EE, Spring Boot is a plus but not mandatory
∙
Good experience in 5+ SQL and NoSQL database development and optimization.
∙Strong hands-on experience with Amazon Redshift, MySQL, MongoDB, and Flyway.
∙In-depth understanding of data warehousing principles and performance tuning techniques.
∙Strong hands-on experience in building complex aggregation pipelines in NoSQL databases such as MongoDB.
∙Proficient in Python or Scala for data processing and automation.
∙3+ years of experience working with AWS-managed database services.
∙3+ years of experience with Power BI or similar BI/reporting platforms.

Job Description:
Title : Python AWS Developer with API
Tech Stack : AWS API gateway, Lambda functionality, Oracle RDS, SQL & database management, (OOPS) principles, Java script, Object relational Mapper, Git, Docker, Java dependency management, CI/CD, AWS cloud & S3, Secret Manager, Python, API frameworks, well-versed with Front and back end programming (python).
Responsibilities:
· Worked on building high-performance APIs using AWS services and Python. Python coding, debugging programs and integrating app with third party web services.
· Troubleshoot and debug non-prod defects, back-end development, API, main focus on coding and monitoring applications.
· Core application logic design.
· Supports dependency teams in UAT testing and perform functional application testing which includes postman testing

🔍 Job Description:
We are looking for an experienced and highly skilled Technical Lead to guide the development and enhancement of a large-scale Data Observability solution built on AWS. This platform is pivotal in delivering monitoring, reporting, and actionable insights across the client's data landscape.
The Technical Lead will drive end-to-end feature delivery, mentor junior engineers, and uphold engineering best practices. The position reports to the Programme Technical Lead / Architect and involves close collaboration to align on platform vision, technical priorities, and success KPIs.
🎯 Key Responsibilities:
- Lead the design, development, and delivery of features for the data observability solution.
- Mentor and guide junior engineers, promoting technical growth and engineering excellence.
- Collaborate with the architect to align on platform roadmap, vision, and success metrics.
- Ensure high quality, scalability, and performance in data engineering solutions.
- Contribute to code reviews, architecture discussions, and operational readiness.
🔧 Primary Must-Have Skills (Non-Negotiable):
- 5+ years in Data Engineering or Software Engineering roles.
- 3+ years in a technical team or squad leadership capacity.
- Deep expertise in AWS Data Services: Glue, EMR, Kinesis, Lambda, Athena, S3.
- Advanced programming experience with PySpark, Python, and SQL.
- Proven experience in building scalable, production-grade data pipelines on cloud platforms.

We are seeking a Senior Laravel Developer with a minimum of 8+ years of experience and a proven track record in developing and maintaining PHP/Laravel-based websites and applications. The ideal candidate should excel in creating high-performance web applications using PHP and MySQL, with expertise in debugging, performance optimisation, and scalability.
Responsibilities:
- Write clean, maintainable code adhering to company coding standards.
- Develop and enhance existing PHP/Laravel/Code-igniter projects.
- Troubleshoot, test, and maintain core product software and databases for optimisation and functionality.
- Contribute to all phases of the development lifecycle.
- Follow industry best practices for secure and scalable development.
Key Skills
Technical Proficiency
- Expertise in PHP, MySQL, and related web technologies.
- Strong experience in debugging, performance optimisation, and scalability.
- Familiarity with resource-intensive application architectures.
- Hands-on experience with AWS services (e.g., EC2, S3, RDS, Lambda) for scalable application deployment.
- Proficiency in setting up and managing CI/CD pipelines using tools like Jenkins, GitLab CI/CD, or AWS CodePipeline.
- Strong knowledge of Git version control and collaborative workflows.
- Experience with microservices architecture and containerization (Docker/Kubernetes) is a plus.
- Familiarity with front-end frameworks (Vue.js, React, or Angular) and modern JavaScript practices is a bonus.
Frameworks
- Hands-on experience with Laravel and CodeIgniter frameworks.
Proactive Problem-Solving
- Ability to identify potential bottlenecks and provide innovative solutions.
Soft Skills
- Adaptability to balance support and development tasks effectively.
- Ability to work independently and collaboratively in teams.
- Exceptional problem-solving skills and attention to detail.
Leadership and Mentorship
- Mentor junior and mid-level developers, providing guidance on coding standards, best practices, and technical challenges.
- Conduct code reviews to ensure high-quality, maintainable codebases.
- Collaborate with stakeholders to define technical requirements and project roadmaps.
Documentation and Best Practices
- Maintain comprehensive technical documentation for projects, processes, and workflows.
- Establish and enforce coding standards, security protocols, and development guidelines across the team.
Preferred Background
- Proven experience managing performance-critical applications.
- Exposure to legacy systems alongside modern development practices.
- Prior experience balancing support and development roles.
- Demonstrated expertise in resource-intensive application development and optimisation.
- Evidence of significant past performance improvements in web applications.
- Proficiency in frameworks such as Laravel and CodeIgniter, and relevant tech stacks.
Qualifications
- Minimum 8+ years of experience managing performance-critical applications.
- BE/B.Tech in Computer Science or equivalent degree required.
- Sound knowledge of OOP principles and best practices.
- Familiarity with Git version control and Agile/Scrum methodologies.
This role offers the opportunity to work on innovative projects while making meaningful contributions to high-scale applications. If you are passionate about creating exceptional web applications and thrive in a fast-paced environment, we encourage you to apply!
About Eazeebox
Eazeebox is India’s first B2B Quick Commerce platform for home electrical goods. We empower electrical retailers with access to 100+ brands, flexible credit options, and 4-hour delivery—making supply chains faster, smarter, and more efficient. Our tech-driven approach enables sub-3 hour inventory-aware fulfilment across micro-markets, with a goal of scaling to 50+ orders/day per store.
About the Role
We’re looking for a DevOps Engineer to help scale and stabilize the cloud-native backbone that powers Eazeebox. You’ll play a critical role in ensuring our microservices architecture remains reliable, responsive, and performant—especially during peak retailer ordering windows. This is a high-ownership role for an "all-rounder" who is passionate about
designing scalable architectures, writing robust code, and ensuring seamless deployments and operations.
What You'll Be Doing
As a critical member of our small, dedicated team, you will take on a versatile role encompassing development, infrastructure, and operations.
Cloud & DevOps Ownership
- Architect and implement containerized services on AWS (S3, EC2, ECS, ECR, CodeBuild, Lambda, Fargate, RDS, CloudWatch) under secure IAM policies.
- Take ownership of CI/CD pipelines, optimizing and managing GitHub Actions workflows.
- Configure and manage microservice versioning and CI/CD deployments.
- Implement secrets rotation and IP-based request rate limiting for enhanced security.
- Configure auto-scaling instances and Kubernetes for high-workload microservices to ensure performance and cost efficiency.
- Hands-on experience with Docker and Kubernetes/EKS fundamentals.
Backend & API Design
- Design, build, and maintain scalable REST/OpenAPI services in Django (DRF), WebSocket implementations, and asynchronous microservices in FastAPI.
- Model relational data in PostgreSQL 17 and optimize with Redis for caching and pub/sub.
- Orchestrate background tasks using Celery or RQ with Redis Streams or Amazon SQS.
- Collaborate closely with the frontend team (React/React Native) to define and build robust APIs.
Testing & Observability
- Ensure code quality via comprehensive testing using Pytest, React Testing Library, and Playwright.
- Instrument applications with CloudWatch metrics, contributing to our observability strategy.
- Maintain a Git-centric development workflow, including branching strategies and pull-request discipline.
Qualifications & Skills
Must-Have
- Experience: 2-4 years of hands-on experience delivering production-level full-stack applications with a strong emphasis on backend and DevOps.
- Backend Expertise: Proficiency in Python, with strong command of Django or FastAPI, including async Python patterns and REST best practices.
- Database Skills: Strong SQL skills with PostgreSQL; practical experience with Redis for caching and messaging.
- Cloud & DevOps Mastery: Hands-on experience with Docker and Kubernetes/EKS fundamentals.
- AWS Proficiency: Experience deploying and managing services on AWS (EC2, S3, RDS, Lambda, ECS Fargate, ECR, SQS).
- CI/CD: Deep experience with GitHub Actions or similar platforms, including semantic-release, Blue-Green Deployments, and artifact signing.
- Automation: Fluency in Python/Bash or Go for automation scripts; comfort with YAML.
- Ownership Mindset: Entrepreneurial spirit, strong sense of ownership, and ability to deliver at scale.
- Communication: Excellent written and verbal communication skills; comfortable in async and distributed team environments.
Nice-to-Have
- Frontend Familiarity: Exposure to React with Redux Toolkit and React Native.
- Event Streaming: Experience with Kafka or Amazon EventBridge.
- Serverless: Knowledge of AWS Lambda, Step Functions, CloudFront Functions, or Cloudflare Workers.
- Observability: Familiarity with Datadog, Posthog, Prometheus/Grafana/Loki.
- Emerging Tech: Interest in GraphQL (Apollo Federation) or generative AI frameworks (Amazon Bedrock, LangChain) and AI/ML.
Key Responsibilities
- Architectural Leadership: Design and lead the technical strategy for migrating our platform from a monolithic to a microservices architecture.
- System Design: Translate product requirements into scalable, secure, and reliable system designs.
- Backend Development: Build and maintain core backend services using Python (Django/FastAPI).
- CI/CD & Deployment: Own and manage CI/CD pipelines for multiple services using GitHub Actions, AWS CodeBuild, and automated deployments.
- Infrastructure & Operations: Deploy production-grade microservices using Docker, Kubernetes, and AWS EKS.
- FinOps & Performance: Drive cloud cost optimization and implement auto-scaling for performance and cost-efficiency.
- Security & Observability: Implement security, monitoring, and compliance using tools like Prometheus, Grafana, Datadog, Posthog, and Loki to ensure 99.99% uptime.
- Collaboration: Work with product and development teams to align technical strategy with business growth plans.
We are looking for someone with a hacker mindset who is ready to pick up new problems and build full stack AI solutions for some of the biggest brands in the country and the world.

Mode of Hire: Permanent
Required Skills Set (Mandatory): Linux, Shell Scripting, Python, AWS, Security best practices, Git
Desired Skills (Good if you have): Ansible, Terraform
Job Responsibilities
- Design, develop, and maintain deployment pipelines and automation tooling to improve platform efficiency, scalability, and reliability.
- Manage infrastructure and services in production AWS environments.
- Drive platform improvements with a focus on security, scalability, and operational excellence.
- Collaborate with engineering teams to enhance development tooling, streamline access workflows, and improve platform usability through feedback.
- Mentor junior engineers and help foster a culture of high-quality engineering and knowledge sharing.
Job Requirements
- Strong foundational understanding of Linux systems.
- Cloud experience (e.g., AWS) with strong problem-solving in cloud-native environments.
- Proven track record of delivering robust, well-documented, and secure automation solutions.
- Comfortable owning end-to-end delivery of infrastructure components and tooling.
Preferred Qualifications
- Advanced system and cloud optimization skills.
- Prior experience in platform teams or DevOps roles at product-focused startups.
- Demonstrated contributions to internal tooling, open-source, or automation projects.
Job Description
We are looking for a hands-on Tech Lead – Java with strong software engineering fundamentals, a deep understanding of Java-based backend systems, and proven experience leading agile teams. This role involves a balance of individual contribution and technical leadership — mentoring developers, designing scalable architectures, and driving the success of product delivery in fast-paced environments.
Key Responsibilities
- Lead the end-to-end design, development, and deployment of Java-based applications and RESTful APIs.
- Collaborate with product managers and architects to define technical solutions and translate business requirements into scalable software.
- Guide and mentor team members in best coding practices, design patterns, and architectural decisions.
- Drive code reviews, technical discussions, and ensure high code quality and performance standards.
- Troubleshoot critical production issues and implement long-term fixes and improvements.
- Advocate for continuous improvement in tools, processes, and systems across the engineering team.
- Stay up to date with modern technologies and recommend their adoption where appropriate.
Required Skills
- 5+ years of experience in Java backend development with expertise in Spring/Spring Boot and RESTful services.
- Solid grasp of Object-Oriented Programming (OOP), system design, and design patterns.
- Proven experience leading a team of engineers or taking ownership of modules/projects.
- Experience with AWS Cloud services (EC2, Lambda, S3, etc.) is a strong advantage.
- Familiarity with Agile/Scrum methodologies and working in cross-functional teams.
- Excellent problem-solving, debugging, and analytical skills.
- Strong communication and leadership skills.
About HummingWave
HummingWave is a leading IT product development company specializing in building full-scale application systems with robust cloud backends, sleek mobile/web frontends, and seamless enterprise integrations. With 50+ digital products delivered across domains for clients in the US, Europe, and Asia-Pacific, we are a team of highly skilled engineers committed to technical excellence and innovation.
Thanks
Create and manage Jenkins Pipelines using Linux groovy
scripting and python
Analyze and fix issues in Jenkins, GitHub, Nexus,
SonarQube and AWS cloud
Perform Jenkins, GitHub, SonarQube and Nexus
administration.
Create resources in AWS environment using
infrastructure-as-code. Analyze and fix issues in AWS
Cloud.
Good-to-Have
AWS Cloud certification
Terraform Certification
- Kubernetes/Docker experience

Job Summary:
We are looking for a skilled and motivated Python AWS Engineer to join our team. The ideal candidate will have strong experience in backend development using Python, cloud infrastructure on AWS, and building serverless or microservices-based architectures. You will work closely with cross-functional teams to design, develop, deploy, and maintain scalable and secure applications in the cloud.
Key Responsibilities:
- Develop and maintain backend applications using Python and frameworks like Django or Flask
- Design and implement serverless solutions using AWS Lambda, API Gateway, and other AWS services
- Develop data processing pipelines using services such as AWS Glue, Step Functions, S3, DynamoDB, and RDS
- Write clean, efficient, and testable code following best practices
- Implement CI/CD pipelines using tools like CodePipeline, GitHub Actions, or Jenkins
- Monitor and optimize system performance and troubleshoot production issues
- Collaborate with DevOps and front-end teams to integrate APIs and cloud-native services
- Maintain and improve application security and compliance with industry standards
Required Skills:
- Strong programming skills in Python
- Solid understanding of AWS cloud services (Lambda, S3, EC2, DynamoDB, RDS, IAM, API Gateway, CloudWatch, etc.)
- Experience with infrastructure as code (e.g., CloudFormation, Terraform, or AWS CDK)
- Good understanding of RESTful API design and microservices architecture
- Hands-on experience with CI/CD, Git, and version control systems
- Familiarity with containerization (Docker, ECS, or EKS) is a plus
- Strong problem-solving and communication skills
Preferred Qualifications:
- Experience with PySpark, Pandas, or data engineering tools
- Working knowledge of Django, Flask, or other Python frameworks
- AWS Certification (e.g., AWS Certified Developer – Associate) is a plus
Educational Qualification:
- Bachelor's or Master’s degree in Computer Science, Engineering, or related field


About Vahan
At Vahan.ai, we are building India’s first AI-powered recruitment marketplace for the 300 million-strong blue-collar workforce — opening doors to economic opportunities and brighter futures.
Already India’s largest recruitment platform, Vahan.ai is backed by marquee investors like Khosla Ventures, Bharti Airtel, Vijay Shekhar Sharma (CEO, Paytm), and leading executives from Google and Facebook.
Our customers include Swiggy, Zomato, Rapido, Zepto, and many more. We leverage cutting-edge technology and AI to recruit for the workforces of some of the most recognized companies in the country.
Our vision is ambitious:
To become the go-to platform for blue-collar professionals worldwide — empowering them with not just earning opportunities but also the tools, benefits, and support they need to thrive. We aim to impact over a billion lives globally, creating a future where everyone has access to economic prosperity.
If our vision excites you, Vahan.ai might just be your next adventure. We’re on the hunt for driven individuals who love tackling big challenges. Dive into the details below and see where you can make your mark.
Role Overview:
We're seeking an experienced Senior Engineering Manager to lead multiple engineering pods/projects (2–3) and drive technical excellence across our product portfolio. This role combines hands-on technical leadership with people management, requiring someone who can scale teams, deliver complex projects, and maintain high engineering standards in a fast-paced startup environment.
Key Responsibilities
Team Leadership & Management
- Lead and manage multiple engineering pods (15–20 engineers in total)
- Hire, onboard, and develop engineering talent across different experience levels
- Conduct performance reviews, provide career guidance, and manage team growth
- Foster a culture of technical excellence, collaboration, and continuous learning
- Build and implement engineering processes that scale with company growth
Technical Leadership
- Guide technical architecture decisions across full-stack applications
- Provide technical mentorship and code review guidance to engineers
- Lead both frontend (React.js, React Native) and backend (Node.js) development initiatives
- Oversee web and mobile application development strategies
- Drive technical problem-solving and debugging for complex production issues
Infrastructure & Operations
- Manage AWS infrastructure, ensuring scalability, reliability, and cost-effectiveness
- Build and maintain production support processes and incident response procedures
- Implement monitoring, alerting, and observability practices
- Lead post-mortem processes and drive continuous improvement initiatives
Cross-Functional Collaboration
- Partner closely with Product, QA, and Business teams to deliver on company objectives
- Translate business requirements into technical roadmaps and delivery plans
- Manage stakeholder expectations and communicate technical decisions effectively
- Balance feature delivery with technical debt management and system scalability
Strategic Planning
- Develop engineering roadmaps aligned with business goals and product strategy
- Make data-driven decisions about technology choices and team structure
- Identify and mitigate technical and operational risks
- Drive engineering metrics and KPIs to measure team performance and product quality
Required Qualifications
Experience
- 7–10 years of software engineering experience
- 3+ years of engineering management experience, preferably in startup environments
- Proven track record of leading teams of 10+ engineers across multiple projects
- Experience hiring and scaling engineering teams in high-growth companies
Technical Skills
- Strong expertise in Node.js, React.js, and React Native (at least one frontend and backend)
- Experience with full-stack web and mobile application development
- Solid understanding of AWS services and cloud infrastructure management
- Knowledge of software architecture, system design, and scalability principles
- Experience with CI/CD pipelines, monitoring tools, and DevOps practices
Leadership & Management
- Demonstrated ability to mentor and develop engineering talent
- Experience conducting performance reviews and managing career development
- Strong stakeholder management and cross-functional collaboration skills
- Proven ability to balance technical decisions with business objectives
- Experience managing production systems and incident response processes
Soft Skills
- Excellent communication and presentation skills
- Strong problem-solving and analytical thinking abilities
- Ability to work effectively in fast-paced, ambiguous environments
- Experience making difficult prioritization decisions under resource constraints
- Cultural fit for startup environment with hands-on, results-oriented approach
Preferred Qualifications
- Experience in Series B+ stage startups or high-growth technology companies
- Background in tech-first / AI-first startups
- AI/ML product knowledge and experience (good to have)
- Experience with microservices architecture and distributed systems
- Contributions to open-source projects or technical community involvement
What We Offer
- Unlimited PTO – Trust and flexibility to manage your time in the way that works best for you
- Comprehensive Medical Insurance – Plans designed to support you and your loved ones
- Monthly Wellness Leaves – Regular time off to recharge and focus on what matters most
- Competitive Pay – Your contributions are recognized and rewarded with a compensation package that reflects your impact
Join us, and be part of something bigger — where your work drives real, positive change in the world.
Job Description :
We are seeking a highly experienced Sr Data Modeler / Solution Architect to join the Data Architecture team at Corporate Office in Bangalore. The ideal candidate will have 4 to 8 years of experience in data modeling and architecture with deep expertise in AWS cloud stack, data warehousing, and enterprise data modeling tools. This individual will be responsible for designing and creating enterprise-grade data models and driving the implementation of Layered Scalable Architecture or Medallion Architecture to support robust, scalable, and high-quality data marts across multiple business units.
This role will involve managing complex datasets from systems like PoS, ERP, CRM, and external sources, while optimizing performance and cost. You will also provide strategic leadership on data modeling standards, governance, and best practices, ensuring the foundation for analytics and reporting is solid and future-ready.
Key Responsibilities:
· Design and deliver conceptual, logical, and physical data models using tools like ERWin.
· Implement Layered Scalable Architecture / Medallion Architecture for building scalable, standardized data marts.
· Optimize performance and cost of AWS-based data infrastructure (Redshift, S3, Glue, Lambda, etc.).
· Collaborate with cross-functional teams (IT, business, analysts) to gather data requirements and ensure model alignment with KPIs and business logic.
· Develop and optimize SQL code, materialized views, stored procedures in AWS Redshift.
· Ensure data governance, lineage, and quality mechanisms are established across systems.
· Lead and mentor technical teams in an Agile project delivery model.
· Manage data layer creation and documentation: data dictionary, ER diagrams, purpose mapping.
· Identify data gaps and availability issues with respect to source systems.
Required Skills & Qualifications:
· Bachelor’s or Master’s degree in Computer Science, IT, or related field (B.E./B.Tech/M.E./M.Tech/MCA).
· Minimum 4 years of experience in data modeling and architecture.
· Proficiency with data modeling tools such as ERWin, with strong knowledge of forward and reverse engineering.
· Deep expertise in SQL (including advanced SQL, stored procedures, performance tuning).
· Strong experience in data warehousing, RDBMS, and ETL tools like AWS Glue, IBM DataStage, or SAP Data Services.
· Hands-on experience with AWS services: Redshift, S3, Glue, RDS, Lambda, Bedrock, and Q.
· Good understanding of reporting tools such as Tableau, Power BI, or AWS QuickSight.
· Exposure to DevOps/CI-CD pipelines, AI/ML, Gen AI, NLP, and polyglot programming is a plus.
· Familiarity with data governance tools (e.g., ORION/EIIG).
· Domain knowledge in Retail, Manufacturing, HR, or Finance preferred.
· Excellent written and verbal communication skills.
Certifications (Preferred):
· AWS Certification (e.g., AWS Certified Solutions Architect or Data Analytics – Specialty)
· Data Governance or Data Modeling Certifications (e.g., CDMP, Databricks, or TOGAF)
Mandatory Skills
aws, Technical Architecture, Aiml, SQL, Data Warehousing, Data Modelling

Supply Wisdom: Full Stack Developer
Location: Hybrid Position based in Bangalore
Reporting to: Tech Lead Manager
Supply Wisdom is a global leader in transformative risk intelligence, offering real-time insights to drive business growth, reduce costs, enhance security and compliance, and identify revenue opportunities. Our AI-based SaaS products cover various risk domains, including financial, cyber, operational, ESG, and compliance. With a diverse workforce that is 57% female, our clients include Fortune 100 and Global 2000 firms in sectors like financial services, insurance, healthcare, and technology.
Objective: We are seeking a skilled Full Stack Developer to design and build scalable software solutions. You will be part of a cross-functional team responsible for the full software development life cycle, from conception to deployment.
As a Full Stack Developer, you should be proficient in both front-end and back-end technologies, development frameworks, and third-party libraries. We’re looking for a team player with strong problem-solving abilities, attention to visual design, and a focus on utility. Familiarity with Agile methodologies, including Scrum and Kanban, is essential.
Responsibilities
- Collaborate with the development team and product manager to ideate software solutions.
- Write effective and secure REST APIs.
- Integrate third-party libraries for product enhancement.
- Design and implement client-side and server-side architecture.
- Work with data scientists and analysts to enhance software using RPA and AI/ML techniques.
- Develop and manage well-functioning databases and applications.
- Ensure software responsiveness and efficiency through testing.
- Troubleshoot, debug, and upgrade software as needed.
- Implement security and data protection settings.
- Create features and applications with mobile-responsive design.
- Write clear, maintainable technical documentation.
- Build front-end applications with appealing, responsive visual design.
Requirements
- Degree in Computer Science (or related field) with 4+ years of hands-on experience in Python development, with strong expertise in the Django framework and Django REST Framework (DRF).
- Proven experience in designing and building RESTful APIs, with a solid understanding of API versioning, authentication (JWT/OAuth2), and best practices.
- Experience with relational databases such as PostgreSQL or MySQL; familiarity with query optimization and database migrations.
- Basic front-end development skills using HTML, CSS, and JavaScript; experience with any JavaScript framework (like React or Next Js) is a plus.
- Good understanding of Object-Oriented Programming (OOP) and design patterns in Python.
- Familiarity with Git and collaborative development workflows (e.g., GitHub, GitLab).
- Knowledge of Docker, CI/CD pipelines.
- Hands-on experience with AWS services, Nginx web server, RabbitMQ (or similar message brokers), event handling, and synchronization.
- Familiarity with Postgres, SSO implementation (desirable), and integration of third-party libraries.
- Experience with unit testing, debugging, and code reviews.
- Experience using tools like Jira and Confluence.
- Ability to work in Agile/Scrum teams with good communication and problem-solving skills.
Our Commitment to You:
We offer a competitive salary and generous benefits. In addition, we offer a vibrant work environment, a global team filled with passionate and fun-loving people coming from diverse cultures and backgrounds.
If you are looking to make an impact in delivering market-leading risk management solutions, empowering our clients, and making the world a better place, then Supply Wisdom is the place for you.
You can learn more at supplywisdom.com and on LinkedIn.
Designing, building, and automating ETL processes using AWS services like Apache Sqoop, AWS S3, AWS CLI, Amazon
EMR, Amazon MSK, Amazon Sagemaker.
∙Developing and maintaining data pipelines to move and transform data from diverse sources into data warehouses or
data lakes.
∙Ensuring data quality and integrity through validation, cleansing, and monitoring ETL processes.
∙Optimizing ETL workflows for performance, scalability, and cost efficiency within the AWS environment.
∙Troubleshooting and resolving issues related to data processing and ETL workflows.
∙Implementing and maintaining security measures and compliance standards for data pipelines and infrastructure.
∙Documenting ETL processes, data mappings, and system architecture.
Job Title: Sr. DevOps Engineer
Experience Required: 2 to 4 years in DevOps or related fields
Employment Type: Full-time
About the Role:
We are seeking a highly skilled and experienced Lead DevOps Engineer. This role will focus on driving the design, implementation, and optimization of our CI/CD pipelines, cloud infrastructure, and operational processes. As a Lead DevOps Engineer, you will play a pivotal role in enhancing the scalability, reliability, and security of our systems while mentoring a team of DevOps engineers to achieve operational excellence.
Key Responsibilities:
Infrastructure Management: Architect, deploy, and maintain scalable, secure, and resilient cloud infrastructure (e.g., AWS, Azure, or GCP).
CI/CD Pipelines: Design and optimize CI/CD pipelines, to improve development velocity and deployment quality.
Automation: Automate repetitive tasks and workflows, such as provisioning cloud resources, configuring servers, managing deployments, and implementing infrastructure as code (IaC) using tools like Terraform, CloudFormation, or Ansible.
Monitoring & Logging: Implement robust monitoring, alerting, and logging systems for enterprise and cloud-native environments using tools like Prometheus, Grafana, ELK Stack, NewRelic or Datadog.
Security: Ensure the infrastructure adheres to security best practices, including vulnerability assessments and incident response processes.
Collaboration: Work closely with development, QA, and IT teams to align DevOps strategies with project goals.
Mentorship: Lead, mentor, and train a team of DevOps engineers to foster growth and technical expertise.
Incident Management: Oversee production system reliability, including root cause analysis and performance tuning.
Required Skills & Qualifications:
Technical Expertise:
Strong proficiency in cloud platforms like AWS, Azure, or GCP.
Advanced knowledge of containerization technologies (e.g., Docker, Kubernetes).
Expertise in IaC tools such as Terraform, CloudFormation, or Pulumi.
Hands-on experience with CI/CD tools, particularly Bitbucket Pipelines, Jenkins, GitLab CI/CD, Github Actions or CircleCI.
Proficiency in scripting languages (e.g., Python, Bash, PowerShell).
Soft Skills:
Excellent communication and leadership skills.
Strong analytical and problem-solving abilities.
Proven ability to manage and lead a team effectively.
Experience:
4 years + of experience in DevOps or Site Reliability Engineering (SRE).
4+ years + in a leadership or team lead role, with proven experience managing distributed teams, mentoring team members, and driving cross-functional collaboration.
Strong understanding of microservices, APIs, and serverless architectures.
Nice to Have:
Certifications like AWS Certified Solutions Architect, Kubernetes Administrator, or similar.
Experience with GitOps tools such as ArgoCD or Flux.
Knowledge of compliance standards (e.g., GDPR, SOC 2, ISO 27001).
Perks & Benefits:
Competitive salary and performance bonuses.
Comprehensive health insurance for you and your family.
Professional development opportunities and certifications, including sponsored certifications and access to training programs to help you grow your skills and expertise.
Flexible working hours and remote work options.
Collaborative and inclusive work culture.
Join us to build and scale world-class systems that empower innovation and deliver exceptional user experiences.
You can directly contact us: Nine three one six one two zero one three two

Job Title: Site Reliability Engineer (SRE)
Experience: 4+ Years
Work Location: Bangalore / Chennai / Pune / Gurgaon
Work Mode: Hybrid or Onsite (based on project need)
Domain Preference: Candidates with past experience working in shoe/footwear retail brands (e.g., Nike, Adidas, Puma) are highly preferred.
🛠️ Key Responsibilities
- Design, implement, and manage scalable, reliable, and secure infrastructure on AWS.
- Develop and maintain Python-based automation scripts for deployment, monitoring, and alerting.
- Monitor system performance, uptime, and overall health using tools like Prometheus, Grafana, or Datadog.
- Handle incident response, root cause analysis, and ensure proactive remediation of production issues.
- Define and implement Service Level Objectives (SLOs) and Error Budgets in alignment with business requirements.
- Build tools to improve system reliability, automate manual tasks, and enforce infrastructure consistency.
- Collaborate with development and DevOps teams to ensure robust CI/CD pipelines and safe deployments.
- Conduct chaos testing and participate in on-call rotations to maintain 24/7 application availability.
✅ Must-Have Skills
- 4+ years of experience in Site Reliability Engineering or DevOps with a focus on reliability, monitoring, and automation.
- Strong programming skills in Python (mandatory).
- Hands-on experience with AWS cloud services (EC2, S3, Lambda, ECS/EKS, CloudWatch, etc.).
- Expertise in monitoring and alerting tools like Prometheus, Grafana, Datadog, CloudWatch, etc.
- Strong background in Linux-based systems and shell scripting.
- Experience implementing infrastructure as code using tools like Terraform or CloudFormation.
- Deep understanding of incident management, SLOs/SLIs, and postmortem practices.
- Prior working experience in footwear/retail brands such as Nike or similar is highly preferred.
Minimum 5 years of experience in a customer-facing role such as pre-sales, solutions engineering or technical architecture.
- Exceptional communication and presentation skills.
- Proven ability in technical integrations and conducting POCs.
- Proficiency in coding with high-level programming languages (Java, Go, Python).
- Solid understanding of Monitoring, Observability, Log Management, SIEM.
- Background in Engineering/DevOps will be considered an advantage.
- Previous experience in Technical Sales of Log Analytics, Monitoring, APM, RUM, SIEM is desirable.
Technical Expertise :
- In-depth knowledge of Kubernetes, AWS, Azure, GCP, Docker, Prometheus, OpenTelemetry.
- Candidates should have hands-on experience and the ability to integrate these technologies into customer environments, providing tailored solutions that meet diverse operational requirements.
Responsibilities
- Provide technology contributions in
- Working in an agile development environment
- Translating business requirements into low-level application design
- Application code development through a collaborative approach
- Doing Full-scale unit testing
- Applying test-driven and behaviour-driven development (TDD/BDD) QA concepts
- Applying continuous integration and continuous deployment (CI/CD) concepts
Soft Skills
- Should be able to contribute as an individual contributor
- Should be able to execute his/her responsibility independently
- Excellent problem-solving skills and attention to detail.
- Focus on self-planning activities
- Firm with communication skills
Mandatory Skills
- Java, Spring Boot, Python and relational / non-relational databases
- Container orchestration - Kubernetes, Docker
- Development experience in Linux environment
- Modern SDLC tooling (Maven, Git)
- Micro services design-oriented application development and deploying the same using Container orchestration in the cloud environment
- Understanding CI/CD pipeline & related system development environment
Nice-to-have Skills
- Front-end technologies (JavaScript, HTML5, CSS, Angular)
Job Title : Senior Software Engineer – Backend
Experience Required : 6 to 12 Years
Location : Bengaluru (Hybrid – 3 Days Work From Office)
Number of Openings : 2
Work Hours : 11:00 AM – 8:00 PM IST
Notice Period : 30 Days Preferred
Work Location : SmartWorks The Cube, Karle Town SEZ, Building No. 5, Nagavara, Bangalore – 560045
Note : Face-to-face interview in Bangalore is mandatory during the second round.
Role Overview :
We are looking for an experienced Senior Backend Developer to join our growing team. This is a hands-on role focused on building cloud-based, scalable applications in the mortgage finance domain.
Key Responsibilities :
- Design, develop, and maintain backend components for cloud-based web applications.
- Contribute to architectural decisions involving microservices and distributed systems.
- Work extensively with Node.js and RESTful APIs.
- Implement scalable solutions using AWS services (e.g., Lambda, SQS, SNS, RDS).
- Utilize both relational and NoSQL databases effectively.
- Collaborate with cross-functional teams to deliver robust and maintainable code.
- Participate in agile development practices and deliver rapid iterations based on feedback.
- Take ownership of system performance, scalability, and reliability.
Core Requirements :
- 5+ Years of total experience in backend development.
- Minimum 3 Years of experience in building scalable microservices or delivering large-scale products.
- Strong expertise in Node.js and REST APIs.
- Solid experience with RDBMS, SQL, and data modeling.
- Good understanding of distributed systems, scalability, and availability.
- Familiarity with AWS infrastructure and services.
- Development experience in Python and/or Java is a plus.
Preferred Skills :
- Experience with frontend frameworks like React.js or AngularJS.
- Working knowledge of Docker and containerized applications.
Interview Process :
- Round 1 : Online technical assessment (1 hour)
- Round 2 : Virtual technical interview
- Round 3 : In-person interview at the Bangalore office (2 hours – mandatory)
- Develop, and maintain Java applications using Core Java, Spring framework, JDBC, and threading concepts.
- Strong understanding of the Spring framework and its various modules.
- Experience with JDBC for database connectivity and manipulation
- Utilize database management systems to store and retrieve data efficiently.
- Proficiency in Core Java8 and thorough understanding of threading concepts and concurrent programming.
- Experience in in working with relational and nosql databases.
- Basic understanding of cloud platforms such as Azure and GCP and gain experience on DevOps practices is added advantage.
- Knowledge of containerization technologies (e.g., Docker, Kubernetes)
- Perform debugging and troubleshooting of applications using log analysis techniques.
- Understand multi-service flow and integration between components.
- Handle large-scale data processing tasks efficiently and effectively.
- Hands on experience using Spark is an added advantage.
- Good problem-solving and analytical abilities.
- Collaborate with cross-functional teams to identify and solve complex technical problems.
- Knowledge of Agile methodologies such as Scrum or Kanban
- Stay updated with the latest technologies and industry trends to improve development processes continuously and methodologies.

Role: Data Engineer (14+ years of experience)
Location: Whitefield, Bangalore
Mode of Work: Hybrid (3 days from office)
Notice period: Immediate/ Serving with 30days left
Location: Candidate should be based out of Bangalore as one round has to be taken F2F
Job Summary:
Role and Responsibilities
● Design and implement scalable data pipelines for ingesting, transforming, and loading data from various tools and sources.
● Design data models to support data analysis and reporting.
● Automate data engineering tasks using scripting languages and tools.
● Collaborate with engineers, process managers, data scientists to understand their needs and design solutions.
● Act as a bridge between the engineering and the business team in all areas related to Data.
● Automate monitoring and alerting mechanism on data pipelines, products and Dashboards and troubleshoot any issues. On call requirements.
● SQL creation and optimization - including modularization and optimization which might need views, table creation in the sources etc.
● Defining best practices for data validation and automating as much as possible; aligning with the enterprise standards
● QA environment data management - e.g Test Data Management etc
Qualifications
● 14+ years of experience as a Data engineer or related role.
● Experience with Agile engineering practices.
● Strong experience in writing queries for RDBMS, cloud-based data warehousing solutions like Snowflake and Redshift.
● Experience with SQL and NoSQL databases.
● Ability to work independently or as part of a team.
● Experience with cloud platforms, preferably AWS.
● Strong experience with data warehousing and data lake technologies (Snowflake)
● Expertise in data modelling
● Experience with ETL/LT tools and methodologies .
● 5+ years of experience in application development including Python, SQL, Scala, or Java
● Experience working on real-time Data Streaming and Data Streaming platform.
NOTE: IT IS MANDATORY TO GIVE ONE TECHNICHAL ROUND FACE TO FACE.
Role: Sr. Java Developer
Experience: 6+ Years
Location: Bangalore (Whitefield)
Work Mode: Hybrid (2-3 days WFO)
Shift Timing: Regular Morning Shift
About the Role:
We are looking for a seasoned Java Developer with 6+ years of experience to join our growing engineering team. The ideal candidate should have strong expertise in Java, Spring Boot, Microservices, and cloud-based deployment using AWS or DevOps tools. This is a hybrid role based out of our Whitefield, Bangalore location.
Key Responsibilities:
- Participate in agile development processes and scrum ceremonies.
- Translate business requirements into scalable and maintainable technical solutions.
- Design and develop applications using Java, Spring Boot, and Microservices architecture.
- Ensure robust and reliable code through full-scale unit testing and TDD/BDD practices.
- Contribute to CI/CD pipeline setup and cloud deployments.
- Work independently and as an individual contributor on complex features.
- Troubleshoot production issues and optimize application performance.
Mandatory Skills:
- Strong Core Java and Spring Boot expertise.
- Proficiency in AWS or DevOps (Docker & Kubernetes).
- Experience with relational and/or non-relational databases (SQL, NoSQL).
- Sound understanding of Microservices architecture and RESTful APIs.
- Containerization experience using Docker and orchestration via Kubernetes.
- Familiarity with Linux-based development environments.
- Exposure to modern SDLC tools – Maven, Git, Jenkins, etc.
- Good understanding of CI/CD pipelines and cloud-based deployment.
Soft Skills:
- Self-driven, proactive, and an individual contributor.
- Strong problem-solving and analytical skills.
- Excellent communication and interpersonal abilities.
- Able to plan, prioritize, and manage tasks independently.
Nice-to-Have Skills:
- Exposure to frontend technologies like Angular, JavaScript, HTML5, and CSS.
Looking for Fresher developers
Responsibilities:
- Implement integrations requested by customers
- Deploy updates and fixes
- Provide Level 2 technical support
- Build tools to reduce occurrences of errors and improve customer experience
- Develop software to integrate with internal back-end systems
- Perform root cause analysis for production errors
- Investigate and resolve technical issues
- Develop scripts to automate visualization
- Design procedures for system troubleshooting and maintenance
Requirements and skill:
Knowledge in DevOps Engineer or similar software engineering role
Good knowledge of Terraform, Kubernetes
Working knowledge on AWS, Google Cloud
You can directly contact me on nine three one six one two zero one three two

Position : Senior Data Analyst
Experience Required : 5 to 8 Years
Location : Hyderabad or Bangalore (Work Mode: Hybrid – 3 Days WFO)
Shift Timing : 11:00 AM – 8:00 PM IST
Notice Period : Immediate Joiners Only
Job Summary :
We are seeking a highly analytical and experienced Senior Data Analyst to lead complex data-driven initiatives that influence key business decisions.
The ideal candidate will have a strong foundation in data analytics, cloud platforms, and BI tools, along with the ability to communicate findings effectively across cross-functional teams. This role also involves mentoring junior analysts and collaborating closely with business and tech teams.
Key Responsibilities :
- Lead the design, execution, and delivery of advanced data analysis projects.
- Collaborate with stakeholders to identify KPIs, define requirements, and develop actionable insights.
- Create and maintain interactive dashboards, reports, and visualizations.
- Perform root cause analysis and uncover meaningful patterns from large datasets.
- Present analytical findings to senior leaders and non-technical audiences.
- Maintain data integrity, quality, and governance in all reporting and analytics solutions.
- Mentor junior analysts and support their professional development.
- Coordinate with data engineering and IT teams to optimize data pipelines and infrastructure.
Must-Have Skills :
- Strong proficiency in SQL and Databricks
- Hands-on experience with cloud data platforms (AWS, Azure, or GCP)
- Sound understanding of data warehousing concepts and BI best practices
Good-to-Have :
- Experience with AWS
- Exposure to machine learning and predictive analytics
- Industry-specific analytics experience (preferred but not mandatory)