50+ Docker Jobs in India
Apply to 50+ Docker Jobs on CutShort.io. Find your next job, effortlessly. Browse Docker Jobs and apply today!


Key Responsibilities:
• Backend Development:
o Design, develop, and maintain microservices and APIs using .NET Core. Should have a good understanding of .NET Framework.
o Implement RESTful APIs, ensuring high performance and security.
o Optimize database queries and design schemas for SQL Server / Snowflake / MongoDB.
• Software Architecture & DevOps:
o Design and implement scalable microservices architecture.
o Work with Docker, Kubernetes, and CI/CD pipelines for deployment and automation.
o Ensure best practices in security, scalability, and performance.
• Collaboration & Agile Development:
o Work closely with UI/UX designers, backend engineers, and product managers.
o Participate in Agile/Scrum ceremonies, code reviews, and knowledge-sharing sessions.
o Write clean, maintainable, and well-documented code.
________________________________________
Required Skills & Qualifications:
o 8+ years of experience as a Full-Stack Developer.
o Strong experience in .NET Core, C#.
o Proficiency in React.js, JavaScript (ES6+), TypeScript.
o Experience with RESTful APIs, Microservices architecture.
o Knowledge of SQL / NoSQL databases (SQL Server, Snowflake, MongoDB).
o Experience with Git, CI/CD pipelines, Docker, and Kubernetes.
o Familiarity with Cloud services (Azure, AWS, or GCP) is a plus.
o Strong debugging and troubleshooting skills.
________________________________________
Nice-to-Have:
• Experience with GraphQL, gRPC, WebSockets.
• Exposure to serverless architecture and cloud-based solutions.
• Knowledge of authentication/authorization frameworks (OAuth, JWT, Identity Server).
• Experience with unit testing and integration testing.
Key Responsibilities
- Team Leadership & Management
- Lead, mentor, and manage a team of 6–10 full-stack engineers (MERN stack).
- Drive performance management, team development, and career progression.
- Collaborate with product managers, designers, and QA to ensure timely delivery of high-quality features.
- Technical Ownership
- Own the architecture, design, and implementation of core fintech product modules using MongoDB, Express.js, React.js, and Node.js.
- Review and enforce coding standards, CI/CD pipelines, and software quality best practices.
- Ensure system scalability, security, performance, and availability.
- Project Delivery
- Translate business requirements into scalable tech solutions.
- Ensure sprint planning, estimation, execution, and stakeholder communication.
- Proactively identify risks and bottlenecks and implement mitigations.
- Product & Innovation Focus
- Work closely with leadership on technology strategy and product roadmaps.
- Foster a culture of innovation, continuous learning, and engineering excellence.
Requirements
- 8+ years of software development experience, with at least 3+ years in team leadership roles.
- Proven track record of working in product-based fintech companies.
- Deep hands-on experience in the MERN stack (MongoDB, Express.js, React.js, Node.js).
- Strong understanding of cloud-native architectures (AWS/GCP/Azure).
- Proficiency in REST APIs, microservices, Docker, and container orchestration.
- Exposure to security best practices in fintech (e.g., data encryption, secure auth flows).
- Strong debugging, optimization, and analytical skills.
- Excellent communication, interpersonal, and stakeholder management skills.
- Ability to work in a fast-paced, startup-like environment with a product ownership mindset.
Nice to Have
- • Experience with TypeScript, GraphQL, or WebSockets.
- • Exposure to DevOps practices and observability tools.
- • Prior experience building lending, payments, or investment platforms.
What you can expect in return:
- ESOPs basis performance
- Health insurance
- Statutory benefits like PF & Gratuity
- Flexible Working Structure
- Professional development opportunities
- Collaborative and inclusive work culture

Job Title: Senior Software Engineer - Backend
About the firm:
Sustainability lies at the core of Stantech AI. Our vision is to empower organizations to derive actionable insights—effectuating a smarter way of working. We operate on the premise that each client is unique and as such requires their own idiosyncratic solutions. Putting this principle into practice, we deliver tailor-made solutions to digitalize, optimize, and strategize fundamental processes underpinning client organizations. For more information, please refer to our website: www.stantech.ai
Job Description:
As a Senior Software Engineer at Stantech AI, you will play a pivotal role in designing, developing, and maintaining enterprise-grade backend services and APIs that cater to the unique needs of our clients. You will be a key member of our engineering team and will contribute to the success of projects by leveraging your expertise in Python, SQL, and modern DevOps practices.
Key Responsibilities:
- Design, develop, and maintain high-performance backend applications and RESTful APIs using Python FastAPI framework.
- Optimize and maintain relational databases with SQL (data modeling, query optimization, and sharding) to ensure data integrity and scalability.
- Create, configure, and manage CI/CD pipelines using GitLab CI for automated build, test, and deployment workflows.
- Collaborate with cross-functional teams (data scientists, frontend engineers, DevOps) to gather requirements and deliver robust, scalable, and user-friendly solutions.
- Participate in architectural and technical decisions to drive innovation, ensure reliability, and improve system performance.
- Conduct code reviews, enforce best practices, and mentor junior engineers.
- Troubleshoot, diagnose, and resolve production issues in a timely manner.
- Stay up-to-date with industry trends, emerging technologies, and best practices.
- Bonus: Hands-on experience with server-level configuration and infrastructure—setting up load balancers, API gateways, and reverse proxies.
Qualifications:
- Bachelor’s or Master’s degree in Computer Science, Engineering, or a related field.
- Minimum 3 years of professional experience in backend development, with strong expertise in Python and SQL.
- Proven track record building and maintaining CI/CD pipelines using GitLab CI.
- Familiarity with containerization and orchestration technologies: Docker, Kubernetes.
- Solid understanding of software development lifecycle (SDLC) best practices, design patterns, and version control (Git).
- Excellent problem-solving, debugging, and communication skills.
- Ability to work independently and collaboratively in a fast-paced environment.
- Plus: Experience with front-end technologies (HTML, CSS, JavaScript) and cloud platforms (AWS, GCP, Azure).
Financial Package:
Competitive salary in line with experience: ₹10–20 Lakhs per annum, contingent on qualifications and experience.

🚀 Hiring: MERN Stack at Deqode
⭐ Experience: 2+ Years
📍 Location: Mumbai
⭐ Work Mode:- 5 Days Work From Office
⏱️ Notice Period: Immediate Joiners or 15 Days
(Only immediate joiners & candidates serving notice period)
MERN Stack (2+ Years of Experience) - Mumbai
🔹 Experience: 2 to 4 Years
🔹Skills: MongoDB, Express, React, Node, Docker, Kubernetes, Kafka

Job Title : Python Backend Engineer (with MLOps & LLMOps Experience)
Experience : 4 to 8 Years
Location : Gurgaon Sector - 43
Employment Type : Full-time
Job Summary :
We are looking for an experienced Python Backend Engineer with a strong background in FastAPI, Django, and hands-on exposure to MLOps and LLMOps practices.
The ideal candidate will be responsible for building scalable backend solutions, integrating AI/ML models into production environments, and implementing efficient pipelines for machine learning and large language model operations.
Mandatory Skills : Python, FastAPI, Django, MLOps, LLMOps, REST API development, Docker, Kubernetes, Cloud (AWS/Azure/GCP), CI/CD.
Key Responsibilities :
- Develop, optimize, and maintain backend services using Python (FastAPI, Django).
- Design and implement API endpoints for high-performance and secure data exchange.
- Collaborate with data science teams to deploy ML/LLM models into production using MLOps/LLMOps best practices.
- Build and manage CI/CD pipelines for ML models and ensure seamless integration with backend systems.
- Implement model monitoring, versioning, and retraining workflows for machine learning and large language models.
- Optimize backend performance for scalability and reliability in AI-driven applications.
- Work with Docker, Kubernetes, and cloud platforms (AWS/Azure/GCP) for deployment and orchestration.
- Ensure best practices in code quality, testing, and security for all backend and model deployment workflows.
Required Skills & Qualifications :
- 4 to 8 years of experience as a Backend Engineer with strong expertise in Python.
- Proficient in FastAPI and Django frameworks for API and backend development.
- Hands-on experience with MLOps and LLMOps workflows (model deployment, monitoring, scaling).
- Familiarity with machine learning model lifecycle and integration into production systems.
- Strong knowledge of RESTful APIs, microservices architecture, and asynchronous programming.
- Experience with Docker, Kubernetes, and cloud environments (AWS, Azure, or GCP).
- Exposure to CI/CD pipelines and DevOps tools.
- Good understanding of Git, version control, and testing frameworks.
Nice to Have :
- Experience with LangChain, Hugging Face, or similar LLM frameworks.
- Knowledge of data pipelines, feature engineering, and ML frameworks (TensorFlow, PyTorch, etc.).
- Understanding of vector databases (Pinecone, Chroma, etc.).
Education :
- Bachelor’s or Master’s degree in Computer Science, Engineering, or related field.
Key Responsibilities
- Backend & API Engineering (60% Focus)
- Design and implement REST APIs and microservices for high-performance AI systems.
- Apply enterprise object-oriented programming best practices for secure, scalable backend services.
- Integrate AI-powered features with cloud-native architectures.
- Generative AI & LLM Development (40% Focus)
- Build LLM-powered features with the OpenAI API or other LLM APIs (reasoning and non-reasoning models, temperature tuning, version control).
- Implement retrieval-augmented generation (RAG).
- Apply advanced prompt engineering and model tuning techniques for optimized results.
- Deploy and manage solutions using Docker and secure integrations (e.g., SSO, Google Drive).
What We’re Looking For
- 10+ years of backend engineering experience (REST APIs, microservices, OO enterprise architecture).
- 3+ years of experience building AI/ML solutions (LLMs, RAG, OpenAI API).
- Strong hands-on Python expertise and object-oriented design patterns.
- Hands-on experience with Langchain, Lambda, Docker, and secure system integrations.
- Proven track record delivering production-ready, scalable applications.

About Us
CLOUDSUFI, a Google Cloud Premier Partner, a Data Science and Product Engineering organization building Products and Solutions for Technology and Enterprise industries. We firmly believe in the power of data to transform businesses and make better decisions. We combine unmatched experience in business processes with cutting edge infrastructure and cloud services. We partner with our customers to monetize their data and make enterprise data dance.
Our Values
We are a passionate and empathetic team that prioritizes human values. Our purpose is to elevate the quality of lives for our family, customers, partners and the community.
Equal Opportunity Statement
CLOUDSUFI is an equal opportunity employer. We celebrate diversity and are committed to creating an inclusive environment for all employees. All qualified candidates receive consideration for employment without regard to race, color, religion, gender, gender identity or expression, sexual orientation, and national origin status. We provide equal opportunities in employment, advancement, and all other areas of our workplace. Please explore more at https://www.cloudsufi.com/.
Role Overview:
As a Senior Data Scientist / AI Engineer, you will be a key player in our technical leadership. You will be responsible for designing, developing, and deploying sophisticated AI and Machine Learning solutions, with a strong emphasis on Generative AI and Large Language Models (LLMs). You will architect and manage scalable AI microservices, drive research into state-of-the-art techniques, and translate complex business requirements into tangible, high-impact products. This role requires a blend of deep technical expertise, strategic thinking, and leadership.
Key Responsibilities:
- Architect & Develop AI Solutions: Design, build, and deploy robust and scalable machine learning models, with a primary focus on Natural Language Processing (NLP), Generative AI, and LLM-based Agents.
- Build AI Infrastructure: Create and manage AI-driven microservices using frameworks like Python FastAPI, ensuring high performance and reliability.
- Lead AI Research & Innovation: Stay abreast of the latest advancements in AI/ML. Lead research initiatives to evaluate and implement state-of-the-art models and techniques for performance and cost optimization.
- Solve Business Problems: Collaborate with product and business teams to understand challenges and develop data-driven solutions that create significant business value, such as building business rule engines or predictive classification systems.
- End-to-End Project Ownership: Take ownership of the entire lifecycle of AI projects—from ideation, data processing, and model development to deployment, monitoring, and iteration on cloud platforms.
- Team Leadership & Mentorship: Lead learning initiatives within the engineering team, mentor junior data scientists and engineers, and establish best practices for AI development.
- Cross-Functional Collaboration: Work closely with software engineers to integrate AI models into production systems and contribute to the overall system architecture.
Required Skills and Qualifications
- Master’s (M.Tech.) or Bachelor's (B.Tech.) degree in Computer Science, Artificial Intelligence, Information Technology, or a related field.
- 6+ years of professional experience in a Data Scientist, AI Engineer, or related role.
- Expert-level proficiency in Python and its core data science libraries (e.g., PyTorch, Huggingface Transformers, Pandas, Scikit-learn).
- Demonstrable, hands-on experience building and fine-tuning Large Language Models (LLMs) and implementing Generative AI solutions.
- Proven experience in developing and deploying scalable systems on cloud platforms, particularly AWS. Experience with GCS is a plus.
- Strong background in Natural Language Processing (NLP), including experience with multilingual models and transcription.
- Experience with containerization technologies, specifically Docker.
- Solid understanding of software engineering principles and experience building APIs and microservices.
Preferred Qualifications
- A strong portfolio of projects. A track record of publications in reputable AI/ML conferences is a plus.
- Experience with full-stack development (Node.js, Next.js) and various database technologies (SQL, MongoDB, Elasticsearch).
- Familiarity with setting up and managing CI/CD pipelines (e.g., Jenkins).
- Proven ability to lead technical teams and mentor other engineers.
- Experience developing custom tools or packages for data science workflows.
Role Overview
We are seeking a DevOps Engineer with 2 years of experience to join our innovative team. The ideal
candidate will bridge the gap between development and operations, implementing and maintaining our
cloud infrastructure while ensuring secure deployment pipelines and robust security practices for our
client projects.
Responsibilities:
- Design, implement, and maintain CI/CD pipelines.
- Containerize applications using Docker and orchestrate deployments
- Manage and optimize cloud infrastructure on AWS and Azure platforms
- Monitor system performance and implement automation for operational tasks to ensure optimal
- performance, security, and scalability.
- Troubleshoot and resolve infrastructure and deployment issues
- Create and maintain documentation for processes and configurations
- Collaborate with cross-functional teams to gather requirements, prioritise tasks, and contribute to project completion.
- Stay informed about emerging technologies and best practices within the fields of DevOps and cloud computing.
Requirements:
- 2+ years of hands-on experience with AWS cloud services
- Strong proficiency in CI/CD pipeline configuration
- Expertise in Docker containerisation and container management
- Proficiency in shell scripting (Bash/Power-Shell)
- Working knowledge of monitoring and logging tools
- Knowledge of network security and firewall configuration
- Strong communication and collaboration skills, with the ability to work effectively within a team
- environment
- Understanding of networking concepts and protocols in AWS and/or Azure
Job Title: Backend Engineer - NodeJS, NestJS, and Python
Location: Hybrid weekly ⅔ days WFO (Bengaluru- India)
About the role:
We are looking for a skilled and passionate Senior Backend Developer to join our dynamic team. The ideal candidate should have strong experience in Node.js and NestJS, along with a solid understanding of database management, query optimization, and microservices architecture. As a backend developer, you will be responsible for developing and maintaining scalable backend systems, building robust APIs, integrating databases, and working closely with frontend and DevOps teams to deliver high-quality software solutions.
What You'll Do 🛠️
- Design, develop, and maintain server-side logic using Node.js, NestJS, and Python.
- Develop and integrate RESTful APIs and microservices to support scalable systems.
- Work with NoSQL and SQL databases (e.g., MongoDB, PostgreSQL, MySQL) to create and manage schemas, write complex queries, and optimize performance.
- Collaborate with cross-functional teams including frontend, DevOps, and QA.
- Ensure code quality, maintainability, and scalability through code reviews, testing, and documentation.
- Monitor and troubleshoot production systems, ensuring high availability and performance.
- Implement security and data protection best practices.
What You'll Bring 💼
- 4 to 6 years of professional experience as a backend developer.
- Strong proficiency in Node.js and NestJS framework.
- Good hands-on experience with Python (Django/Flask experience is a plus).
- Solid understanding of relational and non-relational databases.
- Proficient in writing complex NoSQL queries and SQL queries
- Experience with microservices architecture and distributed systems.
- Familiarity with version control systems like Git.
- Basic understanding of containerization (e.g., Docker) and cloud services is a plus.
- Excellent problem-solving skills and a collaborative mindset.
Bonus Points ➕
- Experience with CI/CD pipelines.
- Exposure to cloud platforms like AWS, GCP or Azure.
- Familiarity with event-driven architecture or message brokers (MQTT, Kafka, RabbitMQ)
Why this role matters
You will help build the company from the ground up—shaping our culture and having an impact from Day 1 as part of the foundational team.

Job Title: Backend Engineer - NodeJS, NestJS, and Python
Location: Hybrid weekly ⅔ days WFO (Bengaluru- India)
About the role:
We are looking for a skilled and passionate Senior Backend Developer to join our dynamic team. The ideal candidate should have strong experience in Node.js and NestJS, along with a solid understanding of database management, query optimization, and microservices architecture. As a backend developer, you will be responsible for developing and maintaining scalable backend systems, building robust APIs, integrating databases, and working closely with frontend and DevOps teams to deliver high-quality software solutions.
What You'll Do 🛠️
- Design, develop, and maintain server-side logic using Node.js, NestJS, and Python.
- Develop and integrate RESTful APIs and microservices to support scalable systems.
- Work with NoSQL and SQL databases (e.g., MongoDB, PostgreSQL, MySQL) to create and manage schemas, write complex queries, and optimize performance.
- Collaborate with cross-functional teams including frontend, DevOps, and QA.
- Ensure code quality, maintainability, and scalability through code reviews, testing, and documentation.
- Monitor and troubleshoot production systems, ensuring high availability and performance.
- Implement security and data protection best practices.
What You'll Bring 💼
- 4 to 6 years of professional experience as a backend developer.
- Strong proficiency in Node.js and NestJS framework.
- Good hands-on experience with Python (Django/Flask experience is a plus).
- Solid understanding of relational and non-relational databases.
- Proficient in writing complex NoSQL queries and SQL queries
- Experience with microservices architecture and distributed systems.
- Familiarity with version control systems like Git.
- Basic understanding of containerization (e.g., Docker) and cloud services is a plus.
- Excellent problem-solving skills and a collaborative mindset.
Bonus Points ➕
- Experience with CI/CD pipelines.
- Exposure to cloud platforms like AWS, GCP or Azure.
- Familiarity with event-driven architecture or message brokers (MQTT, Kafka, RabbitMQ)
Why this role matters
You will help build the company from the ground up—shaping our culture and having an impact from Day 1 as part of the foundational team.
Role Description
This is a full-time on-site role for a Cloud & DevOps Engineer located in Coimbatore. The Cloud & DevOps Engineer will be responsible for managing and optimizing cloud infrastructure, implementing continuous integration and deployment processes, and ensuring the smooth operation of services. The role will involve working with Kubernetes for container orchestration and managing Linux environments.
Qualifications
- Proficiency in Software Development skills
- Experience with Continuous Integration and Deployment processes
- Skills in Kubernetes management and orchestration
- Strong understanding of Linux operating systems
- Ability to work collaboratively in a team environment
- Relevant certifications in cloud and DevOps technologies are a plus

Job Title: Sr Dev Ops Engineer
Location: Bengaluru- India (Hybrid work type)
Reports to: Sr Engineer manager
About Our Client :
We are a solution-based, fast-paced tech company with a team that thrives on collaboration and innovative thinking. Our Client's IoT solutions provide real-time visibility and actionable insights for logistics and supply chain management. Cloud-based, AI-enhanced metrics coupled with patented hardware optimize processes, inform strategic decision making and enable intelligent supply chains without the costly infrastructure
About the role : We're looking for a passionate DevOps Engineer to optimize our software delivery and infrastructure. You'll build and maintain CI/CD pipelines for our microservices, automate infrastructure, and ensure our systems are reliable, scalable, and secure. If you thrive on enhancing performance and fostering operational excellence, this role is for you.
What You'll Do 🛠️
- Cloud Platform Management: Administer and optimize AWS resources, ensuring efficient billing and cost management.
- Billing & Cost Optimization: Monitor and optimize cloud spending.
- Containerization & Orchestration: Deploy and manage applications and orchestrate them.
- Database Management: Deploy, manage, and optimize database instances and their lifecycles.
- Authentication Solutions: Implement and manage authentication systems.
- Backup & Recovery: Implement robust backup and disaster recovery strategies, for Kubernetes cluster and database backups.
- Monitoring & Alerting: Set up and maintain robust systems using tools for application and infrastructure health and integrate with billing dashboards.
- Automation & Scripting: Automate repetitive tasks and infrastructure provisioning.
- Security & Reliability: Implement best practices and ensure system performance and security across all deployments.
- Collaboration & Support: Work closely with development teams, providing DevOps expertise and support for their various application stacks.
What You'll Bring 💼
- Minimum of 4 years of experience in a DevOps or SRE role.
- Strong proficiency in AWS Cloud, including services like Lambda, IoT Core, ElastiCache, CloudFront, and S3.
- Solid understanding of Linux fundamentals and command-line tools.
- Extensive experience with CI/CD tools, GitLab CI.
- Hands-on experience with Docker and Kubernetes, specifically AWS EKS.
- Proven experience deploying and managing microservices.
- Expertise in database deployment, optimization, and lifecycle management (MongoDB, PostgreSQL, and Redis).
- Experience with Identity and Access management solutions like Keycloak.
- Experience implementing backup and recovery solutions.
- Familiarity with optimizing scaling, ideally with Karpenter.
- Proficiency in scripting (Python, Bash).
- Experience with monitoring tools such as Prometheus, Grafana, AWS CloudWatch, Elastic Stack.
- Excellent problem-solving and communication skills.
Bonus Points ➕
- Basic understanding of MQTT or general IoT concepts and protocols.
- Direct experience optimizing React.js (Next.js), Node.js (Express.js, Nest.js) or Python (Flask) deployments in a containerized environment.
- Knowledge of specific AWS services relevant to application stacks.
- Contributions to open-source projects related to Kubernetes, MongoDB, or any of the mentioned frameworks.
- AWS Certifications (AWS Certified DevOps Engineer, AWS Certified Solutions Architect, AWS Certified SysOps Administrator, AWS Certified Advanced Networking).
Why this role:
•You will help build the company from the ground up—shaping our culture and having an impact from Day 1 as part of the foundational team.
Job Title: Sr Dev Ops Engineer
Location: Bengaluru- India (Hybrid work type)
Reports to: Sr Engineer manager
About Our Client :
We are a solution-based, fast-paced tech company with a team that thrives on collaboration and innovative thinking. Our Client's IoT solutions provide real-time visibility and actionable insights for logistics and supply chain management. Cloud-based, AI-enhanced metrics coupled with patented hardware optimize processes, inform strategic decision making and enable intelligent supply chains without the costly infrastructure
About the role : We're looking for a passionate DevOps Engineer to optimize our software delivery and infrastructure. You'll build and maintain CI/CD pipelines for our microservices, automate infrastructure, and ensure our systems are reliable, scalable, and secure. If you thrive on enhancing performance and fostering operational excellence, this role is for you.
What You'll Do 🛠️
- Cloud Platform Management: Administer and optimize AWS resources, ensuring efficient billing and cost management.
- Billing & Cost Optimization: Monitor and optimize cloud spending.
- Containerization & Orchestration: Deploy and manage applications and orchestrate them.
- Database Management: Deploy, manage, and optimize database instances and their lifecycles.
- Authentication Solutions: Implement and manage authentication systems.
- Backup & Recovery: Implement robust backup and disaster recovery strategies, for Kubernetes cluster and database backups.
- Monitoring & Alerting: Set up and maintain robust systems using tools for application and infrastructure health and integrate with billing dashboards.
- Automation & Scripting: Automate repetitive tasks and infrastructure provisioning.
- Security & Reliability: Implement best practices and ensure system performance and security across all deployments.
- Collaboration & Support: Work closely with development teams, providing DevOps expertise and support for their various application stacks.
What You'll Bring 💼
- Minimum of 4 years of experience in a DevOps or SRE role.
- Strong proficiency in AWS Cloud, including services like Lambda, IoT Core, ElastiCache, CloudFront, and S3.
- Solid understanding of Linux fundamentals and command-line tools.
- Extensive experience with CI/CD tools, GitLab CI.
- Hands-on experience with Docker and Kubernetes, specifically AWS EKS.
- Proven experience deploying and managing microservices.
- Expertise in database deployment, optimization, and lifecycle management (MongoDB, PostgreSQL, and Redis).
- Experience with Identity and Access management solutions like Keycloak.
- Experience implementing backup and recovery solutions.
- Familiarity with optimizing scaling, ideally with Karpenter.
- Proficiency in scripting (Python, Bash).
- Experience with monitoring tools such as Prometheus, Grafana, AWS CloudWatch, Elastic Stack.
- Excellent problem-solving and communication skills.
Bonus Points ➕
- Basic understanding of MQTT or general IoT concepts and protocols.
- Direct experience optimizing React.js (Next.js), Node.js (Express.js, Nest.js) or Python (Flask) deployments in a containerized environment.
- Knowledge of specific AWS services relevant to application stacks.
- Contributions to open-source projects related to Kubernetes, MongoDB, or any of the mentioned frameworks.
- AWS Certifications (AWS Certified DevOps Engineer, AWS Certified Solutions Architect, AWS Certified SysOps Administrator, AWS Certified Advanced Networking).
Why this role:
•You will help build the company from the ground up—shaping our culture and having an impact from Day 1 as part of the foundational team.
🚀 Why join Bound AI (OIP Insurtech)?
We build real-world AI workflows that transform insurance operations—from underwriting to policy issuance. You’ll join a fast-growing, global team of engineers and innovators tackling the toughest problems in document intelligence and agent orchestration. We move fast, ship impact, and value autonomy over bureaucracy.
🧭 What You'll Be Doing
- Design and deliver end‑to‑end AI solutions: from intake of SOVs, loss runs, and documents to deployed intelligent agent workflows.
- Collaborate closely with stakeholders (product, operations, engineering) to architect scalable ML & GenAI systems that solve real insurance challenges.
- Translate business needs into architecture diagrams, data flows, and system integrations.
- Choose and integrate components such as RAG pipelines, LLM orchestration (LangChain, DSPy), vector databases, and MLOps tooling.
- Oversee technical proof-of-concepts, pilot projects, and production rollout strategies.
- Establish governance and best practices for model lifecycle, monitoring, error handling, and versioning.
- Act as a trusted advisor and technical leader—mentor engineers and evangelize design principles across teams.
🎯 What We’re Looking For
- 6+ years of experience delivering technical solutions in machine learning, AI engineering or solution architecture.
- Proven track record leading design, deployment, and integration of GenAI-based systems (LLM tuning, RAG, multi-agent orchestration).
- Fluency with Python production code, cloud platforms (AWS, GCP, Azure), and container orchestration tools.
- Excellent communication skills—able to bridge technical strategy and business outcomes with clarity.
- Startup mindset—resourceful, proactive, and hands‑on when needed.
- Bonus: experience with insurance-specific workflows or document intelligence domains (SOVs, loss runs, ACORD forms).
🛠️ Core Skills & Tools
- Foundation models, LLM pipelines, and vector-based retrieval (embedding search, RAG).
- Architecture modeling and integration: APIs, microservices, orchestration frameworks (LangChain, Haystack, DSPy).
- MLOps: CI/CD for models, monitoring, feedback loops, and retraining pipelines.
- Data engineering: preprocessing, structured/unstructured data integration, pipelines.
- Infrastructure: Kubernetes, Docker, cloud deployment, serverless components.
📈 Why This Role Matters
As an AI Solution Architect, you’ll shape the blueprint for how AI transforms insurance workflows—aligning product strategy, operational impact, and technical scalability. You're not just writing code; you’re orchestrating systems that make labor-intensive processes smarter, faster, and more transparent.

💼 Job Title: Full Stack Developer (*Fresher/experienced*)
🏢 Company: SDS Softwares
💻 Location: Work from Home
💸 Salary range: ₹7,000 - ₹18,000 per month (based on knowledge and interview)
🕛 Shift Timings: 12 PM to 9 PM
About the role: As a Full Stack Developer, you will work on both the front-end and back-end of web applications. You will be responsible for developing user-friendly interfaces and maintaining the overall functionality of our projects.
⚜️ Key Responsibilities:
- Collaborate with cross-functional teams to define, design, and ship new features.
- Develop and maintain high-quality web applications (frontend + backend )
- Troubleshoot and debug applications to ensure peak performance.
- Participate in code reviews and contribute to the team’s knowledge base.
⚜️ Required Skills:
- Proficiency in HTML, CSS, JavaScript, React.js for front-end development. ✅
- Understanding of server-side languages such as Node.js, Python, or PHP. ✅
- Familiarity with database technologies such as MySQL, MongoDB, or ✅ PostgreSQL.
- Basic knowledge of version control systems, particularly Git.
- Strong problem-solving skills and attention to detail.
- Excellent communication skills and a team-oriented mindset.
💠 Qualifications:
- Recent graduates or individuals with internship experience (6 months to 1.5years) in software development.
- Must have a personal laptop and stable internet connection.
- Ability to join immediately is preferred.
If you are passionate about coding and eager to learn, we would love to hear from you. 👍


🚀 Hiring: Dot net full stack at Deqode
⭐ Experience: 8+ Years
📍 Location: Bangalore | Mumbai | Pune | Gurgaon | Chennai | Hyderabad
⭐ Work Mode:- Hybrid
⏱️ Notice Period: Immediate Joiners
(Only immediate joiners & candidates serving notice period)
We’re looking for an experienced Dotnet Full Stack Developer with strong hands-on skills in ReactJS, .NET Core, and Azure Cloud Services (Azure Functions, Azure SQL, APIM, etc.).
⭐ Must-Have Skills:-
➡️ Design and develop scalable web applications using ReactJS, C#, and .NET Core.
➡️Azure (Functions, App Services, SQL, APIM, Service Bus)
➡️Familiarity with DevOps practices, CI/CD pipelines, Docker, and Kubernetes.
➡️Advanced experience in Entity Framework Core and SQL Server.
➡️Expertise in RESTful API development and microservices.
Job Summary:
We are seeking a highly skilled and proactive DevOps Engineer with 4+ years of experience to join our dynamic team. This role requires strong technical expertise across cloud infrastructure, CI/CD pipelines, container orchestration, and infrastructure as code (IaC). The ideal candidate should also have direct client-facing experience and a proactive approach to managing both internal and external stakeholders.
Key Responsibilities:
- Collaborate with cross-functional teams and external clients to understand infrastructure requirements and implement DevOps best practices.
- Design, build, and maintain scalable cloud infrastructure on AWS (EC2, S3, RDS, ECS, etc.).
- Develop and manage infrastructure using Terraform or CloudFormation.
- Manage and orchestrate containers using Docker and Kubernetes (EKS).
- Implement and maintain CI/CD pipelines using Jenkins or GitHub Actions.
- Write robust automation scripts using Python and Shell scripting.
- Monitor system performance and availability, and ensure high uptime and reliability.
- Execute and optimize SQL queries for MSSQL and PostgreSQL databases.
- Maintain clear documentation and provide technical support to stakeholders and clients.
Required Skills:
- Minimum 4+ years of experience in a DevOps or related role.
- Proven experience in client-facing engagements and communication.
- Strong knowledge of AWS services – EC2, S3, RDS, ECS, etc.
- Proficiency in Infrastructure as Code using Terraform or CloudFormation.
- Hands-on experience with Docker and Kubernetes (EKS).
- Strong experience in setting up and maintaining CI/CD pipelines with Jenkins or GitHub.
- Solid understanding of SQL and working experience with MSSQL and PostgreSQL.
- Proficient in Python and Shell scripting.
Preferred Qualifications:
- AWS Certifications (e.g., AWS Certified DevOps Engineer) are a plus.
- Experience working in Agile/Scrum environments.
- Strong problem-solving and analytical skills.
Work Mode & Timing:
- Hybrid – Pune-based candidates preferred.
- Working hours: 12:30 PM to 9:30 PM IST to align with client time zones.
Requirements:
• Strong expertise in Java (Java 8 and Java 17 or higher).
• Proficiency in frameworks like Spring Boot, Microservice Architecture.
• Experience in cloud-native development and deployment on Microsoft Azure.
• Hands-on experience with Azure services such as Azure App Services, Functions,
• Kubernetes (AKS), Azure DevOps, Blob Storage, and Service Bus.
• Knowledge of RESTful APIs, SOAP, and microservices architecture.
• Solid understanding of database technologies (e.g., Azure SQL)
Springer Capital is a cross-border asset management firm focused on real estate investment banking in China and the USA. We are offering a remote internship for individuals passionate about automation, cloud infrastructure, and CI/CD pipelines. Start and end dates are flexible, and applicants may be asked to complete a short technical quiz or assignment as part of the application process.
Responsibilities:
▪ Assist in building and maintaining CI/CD pipelines to automate development workflows
▪ Monitor and improve system performance, reliability, and scalability
▪ Manage cloud-based infrastructure (e.g., AWS, Azure, or GCP)
▪ Support containerization and orchestration using Docker and Kubernetes
▪ Implement infrastructure as code using tools like Terraform or CloudFormation
▪ Collaborate with software engineering and data teams to streamline deployments
▪ Troubleshoot system and deployment issues across development and production environments
Site Reliability Engineer 2 (SRE 2)
Overview:
As a Site Reliability Engineer 2 (SRE 2), you play a dual role of a senior technical contributor and a team leader within the SRE team. In addition to ensuring system reliability, scalability, and performance, you will manage shift schedules, guide SRE 1 engineers, and ensure compliance with ITSM processes. Your focus will be both technical execution and operational excellence, ensuring that the team delivers high-quality, consistent support and reliability across environments.
Key Responsibilities:
1. Infrastructure Reliability and Technical Leadership
· Ensure high availability, scalability, and performance of systems through proactive monitoring, automation, and continuous improvement.
· Lead efforts in improving infrastructure observability using tools like Prometheus, Alertmanager, Grafana, and other telemetry systems.
· Serve as an escalation point for complex technical incidents and outages, providing guidance to SRE 1 engineers.
2. Team Oversight and Performance Management
· Provide technical and operational leadership to SRE 1 engineers, ensuring daily tasks are executed as per standards.
· Review SRE 1 work regularly to ensure adherence to best practices, SOPs, and incident response protocols.
· Mentor and train junior team members to enhance their technical skills and operational understanding.
· Conduct regular feedback sessions and contribute to performance evaluations.
3. Shift Management and 24/7 Coverage
· Design, implement, and manage rotating shift schedules to ensure optimal 24/7 support coverage.
· Monitor shift adherence, workload distribution, and overall team health.
· Ensure proper handovers between shifts with complete documentation and context sharing.
4. ITSM Process and Compliance
· Own and enforce ITSM processes, including Incident Management, Change Management, Problem Management, and Service Request Fulfillment.
· Ensure that all incidents, changes, and problems are logged, categorized, and resolved or escalated as per SLA.
· Continuously assess and improve ITSM processes in collaboration with internal stakeholders and audit teams.
5. Incident and Problem Management
· Lead major incident investigations and coordinate response efforts across teams.
· Oversee root cause analysis and implementation of long-term fixes for recurring issues.
· Maintain detailed incident logs and postmortem reports for high-priority incidents.
6. Change and Maintenance Oversight
· Review and approve change requests initiated by SRE 1 or other team members.
· Ensure execution of maintenance tasks adheres to predefined SOPs and does not impact system stability.
· Track and analyze impact of changes to continuously improve reliability metrics.
7. Reporting and Stakeholder Communication
· Create and present weekly/monthly reports on SRE metrics, team performance, incident trends, and capacity planning.
· Collaborate with cross-functional teams, including engineering, QA, support, and product, to align operational goals.
· Provide updates to leadership on key incidents, system health, and team productivity.
About Us
CLOUDSUFI, a Google Cloud Premier Partner, is a global leading provider of data-driven digital transformation across cloud-based enterprises. With a global presence and focus on Software & Platforms, Life sciences and Healthcare, Retail, CPG, financial services, and supply chain, CLOUDSUFI is positioned to meet customers where they are in their data monetization journey.
Our Values
We are a passionate and empathetic team that prioritizes human values. Our purpose is to elevate the quality of lives for our family, customers, partners and the community.
Equal Opportunity Statement
CLOUDSUFI is an equal opportunity employer. We celebrate diversity and are committed to creating an inclusive environment for all employees. All qualified candidates receive consideration for employment without regard to race, color, religion, gender, gender identity or expression, sexual orientation, and national origin status. We provide equal opportunities in employment, advancement, and all other areas of our workplace. Please explore more at https://www.cloudsufi.com/
What are we looking for
We are seeking a highly skilled and experienced Senior DevOps Engineer to join our team. The ideal candidate will have extensive expertise in modern DevOps tools and practices, particularly in managing CI/CD pipelines, infrastructure as code, and cloud-native environments. This role involves designing, implementing, and maintaining robust, scalable, and efficient infrastructure and deployment pipelines to support our development and operations teams.
Required Skills and Experience:
- 7+ years of experience in DevOps, infrastructure automation, or related fields.
- Advanced expertise in Terraform for infrastructure as code.
- Solid experience with Helm for managing Kubernetes applications.
- Proficient with GitHub for version control, repository management, and workflows.
- Extensive experience with Kubernetes for container orchestration and management.
- In-depth understanding of Google Cloud Platform (GCP) services and architecture.
- Strong scripting and automation skills (e.g., Python, Bash, or equivalent).
- Excellent problem-solving skills and attention to detail. - Strong communication and collaboration abilities in agile development environments.
Preferred Qualifications:
- Experience with other CI/CD tools (e.g., Jenkins, GitLab CI/CD).
- Knowledge of additional cloud platforms (e.g., AWS, Azure).
- Certification in Kubernetes (CKA/CKAD) or Google Cloud (GCP Professional DevOps Engineer).
Behavioral Competencies
• Must have worked with US/Europe based clients in onsite/offshore delivery models.
• Should have very good verbal and written communication, technical articulation, listening and presentation skills.
• Should have proven analytical and problem solving skills.
• Should have collaborative mindset for cross-functional team work
• Passion for solving complex search problems
• Should have demonstrated effective task prioritization, time management and internal/external stakeholder management skills.
• Should be a quick learner, self starter, go-getter and team player.
• Should have experience of working under stringent deadlines in a Matrix organization structure.

Job Title : Senior Python Developer – Product Engineering
Experience : 5 to 8 Years
Location : Pune, India (Hybrid – 3-4 days WFO, 1-2 days WFH)
Employment Type : Full-time
Commitment : Minimum 3 years (with end-of-term bonus)
Openings : 2 positions
- Junior : 3 to 5 Years
- Senior : 5 to 8 Years
Mandatory Skills : Python 3.x, REST APIs, multithreading, Celery, encryption (OpenSSL/cryptography.io), PostgreSQL/Redis, Docker/K8s, secure coding
Nice to Have : Experience with EFSS/DRM/DLP platforms, delta sync, file systems, LDAP/AD/SIEM integrations
🎯 Roles & Responsibilities :
- Design and develop backend services for DRM enforcement, file synchronization, and endpoint telemetry.
- Build scalable Python-based APIs interacting with file systems, agents, and enterprise infra.
- Implement encryption workflows, secure file handling, delta sync, and file versioning.
- Integrate with 3rd-party platforms: LDAP, AD, DLP, CASB, SIEM.
- Collaborate with DevOps to ensure high availability and performance of hybrid deployments.
- Participate in code reviews, architectural discussions, and mentor junior developers.
- Troubleshoot production issues and continuously optimize performance.
✅ Required Skills :
- 5 to 8 years of hands-on experience in Python 3.x development.
- Expertise in REST APIs, Celery, multithreading, and file I/O.
- Proficient in encryption libraries (OpenSSL, cryptography.io) and secure coding.
- Experience with PostgreSQL, Redis, SQLite, and Linux internals.
- Strong command over Docker, Kubernetes, CI/CD, and Git workflows.
- Ability to write clean, testable, and scalable code in production environments.
➕ Preferred Skills :
- Background in DRM, EFSS, DLP, or enterprise security platforms.
- Familiarity with file diffing, watermarking, or agent-based tools.
- Knowledge of compliance frameworks (GDPR, DPDP, RBI-CSF) is a plus.

Location: Hybrid/ Remote
Type: Contract / Full‑Time
Experience: 5+ Years
Qualification: Bachelor’s or Master’s in Computer Science or a related technical field
Responsibilities:
- Architect & implement the RAG pipeline: embeddings ingestion, vector search (MongoDB Atlas or similar), and context-aware chat generation.
- Design and build Python‑based services (FastAPI) for generating and updating embeddings.
- Host and apply LoRA/QLoRA adapters for per‑user fine‑tuning.
- Automate data pipelines to ingest daily user logs, chunk text, and upsert embeddings into the vector store.
- Develop Node.js/Express APIs that orchestrate embedding, retrieval, and LLM inference for real‑time chat.
- Manage vector index lifecycle and similarity metrics (cosine/dot‑product).
- Deploy and optimize on AWS (Lambda, EC2, SageMaker), containerization (Docker), and monitoring for latency, costs, and error rates.
- Collaborate with frontend engineers to define API contracts and demo endpoints.
- Document architecture diagrams, API specifications, and runbooks for future team onboarding.
Required Skills
- Strong Python expertise (FastAPI, async programming).
- Proficiency with Node.js and Express for API development.
- Experience with vector databases (MongoDB Atlas Vector Search, Pinecone, Weaviate) and similarity search.
- Familiarity with OpenAI’s APIs (embeddings, chat completions).
- Hands‑on with parameters‑efficient fine‑tuning (LoRA, QLoRA, PEFT/Hugging Face).
- Knowledge of LLM hosting best practices on AWS (EC2, Lambda, SageMaker).
Containerization skills (Docker):
- Good understanding of RAG architectures, prompt design, and memory management.
- Strong Git workflow and collaborative development practices (GitHub, CI/CD).
Nice‑to‑Have:
- Experience with Llama family models or other open‑source LLMs.
- Familiarity with MongoDB Atlas free tier and cluster management.
- Background in data engineering for streaming or batch processing.
- Knowledge of monitoring & observability tools (Prometheus, Grafana, CloudWatch).
- Frontend skills in React to prototype demo UIs.
We are looking for a highly skilled DevOps/Cloud Engineer with over 6 years of experience in infrastructure automation, cloud platforms, networking, and security. If you are passionate about designing scalable systems and love solving complex cloud and DevOps challenges—this opportunity is for you.
Key Responsibilities
- Design, deploy, and manage cloud-native infrastructure using Kubernetes (K8s), Helm, Terraform, and Ansible
- Automate provisioning and orchestration workflows for cloud and hybrid environments
- Manage and optimize deployments on AWS, Azure, and GCP for high availability and cost efficiency
- Troubleshoot and implement advanced network architectures including VPNs, firewalls, load balancers, and routing protocols
- Implement and enforce security best practices: IAM, encryption, compliance, and vulnerability management
- Collaborate with development and operations teams to improve CI/CD workflows and system observability
Required Skills & Qualifications
- 6+ years of experience in DevOps, Infrastructure as Code (IaC), and cloud-native systems
- Expertise in Helm, Terraform, and Kubernetes
- Strong hands-on experience with AWS and Azure
- Solid understanding of networking, firewall configurations, and security protocols
- Experience with CI/CD tools like Jenkins, GitHub Actions, or similar
- Strong problem-solving skills and a performance-first mindset
Why Join Us?
- Work on cutting-edge cloud infrastructure across diverse industries
- Be part of a collaborative, forward-thinking team
- Flexible hybrid work model – work from anywhere while staying connected
- Opportunity to take ownership and lead critical DevOps initiatives

Key Responsibilities
- Design, develop, and maintain automated test scripts using Python, pytest, and Selenium for Salesforce and web applications.
- Create and manage test environments using Docker to ensure consistent testing conditions.
- Collaborate with developers, business analysts, and stakeholders to understand requirements and define test scenarios.
- Execute automated and manual tests, analyze results, and report defects using GitLab or other tracking tools.
- Perform regression, functional, and integration testing for Salesforce applications and customizations.
- Ensure test coverage for Salesforce features, including custom objects, workflows, and Apex code.
- Contribute to continuous integration/continuous deployment (CI/CD) pipelines in GitLab for automated testing.
- Document test cases, processes, and results to maintain a comprehensive testing repository.
- Stay updated on Salesforce updates, testing tools, and industry best practices.
Required Qualifications
- 1-3 years of experience in automation testing, preferably with exposure to Salesforce applications.
- Proficiency in Python, pytest, Selenium, Docker, and GitLab for test automation and version control.
- Understanding of software testing methodologies, including functional, regression, and integration testing.
- Bachelor’s degree in Computer Science, Information Technology, or a related field.
- Strong problem-solving skills and attention to detail.
- Excellent verbal and written communication skills.
- Ability to work in a collaborative, fast-paced team environment.
Preferred Qualifications
- Experience with Salesforce platform testing, including Sales Cloud, Service Cloud, or Marketing Cloud.
- Active Salesforce Trailhead profile with demonstrated learning progress (please include Trailhead profile link in application).
- Salesforce certifications (e.g., Salesforce Administrator or Platform Developer) are a plus.
- Familiarity with testing Apex code, Lightning components, or Salesforce integrations.
- Experience with Agile/Scrum methodologies.
- Knowledge of Webkul’s product ecosystem or e-commerce platforms is an advantage.
Job Title: Backend Engineer – SDE II
Location: Delhi
Employment Type: Full-Time
About the Role:
We are looking for a passionate and experienced Backend Engineer (SDE II) to join our growing team. As a core backend contributor, you'll be responsible for building scalable, secure, and high-performance backend systems. You’ll collaborate closely with product, DevOps, and frontend teams to deliver best-in-class technology solutions.
Key Responsibilities:
- Design, develop, and maintain robust backend systems using Golang and PostgreSQL.
- Build and manage scalable microservices and APIs for high availability and performance.
- Ensure secure authentication and authorization using AWS services.
- Work with Docker, Kubernetes (EKS) to build and deploy containerized applications.
- Implement messaging and real-time systems using Kafka and WebSocket's.
- Maintain best practices in performance tuning, monitoring, and CI/CD.
- Collaborate with cross-functional teams to gather requirements and translate into technical solutions.
- Write clean, maintainable, and well-tested code.
- Contribute to architectural decisions and code reviews.
Required Skills & Qualifications:
- Bachelor's degree in Computer Science, Engineering, or related field.
- Proficiency in Golang and strong experience with PostgreSQL.
- Solid understanding of AWS Cloud, EKS, and Kubernetes.
- Hands-on experience with Docker and container orchestration tools.
- Strong experience in building and maintaining RESTful APIs and real-time systems using Kafka/WebSocket.
- Expertise in building and scaling microservices-based architecture.
- Good understanding of Go concurrency and parallelism.
What We Offer:
- Opportunity to work in a fast-paced, innovative environment.
- Flexible work culture and flat team structure.
- Competitive salary and benefits.
- Work with cutting-edge technologies and a highly talented team.
Responsibilities:
- Design, develop, and maintain scalable applications using Java/Kotlin Core Concepts and Spring Boot MVC.
- Build and optimize REST APIs for seamless client-server communication.
- Develop and ensure efficient HTTP/HTTPS request-response mechanisms.
- Handle Java/Kotlin version upgrades confidently, ensuring code compatibility and leveraging the latest features.
- Solve complex business logic challenges with a methodical and innovative approach.
- Write optimized SQL queries with Postgres DB.
- Ensure code quality through adherence to design patterns (e.g., Singleton, Factory, Observer, MVC) and unit testing frameworks like JUnit.
- Integrate third-party APIs and develop large-scale systems with technical precision.
- Debug and troubleshoot production issues.
Requirements:
- 2 to 4 years of hands-on experience in Java/Kotlin Spring Boot development.
- Proven expertise in handling version upgrades for Java and Kotlin with confidence.
- Strong logical thinking and problem-solving skills, especially in implementing complex algorithms.
- Proficiency with Git, JIRA, and managing software package versions.
- Familiarity with SaaS-based products, XML parsing/generation, and generating PDFs, XLS, CSVs using Spring Boot.
- Strong understanding of JPA, Hibernate, and core Java concepts (OOP).
Skills (Good to Have):
- Exposure to Docker, Redis, and Elasticsearch.
- Knowledge of transaction management and solving computational problems.
- Eagerness to explore new technologies.
About Eazeebox
Eazeebox is India’s first B2B Quick Commerce platform for home electrical goods. We empower electrical retailers with access to 100+ brands, flexible credit options, and 4-hour delivery—making supply chains faster, smarter, and more efficient. Our tech-driven approach enables sub-3 hour inventory-aware fulfilment across micro-markets, with a goal of scaling to 50+ orders/day per store.
About the Role
We’re looking for a DevOps Engineer to help scale and stabilize the cloud-native backbone that powers Eazeebox. You’ll play a critical role in ensuring our microservices architecture remains reliable, responsive, and performant—especially during peak retailer ordering windows. This is a high-ownership role for an "all-rounder" who is passionate about
designing scalable architectures, writing robust code, and ensuring seamless deployments and operations.
What You'll Be Doing
As a critical member of our small, dedicated team, you will take on a versatile role encompassing development, infrastructure, and operations.
Cloud & DevOps Ownership
- Architect and implement containerized services on AWS (S3, EC2, ECS, ECR, CodeBuild, Lambda, Fargate, RDS, CloudWatch) under secure IAM policies.
- Take ownership of CI/CD pipelines, optimizing and managing GitHub Actions workflows.
- Configure and manage microservice versioning and CI/CD deployments.
- Implement secrets rotation and IP-based request rate limiting for enhanced security.
- Configure auto-scaling instances and Kubernetes for high-workload microservices to ensure performance and cost efficiency.
- Hands-on experience with Docker and Kubernetes/EKS fundamentals.
Backend & API Design
- Design, build, and maintain scalable REST/OpenAPI services in Django (DRF), WebSocket implementations, and asynchronous microservices in FastAPI.
- Model relational data in PostgreSQL 17 and optimize with Redis for caching and pub/sub.
- Orchestrate background tasks using Celery or RQ with Redis Streams or Amazon SQS.
- Collaborate closely with the frontend team (React/React Native) to define and build robust APIs.
Testing & Observability
- Ensure code quality via comprehensive testing using Pytest, React Testing Library, and Playwright.
- Instrument applications with CloudWatch metrics, contributing to our observability strategy.
- Maintain a Git-centric development workflow, including branching strategies and pull-request discipline.
Qualifications & Skills
Must-Have
- Experience: 2-4 years of hands-on experience delivering production-level full-stack applications with a strong emphasis on backend and DevOps.
- Backend Expertise: Proficiency in Python, with strong command of Django or FastAPI, including async Python patterns and REST best practices.
- Database Skills: Strong SQL skills with PostgreSQL; practical experience with Redis for caching and messaging.
- Cloud & DevOps Mastery: Hands-on experience with Docker and Kubernetes/EKS fundamentals.
- AWS Proficiency: Experience deploying and managing services on AWS (EC2, S3, RDS, Lambda, ECS Fargate, ECR, SQS).
- CI/CD: Deep experience with GitHub Actions or similar platforms, including semantic-release, Blue-Green Deployments, and artifact signing.
- Automation: Fluency in Python/Bash or Go for automation scripts; comfort with YAML.
- Ownership Mindset: Entrepreneurial spirit, strong sense of ownership, and ability to deliver at scale.
- Communication: Excellent written and verbal communication skills; comfortable in async and distributed team environments.
Nice-to-Have
- Frontend Familiarity: Exposure to React with Redux Toolkit and React Native.
- Event Streaming: Experience with Kafka or Amazon EventBridge.
- Serverless: Knowledge of AWS Lambda, Step Functions, CloudFront Functions, or Cloudflare Workers.
- Observability: Familiarity with Datadog, Posthog, Prometheus/Grafana/Loki.
- Emerging Tech: Interest in GraphQL (Apollo Federation) or generative AI frameworks (Amazon Bedrock, LangChain) and AI/ML.
Key Responsibilities
- Architectural Leadership: Design and lead the technical strategy for migrating our platform from a monolithic to a microservices architecture.
- System Design: Translate product requirements into scalable, secure, and reliable system designs.
- Backend Development: Build and maintain core backend services using Python (Django/FastAPI).
- CI/CD & Deployment: Own and manage CI/CD pipelines for multiple services using GitHub Actions, AWS CodeBuild, and automated deployments.
- Infrastructure & Operations: Deploy production-grade microservices using Docker, Kubernetes, and AWS EKS.
- FinOps & Performance: Drive cloud cost optimization and implement auto-scaling for performance and cost-efficiency.
- Security & Observability: Implement security, monitoring, and compliance using tools like Prometheus, Grafana, Datadog, Posthog, and Loki to ensure 99.99% uptime.
- Collaboration: Work with product and development teams to align technical strategy with business growth plans.
We are looking for a proactive and detail-oriented AWS Project Manager / AWS Administrator to join our team. The ideal candidate will be responsible for configuring, monitoring, and managing AWS cloud services while optimizing usage, ensuring compliance, and driving automation across deployments.
Key Responsibilities:
1) Configure and manage AWS Cloud services including VPCs, URL proxies, Bastion Hosts, and C2S access points.
2) Monitor AWS resources and ensure high availability and performance.
3) Automate deployment processes to reduce manual efforts and enhance operational efficiency.
4) Daily cost monitoring and backup setup for EC2 source code to S3.
5) Manage AWS Elastic Beanstalk (EB), RDS, CloudFront, Load Balancer, and NAT Gateway traffic.
6) Implement security and compliance best practices using AWS Secret Manager, IAM, and organizational policies.
7) Perform system setup for CBT (Computer-Based Test) exams.
8) Utilize AWS Pricing Calculator for resource estimation and budgeting.
9) Implement RI (Reserved Instance) and Spot Instance strategies for cost savings.
10) Create and manage S3 bucket policies with controlled access, especially for specific websites.
11) Automate disaster recovery and backup strategies for RDS and critical infrastructure.
12) Configure and manage AWS environments using EB, RDS, and routing setups.
13) Create and maintain AWS CloudFormation templates for infrastructure as code.
14) Develop comprehensive documentation for all AWS-related activities and configurations.
15) Ensure adherence to AWS best practices in all deployments and services.
16) Handle account creation, policy management, and SSO configuration for organizational identity management.
17) Ability to configure and design AWS architecture.
Required Skills:
· Strong understanding of AWS services and architecture.
· Hands-on experience with:
1) Elastic Beanstalk (EB)
2) RDS
3) EC2
4) S3
5) CloudFront
6) Load Balancer & NAT Gateway
7) CloudFormation
8) IAM, SSO & Identity Management
· Proficient in AWS cost optimization techniques.
· Ability to automate regular tasks and create efficient workflows.
· Clear and structured documentation skills.
Preferred Qualifications:
AWS Certified Solutions Architect (required).
Must have 5+ years of experience in this field.
Previous experience in managed AWS services.

Full Stack Developer (Node.js & React)
Location: Pune, India (Local or Ready to Relocate)
Employment Type: 6–8 Month Contract (Potential Conversion to FTE Based on Performance)
About the Role
We are seeking a highly skilled Full Stack Developer with expertise in Node.js and React to join our dynamic team in Pune. This role involves designing, developing, and deploying scalable web applications. You will collaborate with cross-functional teams to deliver high-impact solutions while adhering to best practices in coding, testing, and security.
Key Responsibilities
- Develop and maintain server-side applications using Node.js (Express/NestJS) and client-side interfaces with React.js (Redux/Hooks).
- Architect RESTful APIs and integrate with databases (SQL/NoSQL) and third-party services.
- Implement responsive UI/UX designs with modern front-end libraries (e.g., Material-UI, Tailwind CSS).
- Write unit/integration tests (Jest, Mocha, React Testing Library) and ensure code quality via CI/CD pipelines.
- Collaborate with product managers, designers, and QA engineers in an Agile environment.
- Troubleshoot performance bottlenecks and optimize applications for scalability.
- Document technical specifications and deployment processes.
Required Skills & Qualifications
- Experience: 3+ years in full-stack development with Node.js and React.
- Backend Proficiency:
- Strong knowledge of Node.js, Express, or NestJS.
- Experience with databases (PostgreSQL, MongoDB, Redis).
- API design (REST/GraphQL) and authentication (JWT/OAuth).
- Frontend Proficiency:
- Expertise in React.js (Functional Components, Hooks, Context API).
- State management (Redux, Zustand) and modern CSS frameworks.
- DevOps & Tools:
- Git, Docker, AWS/Azure, and CI/CD tools (Jenkins/GitHub Actions).
- Testing frameworks (Jest, Cypress, Mocha).
- Soft Skills:
- Problem-solving mindset and ability to work in a fast-paced environment.
- Excellent communication and collaboration skills.
- Location: Based in Pune or willing to relocate immediately.
Preferred Qualifications
- Experience with TypeScript, Next.js, or serverless architectures.
- Knowledge of microservices, message brokers (Kafka/RabbitMQ), or container orchestration (Kubernetes).
- Familiarity with Agile/Scrum methodologies.
- Contributions to open-source projects or a strong GitHub portfolio.
What We Offer
- Competitive Contract Compensation with timely payouts.
- Potential for FTE Conversion: Performance-based path to a full-time role.
- Hybrid Work Model: Flexible in-office (Pune) and remote options.
- Learning Opportunities: Access to cutting-edge tools and mentorship.
- Collaborative Environment: Work with industry experts on innovative projects.
Apply Now!
Ready to make an impact? Send your resume and GitHub/Portfolio links with the subject line:
"Full Stack Developer (Node/React) - Pune".
Local candidates or those relocating to Pune will be prioritized. Applications without portfolios will not be considered.
Equal Opportunity Employer
We celebrate diversity and are committed to creating an inclusive environment for all employees.
AccioJob is conducting a Walk-In Hiring Drive with B2B Automation Platform for the position of DevOps Engineer.
To apply, register and select your slot here: https://go.acciojob.com/aCZCSM
Required Skills: CI/CD Workflows, Docker, Jenkins, Cloud (AWS preferred)
Eligibility:
- Degree: BTech./BE, MCA
- Branch: Computer Science/CSE/Other CS related branch, IT, Electrical/Other electrical related branches, Communications
- Graduation Year: 2023, 2024, 2025
Work Details:
- Work Location: Noida (Onsite)
- CTC: ₹4 LPA to ₹5 LPA
Evaluation Process:
Round 1: Offline Assessment at AccioJob Noida Centre
Further Rounds (for shortlisted candidates only):
Profile & Background Screening Round, Technical Interviews Round 1, Technical Interviews Round 2, Tech + Managerial Round
Important Note: Bring your laptop & earphones for the test.
Register here: https://go.acciojob.com/aCZCSM
Or skip the hassle and apply instantly via our app
https://go.acciojob.com/PDvyDp
Create and manage Jenkins Pipelines using Linux groovy
scripting and python
Analyze and fix issues in Jenkins, GitHub, Nexus,
SonarQube and AWS cloud
Perform Jenkins, GitHub, SonarQube and Nexus
administration.
Create resources in AWS environment using
infrastructure-as-code. Analyze and fix issues in AWS
Cloud.
Good-to-Have
AWS Cloud certification
Terraform Certification
- Kubernetes/Docker experience

*Must speak fluent english*
Video Pre-Screen Interview Required... After submitting please also visit here to apply & start the initial video interview:
Job description
We believe cloud based technology has changed how we work. Our commercial lending SaaS platform empowers our client base of brokers and lenders to create a seamless borrower experience, manage a network of brokers, & scale their entire lending organization in the cloud powered by LendingWise
Company Description
Do you want to work for a company powering the digital lending transformation across numerous lending verticals? We work to make a lending platform that is robust, customizable, & turnkey for numerous business purpose lending products like residential/commercial real estate lending, business funding, SBA, Equipment, MCA and more. The lending suite of products includes a CRM, LOS, Marketplace, and Servicing capabilities, ALL-IN-ONE platform. It’s the holy grail for every private lender/commercial lender/bank/Credit Union or broker of any size. We launched in early 2018 and have seen explosive growth nationwide.
Position Summary
LendingWise is looking for a Full Stack Developer/Architect who has the ability to define, develop and manage the existing architecture strategy for the LendingWise platform & its inter-connected products related to CRM, LOS, Servicing & Marketplace of lenders. You will be responsible for introducing new applications, services, and technologies into the LendingWise technical landscape and you will take responsibility for the initiatives you designed. You will need expert level knowledge of a variety of business domains including, but not limited to, CRM, Loan Origination, Loan Servicing, Payment integration, & API connectivity for various integrations for mortgage vendors (Plaid, Corelogic, DocMagic, Docusign, etc… & CRMs like hubspot, salesforce, etc… You need to manage input from our product owner/manager & CEO and assume overall responsibility for the execution of multiple, large-scale, mission critical initiatives. You will be accountable to a variety of primary business stakeholders for the quality and success of programs assigned to you. An acute sense of attention to detail is necessary. This role requires working with teams within and outside the USA. Flexibility in working hours will be required.
Key Responsibilities
Help plan, architect, & review various phases, features, bugs, & enhancements of the CRM, Loan Origination, & Loan Servicing platform & API connectivity hub
Introduce new tools and fresh insights to assist with development goals
Work on new features, enhancements, bugs & refactor code to MVC structure within the LendingWise platform
Requirements
Deep knowledge of Core PHP & other PHP frameworks, including strong experience with MVC patterns
JavaScript, HTML, Bootstrap
MySQL & Stored Procedures
Excellent skills in object-oriented programming and MVC design patterns
Comfortable working with SSH command line & GitHub version control, & PHP Storm
Solid knowledge of RESTful API endpoints
Multi-tenant CRM, ERP, Real Estate, Loan or FinTech Experience required
Compensation:
Competitive Salary & Bonuses
2 weeks paid time off annually
Flexible work schedule
Job Type: Full-time
Location:
Remote
Job Role: Senior Cypress Testing Engineer
Location: Remote
Type: Full-time
Experience: 4+ years
Salary: Competitive, based on experience
Video Pre-Screen Interview Required... After submitting please also visit here to apply & start the initial video interview: https://app.xinterview.ai/direct_invite/1fedac2d-4cb7-4c33-a2b4-837c17c66fda
About Us
LendingWise is a fast-growing B2B SaaS platform serving the private lending industry with a robust CRM & LOS solution. Built on a custom PHP/MySQL stack with modern front-end technologies (HTML, JS, Bootstrap), we power 250+ customers across the U.S. and manage over $2.3M in ARR. We're scaling our engineering operations and looking for top-tier QA talent to join our distributed team.
What You'll Do
We’re looking for a Cypress Test Engineer to help us increase coverage of our end-to-end test suite across our core CRM & LOS platform. You will collaborate closely with developers, DevOps, and product managers to write automated tests, streamline quality checks, and contribute to our CI/CD workflows.
Key Responsibilities
Design, write, and maintain Cypress end-to-end test suites for core platform features.
Work with PHP developers to integrate testing into GitHub Actions-based CI/CD pipelines.
Assist with configuration and optimization of test environments using Docker containers.
Collaborate with DevOps and developers to integrate test stages into deepsource, PHPUnit, and other code quality tools.
Write test cases that cover user workflows across HTML/Bootstrap/JS interfaces.
Support debugging and analysis of failed tests, helping the team proactively identify regressions.
Recommend improvements to testing strategy and frameworks as the platform evolves.
Requirements
4+ years experience in automated testing with Cypress.
Strong understanding of end-to-end testing principles and frontend behavior validation.
Comfortable working with PHP, MySQL, and legacy codebases.
Familiarity with CI/CD tools like GitHub Actions, Docker, and PHPUnit.
Experience with deepsource or similar code quality and linting tools is a plus.
Strong ENGLISH communication skills and the ability to work independently in a remote environment.
Experience with web technologies such as JavaScript, HTML, and Bootstrap.
Nice to Have
Prior experience working in CRM or fintech product team.
Exposure to testing multi-tenant or enterprise applications.
Why Join LendingWise?
Work on a mission-critical platform used daily by lenders, brokers, and banks.
Fully remote team with flexible working hours.
Opportunity to make a real impact on quality and release reliability.
Collaborative, agile culture with modern tools and engineering practices.

We are looking for a Senior Software Engineer to join our team and contribute to key business functions. The ideal candidate will bring relevant experience, strong problem-solving skills, and a collaborative
mindset.
Responsibilities:
- Design, build, and maintain high-performance systems using modern C++
- Architect and implement containerized services using Docker, with orchestration via Kubernetes or ECS
- Build, monitor, and maintain data ingestion, transformation, and enrichment pipelines
- Deep understanding of cloud platforms (preferably AWS) and hands-on experience in deploying and
- managing applications in the cloud.
- Implement and maintain modern CI/CD pipelines, ensuring seamless integration, testing, and delivery
- Participate in system design, peer code reviews, and performance tuning
Qualifications:
- 5+ years of software development experience, with strong command over modern C++
- Deep understanding of cloud platforms (preferably AWS) and hands-on experience in deploying and managing applications in the cloud.
- Apache Airflow for orchestrating complex data workflows.
- EKS (Elastic Kubernetes Service) for managing containerized workloads.
- Proven expertise in designing and managing robust data pipelines & Microservices.
- Proficient in building and scaling data processing workflows and working with structured/unstructured data
- Strong hands-on experience with Docker, container orchestration, and microservices architecture
- Working knowledge of CI/CD practices, Git, and build/release tools
- Strong problem-solving, debugging, and cross-functional collaboration skills
This position description is intended to describe the duties most frequently performed by an individual in this position. It is not intended to be a complete list of assigned duties but to describe a position level.
We’re looking for an Engineering Manager to guide our micro-service platform and mentor a fully remote backend team. You’ll blend hands-on technical ownership with people leadership—shaping architecture, driving cloud best practices, and coaching engineers in their careers and craft.
Key Responsibilities:
Area
What You’ll Own
Architecture & Delivery
• Define and evolve backend architecture built on Java 17+, Spring Boot 3, AWS (Containers, Lambdas, SQS, S3), Elasticsearch, PostgreSQL/MySQL, Databricks, Redis etc...
• Lead design and code reviews; enforce best practices for testing, CI/CD, observability, security, and cost-efficient cloud operations.
• Drive technical roadmaps, ensuring scalability (billions of events, 99.9 %+ uptime) and rapid feature delivery.
Team Leadership & Growth
• Manage and inspire a distributed team of 6-10 backend engineers across multiple time zones.
• Set clear growth objectives, run 1-on-1s, deliver feedback, and foster an inclusive, high-trust culture.
• Coach the team on AI-assisted development workflows (e.g., GitHub Copilot, LLM-based code review) to boost productivity and code quality.
Stakeholder Collaboration
• Act as technical liaison to Product, Frontend, SRE, and Data teams, translating business goals into resilient backend solutions.
• Communicate complex concepts to both technical and non-technical audiences; influence cross-functional decisions.
Technical Vision & Governance
• Own coding standards, architectural principles, and technology selection.
• Evaluate emerging tools and frameworks (especially around GenAI and cloud-native patterns) and create adoption strategies.
• Balance technical debt and new feature delivery through data-driven prioritization.
Required Qualifications:
● 8+ years designing, building, and operating distributed backend systems with Java & Spring Boot
● Proven experience leading or mentoring engineers; direct people-management a plus
● Expert knowledge of AWS services and cloud-native design patterns
● Hands-on mastery of Elasticsearch, PostgreSQL/MySQL, and Redis for high-volume, low-latency workloads
● Demonstrated success scaling systems to millions of users or billions of events Strong grasp of DevOps practices: containerization (Docker), CI/CD (GitHub Actions), observability stacks
● Excellent communication and stakeholder-management skills in a remote-fi rst environment
Nice-to-Have:
● Hands-on experience with Datadog (APM, Logs, RUM) and a data-driven approach to debugging/performance tuning
● Startup experience—comfortable wearing multiple hats and juggling several projects simultaneously
● Prior title of Principal Engineer, Staff Engineer, or Engineering Manager in a high-growth SaaS company
● Familiarity with AI-assisted development tools (Copilot, CodeWhisperer, Cursor) and a track record of introducing them safely
Job Title: Engineering Manager (Java / Spring Boot, AWS) – Remote
Leadership Role
Location: Remote
Employment Type: Full-time


Build a dynamic solution empowering companies to optimize promotional activities for maximum impact. It collects and validates data, analyzes promotion effectiveness, plans calendars, and integrates seamlessly with existing systems. The tool enhances vendor collaboration, negotiates better deals, and employs machine learning to optimize promotional plans, enabling companies to make informed decisions and maximize return on investment.
Technology Stack: Scala, Go-Lang, Docker, Kubernetes, Databricks, Python optional.
Working Time Zone: EU
Specialty: Data Science
Level of the Candidate: more then 5 years of experience
Language Proficiency: English Upper-Intermediate
Required Soft Skills:
- Prefer problem solving style over experience
- Ability to clarify requirements with the customer
- Willingness to pair with other engineers when solving complex issues
- Good communication skills
Hard Skills / Need to Have:
- Experience in Scala and/or Go, designing and building scalable high-performing applications
- Experience in containerization and microservices orchestration using Docker and Kubernetes
- Experience in building data pipelines and ETL solutions using Databricks
- Experience in data storage and retrieval with PostgreSQL and Elasticsearch
- Experience in deploying and maintaining solutions in the Azure cloud environment
- Experience in Python is nice to have
Responsibilities and Tasks:
- Develop and maintain distributed systems using Scala and/or Go
- Work with Docker and Kubernetes for containerization and microservices orchestration
- Build data pipelines and ETL solutions using Databricks
- Work with PostgreSQL and Elasticsearch for data storage and retrieval
- Deploy and maintain solutions in the Azure cloud environment
Position Overview: We are looking for an experienced and highly skilled Senior Data Engineer to join our team and help design, implement, and optimize data systems that support high-end analytical solutions for our clients. As a customer-centric Data Engineer, you will work closely with clients to understand their business needs and translate them into robust, scalable, and efficient technical solutions. You will be responsible for end-to-end data modelling, integration workflows, and data transformation processes while ensuring security, privacy, and compliance.In this role, you will also leverage the latest advancements in artificial intelligence, machine learning, and large language models (LLMs) to deliver high-impact solutions that drive business success. The ideal candidate will have a deep understanding of data infrastructure, optimization techniques, and cost-effective data management
Key Responsibilities:
• Customer Collaboration:
– Partner with clients to gather and understand their business
requirements, translating them into actionable technical specifications.
– Act as the primary technical consultant to guide clients through data challenges and deliver tailored solutions that drive value.
•Data Modeling & Integration:
– Design and implement scalable, efficient, and optimized data models to support business operations and analytical needs.
– Develop and maintain data integration workflows to seamlessly extract, transform, and load (ETL) data from various sources into data repositories.
– Ensure smooth integration between multiple data sources and platforms, including cloud and on-premise systems
• Data Processing & Optimization:
– Develop, optimize, and manage data processing pipelines to enable real-time and batch data processing at scale.
– Continuously evaluate and improve data processing performance, optimizing for throughput while minimizing infrastructure costs.
• Data Governance & Security:
–Implement and enforce data governance policies and best practices, ensuring data security, privacy, and compliance with relevant industry regulations (e.g., GDPR, HIPAA).
–Collaborate with security teams to safeguard sensitive data and maintain privacy controls across data environments.
• Cross-Functional Collaboration:
– Work closely with data engineers, data scientists, and business
analysts to ensure that the data architecture aligns with organizational objectives and delivers actionable insights.
– Foster collaboration across teams to streamline data workflows and optimize solution delivery.
• Leveraging Advanced Technologies:
– Utilize AI, machine learning models, and large language models (LLMs) to automate processes, accelerate delivery, and provide
smart, data-driven solutions to business challenges.
– Identify opportunities to apply cutting-edge technologies to improve the efficiency, speed, and quality of data processing and analytics.
• Cost Optimization:
–Proactively manage infrastructure and cloud resources to optimize throughput while minimizing operational costs.
–Make data-driven recommendations to reduce infrastructure overhead and increase efficiency without sacrificing performance.
Qualifications:
• Experience:
– Proven experience (5+ years) as a Data Engineer or similar role, designing and implementing data solutions at scale.
– Strong expertise in data modelling, data integration (ETL), and data transformation processes.
– Experience with cloud platforms (AWS, Azure, Google Cloud) and big data technologies (e.g., Hadoop, Spark).
• Technical Skills:
– Advanced proficiency in SQL, data modelling tools (e.g., Erwin,PowerDesigner), and data integration frameworks (e.g., Apache
NiFi, Talend).
– Strong understanding of data security protocols, privacy regulations, and compliance requirements.
– Experience with data storage solutions (e.g., data lakes, data warehouses, NoSQL, relational databases).
• AI & Machine Learning Exposure:
– Familiarity with leveraging AI and machine learning technologies (e.g., TensorFlow, PyTorch, scikit-learn) to optimize data processing and analytical tasks.
–Ability to apply advanced algorithms and automation techniques to improve business processes.
• Soft Skills:
– Excellent communication skills to collaborate with clients, stakeholders, and cross-functional teams.
– Strong problem-solving ability with a customer-centric approach to solution design.
– Ability to translate complex technical concepts into clear, understandable terms for non-technical audiences.
• Education:
– Bachelor’s or Master’s degree in Computer Science, Information Systems, Data Science, or a related field (or equivalent practical experience).
LIFE AT FOUNTANE:
- Fountane offers an environment where all members are supported, challenged, recognized & given opportunities to grow to their fullest potential.
- Competitive pay
- Health insurance for spouses, kids, and parents.
- PF/ESI or equivalent
- Individual/team bonuses
- Employee stock ownership plan
- Fun/challenging variety of projects/industries
- Flexible workplace policy - remote/physical
- Flat organization - no micromanagement
- Individual contribution - set your deadlines
- Above all - culture that helps you grow exponentially!
A LITTLE BIT ABOUT THE COMPANY:
Established in 2017, Fountane Inc is a Ventures Lab incubating and investing in new competitive technology businesses from scratch. Thus far, we’ve created half a dozen multi-million valuation companies in the US and a handful of sister ventures for large corporations, including Target, US Ventures, and Imprint Engine.
We’re a team of 120+ strong from around the world that are radically open-minded and believes in excellence, respecting one another, and pushing our boundaries to the furthest it's ever been.

About Astra:
Astra is a cyber security SaaS company that makes otherwise chaotic pentests a breeze with its one of a kind Pentest Platform. Astra's continuous vulnerability scanner emulates hacker behavior to scan applications for 9300+ security tests. CTOs & CISOs love Astra because it helps them fix vulnerabilities in record time and move from DevOps to DevSecOps with Astra's CI/CD integrations.
Astra is loved by 650+ companies across the globe. In 2023 Astra uncovered 2 million+ vulnerabilities for its customers, saving customers $69M+ in potential losses due to security vulnerabilities.
We've been awarded by the President of France Mr. François Hollande at the La French Tech program and Prime Minister of India Shri Narendra Modi at the Global Conference on Cyber Security. Loom, MamaEarth, Muthoot Finance, Canara Robeco, ScripBox etc. are a few of Astra’s customers.
Role Overview:
As an SDE 2 back-end engineer you will be responsible for leading the development and maintenance of the platform/dashboard backend. This would involve architecting, developing & maintaining RESTful services for vulnerability management, scan orchestration, inventory management and platform features such as on-boarding, trust centers, certificates, payment integrations, vulnerability ingestion, authentication etc.
You should have a strong background in backend programming (Symfony preferred, or Laravel) and have experience with an event driven, async & distributed architecture.
During the first 6 months of your role, you will be involved in the 0->1 journey of two of our upcoming products in our platform along with the respective product owners.
Love solving hard problems? Want to build high impact products rooted in first principles? Is coding your poetry? Join us in shaping the future of cyber security.
Roles & Responsibilities:
- Lead the design, development, and maintenance of backend services and APIs using Symfony PHP framework. What you create is also what you own.
- Collaborate with front-end developers to integrate user-facing elements with server-side logic.
- Collaborate with scanner teams to orchestrate scans, ingest vulnerabilities, configure scanners etc.
- Write clean, well-documented, and efficient code.
- Optimize and refactor existing code to improve performance and reliability.
- Implement security and data protection measures.
- Triage, troubleshoot and upgrade existing systems.
- Ship code to production multiple times a day/week.
- Ensure timely delivery of the features.
- Test your own features, and write test cases for continuous automated testing.
- Troubleshoot complex issues, debug, and upgrade existing systems.
- Mentor junior engineers and lead code reviews, promoting best practices and high standards.
- Drive architectural decisions and technical strategy for backend systems.
What we are looking for:
- 4-6 years experience in a similar role or similar working experience.
- Advanced proficiency in PHP backend programming. Symfony preferred, or Laravel.
- Strong understanding of system design, software architecture principles and performance optimization.
- Experience with an event driven (subscribers/listeners), async (MessageHandlers & job queues) & distributed (load-balanced) architecture.
- Experience with PostgreSQL database and data modeling.
- Understanding of concepts such as Dependency Injection, ORM, data validation, error handling etc.
- Experience with Docker, Kubernetes, GitHub Actions (good to have).
- Experience with Agile methodologies.
- Excellent problem-solving skills and the ability to think strategically about technical solutions.
- Strong communication and interpersonal skills, with the ability to collaborate effectively with cross-functional remote teams.
- Demonstrated track record of delivering high-quality software products on schedule.
- Knowledge of industry best practices in software development, security, and compliance.
- Knowledge of application & infrastructure security - helps you stand out.
What We offer:
- Adrenalin rush of being a part of a fast-growing company, and working on hard problems that matter.
- Fully remote, agile working environment.
- Good engineering culture with full ownership in design, development, release lifecycle.
- A wholesome opportunity where you get to build things from scratch, improve and ship code to production in hours, not weeks.
- Holistic understanding of SaaS and enterprise security business.
- Experience with security side of things.
- Annual trips to beaches or mountains (next one is to Wayanad).
- Open and supportive culture.
- Health insurance & other benefits for you and your spouse. Maternity benefits included.

Role Overview:
As a Backend Developer at LearnTube.ai, you will ship the backbone that powers 2.3 million learners in 64 countries—owning APIs that crunch 1 billion learning events & the AI that supports it with <200 ms latency.
Skip the wait and get noticed faster by completing our AI-powered screening. Click this link to start your quick interview. It only takes a few minutes and could be your shortcut to landing the job! - https://bit.ly/LT_Python
What You'll Do:
At LearnTube, we’re pushing the boundaries of Generative AI to revolutionise how the world learns. As a Backend Engineer, you will be building the backend for an AI system and working directly on AI. Your roles and responsibilities will include:
- Ship Micro-services – Build FastAPI services that handle ≈ 800 req/s today and will triple within a year (sub-200 ms p95).
- Power Real-Time Learning – Drive the quiz-scoring & AI-tutor engines that crunch millions of events daily.
- Design for Scale & Safety – Model data (Postgres, Mongo, Redis, SQS) and craft modular, secure back-end components from scratch.
- Deploy Globally – Roll out Dockerised services behind NGINX on AWS (EC2, S3, SQS) and GCP (GKE) via Kubernetes.
- Automate Releases – GitLab CI/CD + blue-green / canary = multiple safe prod deploys each week.
- Own Reliability – Instrument with Prometheus / Grafana, chase 99.9 % uptime, trim infra spend.
- Expose Gen-AI at Scale – Publish LLM inference & vector-search endpoints in partnership with the AI team.
- Ship Fast, Learn Fast – Work with founders, PMs, and designers in weekly ship rooms; take a feature from Figma to prod in < 2 weeks.
What makes you a great fit?
Must-Haves:
- 2+ yrs Python back-end experience (FastAPI)
- Strong with Docker & container orchestration
- Hands-on with GitLab CI/CD, AWS (EC2, S3, SQS) or GCP (GKE / Compute) in production
- SQL/NoSQL (Postgres, MongoDB) + You’ve built systems from scratch & have solid system-design fundamentals
Nice-to-Haves
- k8s at scale, Terraform,
- Experience with AI/ML inference services (LLMs, vector DBs)
- Go / Rust for high-perf services
- Observability: Prometheus, Grafana, OpenTelemetry
About Us:
At LearnTube, we’re on a mission to make learning accessible, affordable, and engaging for millions of learners globally. Using Generative AI, we transform scattered internet content into dynamic, goal-driven courses with:
- AI-powered tutors that teach live, solve doubts in real time, and provide instant feedback.
- Seamless delivery through WhatsApp, mobile apps, and the web, with over 1.4 million learners across 64 countries.
Meet the Founders:
LearnTube was founded by Shronit Ladhani and Gargi Ruparelia, who bring deep expertise in product development and ed-tech innovation. Shronit, a TEDx speaker, is an advocate for disrupting traditional learning, while Gargi’s focus on scalable AI solutions drives our mission to build an AI-first company that empowers learners to achieve career outcomes. We’re proud to be recognised by Google as a Top 20 AI Startup and are part of their 2024 Startups Accelerator: AI First Program, giving us access to cutting-edge technology, credits, and mentorship from industry leaders.
Why Work With Us?
At LearnTube, we believe in creating a work environment that’s as transformative as the products we build. Here’s why this role is an incredible opportunity:
- Cutting-Edge Technology: You’ll work on state-of-the-art generative AI applications, leveraging the latest advancements in LLMs, multimodal AI, and real-time systems.
- Autonomy and Ownership: Experience unparalleled flexibility and independence in a role where you’ll own high-impact projects from ideation to deployment.
- Rapid Growth: Accelerate your career by working on impactful projects that pack three years of learning and growth into one.
- Founder and Advisor Access: Collaborate directly with founders and industry experts, including the CTO of Inflection AI, to build transformative solutions.
- Team Culture: Join a close-knit team of high-performing engineers and innovators, where every voice matters, and Monday morning meetings are something to look forward to.
- Mission-Driven Impact: Be part of a company that’s redefining education for millions of learners and making AI accessible to everyone.

A cloud tech firm offering secure, scalable hybrid storage s

Role: Python Developer (Immediate Joiner)
Location: Gurugram, India (Onsite)
Experience: 4+ years
Company: Mizzle Cloud Pvt Ltd
Working Days: 6 Days( 5 days- WFO, Sat- WFH)
Job Summary
We are seeking a skilled Python Django Developer with expertise in building robust, scalable, and efficient web applications. Must have 3+ years of core work experience. The ideal candidate will have hands-on experience with RabbitMQ, Redis, Celery, and PostgreSQL to ensure seamless background task management, caching, and database performance.
Key Responsibilities
- Develop, maintain, and enhance Django-based web applications and APIs.
- Design and implement message broker solutions using RabbitMQ to manage asynchronous communication.
- Integrate Redis for caching and session storage to optimize performance.
- Implement and manage task queues using Celery for background processes.
- Work with PostgreSQL for database design, optimization, and query tuning.
- Collaborate with front-end developers, DevOps engineers, and stakeholders to deliver high-quality software solutions.
- Write clean, modular, and well-documented code following best practices and standards.
- Debug, troubleshoot, and resolve issues across the application stack.
- Participate in code reviews, system design discussions, and team meetings.
- Ensure scalability, reliability, and security of applications.
Technical Skills:
- Must have minimum 4+ years of relevant work experience.
- Strong proficiency in Python and Django framework.
- Experience with message brokers, particularly RabbitMQ.
- Familiarity with Redis for caching and session management.
- Hands-on experience with Celery for distributed task queues.
- Proficiency in PostgreSQL, including database design and optimization.
- Knowledge of RESTful API design and development.
- Understanding of Docker and containerized applications.
Preferred Skills:
- Experience with CI/CD pipelines.
- Familiarity with cloud platforms (AWS, GCP).
- Knowledge of Django ORM and its optimizations.
- Basic understanding of front-end technologies (HTML, CSS, JavaScript).
Soft Skills:
- Strong problem-solving skills and attention to detail.
- Excellent communication and teamwork abilities.
- Ability to work in an agile environment and adapt to changing requirements.
Educational Requirements
- Bachelor’s degree in Computer Science, Software Engineering, or a related field.

Job Title: Full Stack Developer (MERN + Python) / Senior Mern Developer
Location: Bangalore
Job Type: Full-time
Experience: 4–8 years
About Miror
Miror is a pioneering FemTech company transforming how midlife women navigate perimenopause and menopause. We offer medically-backed solutions, expert consultations, community engagement, and wellness products to empower women in their health journey. Join us to make a meaningful difference through technology.
Role Overview
· We are seeking a passionate and experienced Full Stack Developer skilled in MERN stack and Python (Django/Flask) to build and scale high-impact features across our web and mobile platforms. You will collaborate with cross-functional teams to deliver seamless user experiences and robust backend systems.
Key Responsibilities
· Design, develop, and maintain scalable web applications using MySQL/Postgres, MongoDB, Express.js, React.js, and Node.js
· Build and manage RESTful APIs and microservices using Python (Django/Flask/FastAPI)
· Integrate with third-party platforms like OpenAI, WhatsApp APIs (Whapi), Interakt, and Zoho
· Optimize performance across the frontend and backend
· Collaborate with product managers, designers, and other developers to deliver high-quality features
· Ensure security, scalability, and maintainability of code
· Write clean, reusable, and well-documented code
· Contribute to DevOps, CI/CD, and server deployment workflows (AWS/Lightsail)
· Participate in code reviews and mentor junior developers if needed
Required Skills
· Strong experience with MERN Stack: MongoDB, Express.js, React.js, Node.js
· Proficiency in Python and web frameworks like Django, Flask, or FastAPI
· Experience working with REST APIs, JWT/Auth, and WebSockets
· Good understanding of frontend design systems, state management (Redux/Context), and responsive UI
· Familiarity with database design and queries (MongoDB, PostgreSQL/MySQL)
· Experience with Git, Docker, and deployment pipelines
· Comfortable working in Linux-based environments (e.g., Ubuntu on AWS)
Bonus Skills
· Experience with AI integrations (e.g., OpenAI, LangChain)
· Familiarity with WooCommerce, WordPress APIs
· Experience in chatbot development or WhatsApp API integration
Who You Are
· You are a problem-solver with a product-first mindset
· You care about user experience and performance
· You enjoy working in a fast-paced, collaborative environment
· You have a growth mindset and are open to learning new technologies
Why Join Us?
· Work at the intersection of healthcare, community, and technology
· Directly impact the lives of women across India and beyond
· Flexible work environment and collaborative team
· Opportunity to grow with a purpose-driven startup
·
In you are interested please apply here and drop me a message here in cutshort.

Job Title : Python Developer (Immediate Joiner)
Experience Required : 3+ Years
Employment Type : Full-time
Location : Gurugram, India (Onsite)
Working Days : 6 Days (5 Days WFO + 1 Day WFH)
Job Summary :
We are seeking a talented and experienced Python Developer with a strong background in Django and a proven ability to build scalable, secure, and high-performance web applications. The ideal candidate will have hands-on experience with RabbitMQ, Redis, Celery, and PostgreSQL, and will play a key role in developing and maintaining robust backend systems. This is an onsite opportunity for immediate joiners.
Mandatory Skills : Python, Django, RabbitMQ, Redis, Celery, PostgreSQL, RESTful APIs, Docker.
Key Responsibilities :
- Design, develop, and maintain Django-based web applications and APIs.
- Implement asynchronous task handling using RabbitMQ and Celery.
- Optimize application performance using Redis for caching and session storage.
- Manage database operations, including schema design and query optimization, using PostgreSQL.
- Collaborate with front-end developers, DevOps teams, and stakeholders to deliver full-featured solutions.
- Write clean, modular, and well-documented code aligned with industry best practices.
- Troubleshoot, debug, and resolve issues across the application stack.
- Participate in architecture discussions, code reviews, and sprint planning.
- Ensure the scalability, performance, and security of backend services.
Required Technical Skills :
- Minimum 3 Years of experience in Python development.
- Strong hands-on experience with the Django framework.
- Proficiency in RabbitMQ for message brokering.
- Practical experience with Redis for caching and session management.
- Experience using Celery for background job/task queue management.
- Solid knowledge of PostgreSQL (database design, indexing, and optimization).
- Understanding of RESTful API development and integration.
- Familiarity with Docker and containerization.
Preferred Skills :
- Exposure to CI/CD tools and pipelines.
- Experience working with cloud platforms such as AWS or GCP.
- Knowledge of Django ORM optimization techniques.
- Basic familiarity with front-end technologies like HTML, CSS, and JavaScript.
Soft Skills :
- Strong analytical and problem-solving capabilities.
- Effective communication and interpersonal skills.
- Ability to thrive in an agile, fast-paced development environment.


About Us
DAITA is a German AI startup revolutionizing the global textile supply chain by digitizing factory-to-brand workflows. We are building cutting-edge AI-powered SaaS and Agentic Systems that automate order management, production tracking, and compliance — making the supply chain smarter, faster, and more transparent.
Fresh off a $500K pre-seed raise, our passionate team is on the ground in India, collaborating directly with factories and brands to build our MVP and create real-world impact. If you’re excited by the intersection of AI, SaaS, and supply chain innovation, join us to help reshape how textiles move from factory floors to global brands.
Role Overview
We’re seeking a versatile Full-Stack Engineer to join our growing engineering team. You’ll be instrumental in designing and building scalable, secure, and high-performance applications that power our AI-driven platform. Working closely with Founders, ML Engineers, and Pilot Customers, you’ll transform complex AI workflows into intuitive, production-ready features.
What You’ll Do
• Design, develop, and deploy backend services, APIs, and microservices powering our platform.
• Build responsive, user-friendly frontend applications tailored for factory and brand users.
• Integrate AI/ML models and agentic workflows into seamless production environments.
• Develop features supporting order parsing, supply chain tracking, compliance, and reporting.
• Collaborate cross-functionally to iterate rapidly, test with users, and deliver impactful releases.
• Optimize applications for performance, scalability, and cost-efficiency on cloud platforms.
• Establish and improve CI/CD pipelines, deployment processes, and engineering best practices.
• Write clear documentation and maintain clean, maintainable code.
Required Skills
• 3–5 years of professional Full-Stack development experience
• Strong backend skills with frameworks like Node.js, Python (FastAPI, Django), Go, or similar
• Frontend experience with React, Vue.js, Next.js, or similar modern frameworks
• Solid knowledge and experience with relational databases (PostgreSQL, MySQL) and NoSQL databases (MongoDB, Redis, Neon)
• Strong API design skills (REST mandatory; GraphQL a plus)
• Containerization expertise with Docker
• Container orchestration and management with Kubernetes (including experience with Helm charts, operators, or custom resource definitions)
• Cloud deployment and infrastructure experience on AWS, GCP or Azure
• Hands-on experience deploying AI/ML models in cloud-native environments (AWS, GCP or Azure) with scalable infrastructure and monitoring.
• Experience with managed AI/ML services like AWS SageMaker, GCP Vertex AI, Azure ML, Together.ai, or similar
• Experience with CI/CD pipelines and DevOps tools such as Jenkins, GitHub Actions, Terraform, Ansible, or ArgoCD
• Familiarity with monitoring, logging, and observability tools like Prometheus, Grafana, ELK stack (Elasticsearch, Logstash, Kibana), or Helicone
Nice-to-have
• Experience with TypeScript for full-stack AI SaaS development
• Use of modern UI frameworks and tooling like Tailwind CSS
• Familiarity with modern AI-first SaaS concepts viz. vector databases for fast ML data retrieval, prompt engineering for LLM integration, integrating with OpenRouter or similar LLM orchestration frameworks etc.
• Knowledge of MLOps tools like Kubeflow, MLflow, or Seldon for model lifecycle management.
• Background in building data pipelines, real-time analytics, and predictive modeling.
• Knowledge of AI-driven security tools and best practices for SaaS compliance.
• Proficiency in cloud automation, cost optimization, and DevOps for AI workflows.
• Ability to design and implement hyper-personalized, adaptive user experiences.
What We Value
• Ownership: You take full responsibility for your work and ship high-quality solutions quickly.
• Bias for Action: You’re pragmatic, proactive, and focused on delivering results.
• Clear Communication: You articulate ideas, challenges, and solutions effectively across teams.
• Collaborative Spirit: You thrive in a cross-functional, distributed team environment.
• Customer Focus: You build with empathy for end users and real-world usability.
• Curiosity & Adaptability: You embrace learning, experimentation, and pivoting when needed.
• Quality Mindset: You write clean, maintainable, and well-tested code.
Why Join DAITA?
• Be part of a mission-driven startup transforming a $1+ Trillion global industry.
• Work closely with founders and AI experts on cutting-edge technology.
• Directly impact real-world supply chains and sustainability.
• Grow your skills in AI, SaaS, and supply chain tech in a fast-paced environment.
DevOps Engineer
AiSensy
Gurugram, Haryana, India (On-site)
About AiSensy
AiSensy is a WhatsApp based Marketing & Engagement platform helping businesses like Adani, Delhi Transport Corporation, Yakult, Godrej, Aditya Birla Hindalco., Wipro, Asian Paints, India Today Group Skullcandy, Vivo, Physicswallah, Cosco grow their revenues via WhatsApp.
- Enabling 100,000+ Businesses with WhatsApp Engagement & Marketing
- 400Crores + WhatsApp Messages done between Businesses and Users via AiSensy per year
- Working with top brands like Delhi Transport Corporation, Vivo, Physicswallah & more
- High Impact as Businesses drive 25-80% Revenues using AiSensy Platform
- Mission-Driven and Growth Stage Startup backed by Marsshot.vc, Bluelotus.vc & 50+ Angel Investors
Now, we’re looking for a DevOps Engineer to help scale our infrastructure and optimize performance for millions of users. 🚀
What You’ll Do (Key Responsibilities)
🔹 CI/CD & Automation:
- Implement, manage, and optimize CI/CD pipelines using AWS CodePipeline, GitHub Actions, or Jenkins.
- Automate deployment processes to improve efficiency and reduce downtime.
🔹 Infrastructure Management:
- Use Terraform, Ansible, Chef, Puppet, or Pulumi to manage infrastructure as code.
- Deploy and maintain Dockerized applications on Kubernetes clusters for scalability.
🔹 Cloud & Security:
- Work extensively with AWS (Preferred) or other cloud platforms to build and maintain cloud infrastructure.
- Optimize cloud costs and ensure security best practices are in place.
🔹 Monitoring & Troubleshooting:
- Set up and manage monitoring tools like CloudWatch, Prometheus, Datadog, New Relic, or Grafana to track system performance and uptime.
- Proactively identify and resolve infrastructure-related issues.
🔹 Scripting & Automation:
- Use Python or Bash scripting to automate repetitive DevOps tasks.
- Build internal tools for system health monitoring, logging, and debugging.
What We’re Looking For (Must-Have Skills)
✅ Version Control: Proficiency in Git (GitLab / GitHub / Bitbucket)
✅ CI/CD Tools: Hands-on experience with AWS CodePipeline, GitHub Actions, or Jenkins
✅ Infrastructure as Code: Strong knowledge of Terraform, Ansible, Chef, or Pulumi
✅ Containerization & Orchestration: Experience with Docker & Kubernetes
✅ Cloud Expertise: Hands-on experience with AWS (Preferred) or other cloud providers
✅ Monitoring & Alerting: Familiarity with CloudWatch, Prometheus, Datadog, or Grafana
✅ Scripting Knowledge: Python or Bash for automation
Bonus Skills (Good to Have, Not Mandatory)
➕ AWS Certifications: Solutions Architect, DevOps Engineer, Security, Networking
➕ Experience with Microsoft/Linux/F5 Technologies
➕ Hands-on knowledge of Database servers

Supply Wisdom: Full Stack Developer
Location: Hybrid Position based in Bangalore
Reporting to: Tech Lead Manager
Supply Wisdom is a global leader in transformative risk intelligence, offering real-time insights to drive business growth, reduce costs, enhance security and compliance, and identify revenue opportunities. Our AI-based SaaS products cover various risk domains, including financial, cyber, operational, ESG, and compliance. With a diverse workforce that is 57% female, our clients include Fortune 100 and Global 2000 firms in sectors like financial services, insurance, healthcare, and technology.
Objective: We are seeking a skilled Full Stack Developer to design and build scalable software solutions. You will be part of a cross-functional team responsible for the full software development life cycle, from conception to deployment.
As a Full Stack Developer, you should be proficient in both front-end and back-end technologies, development frameworks, and third-party libraries. We’re looking for a team player with strong problem-solving abilities, attention to visual design, and a focus on utility. Familiarity with Agile methodologies, including Scrum and Kanban, is essential.
Responsibilities
- Collaborate with the development team and product manager to ideate software solutions.
- Write effective and secure REST APIs.
- Integrate third-party libraries for product enhancement.
- Design and implement client-side and server-side architecture.
- Work with data scientists and analysts to enhance software using RPA and AI/ML techniques.
- Develop and manage well-functioning databases and applications.
- Ensure software responsiveness and efficiency through testing.
- Troubleshoot, debug, and upgrade software as needed.
- Implement security and data protection settings.
- Create features and applications with mobile-responsive design.
- Write clear, maintainable technical documentation.
- Build front-end applications with appealing, responsive visual design.
Requirements
- Degree in Computer Science (or related field) with 4+ years of hands-on experience in Python development, with strong expertise in the Django framework and Django REST Framework (DRF).
- Proven experience in designing and building RESTful APIs, with a solid understanding of API versioning, authentication (JWT/OAuth2), and best practices.
- Experience with relational databases such as PostgreSQL or MySQL; familiarity with query optimization and database migrations.
- Basic front-end development skills using HTML, CSS, and JavaScript; experience with any JavaScript framework (like React or Next Js) is a plus.
- Good understanding of Object-Oriented Programming (OOP) and design patterns in Python.
- Familiarity with Git and collaborative development workflows (e.g., GitHub, GitLab).
- Knowledge of Docker, CI/CD pipelines.
- Hands-on experience with AWS services, Nginx web server, RabbitMQ (or similar message brokers), event handling, and synchronization.
- Familiarity with Postgres, SSO implementation (desirable), and integration of third-party libraries.
- Experience with unit testing, debugging, and code reviews.
- Experience using tools like Jira and Confluence.
- Ability to work in Agile/Scrum teams with good communication and problem-solving skills.
Our Commitment to You:
We offer a competitive salary and generous benefits. In addition, we offer a vibrant work environment, a global team filled with passionate and fun-loving people coming from diverse cultures and backgrounds.
If you are looking to make an impact in delivering market-leading risk management solutions, empowering our clients, and making the world a better place, then Supply Wisdom is the place for you.
You can learn more at supplywisdom.com and on LinkedIn.

Position : Tech Lead – Fullstack Developer
Experience : 7 to 15 Years
Location : MG Road, Bengaluru (Hybrid – 3 Days in Office)
Notice Period : Immediate / Serving / 15 Days or Less
About the Opportunity :
We are hiring a Tech Lead – Fullstack Developer for a well-funded product startup building an enterprise-grade SaaS platform in the Cybersecurity domain.
The role involves designing and delivering scalable microservices and cloud-native applications in a high-performing, Agile engineering environment.
You'll work alongside industry veterans from billion-dollar digital firms, contributing to technical design, product architecture, and engineering best practices.
Mandatory Skills : Java, Spring Boot, ReactJS (or any modern JavaScript framework), RESTful APIs, PostgreSQL, Docker, Kubernetes, CI/CD, Hibernate/JPA, Multithreading, and Microservices architecture.
Role Highlights :
- Lead fullstack product development using Java (Spring Boot) and ReactJS (or similar frameworks).
- Design, develop, test, and deploy scalable microservices and RESTful APIs.
- Collaborate with Product, DevOps, and QA teams in a fast-paced Agile environment.
- Write modular, secure, and efficient code optimized for performance and maintainability.
- Mentor junior developers and influence architecture decisions across the team.
- Participate in all stages of the product lifecycle, from design to deployment.
- Create technical documentation, UML diagrams, and contribute to knowledge-sharing through blogs or whitepapers.
Key Skills Required :
- Strong expertise in Java (mandatory) and Spring Boot.
- Proficient in frontend development using ReactJS or similar frameworks.
- Hands-on experience building and consuming RESTful APIs.
- Solid knowledge of PostgreSQL, Hibernate/JPA, and transaction management.
- Familiarity with Docker, Kubernetes, and cloud platforms (Azure/GCP).
- Understanding of API Gateway, ACID properties, multithreading, and performance tuning.
- Experience with CI/CD tools (Jenkins, GitHub Actions, GitLab CI) and Agile methodologies.
- Strong debugging and profiling skills for performance bottlenecks.
Nice to Have :
- Experience with data integration tools (e.g., Pentaho, Apache NiFi).
- Exposure to the Healthcare or Cybersecurity domain.
- Familiarity with OpenAI APIs or real-time analytics tools.
- Willingness to contribute to internal documentation, blog posts, or whitepapers.
Perks & Benefits :
- Opportunity to build a product from scratch.
- Flat hierarchy and direct access to leadership.
- Strong focus on learning, mentorship, and technical innovation.
- Collaborative startup culture with long-term growth opportunities.
Interview Process :
- Technical Round – Technical Assessment
- Technical Interview – Core Development
- Advanced Technical Interview – Design & Problem Solving
- Final Round – CTO Discussion

Job Description :
We are seeking a talented and experienced Full Stack Developer to join our dynamic team in Hyderabad. The ideal candidate will have a passion for building scalable and efficient web applications, a strong understanding of modern frameworks and technologies, and a keen eye for user experience and design.
Key Responsibilities :
- Design, develop, and maintain web-based applications using React JS, NodeJS, Angular, React Native, and other modern frameworks.
- Develop hybrid mobile applications and responsive web interfaces using Bootstrap and JavaScript.
- Build and optimize back-end services with frameworks such as Express.js or Restify.
- Work with SQL databases, including schema design and query optimization.
- Utilize ORM tools like Sequelize for database management.
- Implement real-time communication features and ensure browser compatibility.
- Collaborate with cross-functional teams to participate in the product development lifecycle, including prototyping, testing, and deployment.
- Adapt to and learn alternative technologies based on project requirements.
Required Skills & Experience :
- 3+ years of experience in full-stack web development.
- Proficient in Angular, NodeJS, React.JS, and JavaScript.
- Strong experience with Express.js or Restify frameworks.
- Solid understanding of SQL databases and ORM tools like Sequelize.
- Knowledge of responsive design principles and hands-on experience in developing responsive web applications.
- Familiarity with React Native for mobile development (a plus)
- Strong understanding of real-time communication technologies.
Additional Skills & Experience :
- Experience with NoSQL databases such as MongoDB or Cassandra.
- Awareness of internationalization (i18n) and the latest trends in UI/UX design.
- Familiarity with other JavaScript libraries/frameworks like VueJS.
- Hands-on experience with implementing payment gateways for different regions.
- Excellent facilitation, verbal, and written communication skills.
- Eagerness to contribute to functional and user experience design discussions.
Education : B.Tech/M.Tech in CSE/IT.ECE
Devops Engineer :
Tech stack : AWS, Gitlab , Python, SNF
Mrproptek is on the lookout for a DevOps Engineer with a passion for cloud infrastructure, automation, and scalable systems. If you're ready to hit the ground running — we want you on our team ASAP!
Location: Mohali, Punjab (On-site)
Experience: 2+ Years
Skills We’re Looking For:
- Strong hands-on experience with AWS
- Expertise in GitLab CI/CD pipelines
- Python scripting proficiency
- Knowledge of SNF (Serverless and Functions) architecture
- Excellent communication and collaboration skills
- Immediate joiners preferred
At MrPropTek, we're redefining the future of property technology with innovative digital solutions. Join a team where your skills truly matter and your ideas shape what’s next.
About the Company
We are hiring for a fast-growing, well-funded product startup backed by a leadership team with a proven track record of building billion-dollar digital businesses. The company is focused on delivering enterprise-grade SaaS products in the Cybersecurity domain for B2B markets. You’ll be part of a passionate and dynamic engineering team building innovative solutions using modern tech stacks.
Key Responsibilities
- Design and develop scalable microservices using Java and Spring Boot
- Build and manage robust RESTful APIs
- Collaborate with cross-functional teams in an Agile setup
- Lead and mentor junior engineers, driving technical excellence
- Contribute to architecture discussions and code reviews
- Work with PostgreSQL, implement data integrity and consistency
- Deploy and manage services on cloud platforms like GCP or Azure
- Utilize Docker/Kubernetes for containerization and orchestration
Must-Have Skills
- Strong backend experience with Java, Spring Boot, REST APIs
- Proficiency in frontend development with React.js
- Experience with PostgreSQL and database optimization
- Hands-on with cloud platforms (GCP or Azure)
- Familiarity with Docker and Kubernetes
- Strong understanding of:
- API Gateways
- Hibernate & JPA
- Transaction management & ACID properties
- Multi-threading and context switching
Good to Have
- Experience in Cybersecurity or Healthcare domain
Exposure to CI/CD pipelines and DevOps practices