Cutshort logo
AWS (Amazon Web Services) Jobs in Pune

50+ AWS (Amazon Web Services) Jobs in Pune | AWS (Amazon Web Services) Job openings in Pune

Apply to 50+ AWS (Amazon Web Services) Jobs in Pune on CutShort.io. Explore the latest AWS (Amazon Web Services) Job opportunities across top companies like Google, Amazon & Adobe.

icon
Searce Inc

at Searce Inc

3 recruiters
Srishti Dani
Posted by Srishti Dani
Mumbai, Pune, Bengaluru (Bangalore)
7 - 10 yrs
Best in industry
Data migration
Datawarehousing
ETL
SQL
Google Cloud Platform (GCP)
+7 more

Lead Data Engineer


What are we looking for

real solver?

Solver? Absolutely. But not the usual kind. We're searching for the architects of the audacious & the pioneers of the possible. If you're the type to dismantle assumptions, re-engineer ‘best practices,’ and build solutions that make the future possible NOW, then you're speaking our language.


Your Responsibilities

What you will wake up to solve.

  • Lead Technical Design & Data Architecture: Architect and lead the end-to-end development of scalable, cloud-native data platforms. You’ll guide the squad on critical architectural decisions—choosing between Batch vs. Streaming or ETL vs. ELT—while remaining 100% hands-on, contributing high-quality, production-grade code.
  • Build High-Velocity Data Pipelines: Drive the implementation of robust data transports and ingestion frameworks using Python, SQL, and Spark. You will build integration layers that connect heterogeneous sources (SaaS, RDBMS, NoSQL) into unified, high-availability environments like BigQuery, Snowflake, or Redshift.
  • Mentor & Elevate the Squad: Foster a culture of technical excellence by mentoring and inspiring a team of data analysts and engineers. Lead deep-dive code reviews, promote best-practice data modeling (Star/Snowflake schema), and ensure the squad adopts modern engineering standards like CI/CD for data.
  • Drive AI-Ready Data Strategy: Be the expert in designing data foundations optimized for AI and Machine Learning. You will champion the use of GCP (Dataflow, Pub/Sub, BigQuery) and AWS (Lambda, Glue, EMR) to create "clean room" environments that fuel advanced analytics and generative AI models.
  • Partner with Clients as a Technical DRI: Act as the Directly Responsible Individual for client success. Translate ambiguous business questions into elegant data services, manage project deliverables using Agile methodologies, and ensure that the data provided is accurate, consistent, and mission-critical.
  • Troubleshoot & Optimize for Scale: Own the reliability of the reporting layer. You will proactively monitor pipelines, troubleshoot complex transformation bottlenecks, and propose ways to improve platform performance and cost-efficiency.
  • Innovate and Build Reusable IP: Spearhead the creation of reusable data frameworks, custom operators, and transformation libraries that accelerate future projects and establish Searce’s unique technical advantage in the market.


Welcome to Searce


The AI-Native tech consultancy that's rewriting the rules.

Searce is an AI-native, engineering-led, modern tech consultancy that empowers clients to futurify their business by delivering intelligent, impactful, real business outcomes. Searce solvers co-innovate with clients as their trusted transformational partners ensuring sustained competitive advantage. Searce clients realize smarter, faster, better business outcomes delivered by AI-native Searce solver squads. 


Functional Skills 

the solver personas.

  • The Data Architect: This persona deconstructs ambiguous business goals into scalable, elegant data blueprints. They don't just move data; they design the foundation—from schema design to partitioning strategies—that allows data scientists and analysts to thrive, foreseeing technical bottlenecks and making pragmatic trade-offs.
  • The Player-Coach: As a hands-on leader, this persona leads from the front by writing exemplary, production-grade SQL and Python while simultaneously mentoring and elevating the skills of the squad. Their success is measured by the team's ability to deliver high-quality, maintainable code and their growth as engineers.
  • The Pragmatic Innovator: This individual balances a passion for modern data tech (like Generative AI and Real-time Streaming) with a sharp focus on business outcomes. They champion new tools where they add real value but are disciplined enough to choose stable, cost-effective solutions to meet deadlines and deliver robust products.
  • The Client-Facing Technologist: This persona acts as the crucial technical bridge between the data squad and the client. They build trust by listening actively, explaining complex data concepts (like data latency or idempotency) in simple terms, and demonstrating how engineering decisions align with the client’s strategic goals.
  • The Quality Craftsman: This individual possesses an unwavering commitment to data integrity and treats data engineering as a craft. They are the guardian of the reporting layer, advocating for robust testing, data validation frameworks, and clean, modular code to ensure the long-term reliability of the data platform.


Experience & Relevance 

  • Engineering Depth: 7-10 years of professional experience in end-to-end data product development. You have a portfolio that proves your ability to build complex, high-velocity pipelines for both Batch and Streaming workloads.
  • Cloud-Native Fluency: Deep, hands-on experience designing and deploying scalable data solutions on at least one major cloud platform (AWS, GCP, or Azure). You are comfortable navigating the nuances of EMR, BigQuery, or Synapse at scale.
  • AI-Native Workflow: You don’t just build for AI; you build with AI. You must be proficient in using AI coding assistants (e.g., GitHub Copilot) to accelerate your delivery and have a track record of building the data foundations required for Generative AI.
  • Architectural Portfolio: Evidence of leading 2-3 large-scale transformations—including platform migrations, data lakehouse builds, or real-time analytics architectures.
  • Client-Facing Acumen: You have direct experience in a consultative, client-facing role. You can confidently translate a CEO’s business vision into a Lead Engineer’s technical specification without losing anything in translation.


Join the ‘real solvers’

ready to futurify?

If you are excited by the possibilities of what an AI-native engineering-led, modern tech consultancy can do to futurify businesses, apply here and experience the ‘Art of the possible’. Don’t Just Send a Resume. Send a Statement.





Read more
Searce Inc

at Searce Inc

3 recruiters
Akanksha Verma
Posted by Akanksha Verma
Pune, Mumbai
4 - 8 yrs
Best in industry
MySQL DBA
skill iconPostgreSQL
skill iconAmazon Web Services (AWS)
Performance tuning
Relational Database (RDBMS)
+1 more


Role Overview:

As a Database Administrator, you will be responsible for the full lifecycle management of our MySQL or PostgreSQL database systems. This includes installation, configuration, performance tuning, security implementation, backup and recovery, and proactive monitoring. You will ensure the reliability, availability, and security of our database infrastructure, supporting our internal operations and client projects.


Job Responsibilities


  • Installation and configuration of database software across diverse operating systems.
  • Designing efficient physical database models derived from logical designs and application specifications, along with configuring database servers according to best practices and workload requirements.
  • Establishing and implementing robust backup and recovery strategies tailored to data volatility and application availability needs.
  • Implementing comprehensive security measures at the OS, database, and network levels to ensure authorized data access and maintain a rigorous security infrastructure with auditing capabilities for compliance.
  • Fine-tuning hardware/VM resources for optimal database performance.
  • Proactive monitoring of the database environment, including performance optimization through adjustments to data structures, SQL, application logic, or the DBMS subsystem.
  • Configuration and implementation of database replication technologies (e.g., Master-Slave, Master-Master, Log Shipping, Mirroring, Always On).
  • Automation of routine DBA tasks utilizing scripting languages such as Shell, PowerShell, Python, or GO.
  • Proficiency in writing general SQL queries.
  • Setting up comprehensive monitoring solutions for databases (OS and database levels) using custom scripts or third-party monitoring tools.
  • Basic Cloud platform knowledge (AWS/GCP/Azure)


Qualification


Experienced DBA (4-8 years) with deep expertise in MySQL or PostgreSQL architecture, configuration, and management.

Proficient in SQL, backup/recovery, security implementation, performance tuning, and replication for both systems.

Skilled in scripting (e.g., Shell, Python) and Linux, possessing strong problem-solving, communication, and teamwork abilities with a proactive approach.

A relevant Bachelor's degree in Computer Science, Information Technology, or a related field.


We identify better ways of doing things.

Solver? Absolutely. But not the usual kind. We are searching for the architects of the

audacious & the pioneers of the possible. If you are the type to dismantle assumptions,

re-engineer ‘best practices,’ and build solutions that make the future possible NOW,

then you are speaking our language.

➔ Improver. Solver. Futurist.

➔ Great sense of humor.

➔ ‘Possible. It is.’ Mindset.

➔ Compassionate collaborator. Bold experimenter. Tireless iterator.

➔ Natural creativity that doesn’t just challenge the norm, but solves to design what’s better.

➔ Thinks in systems. Solves at scale.

This Isn’t for Everyone. But if you’re the kind who questions why things are done a

certain way—and then identifies 3 better ways to do it — we’d love to chat with you.

Read more
Searce Inc

at Searce Inc

3 recruiters
Jatin Gereja
Posted by Jatin Gereja
Bengaluru (Bangalore), Mumbai, Pune
10 - 18 yrs
Best in industry
Google Cloud Platform (GCP)
skill iconAmazon Web Services (AWS)
Enterprise Data Warehouse (EDW)
Data modeling
Big Data
+9 more

Director - Data engineering


What are we looking for

real solver?

Solver? Absolutely. But not the usual kind. We're searching for the architects of the audacious & the pioneers of the possible. If you're the type to dismantle assumptions, re-engineer ‘best practices,’ and build solutions that make the future possible NOW, then you're speaking our language.


Your Responsibilities

what you will wake up to solve.

1. Delivery & Tactical Rigor

  • Methodology Implementation: Implement and manage a unified, 'DataOps-First' methodology for data engineering delivery (ETL/ELT pipelines, Data Modeling, MLOps, Data Governance) within assigned business units. This ensures predictable outcomes and trusted data integrity by reducing architecture variability at the project level.
  • Operational Stewardship: Drive initiatives to optimize team utilization and enhance operational efficiency within the practice. You manage the commercial success of your squads, ensuring data delivery models (from migration to modern data stack implementation) are executed profitably, scalably, and cost-effectively.
  • Execution & Technical Resolution
  • Technical Escalation: Serve as the primary escalation point for delivery issues, personally leading the resolution of complex data integration bottlenecks and pipeline failures to protect client timelines and data reliability standards.
  • Quality Enforcement
  • Quality Oversight: Execute and monitor technical data quality standards, ensuring engineering teams adhere to strict policies regarding data lineage, automated quality checks (observability), security/privacy compliance (GDPR/CCPA/PII), and active catalog management.

2. Strategic Growth & Practice Scaling

  • Talent & Scaling Execution: Execute the strategy for data engineering talent acquisition and development within your business units. Implement objective metrics to assess and grow the 'Data-Native' DNA of your teams, ensuring squads are consistently equipped to handle petabyte-scale environments and high-impact delivery.
  • Offerings Alignment: Drive the adoption of standardized regional offerings (e.g., Modern Data Platform, Data Mesh, Lakehouse Implementation). Ensure your teams leverage the profitable frameworks defined by the practice to accelerate time-to-insight and eliminate architectural fragmentation in client environments.
  • Innovation & IP Development: Lead the practical integration of Vector Databases and LLM-ready architectures into project delivery. Champion the hands-on development of IP and reusable accelerators (e.g., automated ingestion engines) that improve delivery speed and enhance data availability across your portfolio.

3. Leadership & Unit Management

  • Unit Leadership: Directly lead, mentor, and manage the Engineering Managers and Lead Architects within your business unit. Hold your teams accountable for project-level operational consistency, technical talent development, and strict adherence to the practice's data governance standards.
  • Stakeholder Communication: Clearly articulate the business unit’s operational performance, technical quality metrics, and delivery progress to the C-suite Stakeholders and regional client leadership, bridging the gap between technical execution and business value.
  • Ecosystem Alignment: Maintain strong technical relationships with key partner contacts (Snowflake, Databricks, AWS/GCP). Align team delivery capabilities with current product roadmaps and ensure squad-level participation in training, certifications, and partner-led enablement opportunities.


Welcome to Searce

The ‘process-first’, AI-native modern tech consultancy that's rewriting the rules.

We don’t do traditional.

As an engineering-led consultancy, we are dedicated to relentlessly improving the real business outcomes. Our solvers co-innovate with clients to futurify operations and make processes smarter, faster & better.


Functional Skills

1. Delivery Management & Operational Excellence

  • Methodology Execution: Expert capability in implementing and enforcing a unified delivery methodology (DataOps, Agile, Mesh Principles) within specific business units. Proven track record of auditing squad-level adherence to ensure consistency across the project lifecycle.
  • Operational Performance: High proficiency in managing day-to-day operational metrics, including squad utilization, resource forecasting, and productivity tracking. Skilled at optimizing team performance to meet profitability and efficiency targets.
  • SOW & Risk Mitigation: Proven experience in operationalizing Statement of Work (SOW) requirements and identifying technical delivery risks early. Expert at mitigating scope creep and data-specific bottlenecks (e.g., latency, ingestion gaps) before they impact client outcomes.
  • Technical Escalation Leadership: Demonstrated ability to lead "war room" efforts to resolve complex pipeline failures or data integrity issues. Skilled at providing clear, rapid remediation plans and communicating technical status directly to regional stakeholders.

2. Architectural Implementation & Technical Oversight

  • Modern Stack Proficiency: Deep, hands-on expertise in implementing Cloud-Native architectures (Lakehouse, Data Mesh, MPP) on Snowflake, Databricks, or hyperscalers. Ability to conduct deep-dive architectural reviews and course-correct design decisions at the squad level to ensure scalability.
  • Operationalizing Governance: Proven experience in embedding data quality and observability (completeness, freshness, accuracy) directly into the CI/CD pipeline. Responsible for technical enforcement of regulatory compliance (GDPR/PII) and maintaining the integrity of data catalogs across active projects.
  • Applied Domain Expertise: Practical experience leading the delivery of high-growth solutions, specifically Generative AI infrastructure (RAG, Vector DBs), Real-Time Streaming, and large-scale platform migrations with a focus on zero-downtime execution.
  • DataOps & Engineering Standards: Expert-level mastery of DataOps, including the setup and management of orchestration frameworks (Airflow, Dagster) and Infrastructure as Code (IaC). You ensure that automation is a baseline requirement, not an afterthought, for all delivery teams.

3. Unit Management & Commercial Execution

  • Unit & Team Management: Proven success in leading and mentoring Engineering Managers and Lead Architects. Responsible for the operational metrics, technical output, and career development of the business unit's talent pool.
  • Offerings Implementation & Scoping: Expertise in translating service offerings (e.g., Data Maturity Assessments, Lakehouse Builds) into accurate project scopes, technical estimates, and resource plans to ensure delivery is both profitable and competitive.
  • Talent Growth & Mentorship: Functional ability to implement growth frameworks for data engineering roles. Focus on hands-on coaching and scaling high-performance technical talent to meet the demands of complex, petabyte-scale environments.
  • Partner Enablement: Functional competence in managing regional technical relationships with major partners (Snowflake, Databricks, GCP/AWS). Drives squad-level certifications, joint technical enablement, and alignment with partner product roadmaps.

Tech Superpowers

  • Modern Data Architect – Reimagines business with the Modern Data Stack (MDS) to deliver data mesh implementations, insights, & real value to clients.
  • End-to-End Ecosystem Thinker – Builds modular, reusable data products across ingestion, transformation (ETL/ELT), governance, and consumption layers.
  • Distributed Compute Savant – Crafts resilient, high-throughput architectures that survive petabyte-scale volume and data skew without breaking the bank.
  • Governance & Integrity Guardian – Embeds data quality, complete lineage, and privacy-by-design (GDPR/PII) into every table, view, and pipeline.
  • AI-Ready Orchestrator – Engineers pipelines that bridge structured data with Unstructured/Vector stores, powering RAG models and Generative AI workflows.
  • Product-Minded Strategist – Balances architectural purity with time-to-insight; treats every dataset as a measurable "Data Product" with clear ROI.
  • Pragmatic Stack Curator – Chooses the simplest tools that compound reliability; fluent in SQL, Python, Spark, dbt, and Cloud Warehouses.
  • Builder @ Heart – Writes, reviews, and optimizes queries daily; proves architectures with cost-performance benchmarks, not slideware. Business-first, data-second, outcome focused technology leader.

Experience & Relevance

  • Executive Experience: Minimum 10+ years of progressive experience in data engineering and analytics, with at least 3 years in a Senior Manager or Director -level role managing multiple technical teams and owning significant operational and efficiency metrics for a large data service line.
  • Delivery Standardization: Demonstrated success in defining and implementing globally consistent, repeatable delivery methodologies (DataOps/Agile Data Warehousing) across diverse teams.
  • Architectural Depth: Must retain deep, current expertise in Modern Data Stack architectures (Lakehouse, MPP, Mesh) and maintain the ability to personally validate high-level architectural and data pipeline design decisions.
  • Operational Leadership: Proven expertise in managing and scaling large professional services organizations, demonstrated ability to optimize utilization, resource allocation, and operational expense.
  • Domain Expertise: Strong background in Enterprise Data Platforms, Applied AI/ML, Generative AI integration, or large-scale Cloud Data Migration.
  • Communication: Exceptional executive-level presentation and negotiation skills, particularly in communicating complex operational, data quality, and governance metrics to C-level stakeholders.

Join the ‘real solvers’

ready to futurify?

If you are excited by the possibilities of what an AI-native engineering-led, modern tech consultancy can do to futurify businesses, apply here and experience the ‘Art of the possible’. Don’t Just Send a Resume. Send a Statement.

Read more
Searce Inc

at Searce Inc

3 recruiters
Tejashree Kokare
Posted by Tejashree Kokare
Bengaluru (Bangalore), Pune, Mumbai
6 - 15 yrs
Best in industry
Google Cloud Platform (GCP)
Data engineering
Data warehouse architecture
Data architecture
Data modeling
+6 more

Solutions Architect - Data Engineering


Modern tech solutions advisory & 'futurify' consulting as a Searce lead fds (‘forward deployed solver’) architecting scalable data platforms and robust data engineering solutions that power intelligent insights and fuel AI innovation.

If you’re a tech-savvy, consultative seller with the brain of a strategist, the heart of a builder, and the charisma of a storyteller — we’ve got a seat for you at the front of the table.

You're not a sales lead. You're the transformation driver.


What are we looking for

real solver?

Solver? Absolutely. But not the usual kind. We're searching for the architects of the audacious & the pioneers of the possible. If you're the type to dismantle assumptions, re-engineer ‘best practices,’ and build solutions that make the future possible NOW, then you're speaking our language.

  • Improver. Solver. Futurist.
  • Great sense of humor.
  • ‘Possible. It is.’ Mindset.
  • Compassionate collaborator. Bold experimenter. Tireless iterator.
  • Natural creativity that doesn’t just challenge the norm, but solves to design what’s better.
  • Thinks in systems. Solves at scale.


This Isn’t for Everyone. But if you’re the kind who questions why things are done a certain way— and then identifies 3 better ways to do it — we’d love to chat with you.


Your Responsibilities

what you will wake up to solve.


You are not just a Solutions Architect; you are a futurifier of our data universe and the primary enabler of our AI ambitions. With a deep-seated passion for data engineering, you will architect and build the foundational data infrastructure that powers the customers entire data intelligence ecosystem.

As the Directly Responsible Individual (DRI) for our enterprise-grade data platforms, you own the outcome, end-to-end. You are the definitive solver for our customer's most complex data challenges, leveraging a powerful tech stack including Snowflake, Databricks, etc. and core GCP & AWS services (BigQuery, Spanner, Airflow, Kafka). This is a hands-on-keys role where you won't just design solutions—you'll build them, break them, and perfect them.


  • Solution Design & Pre-sales Excellence:Collaborate with cross-functional teams, including sales, engineering, and operations, to ensure successful project delivery.
  • Design Core Data Engineering: Master data modeling, architecting high-performance data ingestion pipelines and ensuring data quality and governance throughout the data lifecycle.
  • Enable Cloud & AI: Design and implement solutions utilizing core GCP data services, building foundational data platforms that efficiently support advanced analytics and AI/ML initiatives.
  • Optimize Performance & Cost: Continuously optimize data architectures and implementations for performance, efficiency, and cost-effectiveness within the cloud environment.
  • Bridge Business & Tech: Translate complex business requirements into clear technical designs, providing technical leadership and guidance to data engineering teams.
  • Stay Ahead of the Curve: Continuously research and evaluate new data technologies, architectural patterns, and industry trends to keep our data platforms at the cutting edge.


Functional Skills:


  • Enterprise Data Architecture Design: Expert ability to design holistic, scalable, and resilient data architectures for complex enterprise environments.
  • Cloud Data Platform Strategy: Proven capability to strategize, design, and implement cloud-native data platforms.
  • Pre-Sales & Technical Storyteller: Crafts compelling, client-ready proposals, architectural decks, and technical demonstrations. Doesn't just present; shapes the strategic technical narrative behind every proposed solution.
  • Advanced Data Modelling: Mastery in designing various data models for analytical, operational, and transactional use cases.
  • Data Ingestion & Pipeline Orchestration: Strong expertise in designing and optimizing robust data ingestion and transformation pipelines.
  • Stakeholder Communication: Exceptional skills in articulating complex technical concepts and architectural decisions to both technical and non-technical stakeholders.
  • Performance & Cost Optimization: Adept at optimizing data solutions for performance, efficiency, and cost within a cloud environment.


Tech Superpowers:


  • Cloud Data Mastery: You're a wizard at leveraging public cloud data services, with deep expertise in GCP (BigQuery, Spanner, etc.) and expert proficiency in modern data warehouse solutions like Snowflake.
  • Data Engineering Core: Highly skilled in designing, implementing, and managing data workflows using tools like Apache Airflow and Apache Kafka. You're also an authority on advanced data modeling and ETL/ELT patterns.
  • AI/ML Data Foundation: You instinctively design data pipelines and structures that efficiently feed and empower Machine Learning and Artificial Intelligence applications.
  • Programming for Data: You have a strong command over key programming languages (Python, SQL) for scripting, automation, and building data processing applications.


Experience & Relevance:


  • Architectural Leadership (8+ Years): You bring extensive experience (7+ years) specifically in a Solutions Architect role, focused on data engineering and platform building.
  • Cloud Data Expertise: You have a proven track record of designing and implementing production-grade data solutions leveraging major public cloud platforms, with significant experience in Google Cloud Platform (GCP).
  • Data Warehousing & Data Platform: Demonstrated hands-on experience in the end-to-end design, implementation, and optimization of modern data warehouses and comprehensive data platforms.
  • Databricks & BigQuery Mastery: You possess significant practical experience with Databricks as a core data warehouse and GCP BigQuery for analytical workloads.
  • Data Ingestion & Orchestration: Proven experience designing and implementing complex data ingestion pipelines and workflow orchestration using tools like Airflow and real-time streaming technologies like Kafka.
  • AI/ML Data Enablement: Experience in building data foundations specifically geared towards supporting Machine Learning and Artificial Intelligence initiatives.


Join the ‘real solvers’

ready to futurify?

If you are excited by the possibilities of what an AI-native engineering-led, modern tech consultancy can do to futurify businesses, apply here and experience the ‘Art of the possible’.


Don’t Just Send a Resume. Send a Statement.


So, If you are passionate about tech, future & what you read above (we really are!), apply here to experience the ‘Art of Possible’

Read more
NonStop io Technologies Pvt Ltd
Kalyani Wadnere
Posted by Kalyani Wadnere
Pune
7 - 10 yrs
Best in industry
skill iconNodeJS (Node.js)
NestJS
skill iconExpress
skill iconPostgreSQL
Microservices
+7 more

About NonStop io Technologies:

NonStop io Technologies is a value-driven company with a strong focus on process-oriented software engineering. We specialize in Product Development and have a decade's worth of experience in building web and mobile applications across various domains. NonStop io Technologies follows core principles that guide its operations and believes in staying invested in a product's vision for the long term. We are a small but proud group of individuals who believe in the 'givers gain' philosophy and strive to provide value in order to seek value. We are committed to and specialize in building cutting-edge technology products and serving as trusted technology partners for startups and enterprises. We pride ourselves on fostering innovation, learning, and community engagement. Join us to work on impactful projects in a collaborative and vibrant environment.

We're seeking an experienced Senior Backend Engineer to join our team. As a senior backend engineer, you will be responsible for designing, developing, and deploying scalable backends for the products we build at NonStop. This includes APIs, databases, and server-side logic.


Responsibilities:

● Design, develop, and deploy backend systems, including APIs, databases, and server-side logic

● Write clean, efficient, and well-documented code that adheres to industry standards and best practices

● Code Quality: Ensure code quality through code reviews, adherence to best practices, and continuous improvement

● Mentorship: Guide and mentor team members, fostering growth and innovation

● Collaboration: Work closely with stakeholders to align technical goals with business objectives

● Problem-Solving: Analyze and resolve technical challenges promptly ● Innovation: Stay updated with the latest technology trends and integrate them into solutions


Requirements:

● At least 7+ years of experience building scalable and reliable backend systems

● Strong expertise in NodeJS/NestJS, Express, PostgreSQL

● Experience with microservices architecture and distributed systems

● Proficiency in database design (SQL and NoSQL)

● Knowledge of cloud platforms (AWS, Azure, or GCP) and CI/CD pipelines

● Deep understanding of design patterns, data structures, and algorithms

● Hands-on experience with containerization technologies like Docker and orchestration tools like Kubernetes

● Exceptional communication and leadership skills

● Strong understanding of object-oriented programming principles and design patterns

● Familiarity with automated testing frameworks and methodologies

● Excellent problem-solving skills and attention to detail

● Strong communication skills and ability to effectively lead and maintain a collaborative team environment

Read more
NonStop io Technologies Pvt Ltd
Kalyani Wadnere
Posted by Kalyani Wadnere
Pune
3 - 5 yrs
Best in industry
skill iconNodeJS (Node.js)
NestJS
skill iconExpress
skill iconPostgreSQL
SQL
+5 more

About NonStop io Technologies:

NonStop io Technologies is a value-driven company with a strong focus on process-oriented software engineering. We specialize in Product Development and have a decade's worth of experience in building web and mobile applications across various domains. NonStop io Technologies follows core principles that guide its operations and believes in staying invested in a product's vision for the long term. We are a small but proud group of individuals who believe in the 'givers gain' philosophy and strive to provide value in order to seek value. We are committed to and specialize in building cutting-edge technology products and serving as trusted technology partners for startups and enterprises. We pride ourselves on fostering innovation, learning, and community engagement. Join us to work on impactful projects in a collaborative and vibrant environment. We're seeking an experienced Backend Software Engineer to join our team. As a backend engineer, you will be responsible for designing, developing, and deploying scalable backends for the products we build at NonStop. This includes APIs, databases, and server-side logic.


Responsibilities:

● Design, develop, and deploy backend systems, including APIs, databases, and server-side logic

● Write clean, efficient, and well-documented code that adheres to industry standards and best practices

● Participate in code reviews and contribute to the improvement of the codebase

● Debug and resolve issues in the existing codebase

● Develop and execute unit tests to ensure high code quality

● Work with DevOps engineers to ensure seamless deployment of software changes

● Monitor application performance, identify bottlenecks, and optimize systems for better scalability and efficiency

● Stay up-to-date with industry trends and emerging technologies; advocate for best practices and new ideas within the team

● Collaborate with cross-functional teams to identify and prioritize project requirements


Requirements:

● At least 3+ years of experience building scalable and reliable backend systems

● Strong expertise in NodeJS/NestJS, Express, PostgreSQL

● Experience with microservices architecture and distributed systems

● Proficiency in database design (SQL and NoSQL)

● Knowledge of cloud platforms (AWS, Azure, or GCP) and CI/CD pipelines

● Deep understanding of design patterns, data structures, and algorithms

● Hands-on experience with containerization technologies like Docker and orchestration tools like Kubernetes

● Exceptional communication and leadership skills

● Strong understanding of object-oriented programming principles and design patterns

● Familiarity with automated testing frameworks and methodologies

● Excellent problem-solving skills and attention to detail

● Strong communication skills and ability to effectively lead and maintain a collaborative team environment

Read more
NonStop io Technologies Pvt Ltd
Kalyani Wadnere
Posted by Kalyani Wadnere
Pune
3 - 5 yrs
Best in industry
skill iconHTML/CSS
skill iconJavascript
skill iconReact.js
skill iconBootstrap
RESTful APIs
+8 more

About NonStop io Technologies:

NonStop io Technologies is a value-driven company with a strong focus on process-oriented software engineering. We specialize in Product Development and have a decade's worth of experience in building web and mobile applications across various domains. NonStop io Technologies follows core principles that guide its operations and believes in staying invested in a product's vision for the long term. We are a small but proud group of individuals who believe in the 'givers gain' philosophy and strive to provide value in order to seek value. We are committed to and specialize in building cutting-edge technology products and serving as trusted technology partners for startups and enterprises. We pride ourselves on fostering innovation, learning, and community engagement. Join us to work on impactful projects in a collaborative and vibrant environment.


Brief Description:

NonStop io is seeking a proficient React.js developer to join our front-end development team. In this position, you’ll be mainly crafting and integrating UI components through React.js methodologies and workflows like Mobx and Redux.


Responsibilities:

● Developing and implementing UI components using React.js

● Collaborating with cross-functional teams to design and ship new features

● Building reusable components and front-end libraries for future use

● Translating designs and wireframes into high-quality code

● Optimizing components for maximum performance across various web browsers

● Troubleshooting and debugging issues to ensure smooth user experiences

● Participating in code reviews to maintain code quality and consistency ● Improvement and optimization of existing codebase

● Keeping up with industry trends

● Identifying issues with technologies and architecture and then implementing solutions

● Assist with ticket creation, refinement, and estimation

● Participate in sprint planning and ticket distribution for frontend team


Qualifications & Skills:

● Proficiency in React.js and its core principles

● Strong JavaScript, TypeScript, HTML5, and CSS3 skills

● Experience with popular React.js state management libraries and approaches (such as Mobx, Redux, and Context API)

● Familiarity with RESTful APIs and integration

● Knowledge of modern authorization mechanisms, such as JSON Web Tokens

● Understanding of front-end build tools and pipelines

● Excellent problem-solving and communication skills

● A strong attention to detail, and a passion for delivering high-quality code

● Expertise in designing scalable and efficient front-end architecture

● Adaptability to changing project requirements and priorities

● Experience with version control systems, particularly Git

● Experience working in Scrum and familiarity with Atlassian (Jira, Confluence, Bitbucket)

● Proven experience in leading teams

● Experience with diverse applications and architectural patterns

● A degree in computer science, software engineering, or a related field

● Strong understanding of object-oriented programming principles and design patterns

● Familiarity with automated testing frameworks and methodologies

● Excellent problem-solving skills and attention to detail

● Strong communication skills and ability to work effectively in a collaborative team environment

Read more
Nirmitee.io

at Nirmitee.io

4 recruiters
Gitashri K
Posted by Gitashri K
Pune
3 - 5 yrs
₹5L - ₹11L / yr
MERN Stack
SQL
skill iconAmazon Web Services (AWS)
Fullstack Developer
skill iconHTML/CSS
+1 more

We are looking for a highly skilled Full Stack Developer (MERN Stack) with 3–5 years of experience to join our growing team. You will have the opportunity to work on cutting-edge technology solutions, build products from scratch, and contribute to scalable systems handling large volumes of data.


Key Responsibilities:

  • Design, develop, and maintain scalable full-stack applications
  • Build responsive and high-performance user interfaces using modern frontend frameworks
  • Develop robust backend services and APIs
  • Ensure seamless system performance while handling large-scale data without slowdowns
  • Collaborate with cross-functional teams (product, design, QA) to meet business goals
  • Optimize applications for maximum speed, scalability, and reliability
  • Participate in architecture discussions and contribute to technical decisions


Required Skills & Qualifications:

Frontend

  • Strong experience in React.js
  • Hands-on experience with Next.js (mandatory)
  • Good understanding of UI/UX principles and responsive design

Backend

  • Solid experience in Node.js
  • Experience with Python or Java is a plus
  • Strong knowledge of RESTful APIs and microservices architecture

Databases

  • Strong experience with SQL (mandatory)
  • Experience with MongoDB is a plus
  • Caching & Messaging
  • Experience with at least one: Redis, Kafka, or Cassandra

Other Requirements


  • Strong problem-solving and analytical skills
  • Ability to work in a fast-paced, collaborative environment
  • Good communication and stakeholder management skills

Good to Have:

  • Cloud certifications (AWS / Azure / GCP)
  • Experience working on high-scale or distributed systems
  • Exposure to DevOps practices and CI/CD pipelines


Why Join Us:

  • Opportunity to work on cutting-edge tech and greenfield projects
  • Ownership and freedom to build solutions from scratch
  • Collaborative and growth-focused work environment
Read more
NonStop io Technologies Pvt Ltd
Kalyani Wadnere
Posted by Kalyani Wadnere
Pune
2 - 5 yrs
Best in industry
skill icon.NET
ASP.NET
skill iconC#
Entity Framework
Web API
+6 more

Company Description:

NonStop io Technologies, founded in August 2015, is a Bespoke Engineering Studio specializing in Product Development. With over 80 satisfied clients worldwide, we serve startups and enterprises across prominent technology hubs, including San Francisco, New York, Houston, Seattle, London, Pune, and Tokyo. Our experienced team brings over 10 years of expertise in building web and mobile products across multiple industries. Our work is grounded in empathy, creativity, collaboration, and clean code, striving to build products that matter and foster an environment of accountability and collaboration.


Brief Description:

NonStop io is seeking a proficient .NET Developer to join our growing team. You will be responsible for developing, enhancing, and maintaining scalable applications using .NET technologies. This role involves working on a healthcare-focused product and requires strong problem-solving skills, attention to detail, and a passion for software development.


Responsibilities:

  • Design, develop, and maintain applications using .NET Core/.NET Framework, C#, and related technologies
  • Write clean, scalable, and efficient code while following best practices
  • Develop and optimize APIs and microservices
  • Work with SQL Server and other databases to ensure high performance and reliability
  • Collaborate with cross-functional teams, including UI/UX designers, QA, and DevOps
  • Participate in code reviews and provide constructive feedback
  • Troubleshoot, debug, and enhance existing applications
  • Ensure compliance with security and performance standards, especially for healthcare-related applications


Qualifications & Skills:

  • Strong experience in .NET Core/.NET Framework and C#
  • Proficiency in building RESTful APIs and microservices architecture
  • Experience with Entity Framework, LINQ, and SQL Server
  • Familiarity with front-end technologies like React, Angular, or Blazor is a plus
  • Knowledge of cloud services (Azure/AWS) is a plus
  • Experience with version control (Git) and CI/CD pipelines
  • Strong understanding of object-oriented programming (OOP) and design patterns
  • Prior experience in healthcare tech or working with HIPAA-compliant systems is a plus


Why Join Us?

  • Opportunity to work on a cutting-edge healthcare product
  • A collaborative and learning-driven environment
  • Exposure to AI and software engineering innovations
  • Excellent work ethics and culture

If you're passionate about technology and want to work on impactful projects, we'd love to hear from you!

Read more
Virtana

at Virtana

3 candid answers
2 recruiters
Krutika Devadiga
Posted by Krutika Devadiga
Pune
5 - 10 yrs
Best in industry
skill iconGo Programming (Golang)
skill iconKubernetes
skill iconDocker
skill iconJava
skill iconAmazon Web Services (AWS)
+5 more

Senior Software Engineer 

Challenge convention and work on cutting edge technology that is transforming the way our customers manage their physical, virtual and cloud computing environments. Virtual Instruments seeks highly talented people to join our growing team, where your contributions will impact the development and delivery of our product roadmap. Our award-winning Virtana Platform provides the only real-time, system-wide, enterprise scale solution for providing visibility into performance, health and utilization metrics, translating into improved performance and availability while lowering the total cost of the infrastructure supporting mission-critical applications.  

We are seeking an individual with expert knowledge in Systems Management and/or Systems Monitoring Software, Observability platforms and/or Performance Management Software and Solutions with insight into integrated infrastructure platforms like Cisco UCS, infrastructure providers like Nutanix, VMware, EMC & NetApp and public cloud platforms like Google Cloud and AWS to expand the depth and breadth of Virtana Products. 


Work Location: Pune/ Chennai


Job Type: Hybrid

 

Role Responsibilities: 

  • The engineer will be primarily responsible for architecture, design and development of software solutions for the Virtana Platform 
  • Partner and work closely with cross functional teams and with other engineers and product managers to architect, design and implement new features and solutions for the Virtana Platform. 
  • Communicate effectively across the departments and R&D organization having differing levels of technical knowledge.  
  • Work closely with UX Design, Quality Assurance, DevOps and Documentation teams. Assist with functional and system test design and deployment automation 
  • Provide customers with complex and end-to-end application support, problem diagnosis and problem resolution 
  • Learn new technologies quickly and leverage 3rd party libraries and tools as necessary to expedite delivery 

 

Required Qualifications:    

  • Minimum of 7+ years of progressive experience with back-end development in a Client Server Application development environment focused on Systems Management, Systems Monitoring and Performance Management Software. 
  • Deep experience in public cloud environment using Kubernetes and other distributed managed services like Kafka etc (Google Cloud and/or AWS) 
  • Experience with CI/CD and cloud-based software development and delivery 
  • Deep experience with integrated infrastructure platforms and experience working with one or more data collection technologies like SNMP, REST, OTEL, WMI, WBEM. 
  • Minimum of 6 years of development experience with one or more of these high level languages like GO, Python, Java. Deep experience with one of these languages is required. 
  • Bachelor’s or Master’s degree in computer science, Computer Engineering or equivalent 
  • Highly effective verbal and written communication skills and ability to lead and participate in multiple projects 
  • Well versed with identifying opportunities and risks in a fast-paced environment and ability to adjust to changing business priorities 
  • Must be results-focused, team-oriented and with a strong work ethic 

 

Desired Qualifications: 

  • Prior experience with other virtualization platforms like OpenShift is a plus 
  • Prior experience as a contributor to engineering and integration efforts with strong attention to detail and exposure to Open-Source software is a plus 
  • Demonstrated ability as a lead engineer who can architect, design and code with strong communication and teaming skills 
  • Deep development experience with the development of Systems, Network and performance Management Software and/or Solutions is a plus 

  

About Virtana:  Virtana delivers the industry’s only broadest and deepest Observability Platform that allows organizations to monitor infrastructure, de-risk cloud migrations, and reduce cloud costs by 25% or more. 

  

Over 200 Global 2000 enterprise customers, such as AstraZeneca, Dell, Salesforce, Geico, Costco, Nasdaq, and Boeing, have valued Virtana’s software solutions for over a decade. 

  

Our modular platform for hybrid IT digital operations includes Infrastructure Performance Monitoring and Management (IPM), Artificial Intelligence for IT Operations (AIOps), Cloud Cost Management (Fin Ops), and Workload Placement Readiness Solutions. Virtana is simplifying the complexity of hybrid IT environments with a single cloud-agnostic platform across all the categories listed above. The $30B IT Operations Management (ITOM) Software market is ripe for disruption, and Virtana is uniquely positioned for success. 

Read more
Cambridge Wealth (Baker Street Fintech)
Sangeeta Bhagwat
Posted by Sangeeta Bhagwat
Pune
1 - 4 yrs
₹3L - ₹7L / yr
SQL
skill iconPython
skill iconAmazon Web Services (AWS)
Spotfire
Qlikview
+12 more

Who are we aka "About Us":

 

We are an early-stage Fintech Startup - working on exciting Fintech Products for some of the Top 5 Global Banks and building our own. If you are looking for a place where you can make a mark and not just be a cog in the wheel, Baker street Fintech Pvt Ltd (Parent Company) might be the place for you. We have a flat, ownership-oriented culture, and deliver world-class quality. You will be working with a founding team that has delivered over 26 industry-leading product experiences and won the Webby awards for Digital Strategy. In short, a bleeding edge team. 

 

As Cambridge Wealth, we are well-established in the wealth and mutual fund distribution segment, having won awards from BSE Star as well as Mutual Fund houses. Our UHNI/HNI/NRI clients include renowned professionals from various industries. 

 

What are we looking for a.k.a “The JD” :

 

We are seeking a skilled and detail-oriented Data Analyst to join our product team. As a Data Analyst, you will play a crucial role in extracting, analysing, and interpreting complex financial data to drive strategic decision-making and optimize our data solutions. The ideal candidate should possess a strong foundation in SQL / NoSQL databases, Python programming, and proficiency in tools like PostgreSQL and Excel. A deep understanding of financial concepts is also a plus. Additionally, having an interest in business intelligence tools and machine learning will be valuable for this role.

 

Responsibilities:

  • Proficient in writing complex SQL Queries
  • Utilize Python for data manipulation, analysis, and visualisation, using libraries such as pandas, matplotlib, psycopg etc.
  • Perform database optimization, indexing, and query tuning to ensure high performance.
  • Monitor and maintain data quality, troubleshoot data-related issues, and implement solutions to optimize data integrity and performance.
  • Design, configure, and maintain PostgreSQL databases
  • Set up and manage database clusters, replication, and backups for disaster recovery

 

Preferred Qualifications:

  • Intermediate-level Excel skills for data analysis and reporting.
  • Strong communication skills to present findings effectively and recommendations to both technical and non-technical stakeholders.
  • Detail-oriented mindset with a commitment to data accuracy and quality.

 

*(Only Applicants who have finished their educational commitments are requested to apply)

 

Not sure whether you should apply? Here's a quick checklist to make things easier. You are someone who:

  • Has worked (1-3 years preferably) or is looking to work specifically with an early-stage startup.
  • You are ready to be a part of a Zero To One Journey which implies that you shall be involved in building fintech products and process from the ground up.
  • You are comfortable to work in an unstructured environment with a small team where you decide what your day looks like and take initiative to take up the right piece of work, own it and work with the founding team on it.
  • This is not an environment where someone will be checking up on you every few hours. It is up to you to schedule check-ins whenever you find the need to, else we assume you are progressing well with your tasks. You will be expected to find solutions to problems and suggest improvements.
  • You want complete ownership for your role & be able to drive it the way you think is right.
  • You can be a self-starter and take ownership of deliverables to develop a consensus with the team on approach and methods and deliver to them.
  • Are looking to stick around for the long term and grow with the company.

 

Read more
NonStop io Technologies Pvt Ltd
Kalyani Wadnere
Posted by Kalyani Wadnere
Pune
5 - 7 yrs
Best in industry
skill iconJava
Selenium
Selenium Web driver
CI/CD
Appium
+11 more

About NonStop io Technologies:

NonStop io Technologies is a value-driven company with a strong focus on process-oriented software engineering. We specialize in Product Development and have a decade's worth of experience in building web and mobile applications across various domains. NonStop io Technologies follows core principles that guide its operations and believes in staying invested in a product's vision for the long term. We are a small but proud group of individuals who believe in the 'givers gain' philosophy and strive to provide value in order to seek value. We are committed to and specialize in building cutting-edge technology products and serving as trusted technology partners for startups and enterprises. We pride ourselves on fostering innovation, learning, and community engagement. Join us to work on impactful projects in a collaborative and vibrant environment.


Brief Description:

We are seeking a highly skilled QA Automation Engineer with strong expertise in Java and Selenium to join our growing engineering team. The ideal candidate will play a key role in designing, developing, and maintaining scalable test automation frameworks while ensuring high product quality across releases.


Roles and Responsibilities:

● Design, develop, and maintain robust automation frameworks using Java and Selenium

● Build automated test scripts for web applications and integrate them into CI CD pipelines

● Collaborate closely with developers, product managers, and business analysts to understand requirements and define effective test strategies

● Participate in sprint planning, requirement reviews, and technical discussions

● Perform root cause analysis for defects and work with engineering teams for resolution

● Improve automation coverage and reduce manual regression effort

● Ensure test environments, test data, and execution reports are maintained and documented

● Mentor junior QA engineers and promote best practices in automation

● Develop, execute, and maintain comprehensive test plans and test cases for manual and automated testing

● Perform functional, regression, performance, and security testing to ensure software quality

● Design and develop automated test scripts using tools such as Selenium, Appium, or similar frameworks

● Identify, document, and track software defects, working closely with development teams for resolution

● Ensure test coverage by working closely with developers, product managers, and other stakeholders

● Establish and maintain continuous integration (CI) and continuous deployment (CD) pipelines for test automation

● Conduct API testing using tools like Postman or RestAssured

● Collaborate with cross-functional teams to enhance the overall quality of the product

● Stay up to date with the latest industry trends and best practices in QA methodologies and automation frameworks


Requirements:

● 5 to 7 years of experience in QA automation

● Strong hands-on experience with Java and Selenium WebDriver

● Experience in building or enhancing automation frameworks from scratch

● Good understanding of TestNG or JUnit

● Experience with Maven or Gradle

● Familiarity with CI CD tools such as Jenkins, GitHub Actions, or similar

● Strong understanding of Agile Scrum methodology

● Experience with API testing tools such as Rest Assured or Postman is a plus

● Knowledge of version control systems like Git

● Strong analytical and problem-solving skills

● Strong understanding of software testing life cycle (STLC) and defect lifecycle management

● Experience with version control systems (e.g., Git)

● Relevant certifications in software testing (e.g., ISTQB) are desirable but not required

● Solid understanding of software testing principles, methodologies, and techniques

● Excellent analytical and problem-solving skills

● Strong attention to detail and a commitment to delivering high-quality software

● Good communication and collaboration skills, with the ability to work effectively in a team environment


Good to Have:

● Experience with performance testing tools

● Exposure to cloud platforms such as AWS or Azure

● Knowledge of containerization tools like Docker

● Experience in BDD frameworks such as Cucumber.


Why Join Us?

● A collaborative and learning-driven environment

● Exposure to AI and software engineering innovations

● Excellent work ethic and culture


If you're passionate about technology and want to work on impactful projects, we'd love to hear from you!

Read more
NonStop io Technologies Pvt Ltd
Kalyani Wadnere
Posted by Kalyani Wadnere
Pune
6 - 10 yrs
Best in industry
skill iconSpring Boot
Spring MVC
MVC Framework
skill iconJava
Hibernate (Java)
+10 more

Company Description

NonStop io Technologies, founded in August 2015, is a Bespoke Engineering Studio specializing in Product Development. With over 80 satisfied clients worldwide, we serve startups and enterprises across prominent technology hubs, including San Francisco, New York, Houston, Seattle, London, Pune, and Tokyo. Our experienced team brings over 10 years of expertise in building web and mobile products across multiple industries. Our work is grounded in empathy, creativity, collaboration, and clean code, striving to build products that matter and foster an environment of accountability and collaboration.


Role Description

This is a full-time hybrid role for a Java Software Engineer, based in Pune. The Java Software Engineer will be responsible for designing, developing, and maintaining software applications. Key responsibilities include working with microservices architecture, implementing and managing the Spring Framework, and programming in Java. Collaboration with cross-functional teams to define, design, and ship new features is also a key aspect of this role.


Responsibilities:

● Develop and Maintain: Write clean, efficient, and maintainable code for Java-based applications 

● Collaborate: Work with cross-functional teams to gather requirements and translate them into technical solutions 

● Code Reviews: Participate in code reviews to maintain high-quality standards 

● Troubleshooting: Debug and resolve application issues in a timely manner 

● Testing: Develop and execute unit and integration tests to ensure software reliability

● Optimize: Identify and address performance bottlenecks to enhance application performance 


Qualifications & Skills:

● Strong knowledge of Java, Spring Framework (Spring Boot, Spring MVC), and Hibernate/JPA 

● Familiarity with RESTful APIs and web services 

● Proficiency in working with relational databases like MySQL or PostgreSQL 

● Practical experience with AWS cloud services and building scalable, microservices-based architectures

● Experience with build tools like Maven or Gradle 

● Understanding of version control systems, especially Git 

● Strong understanding of object-oriented programming principles and design patterns 

● Familiarity with automated testing frameworks and methodologies 

● Excellent problem-solving skills and attention to detail 

● Strong communication skills and ability to work effectively in a collaborative team environment 


Why Join Us? 

● Opportunity to work on cutting-edge technology products 

● A collaborative and learning-driven environment 

● Exposure to AI and software engineering innovations 

● Excellent work ethic and culture 


If you're passionate about technology and want to work on impactful projects, we'd love to hear from you

Read more
NonStop io Technologies Pvt Ltd
Kalyani Wadnere
Posted by Kalyani Wadnere
Pune
4 - 7 yrs
Best in industry
DevOps
skill iconAmazon Web Services (AWS)
Terraform
Windows Azure
Google Cloud Platform (GCP)
+9 more

About NonStop io Technologies:

NonStop io Technologies is a value-driven company with a strong focus on process-oriented software engineering. We specialize in Product Development and have a decade's worth of experience in building web and mobile applications across various domains. NonStop io Technologies follows core principles that guide its operations and believes in staying invested in a product's vision for the long term. We are a small but proud group of individuals who believe in the 'givers gain' philosophy and strive to provide value in order to seek value. We are committed to and specialize in building cutting-edge technology products and serving as trusted technology partners for startups and enterprises. We pride ourselves on fostering innovation, learning, and community engagement. Join us to work on impactful projects in a collaborative and vibrant environment.


Brief Description:

We are looking for a skilled and proactive DevOps Engineer to join our growing engineering team. The ideal candidate will have hands-on experience in building, automating, and managing scalable infrastructure and CI CD pipelines. You will work closely with development, QA, and product teams to ensure reliable deployments, performance, and system security.


Roles and Responsibilities:

● Design, implement, and manage CI CD pipelines for multiple environments

● Automate infrastructure provisioning using Infrastructure as Code tools

● Manage and optimize cloud infrastructure on AWS, Azure, or GCP

● Monitor system performance, availability, and security

● Implement logging, monitoring, and alerting solutions

● Collaborate with development teams to streamline release processes

● Troubleshoot production issues and ensure high availability

● Implement containerization and orchestration solutions such as Docker and Kubernetes

● Enforce DevOps best practices across the engineering lifecycle

● Ensure security compliance and data protection standards are maintained


Requirements:

● 4 to 7 years of experience in DevOps or Site Reliability Engineering

● Strong experience with cloud platforms such as AWS, Azure, or GCP - Relevant Certifications will be a great advantage

● Hands-on experience with CI CD tools like Jenkins, GitHub Actions, GitLab CI, or Azure DevOps

● Experience working in microservices architecture

● Exposure to DevSecOps practices

● Experience in cost optimization and performance tuning in cloud environments

● Experience with Infrastructure as Code tools such as Terraform, CloudFormation, or ARM

● Strong knowledge of containerization using Docker

● Experience with Kubernetes in production environments

● Good understanding of Linux systems and shell scripting

● Experience with monitoring tools such as Prometheus, Grafana, ELK, or Datadog

● Strong troubleshooting and debugging skills

● Understanding of networking concepts and security best practices


Why Join Us?

● Opportunity to work on a cutting-edge healthcare product

● A collaborative and learning-driven environment

● Exposure to AI and software engineering innovations

● Excellent work ethic and culture


If you're passionate about technology and want to work on impactful projects, we'd love to hear from you!

Read more
StarApps Studio

at StarApps Studio

2 candid answers
4 products
Shivani Kawade
Posted by Shivani Kawade
Pune
4 - 8 yrs
₹15L - ₹30L / yr
skill iconReact.js
TypeScript
skill iconJavascript
skill iconHTML/CSS
skill iconRuby on Rails (ROR)
+9 more

StarApps Studio is a product-driven SaaS company building Shopify apps that power thousands of online stores. We’ve developed 6 highly-rated Shopify apps (averaging 4.9★) used by 30,000+ Shopify merchants worldwide, including over 1,000 Shopify Plus stores. In just a few years, our bootstrapped team grew from $5.5M to $10M in Annual Recurring Revenue (ARR) by obsessing over quality and merchant success. We’re a tight-knit, 20-person team based in Baner, Pune, on a mission to help e-commerce brands create world-class shopping experiences.


Role Overview

We are looking for a Full Stack Developer who will own features end-to-end with an emphasis on backend excellence. In this role, you will design and optimize complex data models and API architectures in Ruby on Rails, implement robust background job queues (e.g. delayed_job) for heavy workloads, and perform rigorous performance tuning to ensure our systems scale. On the frontend, you'll build and integrate React components to deliver complete, user-friendly features. This is a role for someone who loves tackling deep technical challenges in the backend while also crafting intuitive user interfaces – an opportunity to leverage your backend expertise while driving full-stack product ownership.


Key Responsibilities

  • Architect & Optimize Backend: Design scalable database schemas and efficient data models. Develop high-performance RESTful APIs and services in Ruby on Rails, ensuring clean, maintainable code and great performance.
  • Backend API Development: Design, implement, and maintain robust backend services and RESTful APIs in Ruby on Rails to support new features and internal tools.
  • End-to-End Performance Tuning: Optimize application performance across the stack – from minimizing frontend load times to improving backend query efficiency – for our high-traffic, data-intensive apps.
  • Collaboration & Agile Delivery: Work closely with designers, product managers, and QA to translate requirements into technical solutions. Participate in sprint planning, code reviews, and daily deployments to ship features continuously and reliably.
  • Quality & Maintenance: Write clean, maintainable code with appropriate test coverage (unit and integration tests) to ensure reliability. Monitor, debug, and resolve issues in production, and continually refactor and improve existing code for stability and performance


What We’re Looking For (Requirements)

  • 4–8 Years Experience: Proven experience as a software developer in a product company (experience in e-commerce or SaaS is highly preferred). You have built real products used by actual customers at scale.
  • Ruby on Rails Expertise: Strong command of Ruby on Rails. Experience designing RESTful APIs, working with MVC architecture, and using Rails best practices. You should understand how to structure large Rails applications for maintainability.
  • Backend Proficiency: Comfortable building server-side applications and APIs with Ruby on Rails. You can implement business logic, integrate with databases, and create RESTful endpoints (bonus if you’ve worked with GraphQL or other backend frameworks).
  • Database Skills: Proficiency with PostgreSQL (or similar RDBMS). Capable of writing complex SQL queries, optimizing queries/indexes, and designing efficient relational schemas. Familiarity with Redis or caching strategies is a plus.
  • Front-End Proficiency: Comfortable building user interfaces with React and modern JavaScript/TypeScript. Able to implement frontend components that consume APIs and provide a smooth user experience.
  • System Design & Quality: Solid understanding of web application architecture, performance tuning, and scalability concerns. Experience with profiling, benchmarking, and optimizing web applications. Commitment to writing clean, maintainable code with proper tests.
  • Product Mindset: You care about the why behind the features. You are comfortable digging into requirements, questioning assumptions, and ensuring that we build solutions that truly solve merchant problems.
  • Adaptability & Collaboration: Excellent problem-solving skills, communication, and ability to work in a fast-paced, collaborative environment. You are a self-starter who can take ownership of tasks and drive them to completion, but also know when to ask for help.


Tech Stack

  • Frontend: React, TypeScript/JavaScript, HTML5, CSS3 (Tailwind/Bootstrap), modern build tools (Webpack, Babel).
  • Backend: Ruby on Rails (REST APIs, background jobs), some services in Python.
  • Database: PostgreSQL.
  • Cloud & DevOps: Amazon Web Services (EC2, S3, RDS, CloudFront), Docker, CI/CD for daily deployments.
  • Tools: Git (GitHub), Agile issue tracking (JIRA/Trello), and a keen use of automated testing.


(Don’t worry if you aren’t familiar with every item – we value willingness to learn. This is our current stack, and we continually adopt new technologies that improve our products.)


Why Join Us

  • High Impact & Ownership: Your work will directly enhance the shopping experience of 50M+ shoppers daily. At StarApps, developers deploy code daily and see the immediate impact on thousands of merchants – you’ll own projects end-to-end and build features that matter.
  • Fast-Growing, Profitable Startup: Join a bootstrapped, profitable company on an exciting growth trajectory (from $4M to $10M ARR). There’s no bureaucracy here – just a passionate team obsessed with product quality and merchant happiness. You’ll be part of our core team as we scale, with ample opportunities to grow into leadership roles.
  • Cutting-Edge Tech & Challenges: Work with modern technologies (React, Rails, AWS) on performance-intensive applications. Tackle complex challenges in scaling, optimization, and UX for a global user base – continuously sharpen your skills in a supportive, learning-focused environment.
  • Collaborative Culture: We are a small 25-person team that operates like a close-knit family. You’ll work side by side with experienced founders and a talented team that values innovation, humility, and continuous improvement. Our culture is open, empathetic, and growth-oriented – every voice is heard, and every team member plays a crucial role in our success.


Growth & Benefits: We invest in our team’s growth. Expect a competitive salary, performance bonuses, and whatever tools you need to do your best work. We sponsor professional development (courses, conferences, books) and encourage knowledge-sharing. You’ll enjoy a flexible leave policy, team off-sites, and the excitement of building a global product from our new office in Baner, Pune.

Read more
Cambridge Wealth (Baker Street Fintech)
Sangeeta Bhagwat
Posted by Sangeeta Bhagwat
Pune
3 - 5 yrs
₹10L - ₹12L / yr
skill iconPython
SQL
skill iconAmazon Web Services (AWS)
skill iconPostgreSQL
pandas
+9 more


Department

Product & Technology

Location

On-site | Prabhat Road, Pune

Experience

3-5 Years in a Data Engineering or Analytics Role

Domain

Fintech / Wealth Management — non-negotiable

Compensation

11-12 LPA Fixed + Performance Bonus

Growth

Title upgrade + salary revision at 12–18 months for strong performers


Why this role is different from most Data Engineer postings

You will work directly with the founding team on a live wealth management platform used by HNI and NRI clients. You will not spend years in a queue waiting to matter your work ships to production, your analysis influences product decisions, and you will guide junior teammates from day one. If you perform, a raise and title upgrade are on the table within 1218 months. This is the kind of early-team role that defines careers.


About Cambridge Wealth

Cambridge Wealth is a fast-growing, award-winning Financial Services and Fintech firm obsessed with quality and exceptional client service. We serve a high-profile clientele NRI, Mass Affluent, HNI, and ultra-HNI professionals and have received multiple awards from major Mutual Fund houses and BSE. We are past the zero-to-one stage and now focused on scaling our features and intelligence layer. You will be joining at exactly the right time.


What You Will Be Doing

This is a central, hands-on data engineering role at the intersection of financial analytics and applied ML. You will own the data pipelines and analytical models that power investment insights for wealth management clients transforming transaction data and portfolio information into measurable, actionable intelligence.

We are not looking for someone who just keeps the lights on. We want someone who looks at a working system and immediately sees how to make it 10x faster, cleaner, and smarter using AI and automation wherever possible.


Key Responsibilities:


Data Engineering & Pipelines

  • Build and optimize PostgreSQL-based pipelines to process large volumes of investment transaction data.
  • Design and maintain database schemas, foreign tables, and analytical structures for performance at scale.
  • Write advanced SQL — window functions, stored procedures, query optimization, index design.
  • Build Python automation scripts for data ingestion, transformation, and scheduled pipeline runs.
  • Monitor AWS RDS workloads and troubleshoot performance issues proactively.


Financial Analytics & Modelling

  • Develop analytical frameworks to evaluate client portfolios against benchmarks and category averages.
  • Build data models covering mutual fund schemes, SIPs, redemptions, switches, and transfer lifecycles.
  • Create materialized views and derived tables optimized for dashboards and internal reporting tools.
  • Analyse client transaction history to surface patterns in investment behaviour and financial discipline.


Applied ML & AI-Driven Development

  • Use Python (Pandas, NumPy, Scikit-learn) for trend analysis, forecasting, and predictive modelling.
  • Implement classification or regression models to support financial pattern detection.
  • Use AI tools — LLMs, Copilots — to accelerate ETL development, code quality, and data cleaning.
  • Identify opportunities to automate repetitive data tasks and advocate for smarter tooling.


Data Quality & Governance

  • Own data integrity end-to-end in a live, high-stakes financial environment.
  • Build and maintain validation and cleaning protocols across all financial datasets.
  • Maintain Excel models, Power Query workflows, and structured reporting outputs.


Collaboration & Junior Mentorship

  • Work directly with Product, Investment Research, and Wealth Advisory teams.
  • Translate open-ended business questions into structured queries and measurable outputs.
  • Guide 1–2 junior trainees — review their work, set code quality standards, and help them grow.
  • Present findings clearly to non-technical stakeholders — no jargon, just clarity.


Skills — What We Need vs. What Helps

Skill / Tool

Requirement


Must-Haves:

SQL & PostgreSQL (window functions, stored procedures, optimization)

Python — Pandas, NumPy for data processing and automation

ML fundamentals — classification or regression (Scikit-learn)

AWS RDS or equivalent cloud database experience

Financial domain knowledge — mutual funds, SIPs, portfolio concepts

Python data visualization — Matplotlib, Seaborn, or Plotly

Strong Advantage

Excel — Power Query, advanced modelling

Materialized views, query planning, index optimization

Experience with BI/dashboard tools

Good to Have

NoSQL databases

Prior fintech or wealth management startup experience


Financial Domain — Non-Negotiable

This is a wealth management platform. You must come in with a working understanding of:

  • Mutual fund structures, scheme types, and NAV-based transactions
  • Investment lifecycle — SIPs, Lump Sum, Redemptions, Switches, and STPs
  • Portfolio allocation and benchmarking against indices (e.g. Nifty 50, category averages)
  • How HNI/NRI clients interact with financial products differently from retail investors

You do not need to be a CFA. But if mutual funds and portfolio analytics are completely new territory, this role is not the right fit right now.


The Culture Fit — Read This Carefully

We are a small, fast-moving team. This is not a place where you wait for a ticket to arrive in your queue. The right person for this role:

  • Has worked at a small startup before and is used to wearing multiple hats
  • Finds broken or slow data systems genuinely irritating and fixes them without being asked
  • Reaches for Python or an LLM when there is a repetitive task — automating is instinctive
  • Is comfortable saying 'I don't know but I'll find out' and follows through independently
  • Wants visibility and ownership, not just a well-defined job description
  • Is looking for a role where strong performance is directly visible and rewarded


Growth Path — What Happens If You Perform

This is not a vague 'growth opportunity' pitch.

If you hit the bar in your first 12–18 months, you will receive a salary revision and a title upgrade to Senior Data Engineer or Lead Data Engineer depending on team expansion. As we scale our Data and AI team, this role is the natural stepping stone to a team lead position. You will also gain direct exposure to founding-team decision-making — the kind of access that is hard to get at larger companies.


Preferred Background

  • 2–4 years in a data engineering or analytics role at a startup or small Fintech
  • Experience in a live product environment where data errors have real consequences
  • Exposure to portfolio analytics, investment research, or wealth management platforms
  • Has mentored or reviewed code for at least one junior team member


Hiring Process

We respect your time. The process is direct and moves fast.

  • Screening Questions — 5 minutes online
  • Online Challenge — MCQ(Data, SQL, AWS, etc), and one applied ML or analytics problem, Communication Skills and Personality (focused, not trick questions)
  • People Round — 30-minute video call, culture and communication
  • Technical Deep-Dive — 1 hour in person, live financial data problems and your past work
  • Founder's Interview — 1 hour in person, growth conversation and mutual fit
  • Offer & Background Verification


Read more
TVARIT GmbH

at TVARIT GmbH

2 candid answers
DrSoumya Sahadevan
Posted by DrSoumya Sahadevan
Pune
7 - 15 yrs
₹20L - ₹30L / yr
skill iconAmazon Web Services (AWS)
Windows Azure
Google Cloud Platform (GCP)
PySpark
databricks
+2 more

About TVARIT

TVARIT GmbH specializes in developing and delivering cutting-edge artificial intelligence (AI) solutions for the metal industry, including steel, aluminum, copper, cast iron, and more. Our software products empower customers to make intelligent, data-driven decisions, driving advancements in Predictive Quality (PsQ), Predictive Maintenance (PdM), and Energy Consumption Reduction (PsE), etc. With a strong portfolio of renowned reference customers, state-of-the-art technology, a talented research team from prestigious universities, and recognition through esteemed awards such as the EU Horizon 2020 AI Prize, TVARIT is recognized as one of the most innovative AI companies in Germany and Europe. We are seeking a self-motivated individual with a positive "can-do" attitude and excellent oral and written communication skills in English to join our team.


Job Description: We are looking for a Senior Data Engineer with strong expertise in Azure Databricks, PySpark, and distributed computing to develop and optimize scalable ETL pipelines for manufacturing analytics. The role involves working with high-frequency industrial data to enable real-time and batch data processing.


Key Responsibilities · Build scalable real-time and batch processing workflows using Azure Databricks, PySpark, and Apache Spark.

· Perform data pre-processing, including cleaning, transformation, deduplication, normalization, encoding, and scaling to ensure high-quality input for downstream analytics.

· Design and maintain cloud-based data architectures, including data lakes, lakehouses, and warehouses, following Medallion Architecture.

· Deploy and optimize data solutions on Azure (preferred), AWS, or GCP with a focus on performance, security, and scalability.

· Develop and optimize ETL/ELT pipelines for structured and unstructured data from IoT, MES, SCADA, LIMS, and ERP systems. · Automate data workflows using CI/CD and DevOps best practices, ensuring security and compliance with industry standards

· Monitor, troubleshoot, and enhance data pipelines for high availability and reliability.

· Utilize Docker and Kubernetes for scalable data processing.

· Collaborate with automation team, data scientists and engineers to provide clean, structured data for AI/ML models.


Desired Skills and Qualifications · Bachelor’s or Master’s degree in Computer Science, Information Technology, or a related field.

· 7+ years of experience in core data engineering, with a strong focus on cloud platforms such as Azure (preferred), AWS, or GCP · Proficiency in PySpark, Azure Databricks, Python and Apache Spark, etc.

. 2 years of team handling experience.

· Expertise in relational databases (e.g., SQL Server, PostgreSQL), time series databases (e.g. Influx DB), and NoSQL databases (e.g., MongoDB, Cassandra) · Experience in containerization (Docker, Kubernetes).

· Strong analytical and problem-solving skills with attention to detail.

· Good to have MLOps, DevOps including model lifecycle management

· Excellent communication and collaboration skills, with a proven ability to work effectively as a team player.

· Comfortable working in a dynamic, fast-paced startup environment, adapting quickly to changing priorities and responsibilities.

Read more
TVARIT GmbH

at TVARIT GmbH

2 candid answers
DrSoumya Sahadevan
Posted by DrSoumya Sahadevan
Pune
5 - 15 yrs
₹20L - ₹38L / yr
skill iconReact.js
API
AWS CloudFormation
skill iconDjango
skill iconNodeJS (Node.js)
+7 more

Availability: Full time 

Location: Pune, India 

Experience: 5- 6 years

 

Tvarit Solutions Private Limited (wholly owned subsidiary of TVARIT GmbH, Germany). TVARIT provides software to reduce manufacturing waste like scrap, energy, and machine downtime using its patented technology. With its software products, and highly competent team from renowned Universities, TVARIT has gained customer trust across 4 continents within a short span of 3 years. TVARIT is awarded among the top 8 out of 490 AI companies by European Data Incubator, apart from many more awards by the German government and industrial organizations making TVARIT one of the most innovative AI companies in Germany and Europe.  

 

We are looking for a passionate Full Stack Developer Level 2 to join our technology team in Pune Centre. You will be responsible for handling architecting, design, development, testing, leading the software development team and working toward infrastructure development that will support the company’s solutions. You will get an opportunity to work closely on projects that will involve the automation of the manufacturing process. 

 

Key Responsibilities 

· Full Stack Development: Design, develop, and maintain scalable web applications using React with TypeScript for the frontend and Node.js/Python for the backend.

· AI Integration: Collaborate with data scientists and ML engineers to integrate AI/ML models into the SaaS platform, ensuring seamless performance and usability.

· API Development & Optimization: Build and optimize high-performance REST APIs in Node.js and Python (Django, Flask, or FastAPI) to support real-time data processing and analytics.

· Database Engineering: Design, manage, and optimize data storage using relational (PostgreSQL), NoSQL (MongoDB/DynamoDB), graph, and vector databases for handling complex industrial data.

· Cloud-Native Deployment: Deploy, monitor, and manage services in containerized environments using Docker and Kubernetes on Linux-based systems (Ubuntu/Debian).

· System Architecture & Design: Contribute to architectural decisions, leveraging OOPs, microservices, domain-driven design, and design patterns to ensure scalability, security, and maintainability.

· Data Handling & Processing: Work with large-scale manufacturing datasets using Python (pandas) to enable predictive analytics and AI-driven insights.

· Collaboration & Agile Delivery: Partner with cross-functional teams—including product managers, manufacturing domain experts, and AI researchers—to translate business needs into technical solutions.

· Performance & Security: Ensure robust, secure, and high-performance software by implementing best practices in algorithms, data structures, and system design.

· Continuous Improvement: Stay updated on emerging technologies in AI, SaaS, and manufacturing systems to propose innovative solutions that enhance product capability.

 

Must have worked on these technologies.

· 5+ years of experience working with React-Typescript, node.js on a production level

· Python, pandas, High performance REST APIs in node and Python (in Django or Flask or Fast API)

· Databases: Relational DB like PostgreSQL, No SQL DB like Mongo or Dynamo DB, Vector databases, Graphs DBs

· OS: Linux flavor like Ubuntu, Debian

· Source Control and CI/CD

· Software Fundamentals: Excellent command on Algorithms and Data Structures

· Software design and Architecture: OOPs, Design Patterns, Micro Services, monolithic architectures, Domain driven Design

· Containers: Docker and Kubernetes

· Cloud: Fundamentals of AWS like S3 buckets, EC2, IAMs, Security groups


Benefits and Perks:

· Be part of the product which is transforming the manufacturing landscape with AI

· Culture of innovation, creativity, learning, and even failure, we believe in bringing out the best in you.

· Progressive leave policy for effective work-life balance.

· Get mentored by highly qualified internal resource groups and opportunities to avail industry-driven mentorship programs.

· Multicultural peer groups and supportive workplace policies. 

· Work from beaches, hills, mountains, and many more with the yearly workcation program; we believe in mixing elements of vacation and work.

 

 

 

How it's like to work for a Startup?

Working for TVARIT (deep-tech German IT Startup) can offer you a unique blend of innovation, collaboration, and growth opportunities. But it's essential to approach it with a willingness to adapt and thrive in a dynamic environment.

 

If this position sparked your interest, do apply today!

Read more
Wissen Technology

at Wissen Technology

4 recruiters
Monika Sekaran
Posted by Monika Sekaran
Pune
7 - 11 yrs
Best in industry
skill iconJava
skill iconSpring Boot
skill iconAmazon Web Services (AWS)
Microservices
Design patterns
+2 more

Key Responsibilities:

  • Design, develop, and maintain scalable backend applications using Java and Spring Boot.
  • Build and consume RESTful APIs and ensure secure, reliable API integrations.
  • Develop microservices-based architecture and deploy applications in cloud environments.
  • Work with cloud platforms such as AWS/Azure/GCP for application deployment and management.
  • Write clean, maintainable, and efficient code following best practices.
  • Implement CI/CD pipelines and support DevOps practices.
  • Optimize applications for performance, scalability, and reliability.
  • Collaborate with cross-functional teams including frontend, QA, DevOps, and product teams.
  • Participate in code reviews, technical design discussions, and architectural decisions.
  • Troubleshoot production issues and provide timely resolution.

Required Skills & Qualifications:

  • 5–10 years of hands-on experience in Java (Java 8 or above).
  • Strong experience with Spring Boot, Spring MVC, Spring Data, Spring Security.
  • Solid understanding of RESTful API design & development.
  • Experience in microservices architecture.
  • Hands-on experience with at least one cloud platform (AWS / Azure / GCP).
  • Knowledge of containerization tools like Docker and orchestration tools like Kubernetes.
  • Experience with relational and/or NoSQL databases (MySQL, PostgreSQL, MongoDB).
  • Familiarity with CI/CD tools (Jenkins, GitHub Actions, etc.).
  • Strong understanding of Git and version control practices.
  • Good understanding of design patterns and object-oriented programming principles.


Read more
Generative AI Persona platform

Generative AI Persona platform

Agency job
via Peak Hire Solutions by Dhara Thakkar
Pune
6 - 7 yrs
₹15L - ₹20L / yr
skill iconMachine Learning (ML)
skill iconPython
ETL
skill iconData Science
ELT
+6 more

Description

We are currently hiring for the position of Data Scientist/ Senior Machine Learning Engineer (6–7 years’ experience).

 

Please find the detailed Job Description attached for your reference. We are looking for candidates with strong experience in:

  • Machine Learning model development
  • Scalable data pipeline development (ETL/ELT)
  • Python and SQL
  • Cloud platforms such as Azure/AWS/Databricks
  • ML deployment environments (SageMaker, Azure ML, etc.)

 

Kindly note:

  • Location: Pune (Work From Office)
  • Immediate joiners preferred

 

While sharing profiles, please ensure the following details are included:

  • Current CTC
  • Expected CTC
  • Notice Period
  • Current Location
  • Confirmation on Pune WFO comfort

 

Must have skills

Machine Learning - 6 years

Python - 6 years

ETL(Extract, Transform, Load) - 6 years

SQL - 6 years

Azure - 6 years

 

Read more
Digital solutions and services company

Digital solutions and services company

Agency job
via Peak Hire Solutions by Dhara Thakkar
Pune
6 - 7 yrs
₹17L - ₹23L / yr
skill iconMachine Learning (ML)
skill iconPython
ETL
skill iconData Science
SQL
+5 more

Data Scientist or Senior Machine Learning Engineer


We are currently hiring for the position of Data Scientist/ Senior Machine Learning Engineer (6–7 years' experience).


Please find the detailed Job Description attached for your reference.

We are looking for candidates with strong experience in:

  • Machine Learning model development
  • Scalable data pipeline development (ETL/ELT)
  • Python and SQL
  • Cloud platforms such as Azure/AWS/Databricks
  • ML deployment environments (SageMaker, Azure ML, etc.)


Kindly note:

  • Location: Pune (Work from Office)
  • Immediate joiners preferred


While sharing profiles, please ensure the following details are included:

  • Current CTC
  • Expected CTC
  • Notice Period
  • Current Location
  • Confirmation on Pune WFO comfort


Must have Skills

  • Machine Learning - 6 Years
  • Python - 6 Years
  • ETL (Extract, Transform, Load) - 6 Years
  • SQL - 6 Years
  • Azure - 6 Years


Request you to share relevant profiles at the earliest. Looking forward to your support.

Read more
Wissen Technology

at Wissen Technology

4 recruiters
Janane Mohanasankaran
Posted by Janane Mohanasankaran
Mumbai, Pune
7 - 13 yrs
Best in industry
skill iconJava
skill iconSpring Boot
Microservices
RESTful APIs
skill iconAmazon Web Services (AWS)
+1 more

JOB DESCRIPTION:


Location: Pune, Mumbai

Mode of Work : 3 days from Office


DSA(Collections, Hash maps, trees, Linkedlist and Arrays, etc), Core Oops concepts(Multithreading, Multi Processing, Polymorphism, Inheritance etc) Annotations in Spring and Spring boot, Java 8 Vital features, database Optimization, Microsevices and Rest API

  • Design, develop, and maintain low-latency, high-performance enterprise applications using Core Java (Java 5.0 and above).
  • Implement and integrate APIs using Spring Framework and Apache CXF.
  • Build microservices-based architecture for scalable and distributed systems.
  • Collaborate with cross-functional teams for high/low-level design, development, and deployment of software solutions.
  • Optimize performance through efficient multithreading, memory management, and algorithm design.
  • Ensure best coding practices, conduct code reviews, and perform unit/integration testing.
  • Work with RDBMS (preferably Sybase) for backend data integration.
  • Analyze complex business problems and deliver innovative technology solutions in the financial/trading domain.
  • Work in Unix/Linux environments for deployment and troubleshooting.
Read more
NonStop io Technologies Pvt Ltd
Kalyani Wadnere
Posted by Kalyani Wadnere
Pune
8 - 15 yrs
Best in industry
skill iconJavascript
skill iconReact.js
skill iconNodeJS (Node.js)
TypeScript
skill iconAmazon Web Services (AWS)
+6 more

About NonStop io Technologies:

NonStop io Technologies is a value-driven company with a strong focus on process-oriented software engineering. We specialize in Product Development and have a decade's worth of experience in building web and mobile applications across various domains. NonStop io Technologies follows core principles that guide its operations and believes in staying invested in a product's vision for the long term. We are a small but proud group of individuals who believe in the 'givers gain' philosophy and strive to provide value in order to seek value. We are committed to and specialize in building cutting-edge technology products and serving as trusted technology partners for startups and enterprises. We pride ourselves on fostering innovation, learning, and community engagement. Join us to work on impactful projects in a collaborative and vibrant environment.


Brief Description:

We are looking for an Engineering Manager who combines technical depth with leadership strength. This role involves leading one or more product engineering pods, driving architecture decisions, ensuring delivery excellence, and working closely with stakeholders to build scalable web and mobile technology solutions. As a key part of our leadership team, you’ll play a pivotal role in mentoring engineers, improving processes, and fostering a culture of ownership, innovation, and continuous learning.


Roles and Responsibilities:

● Team Management: Lead, coach, and grow a team of 15-20 software engineers, tech leads, and QA engineers

● Technical Leadership: Guide the team in building scalable, high-performance web and mobile applications using modern frameworks and technologies

● Architecture Ownership: Architect robust, secure, and maintainable technology solutions aligned with product goals

● Project Execution: Ensure timely and high-quality delivery of projects by driving engineering best practices, agile processes, and cross-functional collaboration

● Stakeholder Collaboration: Act as a bridge between business stakeholders, product managers, and engineering teams to translate requirements into technology plans

● Culture & Growth: Help build and nurture a culture of technical excellence, accountability, and continuous improvement

● Hiring & Onboarding: Contribute to recruitment efforts, onboarding, and learning & development of team members.


Requirements:

● 8+ years of software development experience, with 2+ years in a technical leadership or engineering manager role

● Proven experience in architecting and building web and mobile applications at scale

● Hands-on knowledge of technologies such as JavaScript/TypeScript, Angular, React, Node.js, .NET, Java, Python, or similar stacks

● Solid understanding of cloud platforms (AWS/Azure/GCP) and DevOps practices

● Strong interpersonal skills with a proven ability to manage stakeholders and lead diverse teams

● Excellent problem-solving, communication, and organizational skills

● Nice to have:

  • Prior experience in working with startups or product-based companies
  • Experience mentoring tech leads and helping shape engineering culture
  • Exposure to AI/ML, data engineering, or platform thinking


Why Join Us?

● Opportunity to work on a cutting-edge healthcare product

● A collaborative and learning-driven environment

● Exposure to AI and software engineering innovations

● Excellent work ethics and culture.



If you're passionate about technology and want to work on impactful projects, we'd love to hear from you!

Read more
NonStop io Technologies Pvt Ltd
Kalyani Wadnere
Posted by Kalyani Wadnere
Pune
2 - 5 yrs
Best in industry
Data Structures
Google Cloud Platform (GCP)
skill iconAmazon Web Services (AWS)
Windows Azure
Scikit-Learn
+3 more

About NonStop io Technologies

NonStop io Technologies is a value-driven company with a strong focus on process-oriented software engineering. We specialize in Product Development and have a decade's worth of experience in building web and mobile applications across various domains. NonStop io Technologies follows core principles that guide its operations and believes in staying invested in a product's vision for the long term. We are a small but proud group of individuals who believe in the 'givers gain' philosophy and strive to provide value in order to seek value. We are committed to and specialize in building cutting-edge technology products and serving as trusted technology partners for startups and enterprises. We pride ourselves on fostering innovation, learning, and community engagement. Join us to work on impactful projects in a collaborative and vibrant environment.


Brief Description:

We're seeking an AI/ML Engineer to join our team. As AI/ML Engineer, you will be responsible for designing, developing, and implementing artificial intelligence (AI) and machine learning (ML) solutions to solve real-world business problems. You will work closely with engineering teams, including software engineers, domain experts, and product managers, to deploy and integrate Applied AI/ML solutions into the products that are being built at NonStop io. Your role will involve researching cutting-edge algorithms and data processing techniques, and implementing scalable solutions to drive innovation and improve the overall user experience.


Responsibilities

● Applied AI/ML engineering; Building engineering solutions on top of the AI/ML tooling available in the industry today. Eg: Engineering APIs around OpenAI

● AI/ML Model Development: Design, develop, and implement machine learning models and algorithms that address specific business challenges, such as natural language processing, computer vision, recommendation systems, anomaly detection, etc.

● Data Preprocessing and Feature Engineering: Cleanse, preprocess, and transform raw data into suitable formats for training and testing AI/ML models. Perform feature engineering to extract relevant features from the data

● Model Training and Evaluation: Train and validate AI/ML models using diverse datasets to achieve optimal performance. Employ appropriate evaluation metrics to assess model accuracy, precision, recall, and other relevant metrics

● Data Visualization: Create clear and insightful data visualizations to aid in understanding data patterns, model behaviour, and performance metrics

● Deployment and Integration: Collaborate with software engineers and DevOps teams to deploy AI/ML models into production environments and integrate them into various applications and systems

● Data Security and Privacy: Ensure compliance with data privacy regulations and implement security measures to protect sensitive information used in AI/ML processes

● Continuous Learning: Stay updated with the latest advancements in AI/ML research, tools, and technologies, and apply them to improve existing models and develop novel solutions

● Documentation: Maintain detailed documentation of the AI/ML development process, including code, models, algorithms, and methodologies for easy understanding and future reference.


Qualifications & Skills

● Bachelor's, Master's, or PhD in Computer Science, Data Science, Machine Learning, or a related field. Advanced degrees or certifications in AI/ML are a plus

● Proven experience as an AI/ML Engineer, Data Scientist, or related role, ideally with a strong portfolio of AI/ML projects

● Proficiency in programming languages commonly used for AI/ML. Preferably Python

● Familiarity with popular AI/ML libraries and frameworks, such as TensorFlow, PyTorch, scikit-learn, etc.

● Familiarity with popular AI/ML Models such as GPT3, GPT4, Llama2, BERT etc.

● Strong understanding of machine learning algorithms, statistics, and data structures

● Experience with data preprocessing, data wrangling, and feature engineering

● Knowledge of deep learning architectures, neural networks, and transfer learning

● Familiarity with cloud platforms and services (e.g., AWS, Azure, Google Cloud) for scalable AI/ML deployment

● Solid understanding of software engineering principles and best practices for writing maintainable and scalable code

● Excellent analytical and problem-solving skills, with the ability to think critically and propose innovative solutions

● Effective communication skills to collaborate with cross-functional teams and present complex technical concepts to non-technical stakeholders

Read more
Global Digital Transformation Solutions Provider

Global Digital Transformation Solutions Provider

Agency job
via Peak Hire Solutions by Dhara Thakkar
Pune, Trivandrum , Thiruvananthapuram
8 - 10 yrs
₹20L - ₹24L / yr
skill iconJava
skill iconPython
API
Google Cloud Platform (GCP)
skill iconAmazon Web Services (AWS)
+13 more

Job Details

Job Title: Lead Software Engineer - Java, Python, API Development

Industry: Global digital transformation solutions provider

Domain - Information technology (IT)

Experience Required: 8-10 years

Employment Type: Full Time

Job Location: Pune & Trivandrum/ Thiruvananthapuram

CTC Range: Best in Industry

 

Job Description

Job Summary

We are seeking a Lead Software Engineer with strong hands-on expertise in Java and Python to design, build, and optimize scalable backend applications and APIs. The ideal candidate will bring deep experience in cloud technologies, large-scale data processing, and leading the design of high-performance, reliable backend systems.

 

Key Responsibilities

  • Design, develop, and maintain backend services and APIs using Java and Python
  • Build and optimize Java-based APIs for large-scale data processing
  • Ensure high performance, scalability, and reliability of backend systems
  • Architect and manage backend services on cloud platforms (AWS, GCP, or Azure)
  • Collaborate with cross-functional teams to deliver production-ready solutions
  • Lead technical design discussions and guide best practices

 

Requirements

  • 8+ years of experience in backend software development
  • Strong proficiency in Java and Python
  • Proven experience building scalable APIs and data-driven applications
  • Hands-on experience with cloud services and distributed systems
  • Solid understanding of databases, microservices, and API performance optimization

 

Nice to Have

  • Experience with Spring Boot, Flask, or FastAPI
  • Familiarity with Docker, Kubernetes, and CI/CD pipelines
  • Exposure to Kafka, Spark, or other big data tools

 

Skills

Java, Python, API Development, Data Processing, AWS Backend

 

Skills: Java, API development, Data Processing, AWS backend, Python,

 

Must-Haves

Java (8+ years), Python (8+ years), API Development (8+ years), Cloud Services (AWS/GCP/Azure), Database & Microservices

8+ years of experience in backend software development

Strong proficiency in Java and Python

Proven experience building scalable APIs and data-driven applications

Hands-on experience with cloud services and distributed systems

Solid understanding of databases, microservices, and API performance optimization

Mandatory Skills: Java API AND AWS

 

******

Notice period - 0 to 15 days only

Job stability is mandatory

Location: Pune, Trivandrum

Read more
NonStop io Technologies Pvt Ltd
Kalyani Wadnere
Posted by Kalyani Wadnere
Pune
3 - 5 yrs
Best in industry
skill iconReact.js
skill iconAngular (2+)
skill iconVue.js
skill iconPython
skill iconJava
+11 more

About NonStop io Technologies:

NonStop io Technologies is a value-driven company with a strong focus on process-oriented software engineering. We specialize in Product Development and have a decade's worth of experience in building web and mobile applications across various domains. NonStop io Technologies follows core principles that guide its operations and believes in staying invested in a product's vision for the long term. We are a small but proud group of individuals who believe in the 'givers gain' philosophy and strive to provide value in order to seek value. We are committed to and specialize in building cutting-edge technology products and serving as trusted technology partners for startups and enterprises. We pride ourselves on fostering innovation, learning, and community engagement. Join us to work on impactful projects in a collaborative and vibrant environment.


Brief Description:

We are looking for a passionate and experienced Full Stack Engineer to join our engineering team. The ideal candidate will have strong experience in both frontend and backend development, with the ability to design, build, and scale high-quality applications. You will collaborate with cross-functional teams to deliver robust and user-centric solutions.


Roles and Responsibilities:

● Design, develop, and maintain scalable web applications

● Build responsive and high-performance user interfaces

● Develop secure and efficient backend services and APIs

● Collaborate with product managers, designers, and QA teams to deliver features

● Write clean, maintainable, and testable code

● Participate in code reviews and contribute to engineering best practices

● Optimize applications for speed, performance, and scalability

● Troubleshoot and resolve production issues

● Contribute to architectural decisions and technical improvements.


Requirements:

● 3 to 5 years of experience in full-stack development

● Strong proficiency in frontend technologies such as React, Angular, or Vue

● Solid experience with backend technologies such as Node.js, .NET, Java, or Python

● Experience in building RESTful APIs and microservices

● Strong understanding of databases such as PostgreSQL, MySQL, MongoDB, or SQL Server

● Experience with version control systems like Git

● Familiarity with CI CD pipelines

● Good understanding of cloud platforms such as AWS, Azure, or GCP

● Strong understanding of software design principles and data structures

● Experience with containerization tools such as Docker

● Knowledge of automated testing frameworks

● Experience working in Agile environments


Why Join Us?

● Opportunity to work on a cutting-edge healthcare product

● A collaborative and learning-driven environment

● Exposure to AI and software engineering innovations

● Excellent work ethic and culture


If you're passionate about technology and want to work on impactful projects, we'd love to hear from you!

Read more
Netra Labs
Poonam Topagi
Posted by Poonam Topagi
Pune
4 - 7 yrs
₹10L - ₹20L / yr
FastAPI
Celery
skill iconRedis
skill iconReact.js
skill iconHTML/CSS
+2 more

Role Description


This is a full-time on-site role for a Python Full Stack Developer located in Pune. You will be responsible for end-to-end development of scalable, AI-driven web applications. Day-to-day tasks involve architecting asynchronous backend services using Python and FastAPI, building dynamic user interfaces with ReactJS, and managing cloud infrastructure on AWS. You will collaborate with data scientists and product teams to integrate AI models into enterprise solutions while ensuring high performance and reliability.


Key Responsibilities


1.      Design and develop high-performance asynchronous APIs using Python and FastAPI.

2.      Build responsive, interactive frontends using ReactJS, HTML, CSS, and Tailwind CSS.

3.      Implement distributed task queues and caching mechanisms using Celery and Redis.

4.      Architect and optimize databases, managing both structured (PostgreSQL) and unstructured (MongoDB) data.

5.      Deploy and manage infrastructure on AWS (EC2, Lambda, S3) and maintain CI/CD pipelines for automated deployment.

6.      Integrate AI/ML models into production workflows and optimize system performance for scalability.

7.      Ensure application security, data privacy, and code quality through best practices and regular testing.


Required Skills & Qualifications


1.      3–5 years of experience in full-stack development with a strong focus on Python.

2.      Proficiency in FastAPI and deep understanding of asynchronous programming (asyncio).

3.      Solid experience with ReactJS, HTML, CSS, JavaScript, and Tailwind CSS.

4.      Hands-on experience with Celery and Redis for background task processing.

5.      Working knowledge of AWS services and containerization tools like Docker.

6.      Proficiency in database management using PostgreSQL and MongoDB.

7.      Experience setting up CI/CD pipelines (Jenkins, GitHub Actions, etc.) and version control (Git).

8.      Strong understanding of RESTful API design, microservices, and security best practices.

Read more
A global engineering and technology group

A global engineering and technology group

Agency job
via Infinium Associate by Toshi Srivastava
Delhi, Gurugram, Noida, Ghaziabad, Faridabad, Bengaluru (Bangalore), Hyderabad, Pune, Chennai
7 - 10 yrs
₹15L - ₹30L / yr
Teamcenter
nx
CAD
skill iconAmazon Web Services (AWS)
configuration management

Job Role: Teamcenter Admin

• Teamcenter and CAD (NX) Configuration Management

• Advanced debugging and root-cause analysis beyond L2

• Code fixes and minor defect remediation

• AWS knowledge, which is foundational to our Teamcenter architecture

• Experience supporting weekend and holiday code deployments

• Operational administration (break/fix, handle ticket escalations, problem management

• Support for project activities

• Deployment and code release support

• Hypercare support following deployment, which is expected to onboard approximately 1,000+ additional users

Read more
Global Digital Transformation Solutions Provider

Global Digital Transformation Solutions Provider

Agency job
via Peak Hire Solutions by Dhara Thakkar
Bengaluru (Bangalore), Chennai, Hyderabad, Kochi (Cochin), Noida, Pune, Thiruvananthapuram
7 - 10 yrs
₹21L - ₹30L / yr
Perforce
DevOps
skill iconGit
skill iconGitHub
skill iconPython
+7 more

JOB DETAILS:

* Job Title: Specialist I - DevOps Engineering

* Industry: Global Digital Transformation Solutions Provider

* Salary: Best in Industry

* Experience: 7-10 years

* Location: Bengaluru (Bangalore), Chennai, Hyderabad, Kochi (Cochin), Noida, Pune, Thiruvananthapuram

 

Job Description

Job Summary:

As a DevOps Engineer focused on Perforce to GitHub migration, you will be responsible for executing seamless and large-scale source control migrations. You must be proficient with GitHub Enterprise and Perforce, possess strong scripting skills (Python/Shell), and have a deep understanding of version control concepts.

The ideal candidate is a self-starter, a problem-solver, and thrives on challenges while ensuring smooth transitions with minimal disruption to development workflows.

 

Key Responsibilities:

  • Analyze and prepare Perforce repositories — clean workspaces, merge streams, and remove unnecessary files.
  • Handle large files efficiently using Git Large File Storage (LFS) for files exceeding GitHub’s 100MB size limit.
  • Use git-p4 fusion (Python-based tool) to clone and migrate Perforce repositories incrementally, ensuring data integrity.
  • Define migration scope — determine how much history to migrate and plan the repository structure.
  • Manage branch renaming and repository organization for optimized post-migration workflows.
  • Collaborate with development teams to determine migration points and finalize migration strategies.
  • Troubleshoot issues related to file sizes, Python compatibility, network connectivity, or permissions during migration.

 

Required Qualifications:

  • Strong knowledge of Git/GitHub and preferably Perforce (Helix Core) — understanding of differences, workflows, and integrations.
  • Hands-on experience with P4-Fusion.
  • Familiarity with cloud platforms (AWS, Azure) and containerization technologies (Docker, Kubernetes).
  • Proficiency in migration tools such as git-p4 fusion — installation, configuration, and troubleshooting.
  • Ability to identify and manage large files using Git LFS to meet GitHub repository size limits.
  • Strong scripting skills in Python and Shell for automating migration and restructuring tasks.
  • Experience in planning and executing source control migrations — defining scope, branch mapping, history retention, and permission translation.
  • Familiarity with CI/CD pipeline integration to validate workflows post-migration.
  • Understanding of source code management (SCM) best practices, including version history and repository organization in GitHub.
  • Excellent communication and collaboration skills for cross-team coordination and migration planning.
  • Proven practical experience in repository migration, large file management, and history preservation during Perforce to GitHub transitions.

 

Skills: Github, Kubernetes, Perforce, Perforce (Helix Core), Devops Tools

 

Must-Haves

Git/GitHub (advanced), Perforce (Helix Core) (advanced), Python/Shell scripting (strong), P4-Fusion (hands-on experience), Git LFS (proficient)

Read more
Virtana

at Virtana

3 candid answers
2 recruiters
Krutika Devadiga
Posted by Krutika Devadiga
Pune
5 - 10 yrs
Best in industry
skill iconPython
skill iconKubernetes
skill iconDocker
skill iconAmazon Web Services (AWS)
Google Cloud Platform (GCP)
+3 more

Role Overview:

Challenge convention and work on cutting edge technology that is transforming the way our customers manage their physical, virtual and cloud computing environments. Virtual Instruments seeks highly talented people to join our growing team, where your contributions will impact the development and delivery of our product roadmap. Our award-winning Virtana Platform provides the only real-time, system-wide, enterprise scale solution for providing visibility into performance, health and utilization metrics, translating into improved performance and availability while lowering the total cost of the infrastructure supporting mission-critical applications.


We are seeking an individual with knowledge in Systems Management and/or Systems Monitoring Software and/or Performance Management Software and Solutions with insight into integrated infrastructure platforms like Cisco UCS, infrastructure providers like Nutanix, VMware, EMC & NetApp and public cloud platforms like Google Cloud and AWS to expand the depth and breadth of Virtana Products.


Work Location: Pune/ Chennai


Job Type: Hybrid


Role Responsibilities:

  • The engineer will be primarily responsible for design and development of software solutions for the Virtana Platform
  • Partner and work closely with team leads, architects and engineering managers to design and implement new integrations and solutions for the Virtana Platform.
  • Communicate effectively with people having differing levels of technical knowledge.
  • Work closely with Quality Assurance and DevOps teams assisting with functional and system testing design and deployment
  • Provide customers with complex application support, problem diagnosis and problem resolution

 

Required Qualifications:

  • Minimum of 4+ years of experience in a Web Application centric Client Server Application development environment focused on Systems Management, Systems Monitoring and Performance Management Software.
  • Able to understand and comprehend integrated infrastructure platforms and experience working with one or more data collection technologies like SNMP, REST, OTEL, WMI, WBEM.
  • Minimum of 4 years of development experience with one of these high level languages like Python, Java, GO is required.
  • Bachelor's (B.E, B.Tech) or Master's degree (M.E, M.Tech. MCA) in computer science, Computer Engineering or equivalent
  •  2 years of development experience in public cloud environment using Kubernetes etc (Google Cloud and/or AWS)

 

Desired Qualifications:

  • Prior experience with other virtualization platforms like OpenShift is a plus
  • Prior experience as a contributor to engineering and integration efforts with strong attention to detail and exposure to Open-Source software is a plus
  • Demonstrated ability as a strong technical engineer who can design and code with strong communication skills
  • Firsthand development experience with the development of Systems, Network and performance Management Software and/or Solutions is a plus
  • Ability to use a variety of debugging tools, simulators and test harnesses is a plus

 

About Virtana:

Virtana delivers the industry's only broadest and deepest Observability Platform that allows organizations to monitor infrastructure, de-risk cloud migrations, and reduce cloud costs by 25% or more.

Over 200 Global 2000 enterprise customers, such as AstraZeneca, Dell, Salesforce, Geico, Costco, Nasdaq, and Boeing, have valued Virtana's software solutions for over a decade.

Our modular platform for hybrid IT digital operations includes Infrastructure Performance Monitoring and Management (IPM), Artificial Intelligence for IT Operations (AIOps), Cloud Cost Management (Fin Ops), and Workload Placement Readiness Solutions. Virtana is simplifying the complexity of hybrid IT environments with a single cloud-agnostic platform across all the categories listed above. The $30BIT Operations Management (ITOM) Software market is ripe for disruption, and Virtana is uniquely positioned for success.

Read more
Deqode

at Deqode

1 recruiter
purvisha Bhavsar
Posted by purvisha Bhavsar
Pune, Delhi, Kolkata, Bengaluru (Bangalore), Kochi (Cochin), Hosur, Trivandrum
7 - 9 yrs
₹5.5L - ₹20L / yr
skill icon.NET
skill iconAmazon Web Services (AWS)
skill iconC#
skill iconReact.js
SQL

Job Description -

Profile: .Net Full Stack Lead

Experience Required: 7–12 Years

Location: Pune, Bangalore, Chennai, Coimbatore, Delhi, Hosur, Hyderabad, Kochi, Kolkata, Trivandrum

Work Mode: Hybrid

Shift: Normal Shift

Key Responsibilities:

  • Design, develop, and deploy scalable microservices using .NET Core and C#
  • Build and maintain serverless applications using AWS services (Lambda, SQS, SNS)
  • Develop RESTful APIs and integrate them with front-end applications
  • Work with both SQL and NoSQL databases to optimize data storage and retrieval
  • Implement Entity Framework for efficient database operations and ORM
  • Lead technical discussions and provide architectural guidance to the team
  • Write clean, maintainable, and testable code following best practices
  • Collaborate with cross-functional teams to deliver high-quality solutions
  • Participate in code reviews and mentor junior developers
  • Troubleshoot and resolve production issues in a timely manner

Required Skills & Qualifications:

  • 7–12 years of hands-on experience in .NET development
  • Strong proficiency in .NET Framework.NET Core, and C#
  • Proven expertise with AWS services (Lambda, SQS, SNS)
  • Solid understanding of SQL and NoSQL databases (SQL Server, MongoDB, DynamoDB, etc.)
  • Experience building and deploying Microservices architecture
  • Proficiency in Entity Framework or EF Core
  • Strong knowledge of RESTful API design and development
  • Experience with React or Angular is a good to have
  • Understanding of CI/CD pipelines and DevOps practices
  • Strong debugging, performance optimization, and problem-solving skills
  • Experience with design patterns, SOLID principles, and best coding practices
  • Excellent communication and team leadership skills


Read more
Deqode

at Deqode

1 recruiter
Samiksha Agrawal
Posted by Samiksha Agrawal
Pune
7 - 10 yrs
₹10L - ₹17L / yr
skill icon.NET
skill iconAmazon Web Services (AWS)
skill iconReact.js
skill iconAngular (2+)


Job Title: Dot Net Full Stack Lead

Experience Required:7-12 Years

Location: Pune, Bangalore, Chennai, Coimbatore, Delhi, Hosur, Hyderabad, Kochi, Kolkata, Trivandrum

Job Type: Full-time

About the Role:

We are looking for a skilled .NET Developer with strong AWS cloud experience to join our engineering team. You will be responsible for designing, developing, and maintaining scalable microservices-based applications using .NET technologies and AWS cloud services.

Key Responsibilities:

  • Design, develop, and deploy microservices using .NET Core and C#
  • Build and maintain serverless applications using AWS Lambda, SQS, and SNS
  • Develop RESTful APIs and integrate them with front-end applications
  • Work with both SQL and NoSQL databases to optimize data storage and retrieval
  • Implement Entity Framework for database operations and ORM
  • Write clean, maintainable, and testable code following best practices
  • Collaborate with cross-functional teams to deliver high-quality solutions
  • Participate in code reviews and contribute to technical documentation
  • Troubleshoot and resolve production issues in a timely manner

Mandatory Skills:

  • Strong proficiency in .NET Framework and .NET Core
  • Expertise in C# programming
  • Hands-on experience with AWS services (Lambda, SQS, SNS)
  • Solid understanding of SQL and NoSQL databases
  • Experience building and deploying Microservices architecture
  • Proficiency in Entity Framework or EF Core
  • Knowledge of RESTful API design and development
  • Understanding of CI/CD pipelines and DevOps practices

Good to Have:

  • Experience with React or Angular for full-stack development
  • Knowledge of containerization (Docker, Kubernetes)
  • Familiarity with other AWS services (EC2, S3, DynamoDB, API Gateway)
  • Experience with message queuing and event-driven architecture
  • Understanding of SOLID principles and design patterns
  • Experience with unit testing and test-driven development (TDD)


Read more
Numino Labs
Disha Kamat
Posted by Disha Kamat
Pune
8 - 18 yrs
₹40L - ₹70L / yr
skill iconPython
skill iconReact.js
skill iconNextJs (Next.js)
skill iconAmazon Web Services (AWS)

Numino Labs 

Business: Software product engineering services: Pune, Goa.

Clients: Software product companies in the USA.

Business model: Exclusive teams for working on client products; direct and daily interactions with clients


Client 

Silicon Valley startup in genAI: 45m+ in funding.

Product: B2B SaaS. 

Core IP: Physics AI foundation model for hardware designers with specific focus on semi-conductor chip design.

Customers: World's top chip manufacturers


Responsibilities

  • Team player: Delivers effectively with teams; interpersonal skills, communication skills, risk management skills
  • Technical Leadership: Works with ambiguous requirements, designs solutions, independently drives delivery to customers
  • Hands on coder: Leverages AI to drive implementation across Reactjs, Python, DB, UnitTest, TestAutomation & Cloud Infra & CI/CD Automation.


Requirements

  • Strong computer science fundamentals: data structures & algorithms, networking, RDBMS, and distributed computing
  • 8-15 years of experience on Python Stack: Behave, PyTest, Python Generators & async operations, multithreading, context managers, decorators, descriptors
  • Python frameworks: FastAPI or Flask or DJango or SQLAlchemy
  • Expertise in Microservices, REST/gRPC APIs design, Authentication, Single Sign-on
  • Experience in high performance delivering solutions on Cloud
  • Some experience in FE: Js & Nextjs/ReactJs
  • Some experience in DevOps, Cloud Infra Automation, Test Automation
Read more
DeepIntent

at DeepIntent

2 candid answers
17 recruiters
Amruta Mundale
Posted by Amruta Mundale
Pune
8 - 10 yrs
Best in industry
Technical support
SQL
Apache
Google Cloud Platform (GCP)
skill iconAmazon Web Services (AWS)
+2 more

What You’ll Do:

We are looking for a Staff Operations Engineer based in Pune, India who can master both DeepIntent’s data architectures and pharma research and analytics methodologies to make significant contributions to how health media is analyzed by our clients. This role requires an Engineer who not only understands DBA functions but also how they impact research objectives and can work with researchers and data scientists to achieve impactful results.  

This role will be in the Engineering Operations team and will require integration and partnership with the Engineering Organization. The ideal candidate is a self-starter who is inquisitive who is not afraid to take on and learn from challenges and will constantly seek to improve the facets of the business they manage. The ideal candidate will also need to demonstrate the ability to collaborate and partner with others.  

  • Serve as the Engineering interface between Analytics and Engineering teams.
  • Develop and standardize all interface points for analysts to retrieve and analyze data with a focus on research methodologies and data-based decision-making.
  • Optimize queries and data access efficiencies, serve as an expert in how to most efficiently attain desired data points.
  • Build “mastered” versions of the data for Analytics-specific querying use cases.
  • Establish a formal data practice for the Analytics practice in conjunction with the rest of DeepIntent.
  • Interpret analytics methodology requirements and apply them to data architecture to create standardized queries and operations for use by analytics teams.
  • Implement DataOps practices.
  • Master existing and new Data Pipelines and develop appropriate queries to meet analytics-specific objectives.
  • Collaborate with various business stakeholders, software engineers, machine learning engineers, and analysts.
  • Operate between Engineers and Analysts to unify both practices for analytics insight creation.

Who You Are:

  • 8+ years of experience in Tech Support (Specialised in Monitoring and maintaining Data pipeline).
  • Adept in market research methodologies and using data to deliver representative insights.
  • Inquisitive, curious, understands how to query complicated data sets, move and combine data between databases.
  • Deep SQL experience is a must.
  • Exceptional communication skills with the ability to collaborate and translate between technical and non-technical needs.
  • English Language Fluency and proven success working with teams in the U.S.
  • Experience in designing, developing and operating configurable Data pipelines serving high-volume and velocity data.
  • Experience working with public clouds like GCP/AWS.
  • Good understanding of software engineering, DataOps, and data architecture, Agile and DevOps methodologies.
  • Proficient with SQL, Python or JVM-based language, Bash.
  • Experience with any of Apache open-source projects such as Spark, Druid, Beam, Airflow etc. and big data databases like BigQuery, Clickhouse, etc. 
  • Ability to think big, take bets and innovate, dive deep, hire and develop the best talent, learn and be curious.
  • Experience in debugging UI and Backend issues will be add on.


Read more
IntegriMart

IntegriMart

Agency job
Pune
5 - 8 yrs
₹12L - ₹15L / yr
Manual testing
Automation
skill iconPython
playwright
skill iconAmazon Web Services (AWS)
+2 more

Hope you are doing great!

We have an Urgent opening for a Senior Automation QA professional to join a global life sciences data platform company. Immediate interview slots available.


🔹 Quick Role Overview

  • Role: Senior Automation QA
  • Location: Pune(Hybrid -3 days work from office)
  • Employment Type: Full-Time
  • Experience Required: 5+ Years
  • Interview Process: 2–3 Rounds
  • Qualification: B.E / B.Tech
  • Notice Period : 0-30 Days


📌 Job Description

IntegriChain is the data and business process platform for life sciences manufacturers, delivering visibility into patient access, affordability, and adherence. The platform enables manufacturers to drive gross-to-net savings, ensure channel integrity, and improve patient outcomes.

We are expanding our Engineering team to strengthen our ability to process large volumes of healthcare and pharmaceutical data at enterprise scale.

The Senior Automation QA will be responsible for ensuring software quality by designing, developing, and maintaining automated test frameworks. This role involves close collaboration with engineering and product teams, ownership of test strategy, mentoring junior QA engineers, and driving best practices to improve product reliability and release efficiency.


🎯 Key Responsibilities

  • Hands-on QA across UI, API, and Database testing – both Automation & Manual
  • Analyze requirements, user stories, and technical documents to design detailed test cases and test data
  • Design, build, execute, and maintain automation scripts using BDD (Gherkin), Pytest, and Playwright
  • Own and maintain QA artifacts: Test Strategy, BRD, defect metrics, leakage reports, quality dashboards
  • Work with stakeholders to review and improve testing approaches using data-backed quality metrics
  • Ensure maximum feasible automation coverage in every sprint
  • Perform functional, integration, and regression testing in Agile & DevOps environments
  • Drive Shift-left testing, identifying defects early and ensuring faster closures
  • Contribute to enhancing automation frameworks with minimal guidance
  • Lead and mentor a QA team (up to 5 members)
  • Support continuous improvement initiatives and institutionalize QA best practices
  • Act as a problem-solver and strong team collaborator in a fast-paced environment


🧩 Desired Skills & Competencies

✅ Must-Have:

  • 5+ years of experience in test planning, test case design, test data preparation, automation & manual testing
  • 3+ years of strong UI & API automation experience using Playwright with Python
  • Solid experience in BDD frameworks (Gherkin, Pytest)
  • Strong database testing skills (Postgres / Snowflake / MySQL / RDS)
  • Hands-on experience with Git and Jenkins (DevOps exposure)
  • Working experience with JMeter
  • Experience in Agile methodologies (Scrum / Kanban)
  • Excellent problem-solving, analytical, communication, and stakeholder management skills

👍 Good to Have:

  • Experience testing AWS / Cloud-hosted applications
  • Exposure to ETL processes and BI reporting systems


Read more
Wissen Technology

at Wissen Technology

4 recruiters
Janane Mohanasankaran
Posted by Janane Mohanasankaran
Mumbai, Pune
4 - 9 yrs
Best in industry
skill iconJava
skill iconSpring Boot
Microservices
skill iconAmazon Web Services (AWS)
RESTful APIs

JOB DESCRIPTION:


Location: Pune, Mumbai

Mode of Work : 3 days from Office


DSA(Collections, Hash maps, trees, Linkedlist and Arrays, etc), Core Oops concepts(Multithreading, Multi Processing, Polymorphism, Inheritance etc) Annotations in Spring and Spring boot, Java 8 Vital features, database Optimization, Microsevices and Rest API

  • Design, develop, and maintain low-latency, high-performance enterprise applications using Core Java (Java 5.0 and above).
  • Implement and integrate APIs using Spring Framework and Apache CXF.
  • Build microservices-based architecture for scalable and distributed systems.
  • Collaborate with cross-functional teams for high/low-level design, development, and deployment of software solutions.
  • Optimize performance through efficient multithreading, memory management, and algorithm design.
  • Ensure best coding practices, conduct code reviews, and perform unit/integration testing.
  • Work with RDBMS (preferably Sybase) for backend data integration.
  • Analyze complex business problems and deliver innovative technology solutions in the financial/trading domain.
  • Work in Unix/Linux environments for deployment and troubleshooting.


Read more
NeoGenCode Technologies Pvt Ltd
Pune
3 - 8 yrs
₹10L - ₹25L / yr
skill iconJava
skill iconSpring Boot
RESTful APIs
Hibernate (Java)
JPA
+10 more

Job Title : Java Backend Developer

Experience : 3 – 8 Years

Location : Pune (Onsite) (Pune candidates Only)

Notice Period : Immediate to 15 Days (or serving NP whose LWD is near)


About the Role :

We are seeking an experienced Java Backend Developer with strong hands-on skills in backend microservices development, API design, cloud platforms, observability, and CI/CD.

The ideal candidate will contribute to building scalable, secure, and reliable applications while working closely with cross-functional teams.


Mandatory Skills : Java 8 / Java 17, Spring Boot 3.x, REST APIs, Hibernate / JPA, MySQL, MongoDB, Prometheus / Grafana / Spring Actuators, AWS, Docker, Jenkins / GitHub Actions, GitHub, Windows 7 / Linux.


Key Responsibilities :

  • Design, develop, and maintain backend microservices and REST APIs
  • Implement data persistence using relational and NoSQL databases
  • Ensure performance, scalability, and security of backend systems
  • Integrate observability and monitoring tools for production environments
  • Work within CI/CD pipelines and containerized deployments
  • Collaborate with DevOps, QA, and product teams for feature delivery
  • Troubleshoot, optimize, and improve existing modules and services

Mandatory Skills :

  • Languages & Frameworks : Java 8, Java 17, Spring Boot 3.x, REST APIs, Hibernate, JPA
  • Databases : MySQL, MongoDB
  • Observability : Prometheus, Grafana, Spring Actuators
  • Cloud Technologies : AWS
  • Containerization Tools : Docker
  • CI/CD Tools : Jenkins, GitHub Actions
  • Version Control : GitHub
  • Operating Systems : Windows 7, Linux

Nice to Have :

  • Strong analytical and debugging abilities
  • Experience working in Agile/Scrum environments
  • Good communication and collaborative skills
Read more
NeoGenCode Technologies Pvt Ltd
Ritika Verma
Posted by Ritika Verma
Pune
3 - 5 yrs
₹13L - ₹20L / yr
skill iconJava
skill iconSpring Boot
Hibernate (Java)
skill iconAmazon Web Services (AWS)
Microservices
+2 more

Job Title: Java Developer (Full-time)

Location: Pune (Onsite)

Experience Required: 3+ Years

Working Days: 5 Days (Mon to Fri)

Key Responsibilities:

  • Design, develop, and deploy scalable microservices using Java and Spring Boot.
  •  Implement RESTful APIs and integrate with external systems and databases.
  • Build and manage services on AWS Cloud using components like ECS, Lambda, S3, RDS,
  • and API Gateway.
  • Collaborate with DevOps to integrate CI/CD pipelines for automated builds, tests, and
  • deployments.
  • Ensure application performance, reliability, and security in a cloud-native environment.
  • Participate in code reviews, troubleshooting, and performance optimization.
  • Work closely with architecture, QA, and product teams to deliver high-quality solutions.

Required Skills & Experience:

  • Strong proficiency in Java, Spring Boot, and microservice architecture. 
  • Hands-on experience with AWS Cloud services (ECS, EKS, Lambda, RDS, CloudWatch,etc.).
  •  Knowledge of Docker, Kubernetes, and CI/CD tools (Jenkins, GitLab, or AWS
  • CodePipeline).
  • Experience with REST APIs, JSON, and message brokers (Kafka, RabbitMQ, or SNS/SQS).
  • Proficiency in SQL and experience with relational databases (Oracle, MySQL, or
  • PostgreSQL).
  •  Familiarity with security best practices, monitoring, and logging in the cloud.


Read more
DeepIntent

at DeepIntent

2 candid answers
17 recruiters
Amruta Mundale
Posted by Amruta Mundale
Pune
5 - 8 yrs
Best in industry
skill iconPython
SQL
Spark
airflow
pandas
+6 more

What You’ll Do:

As a Sr. Data Scientist, you will work closely across DeepIntent Data Science teams located in New York, India, and Bosnia. The role will focus on building predictive models, implementing data-driven solutions to maximize ad effectiveness. You will also lead efforts in generating analyses and insights related to the measurement of campaign outcomes, Rx, patient journey, and supporting the evolution of the DeepIntent product suite. Activities in this position include developing and deploying models in production, reading campaign results, analyzing medical claims, clinical, demographic and clickstream data, performing analysis and creating actionable insights, summarizing, and presenting results and recommended actions to internal stakeholders and external clients, as needed.

  • Explore ways to create better predictive models.
  • Analyze medical claims, clinical, demographic and clickstream data to produce and present actionable insights.
  • Explore ways of using inference, statistical, and machine learning techniques to improve the performance of existing algorithms and decision heuristics.
  • Design and deploy new iterations of production-level code.
  • Contribute posts to our upcoming technical blog.

Who You Are:

  • Bachelor’s degree in a STEM field, such as Statistics, Mathematics, Engineering, Biostatistics, Econometrics, Economics, Finance, or Data Science.
  • 5+ years of working experience as a Data Scientist or Researcher in digital marketing, consumer advertisement, telecom, or other areas requiring customer-level predictive analytics.
  • Advanced proficiency in performing statistical analysis in Python, including relevant libraries, is required.
  • Experience working with data processing, transformation and building model pipelines using tools such as Spark, Airflow, and Docker.
  • You have an understanding of the ad-tech ecosystem, digital marketing and advertising data and campaigns or familiarity with the US healthcare patient and provider systems (e.g. medical claims, medications).
  • You have varied and hands-on predictive machine learning experience (deep learning, boosting algorithms, inference…).
  • You are interested in translating complex quantitative results into meaningful findings and interpretable deliverables, and communicating with less technical audiences orally and in writing.
  • You can write production level code, work with Git repositories.
  • Active Kaggle participant.
  • Working experience with SQL.
  • Familiar with medical and healthcare data (medical claims, Rx, preferred).
  • Conversant with cloud technologies such as AWS or Google Cloud.
Read more
AdElement

at AdElement

2 recruiters
Ritisha Nigam
Posted by Ritisha Nigam
Pune
3 - 10 yrs
₹5L - ₹10L / yr
skill iconJava
skill iconAmazon Web Services (AWS)
Spark
Hadoop
Amazon Redshift
+3 more

We are looking for Senior Software Engineers responsible for designing, developing, and maintaining large scale distributed ad technology systems. This would entail working on several different systems, platforms and technologies.Collaborate with various engineering teams to meet a range of technological challenges. You will work with our product team to contribute and influence the roadmap of our products and technologies and also influence and inspire team members.


Experience

  • 3 - 10 Years


Required Skills

  • 3+ years of work experience and a degree in computer science or a similar field
  • Knowledgeable about computer science fundamentals including data structures, algorithms, and coding
  • Enjoy owning projects from creation to completion and wearing multiple hats
  • Product focused mindset
  • Experience building distributed systems capable of handling large volumes of traffic
  • Fluency with Java, Vertex, Redis, Relational Databases
  • Possess good communication skills
  • Enjoy working in a team-oriented environment that values excellence
  • Have a knack for solving very challenging problems
  • (Preferred) Previous experience in advertising technology or gaming apps
  • (Preferred) Hands-on experience with Spark, Kafka or similar open-source software

Responsibilities

  • Creating design and architecture documents
  • Conducting code reviews
  • Collaborate with others in the engineering teams to meet a range of technological challenges
  • Build, Design and Develop large scale advertising technology system capable of handling tens of billions of events daily

Education

  • UG - B.Tech/B.E. - Computers; PG - M.Tech - Computer


What We Offer:

  • Competitive salary and benefits package.
  • Opportunities for professional growth and development.
  • A collaborative and inclusive work environment.


Salary budget upto 50 LPA or hike20% on current ctc

you can text me over linkedin for quick response


Read more
BharatGo

at BharatGo

2 candid answers
1 video
Pravin Adik
Posted by Pravin Adik
Pune
0 - 2 yrs
₹5000 - ₹10000 / mo
skill iconReact.js
skill iconNodeJS (Node.js)
skill iconNextJs (Next.js)
skill iconPostgreSQL
RESTful APIs
+1 more

Role & Responsibilities:

As a Full Stack Developer Intern, you will take on significant responsibilities in the design, development, and maintenance of web applications using Next.js, React.js, Node.js, PostgreSQL, and AWS Cloud services. We seek individuals who are self-motivated, energetic, and capable of delivering high-quality work with minimal supervision.

  • Develop user-friendly web applications using Next.js and React.js.
  • Create and implement RESTful APIs using Node.js.
  • Write high-quality, maintainable code while adhering to best practices in software development.
  • Deliver projects on time while maintaining a strong focus on performance and user experience.
  • Manage data effectively using PostgreSQL databases.
  • Code Quality & Reviews: Maintain code quality standards and conduct regular code reviews to ensure the delivery of high-quality, error-free code.
  • Performance Optimization: Identify and troubleshoot performance bottlenecks to ensure a seamless and lightning-fast platform experience.
  • Bug Fixing & Maintenance: Monitor platform performance and proactively address any issues or bugs to keep the platform running flawlessly.
  • Contribute innovative ideas and solutions during team discussions and brainstorming sessions.
  • Communicate openly and honestly with team members, sharing insights and feedback constructively.
  • Stay updated on emerging technologies and demonstrate a willingness to learn more.


Qualification:

  • Graduate/Post-Graduate with a degree in Computer Science, Software Engineering, or a related field.
  • Proficiency in HTML, CSS, JavaScript, and modern front-end frameworks (specifically Next.js and React.js).
  • Strong knowledge of back-end technologies such as Node.js and Express.js.
  • Experience with relational databases, particularly PostgreSQL.
  • Familiarity with AWS Cloud services is a plus.
  • Excellent problem-solving skills with a proactive approach to challenges.
  • Proven ability to troubleshoot and resolve complex technical issues.
  • Strong communication skills with the confidence to share ideas openly.
  • High energy level and passion for contributing to the company’s success with integrity and honesty.
  • Startup Enthusiast: Embrace the fast-paced and dynamic environment of a startup, driven by a passion for making a positive impact.


Read more
Wissen Technology

at Wissen Technology

4 recruiters
Janane Mohanasankaran
Posted by Janane Mohanasankaran
Bengaluru (Bangalore), Mumbai, Pune
4 - 7 yrs
Best in industry
skill iconPython
pandas
NumPy
SQL
skill iconHTML/CSS
+4 more

Specific Knowledge/Skills


  1. 4-6 years of experience
  2. Proficiency in Python programming.
  3. Basic knowledge of front-end development.
  4. Basic knowledge of Data manipulation and analysis libraries
  5. Code versioning and collaboration. (Git)
  6. Knowledge for Libraries for extracting data from websites.
  7. Knowledge of SQL and NoSQL databases
  8. Familiarity with RESTful APIs
  9. Familiarity with Cloud (Azure /AWS) technologies
Read more
Global digital transformation solutions provider

Global digital transformation solutions provider

Agency job
via Peak Hire Solutions by Dhara Thakkar
Pune
6 - 9 yrs
₹15L - ₹25L / yr
Data engineering
Apache Kafka
skill iconPython
skill iconAmazon Web Services (AWS)
AWS Lambda
+11 more

Job Details

- Job Title: Lead I - Data Engineering 

- Industry: Global digital transformation solutions provider

- Domain - Information technology (IT)

- Experience Required: 6-9 years

- Employment Type: Full Time

- Job Location: Pune

- CTC Range: Best in Industry


Job Description

Job Title: Senior Data Engineer (Kafka & AWS)

Responsibilities:

  • Develop and maintain real-time data pipelines using Apache Kafka (MSK or Confluent) and AWS services.
  • Configure and manage Kafka connectors, ensuring seamless data flow and integration across systems.
  • Demonstrate strong expertise in the Kafka ecosystem, including producers, consumers, brokers, topics, and schema registry.
  • Design and implement scalable ETL/ELT workflows to efficiently process large volumes of data.
  • Optimize data lake and data warehouse solutions using AWS services such as Lambda, S3, and Glue.
  • Implement robust monitoring, testing, and observability practices to ensure reliability and performance of data platforms.
  • Uphold data security, governance, and compliance standards across all data operations.

 

Requirements:

  • Minimum of 5 years of experience in Data Engineering or related roles.
  • Proven expertise with Apache Kafka and the AWS data stack (MSK, Glue, Lambda, S3, etc.).
  • Proficient in coding with Python, SQL, and Java — with Java strongly preferred.
  • Experience with Infrastructure-as-Code (IaC) tools (e.g., CloudFormation) and CI/CD pipelines.
  • Excellent problem-solvingcommunication, and collaboration skills.
  • Flexibility to write production-quality code in both Python and Java as required.

 

Skills: Aws, Kafka, Python


Must-Haves

Minimum of 5 years of experience in Data Engineering or related roles.

Proven expertise with Apache Kafka and the AWS data stack (MSK, Glue, Lambda, S3, etc.).

Proficient in coding with Python, SQL, and Java — with Java strongly preferred.

Experience with Infrastructure-as-Code (IaC) tools (e.g., CloudFormation) and CI/CD pipelines.

Excellent problem-solving, communication, and collaboration skills.

Flexibility to write production-quality code in both Python and Java as required.

Skills: Aws, Kafka, Python

Notice period - 0 to 15days only

Read more
Global digital transformation solutions provider.

Global digital transformation solutions provider.

Agency job
via Peak Hire Solutions by Dhara Thakkar
Chennai, Kochi (Cochin), Pune, Trivandrum, Thiruvananthapuram
5 - 7 yrs
₹10L - ₹25L / yr
Google Cloud Platform (GCP)
skill iconJenkins
CI/CD
skill iconDocker
skill iconKubernetes
+15 more

Job Description

We are seeking a highly skilled Site Reliability Engineer (SRE) with strong expertise in Google Cloud Platform (GCP) and CI/CD automation to lead cloud infrastructure initiatives. The ideal candidate will design and implement robust CI/CD pipelines, automate deployments, ensure platform reliability, and drive continuous improvement in cloud operations and DevOps practices.


Key Responsibilities:

  • Design, develop, and optimize end-to-end CI/CD pipelines using Jenkins, with a strong focus on Declarative Pipeline syntax.
  • Automate deployment, scaling, and management of applications across various GCP services including GKE, Cloud Run, Compute Engine, Cloud SQL, Cloud Storage, VPC, and Cloud Functions.
  • Collaborate closely with development and DevOps teams to ensure seamless integration of applications into the CI/CD pipeline and GCP environment.
  • Implement and manage monitoring, logging, and ing solutions to maintain visibility, reliability, and performance of cloud infrastructure and applications.
  • Ensure compliance with security best practices and organizational policies across GCP environments.
  • Document processes, configurations, and architectural decisions to maintain operational transparency.
  • Stay updated with the latest GCP services, DevOps, and SRE best practices to enhance infrastructure efficiency and reliability.


Mandatory Skills:

  • Google Cloud Platform (GCP) – Hands-on experience with core GCP compute, networking, and storage services.
  • Jenkins – Expertise in Declarative Pipeline creation and optimization.
  • CI/CD – Strong understanding of automated build, test, and deployment workflows.
  • Solid understanding of SRE principles including automation, scalability, observability, and system reliability.
  • Familiarity with containerization and orchestration tools (Docker, Kubernetes – GKE).
  • Proficiency in scripting languages such as Shell, Python, or Groovy for automation tasks.


Preferred Skills:

  • Experience with TerraformAnsible, or Cloud Deployment Manager for Infrastructure as Code (IaC).
  • Exposure to monitoring and observability tools like Stackdriver, Prometheus, or Grafana.
  • Knowledge of multi-cloud or hybrid environments (AWS experience is a plus).
  • GCP certification (Professional Cloud DevOps Engineer / Cloud Architect) preferred.


Skills

Gcp, Jenkins, CICD Aws,


Nice to Haves

Experience with Terraform, Ansible, or Cloud Deployment Manager for Infrastructure as Code (IaC).

Exposure to monitoring and observability tools like Stackdriver, Prometheus, or Grafana.

Knowledge of multi-cloud or hybrid environments (AWS experience is a plus).

GCP certification (Professional Cloud DevOps Engineer / Cloud Architect) preferred.

 

******

Notice period - 0 to 15days only

Location – Pune, Trivandrum, Kochi, Chennai

Read more
DeepIntent

at DeepIntent

2 candid answers
17 recruiters
Amruta Mundale
Posted by Amruta Mundale
Pune
4 - 8 yrs
Best in industry
skill iconJava
SQL
skill iconSpring Boot
Apache
skill iconAmazon Web Services (AWS)
+1 more

What You’ll Do:

  • Setting up formal data practices for the company.
  • Building and running super stable and scalable data architectures.
  • Making it easy for folks to add and use new data with self-service pipelines.
  • Getting DataOps practices in place.
  • Designing, developing, and running data pipelines to help out Products, Analytics, data scientists and machine learning engineers.
  • Creating simple, reliable data storage, ingestion, and transformation solutions that are a breeze to deploy and manage.
  • Writing and Managing reporting API for different products.
  • Implementing different methodologies for different reporting needs.
  • Teaming up with all sorts of people – business folks, other software engineers, machine learning engineers, and analysts.

Who You Are:

  • Bachelor’s degree in engineering (CS / IT) or equivalent degree from a well-known Institute / University.
  • 3.5+ years of experience in building and running data pipelines for tons of data.
  • Experience with public clouds like GCP or AWS.
  • Experience with Apache open-source projects like Spark, Druid, Airflow, and big data databases like BigQuery, Clickhouse.
  • Experience making data architectures that are optimised for both performance and cost.
  • Good grasp of software engineering, DataOps, data architecture, Agile, and DevOps.
  • Proficient in SQL, Java, Spring Boot, Python, and Bash.
  • Good communication skills for working with technical and non-technical people.
  • Someone who thinks big, takes chances, innovates, dives deep, gets things done, hires and develops the best, and is always learning and curious.


Read more
NeoGenCode Technologies Pvt Ltd
Shivank Bhardwaj
Posted by Shivank Bhardwaj
Pune
6 - 8 yrs
₹12L - ₹22L / yr
skill iconNodeJS (Node.js)
skill iconReact.js
skill iconJavascript
skill iconGo Programming (Golang)
Elixir
+10 more


Job Description – Full Stack Developer (React + Node.js)

Experience: 5–8 Years

Location: Pune

Work Mode: WFO

Employment Type: Full-time


About the Role

We are looking for an experienced Full Stack Developer with strong hands-on expertise in React and Node.js to join our engineering team. The ideal candidate should have solid experience building scalable applications, working with production systems, and collaborating in high-performance tech environments.


Key Responsibilities

  • Design, develop, and maintain scalable full-stack applications using React and Node.js.
  • Collaborate with cross-functional teams to define, design, and deliver new features.
  • Write clean, maintainable, and efficient code following OOP/FP and SOLID principles.
  • Work with relational databases such as PostgreSQL or MySQL.
  • Deploy and manage applications in cloud environments (preferably GCP or AWS).
  • Optimize application performance, troubleshoot issues, and ensure high availability in production systems.
  • Utilize containerization tools like Docker for efficient development and deployment workflows.
  • Integrate third-party services and APIs, including AI APIs and tools.
  • Contribute to improving development processes, documentation, and best practices.


Required Skills

  • Strong experience with React.js (frontend).
  • Solid hands-on experience with Node.js (backend).
  • Good understanding of relational databases: PostgreSQL / MySQL.
  • Experience working in production environments and debugging live systems.
  • Strong understanding of OOP or Functional Programming, and clean coding standards.
  • Knowledge of Docker or other containerization tools.
  • Experience with cloud platforms (GCP or AWS).
  • Excellent written and verbal communication skills.


Good to Have

  • Experience with Golang or Elixir.
  • Familiarity with Kubernetes, RabbitMQ, Redis, etc.
  • Contributions to open-source projects.
  • Previous experience working with AI APIs or machine learning tools.


Read more
Global Digital Transformation Solutions Provider

Global Digital Transformation Solutions Provider

Agency job
via Peak Hire Solutions by Dhara Thakkar
Pune, Hyderabad
6 - 10 yrs
₹24L - ₹40L / yr
skill iconMachine Learning (ML)
skill iconAmazon Web Services (AWS)
MLOps
ECS
skill iconKubernetes
+16 more

Core Responsibilities:

  • The MLE will design, build, test, and deploy scalable machine learning systems, optimizing model accuracy and efficiency
  • Model Development: Algorithms and architectures span traditional statistical methods to deep learning along with employing LLMs in modern frameworks.
  • Data Preparation: Prepare, cleanse, and transform data for model training and evaluation.
  • Algorithm Implementation: Implement and optimize machine learning algorithms and statistical models.
  • System Integration: Integrate models into existing systems and workflows.
  • Model Deployment: Deploy models to production environments and monitor performance.
  • Collaboration: Work closely with data scientists, software engineers, and other stakeholders.
  • Continuous Improvement: Identify areas for improvement in model performance and systems.

 

Skills:

  • Programming and Software Engineering: Knowledge of software engineering best practices (version control, testing, CI/CD).
  • Data Engineering: Ability to handle data pipelines, data cleaning, and feature engineering. Proficiency in SQL for data manipulation + Kafka, Chaossearch logs, etc for troubleshooting; Other tech touch points are ScyllaDB (like BigTable), OpenSearch, Neo4J graph
  • Model Deployment and Monitoring: MLOps Experience in deploying ML models to production environments.
  • Knowledge of model monitoring and performance evaluation.

 

Required experience:

  • Amazon SageMaker: Deep understanding of SageMaker's capabilities for building, training, and deploying ML models; understanding of the Sagemaker pipeline with ability to analyze gaps and recommend/implement improvements
  • AWS Cloud Infrastructure: Familiarity with S3, EC2, Lambda and using these services in ML workflows
  • AWS data: Redshift, Glue
  • Containerization and Orchestration: Understanding of Docker and Kubernetes, and their implementation within AWS (EKS, ECS)

 

Skills: Aws, Aws Cloud, Amazon Redshift, Eks

 

Must-Haves

Amazon SageMaker, AWS Cloud Infrastructure (S3, EC2, Lambda), Docker and Kubernetes (EKS, ECS), SQL, AWS data (Redshift, Glue)

Skills : Machine Learning, MLOps, AWS Cloud, Redshift OR Glue, Kubernetes, Sage maker


******

Notice period - 0 to 15 days only 

Location : Pune & Hyderabad only

Read more
Virtana

at Virtana

3 candid answers
2 recruiters
Krutika Devadiga
Posted by Krutika Devadiga
Pune
8 - 13 yrs
Best in industry
skill iconJava
skill iconKubernetes
skill iconAmazon Web Services (AWS)
skill iconSpring Boot
skill iconGo Programming (Golang)
+13 more

Company Overview:

Virtana delivers the industry’s only unified platform for Hybrid Cloud Performance, Capacity and Cost Management. Our platform provides unparalleled, real-time visibility into the performance, utilization, and cost of infrastructure across the hybrid cloud – empowering customers to manage their mission critical applications across physical, virtual, and cloud computing environments. Our SaaS platform allows organizations to easily manage and optimize their spend in the public cloud, assure resources are performing properly through real-time monitoring, and provide the unique ability to plan migrations across the hybrid cloud. 

As we continue to expand our portfolio, we are seeking a highly skilled and hands-on Staff Software Engineer in backend technologies to contribute to the futuristic development of our sophisticated monitoring products.

 

Position Overview:

As a Staff Software Engineer specializing in backend technologies for Storage and Network monitoring in an AI enabled Data center as well as Cloud, you will play a critical role in designing, developing, and delivering high-quality features within aggressive timelines. Your expertise in microservices-based streaming architectures and strong hands-on development skills are essential to solve complex problems related to large-scale data processing. Proficiency in backend technologies such as Java, Python is crucial.



Work Location: Pune


Job Type: Hybrid

 

Key Responsibilities:

  • Hands-on Development: Actively participate in the design, development, and delivery of high-quality features, demonstrating strong hands-on expertise in backend technologies like Java, Python, Go or related languages.
  • Microservices and Streaming Architectures: Design and implement microservices-based streaming architectures to efficiently process and analyze large volumes of data, ensuring real-time insights and optimal performance.
  • Agile Development: Collaborate within an agile development environment to deliver features on aggressive schedules, maintaining a high standard of quality in code, design, and architecture.
  • Feature Ownership: Take ownership of features from inception to deployment, ensuring they meet product requirements and align with the overall product vision.
  • Problem Solving and Optimization: Tackle complex technical challenges related to data processing, storage, and real-time monitoring, and optimize backend systems for high throughput and low latency.
  • Code Reviews and Best Practices: Conduct code reviews, provide constructive feedback, and promote best practices to maintain a high-quality and maintainable codebase.
  • Collaboration and Communication: Work closely with cross-functional teams, including UI/UX designers, product managers, and QA engineers, to ensure smooth integration and alignment with product goals.
  • Documentation: Create and maintain technical documentation, including system architecture, design decisions, and API documentation, to facilitate knowledge sharing and onboarding.


Qualifications:

  • Bachelor's or Master's degree in Computer Science, Software Engineering, or a related field.
  • 8+ years of hands-on experience in backend development, demonstrating expertise in Java, Python or related technologies.
  • Strong domain knowledge in Storage and Networking, with exposure to monitoring technologies and practices.
  • Experience is handling the large data-lakes with purpose-built data stores (Vector databases, no-SQL, Graph, Time-series).
  • Practical knowledge of OO design patterns and Frameworks like Spring, Hibernate.
  • Extensive experience with cloud platforms such as AWS, Azure or GCP and development expertise on Kubernetes, Docker, etc.
  • Solid experience designing and delivering features with high quality on aggressive schedules.
  • Proven experience in microservices-based streaming architectures, particularly in handling large amounts of data for storage and networking monitoring.
  • Familiarity with performance optimization techniques and principles for backend systems.
  • Excellent problem-solving and critical-thinking abilities.
  • Outstanding communication and collaboration skills.


Why Join Us:

  • Opportunity to be a key contributor in the development of a leading performance monitoring company specializing in AI-powered Storage and Network monitoring.
  • Collaborative and innovative work environment.
  • Competitive salary and benefits package.
  • Professional growth and development opportunities.
  • Chance to work on cutting-edge technology and products that make a real impact.


If you are a hands-on technologist with a proven track record of designing and delivering high-quality features on aggressive schedules and possess strong expertise in microservices-based streaming architectures, we invite you to apply and help us redefine the future of performance monitoring.

Read more
Financial Services Company

Financial Services Company

Agency job
via Peak Hire Solutions by Dhara Thakkar
Pune
4 - 8 yrs
₹10L - ₹13L / yr
SQL
databricks
PowerBI
Windows Azure
Data engineering
+9 more

Review Criteria

  • Strong Senior Data Engineer profile
  • 4+ years of hands-on Data Engineering experience
  • Must have experience owning end-to-end data architecture and complex pipelines
  • Must have advanced SQL capability (complex queries, large datasets, optimization)
  • Must have strong Databricks hands-on experience
  • Must be able to architect solutions, troubleshoot complex data issues, and work independently
  • Must have Power BI integration experience
  • CTC has 80% fixed and 20% variable in their ctc structure


Preferred

  • Worked on Call center data, understand nuances of data generated in call centers
  • Experience implementing data governance, quality checks, or lineage frameworks
  • Experience with orchestration tools (Airflow, ADF, Glue Workflows), Python, Delta Lake, Lakehouse architecture


Job Specific Criteria

  • CV Attachment is mandatory
  • Are you Comfortable integrating with Power BI datasets?
  • We have an alternate Saturdays working. Are you comfortable to WFH on 1st and 4th Saturday?


Role & Responsibilities

We are seeking a highly experienced Senior Data Engineer with strong architectural capability, excellent optimisation skills, and deep hands-on experience in modern data platforms. The ideal candidate will have advanced SQL skills, strong expertise in Databricks, and practical experience working across cloud environments such as AWS and Azure. This role requires end-to-end ownership of complex data engineering initiatives, including architecture design, data governance implementation, and performance optimisation. You will collaborate with cross-functional teams to build scalable, secure, and high-quality data solutions.

 

Key Responsibilities-

  • Lead the design and implementation of scalable data architectures, pipelines, and integration frameworks.
  • Develop, optimise, and maintain complex SQL queries, transformations, and Databricks-based data workflows.
  • Architect and deliver high-performance ETL/ELT processes across cloud platforms.
  • Implement and enforce data governance standards, including data quality, lineage, and access control.
  • Partner with analytics, BI (Power BI), and business teams to enable reliable, governed, and high-value data delivery.
  • Optimise large-scale data processing, ensuring efficiency, reliability, and cost-effectiveness.
  • Monitor, troubleshoot, and continuously improve data pipelines and platform performance.
  • Mentor junior engineers and contribute to engineering best practices, standards, and documentation.


Ideal Candidate

  • Proven industry experience as a Senior Data Engineer, with ownership of high-complexity projects.
  • Advanced SQL skills with experience handling large, complex datasets.
  • Strong expertise with Databricks for data engineering workloads.
  • Hands-on experience with major cloud platforms — AWS and Azure.
  • Deep understanding of data architecture, data modelling, and optimisation techniques.
  • Familiarity with BI and reporting environments such as Power BI.
  • Strong analytical and problem-solving abilities with a focus on data quality and governance
  • Proficiency in python or another programming language in a plus.
Read more
Non-Banking Financial Company

Non-Banking Financial Company

Agency job
via Peak Hire Solutions by Dhara Thakkar
Pune
4 - 8 yrs
₹8L - ₹13L / yr
SQL
databricks
PowerBI
Data engineering
Data architecture
+7 more

ROLES AND RESPONSIBILITIES:

We are seeking a highly experienced Senior Data Engineer with strong architectural capability, excellent optimisation skills, and deep hands-on experience in modern data platforms. The ideal candidate will have advanced SQL skills, strong expertise in Databricks, and practical experience working across cloud environments such as AWS and Azure. This role requires end-to-end ownership of complex data engineering initiatives, including architecture design, data governance implementation, and performance optimisation. You will collaborate with cross-functional teams to build scalable, secure, and high-quality data solutions.


Key Responsibilities-

  • Lead the design and implementation of scalable data architectures, pipelines, and integration frameworks.
  • Develop, optimise, and maintain complex SQL queries, transformations, and Databricks-based data workflows.
  • Architect and deliver high-performance ETL/ELT processes across cloud platforms.
  • Implement and enforce data governance standards, including data quality, lineage, and access control.
  • Partner with analytics, BI (Power BI), and business teams to enable reliable, governed, and high-value data delivery.
  • Optimise large-scale data processing, ensuring efficiency, reliability, and cost-effectiveness.
  • Monitor, troubleshoot, and continuously improve data pipelines and platform performance.
  • Mentor junior engineers and contribute to engineering best practices, standards, and documentation.


IDEAL CANDIDATE:

  • Proven industry experience as a Senior Data Engineer, with ownership of high-complexity projects.
  • Advanced SQL skills with experience handling large, complex datasets.
  • Strong expertise with Databricks for data engineering workloads.
  • Hands-on experience with major cloud platforms — AWS and Azure.
  • Deep understanding of data architecture, data modelling, and optimisation techniques.
  • Familiarity with BI and reporting environments such as Power BI.
  • Strong analytical and problem-solving abilities with a focus on data quality and governance
  • Proficiency in python or another programming language in a plus.


PERKS, BENEFITS AND WORK CULTURE:

Our people define our passion and our audacious, incredibly rewarding achievements. The company is one of India’s most diversified Non-banking financial companies, and among Asia’s top 10 Large workplaces. If you have the drive to get ahead, we can help find you an opportunity at any of the 500+ locations we’re present in India.

Read more
Get to hear about interesting companies hiring right now
Company logo
Company logo
Company logo
Company logo
Company logo
Linkedin iconFollow Cutshort
Why apply via Cutshort?
Connect with actual hiring teams and get their fast response. No spam.
Find more jobs
Get to hear about interesting companies hiring right now
Company logo
Company logo
Company logo
Company logo
Company logo
Linkedin iconFollow Cutshort