Cutshort logo
Dataflow architecture Jobs in Pune

11+ Dataflow architecture Jobs in Pune | Dataflow architecture Job openings in Pune

Apply to 11+ Dataflow architecture Jobs in Pune on CutShort.io. Explore the latest Dataflow architecture Job opportunities across top companies like Google, Amazon & Adobe.

icon
MNC Pune based IT company

MNC Pune based IT company

Agency job
via Bhs Staffing Solutions Pvt Ltd by Bhagyesh Shinde
Pune
10 - 18 yrs
₹35L - ₹40L / yr
Google Cloud Platform (GCP)
Dataflow architecture
Data migration
Data processing
Big Data
+4 more

CANDIDATE WILL BE DEPLOYED IN A FINANCIAL CAPTIVE ORGANIZATION @ PUNE (KHARADI)

 

Below are the job Details :-

 

Experience 10 to 18 years

 

Mandatory skills –

  • data migration,
  • data flow

The ideal candidate for this role will have the below experience and qualifications:  

  • Experience of building a range of Services in a Cloud Service provider (ideally GCP)  
  • Hands-on design and development of Google Cloud Platform (GCP), across a wide range of GCP services including hands on experience of GCP storage & database technologies. 
  • Hands-on experience in architecting, designing or implementing solutions on GCP, K8s, and other Google technologies. Security and Compliance, e.g. IAM and cloud compliance/auditing/monitoring tools 
  • Desired Skills within the GCP stack - Cloud Run, GKE, Serverless, Cloud Functions, Vision API, DLP, Data Flow, Data Fusion 
  • Prior experience of migrating on-prem applications to cloud environments. Knowledge and hands on experience on Stackdriver, pub-sub, VPC, Subnets, route tables, Load balancers, firewalls both for on premise and the GCP.  
  • Integrate, configure, deploy and manage centrally provided common cloud services (e.g. IAM, networking, logging, Operating systems, Containers.)  
  • Manage SDN in GCP Knowledge and experience of DevOps technologies around Continuous Integration & Delivery in GCP using Jenkins.  
  • Hands on experience of Terraform, Kubernetes, Docker, Stackdriver, Terraform  
  • Programming experience in one or more of the following languages: Python, Ruby, Java, JavaScript, Go, Groovy, Scala  
  • Knowledge or experience in DevOps tooling such as Jenkins, Git, Ansible, Splunk, Jira or Confluence, AppD, Docker, Kubernetes  
  • Act as a consultant and subject matter expert for internal teams to resolve technical deployment obstacles, improve product's vision. Ensure compliance with centrally defined Security 
  • Financial experience is preferred 
  • Ability to learn new technologies and rapidly prototype newer concepts 
  • Top-down thinker, excellent communicator, and great problem solver

 

Exp:- 10  to 18 years

 

Location:- Pune

 

Candidate must have experience in below.

  • GCP Data Platform
  • Data Processing:- Data Flow, Data Prep, Data Fusion
  • Data Storage:- Big Query, Cloud Sql,
  • Pub Sub, GCS Bucket
Read more
Pune
10 - 15 yrs
₹30L - ₹40L / yr
skill iconData Analytics
skill iconGoogle Analytics

Strong Lead – User Research & Analyst profile (behavioural/user/product/ux analytics)

Mandatory (Experience 1): Must have 10+ years of experience in Behavioral Data Analytics, User Research, or Product Insights, driving data-informed decision-making for B2C digital products (web and app).

Mandatory (Experience 2): Must have 6 months+ experience in analyzing user journeys, clickstream, and behavioral data using tools such as Google Analytics, Mixpanel, CleverTap, Firebase, or Amplitude.

Mandatory (Experience 3): Experience in leading cross-functional user research and analytics initiatives in collaboration with Product, Design, Engineering, and Business teams to translate behavioral insights into actionable strategies.

Mandatory (Skills 1): Strong expertise in A/B testing and experimentation, including hypothesis design, execution, statistical validation, and impact interpretation.

Mandatory (Skills 2): Ability to identify behavioral patterns, funnel drop-offs, engagement trends, and user journey anomalies using large datasets and mixed-method analysis.

Mandatory (Skills 3): Hands-on proficiency in SQL, Excel, and data visualization/storytelling tools such as Tableau, Power BI, or Looker for executive reporting and dashboard creation.

Mandatory (Skills 4): Deep understanding of UX principles, customer journey mapping, and product experience design, with experience integrating qualitative and quantitative insights.

Mandatory (Company): B2C product organizations (fintech, e-commerce, edtech, or consumer platforms) with large-scale user datasets and analytics maturity.

Mandatory (Note): Don’t want data analysts; looking for strategic behavioral insight leaders or research-driven analytics professionals focused on user behavior and product decision-making.

Read more
With a global provider of Business Process Management.

With a global provider of Business Process Management.

Agency job
via Jobdost by Saida Pathan
Bengaluru (Bangalore), Mumbai, Pune, Chennai, Nashik, Visakhapatnam, Gurugram
2 - 8 yrs
₹8L - ₹20L / yr
Appian
BPM
Job Description:
Extensive experience in Appian BPM application development
· Knowledge of Appian architecture and its objects best practices
· Participate in analysis, design, and new development of Appian based applications
· Team leadership and provide technical leadership to Scrum teams
· Must be able to multi-task, work in a fast-paced environment and ability to resolve problems faced
by team
· Build applications: interfaces, process flows, expressions, data types, sites, integrations, etc.
· Proficient with SQL queries and with accessing data present in DB tables and views
· Experience in Analysis, Designing process models, Records, Reports, SAIL, forms, gateways, smart
services, integration services and web services
· Experience working with different Appian Object types, query rules, constant rules and expression
rules
Qualifications
· At least 6 years of experience in Implementing BPM solutions using Appian 19.x or higher
· Over 8 years in Implementing IT solutions using BPM or integration technologies
· Certification Mandatory- L1 and L2 a
· Experience in Scrum/Agile methodologies with Enterprise level application development projects
· Good understanding of database concepts and strong working knowledge any one of the major
databases e g Oracle SQL Server MySQL
Additional information
Skills Required
· Appian BPM application development on version 19.x or higher
· Experience of integrations using web services e g XML REST WSDL SOAP API JDBC JMS
· Good leadership skills and the ability to lead a team of software engineers technically
· Experience working in Agile Scrum teams
· Good Communication skills.
Read more
Deqode

at Deqode

1 recruiter
Roshni Maji
Posted by Roshni Maji
Pune, Bengaluru (Bangalore), Gurugram, Chennai, Mumbai
5 - 7 yrs
₹6L - ₹20L / yr
skill iconAmazon Web Services (AWS)
Amazon Redshift
AWS Glue
skill iconPython
PySpark

Position: AWS Data Engineer

Experience: 5 to 7 Years

Location: Bengaluru, Pune, Chennai, Mumbai, Gurugram

Work Mode: Hybrid (3 days work from office per week)

Employment Type: Full-time

About the Role:

We are seeking a highly skilled and motivated AWS Data Engineer with 5–7 years of experience in building and optimizing data pipelines, architectures, and data sets. The ideal candidate will have strong experience with AWS services including Glue, Athena, Redshift, Lambda, DMS, RDS, and CloudFormation. You will be responsible for managing the full data lifecycle from ingestion to transformation and storage, ensuring efficiency and performance.

Key Responsibilities:

  • Design, develop, and optimize scalable ETL pipelines using AWS Glue, Python/PySpark, and SQL.
  • Work extensively with AWS services such as Glue, Athena, Lambda, DMS, RDS, Redshift, CloudFormation, and other serverless technologies.
  • Implement and manage data lake and warehouse solutions using AWS Redshift and S3.
  • Optimize data models and storage for cost-efficiency and performance.
  • Write advanced SQL queries to support complex data analysis and reporting requirements.
  • Collaborate with stakeholders to understand data requirements and translate them into scalable solutions.
  • Ensure high data quality and integrity across platforms and processes.
  • Implement CI/CD pipelines and best practices for infrastructure as code using CloudFormation or similar tools.

Required Skills & Experience:

  • Strong hands-on experience with Python or PySpark for data processing.
  • Deep knowledge of AWS Glue, Athena, Lambda, Redshift, RDS, DMS, and CloudFormation.
  • Proficiency in writing complex SQL queries and optimizing them for performance.
  • Familiarity with serverless architectures and AWS best practices.
  • Experience in designing and maintaining robust data architectures and data lakes.
  • Ability to troubleshoot and resolve data pipeline issues efficiently.
  • Strong communication and stakeholder management skills.


Read more
DataMetica

at DataMetica

1 video
7 recruiters
Sayali Kachi
Posted by Sayali Kachi
Pune
2 - 6 yrs
₹3L - ₹10L / yr
SQL
Linux/Unix
Data Warehouse (DWH)
Shell Scripting

Experience - 2 to 6 Years

 

Work Location - Pune

 

Datametica is looking for talented SQL engineers who would get training & the opportunity to work on Cloud and Big Data Analytics.

 

Mandatory Skills:

  • Strong in SQL development
  • Hands-on at least one scripting language - preferably shell scripting
  • Development experience in Data warehouse projects

 

Opportunities:

  • Selected candidates will be provided learning opportunities on one or more of the following: Google Cloud, AWS, DevOps Tools, Big Data technologies like Hadoop, Pig, Hive, Spark, Sqoop, Flume, and KafkaWould get a chance to be part of the enterprise-grade implementation of Cloud and Big Data systems
  • Will play an active role in setting up the Modern data platform based on Cloud and Big Data
  • Would be part of teams with rich experience in various aspects of distributed systems and computing
Read more
Media.net

at Media.net

21 recruiters
Agency job
via Volks Consulting by SHUBHAM MAGDUM
Remote, Bengaluru (Bangalore), Pune, Mumbai
2 - 6 yrs
₹20L - ₹45L / yr
skill iconJava
Spring
Data Structures
Algorithms
Apache Kafka
+4 more
  •  2 - 6 years of software development experience
  •  Good grasp on programming fundamentals including OOP, Design Patterns and Data Structures
  •  Excellent analytical, logical and problem-solving skills
  • Software Development Engineer
  • Good understanding of complexities involved in designing/developing large scale systems
  • Strong system design skills
  •  Experience in technologies like Elasticsearch, Redis, Kafka etc
  • Good knowledge of relational and NoSQL databases
  • Familiarity with common machine learning algorithms. In-depth knowledge is a plus
  • Experience of working with big data technologies like Hadoop, Spark, Hive is a big plus
  • Ability to understand business requirements and take ownership of the work
  • Exhibit passion and enthusiasm for building and maintaining large scale platforms
Read more
KodeYoga India Pvt Ltd

at KodeYoga India Pvt Ltd

1 recruiter
Manav Prasad
Posted by Manav Prasad
Pune
3 - 8 yrs
₹6L - ₹30L / yr
skill iconJava
J2EE
Struts
Spring
Hibernate (Java)
+6 more

                                                                                   

Who are we looking for?

A passionate developer

What’s non-negotiable?

  • Strong working knowledge of OOPS
  • Functional programming principles
  • Strong believer of Clean Code practices
  • An advocate of TDD, BDD, SOLID, CI/CD, Lean development

    You’ll easily settle in if:
    You come from a strong Java/J2EE background with web application frameworks like Spring Boot or Drop wizard
    You have ample work experience in Caching, Multi-Threading, Asynchronous APIs, Exception Management and use of collections, Mocking, Unit testing tools like Junit & TestNG"
    You are fluent with version control tools like GIT and Bitbucket
     Experience with Continuous Integration, Continuous Deployment, Static Code Analysis, Jenkins and SonarQube
    Willingness to take ownership of the technical solution and ensure technical expectations of deliverables are met.
     Exposure to AWS/Azure cloud and containerization
    Have a good Understanding of Distributed Application Architecture.
    You ‘ll love:
    Awesome opportunity to work with Micro Services architecture getting shipped on Cloud.
    Experience in working with automated build deploy powered by Code Analysis, Automated Tests, Functional and Nonfunctional analysis, Blue Green deployment and much more
     First-hand experience on broader enterprise application concerns like Message Bus, Queues, Caches, Concurrency and Parallelization
Read more
1CH

at 1CH

1 recruiter
Sathish Sukumar
Posted by Sathish Sukumar
Chennai, Bengaluru (Bangalore), Hyderabad, NCR (Delhi | Gurgaon | Noida), Mumbai, Pune
4 - 15 yrs
₹10L - ₹25L / yr
Data engineering
Data engineer
ETL
SSIS
ADF
+3 more
  • Expertise in designing and implementing enterprise scale database (OLTP) and Data warehouse solutions.
  • Hands on experience in implementing Azure SQL Database, Azure SQL Date warehouse (Azure Synapse Analytics) and big data processing using Azure Databricks and Azure HD Insight.
  • Expert in writing T-SQL programming for complex stored procedures, functions, views and query optimization.
  • Should be aware of Database development for both on-premise and SAAS Applications using SQL Server and PostgreSQL.
  • Experience in ETL and ELT implementations using Azure Data Factory V2 and SSIS.
  • Experience and expertise in building machine learning models using Logistic and linear regression, Decision tree  and Random forest Algorithms.
  • PolyBase queries for exporting and importing data into Azure Data Lake.
  • Building data models both tabular and multidimensional using SQL Server data tools.
  • Writing data preparation, cleaning and processing steps using Python, SCALA, and R.
  • Programming experience using python libraries NumPy, Pandas and Matplotlib.
  • Implementing NOSQL databases and writing queries using cypher.
  • Designing end user visualizations using Power BI, QlikView and Tableau.
  • Experience working with all versions of SQL Server 2005/2008/2008R2/2012/2014/2016/2017/2019
  • Experience using the expression languages MDX and DAX.
  • Experience in migrating on-premise SQL server database to Microsoft Azure.
  • Hands on experience in using Azure blob storage, Azure Data Lake Storage Gen1 and Azure Data Lake Storage Gen2.
  • Performance tuning complex SQL queries, hands on experience using SQL Extended events.
  • Data modeling using Power BI for Adhoc reporting.
  • Raw data load automation using T-SQL and SSIS
  • Expert in migrating existing on-premise database to SQL Azure.
  • Experience in using U-SQL for Azure Data Lake Analytics.
  • Hands on experience in generating SSRS reports using MDX.
  • Experience in designing predictive models using Python and SQL Server.
  • Developing machine learning models using Azure Databricks and SQL Server
Read more
Equal Experts

at Equal Experts

13 recruiters
Jenny John
Posted by Jenny John
Pune
8 - 18 yrs
₹18L - ₹40L / yr
RESTful APIs
skill iconJava
Test driven development (TDD)
Spring
Microservices
Role: The role of lead is not a textbook checklist, however, there are technical responsibilities that a team must fulfill and we expect the tech lead to ensure these responsibilities are covered and be able to cover them themselves if needed. We expect tech leads to taking a collaborative approach to leading their team. This is especially important when considering the amount of experience that each of our consultants brings. Given this, we expect you to: ● Respect the other members of your team and recognise you don’t always know best. ● Spot gaps in team capability and figure out how to fix them as a team. ● Be hands on, able and willing to contribute to development, however, don’t expect to be coding all of the time. ● Encourage the team to be proactive, give them responsibility. Responsibilities ● Have a clear understanding of the deployment architecture ● Have a clear understanding of the build pipeline ● Understand how you get changes into production ● Understand how all parts of the system work together ● Facilitate technical communication with other teams, both within your engagement and across other EE clients. ● Actively seek to remove knowledge silos within the team ● Ensure you have a release / branching strategy in place ● Act as the primary point of contact for your team when communicating with other teams ● Ensure there is a technical vision for the team ● Liaise with environment specialists to ensure smooth deployments to production ● Encourage the team to follow good development practices aligned to EE technical values ● Feedback to the delivery lead or engagement manager on the quality of your team (good and bad) ● Recognise team members that have the potential to grow into team leads ● Ensure the use of new technologies or dependencies does not block the team. ● Ensure the team keeps necessary architectural documents up to date ● Keep an eye on the long term consequences of architectural choices, and remind others when necessary ● Build good relationships with your team members. Act as a mentor when required ● Keep the client informed and engaged in the technical side of the project ● Build relationships across your client community Technologies / Experience The successful candidate must have the following experience: ● Worked as the tech lead of a development/delivery team in a large organisation ● Have worked with a variety of different technical architect roles ● Be deeply proficient in at least one programming language ● Be comfortable using other languages and evidence using multiple languages ● Have hands-on experience with some form of configuration management tooling, e.g. Ansible, Chef, Puppet ● Have hands-on experience of at least one continuous integration and continuous delivery technology, e.g. Jenkins, Go, Team City or Bamboo. ● Full stack development experience from user interface through to data persistence ● A strong proponent of XP practices such as TDD ● Working with a delivery team to formulate an automated test strategy ● Worked as part of a number of agile delivery teams and seen a number of different approaches to delivery ● Good appreciation of secure coding practices and end to end system security The following exposure will also be looked on favorably: ● Performing an ‘architect’ role, while retaining hands-on involvement ● Working with cloud hosting platforms such as AWS, Rackspace, Azure etc. ● Infrastructure management technologies such as Cloud Formation or Terraform
Read more
Arc Digitech

at Arc Digitech

1 recruiter
Samrat Biswajit
Posted by Samrat Biswajit
Pune
0 - 3 yrs
₹1L - ₹2L / yr
Graphic Designing
User Experience (UX) Design
skill iconHTML/CSS
Hi, We are looking for someone who can fulfill our needs of graphic designing. It includes designing images for our social media platforms, logos, several campaign flyers, infographics as required. We expect the candidate to be creative and have knowledge of all required software. Thank you.
Read more
Mooshak

at Mooshak

1 recruiter
Anurag Gaur
Posted by Anurag Gaur
Pune
1 - 5 yrs
₹6L - ₹10L / yr
skill iconNodeJS (Node.js)
NOSQL Databases
skill iconJava
Apache Storm
Apache Kafka
+1 more
ABOUT MOOSHAK We're at a point where the urban English-speaking Indian population is almost all online.The next billion Indians online all communicate via Indian languages. Mooshak was created with the singular aim of making the Internet fun and relevant for this large, untapped segment. At Mooshak, we want to connect and engage Indians in their own language. And that presents problems in various domains, from creativity in content creation, to creating a highly scalable platform, to applying techniques in AI and NLP in Indian languages to understand what people are saying and react to what they want. Mooshak is architected to scale. Irrespective of the number of followers, the read time for a feed remains constant. We achieve this by using distributed message queues and a distributed computing engine and some nifty caching! TECHNICAL RESPONSIBILITIES Mooshak’s Tech Stack Java Node.js Mongo DB Redis Apache Kafka & Apache Storm Nginx / Jenkins Server Developer’s Roles and Responsibilities You are expected to know at least 4 of these technologies with the ability to quickly learn the others. You will play the leading role in all stages of server development Architecture Coding Final testing Shipping The APIs are written and the product works fine. You are expected to understand the architecture and enhance product functionality. Sometimes you may be required to double up as the Dev Ops guy should the servers fail or the product not be working as expected. The core APIs are written in Node.js The distributed message queue (Kafka) and compute engine (Storm) are implemented in Java. Understanding of Angular 2 is a big plus as our Web app is built on the same. NON TECHNICAL RESPONSIBILITIES We are a startup. This means that: You will be expected to be someone who comes up with solutions instead of problems. You will be expected to work non stop including weekends if the servers crash. But otherwise we are quite chill! You will be expected to talk to multiple stakeholders customers, designer, client side developer to achieve user and business needs. A high aptitude and a positive attitude are a must You should be comfortable working independently as well as in a team. We are a lean team right now, with you as the only server developer (assisted by the folks who built the platform) JOB LOCATION You would be working out of our office in Pune. You may be required to travel occasionally to Mumbai or Bangalore to interact with some other team members.
Read more
Get to hear about interesting companies hiring right now
Company logo
Company logo
Company logo
Company logo
Company logo
Linkedin iconFollow Cutshort
Why apply via Cutshort?
Connect with actual hiring teams and get their fast response. No spam.
Find more jobs
Get to hear about interesting companies hiring right now
Company logo
Company logo
Company logo
Company logo
Company logo
Linkedin iconFollow Cutshort