4+ Data flow Jobs in India
Apply to 4+ Data flow Jobs on CutShort.io. Find your next job, effortlessly. Browse Data flow Jobs and apply today!
Senior Data Engineer
About Pro Football Focus (PFF)
Pro Football Focus (PFF) is a global leader in sports analytics, delivering the most comprehensive football data and insights in the industry. Trusted by all 32 NFL teams and numerous collegiate programs, PFF transforms game film into advanced analytics and decision-support tools that power winning strategies.
With a strong focus on data science, engineering excellence, and innovation, PFF builds scalable data platforms that process massive volumes of structured and unstructured football data — enabling teams, media, and fans to make smarter, data-driven decisions.
Key Responsibilities
- Design and build scalable ETL pipelines for large-scale data transformation and ingestion
- Handle both one-time bulk data migrations and real-time streaming data updates
- Develop and maintain data workflows that ensure reliability, quality, and performance
- Design efficient data storage strategies using relational and columnar databases
- Make architectural decisions on when to use columnar vs relational systems
- Work with and contribute to data lake architectures and data warehousing solutions
- Optimize query performance and data modeling strategies
- Collaborate with backend engineers, data scientists, and product teams
- Contribute to API layers including GraphQL integrations where required
- Ensure best practices around data governance, scalability, and observability
Required Skills & Experience
- 5+ years of relevant experience in Data Engineering
- Strong experience writing and maintaining ETL pipelines
- Hands-on experience with large-scale data transformation and storage
- Solid understanding of:
- Relational databases (e.g., PostgreSQL, MySQL)
- Columnar databases (e.g., Redshift, Snowflake, ClickHouse)
- Knowing when and why to use each
- Strong understanding of Data Lake concepts (S3-based architectures, partitioning, storage layers, metadata management)
- Experience with Data Warehousing concepts (dimensional modeling, star schema, fact/dimension tables)
- Familiarity with GraphQL and data API exposure
- Experience working in AWS ecosystem (S3, Redshift, Glue, Lambda, etc.) preferred
- Strong SQL skills and query optimization knowledge
- Good understanding of performance tuning and scalable system design
Nice to Have
- Experience working with streaming systems (Kafka/Kinesis)
- Experience with infrastructure-as-code
- Exposure to analytics-heavy or sports/media domains
- Experience building or contributing to data lake architecture
Review Criteria
- Strong Data / ETL Test Engineer
- 5+ years of overall experience in Testing/QA
- 3+ years of hands-on end-to-end data testing/ETL testing experience, covering data extraction, transformation, loading validation, reconciliation, working across BI / Analytics / Data Warehouse / e-Governance platforms
- Must have strong understanding and hands-on exposure to Data Warehouse concepts and processes, including fact & dimension tables, data models, data flows, aggregations, and historical data handling.
- Must have experience in Data Migration Testing, including validation of completeness, correctness, reconciliation, and post-migration verification from legacy platforms to upgraded/cloud-based data platforms.
- Must have independently handled test strategy, test planning, test case design, execution, defect management, and regression cycles for ETL and BI testing
- Hands-on experience with ETL tools and SQL-based data validation is mandatory (Working knowledge or hands-on exposure to Redshift and/or Qlik will be considered sufficient)
- Must hold a Bachelor’s degree B.E./B.Tech else should have master's in M.Tech/MCA/M.Sc/MS
- Must demonstrate strong verbal and written communication skills, with the ability to work closely with business stakeholders, data teams, and QA leadership
- Mandatory Location: Candidate must be based within Delhi NCR (100 km radius)
Preferred
- Relevant certifications such as ISTQB or Data Analytics / BI certifications (Power BI, Snowflake, AWS, etc.)
Job Specific Criteria
- CV Attachment is mandatory
- Do you have experience working on Government projects/companies, mention brief about project?
- Do you have experience working on enterprise projects/companies, mention brief about project?
- Please mention the names of 2 key projects you have worked on related to Data Warehouse / ETL / BI testing?
- Do you hold any ISTQB or Data / BI certifications (Power BI, Snowflake, AWS, etc.)?
- Do you have exposure to BI tools such as Qlik?
- Are you willing to relocate to Delhi and why (if not from Delhi)?
- Are you available for a face-to-face round?
Role & Responsibilities
- 5 years’ experience in Data Testing across BI Analytics platforms with at least 2 largescale enterprise Data Warehouse Analytics eGovernance programs
- Proficiency in ETL Data Warehouse and BI report dashboard validation including test planning data reconciliation acceptance criteria definition defect triage and regression cycle management for BI landscapes
- Proficient in analyzing business requirements and data mapping specifications BRDs Data Models Source to Target Mappings User Stories Reports Dashboards to define comprehensive test scenarios and test cases
- Ability to review high level and low-level data models ETL workflows API specifications and business logic implementations to design test strategies ensuring accuracy consistency and performance of data pipelines
- Ability to test and validate the migrated data from old platform to an upgraded platform and ensure the completeness and correctness of migration
- Experience of conducting test of migrated data and defining test scenarios and test cases for the same
- Experience with BI tools like Qlik ETL platforms Data Lake platforms Redshift to support end to end validation
- Exposure to Data Quality Metadata Management and Data Governance frameworks ensuring KPIs metrics and dashboards align with business expectations
Ideal Candidate
- 5 years’ experience in Data Testing across BI Analytics platforms with at least 2 largescale enterprise Data Warehouse Analytics eGovernance programs
- Proficiency in ETL Data Warehouse and BI report dashboard validation including test planning data reconciliation acceptance criteria definition defect triage and regression cycle management for BI landscapes
- Proficient in analyzing business requirements and data mapping specifications BRDs Data Models Source to Target Mappings User Stories Reports Dashboards to define comprehensive test scenarios and test cases
- Ability to review high level and low-level data models ETL workflows API specifications and business logic implementations to design test strategies ensuring accuracy consistency and performance of data pipelines
- Ability to test and validate the migrated data from old platform to an upgraded platform and ensure the completeness and correctness of migration
- Experience of conducting test of migrated data and defining test scenarios and test cases for the same
- Experience with BI tools like Qlik ETL platforms Data Lake platforms Redshift to support end to end validation
- Exposure to Data Quality Metadata Management and Data Governance frameworks ensuring KPIs metrics and dashboards align with business expectations
Duration : Full Time
Location : Vishakhapatnam, Bangalore, Chennai
years of experience : 3+ years
Job Description :
- 3+ Years of working as a Data Engineer with thorough understanding of data frameworks that collect, manage, transform and store data that can derive business insights.
- Strong communications (written and verbal) along with being a good team player.
- 2+ years of experience within the Big Data ecosystem (Hadoop, Sqoop, Hive, Spark, Pig, etc.)
- 2+ years of strong experience with SQL and Python (Data Engineering focused).
- Experience with GCP Data Services such as BigQuery, Dataflow, Dataproc, etc. is an added advantage and preferred.
- Any prior experience in ETL tools such as DataStage, Informatica, DBT, Talend, etc. is an added advantage for the role.
● Create and maintain optimal data pipeline architecture.
● Assemble large, complex data sets that meet functional / non-functional
business requirements.
● Building and optimizing ‘big data’ data pipelines, architectures and data sets.
● Maintain, organize & automate data processes for various use cases.
● Identifying trends, doing follow-up analysis, preparing visualizations.
● Creating daily, weekly and monthly reports of product KPIs.
● Create informative, actionable and repeatable reporting that highlights
relevant business trends and opportunities for improvement.
Required Skills And Experience:
● 2-5 years of work experience in data analytics- including analyzing large data sets.
● BTech in Mathematics/Computer Science
● Strong analytical, quantitative and data interpretation skills.
● Hands-on experience with Python, Apache Spark, Hadoop, NoSQL
databases(MongoDB preferred), Linux is a must.
● Experience building and optimizing ‘big data’ data pipelines, architectures and data sets.
● Experience with Google Cloud Data Analytics Products such as BigQuery, Dataflow, Dataproc etc. (or similar cloud-based platforms).
● Experience working within a Linux computing environment, and use of
command-line tools including knowledge of shell/Python scripting for
automating common tasks.
● Previous experience working at startups and/or in fast-paced environments.
● Previous experience as a data engineer or in a similar role.


