50+ Data modeling Jobs in India
Apply to 50+ Data modeling Jobs on CutShort.io. Find your next job, effortlessly. Browse Data modeling Jobs and apply today!


Role Overview
We're looking for experienced Data Engineers who can independently design, build, and manage scalable data platforms. You'll work directly with clients and internal teams to develop robust data pipelines that support analytics, AI/ML, and operational systems.
You’ll also play a mentorship role and help establish strong engineering practices across our data projects.
Key Responsibilities
- Design and develop large-scale, distributed data pipelines (batch and streaming)
- Implement scalable data models, warehouses/lakehouses, and data lakes
- Translate business requirements into technical data solutions
- Optimize data pipelines for performance and reliability
- Ensure code is clean, modular, tested, and documented
- Contribute to architecture, tooling decisions, and platform setup
- Review code/design and mentor junior engineers
Must-Have Skills
- Strong programming skills in Python and advanced SQL
- Solid grasp of ETL/ELT, data modeling (OLTP & OLAP), and stream processing
- Hands-on experience with frameworks like Apache Spark, Flink, etc.
- Experience with orchestration tools like Airflow
- Familiarity with CI/CD pipelines and Git
- Ability to debug and scale data pipelines in production
Preferred Skills
- Experience with cloud platforms (AWS preferred, GCP or Azure also fine)
- Exposure to Databricks, dbt, or similar tools
- Understanding of data governance, quality frameworks, and observability
- Certifications (e.g., AWS Data Analytics, Solutions Architect, Databricks) are a bonus
What We’re Looking For
- Problem-solver with strong analytical skills and attention to detail
- Fast learner who can adapt across tools, tech stacks, and domains
- Comfortable working in fast-paced, client-facing environments
- Willingness to travel within India when required

Data Analytics Lead
Responsibilities:
· Oversee the design, development, and implementation of data analysis solutions to meet business needs.
· Work closely with business stakeholders and the Aviation SME to define data requirements, project scope, and deliverables.
· Drive the design and development of analytics data models and data warehouse designs.
· Develop and maintain data quality standards and procedures.
· Manage and prioritize data analysis projects, ensuring timely completion.
· Identify opportunities to improve data analysis processes and tools.
· Collaborate with Data Engineers and Data Architects to ensure data solutions align with the overall data platform architecture.
· Evaluate and recommend new data analysis tools and technologies.
· Contribute to the development of best practices for data analysis.
· Participate in project meetings and provide input on data-related issues, risks and requirements.
Qualifications
· 8+ years of experience as a Data Analytics Lead, with experience leading or mentoring a team.
· Extensive experience with cloud-based data modelling and data warehousing solutions, using Azure Data Bricks.
· Proven experience in data technologies and platforms, ETL processes and tools, preferably using Azure Data Factory, Azure Databricks (Spark), Delta Lake.
· Advanced proficiency in data visualization tools such as Power BI.
Data Analysis and Visualization:
- Experience in data analysis, statistical modelling, and machine learning techniques.
- Proficiency in analytical tools like Python, R, and libraries such as Pandas, NumPy for data analysis and modelling.
- Strong expertise in Power BI, Superset, Tablue for data visualization, data modelling, and DAX queries, with knowledge of best practices.
- Experience in implementing Row-Level Security in Power BI.
- Ability to work with medium-complex data models and quickly understand application data design and processes.
- Familiar with industry best practices for Power BI and experienced in performance optimization of existing implementations.
- Understanding of machine learning algorithms, including supervised, unsupervised, and deep learning techniques.
Data Handling and Processing:
- Proficient in SQL Server and query optimization.
- Expertise in application data design and process management.
- Extensive knowledge of data modelling.
- Hands-on experience with Azure Data Factory,Azure Databricks.
- Expertise in data warehouse development, including experience with SSIS (SQL Server Integration Services) and SSAS (SQL Server Analysis Services).
- Proficiency in ETL processes (data extraction, transformation, and loading), including data cleaning and normalization.
- Familiarity with big data technologies (e.g., Hadoop, Spark, Kafka) for large-scale data processing.
Understanding of data governance, compliance, and security measures within Azure environments.
Job Title: Lead Data Engineer
📍 Location: Pune
🧾 Experience: 10+ Years
💰 Budget: Up to 1.7 LPM
Responsibilities
- Collaborate with Data & ETL teams to review, optimize, and scale data architectures within Snowflake.
- Design, develop, and maintain efficient ETL/ELT pipelines and robust data models.
- Optimize SQL queries for performance and cost efficiency.
- Ensure data quality, reliability, and security across pipelines and datasets.
- Implement Snowflake best practices for performance, scaling, and governance.
- Participate in code reviews, knowledge sharing, and mentoring within the data engineering team.
- Support BI and analytics initiatives by enabling high-quality, well-modeled datasets.
Exp: 10+ Years
CTC: 1.7 LPM
Location: Pune
SnowFlake Expertise Profile
Should hold 10 + years of experience with strong skills with core understanding of cloud data warehouse principles and extensive experience in designing, building, optimizing, and maintaining robust and scalable data solutions on the Snowflake platform.
Possesses a strong background in data modelling, ETL/ELT, SQL development, performance tuning, scaling, monitoring and security handling.
Responsibilities:
* Collaboration with Data and ETL team to review code, understand current architecture and help improve it based on Snowflake offerings and experience
* Review and implement best practices to design, develop, maintain, scale, efficiently monitor data pipelines and data models on the Snowflake platform for
ETL or BI.
* Optimize complex SQL queries for data extraction, transformation, and loading within Snowflake.
* Ensure data quality, integrity, and security within the Snowflake environment.
* Participate in code reviews and contribute to the team's development standards.
Education:
* Bachelor’s degree in computer science, Data Science, Information Technology, or anything equivalent.
* Relevant Snowflake certifications are a plus (e.g., Snowflake certified Pro / Architecture / Advanced).

Springer Capital is a cross-border asset management firm specializing in real estate investment banking between China and the USA. We are offering a remote internship for aspiring data engineers interested in data pipeline development, data integration, and business intelligence. The internship offers flexible start and end dates. A short quiz or technical task may be required as part of the selection process.
Responsibilities:
-Design, build, and maintain scalable data pipelines for
structured and unstructured data sources
-Develop ETL processes to collect, clean, and transform data
from internal and external systems. Support integration of data into
dashboards, analytics tools, and reporting systems
-Collaborate with data analysts and software developers to
improve data accessibility and performance.
-Document workflows and maintain data infrastructure best
practices.
-Assist in identifying opportunities to automate repetitive data
tasks
Please send your resume to talent@springer. capital

Springer Capital is a cross-border asset management firm specializing in real estate investment banking between China and the USA. We are offering a remote internship for aspiring data engineers interested in data pipeline development, data integration, and business intelligence. The internship offers flexible start and end dates. A short quiz or technical task may be required as part of the selection process.
Responsibilities:
-Design, build, and maintain scalable data pipelines for
structured and unstructured data sources
-Develop ETL processes to collect, clean, and transform data
from internal and external systems. Support integration of data into
dashboards, analytics tools, and reporting systems
-Collaborate with data analysts and software developers to
improve data accessibility and performance.
-Document workflows and maintain data infrastructure best
practices.
-Assist in identifying opportunities to automate repetitive data
tasks
Please send your resume to talent@springer. capital

Job Description: Oracle BI Publisher Developer
Position Type
• Work Type: Full-time
• Employment Type: Contract
Experience Required
• Minimum 1 year of hands-on experience in:
o Oracle Database: SQL development, performance tuning, data modelling
o Oracle BI Publisher: Report design, template customization, data source integration
Technical Skills
• Mandatory:
o Oracle SQL & PL/SQL
o BI Publisher report development and deployment
o XML and XSLT for template customization
• Preferred:
o Experience with Oracle E-Business Suite or Fusion Applications
o Familiarity with data visualization principles
o Basic understanding of performance metrics and report optimization
Responsibilities
• Design, develop, and maintain BI Publisher reports based on business requirements
• Write and optimize SQL queries for data extraction and transformation
• Collaborate with stakeholders to ensure report accuracy and usability
• Troubleshoot and resolve issues related to data and report performance
• Document technical specifications and maintain version control



We’re seeking a highly skilled, execution-focused Senior Data Scientist with a minimum of 5 years of experience. This role demands hands-on expertise in building, deploying, and optimizing machine learning models at scale, while working with big data technologies and modern cloud platforms. You will be responsible for driving data-driven solutions from experimentation to production, leveraging advanced tools and frameworks across Python, SQL, Spark, and AWS. The role requires strong technical depth, problem-solving ability, and ownership in delivering business impact through data science.
Responsibilities
- Design, build, and deploy scalable machine learning models into production systems.
- Develop advanced analytics and predictive models using Python, SQL, and popular ML/DL frameworks (Pandas, Scikit-learn, TensorFlow, PyTorch).
- Leverage Databricks, Apache Spark, and Hadoop for large-scale data processing and model training.
- Implement workflows and pipelines using Airflow and AWS EMR for automation and orchestration.
- Collaborate with engineering teams to integrate models into cloud-based applications on AWS.
- Optimize query performance, storage usage, and data pipelines for efficiency.
- Conduct end-to-end experiments, including data preprocessing, feature engineering, model training, validation, and deployment.
- Drive initiatives independently with high ownership and accountability.
- Stay up to date with industry best practices in machine learning, big data, and cloud-native deployments.
Requirements:
- Minimum 5 years of experience in Data Science or Applied Machine Learning.
- Strong proficiency in Python, SQL, and ML libraries (Pandas, Scikit-learn, TensorFlow, PyTorch).
- Proven expertise in deploying ML models into production systems.
- Experience with big data platforms (Hadoop, Spark) and distributed data processing.
- Hands-on experience with Databricks, Airflow, and AWS EMR.
- Strong knowledge of AWS cloud services (S3, Lambda, SageMaker, EC2, etc.).
- Solid understanding of query optimization, storage systems, and data pipelines.
- Excellent problem-solving skills, with the ability to design scalable solutions.
- Strong communication and collaboration skills to work in cross-functional teams.
Benefits:
- Best in class salary: We hire only the best, and we pay accordingly.
- Proximity Talks: Meet other designers, engineers, and product geeks — and learn from experts in the field.
- Keep on learning with a world-class team: Work with the best in the field, challenge yourself constantly, and learn something new every day.
About Us:
Proximity is the trusted technology, design, and consulting partner for some of the biggest Sports, Media, and Entertainment companies in the world! We’re headquartered in San Francisco and have offices in Palo Alto, Dubai, Mumbai, and Bangalore. Since 2019, Proximity has created and grown high-impact, scalable products used by 370 million daily users, with a total net worth of $45.7 billion among our client companies.
Today, we are a global team of coders, designers, product managers, geeks, and experts. We solve complex problems and build cutting-edge tech, at scale. Our team of Proxonauts is growing quickly, which means your impact on the company’s success will be huge. You’ll have the chance to work with experienced leaders who have built and led multiple tech, product, and design teams.

We are looking for experienced Data Engineers who can independently build, optimize, and manage scalable data pipelines and platforms.
In this role, you’ll:
- Work closely with clients and internal teams to deliver robust data solutions powering analytics, AI/ML, and operational systems.
- Mentor junior engineers and bring engineering discipline into our data engagements.
Key Responsibilities
- Design, build, and optimize large-scale, distributed data pipelines for both batch and streaming use cases.
- Implement scalable data models, warehouses/lakehouses, and data lakes to support analytics and decision-making.
- Collaborate with stakeholders to translate business requirements into technical solutions.
- Drive performance tuning, monitoring, and reliability of data pipelines.
- Write clean, modular, production-ready code with proper documentation and testing.
- Contribute to architectural discussions, tool evaluations, and platform setup.
- Mentor junior engineers and participate in code/design reviews.
Must-Have Skills
- Strong programming skills in Python and advanced SQL expertise.
- Deep understanding of ETL/ELT, data modeling (OLTP & OLAP), warehousing, and stream processing.
- Hands-on with distributed data processing frameworks (Apache Spark, Flink, or similar).
- Experience with orchestration tools like Airflow (or similar).
- Familiarity with CI/CD pipelines and Git.
- Ability to debug, optimize, and scale data pipelines in production.
Good to Have
- Experience with cloud platforms (AWS preferred; GCP/Azure also welcome).
- Exposure to Databricks, dbt, or similar platforms.
- Understanding of data governance, quality frameworks, and observability.
- Certifications (e.g., AWS Data Analytics, Solutions Architect, or Databricks).
Other Expectations
- Comfortable working in fast-paced, client-facing environments.
- Strong analytical and problem-solving skills with attention to detail.
- Ability to adapt across tools, stacks, and business domains.
- Willingness to travel within India for short/medium-term client engagements, as needed.

Role Overview:
We are seeking a talented and experienced Data Architect with strong data visualization capabilities to join our dynamic team in Mumbai. As a Data Architect, you will be responsible for designing, building, and managing our data infrastructure, ensuring its reliability, scalability, and performance. You will also play a crucial role in transforming complex data into insightful visualizations that drive business decisions. This role requires a deep understanding of data modeling, database technologies (particularly Oracle Cloud), data warehousing principles, and proficiency in data manipulation and visualization tools, including Python and SQL.
Responsibilities:
- Design and implement robust and scalable data architectures, including data warehouses, data lakes, and operational data stores, primarily leveraging Oracle Cloud services.
- Develop and maintain data models (conceptual, logical, and physical) that align with business requirements and ensure data integrity and consistency.
- Define data governance policies and procedures to ensure data quality, security, and compliance.
- Collaborate with data engineers to build and optimize ETL/ELT pipelines for efficient data ingestion, transformation, and loading.
- Develop and execute data migration strategies to Oracle Cloud.
- Utilize strong SQL skills to query, manipulate, and analyze large datasets from various sources.
- Leverage Python and relevant libraries (e.g., Pandas, NumPy) for data cleaning, transformation, and analysis.
- Design and develop interactive and insightful data visualizations using tools like [Specify Visualization Tools - e.g., Tableau, Power BI, Matplotlib, Seaborn, Plotly] to communicate data-driven insights to both technical and non-technical stakeholders.
- Work closely with business analysts and stakeholders to understand their data needs and translate them into effective data models and visualizations.
- Ensure the performance and reliability of data visualization dashboards and reports.
- Stay up-to-date with the latest trends and technologies in data architecture, cloud computing (especially Oracle Cloud), and data visualization.
- Troubleshoot data-related issues and provide timely resolutions.
- Document data architectures, data flows, and data visualization solutions.
- Participate in the evaluation and selection of new data technologies and tools.
Qualifications:
- Bachelor's or Master's degree in Computer Science, Data Science, Information Systems, or a related field.
- Proven experience (typically 5+ years) as a Data Architect, Data Modeler, or similar role.
- Deep understanding of data warehousing concepts, dimensional modeling (e.g., star schema, snowflake schema), and ETL/ELT processes.
- Extensive experience working with relational databases, particularly Oracle, and proficiency in SQL.
- Hands-on experience with Oracle Cloud data services (e.g., Autonomous Data Warehouse, Object Storage, Data Integration).
- Strong programming skills in Python and experience with data manipulation and analysis libraries (e.g., Pandas, NumPy).
- Demonstrated ability to create compelling and effective data visualizations using industry-standard tools (e.g., Tableau, Power BI, Matplotlib, Seaborn, Plotly).
- Excellent analytical and problem-solving skills with the ability to interpret complex data and translate it into actionable insights.
- Strong communication and presentation skills, with the ability to effectively communicate technical concepts to non-technical audiences.
- Experience with data governance and data quality principles.
- Familiarity with agile development methodologies.
- Ability to work independently and collaboratively within a team environment.
Application Link- https://forms.gle/km7n2WipJhC2Lj2r5

The Opportunity
We’re looking for a Senior Data Engineer to join our growing Data Platform team. This role is a hybrid of data engineering and business intelligence, ideal for someone who enjoys solving complex data challenges while also building intuitive and actionable reporting solutions.
You’ll play a key role in designing and scaling the infrastructure and pipelines that power analytics, dashboards, machine learning, and decision-making across Sonatype. You’ll also be responsible for delivering clear, compelling, and insightful business intelligence through tools like Looker Studio and advanced SQL queries.
What You’ll Do
- Design, build, and maintain scalable data pipelines and ETL/ELT processes.
- Architect and optimize data models and storage solutions for analytics and operational use.
- Create and manage business intelligence reports and dashboards using tools like Looker Studio, Power BI, or similar.
- Collaborate with data scientists, analysts, and stakeholders to ensure datasets are reliable, meaningful, and actionable.
- Own and evolve parts of our data platform (e.g., Airflow, dbt, Spark, Redshift, or Snowflake).
- Write complex, high-performance SQL queries to support reporting and analytics needs.
- Implement observability, alerting, and data quality monitoring for critical pipelines.
- Drive best practices in data engineering and business intelligence, including documentation, testing, and CI/CD.
- Contribute to the evolution of our next-generation data lakehouse and BI architecture.
What We’re Looking For
Minimum Qualifications
- 5+ years of experience as a Data Engineer or in a hybrid data/reporting role.
- Strong programming skills in Python, Java, or Scala.
- Proficiency with data tools such as Databricks, data modeling techniques (e.g., star schema, dimensional modeling), and data warehousing solutions like Snowflake or Redshift.
- Hands-on experience with modern data platforms and orchestration tools (e.g., Spark, Kafka, Airflow).
- Proficient in SQL with experience in writing and optimizing complex queries for BI and analytics.
- Experience with BI tools such as Looker Studio, Power BI, or Tableau.
- Experience in building and maintaining robust ETL/ELT pipelines in production.
- Understanding of data quality, observability, and governance best practices.
Bonus Points
- Experience with dbt, Terraform, or Kubernetes.
- Familiarity with real-time data processing or streaming architectures.
- Understanding of data privacy, compliance, and security best practices in analytics and reporting.
Why You’ll Love Working Here
- Data with purpose: Work on problems that directly impact how the world builds secure software.
- Full-spectrum impact: Use both engineering and analytical skills to shape product, strategy, and operations.
- Modern tooling: Leverage the best of open-source and cloud-native technologies.
- Collaborative culture: Join a passionate team that values learning, autonomy, and real-world impact.

About the Role
We’re hiring a Data Engineer to join our Data Platform team. You’ll help build and scale the systems that power analytics, reporting, and data-driven features across the company. This role works with engineers, analysts, and product teams to make sure our data is accurate, available, and usable.
What You’ll Do
- Build and maintain reliable data pipelines and ETL/ELT workflows.
- Develop and optimize data models for analytics and internal tools.
- Work with team members to deliver clean, trusted datasets.
- Support core data platform tools like Airflow, dbt, Spark, Redshift, or Snowflake.
- Monitor data pipelines for quality, performance, and reliability.
- Write clear documentation and contribute to test coverage and CI/CD processes.
- Help shape our data lakehouse architecture and platform roadmap.
What You Need
- 2–4 years of experience in data engineering or a backend data-related role.
- Strong skills in Python or another backend programming language.
- Experience working with SQL and distributed data systems (e.g., Spark, Kafka).
- Familiarity with NoSQL stores like HBase or similar.
- Comfortable writing efficient queries and building data workflows.
- Understanding of data modeling for analytics and reporting.
- Exposure to tools like Airflow or other workflow schedulers.
Bonus Points
- Experience with DBT, Databricks, or real-time data pipelines.
- Familiarity with cloud infrastructure tools like Terraform or Kubernetes.
- Interest in data governance, ML pipelines, or compliance standards.
Why Join Us?
- Work on data that supports meaningful software security outcomes.
- Use modern tools in a cloud-first, open-source-friendly environment.
- Join a team that values clarity, learning, and autonomy.
If you're excited about building impactful software and helping others do the same, this is an opportunity to grow as a technical leader and make a meaningful impact.
Advanced SQL, data modeling skills - designing Dimensional Layer, 3NF, denormalized views & semantic layer, Expertise in GCP services
Role & Responsibilities:
● Design and implement robust semantic layers for data systems on Google Cloud Platform (GCP)
● Develop and maintain complex data models, including dimensional models, 3NF structures, and denormalized views
● Write and optimize advanced SQL queries for data extraction, transformation, and analysis
● Utilize GCP services to create scalable and efficient data architectures
● Collaborate with cross-functional teams to translate business requirements(specified in mapping sheets or Legacy
Datastage jobs) into effective data models
● Implement and maintain data warehouses and data lakes on GCP
● Design and optimize ETL/ELT processes for large-scale data integration
● Ensure data quality, consistency, and integrity across all data models and semantic layers
● Develop and maintain documentation for data models, semantic layers, and data flows
● Participate in code reviews and implement best practices for data modeling and database design
● Optimize database performance and query execution on GCP
● Provide technical guidance and mentorship to junior team members
● Stay updated with the latest trends and advancements in data modeling, GCP services, and big data technologies
● Collaborate with data scientists and analysts to enable efficient data access and analysis
● Implement data governance and security measures within the semantic layer and data model
Springer Capital is a cross-border asset management firm specializing in real estate investment banking between China and the USA. We are offering a remote internship for aspiring data engineers interested in data pipeline development, data integration, and business intelligence.
The internship offers flexible start and end dates. A short quiz or technical task may be required as part of the selection process.
Responsibilities:
- Design, build, and maintain scalable data pipelines for structured and unstructured data sources
- Develop ETL processes to collect, clean, and transform data from internal and external systems
- Support integration of data into dashboards, analytics tools, and reporting systems
- Collaborate with data analysts and software developers to improve data accessibility and performance
- Document workflows and maintain data infrastructure best practices
- Assist in identifying opportunities to automate repetitive data tasks
Springer Capital is a cross-border asset management firm specializing in real estate investment banking between China and the USA. We are offering a remote internship for aspiring data engineers interested in data pipeline development, data integration, and business intelligence. The internship offers flexible start and end dates. A short quiz or technical task may be required as part of the selection process.
Responsibilities:
▪ Design, build, and maintain scalable data pipelines for structured and unstructured data sources
▪ Develop ETL processes to collect, clean, and transform data from internal and external systems
▪ Support integration of data into dashboards, analytics tools, and reporting systems
▪ Collaborate with data analysts and software developers to improve data accessibility and performance
▪ Document workflows and maintain data infrastructure best practices
▪ Assist in identifying opportunities to automate repetitive data tasks
Job Title : Informatica MDM Developer
Experience : 7 to 10 Years
Location : Bangalore (3 Days Work From Office – ITPL Main Road, Mahadevapura)
Job Type : Full-time / Contract
Job Overview :
We are hiring an experienced Informatica MDM Developer to join our team in Bangalore. The ideal candidate will play a key role in implementing and customizing Master Data Management (MDM) solutions using Informatica MDM (Multi-Domain Edition), ensuring a trusted, unified view of enterprise data.
Mandatory Skills :
Informatica MDM (Multi-Domain Edition), ActiveVOS workflows, Java (User Exits), Services Integration Framework (SIF) APIs, SQL/PLSQL, Data Modeling, Informatica Data Quality (IDQ), MDM concepts (golden record, survivorship, trust, hierarchy).
Key Responsibilities :
- Configure Informatica MDM Hub : subject area models, base objects, relationships.
- Develop match/merge rules, trust/survivorship logic to create golden records.
- Design workflows using ActiveVOS for data stewardship and exception handling.
- Integrate with source/target systems (ERP, CRM, Data Lakes, APIs).
- Customize user exits (Java), SIF APIs, and business entity services.
- Implement and maintain data quality validations using IDQ.
- Collaborate with cross-functional teams for governance alignment.
- Support MDM jobs, synchronization, batch groups, and performance tuning.
Must-Have Skills :
- 7 to 10 years of experience in Data Engineering or MDM.
- 5+ years hands-on with Informatica MDM (Multi-Domain Edition).
- Strong in MDM concepts : golden record, trust, survivorship, hierarchy.
Proficient in :
- Informatica MDM Hub Console, Provisioning Tool, SIF.
- ActiveVOS workflows, Java-based user exits.
- SQL, PL/SQL, and data modeling.
- Experience with system integration and Informatica Data Quality (IDQ).
Nice-to-Have :
- Knowledge of Informatica EDC, Axon, cloud MDM (AWS/GCP/Azure).
- Understanding of data lineage, GDPR/HIPAA compliance, and DevOps tools.
Required Skills:
● 6+ years of experience with hybrid data environments that leverage both distributed and relational database technologies to support analytics services (Oracle, IMB DB2, GCP)
● Solid understanding of data warehousing principles, architecture, and its implementation in complex environments.
● Good experience in OLTP and OLAP systems
● Excellent Data Analysis skills
● Good understanding of one or more ETL tools and data ingestion frameworks.
● Experience as a designer of complex Dimensional data models for analytics services
● Experience with various testing methodologies and user acceptance testing.
● Experience on one or more cloud platforms (e.g. AWS, Azure, GCP) ● Understanding of Data Quality and Data Governance
● Understanding of Industry Data Models
● Experience in leading the large teams
● Experience with processing large datasets from multiple sources.
● Ability to operate effectively and independently in a dynamic, fluid environment.
● Good understanding of agile methodology
● Strong verbal and written communications skills with experience in relating complex concepts to non-technical users.
● Demonstrated ability to exchange ideas and convey complex information clearly and concisely
● Proven ability to lead and drive projects and assignments to completion
● Exposure to Data Modeling Tools
○ ERwin ○ Power Designer ○ Business Glossary ○ ER/Studio ○ Enterprise Architect, ○ MagicDraw
Job Summary:
Position : Senior Power BI Developer
Experience : 4+Years
Location : Ahmedabad - WFO
Key Responsibilities:
- Design, develop, and maintain interactive and user-friendly Power BI dashboards and
- reports.
- Translate business requirements into functional and technical specifications.
- Perform data modeling, DAX calculations, and Power Query transformations.
- Integrate data from multiple sources including SQL Server, Excel, SharePoint, and APIs.
- Optimize Power BI datasets, reports, and dashboards for performance and usability.
- Collaborate with business analysts, data engineers, and stakeholders to ensure data accuracy
- and relevance.
- Ensure security and governance best practices in Power BI workspaces and datasets.
- Provide ongoing support and troubleshooting for existing Power BI solutions.
- Stay updated with Power BI updates, best practices, and industry trends.
Required Skills & Qualifications:
- Bachelor’s degree in Computer Science, Information Technology, Data Analytics, or a
- related field.
- 4+ years of professional experience in data analytics or business intelligence.
- 3+ years of hands-on experience with Power BI (Power BI Desktop, Power BI Service).
- Strong expertise in DAX, Power Query (M Language), and data modeling (star/snowflake schema).
- Proficiency in writing complex SQL queries and optimizing them for performance.
- Experience in working with large and complex datasets.
- Experience in BigQuery, MySql, Looker Studio is a plus.
- Ecommerce Industry Experience will be an added advantage.
- Solid understanding of data warehousing concepts and ETL processes.
- Experience with version control tools such as Power Apps & Power Automate would be a plus.
Preferred Qualifications:
- Microsoft Power BI Certification (PL-300 or equivalent is a plus). Experience with Azure Data Services (Azure Data Factory, Azure SQL, Synapse).
- Knowledge of other BI tools (Tableau, Qlik) is a plus.
- Familiarity with scripting languages (Python, R) for data analysis is a bonus.
- Experience integrating Power BI into web portals using Power BI Embedded.
Job Title : Senior Data Engineer
Experience : 6 to 10 Years
Location : Gurgaon (Hybrid – 3 days office / 2 days WFH)
Notice Period : Immediate to 30 days (Buyout option available)
About the Role :
We are looking for an experienced Senior Data Engineer to join our Digital IT team in Gurgaon.
This role involves building scalable data pipelines, managing data architecture, and ensuring smooth data flow across the organization while maintaining high standards of security and compliance.
Mandatory Skills :
Azure Data Factory (ADF), Azure Cloud Services, SQL, Data Modelling, CI/CD tools, Git, Data Governance, RDBMS & NoSQL databases (e.g., SQL Server, PostgreSQL, Redis, ElasticSearch), Data Lake migration.
Key Responsibilities :
- Design and develop secure, scalable end-to-end data pipelines using Azure Data Factory (ADF) and Azure services.
- Build and optimize data architectures (including Medallion Architecture).
- Collaborate with cross-functional teams on cybersecurity, data privacy (e.g., GDPR), and governance.
- Manage structured/unstructured data migration to Data Lake.
- Ensure CI/CD integration for data workflows and version control using Git.
- Identify and integrate data sources (internal/external) in line with business needs.
- Proactively highlight gaps and risks related to data compliance and integrity.
Required Skills :
- Azure Data Factory (ADF) – Mandatory
- Strong SQL and Data Modelling expertise.
- Hands-on with Azure Cloud Services and data architecture.
- Experience with CI/CD tools and version control (Git).
- Good understanding of Data Governance practices.
- Exposure to ETL/ELT pipelines and Data Lake migration.
- Working knowledge of RDBMS and NoSQL databases (e.g., SQL Server, PostgreSQL, Redis, ElasticSearch).
- Understanding of RESTful APIs, deployment on cloud/on-prem infrastructure.
- Strong problem-solving, communication, and collaboration skills.
Additional Info :
- Work Mode : Hybrid (No remote); relocation to Gurgaon required for non-NCR candidates.
- Communication : Above-average verbal and written English skills.
Perks & Benefits :
- 5 Days work week
- Global exposure and leadership collaboration.
- Health insurance, employee-friendly policies, training and development.
Job Description :
We are seeking a highly experienced Sr Data Modeler / Solution Architect to join the Data Architecture team at Corporate Office in Bangalore. The ideal candidate will have 4 to 8 years of experience in data modeling and architecture with deep expertise in AWS cloud stack, data warehousing, and enterprise data modeling tools. This individual will be responsible for designing and creating enterprise-grade data models and driving the implementation of Layered Scalable Architecture or Medallion Architecture to support robust, scalable, and high-quality data marts across multiple business units.
This role will involve managing complex datasets from systems like PoS, ERP, CRM, and external sources, while optimizing performance and cost. You will also provide strategic leadership on data modeling standards, governance, and best practices, ensuring the foundation for analytics and reporting is solid and future-ready.
Key Responsibilities:
· Design and deliver conceptual, logical, and physical data models using tools like ERWin.
· Implement Layered Scalable Architecture / Medallion Architecture for building scalable, standardized data marts.
· Optimize performance and cost of AWS-based data infrastructure (Redshift, S3, Glue, Lambda, etc.).
· Collaborate with cross-functional teams (IT, business, analysts) to gather data requirements and ensure model alignment with KPIs and business logic.
· Develop and optimize SQL code, materialized views, stored procedures in AWS Redshift.
· Ensure data governance, lineage, and quality mechanisms are established across systems.
· Lead and mentor technical teams in an Agile project delivery model.
· Manage data layer creation and documentation: data dictionary, ER diagrams, purpose mapping.
· Identify data gaps and availability issues with respect to source systems.
Required Skills & Qualifications:
· Bachelor’s or Master’s degree in Computer Science, IT, or related field (B.E./B.Tech/M.E./M.Tech/MCA).
· Minimum 4 years of experience in data modeling and architecture.
· Proficiency with data modeling tools such as ERWin, with strong knowledge of forward and reverse engineering.
· Deep expertise in SQL (including advanced SQL, stored procedures, performance tuning).
· Strong experience in data warehousing, RDBMS, and ETL tools like AWS Glue, IBM DataStage, or SAP Data Services.
· Hands-on experience with AWS services: Redshift, S3, Glue, RDS, Lambda, Bedrock, and Q.
· Good understanding of reporting tools such as Tableau, Power BI, or AWS QuickSight.
· Exposure to DevOps/CI-CD pipelines, AI/ML, Gen AI, NLP, and polyglot programming is a plus.
· Familiarity with data governance tools (e.g., ORION/EIIG).
· Domain knowledge in Retail, Manufacturing, HR, or Finance preferred.
· Excellent written and verbal communication skills.
Certifications (Preferred):
· AWS Certification (e.g., AWS Certified Solutions Architect or Data Analytics – Specialty)
· Data Governance or Data Modeling Certifications (e.g., CDMP, Databricks, or TOGAF)
Mandatory Skills
aws, Technical Architecture, Aiml, SQL, Data Warehousing, Data Modelling
Job Description:
We are seeking a skilled Power BI Developer with a strong understanding of Capital Markets to join our data analytics team. The ideal candidate will be responsible for designing, developing, and maintaining interactive dashboards and reports that provide insights into trading, risk, and financial performance. This role requires experience working with capital market data sets and a solid grasp of financial instruments and market operations.
Key Responsibilities:
- Develop interactive Power BI dashboards and reports tailored to capital markets (e.g., equities, derivatives, fixed income).
- Connect to and integrate data from various sources such as Bloomberg, Reuters, SQL databases, and Excel.
- Translate business requirements into data models and visualizations that provide actionable insights.
- Optimize Power BI reports for performance, usability, and scalability.
- Work closely with business stakeholders (trading, risk, compliance) to understand KPIs and analytics needs.
- Implement row-level security and data access controls.
- Maintain data quality, lineage, and versioning documentation.
Required Skills & Qualifications:
- 3+ years of experience with Power BI (Power Query, DAX, data modeling).
- Strong understanding of capital markets: trading workflows, market data, instruments (equities, bonds, derivatives, etc.).
- Experience with SQL and working with large financial datasets.
- Familiarity with risk metrics, trade lifecycle, and financial statement analysis.
- Knowledge of data governance, security, and performance tuning in BI environments.
- Excellent communication skills and ability to work with cross-functional teams.
Preferred Qualifications:
- Experience with Python or R for data analysis.
- Knowledge of investment banking or asset management reporting frameworks.
- Exposure to cloud platforms like Azure, AWS, or GCP.
- Certifications in Power BI or Capital Markets.
Job Title : Solution Architect – Denodo
Experience : 10+ Years
Location : Remote / Work from Home
Notice Period : Immediate joiners preferred
Job Overview :
We are looking for an experienced Solution Architect – Denodo to lead the design and implementation of data virtualization solutions. In this role, you will work closely with cross-functional teams to ensure our data architecture aligns with strategic business goals. The ideal candidate will bring deep expertise in Denodo, strong technical leadership, and a passion for driving data-driven decisions.
Mandatory Skills : Denodo, Data Virtualization, Data Architecture, SQL, Data Modeling, ETL, Data Integration, Performance Optimization, Communication Skills.
Key Responsibilities :
- Architect and design scalable data virtualization solutions using Denodo.
- Collaborate with business analysts and engineering teams to understand requirements and define technical specifications.
- Ensure adherence to best practices in data governance, performance, and security.
- Integrate Denodo with diverse data sources and optimize system performance.
- Mentor and train team members on Denodo platform capabilities.
- Lead tool evaluations and recommend suitable data integration technologies.
- Stay updated with emerging trends in data virtualization and integration.
Required Qualifications :
- Bachelor’s degree in Computer Science, IT, or a related field.
- 10+ Years of experience in data architecture and integration.
- Proven expertise in Denodo and data virtualization frameworks.
- Strong proficiency in SQL and data modeling.
- Hands-on experience with ETL processes and data integration tools.
- Excellent communication, presentation, and stakeholder management skills.
- Ability to lead technical discussions and influence architectural decisions.
- Denodo or data architecture certifications are a strong plus.
Job Summary:
Position : Senior Power BI Developer
Experience : 4+Years
Location : Ahmedabad - WFO
Key Responsibilities:
- Design, develop, and maintain interactive and user-friendly Power BI dashboards and
- reports.
- Translate business requirements into functional and technical specifications.
- Perform data modeling, DAX calculations, and Power Query transformations.
- Integrate data from multiple sources including SQL Server, Excel, SharePoint, and APIs.
- Optimize Power BI datasets, reports, and dashboards for performance and usability.
- Collaborate with business analysts, data engineers, and stakeholders to ensure data accuracy
- and relevance.
- Ensure security and governance best practices in Power BI workspaces and datasets.
- Provide ongoing support and troubleshooting for existing Power BI solutions.
- Stay updated with Power BI updates, best practices, and industry trends.
Required Skills & Qualifications:
- Bachelor’s degree in Computer Science, Information Technology, Data Analytics, or a
- related field.
- 4+ years of professional experience in data analytics or business intelligence.
- 3+ years of hands-on experience with Power BI (Power BI Desktop, Power BI Service).
- Strong expertise in DAX, Power Query (M Language), and data modeling (star/snowflake schema).
- Proficiency in writing complex SQL queries and optimizing them for performance.
- Experience in working with large and complex datasets.
- Experience in BigQuery, MySql, Looker Studio is a plus.
- Ecommerce Industry Experience will be an added advantage.
- Solid understanding of data warehousing concepts and ETL processes.
- Experience with version control tools such as Power Apps & Power Automate would be a plus.
Preferred Qualifications:
- Microsoft Power BI Certification (PL-300 or equivalent is a plus). Experience with Azure Data Services (Azure Data Factory, Azure SQL, Synapse).
- Knowledge of other BI tools (Tableau, Qlik) is a plus.
- Familiarity with scripting languages (Python, R) for data analysis is a bonus.
- Experience integrating Power BI into web portals using Power BI Embedded.
Job Title : Senior Software Engineer – Backend
Experience Required : 6 to 12 Years
Location : Bengaluru (Hybrid – 3 Days Work From Office)
Number of Openings : 2
Work Hours : 11:00 AM – 8:00 PM IST
Notice Period : 30 Days Preferred
Work Location : SmartWorks The Cube, Karle Town SEZ, Building No. 5, Nagavara, Bangalore – 560045
Note : Face-to-face interview in Bangalore is mandatory during the second round.
Role Overview :
We are looking for an experienced Senior Backend Developer to join our growing team. This is a hands-on role focused on building cloud-based, scalable applications in the mortgage finance domain.
Key Responsibilities :
- Design, develop, and maintain backend components for cloud-based web applications.
- Contribute to architectural decisions involving microservices and distributed systems.
- Work extensively with Node.js and RESTful APIs.
- Implement scalable solutions using AWS services (e.g., Lambda, SQS, SNS, RDS).
- Utilize both relational and NoSQL databases effectively.
- Collaborate with cross-functional teams to deliver robust and maintainable code.
- Participate in agile development practices and deliver rapid iterations based on feedback.
- Take ownership of system performance, scalability, and reliability.
Core Requirements :
- 5+ Years of total experience in backend development.
- Minimum 3 Years of experience in building scalable microservices or delivering large-scale products.
- Strong expertise in Node.js and REST APIs.
- Solid experience with RDBMS, SQL, and data modeling.
- Good understanding of distributed systems, scalability, and availability.
- Familiarity with AWS infrastructure and services.
- Development experience in Python and/or Java is a plus.
Preferred Skills :
- Experience with frontend frameworks like React.js or AngularJS.
- Working knowledge of Docker and containerized applications.
Interview Process :
- Round 1 : Online technical assessment (1 hour)
- Round 2 : Virtual technical interview
- Round 3 : In-person interview at the Bangalore office (2 hours – mandatory)

Role: Data Engineer (14+ years of experience)
Location: Whitefield, Bangalore
Mode of Work: Hybrid (3 days from office)
Notice period: Immediate/ Serving with 30days left
Location: Candidate should be based out of Bangalore as one round has to be taken F2F
Job Summary:
Role and Responsibilities
● Design and implement scalable data pipelines for ingesting, transforming, and loading data from various tools and sources.
● Design data models to support data analysis and reporting.
● Automate data engineering tasks using scripting languages and tools.
● Collaborate with engineers, process managers, data scientists to understand their needs and design solutions.
● Act as a bridge between the engineering and the business team in all areas related to Data.
● Automate monitoring and alerting mechanism on data pipelines, products and Dashboards and troubleshoot any issues. On call requirements.
● SQL creation and optimization - including modularization and optimization which might need views, table creation in the sources etc.
● Defining best practices for data validation and automating as much as possible; aligning with the enterprise standards
● QA environment data management - e.g Test Data Management etc
Qualifications
● 14+ years of experience as a Data engineer or related role.
● Experience with Agile engineering practices.
● Strong experience in writing queries for RDBMS, cloud-based data warehousing solutions like Snowflake and Redshift.
● Experience with SQL and NoSQL databases.
● Ability to work independently or as part of a team.
● Experience with cloud platforms, preferably AWS.
● Strong experience with data warehousing and data lake technologies (Snowflake)
● Expertise in data modelling
● Experience with ETL/LT tools and methodologies .
● 5+ years of experience in application development including Python, SQL, Scala, or Java
● Experience working on real-time Data Streaming and Data Streaming platform.
NOTE: IT IS MANDATORY TO GIVE ONE TECHNICHAL ROUND FACE TO FACE.

Location: Mumbai
Job Type: Full-Time (Hybrid – 3 days in office, 2 days WFH)
Job Overview:
We are looking for a skilled Azure Data Engineer with strong experience in data modeling, pipeline development, and SQL/Spark expertise. The ideal candidate will work closely with the Data Analytics & BI teams to implement robust data solutions on Azure Synapse and ensure seamless data integration with third-party applications.
Key Responsibilities:
- Design, develop, and maintain Azure data pipelines using Azure Synapse (SQL dedicated pools or Apache Spark pools).
- Implement data models in collaboration with the Data Analytics and BI teams.
- Optimize and manage large-scale SQL and Spark-based data processing solutions.
- Ensure data availability and reliability for third-party application consumption.
- Collaborate with cross-functional teams to translate business requirements into scalable data solutions.
Required Skills & Experience:
3–5 years of hands-on experience in:
- Azure data services
- Data Modeling
- SQL development and tuning
- Apache Spark
- Strong knowledge of Azure Synapse Analytics.
- Experience in designing data pipelines and ETL/ELT processes.
- Ability to troubleshoot and optimize complex data workflows.
Preferred Qualifications:
- Experience with data governance, security, and data quality practices.
- Familiarity with DevOps practices in a data engineering context.
- Effective communication skills and the ability to work in a collaborative team environment.
Involved in capex,modelling, Budgeting,investment decision making
Should be able to converse well with global stakeholders
O&G, Metals and mining heavy manufacturing Preferred
Others - not a frequent job hopper
Margin optimization -holistic understanding of the O&G P&L (not limited to 1-2 line items only)
Job Title : Cognos BI Developer
Experience : 6+ Years
Location : Bangalore / Hyderabad (Hybrid)
Notice Period : Immediate Joiners Preferred (Candidates serving notice with 10–15 days left can be considered)
Interview Mode : Virtual
Job Description :
We are seeking an experienced Cognos BI Developer with strong data modeling, dashboarding, and reporting expertise to join our growing team. The ideal candidate should have a solid background in business intelligence, data visualization, and performance analysis, and be comfortable working in a hybrid setup from Bangalore or Hyderabad.
Mandatory Skills :
Cognos BI, Framework Manager, Cognos Dashboarding, SQL, Data Modeling, Report Development (charts, lists, cross tabs, maps), ETL Concepts, KPIs, Drill-through, Macros, Prompts, Filters, Calculations.
Key Responsibilities :
- Understand business requirements in the BI context and design data models using Framework Manager to transform raw data into meaningful insights.
- Develop interactive dashboards and reports using Cognos Dashboard.
- Identify and define KPIs and create reports to monitor them effectively.
- Analyze data and present actionable insights to support business decision-making.
- Translate business requirements into technical specifications and determine timelines for execution.
- Design and develop models in Framework Manager, publish packages, manage security, and create reports based on these packages.
- Develop various types of reports, including charts, lists, cross tabs, and maps, and design dashboards combining multiple reports.
- Implement reports using macros, prompts, filters, and calculations.
- Perform data warehouse development activities and ensure seamless data flow.
- Write and optimize SQL queries to investigate data and resolve performance issues.
- Utilize Cognos features such as master-detail reports, drill-throughs, bookmarks, and page sets.
- Analyze and improve ETL processes to enhance data integration.
- Apply technical enhancements to existing BI systems to improve their performance and usability.
- Possess solid understanding of database fundamentals, including relational and multidimensional database design.
- Hands-on experience with Cognos Data Modules (data modeling) and dashboarding.
Job Title : Data Engineer – Snowflake Expert
Location : Pune (Onsite)
Experience : 10+ Years
Employment Type : Contractual
Mandatory Skills : Snowflake, Advanced SQL, ETL/ELT (Snowpipe, Tasks, Streams), Data Modeling, Performance Tuning, Python, Cloud (preferably Azure), Security & Data Governance.
Job Summary :
We are seeking a seasoned Data Engineer with deep expertise in Snowflake to design, build, and maintain scalable data solutions.
The ideal candidate will have a strong background in data modeling, ETL/ELT, SQL optimization, and cloud data warehousing principles, with a passion for leveraging Snowflake to drive business insights.
Responsibilities :
- Collaborate with data teams to optimize and enhance data pipelines and models on Snowflake.
- Design and implement scalable ELT pipelines with performance and cost-efficiency in mind.
- Ensure high data quality, security, and adherence to governance frameworks.
- Conduct code reviews and align development with best practices.
Qualifications :
- Bachelor’s in Computer Science, Data Science, IT, or related field.
- Snowflake certifications (Pro/Architect) preferred.
- Extract Transform Load (ETL) and ETL Tools skills
- Data Modeling and Data Integration expertise
- Data Warehousing knowledge
- Experience in working with SQL databases
- Strong analytical and problem-solving abilities
- Excellent communication and interpersonal skills
- Bachelor's degree in Computer Science, Information Systems, or related field
- Relevant certifications in ETL Testing or Data Warehousing

About the Role:
We are seeking a talented Lead Data Engineer to join our team and play a pivotal role in transforming raw data into valuable insights. As a Data Engineer, you will design, develop, and maintain robust data pipelines and infrastructure to support our organization's analytics and decision-making processes.
Responsibilities:
- Data Pipeline Development: Build and maintain scalable data pipelines to extract, transform, and load (ETL) data from various sources (e.g., databases, APIs, files) into data warehouses or data lakes.
- Data Infrastructure: Design, implement, and manage data infrastructure components, including data warehouses, data lakes, and data marts.
- Data Quality: Ensure data quality by implementing data validation, cleansing, and standardization processes.
- Team Management: Able to handle team.
- Performance Optimization: Optimize data pipelines and infrastructure for performance and efficiency.
- Collaboration: Collaborate with data analysts, scientists, and business stakeholders to understand their data needs and translate them into technical requirements.
- Tool and Technology Selection: Evaluate and select appropriate data engineering tools and technologies (e.g., SQL, Python, Spark, Hadoop, cloud platforms).
- Documentation: Create and maintain clear and comprehensive documentation for data pipelines, infrastructure, and processes.
Skills:
- Strong proficiency in SQL and at least one programming language (e.g., Python, Java).
- Experience with data warehousing and data lake technologies (e.g., Snowflake, AWS Redshift, Databricks).
- Knowledge of cloud platforms (e.g., AWS, GCP, Azure) and cloud-based data services.
- Understanding of data modeling and data architecture concepts.
- Experience with ETL/ELT tools and frameworks.
- Excellent problem-solving and analytical skills.
- Ability to work independently and as part of a team.
Preferred Qualifications:
- Experience with real-time data processing and streaming technologies (e.g., Kafka, Flink).
- Knowledge of machine learning and artificial intelligence concepts.
- Experience with data visualization tools (e.g., Tableau, Power BI).
- Certification in cloud platforms or data engineering.
Bangalore / Chennai
- Hands-on data modelling for OLTP and OLAP systems
- In-depth knowledge of Conceptual, Logical and Physical data modelling
- Strong understanding of Indexing, partitioning, data sharding, with practical experience of having done the same
- Strong understanding of variables impacting database performance for near-real-time reporting and application interaction.
- Should have working experience on at least one data modelling tool, preferably DBSchema, Erwin
- Good understanding of GCP databases like AlloyDB, CloudSQL, and BigQuery.
- People with functional knowledge of the mutual fund industry will be a plus
Role & Responsibilities:
● Work with business users and other stakeholders to understand business processes.
● Ability to design and implement Dimensional and Fact tables
● Identify and implement data transformation/cleansing requirements
● Develop a highly scalable, reliable, and high-performance data processing pipeline to extract, transform and load data from various systems to the Enterprise Data Warehouse
● Develop conceptual, logical, and physical data models with associated metadata including data lineage and technical data definitions
● Design, develop and maintain ETL workflows and mappings using the appropriate data load technique
● Provide research, high-level design, and estimates for data transformation and data integration from source applications to end-user BI solutions.
● Provide production support of ETL processes to ensure timely completion and availability of data in the data warehouse for reporting use.
● Analyze and resolve problems and provide technical assistance as necessary. Partner with the BI team to evaluate, design, develop BI reports and dashboards according to functional specifications while maintaining data integrity and data quality.
● Work collaboratively with key stakeholders to translate business information needs into well-defined data requirements to implement the BI solutions.
● Leverage transactional information, data from ERP, CRM, HRIS applications to model, extract and transform into reporting & analytics.
● Define and document the use of BI through user experience/use cases, prototypes, test, and deploy BI solutions.
● Develop and support data governance processes, analyze data to identify and articulate trends, patterns, outliers, quality issues, and continuously validate reports, dashboards and suggest improvements.
● Train business end-users, IT analysts, and developers.
Required Skills:
● Bachelor’s degree in Computer Science or similar field or equivalent work experience.
● 5+ years of experience on Data Warehousing, Data Engineering or Data Integration projects.
● Expert with data warehousing concepts, strategies, and tools.
● Strong SQL background.
● Strong knowledge of relational databases like SQL Server, PostgreSQL, MySQL.
● Strong experience in GCP & Google BigQuery, Cloud SQL, Composer (Airflow), Dataflow, Dataproc, Cloud Function and GCS
● Good to have knowledge on SQL Server Reporting Services (SSRS), and SQL Server Integration Services (SSIS).
● Knowledge of AWS and Azure Cloud is a plus.
● Experience in Informatica Power exchange for Mainframe, Salesforce, and other new-age data sources.
● Experience in integration using APIs, XML, JSONs etc.
📢 Job Title : MDM Business Analyst
📍 Location : On-site, Sydney, Australia
💼 Experience : 5+ Years
🗓 Notice Period : Immediate
Summary :
We are seeking a skilled MDM Business Analyst with hands-on experience in Atacama or Informatica MDM solutions. The ideal candidate will work closely with business and technical teams to define MDM strategies, ensure data quality, and drive enterprise data governance.
🔧 Key Responsibilities :
- Gather, analyze, and document business and data governance requirements.
- Design and support implementation of MDM solutions in collaboration with technical teams.
- Prepare BRDs, FRSs, and TDDs to translate business needs into technical specifications.
- Lead data profiling, modeling, cleansing, and mapping activities.
- Ensure effective MDM performance, hierarchy management, and workflow automation.
- Support UAT by defining test cases and validating data quality.
- Act as a liaison between business, IT, and vendors to ensure smooth project delivery.
✅ Required Skills :
- Bachelor’s in Computer Science, IT, or related field.
- 5+ years of experience in MDM-focused Business Analyst roles.
- Expertise in Atacama MDM / Informatica MDM (or similar tools).
- Strong SQL, data modeling, and data governance experience.
- Familiarity with relational databases (Oracle, SQL Server, PostgreSQL).
- Excellent communication, documentation, and stakeholder management skills.
➕ Nice to Have :
- Familiarity with Agile (Scrum/Kanban) methodologies.
- Experience with BI tools (e.g., Power BI, Tableau).
- Certifications in MDM, CBAP, PMP, or Informatica.
- Exposure to cloud MDM (AWS, Azure, GCP).
- Experience in finance, healthcare, or retail domains.

Key Responsibilities :
- Centralize structured and unstructured data
- Contribute to data strategy through data modeling, management, and governance
- Build, optimize, and maintain data pipelines and management frameworks
- Collaborate with cross-functional teams to develop scalable data and AI-driven solutions
- Take ownership of projects from ideation to production
Ideal Qualifications and Skills :
- Bachelor's degree in Computer Science or equivalent experience
- 8+ years of industry experience
- Strong expertise in data modeling and management concepts
- Experience with Snowflake, data warehousing, and data pipelines
- Proficiency in Python or another programming language
- Excellent communication, collaboration, and ownership mindset
- Foundational knowledge of API development and integration
Nice to Have :
- Experience with Tableau, Alteryx
- Master data management implementation experience
Success Factors :
- Strong technical foundation
- Collaborative mindset
- Ability to navigate complex data challenges
- Ownership mindset and startup-like culture fit
1. GCP - GCS, PubSub, Dataflow or DataProc, Bigquery, BQ optimization, Airflow/Composer, Python(preferred)/Java
2. ETL on GCP Cloud - Build pipelines (Python/Java) + Scripting, Best Practices, Challenges
3. Knowledge of Batch and Streaming data ingestion, build End to Data pipelines on GCP
4. Knowledge of Databases (SQL, NoSQL), On-Premise and On-Cloud, SQL vs No SQL, Types of No-SQL DB (At Least 2 databases)
5. Data Warehouse concepts - Beginner to Intermediate level
6.Data Modeling, GCP Databases, DB Schema(or similar)
7.Hands-on data modelling for OLTP and OLAP systems
8.In-depth knowledge of Conceptual, Logical and Physical data modelling
9.Strong understanding of Indexing, partitioning, data sharding, with practical experience of having done the same
10.Strong understanding of variables impacting database performance for near-real-time reporting and application interaction.
11.Should have working experience on at least one data modelling tool,
preferably DBSchema, Erwin
12Good understanding of GCP databases like AlloyDB, CloudSQL, and
BigQuery.
13.People with functional knowledge of the mutual fund industry will be a plus Should be willing to work from Chennai, office presence is mandatory
Role & Responsibilities:
● Work with business users and other stakeholders to understand business processes.
● Ability to design and implement Dimensional and Fact tables
● Identify and implement data transformation/cleansing requirements
● Develop a highly scalable, reliable, and high-performance data processing pipeline to extract, transform and load data from various systems to the Enterprise Data Warehouse
● Develop conceptual, logical, and physical data models with associated metadata including data lineage and technical data definitions
● Design, develop and maintain ETL workflows and mappings using the appropriate data load technique
● Provide research, high-level design, and estimates for data transformation and data integration from source applications to end-user BI solutions.
● Provide production support of ETL processes to ensure timely completion and availability of data in the data warehouse for reporting use.
● Analyze and resolve problems and provide technical assistance as necessary. Partner with the BI team to evaluate, design, develop BI reports and dashboards according to functional specifications while maintaining data integrity and data quality.
● Work collaboratively with key stakeholders to translate business information needs into well-defined data requirements to implement the BI solutions.
● Leverage transactional information, data from ERP, CRM, HRIS applications to model, extract and transform into reporting & analytics.
● Define and document the use of BI through user experience/use cases, prototypes, test, and deploy BI solutions.
● Develop and support data governance processes, analyze data to identify and articulate trends, patterns, outliers, quality issues, and continuously validate reports, dashboards and suggest improvements.
● Train business end-users, IT analysts, and developers.
Role & Responsibilities
Data Organization and Governance: Define and maintain governance standards that span multiple systems (AWS, Fivetran, Snowflake, PostgreSQL, Salesforce/nCino, Looker), ensuring that data remains accurate, accessible, and organized across the organization.
Solve Data Problems Proactively: Address recurring data issues that sidetrack operational and strategic initiatives by implementing processes and tools to anticipate, identify, and resolve root causes effectively.
System Integration: Lead the integration of diverse systems into a cohesive data environment, optimizing workflows and minimizing manual intervention.
Hands-On Problem Solving: Take a hands-on approach to resolving reporting issues and troubleshooting data challenges when necessary, ensuring minimal disruption to business operations.
Collaboration Across Teams: Work closely with business and technical stakeholders to understand and solve our biggest challenges
Mentorship and Leadership: Guide and mentor team members, fostering a culture of accountability and excellence in data management practices.
Strategic Data Support: Ensure that marketing, analytics, and other strategic initiatives are not derailed by data integrity issues, enabling the organization to focus on growth and innovation.
We are seeking a highly skilled and experienced Power BI Lead / Architect to join our growing team. The ideal candidate will have a strong understanding of data warehousing, data modeling, and business intelligence best practices. This role will be responsible for leading the design, development, and implementation of complex Power BI solutions that provide actionable insights to key stakeholders across the organization.
Location - Pune (Hybrid 3 days)
Responsibilities:
Lead the design, development, and implementation of complex Power BI dashboards, reports, and visualizations.
Develop and maintain data models (star schema, snowflake schema) for optimal data analysis and reporting.
Perform data analysis, data cleansing, and data transformation using SQL and other ETL tools.
Collaborate with business stakeholders to understand their data needs and translate them into effective and insightful reports.
Develop and maintain data pipelines and ETL processes to ensure data accuracy and consistency.
Troubleshoot and resolve technical issues related to Power BI dashboards and reports.
Provide technical guidance and mentorship to junior team members.
Stay abreast of the latest trends and technologies in the Power BI ecosystem.
Ensure data security, governance, and compliance with industry best practices.
Contribute to the development and improvement of the organization's data and analytics strategy.
May lead and mentor a team of junior Power BI developers.
Qualifications:
8-12 years of experience in Business Intelligence and Data Analytics.
Proven expertise in Power BI development, including DAX, advanced data modeling techniques.
Strong SQL skills, including writing complex queries, stored procedures, and views.
Experience with ETL/ELT processes and tools.
Experience with data warehousing concepts and methodologies.
Excellent analytical, problem-solving, and communication skills.
Strong teamwork and collaboration skills.
Ability to work independently and proactively.
Bachelor's degree in Computer Science, Information Systems, or a related field preferred.

Senior Data Analyst
Experience: 8+ Years
Work Mode: Remote Full Time
Responsibilities:
• Analyze large datasets to uncover trends, patterns, and insights to support business goals.
• Design, develop, and manage interactive dashboards and reports using Power BI.
• Utilize DAX and SQL for advanced data querying and data modeling.
• Create and manage complex SQL queries for data extraction, transformation, and loading processes.
• Collaborate with cross-functional teams to understand data requirements and translate them into actionable solutions.
• Maintain data accuracy and integrity across projects, ensuring reliable data-driven insights.
• Present findings to stakeholders, translating complex data insights into simple, actionable business recommendations.
Skills:
Power BI, DAX (Data Analysis Expressions), SQL, Data Modeling, Python
Preferred Skills:
• Machine Learning: Exposure to machine learning models and their integration within analytical solutions.
• Microsoft Fabric: Familiarity with Microsoft Fabric for enhanced data integration and management.

Job Description for Data Engineer Role:-
Must have:
Experience working with Programming languages. Solid foundational and conceptual knowledge is expected.
Experience working with Databases and SQL optimizations
Experience as a team lead or a tech lead and is able to independently drive tech decisions and execution and motivate the team in ambiguous problem spaces.
Problem Solving, Judgement and Strategic decisioning skills to be able to drive the team forward.
Role and Responsibilities:
- Share your passion for staying on top of tech trends, experimenting with and learning new technologies, participating in internal & external technology communities, mentoring other members of the engineering community, and from time to time, be asked to code or evaluate code
- Collaborate with digital product managers, and leaders from other team to refine the strategic needs of the project
- Utilize programming languages like Java, Python, SQL, Node, Go, and Scala, Open Source RDBMS and NoSQL databases
- Defining best practices for data validation and automating as much as possible; aligning with the enterprise standards
Qualifications -
- Experience with SQL and NoSQL databases.
- Experience with cloud platforms, preferably AWS.
- Strong experience with data warehousing and data lake technologies (Snowflake)
- Expertise in data modelling
- Experience with ETL/LT tools and methodologies
- Experience working on real-time Data Streaming and Data Streaming platforms
- 2+ years of experience in at least one of the following: Java, Scala, Python, Go, or Node.js
- 2+ years working with SQL and NoSQL databases, data modeling and data management
- 2+ years of experience with AWS, GCP, Azure, or another cloud service.
The Sr. Analytics Engineer would provide technical expertise in needs identification, data modeling, data movement, and transformation mapping (source to target), automation and testing strategies, translating business needs into technical solutions with adherence to established data guidelines and approaches from a business unit or project perspective.
Understands and leverages best-fit technologies (e.g., traditional star schema structures, cloud, Hadoop, NoSQL, etc.) and approaches to address business and environmental challenges.
Provides data understanding and coordinates data-related activities with other data management groups such as master data management, data governance, and metadata management.
Actively participates with other consultants in problem-solving and approach development.
Responsibilities :
Provide a consultative approach with business users, asking questions to understand the business need and deriving the data flow, conceptual, logical, and physical data models based on those needs.
Perform data analysis to validate data models and to confirm the ability to meet business needs.
Assist with and support setting the data architecture direction, ensuring data architecture deliverables are developed, ensuring compliance to standards and guidelines, implementing the data architecture, and supporting technical developers at a project or business unit level.
Coordinate and consult with the Data Architect, project manager, client business staff, client technical staff and project developers in data architecture best practices and anything else that is data related at the project or business unit levels.
Work closely with Business Analysts and Solution Architects to design the data model satisfying the business needs and adhering to Enterprise Architecture.
Coordinate with Data Architects, Program Managers and participate in recurring meetings.
Help and mentor team members to understand the data model and subject areas.
Ensure that the team adheres to best practices and guidelines.
Requirements :
- Strong working knowledge of at least 3 years of Spark, Java/Scala/Pyspark, Kafka, Git, Unix / Linux, and ETL pipeline designing.
- Experience with Spark optimization/tuning/resource allocations
- Excellent understanding of IN memory distributed computing frameworks like Spark and its parameter tuning, writing optimized workflow sequences.
- Experience of relational databases (e.g., PostgreSQL, MySQL) and NoSQL databases (e.g., Redshift, Bigquery, Cassandra, etc).
- Familiarity with Docker, Kubernetes, Azure Data Lake/Blob storage, AWS S3, Google Cloud storage, etc.
- Have a deep understanding of the various stacks and components of the Big Data ecosystem.
- Hands-on experience with Python is a huge plus
Responsibilities:
- Design, implement, and maintain scalable and reliable database solutions on the AWS platform.
- Architect, deploy, and optimize DynamoDB databases for performance, scalability, and cost-efficiency.
- Configure and manage AWS OpenSearch (formerly Amazon Elasticsearch Service) clusters for real-time search and analytics capabilities.
- Design and implement data processing and analytics solutions using AWS EMR (Elastic MapReduce) for large-scale data processing tasks.
- Collaborate with cross-functional teams to gather requirements, design database solutions, and implement best practices.
- Perform performance tuning, monitoring, and troubleshooting of database systems to ensure high availability and performance.
- Develop and maintain documentation, including architecture diagrams, configurations, and operational procedures.
- Stay current with the latest AWS services, database technologies, and industry trends to provide recommendations for continuous improvement.
- Participate in the evaluation and selection of new technologies, tools, and frameworks to enhance database capabilities.
- Provide guidance and mentorship to junior team members, fostering knowledge sharing and skill development.
Requirements:
- Bachelor’s degree in computer science, Information Technology, or related field.
- Proven experience as an AWS Architect or similar role, with a focus on database technologies.
- Hands-on experience designing, implementing, and optimizing DynamoDB databases in production environments.
- In-depth knowledge of AWS OpenSearch (Elasticsearch) and experience configuring and managing clusters for search and analytics use cases.
- Proficiency in working with AWS EMR (Elastic MapReduce) for big data processing and analytics.
- Strong understanding of database concepts, data modelling, indexing, and query optimization.
- Experience with AWS services such as S3, EC2, RDS, Redshift, Lambda, and CloudFormation.
- Excellent problem-solving skills and the ability to troubleshoot complex database issues.
- Solid understanding of cloud security best practices and experience implementing security controls in AWS environments.
- Strong communication and collaboration skills with the ability to work effectively in a team environment.
- AWS certifications such as AWS Certified Solutions Architect, AWS Certified Database - Specialty, or equivalent certifications are a plus.
What’s in it for you?
Opportunity To Unlock Your Creativity
Think of all the times you were held back from trying new ideas because you were boxed in by bureaucratic legacy processes or old school tactics. Having a growth mindset is deeply ingrained into our company culture since day 1 so Fictiv is an environment where you have the creative liberty and support of the team to try big bold ideas to achieve our sales and customer goals.
Opportunity To Grow Your Career
At Fictiv, you'll be surrounded by supportive teammates who will push you to be your best through their curiosity and passion.
Opportunity To Unlock Your Creativity
Think of all the times you were held back from trying new ideas because you were boxed in by bureaucratic legacy processes or old school tactics. Having a growth mindset is deeply ingrained into our company culture since day 1 so Fictiv is an environment where you have the creative liberty and support of the team to try big bold ideas to achieve our sales and customer goals.
Impact In This Role
Excellent problem solving, decision-making and critical thinking skills.
Collaborative, a team player.
Excellent verbal and written communication skills.
Exhibits initiative, integrity and empathy.
Enjoy working with a diverse group of people in multiple regions.
Comfortable not knowing answers, but resourceful and able to resolve issues.
Self-starter; comfortable with ambiguity, asking questions and constantly learning.
Customer service mentality; advocates for another person's point of view.
Methodical and thorough in written documentation and communication.
Culture oriented; wants to work with people rather than in isolation.
You will report to the Director of IT Engineering
What You’ll Be Doing
- Interface with Business Analysts and Stakeholders to understand & clarify requirements
- Develop technical design for solutions Development
- Implement high quality, scalable solutions following best practices, including configuration and code.
- Deploy solutions and code using automated deployment tools
- Take ownership of technical deliverables, ensure that quality work is completed, fully tested, delivered on time.
- Conduct code reviews, optimization, and refactoring to minimize technical debt within Salesforce implementations.
- Collaborate with cross-functional teams to integrate Salesforce with other systems and platforms, ensuring seamless data flow and system interoperability.
- Identify opportunities for process improvements, mentor and support other developers/team members as needed.
- Stay updated on new Salesforce features and functionalities and provide recommendations for process improvements.
Desired Traits
- 8-10 years of experience in Salesforce development
- Proven experience in developing Salesforce solutions with a deep understanding of Apex, Visualforce, Lightning Web Components, and Salesforce APIs.
- Have worked in Salesforce CPQ, Sales/Manufacturing Cloud, Case Management
- Experienced in designing and implementing custom solutions that align with business needs.
- Strong knowledge of Salesforce data modeling, reporting, and database design.
- Demonstrated experience in building and maintaining integrations between Salesforce and external applications.
- Strong unit testing, functional testing and debugging skills
- Strong understanding of best practices
- Active Salesforce Certifications are desirable.
- Experience in Mulesoft is a plus
- Excellent communication skills and the ability to translate complex technical requirements into actionable solutions.
Interested in learning more? We look forward to hearing from you soon.
Responsibilities -
- Collaborate with the development team to understand data requirements and identify potential scalability issues.
- Design, develop, and implement scalable data pipelines and ETL processes to ingest, process, and analyse large - volumes of data from various sources.
- Optimize data models and database schemas to improve query performance and reduce latency.
- Monitor and troubleshoot the performance of our Cassandra database on Azure Cosmos DB, identifying bottlenecks and implementing optimizations as needed.
- Work with cross-functional teams to ensure data quality, integrity, and security.
- Stay up to date with emerging technologies and best practices in data engineering and distributed systems.
Qualifications & Requirements -
- Proven experience as a Data Engineer or similar role, with a focus on designing and optimizing large-scale data systems.
- Strong proficiency in working with NoSQL databases, particularly Cassandra.
- Experience with cloud-based data platforms, preferably Azure Cosmos DB.
- Solid understanding of Distributed Systems, Data modelling, Data Warehouse Designing, and ETL Processes.
- Detailed understanding of Software Development Life Cycle (SDLC) is required.
- Good to have knowledge on any visualization tool like Power BI, Tableau.
- Good to have knowledge on SAP landscape (SAP ECC, SLT, BW, HANA etc).
- Good to have experience on Data Migration Project.
- Knowledge of Supply Chain domain would be a plus.
- Familiarity with software architecture (data structures, data schemas, etc.)
- Familiarity with Python programming language is a plus.
- The ability to work in a dynamic, fast-paced, work environment.
- A passion for data and information with strong analytical, problem solving, and organizational skills.
- Self-motivated with the ability to work under minimal direction.
- Strong communication and collaboration skills, with the ability to work effectively in a cross-functional team environment.
Role Title: Developer - Guidewire Integration-Config
Role Purpose
We are looking for a Developer for our Claims Guidewire team, who is a technology enthusiast, and eager to be part of a culture of modern software engineering practices, continuous improvement, and innovation.
As a Developer, you will be part of a dynamic engineering team and work on development, maintenance, and transformation of our strategic Claims Guidewire platform. You will learn about software applications, technology stack, ways of working and standards.
Key Accountabilities
· Deliver software development tasks for Claims Guidewire applications, in the areas of Integration and Configuration, with expected quality measures and timeframe, e.g., coding, writing unit test cases (G-Unit) and unit testing, debugging and defect fixing, providing test support, providing release support.
· Communicate with technical leads and IT groups for understanding the project’s technical implications, dependencies, and potential conflicts.
· Research issues reported in Production, perform root cause analysis and document them, respond to and resolve technical issues in a timely manner.
· Perform versioning of the release updates and resolve the code conflicts while merging and promoting the code to higher environments.
· Develop their technical and functional knowledge on Claims Digital Guidewire platform.
· Understand and follow Guidewire’s cloud standards for application development.
· Active participation in team meetings like daily stand-ups, risk forums, planning sessions and retrospectives.
Skills & Experience
· 3+ years of development experience on Guidewire cloud platform and applications, Guidewire certification preferred.
· Hands on development expertise on Guidewire ClaimCentre with configuration and integration
· Experience in Guidewire platform (Gosu scripting / Edge APIs / UI / Data Model)
· Should have knowledge on Admin data loading, Assignment and Segmentation Rules, Pre-update and Validation rules, Authority limits, Financials (checks, reserves, recoveries …)
· Good experience on LOB configuration and related type-lists
· Good experience on integration components including plug-ins, messaging (and supporting business rules), batches, REST APIs and programs that call the Guidewire application APIs.
· Experience on any database Oracle / SQL Server and well versed in SQL.
· Experience of working in a CI/CD setup and related tools/technologies
· Insurance domain knowledge with Property & Casualty background preferred.
Location- Gurugram
CTC- Upto 25lpa

Key Roles/Responsibilities: –
• Develop an understanding of business obstacles, create
• solutions based on advanced analytics and draw implications for
• model development
• Combine, explore and draw insights from data. Often large and
• complex data assets from different parts of the business.
• Design and build explorative, predictive- or prescriptive
• models, utilizing optimization, simulation and machine learning
• techniques
• Prototype and pilot new solutions and be a part of the aim
• of ‘productifying’ those valuable solutions that can have impact at a
• global scale
• Guides and coaches other chapter colleagues to help solve
• data/technical problems at an operational level, and in
• methodologies to help improve development processes
• Identifies and interprets trends and patterns in complex data sets to
• enable the business to take data-driven decisions
Minimum of 8 years of experience of which, 4 years should be of applied data mining
experience in disciplines such as Call Centre Metrics.
Strong experience in advanced statistics and analytics including segmentation, modelling, regression, forecasting etc.
Experience with leading and managing large teams.
Demonstrated pattern of success in using advanced quantitative analytic methods to solve business problems.
Demonstrated experience with Business Intelligence/Data Mining tools to work with
data, investigate anomalies, construct data sets, and build models.
Critical to share details on projects undertaken (preferably on telecom industry)
specifically through analysis from CRM.

About the role:
Hopscotch is looking for a passionate Data Engineer to join our team. You will work closely with other teams like data analytics, marketing, data science and individual product teams to specify, validate, prototype, scale, and deploy data pipelines features and data architecture.
Here’s what will be expected out of you:
➢ Ability to work in a fast-paced startup mindset. Should be able to manage all aspects of data extraction transfer and load activities.
➢ Develop data pipelines that make data available across platforms.
➢ Should be comfortable in executing ETL (Extract, Transform and Load) processes which include data ingestion, data cleaning and curation into a data warehouse, database, or data platform.
➢ Work on various aspects of the AI/ML ecosystem – data modeling, data and ML pipelines.
➢ Work closely with Devops and senior Architect to come up with scalable system and model architectures for enabling real-time and batch services.
What we want:
➢ 5+ years of experience as a data engineer or data scientist with a focus on data engineering and ETL jobs.
➢ Well versed with the concept of Data warehousing, Data Modelling and/or Data Analysis.
➢ Experience using & building pipelines and performing ETL with industry-standard best practices on Redshift (more than 2+ years).
➢ Ability to troubleshoot and solve performance issues with data ingestion, data processing & query execution on Redshift.
➢ Good understanding of orchestration tools like Airflow.
➢ Strong Python and SQL coding skills.
➢ Strong Experience in distributed systems like spark.
➢ Experience with AWS Data and ML Technologies (AWS Glue,MWAA, Data Pipeline,EMR,Athena, Redshift,Lambda etc).
➢ Solid hands on with various data extraction techniques like CDC or Time/batch based and the related tools (Debezium, AWS DMS, Kafka Connect, etc) for near real time and batch data extraction.
Note :
Product based companies, Ecommerce companies is added advantage

Requirements:
- 2+ years of experience (4+ for Senior Data Engineer) with system/data integration, development or implementation of enterprise and/or cloud software Engineering degree in Computer Science, Engineering or related field.
- Extensive hands-on experience with data integration/EAI technologies (File, API, Queues, Streams), ETL Tools and building custom data pipelines.
- Demonstrated proficiency with Python, JavaScript and/or Java
- Familiarity with version control/SCM is a must (experience with git is a plus).
- Experience with relational and NoSQL databases (any vendor) Solid understanding of cloud computing concepts.
- Strong organisational and troubleshooting skills with attention to detail.
- Strong analytical ability, judgment and problem-solving techniques Interpersonal and communication skills with the ability to work effectively in a cross functional team.


Lifespark is looking for individuals with a passion for impacting real lives through technology. Lifespark is one of the most promising startups in the Assistive Tech space in India, and has been honoured with several National and International awards. Our mission is to create seamless, persistent and affordable healthcare solutions. If you are someone who is driven to make a real impact in this world, we are your people.
Lifespark is currently building solutions for Parkinson’s Disease, and we are looking for a ML lead to join our growing team. You will be working directly with the founders on high impact problems in the Neurology domain. You will be solving some of the most fundamental and exciting challenges in the industry and will have the ability to see your insights turned into real products every day
Essential experience and requirements:
1. Advanced knowledge in the domains of computer vision, deep learning
2. Solid understand of Statistical / Computational concepts like Hypothesis Testing, Statistical Inference, Design of Experiments and production level ML system design
3. Experienced with proper project workflow
4. Good at collating multiple datasets (potentially from different sources)
5. Good understanding of setting up production level data pipelines
6. Ability to independently develop and deploy ML systems to various platforms (local and cloud)
7. Fundamentally strong with time-series data analysis, cleaning, featurization and visualisation
8. Fundamental understanding of model and system explainability
9. Proactive at constantly unlearning and relearning
10. Documentation ninja - can understand others documentation as well as create good documentation
Responsibilities :
1. Develop and deploy ML based systems built upon healthcare data in the Neurological domain
2. Maintain deployed systems and upgrade them through online learning
3. Develop and deploy advanced online data pipelines
- Big data developer with 8+ years of professional IT experience with expertise in Hadoop ecosystem components in ingestion, Data modeling, querying, processing, storage, analysis, Data Integration and Implementing enterprise level systems spanning Big Data.
- A skilled developer with strong problem solving, debugging and analytical capabilities, who actively engages in understanding customer requirements.
- Expertise in Apache Hadoop ecosystem components like Spark, Hadoop Distributed File Systems(HDFS), HiveMapReduce, Hive, Sqoop, HBase, Zookeeper, YARN, Flume, Pig, Nifi, Scala and Oozie.
- Hands on experience in creating real - time data streaming solutions using Apache Spark core, Spark SQL & DataFrames, Kafka, Spark streaming and Apache Storm.
- Excellent knowledge of Hadoop architecture and daemons of Hadoop clusters, which include Name node,Data node, Resource manager, Node Manager and Job history server.
- Worked on both Cloudera and Horton works in Hadoop Distributions. Experience in managing Hadoop clustersusing Cloudera Manager tool.
- Well versed in installation, Configuration, Managing of Big Data and underlying infrastructure of Hadoop Cluster.
- Hands on experience in coding MapReduce/Yarn Programs using Java, Scala and Python for analyzing Big Data.
- Exposure to Cloudera development environment and management using Cloudera Manager.
- Extensively worked on Spark using Scala on cluster for computational (analytics), installed it on top of Hadoop performed advanced analytical application by making use of Spark with Hive and SQL/Oracle .
- Implemented Spark using PYTHON and utilizing Data frames and Spark SQL API for faster processing of data and handled importing data from different data sources into HDFS using Sqoop and performing transformations using Hive, MapReduce and then loading data into HDFS.
- Used Spark Data Frames API over Cloudera platform to perform analytics on Hive data.
- Hands on experience in MLlib from Spark which are used for predictive intelligence, customer segmentation and for smooth maintenance in Spark streaming.
- Experience in using Flume to load log files into HDFS and Oozie for workflow design and scheduling.
- Experience in optimizing MapReduce jobs to use HDFS efficiently by using various compression mechanisms.
- Working on creating data pipeline for different events of ingestion, aggregation, and load consumer response data into Hive external tables in HDFS location to serve as feed for tableau dashboards.
- Hands on experience in using Sqoop to import data into HDFS from RDBMS and vice-versa.
- In-depth Understanding of Oozie to schedule all Hive/Sqoop/HBase jobs.
- Hands on expertise in real time analytics with Apache Spark.
- Experience in converting Hive/SQL queries into RDD transformations using Apache Spark, Scala and Python.
- Extensive experience in working with different ETL tool environments like SSIS, Informatica and reporting tool environments like SQL Server Reporting Services (SSRS).
- Experience in Microsoft cloud and setting cluster in Amazon EC2 & S3 including the automation of setting & extending the clusters in AWS Amazon cloud.
- Extensively worked on Spark using Python on cluster for computational (analytics), installed it on top of Hadoop performed advanced analytical application by making use of Spark with Hive and SQL.
- Strong experience and knowledge of real time data analytics using Spark Streaming, Kafka and Flume.
- Knowledge in installation, configuration, supporting and managing Hadoop Clusters using Apache, Cloudera (CDH3, CDH4) distributions and on Amazon web services (AWS).
- Experienced in writing Ad Hoc queries using Cloudera Impala, also used Impala analytical functions.
- Experience in creating Data frames using PySpark and performing operation on the Data frames using Python.
- In depth understanding/knowledge of Hadoop Architecture and various components such as HDFS and MapReduce Programming Paradigm, High Availability and YARN architecture.
- Establishing multiple connections to different Redshift clusters (Bank Prod, Card Prod, SBBDA Cluster) and provide the access for pulling the information we need for analysis.
- Generated various kinds of knowledge reports using Power BI based on Business specification.
- Developed interactive Tableau dashboards to provide a clear understanding of industry specific KPIs using quick filters and parameters to handle them more efficiently.
- Well Experience in projects using JIRA, Testing, Maven and Jenkins build tools.
- Experienced in designing, built, and deploying and utilizing almost all the AWS stack (Including EC2, S3,), focusing on high-availability, fault tolerance, and auto-scaling.
- Good experience with use-case development, with Software methodologies like Agile and Waterfall.
- Working knowledge of Amazon's Elastic Cloud Compute( EC2 ) infrastructure for computational tasks and Simple Storage Service ( S3 ) as Storage mechanism.
- Good working experience in importing data using Sqoop, SFTP from various sources like RDMS, Teradata, Mainframes, Oracle, Netezza to HDFS and performed transformations on it using Hive, Pig and Spark .
- Extensive experience in Text Analytics, developing different Statistical Machine Learning solutions to various business problems and generating data visualizations using Python and R.
- Proficient in NoSQL databases including HBase, Cassandra, MongoDB and its integration with Hadoop cluster.
- Hands on experience in Hadoop Big data technology working on MapReduce, Pig, Hive as Analysis tool, Sqoop and Flume data import/export tools.