4+ ETL Jobs in Kochi (Cochin) | ETL Job openings in Kochi (Cochin)
Apply to 4+ ETL Jobs in Kochi (Cochin) on CutShort.io. Explore the latest ETL Job opportunities across top companies like Google, Amazon & Adobe.
Job Description & Specification:
Post Title: Data Engineer
Work Mode: Kochi Onsite - UK Time zone
Role Overview:
We are seeking a talented and experienced Data Engineer to join our team. The ideal candidate will have expertise in technologies such as Metabases, Dbt, Stitch, Snowflake, Avo, and MongoDB. As a Data Engineer, you will play a crucial role in designing, developing, and maintaining our data infrastructure to support our analytics and data-driven decision-making processes.
Responsibilities:
- Designing, developing and implementing scalable data pipelines and ETL processes using tools such as Stitch and Dbt to ingest, transform, and load data from various sources into our data warehouse (Snowflake).
- Implement data modeling best practices and standards using Dbt to create and manage data models for reporting and analytics.
- Collaborating with cross-functional teams to understand data requirements and deliver solutions that meet business needs.
- Develop and maintain dashboards and visualizations in Metabases to enable self-service analytics and data exploration for internal teams.
- Building and optimizing ETL processes to ensure data quality and integrity.
- Optimizing data processing and storage solutions for performance, scalability and reliability, leveraging cloud-based technologies.
- Implementing monitoring and alerting systems to proactively identify and address data issues.
- Implementing data quality checks and monitoring processes to ensure the accuracy, completeness, and integrity of data.
- Managing and optimizing databases (like MongoDB for performance and scalability).
- Developing and maintaining documentation, best practices, and standards for data engineering processes and workflows.
- Stay up to date with emerging technologies and trends in data engineering, machine learning, and analytics, and evaluate their potential impact on data strategy and architecture.
Requirements:
- Bachelor's or Master's degree in Computer Science.
- Minimum of 4 years of experience working as a data engineer with expertise in Metabases, Dbt, Stitch, Snowflake, Avo, MongoDB.
- Strong programming skills in languages like Python, and experience with SQL and database technologies (e.g., PostgreSQL, MySQL, MongoDB).
- Hands-on experience with data integration tools (e.g., Stitch), data modeling tools (e.g., Dbt), and BI platforms (e.g., Metabases).
- Experience with cloud platforms such as AWS.
- Strong understanding of data modeling concepts, database design, and data warehousing principles
- Experience with big data technologies and frameworks (e.g., Hadoop, Spark, Kafka) and cloud-based data platforms (e.g., AWS EMR, Azure Databricks, Google BigQuery).
- Familiarity with data integration tools, ETL processes, and workflow orchestration tools (e.g., Apache Airflow, Apache NiFi).
- Excellent problem-solving skills and attention to detail.
- Strong communication skills with the ability to work effectively in a global team environment.
- Experience in the education or Edtech industry is a plus.
- Knowledge of Avo for schema management and versioning will be an added advantage.
- Familiarity with machine learning algorithms, data science workflows, and analytics tools (e.g., TensorFlow, PyTorch, scikit-learn, Tableau).
- Knowledge of distributed computing concepts and containerization technologies.
- Experience with version control systems (e.g., Git) and CI/CD pipelines.
- Certifications in cloud computing (e.g., AWS Certified Developer, Google Cloud Professional Data Engineer) or data engineering (e.g., Databricks Certified Associate Developer) are desirable.
Benefits:
- Competitive salary and bonus structure based on performance and achievement of goals.
- Comprehensive benefits package including medical insurance.
Join us in shaping the future of technology by applying your expertise as a Data Engineer. If you are passionate about driving innovation and delivering impactful solutions, we invite you to be part of our dynamic team. Apply now!!
JOB DETAILS:
* Job Title: Associate III - Data Engineering
* Industry: Global digital transformation solutions provide
* Salary: Best in Industry
* Experience: 4-6 years
* Location: Trivandrum, Kochi
Job Description
Job Title:
Data Services Engineer – AWS & Snowflake
Job Summary:
As a Data Services Engineer, you will be responsible for designing, developing, and maintaining robust data solutions using AWS cloud services and Snowflake.
You will work closely with cross-functional teams to ensure data is accessible, secure, and optimized for performance.
Your role will involve implementing scalable data pipelines, managing data integration, and supporting analytics initiatives.
Responsibilities:
• Design and implement scalable and secure data pipelines on AWS and Snowflake (Star/Snowflake schema)
• Optimize query performance using clustering keys, materialized views, and caching
• Develop and maintain Snowflake data warehouses and data marts.
• Build and maintain ETL/ELT workflows using Snowflake-native features (Snowpipe, Streams, Tasks).
• Integrate Snowflake with cloud platforms (AWS, Azure, GCP) and third-party tools (Airflow, dbt, Informatica)
• Utilize Snowpark and Python/Java for complex transformations
• Implement RBAC, data masking, and row-level security.
• Optimize data storage and retrieval for performance and cost-efficiency.
• Collaborate with stakeholders to gather data requirements and deliver solutions.
• Ensure data quality, governance, and compliance with industry standards.
• Monitor, troubleshoot, and resolve data pipeline and performance issues.
• Document data architecture, processes, and best practices.
• Support data migration and integration from various sources.
Qualifications:
• Bachelor’s degree in Computer Science, Information Technology, or a related field.
• 3 to 4 years of hands-on experience in data engineering or data services.
• Proven experience with AWS data services (e.g., S3, Glue, Redshift, Lambda).
• Strong expertise in Snowflake architecture, development, and optimization.
• Proficiency in SQL and Python for data manipulation and scripting.
• Solid understanding of ETL/ELT processes and data modeling.
• Experience with data integration tools and orchestration frameworks.
• Excellent analytical, problem-solving, and communication skills.
Preferred Skills:
• AWS Glue, AWS Lambda, Amazon Redshift
• Snowflake Data Warehouse
• SQL & Python
Skills: Aws Lambda, AWS Glue, Amazon Redshift, Snowflake Data Warehouse
Must-Haves
AWS data services (4-6 years), Snowflake architecture (4-6 years), SQL (proficient), Python (proficient), ETL/ELT processes (solid understanding)
Skills: AWS, AWS lambda, Snowflake, Data engineering, Snowpipe, Data integration tools, orchestration framework
Relevant 4 - 6 Years
python is mandatory
******
Notice period - 0 to 15 days only (Feb joiners’ profiles only)
Location: Kochi
F2F Interview 7th Feb

Global digital transformation solutions provider.
Job Description – Senior Technical Business Analyst
Location: Trivandrum (Preferred) | Open to any location in India
Shift Timings - 8 hours window between the 7:30 PM IST - 4:30 AM IST
About the Role
We are seeking highly motivated and analytically strong Senior Technical Business Analysts who can work seamlessly with business and technology stakeholders to convert a one-line problem statement into a well-defined project or opportunity. This role is ideal for fresh graduates who have a strong foundation in data analytics, data engineering, data visualization, and data science, along with a strong drive to learn, collaborate, and grow in a dynamic, fast-paced environment.
As a Technical Business Analyst, you will be responsible for translating complex business challenges into actionable user stories, analytical models, and executable tasks in Jira. You will work across the entire data lifecycle—from understanding business context to delivering insights, solutions, and measurable outcomes.
Key Responsibilities
Business & Analytical Responsibilities
- Partner with business teams to understand one-line problem statements and translate them into detailed business requirements, opportunities, and project scope.
- Conduct exploratory data analysis (EDA) to uncover trends, patterns, and business insights.
- Create documentation including Business Requirement Documents (BRDs), user stories, process flows, and analytical models.
- Break down business needs into concise, actionable, and development-ready user stories in Jira.
Data & Technical Responsibilities
- Collaborate with data engineering teams to design, review, and validate data pipelines, data models, and ETL/ELT workflows.
- Build dashboards, reports, and data visualizations using leading BI tools to communicate insights effectively.
- Apply foundational data science concepts such as statistical analysis, predictive modeling, and machine learning fundamentals.
- Validate and ensure data quality, consistency, and accuracy across datasets and systems.
Collaboration & Execution
- Work closely with product, engineering, BI, and operations teams to support the end-to-end delivery of analytical solutions.
- Assist in development, testing, and rollout of data-driven solutions.
- Present findings, insights, and recommendations clearly and confidently to both technical and non-technical stakeholders.
Required Skillsets
Core Technical Skills
- 6+ years of Technical Business Analyst experience within an overall professional experience of 8+ years
- Data Analytics: SQL, descriptive analytics, business problem framing.
- Data Engineering (Foundational): Understanding of data warehousing, ETL/ELT processes, cloud data platforms (AWS/GCP/Azure preferred).
- Data Visualization: Experience with Power BI, Tableau, or equivalent tools.
- Data Science (Basic/Intermediate): Python/R, statistical methods, fundamentals of ML algorithms.
Soft Skills
- Strong analytical thinking and structured problem-solving capability.
- Ability to convert business problems into clear technical requirements.
- Excellent communication, documentation, and presentation skills.
- High curiosity, adaptability, and eagerness to learn new tools and techniques.
Educational Qualifications
- BE/B.Tech or equivalent in:
- Computer Science / IT
- Data Science
What We Look For
- Demonstrated passion for data and analytics through projects and certifications.
- Strong commitment to continuous learning and innovation.
- Ability to work both independently and in collaborative team environments.
- Passion for solving business problems using data-driven approaches.
- Proven ability (or aptitude) to convert a one-line business problem into a structured project or opportunity.
Why Join Us?
- Exposure to modern data platforms, analytics tools, and AI technologies.
- A culture that promotes innovation, ownership, and continuous learning.
- Supportive environment to build a strong career in data and analytics.
Skills: Data Analytics, Business Analysis, Sql
Must-Haves
Technical Business Analyst (6+ years), SQL, Data Visualization (Power BI, Tableau), Data Engineering (ETL/ELT, cloud platforms), Python/R
• Lead the Development Team.
• Strong in SQL Coding and performance tuning.
• Strong in Data warehousing concepts and design.
• Strong in ETL process and implementation.
• Participate in database design and architecture to support application development
activities.
• Experience in writing and reviewing complex functions, stored procedures, and custom
scripts to support application development.
• Tuning application queries by analysing execution plans.
• Experience in data migration between different DB systems.
• Expertise in trouble shooting day to day production issues in DB.
• Oversee progress of development team to ensure consistency with initial design
Skills and Qualifications
• 8+ years of professional experience with at least 2 years as an architect.
• Deep knowledge in SQL Server is a must. Knowledge in MySQL and Vertica would be an
asset.
• Ability to work in a very dynamic environment & ready to support production
environment.
• Organised, self-motivated, team player, actions and result oriented.
• Ability to successfully work under tight timelines.
• Good verbal and written communication skills.
• Ability to meet production targets and milestones.
