Senior Data Engineer (Integration)

at Top Multinational Fintech Startup

Agency job
Bengaluru (Bangalore)
4 - 7 yrs
₹20L - ₹30L / yr
Full time
Data engineering
Big Data
Elastic Search
Roles & Responsibilties
What will you do?
  • Deliver plugins for our Python-based ETL pipelines
  • Deliver Python microservices for provisioning and managing cloud infrastructure
  • Implement algorithms to analyse large data sets
  • Draft design documents that translate requirements into code
  • Effectively manage challenges associated with handling large volumes of data working to tight deadlines
  • Manage expectations with internal stakeholders and context-switch in a fast-paced environment
  • Thrive in an environment that uses AWS and Elasticsearch extensively
  • Keep abreast of technology and contribute to the engineering strategy
  • Champion best development practices and provide mentorship to others
What are we looking for?
  • First and foremost you are a Python developer, experienced with the Python Data stack
  • You love and care about data
  • Your code is an artistic manifest reflecting how elegant you are in what you do
  • You feel sparks of joy when a new abstraction or pattern arises from your code
  • You support the manifests DRY (Don’t Repeat Yourself) and KISS (Keep It Short and Simple)
  • You are a continuous learner
  • You have a natural willingness to automate tasks
  • You have critical thinking and an eye for detail
  • Excellent ability and experience of working to tight deadlines
  • Sharp analytical and problem-solving skills
  • Strong sense of ownership and accountability for your work and delivery
  • Excellent written and oral communication skills
  • Mature collaboration and mentoring abilities
  • We are keen to know your digital footprint (community talks, blog posts, certifications, courses you have participated in or you are keen to, your personal projects as well as any kind of contributions to the open-source communities if any)
Nice to have:
  • Delivering complex software, ideally in a FinTech setting
  • Experience with CI/CD tools such as Jenkins, CircleCI
  • Experience with code versioning (git / mercurial / subversion)
Why apply to jobs via Cutshort
Personalized job matches
Stop wasting time. Get matched with jobs that meet your skills, aspirations and preferences.
Verified hiring teams
See actual hiring teams, find common social connections or connect with them directly. No 3rd party agencies here.
Move faster with AI
We use AI to get you faster responses, recommendations and unmatched user experience.
Matches delivered
Network size
Companies hiring

Similar jobs

Sr. AL Engineer

at Matellio India Private Limited

Founded 1998  •  Services  •  100-1000 employees  •  Profitable
Machine Learning (ML)
Data Science
Natural Language Processing (NLP)
Computer Vision
Deep Learning
Linear regression
Linear algebra
Big Data
Artificial Intelligence (AI)
Remote only
8 - 15 yrs
₹10L - ₹27L / yr

Responsibilities include: 

  • Convert the machine learning models into application program interfaces (APIs) so that other applications can use it
  • Build AI models from scratch and help the different components of the organization (such as product managers and stakeholders) understand what results they gain from the model
  • Build data ingestion and data transformation infrastructure
  • Automate infrastructure that the data science team uses
  • Perform statistical analysis and tune the results so that the organization can make better-informed decisions
  • Set up and manage AI development and product infrastructure
  • Be a good team player, as coordinating with others is a must
Job posted by
Harshit Sharma

Analytics Manager


Founded 2014  •  Services  •  100-1000 employees  •  Profitable
Amazon Redshift
Bengaluru (Bangalore)
7 - 12 yrs
Best in industry

Company Overview:


At Porter, we are passionate about improving productivity. We want to help businesses, large and small, optimize their last-mile operations and empower them to unleash the growth of their core functions. Last mile delivery logistics is one of the biggest and fastest growing sectors of the economy with a market cap upwards of 50 billion USD and a growth rate exceeding 15% CAGR.


Porter is the fastest growing leader in this sector with operations in 14 major cities, a fleet size exceeding 1L registered and 50k active driver partners and a customer base with 3.5M being monthly active. Our industry-best technology platform has raised over 50 million USD from investors including Sequoia Capital, Kae Capital, Mahindra group and LGT Aspada.


We are addressing a massive problem and going after a huge market. We’re trying to create a household name in transportation and our ambition is to disrupt all facets of last mile logistics including warehousing and LTL transportation. At Porter, we’re here to do the best work of our lives.


If you want to do the same and love the challenges and opportunities of a fast paced work environment, then we believe Porter is the right place for you.



Company URL:


  • This role requires a person to support & drive business charters & accompanying products by aligning with business & product stakeholders and building & nurturing a team to aid in this. Responsibilities include


Business and product alignment | Analytics scoping and planning

  • Understand and align business and product initiatives working with stakeholders.
  • Help shape business / product requirements formulation into analytics / data scope.
  • Formalize requirements into roadmaps and iterations - accounting for dependencies.

Project Management

  • Drive project execution using effective prioritization and resource allocation.
  • Resolve blockers through technical expertise, negotiation, and delegation.
  • Strive for completeness of solution through stand-ups and course-correction.


KPIs and metrics

  • Drive business-level KPIs
  • Ensure all initiatives and product features resonate with and amplify those KPIs.
  • Responsible for high-level design and data architecture impacting performance.
  • Envision data APIs and set up proper ownership structures.

Team Management

  • Manage a team of 5-8 members and a portfolio of 1-2 business charters.
  • Do regular one-on-ones with reportees to ensure resource welfare.
  • Periodic assessment and actionable feedback for progress.
  • Recruit new members with a view to long-term resource planning through effective collaboration

with the hiring team.

Process design

  • Formulate SLA benchmarks and set up processes to ensure on-time ad-hoc requests resolution + issues tracking + solving for inefficiencies and effort duplication.
  • Establish good code review practices - using this as a nurturing tool.
  • Set up communication channels (blog / newsletter) and feedback loops for knowledge sharing and stakeholder management. 
  • Explore the latest best practices and tools for constant upskilling.

Analytics Stack

  • Analytics : Python / R / SQL + Excel / PPT, Colab notebooks
  • Database : PostgreSQL, Amazon Redshift, DynamoDB, Aerospike
  • Warehouse : Amazon Redshift
  • ETL : Lots of Python + custom-made
  • Business Intelligence / Visualization : Metabase + Python/R libraries (location data) + Dash
  • Deployment pipeline : Docker, Jenkins, AWS Lambda
  • Collaboration : Git, Dropbox Paper

Qualification Prerequisites

  • Analytics experience of minimum 7 years
  • Experience managing an Analytics team of at least 4 people end-to-end
  • Exposure to consumer facing products (product org experience preferred)
  • Strong technical background and ability to contribute to design and review
  • Familiarity with our current or a similar analytics stack
Job posted by
Satyajit Mittra

Database Developer

at Mobile Programming LLC

Founded 1998  •  Services  •  100-1000 employees  •  Profitable
Data engineering
Big Data
Google Cloud Platform (GCP)
Remote only
4 - 7 yrs
₹10L - ₹14L / yr

Job ID: RP100

Work Location: Remote

Required Experience: 4 to 7 years

Job Description

  • Must have Google Cloud Big Query experience
  • Strong experience with data analysis, data modeling and governance, with excellent analytical and problem-solving abilities
  • Good knowledge of Data Warehouses, data flow ETL pipelines
  • Design, configuration/administration of database software in Cloud platform.
  • Monitoring, Troubleshooting, and Performance tuning the DB objects.
  • Experience on Table Partition, Clustered Table, Materialized View, External Tables etc.

Anyone RDBMS technologies

  • Good experience in DB design with knowledge of ER Diagram, PK/FK, Stored procedure, Function, Triggers, and Indexes.
  • Understanding the requirement of the App team and creating the necessary DB objects by following the best practices.
  • Managing logins and database users, as well as database roles, application roles, and other security principles within the database.
  • Deep knowledge about Indexes, Performance tuning, and Complex SQL Query patterns.
  • Monitoring, Tuning, and Troubleshooting the database-related issues.

About Us:

Mobile programming LLC is a US-based digital transformation company. We help enterprises transform ideas into innovative and intelligent solutions, governing the Internet of Things, Digital Commerce, Business Intelligence Analytics, and Cloud Programming. Bring your challenges to us, we will give you the smartest solutions. From conceptualizing and engineering to advanced manufacturing, we help customers build and scale products fit for the global marketplace.

Mobile programming LLC has offices located in Los Angeles, San Jose, Glendale, San Diego, Phoenix, Plano, New York, Fort Lauderdale, and Boston. Mobile programming is SAP Preferred Vendor, Apple Adjunct Partner, Google Empaneled Mobile Vendor, and Microsoft Gold Certified Partner.

Job posted by
Garima Walia

Data Engineer

at surusha technology Pvt Ltd

Founded 2016  •  Products & Services  •  20-100 employees  •  Profitable
SQL Azure
Data engineering
Remote only
3 - 6 yrs
₹3L - ₹6L / yr
Role - Data Engineer

Skillsets-Azure, Olap, Etl, sql, python, c#

exp range - 3+ to 4 years

salary-best in industry

notice period - Currently serving notice period (Immediate joiners are preferred)

location- remote work

job type -permanent role

it is full time and totally remote based

Note: For the interview 3 rounds are there -technical round, manager/client round, hr round
Job posted by
subham kumar

Data Lead

at Blue Sky Analytics

Founded 2018  •  Product  •  20-100 employees  •  Raised funding
Remote sensing
Data Science
GIS analysis
Remote only
5 - 10 yrs
₹8L - ₹25L / yr

About the Company

Blue Sky Analytics is a Climate Tech startup that combines the power of AI & Satellite data to aid in the creation of a global environmental data stack. Our funders include Beenext and Rainmatter. Over the next 12 months, we aim to expand to 10 environmental data-sets spanning water, land, heat, and more!

We are looking for a Data Lead - someone who works at the intersection of data science, GIS, and engineering. We want a leader who not only understands environmental data but someone who can quickly assemble large scale datasets that are crucial to the well being of our planet. Come save the planet with us!

Your Role

Manage: As a leadership position, this requires long term strategic thinking. You will be in charge of daily operations of the data team. This would include running team standups, planning the execution of data generation and ensuring the algorithms are put in production. You will also be the person in charge to dumb down the data science for the rest of us who do not know what it means.

Love and Live Data: You will also be taking all the responsibility of ensuring that the data we generate is accurate, clean, and is ready to use for our clients. This would entail that you understand what the market needs, calculate feasibilities and build data pipelines. You should understand the algorithms that we use or need to use and take decisions on what would serve the needs of our clients well. We also want our Data Lead to be constantly probing for newer and optimized ways of generating datasets. It would help if they were abreast of all the latest developments in the data science and environmental worlds. The Data Lead also has to be able to work with our Platform team on integrating the data on our platform and API portal.

Collaboration: We use Clubhouse to track and manage our projects across our organization - this will require you to collaborate with the team and follow up with members on a regular basis. About 50% of the work, needs to be the pulse of the platform team. You'll collaborate closely with peers from other functions—Design, Product, Marketing, Sales, and Support to name a few—on our overall product roadmap, on product launches, and on ongoing operations. You will find yourself working with the product management team to define and execute the feature roadmap. You will be expected to work closely with the CTO, reporting on daily operations and development. We don't believe in a top-down hierarchical approach and are transparent with everyone. This means honest and mutual feedback and ability to adapt.

Teaching: Not exactly in the traditional sense. You'll recruit, coach, and develop engineers while ensuring that they are regularly receiving feedback and making rapid progress on personal and professional goals.

Humble and cool: Look we will be upfront with you about one thing - our team is fairly young and is always buzzing with work. In this fast-paced setting, we are looking for someone who can stay cool, is humble, and is willing to learn. You are adaptable, can skill up fast, and are fearless at trying new methods. After all, you're in the business of saving the planet!


  • A minimum of 5 years of industry experience.
  • Hyper-curious!
  • Exceptional at Remote Sensing Data, GIS, Data Science.
  • Must have big data & data analytics experience
  • Very good in documentation & speccing datasets
  • Experience with AWS Cloud, Linux, Infra as Code & Docker (containers) is a must
  • Coordinate with cross-functional teams (DevOPS, QA, Design etc.) on planning and execution
  • Lead, mentor and manage deliverables of a team of talented and highly motivated team of developers
  • Must have experience in building, managing, growing & hiring data teams. Has built large-scale datasets from scratch
  • Managing work on team's Clubhouse & follows up with the team. ~ 50% of work, needs to be the pulse of the platform team
  • Exceptional communication skills & ability to abstract away problems & build systems. Should be able to explain to the management anything & everything
  • Quality control - you'll be responsible for maintaining a high quality bar for everything your team ships. This includes documentation and data quality
  • Experience of having led smaller teams, would be a plus.


  • Work from anywhere: Work by the beach or from the mountains.
  • Open source at heart: We are building a community where you can use, contribute and collaborate on.
  • Own a slice of the pie: Possibility of becoming an owner by investing in ESOPs.
  • Flexible timings: Fit your work around your lifestyle.
  • Comprehensive health cover: Health cover for you and your dependents to keep you tension free.
  • Work Machine of choice: Buy a device and own it after completing a year at BSA.
  • Quarterly Retreats: Yes there's work-but then there's all the non-work+fun aspect aka the retreat!
  • Yearly vacations: Take time off to rest and get ready for the next big assignment by availing the paid leaves.
Job posted by
Balahun Khonglanoh

Machine Learning Engineer

at Essenvia

Founded 2018  •  Product  •  0-20 employees  •  Raised funding
Machine Learning (ML)
Deep Learning
Natural Language Processing (NLP)
Optical character recognition
Text mining
Elastic Search
Bengaluru (Bangalore)
3 - 6 yrs
₹12L - ₹20L / yr

Essenvia is an cloud based SaaS platform that helps medical device companies reduce the time and cost of bringing Medical Devices to market. It’s product suite includes collaborative multiuser  platform to prepare regulatory submissions, document management system, streamline the Medical Device regulatory pathway.

We are looking for a savvy Machine learning Engineer to join our team based out of Bangalore. The hire will be responsible for creating and managing proprietary data set for machine learning algorithms using various conventional and non-conventional data sources. The Engineer will support initiatives and will ensure optimal data delivery architecture for machine learning models. The right candidate will be excited by the prospect of becoming a key member in designing the data architecture to support our next generation of products, must be self-driven and  able to work on tight time line in start-up culture


Extract key information from various data sources

Process documents using OCR and extract key entities
Extract blocks of relevant texts using pattern recognition

Prepare structured and unstructured data pipeline for machine learning models
Assemble large, complex data sets using various data sets.

Mandatory Skills

Knowledge of algorithm and data structure
Programming Knowledge in Python and Java
Knowledge of Text mining/ Text extraction/ Regex matching
Knowledge of OCR

Experience in data cleaning, ETL, pipeline building and model-maintenance using Airflow
Knowledge Elastic search, Neo4j and GraphQL

Desirable Skills

Knowledge of NLP
Knowledge of preparing and using custom corpora

Prior experience in medical science datasets

Exposure to Deep Learning applications and tools like TensorFlow, Theano is preferred

Job posted by
Basant Sahoo

ELK Developer at NSEIT @Bangalore


Founded 1999  •  Products & Services  •  100-1000 employees  •  Profitable
ELK Stack
Elastic Search
Bengaluru (Bangalore)
3 - 5 yrs
₹6L - ₹20L / yr
• Introduction: ELK(Elasticsearch, Logstash, and Kibana) stack. ELK (Elasticsearch, Logstash and Kibana) stack is an end-to-end stack that delivers actionable insights in real time from almost any type of structured and unstructured data source. ELK Stack is the most popular log management platform.

• Responsibilities:
o Should be able to work with API, shards etc in Elasticsearch.
o Write parser in Logstash
o Create Dashboards in Kibana

• Mandatory Experience.
o Must have very good understanding of Log Analytics
o Hands on experience in Elasticsearch, logstash & Kibana should be at expert level
o Elasticsearch : Should be able to write Kibana API
o Logstash : Should be able to write parsers.
o Kibana : Create different visualization and dashboards according to the Client needs
o Scripts : Should be able to write scripts in linux.
Job posted by
Akansha Singh

Credit Risk Analyst

at Niro

Founded 2021  •  Product  •  20-100 employees  •  Raised funding
Risk assessment
Risk Management
Risk analysis
Bengaluru (Bangalore)
2 - 4 yrs
₹7L - ₹15L / yr
  • Gather information from multiple data sources make Approval Decisions mechanically
  • Read and interpret credit related information to the borrowers
  • Interpret, analyze and assess all forms of complex information
  • Embark on risk assessment analysis
  • Maintain the credit exposure of the company within certain risk level with set limit in mind
  • Build strategies to minimize risk and increase approval rates
  • Design Champion and Challenger tests, implement and read test results
  • Build Line assignment strategies
Skills required:
- Credit Risk Modeling
- Statistical Data Understanding and interpretation
- Basic Regression and Advanced Machine Learning Models
- Conversant with coding on Python using libraries like Sklearn etc.
- Build and understand decision trees
Job posted by
Vinay Gurram

Artificial Intelligence Developers

at Precily Private Limited

Founded 2016  •  Product  •  20-100 employees  •  Raised funding
Data Science
Artificial Neural Network (ANN)
Artificial Intelligence (AI)
Machine Learning (ML)
Natural Language Processing (NLP)
Big Data
NCR (Delhi | Gurgaon | Noida)
1 - 3 yrs
₹3L - ₹9L / yr
-Precily AI: Automatic summarization, shortening a business document, book with our AI. Create a summary of the major points of the original document. AI can make a coherent summary taking into account variables such as length, writing style, and syntax. We're also working in the legal domain to reduce the high number of pending cases in India. We use Artificial Intelligence and Machine Learning capabilities such as NLP, Neural Networks in Processing the data to provide solutions for various industries such as Enterprise, Healthcare, Legal.
Job posted by
Bharath Rao

Big Data Developer

at MediaMelon Inc

Founded 2008  •  Product  •  20-100 employees  •  Raised funding
Spark Streaming
Aero spike
Apache Kafka
Big Data
Elastic Search
Bengaluru (Bangalore), Bengaluru (Bangalore)
1 - 7 yrs
₹0L / yr
Develop analytic tools, working on BigData and Distributed systems. - Provide technical leadership on developing our core Analytic platform - Lead development efforts on product features using Scala/Java -Demonstrable excellence in innovation, problem solving, analytical skills, data structures and design patterns - Expert in building applications using Spark and Spark Streaming -Exposure to NoSQL: HBase/Cassandra, Hive and Pig -Latin, Mahout -Extensive experience with Hadoop and Machine learning algorithms
Job posted by
Katreddi Kiran Kumar
Did not find a job you were looking for?
Search for relevant jobs from 10000+ companies such as Google, Amazon & Uber actively hiring on Cutshort.
Get to hear about interesting companies hiring right now
iconFollow Cutshort
Want to apply to this role at Top Multinational Fintech Startup?
Why apply via Cutshort?
Connect with actual hiring teams and get their fast response. No spam.
Learn more
Get to hear about interesting companies hiring right now
iconFollow Cutshort