Data Scientist

at Top startup of India - News App

Agency job
icon
Noida
icon
6 - 10 yrs
icon
₹35L - ₹65L / yr (ESOP available)
icon
Full time
Skills
Data Science
Machine Learning (ML)
Natural Language Processing (NLP)
Computer Vision
TensorFlow
Deep Learning
Python
PySpark
MongoDB
Hadoop
Spark
This will be an individual contributor role and people from Tier 1/2 and Product based company can only apply.

Requirements-

● B.Tech/Masters in Mathematics, Statistics, Computer Science or another quantitative field
● 2-3+ years of work experience in ML domain ( 2-5 years experience )
● Hands-on coding experience in Python
● Experience in machine learning techniques such as Regression, Classification,Predictive modeling, Clustering, Deep Learning stack, NLP.
● Working knowledge of Tensorflow/PyTorch
Optional Add-ons-
● Experience with distributed computing frameworks: Map/Reduce, Hadoop, Spark etc.
● Experience with databases: MongoDB
Read more
Why apply to jobs via Cutshort
Personalized job matches
Stop wasting time. Get matched with jobs that meet your skills, aspirations and preferences.
Verified hiring teams
See actual hiring teams, find common social connections or connect with them directly. No 3rd party agencies here.
Move faster with AI
We use AI to get you faster responses, recommendations and unmatched user experience.
2101133
Matches delivered
3712187
Network size
15000
Companies hiring

Similar jobs

GCP Data Engineer

at Beyond Human Resource

Founded 2017  •  Services  •  employees  •  Raised funding
Google Cloud Platform (GCP)
Data flow
Python
SQL server
Apache Beam
BigQuery
icon
Remote only
icon
4 - 6 yrs
icon
₹8L - ₹12L / yr
Primary Skills:
1. Working knowledge on GCP (Cloud Storage, Cloud functions, Firestore, DAG, Airflow/Cloud Composer, python, apache beam, bigquery)
2. Knowledge on Google BQ and DBT
3. Good to have Python scripting knowledge for Data Engineering
4. Terraform knowledge will be an added advantage
5. Knowledge on Data warehousing is a must

Secondary Skills:

6. Data Analysis, Knowledge on Teradata (BTEQ, Mload)
7. ETL or ELT process
8. Building CI/CD pipeline, containerization etc
9. Agile ways of working

Team Lead and Process:

1. Ensuring adherence to schedule and quality of activities related to design, build, testing and implementation of deliverables.
2. Participate in requirement elicitation, validation of architecture, creation and review of design.
3. Provide support to the team like pseudocode to team, coordinating with architects to resolve blockers, assigning and reviewing tasks and ensure quality and timelines are met.
Read more
Job posted by
Priya Sahni

Data Engineering Manager

at Porter.in

Founded 2014  •  Services  •  100-1000 employees  •  Profitable
Python
SQL
Spark
Amazon Web Services (AWS)
Team Management
icon
Bengaluru (Bangalore)
icon
7 - 12 yrs
icon
₹25L - ₹35L / yr

Manager | Data Engineering

Bangalore | Full Time

Company Overview:

At Porter, we are passionate about improving productivity. We want to help businesses, large and small, optimize their last-mile operations and empower them to unleash the growth of their core functions. Last mile delivery logistics is one of the biggest and fastest growing sectors of the economy with a market cap upwards of 50 billion USD and a growth rate exceeding 15% CAGR.

Porter is the fastest growing leader in this sector with operations in 14 major cities, a fleet size exceeding 1L registered and 50k active driver partners and a customer base with 3.5M being monthly active. Our industry-best technology platform has raised over 50 million USD from investors including Sequoia Capital, Kae Capital, Mahindra group and LGT Aspada. We are addressing a massive problem and going after a huge market.

We’re trying to create a household name in transportation and our ambition is to disrupt all facets of last mile logistics including warehousing and LTL transportation. At Porter, we’re here to do the best work of our lives.

If you want to do the same and love the challenges and opportunities of a fast paced work environment, then we believe Porter is the right place for you.

 

Responsibilities

Data Strategy and Alignment

  • Work closely with data analysts and business / product teams to understand requirements and provide data ready for analysis and reporting.
  • Apply, help define, and champion data governance : data quality, testing, documentation, coding best practices and peer reviews.
  • Continuously discover, transform, test, deploy, and document data sources and data models.
  • Work closely with the Infrastructure team to build and improve our Data Infrastructure.
  • Develop and execute data roadmap (and sprints) - with a keen eye on industry trends and direction.

 

 

 

Data Stores and System Development

  • Design and implement high-performance, reusable, and scalable data models for our data warehouse to ensure our end-users get consistent and reliable answers when running their own analyses.
  • Focus on test driven design and results for repeatable and maintainable processes and tools.
  • Create and maintain optimal data pipeline architecture - and data flow logging framework.
  • Build the data products, features, tools, and frameworks that enable and empower Data, and Analytics teams across Porter.

Project Management

  • Drive project execution using effective prioritization and resource allocation.
  • Resolve blockers through technical expertise, negotiation, and delegation.
  • Strive for on-time complete solutions through stand-ups and course-correction.

Team Management

  • Manage and elevate team of 5-8 members.
  • Do regular one-on-ones with teammates to ensure resource welfare.
  • Periodic assessment and actionable feedback for progress.
  • Recruit new members with a view to long-term resource planning through effective collaboration with the hiring team.

Process design

  • Set the bar for the quality of technical and data-based solutions the team ships.
  • Enforce code quality standards and establish good code review practices - using this as a nurturing tool.
  • Set up communication channels and feedback loops for knowledge sharing and stakeholder management.
  • Explore the latest best practices and tools for constant up-skilling.

 

Data Engineering Stack

  • Analytics : Python / R / SQL + Excel / PPT, Google Colab
  • Database : PostgreSQL, Amazon Redshift, DynamoDB, Aerospike
  • Warehouse : Redshift, S3
  • ETL : Airflow + DBT + Custom-made Python + Amundsen (Discovery)
  • Business Intelligence / Visualization : Metabase + Google Data Studio
  • Frameworks : Spark + Dash + StreamLit
  • Collaboration : Git, Notion
Read more
Job posted by
Satyajit Mittra
PySpark
Data engineering
Big Data
Hadoop
Spark
SQL
Python
Microsoft SQL Server DBA
ELT
icon
Remote only
icon
7 - 13 yrs
icon
₹15L - ₹35L / yr
Experience
Experience Range

2 Years - 10 Years

Function Information Technology
Desired Skills
Must Have Skills:
• Good experience in Pyspark - Including Dataframe core functions and Spark SQL
• Good experience in SQL DBs - Be able to write queries including fair complexity.
• Should have excellent experience in Big Data programming for data transformation and aggregations
• Good at ELT architecture. Business rules processing and data extraction from Data Lake into data streams for business consumption.
• Good customer communication.
• Good Analytical skills
Education
Education Type Engineering
Degree / Diploma Bachelor of Engineering, Bachelor of Computer Applications, Any Engineering
Specialization / Subject Any Specialisation
Job Type Full Time
Job ID 000018
Department Software Development
Read more
Job posted by
Minakshi Kumari

Data Engineer

at AI-powered cloud-based SaaS solution

Agency job
via wrackle
Data engineering
Big Data
Data Engineer
Big Data Engineer
Hibernate (Java)
Data Structures
Agile/Scrum
SaaS
Cassandra
Spark
Python
NOSQL Databases
Hadoop
HDFS
MapReduce
AWS CloudFormation
EMR
Amazon S3
Apache Kafka
Apache ZooKeeper
Systems Development Life Cycle (SDLC)
Java
YARN
icon
Bengaluru (Bangalore)
icon
2 - 10 yrs
icon
₹15L - ₹50L / yr
Responsibilities

● Able contribute to the gathering of functional requirements, developing technical
specifications, and project & test planning
● Demonstrating technical expertise, and solving challenging programming and design
problems
● Roughly 80% hands-on coding
● Generate technical documentation and PowerPoint presentations to communicate
architectural and design options, and educate development teams and business users
● Resolve defects/bugs during QA testing, pre-production, production, and post-release
patches
● Work cross-functionally with various bidgely teams including: product management,
QA/QE, various product lines, and/or business units to drive forward results

Requirements
● BS/MS in computer science or equivalent work experience
● 2-4 years’ experience designing and developing applications in Data Engineering
● Hands-on experience with Big data Eco Systems.
● Hadoop,Hdfs,Map Reduce,YARN,AWS Cloud, EMR, S3, Spark, Cassandra, Kafka,
Zookeeper
● Expertise with any of the following Object-Oriented Languages (OOD): Java/J2EE,Scala,
Python
● Strong leadership experience: Leading meetings, presenting if required
● Excellent communication skills: Demonstrated ability to explain complex technical
issues to both technical and non-technical audiences
● Expertise in the Software design/architecture process
● Expertise with unit testing & Test-Driven Development (TDD)
● Experience on Cloud or AWS is preferable
● Have a good understanding and ability to develop software, prototypes, or proofs of
concepts (POC's) for various Data Engineering requirements.
Read more
Job posted by
Naveen Taalanki

Hadoop Developer

at Recro

Founded 2014  •  Products & Services  •  100-1000 employees  •  Profitable
Hadoop
Apache Hadoop
Apache Hive
Apache Sqoop
Apache HBase
Java
MySQL
Spark
icon
Remote only
icon
2 - 6 yrs
icon
₹2L - ₹6L / yr

● BTech/BE/MS in Computer Science or related technical discipline with 2+ years of experience
● Strong analytical ability and problem solving skills, structured thought process and systematic approach towards handling fuzzy problems
● Fast learner, ability to adapt to situations, fast paced environments and never give up attitude
● Extreme proficiency in Java and Api (mandatory skill)
● Strong hand-on on querying and exposure to at least one relational or no sql database
● Hands-on on in Big Data technologies like Hadoop & Hive, experience in Spark is an added advantage
● Good communication both written and oral
● Sound understanding and adherence to SDLC processes and practices.
Read more
Job posted by
Agnish Banerjee

Data Model & UDX

at They provide both wholesale and retail funding. (PM1)

Agency job
via Multi Recruit
Teradata
Vertica
Python
DBA
Redshift
Synapse
Snowflake
Dynamo DB
UDX
FSLDM
Cosmos DB
OLAP
Data modeling
icon
Mumbai
icon
5 - 7 yrs
icon
₹20L - ₹25L / yr
  • Key responsibility is to design, develop & maintain efficient Data models for the organization maintained to ensure optimal query performance by the consumption layer.
  • Developing, Deploying & maintaining a repository of UDXs written in Java / Python.
  • Develop optimal Data Model design, analyzing complex distributed data deployments, and making recommendations to optimize performance basis data consumption patterns, performance expectations, the query is executed on the tables/databases, etc.
  • Periodic Database health check and maintenance
  • Designing collections in a no-SQL Database for efficient performance
  • Document & maintain data dictionary from various sources to enable data governance
  • Coordination with Business teams, IT, and other stakeholders to provide best-in-class data pipeline solutions, exposing data via APIs, loading in down streams, No-SQL Databases, etc
  • Data Governance Process Implementation and ensuring data security

Requirements

  • Extensive working experience in Designing & Implementing Data models in OLAP Data Warehousing solutions (Redshift, Synapse, Snowflake, Teradata, Vertica, etc).
  • Programming experience using Python / Java.
  • Working knowledge in developing & deploying User-defined Functions (UDXs) using Java / Python.
  • Strong understanding & extensive working experience in OLAP Data Warehousing (Redshift, Synapse, Snowflake, Teradata, Vertica, etc) architecture and cloud-native Data Lake (S3, ADLS, BigQuery, etc) Architecture.
  • Strong knowledge in Design, Development & Performance tuning of 3NF/Flat/Hybrid Data Model.
  • Extensive technical experience in SQL including code optimization techniques.
  • Strung knowledge of database performance and tuning, troubleshooting, and tuning.
  • Knowledge of collection design in any No-SQL DB (DynamoDB, MongoDB, CosmosDB, etc), along with implementation of best practices.
  • Ability to understand business functionality, processes, and flows.
  • Good combination of technical and interpersonal skills with strong written and verbal communication; detail-oriented with the ability to work independently.
  • Any OLAP DWH DBA Experience and User Management will be added advantage.
  • Knowledge in financial industry-specific Data models such as FSLDM, IBM Financial Data Model, etc will be added advantage.
  • Experience in Snowflake will be added advantage.
  • Working experience in BFSI/NBFC & data understanding of Loan/Mortgage data will be added advantage.

Functional knowledge

  • Data Governance & Quality Assurance
  • Modern OLAP Database Architecture & Design
  • Linux
  • Data structures, algorithm & data modeling techniques
  • No-SQL database architecture
  • Data Security

 

Read more
Job posted by
Sapna Deb

Data Engineer

at VIMANA

Founded 2009  •  Product  •  20-100 employees  •  Profitable
Data engineering
Data Engineer
Apache Kafka
Big Data
Java
NodeJS (Node.js)
Elastic Search
Test driven development (TDD)
Python
icon
Remote, Chennai
icon
2 - 5 yrs
icon
₹10L - ₹20L / yr

We are looking for passionate, talented and super-smart engineers to join our product development team. If you are someone who innovates, loves solving hard problems, and enjoys end-to-end product development, then this job is for you! You will be working with some of the best developers in the industry in a self-organising, agile environment where talent is valued over job title or years of experience.

 

Responsibilities:

  • You will be involved in end-to-end development of VIMANA technology, adhering to our development practices and expected quality standards.
  • You will be part of a highly collaborative Agile team which passionately follows SAFe Agile practices, including pair-programming, PR reviews, TDD, and Continuous Integration/Delivery (CI/CD).
  • You will be working with cutting-edge technologies and tools for stream processing using Java, NodeJS and Python, using frameworks like Spring, RxJS etc.
  • You will be leveraging big data technologies like Kafka, Elasticsearch and Spark, processing more than 10 Billion events per day to build a maintainable system at scale.
  • You will be building Domain Driven APIs as part of a micro-service architecture.
  • You will be part of a DevOps culture where you will get to work with production systems, including operations, deployment, and maintenance.
  • You will have an opportunity to continuously grow and build your capabilities, learning new technologies, languages, and platforms.

 

Requirements:

  • Undergraduate degree in Computer Science or a related field, or equivalent practical experience.
  • 2 to 5 years of product development experience.
  • Experience building applications using Java, NodeJS, or Python.
  • Deep knowledge in Object-Oriented Design Principles, Data Structures, Dependency Management, and Algorithms.
  • Working knowledge of message queuing, stream processing, and highly scalable Big Data technologies.
  • Experience in working with Agile software methodologies (XP, Scrum, Kanban), TDD and Continuous Integration (CI/CD).
  • Experience using no-SQL databases like MongoDB or Elasticsearch.
  • Prior experience with container orchestrators like Kubernetes is a plus.
About VIMANA

We build products and platforms for the Industrial Internet of Things. Our technology is being used around the world in mission-critical applications - from improving the performance of manufacturing plants, to making electric vehicles safer and more efficient, to making industrial equipment smarter.

Please visit https://govimana.com/ to learn more about what we do.

Why Explore a Career at VIMANA
  • We recognize that our dedicated team members make us successful and we offer competitive salaries.
  • We are a workplace that values work-life balance, provides flexible working hours, and full time remote work options.
  • You will be part of a team that is highly motivated to learn and work on cutting edge technologies, tools, and development practices.
  • Bon Appetit! Enjoy catered breakfasts, lunches and free snacks!

VIMANA Interview Process
We usually target to complete all the interviews in a week's time and would provide prompt feedback to the candidate. As of now, all the interviews are conducted online due to covid situation.

1.Telephonic screening (30 Min )

A 30 minute telephonic interview to understand and evaluate the candidate's fit with the job role and the company.
Clarify any queries regarding the job/company.
Give an overview about further interview rounds

2. Technical Rounds

This would be deep technical round to evaluate the candidate's technical capability pertaining to the job role.

3. HR Round

Candidate's team and cultural fit will be evaluated during this round

We would proceed with releasing the offer if the candidate clears all the above rounds.

Note: In certain cases, we might schedule additional rounds if needed before releasing the offer.
Read more
Job posted by
Loshy Chandran

Data Scientist

at Vestibulum Technologies Pvt Ltd

Founded 2017  •  Products & Services  •  0-20 employees  •  Bootstrapped
Data Science
Machine Learning (ML)
Deep Learning
Python
Computer Vision
Natural Language Processing (NLP)
TensorFlow
PyTorch
icon
Bengaluru (Bangalore)
icon
0.3 - 0.6 yrs
icon
₹15,000 - ₹25,000 / mo
Hiring for Machine Learning Deep Learning, who have build solutions for computer vision ,nlp, Tensorflow, Pytorch, Tensorflow Lite for Android phone 
Read more
Job posted by
Srinivasan Saripalli

Big Data Developer

at Intelliswift

Founded 2001  •  Products & Services  •  100-1000 employees  •  Profitable
Big Data
Spark
Scala
SQL
icon
Chennai
icon
4 - 8 yrs
icon
₹8L - ₹17L / yr
Greetings from Intelliswift! Intelliswift Software Inc. is a premier software solutions and Services Company headquartered in the Silicon Valley, with offices across the United States, India, and Singapore. The company has a proven track record of delivering results through its global delivery centers and flexible engagement models for over 450 brands ranging from Fortune 100 to growing companies. Intelliswift provides a variety of services including Enterprise Applications, Mobility, Big Data / BI, Staffing Services, and Cloud Solutions. Growing at an outstanding rate, it has been recognized as the second largest private IT Company in the East Bay. Domains: IT, Retail, Pharma, Healthcare, BFSI, and Internet & E-commerce website https://www.intelliswift.com/ Experience: 4-8 Years Job Location: Chennai Job Description: Skills: Spark, Scala, Big data, Hive · Strong Working experience in Spark, Scala, big data, h base and hive. · Should have good working experience in SQL and Spark SQL. · Good to have knowledge or experience in Teradata. · Familiar with General engineering Git, jenkins, sbt, maven.
Read more
Job posted by
Pratish Mishra

Big Data Developer

at GeakMinds Technologies Pvt Ltd

Founded 2011  •  Services  •  100-1000 employees  •  Profitable
Hadoop
Big Data
HDFS
Apache Sqoop
Apache Flume
Apache HBase
Apache Kafka
icon
Chennai
icon
1 - 5 yrs
icon
₹1L - ₹6L / yr
• Looking for Big Data Engineer with 3+ years of experience. • Hands-on experience with MapReduce-based platforms, like Pig, Spark, Shark. • Hands-on experience with data pipeline tools like Kafka, Storm, Spark Streaming. • Store and query data with Sqoop, Hive, MySQL, HBase, Cassandra, MongoDB, Drill, Phoenix, and Presto. • Hands-on experience in managing Big Data on a cluster with HDFS and MapReduce. • Handle streaming data in real time with Kafka, Flume, Spark Streaming, Flink, and Storm. • Experience with Azure cloud, Cognitive Services, Databricks is preferred.
Read more
Job posted by
John Richardson
Did not find a job you were looking for?
icon
Search for relevant jobs from 10000+ companies such as Google, Amazon & Uber actively hiring on Cutshort.
Get to hear about interesting companies hiring right now
iconFollow Cutshort
Want to apply to this role at Top startup of India - News App?
Why apply via Cutshort?
Connect with actual hiring teams and get their fast response. No spam.
Learn more
Get to hear about interesting companies hiring right now
iconFollow Cutshort