Technical Architect

at CarDekho

DP
Posted by Shilpa Jangid
icon
NCR (Delhi | Gurgaon | Noida)
icon
6 - 10 yrs
icon
₹50L - ₹70L / yr (ESOP available)
icon
Full time
Skills
Big Data
Hadoop
Spark
Apache Hive
Python
Amazon Web Services (AWS)
Go Programming (Golang)
NodeJS (Node.js)
Elastic Search
Cassandra
Microservices
Amazon Redshift
Data Warehouse (DWH)
We at CarDekho are looking for a self-motivated technology leader in big data and data warehouse, search, and analytics technologies who have hands-on development experience in architecting scalable backend systems with a high volume of data.

Responsibilities: 
  • Architect and design for our customers' data-driven applications and solutions and own back-end technology
  • stack
  • Develop architectures that are inherently secure, robust, scalable, modular, and API-centric
  • Build distributed backend systems serving real-time analytics and machine learning features at scale
  • Own the scalability, performance, and performance metrics of complex distributed systems.
  • Apply architecture best practices that help increase execution velocity
  • Collaborate with the key stakeholders, like business, product, and other technology teams
  • Mentor junior members in the team
Requirement: 
  • Excellent Academic Background (MS/B.Tech from a top tier university)
  • 6-10 years of experience in backend architecture and development with large data volumes
  • Extensive hands-on experience in the Big Data Ecosystem (like Hadoop, Spark, Presto, Hive), Database (like
  • MySQL, PostgreSQL), NoSQL (like MongoDB, Cassandra), and Data Warehousing like Redshift
  • Have deep expertise with search engines like Elastic Search and javascript environment like Node.js
  • Experience in cloud-based technology solutions with scale and robustness
  • Strong data management and migration experience including proficiency in data warehousing, data quality, and analysis.
  • Experience in the development of microservices/REST APIs
  • Experience with Agile and DevOps development methodology and tools like Jira, Confluence
  • Understanding/exposure to complete product development cycle
Read more

About CarDekho

Thinking of buying a car? At CarDekho.com, buy new and used cars, search by filter and preferences, compare cars, read latest news and updates, see 360 views & more!
Read more
Founded
2008
Type
Products & Services
Size
100-1000 employees
Stage
Profitable
View full company details
Why apply to jobs via Cutshort
Personalized job matches
Stop wasting time. Get matched with jobs that meet your skills, aspirations and preferences.
Verified hiring teams
See actual hiring teams, find common social connections or connect with them directly. No 3rd party agencies here.
Move faster with AI
We use AI to get you faster responses, recommendations and unmatched user experience.
2101133
Matches delivered
3712187
Network size
15000
Companies hiring

Similar jobs

Big Data
Scala
Spark
Hadoop
Python
Amazon Web Services (AWS)
icon
Bengaluru (Bangalore), Hyderabad, Pune
icon
9 - 16 yrs
icon
₹7L - ₹32L / yr
Greetings..
 
We have urgent requirement for the post of Big Data Architect in reputed MNC company
 
 


Location:  Pune/Nagpur,Goa,Hyderabad/Bangalore

Job Requirements:

  • 9 years and above of total experience preferably in bigdata space.
  • Creating spark applications using Scala to process data.
  • Experience in scheduling and troubleshooting/debugging Spark jobs in steps.
  • Experience in spark job performance tuning and optimizations.
  • Should have experience in processing data using Kafka/Pyhton.
  • Individual should have experience and understanding in configuring Kafka topics to optimize the performance.
  • Should be proficient in writing SQL queries to process data in Data Warehouse.
  • Hands on experience in working with Linux commands to troubleshoot/debug issues and creating shell scripts to automate tasks.
  • Experience on AWS services like EMR.
Read more
Job posted by
Haina khan

Big Data Engineer- Pyspark & cloud

at Cactus Communications

Founded 2002  •  Product  •  1000-5000 employees  •  Profitable
PySpark
Hadoop
Amazon Web Services (AWS)
HDFS
Elastic Search
Natural Language Processing (NLP)
Machine Learning (ML)
icon
Remote only
icon
1 - 3 yrs
icon
Best in industry

Please note - This is a 100% remote opportunity and you can work from any location.

 

About the team:

You will be a part of Cactus Labs which is the R&D Cell of Cactus Communications. Cactus Labs is a high impact cell that works to solve complex technical and business problems that help keep us strategically competitive in the industry. We are a multi-cultural team spread across multiple countries. We work in the domain of AI/ML especially with Text (NLP - Natural Language Processing), Language Understanding, Explainable AI, Big Data, AR/VR etc.

 

The opportunity: Within Cactus Labs you will work with the Big Data team. This team manages Terabytes of data coming from different sources. We are re-orchestrating data pipelines to handle this data at scale and improve visibility and robustness. We operate across all the three Cloud Platforms and leverage the best of them.

 

In this role, you will get to own a component end to end. You will also get to work on cloud platform and learn to design distributed data processing systems to operate at scale.

 

Responsibilities:

  • Collaborate with a team of Big Data Engineers, Big Data and Cloud Architects and Domain SMEs to drive the product ahead
  • Stay up to date with the progress of in the domain since We work on cutting-edge technologies and are constantly trying new things out
  • Build solutions for massive scale. This requires extensive benchmarking to pick the right approach
  • Understand the data in and out, and make sense of it. You will at times need to draw conclusions and present it to the business users
  • Be independent, self-driven and highly motivated. While you will have the best people to learn from and access to various courses or training materials, we expect you to take charge of your growth and learning.

 

Expectations from you:

  • 1-3 Year of relevant experience in Big Data with pySpark
  • Hands on experience of distributed computing and Big Data Ecosystem - Hadoop, HDFS, Spark etc
  • Good understanding of data lake and their importance in a Big Data Ecosystem
  • Experience of working in the Cloud Environment (AWS, Azure or GCP)
  • You like to work without a lot of supervision or micromanagement.
  • Above all, you get excited by data. You like to dive deep, mine patterns and draw conclusions. You believe in making data driven decisions and helping the team look for the pattern as well.

 

Preferred skills:

  • Familiarity with search engines like Elasticsearch and Bigdata warehouses systems like AWS Athena, Google Big Query etc
  • Building data pipelines using Airflow
  • Experience of working in AWS Cloud Environment
  • Knowledge of NLP and ML
Read more
Job posted by
Hemal Kamble

Capabilities & Insights Analyst

at Top Managment Consulting Firm

Market Research
Data Analytics
Python
R Programming
Linear regression
Clustering
Forecasting
Regression analysis
Market segmentation
icon
Bengaluru (Bangalore), Gurugram
icon
1 - 6 yrs
icon
₹7L - ₹15L / yr

Company Profile:

The company is World's No1 Global management consulting firm.


Job Qualifications
 Graduate or post graduate degree in statistics, economics, econometrics, computer science,
engineering, or mathematics
 2-5 years of relevant experience
 Adept in forecasting, regression analysis and segmentation work
 Understanding of modeling techniques, specifically logistic regression, linear regression, cluster
analysis, CHAID, etc.
 Statistical programming software experience in R & Python, comfortable working with large data
sets; SAS & SQL are also preferred
 Excellent analytical and problem-solving skills, including the ability to disaggregate issues, identify
root causes and recommend solutions
 Excellent time management skills
 Good written and verbal communication skills; understanding of both written and spoken English
 Strong interpersonal skills
 Ability to act autonomously, bringing structure and organization to work
 Creative and action-oriented mindset
 Ability to interact in a fluid, demanding and unstructured environment where priorities evolve
constantly and methodologies are regularly challenged
 Ability to work under pressure and deliver on tight deadlines
Read more
Job posted by
Jayaraj E

Senior Data Engineer

at Balbix India Pvt Ltd

Founded 2015  •  Product  •  100-500 employees  •  Profitable
NOSQL Databases
MongoDB
Elastic Search
API
SQL
Python
Scala
Time series
InfluxDB
icon
Gurgaon, Bengaluru (Bangalore)
icon
3 - 10 yrs
icon
₹15L - ₹25L / yr
WHO WE ARE
Balbix is the world's leading platform for cybersecurity posture automation. Using Balbix, organizations can discover, prioritize and mitigate unseen risks and vulnerabilities at high velocity. With seamless data collection and petabyte-scale analysis capabilities, Balbix is deployed and operational within hours, and helps to decrease breach risk immediately. 
 
Balbix counts many global 1000 companies among its rapidly growing customer base. We are backed by John Chambers (the former CEO and Chairman of Cisco)top Silicon Valley VCs and global investors. We have been called magical, and have received raving reviews as well as customer testimonialsnumerous industry awards, and recognition by Gartner as a Cool Vendor, and by Frost & Sullivan.

ABOUT THIS ROLE
As a senior data engineer you will work on problems related to storing, analyzing, and manipulating very large cybersecurity and IT data sets. You will collaborate closely with our data scientists, threat researchers and network experts to solve real-world problems plaguing cybersecurity. This role requires excellent architecture, design, testing and programming skills as well as experience in large-scale data engineering.

You will:
  • Architect and implement modules for ingesting, storing and manipulating large data sets for a variety of cybersecurity use-cases. 
  • Write code to provide backend support for data-driven UI widgets, web dashboards, workflows, search and API connectors.  
  • Design and implement high performance APIs between our frontend and backend components, and between different backend components. 
  • Build production quality solutions that balance complexity and performance
  • Participate in the engineering life-cycle at Balbix, including designing high quality UI components, writing production code, conducting code reviews and working alongside our backend infrastructure and reliability teams
  • Stay current on the ever-evolving technology landscape of web based UIs and recommend new systems for incorporation in our technology stack.
You are:
  • Product-focused and passionate about building truly usable systems
  • Collaborative and comfortable working with across teams including data engineering, front end, product management, and DevOps
  • Responsible and like to take ownership of challenging problems
  • A good communicator, and facilitate teamwork via good documentation practices
  • Comfortable with ambiguity and able to iterate quickly in response to an evolving understanding of customer needs
  • Curious about the world and your profession, and a constant learner 
You have:
  • BS in Computer Science or related field
  • Atleast 3+ years of experience in the backend web stack (Node.js, MongoDB, Redis, Elastic Search, Postgres, Java, Python, Docker, Kubernetes, etc.)
  • SQL, no-SQL database experience
  • Experience building API (development experience using GraphQL is a plus)
  • Familiarity with issues of web performance, availability, scalability, reliability, and maintainability 
Read more
Job posted by
Garima Saraswat

Data Scientist (Forecasting)

at Anicaa Data

Agency job
via wrackle
TensorFlow
PyTorch
Machine Learning (ML)
Data Science
data scientist
Forecasting
C++
Python
Artificial Neural Network (ANN)
moving average
ARIMA
Big Data
Data Analytics
Amazon Web Services (AWS)
azure
Google Cloud Platform (GCP)
icon
Bengaluru (Bangalore)
icon
3 - 6 yrs
icon
₹10L - ₹25L / yr

Job Title – Data Scientist (Forecasting)

Anicca Data is seeking a Data Scientist (Forecasting) who is motivated to apply his/her/their skill set to solve complex and challenging problems. The focus of the role will center around applying deep learning models to real-world applications.  The candidate should have experience in training, testing deep learning architectures.  This candidate is expected to work on existing codebases or write an optimized codebase at Anicca Data. The ideal addition to our team is self-motivated, highly organized, and a team player who thrives in a fast-paced environment with the ability to learn quickly and work independently.

 

Job Location: Remote (for time being) and Bangalore, India (post-COVID crisis)

 

Required Skills:

  • At least 3+ years of experience in a Data Scientist role
  • Bachelor's/Master’s degree in Computer Science, Engineering, Statistics, Mathematics, or similar quantitative discipline. D. will add merit to the application process
  • Experience with large data sets, big data, and analytics
  • Exposure to statistical modeling, forecasting, and machine learning. Deep theoretical and practical knowledge of deep learning, machine learning, statistics, probability, time series forecasting
  • Training Machine Learning (ML) algorithms in areas of forecasting and prediction
  • Experience in developing and deploying machine learning solutions in a cloud environment (AWS, Azure, Google Cloud) for production systems
  • Research and enhance existing in-house, open-source models, integrate innovative techniques, or create new algorithms to solve complex business problems
  • Experience in translating business needs into problem statements, prototypes, and minimum viable products
  • Experience managing complex projects including scoping, requirements gathering, resource estimations, sprint planning, and management of internal and external communication and resources
  • Write C++ and Python code along with TensorFlow, PyTorch to build and enhance the platform that is used for training ML models

Preferred Experience

  • Worked on forecasting projects – both classical and ML models
  • Experience with training time series forecasting methods like Moving Average (MA) and Autoregressive Integrated Moving Average (ARIMA) with Neural Networks (NN) models as Feed-forward NN and Nonlinear Autoregressive
  • Strong background in forecasting accuracy drivers
  • Experience in Advanced Analytics techniques such as regression, classification, and clustering
  • Ability to explain complex topics in simple terms, ability to explain use cases and tell stories
Read more
Job posted by
Naveen Taalanki

Data Scientist

at Srijan Technologies

Founded 2002  •  Products & Services  •  100-1000 employees  •  Profitable
Data Science
Machine Learning (ML)
R Programming
Python
Deep Learning
Clustering
icon
Remote only
icon
3 - 8 yrs
icon
₹10L - ₹25L / yr

Job Responsibilities:-

  • Develop robust, scalable and maintainable machine learning models to answer business problems against large data sets.
  • Build methods for document clustering, topic modeling, text classification, named entity recognition, sentiment analysis, and POS tagging.
  • Perform elements of data cleaning, feature selection and feature engineering and organize experiments in conjunction with best practices.
  • Benchmark, apply, and test algorithms against success metrics. Interpret the results in terms of relating those metrics to the business process.
  • Work with development teams to ensure models can be implemented as part of a delivered solution replicable across many clients.
  • Knowledge of Machine Learning, NLP, Document Classification, Topic Modeling and Information Extraction with a proven track record of applying them to real problems.
  • Experience working with big data systems and big data concepts.
  • Ability to provide clear and concise communication both with other technical teams and non-technical domain specialists.
  • Strong team player; ability to provide both a strong individual contribution but also work as a team and contribute to wider goals is a must in this dynamic environment.
  • Experience with noisy and/or unstructured textual data.

knowledge graph and NLP including summarization, topic modelling etc

  • Strong coding ability with statistical analysis tools in Python or R, and general software development skills (source code management, debugging, testing, deployment, etc.)
  • Working knowledge of various text mining algorithms and their use-cases such as keyword extraction, PLSA, LDA, HMM, CRF, deep learning & recurrent ANN, word2vec/doc2vec, Bayesian modeling.
  • Strong understanding of text pre-processing and normalization techniques, such as tokenization,
  • POS tagging and parsing and how they work at a low level.
  • Excellent problem solving skills.
  • Strong verbal and written communication skills
  • Masters or higher in data mining or machine learning; or equivalent practical analytics / modelling experience
  • Practical experience in using NLP related techniques and algorithms
  • Experience in open source coding and communities desirable.

Able to containerize Models and associated modules and work in a Microservices environment

Read more
Job posted by
Adyasha Satpathy

Data Engineer

at aquilae Technologies Pvt Ltd

Agency job
via SHASHWATH SOLUTION
Big Data
Hadoop
Spark
PySpark
icon
Remote only
icon
4 - 8 yrs
icon
₹8L - ₹35L / yr

Day in the life

·         In this role (senior data engineer) you'll get ...

 

·         Being one of the very first and part of core team member for data platform, setup platform foundation while adhering all required quality standards and design patterns

·         Write efficient and quality code that can scale

·         Adopt quality standards, recommend process standards and best practices

·         Research, learn & adapt new technologies to solve problems & improve existing solutions

·         Contribute to engineering excellence backlog Identify performance issues

·         Effective code and design reviews

·         Improve reliability of overall production system by proactively identifying patterns of failure

·         Leading and mentoring junior engineers by example

·         End-to-end ownership of stories (including design, serviceability, performance, failure handling)

·         Strive hard to provide the best experience to anyone using our products

·         Conceptualise innovative and elegant solutions to solve challenging big data problems

·         Engage with Product Management and Business to drive the agenda, set your priorities and deliver awesome products

·         Adhere to company policies, procedures, mission, values, and standards of ethics and integrity

 

Must-have skills

·         On day one we'll expect you to...

 

·         B.E/B.Tech from a reputed institution

·         Minimum 5 years of software development experience and at least a year experience in leading/guiding people

·         Expert coding skills in Java/Scala or Python

·         Deep understanding in Big Data Ecosystem - Hadoop and Spark

·         Must have project experience with Spark

·         Ability to independently troubleshoot Spark jobs Good understanding of distributed systems

·         Fast learner and quickly adapt to new technologies Prefer individuals with high ownership and commitment Expert hands on experience with RDBMS

·         Fast learner and quickly adapt to new technologies Prefer individuals with high ownership and commitment

·         Ability to work independently as well as working collaboratively in a team

 

 

What separates the best from the rest

Added bonuses you have...

 

Hands on experience with EMR/Glue/Data bricks Hand on experience with Airflow

Hands on experience with AWS Big Data ecosystem

 

 

Team culture

We are looking for passionate Engineers who are always hungry for challenging problems. We believe in creating opportunistic, yet balanced, work environment for savvy, entrepreneurial tech individuals. We are thriving on remote work with team working across multiple timezones.

 

 

 

Read more
Job posted by
karthik kashyap
PySpark
Data engineering
Big Data
Hadoop
Spark
Python
AWS Lambda
SQL
hadoop
kafka
icon
Bengaluru (Bangalore)
icon
6 - 8 yrs
icon
₹8L - ₹15L / yr
6-8years of experience in data engineer
Spark
Hadoop
Big Data
Data engineering
PySpark
Python
AWS Lambda
SQL
hadoop
kafka
Read more
Job posted by
Prashma S R

Data Engineer

at Mobile Programming LLC

Founded 1998  •  Services  •  100-1000 employees  •  Profitable
Data engineering
Data engineer
Spark
Apache Spark
Apache Kafka
Python
SQL
Linux/Unix
Shell Scripting
DevOps
CI/CD
Docker
Kubernetes
Javascript
Scala
Data integration
Spark Kafka
Pentaho
icon
Mohali, Gurugram, Pune, Bengaluru (Bangalore), Hyderabad, Chennai
icon
3 - 8 yrs
icon
₹2L - ₹9L / yr

Responsibilities for Data Engineer

  • Create and maintain optimal data pipeline architecture,
  • Assemble large, complex data sets that meet functional / non-functional business requirements.
  • Identify, design, and implement internal process improvements: automating manual processes, optimizing data delivery, re-designing infrastructure for greater scalability, etc.
  • Build the infrastructure required for optimal extraction, transformation, and loading of data from a wide variety of data sources using SQL and AWS ‘big data’ technologies.
  • Build analytics tools that utilize the data pipeline to provide actionable insights into customer acquisition, operational efficiency and other key business performance metrics.
  • Work with stakeholders including the Executive, Product, Data and Design teams to assist with data-related technical issues and support their data infrastructure needs.
  • Keep our data separated and secure across national boundaries through multiple data centers and AWS regions.
  • Create data tools for analytics and data scientist team members that assist them in building and optimizing our product into an innovative industry leader.
  • Work with data and analytics experts to strive for greater functionality in our data systems.

Qualifications for Data Engineer

  • Advanced working SQL knowledge and experience working with relational databases, query authoring (SQL) as well as working familiarity with a variety of databases.
  • Experience building and optimizing ‘big data’ data pipelines, architectures and data sets.
  • Experience performing root cause analysis on internal and external data and processes to answer specific business questions and identify opportunities for improvement.
  • Strong analytic skills related to working with unstructured datasets.
  • Build processes supporting data transformation, data structures, metadata, dependency and workload management.
  • A successful history of manipulating, processing and extracting value from large disconnected datasets.
  • Working knowledge of message queuing, stream processing, and highly scalable ‘big data’ data stores.
  • Strong project management and organizational skills.
  • Experience supporting and working with cross-functional teams in a dynamic environment.
  • We are looking for a candidate with 5+ years of experience in a Data Engineer role, who has attained a Graduate degree in Computer Science, Statistics, Informatics, Information Systems or another quantitative field. They should also have experience using the following software/tools:

  • Experience with big data tools: Hadoop, Spark, Kafka, etc.
  • Experience with relational SQL and NoSQL databases, including Postgres and Cassandra.
  • Experience with data pipeline and workflow management tools: Azkaban, Luigi, Airflow, etc.
  • Experience with AWS cloud services: EC2, EMR, RDS, Redshift
  • Experience with stream-processing systems: Storm, Spark-Streaming, etc.
  • Experience with object-oriented/object function scripting languages: Python, Java, C++, Scala, etc.
Read more
Job posted by
Apurva kalsotra

Data Engineer

at Codalyze Technologies

Founded 2016  •  Products & Services  •  20-100 employees  •  Profitable
Hadoop
Big Data
Scala
Spark
Amazon Web Services (AWS)
Java
Python
Apache Hive
icon
Mumbai
icon
3 - 7 yrs
icon
₹7L - ₹20L / yr
Job Overview :

Your mission is to help lead team towards creating solutions that improve the way our business is run. Your knowledge of design, development, coding, testing and application programming will help your team raise their game, meeting your standards, as well as satisfying both business and functional requirements. Your expertise in various technology domains will be counted on to set strategic direction and solve complex and mission critical problems, internally and externally. Your quest to embracing leading-edge technologies and methodologies inspires your team to follow suit.

Responsibilities and Duties :

- As a Data Engineer you will be responsible for the development of data pipelines for numerous applications handling all kinds of data like structured, semi-structured &
unstructured. Having big data knowledge specially in Spark & Hive is highly preferred.

- Work in team and provide proactive technical oversight, advice development teams fostering re-use, design for scale, stability, and operational efficiency of data/analytical solutions

Education level :

- Bachelor's degree in Computer Science or equivalent

Experience :

- Minimum 5+ years relevant experience working on production grade projects experience in hands on, end to end software development

- Expertise in application, data and infrastructure architecture disciplines

- Expert designing data integrations using ETL and other data integration patterns

- Advanced knowledge of architecture, design and business processes

Proficiency in :

- Modern programming languages like Java, Python, Scala

- Big Data technologies Hadoop, Spark, HIVE, Kafka

- Writing decently optimized SQL queries

- Orchestration and deployment tools like Airflow & Jenkins for CI/CD (Optional)

- Responsible for design and development of integration solutions with Hadoop/HDFS, Real-Time Systems, Data Warehouses, and Analytics solutions

- Knowledge of system development lifecycle methodologies, such as waterfall and AGILE.

- An understanding of data architecture and modeling practices and concepts including entity-relationship diagrams, normalization, abstraction, denormalization, dimensional
modeling, and Meta data modeling practices.

- Experience generating physical data models and the associated DDL from logical data models.

- Experience developing data models for operational, transactional, and operational reporting, including the development of or interfacing with data analysis, data mapping,
and data rationalization artifacts.

- Experience enforcing data modeling standards and procedures.

- Knowledge of web technologies, application programming languages, OLTP/OLAP technologies, data strategy disciplines, relational databases, data warehouse development and Big Data solutions.

- Ability to work collaboratively in teams and develop meaningful relationships to achieve common goals

Skills :

Must Know :

- Core big-data concepts

- Spark - PySpark/Scala

- Data integration tool like Pentaho, Nifi, SSIS, etc (at least 1)

- Handling of various file formats

- Cloud platform - AWS/Azure/GCP

- Orchestration tool - Airflow
Read more
Job posted by
Aishwarya Hire
Did not find a job you were looking for?
icon
Search for relevant jobs from 10000+ companies such as Google, Amazon & Uber actively hiring on Cutshort.
Get to hear about interesting companies hiring right now
iconFollow Cutshort
Want to apply to this role at CarDekho?
Why apply via Cutshort?
Connect with actual hiring teams and get their fast response. No spam.
Learn more
Get to hear about interesting companies hiring right now
iconFollow Cutshort