Cutshort logo
Lead Artificial Intelligence Engineer

Lead Artificial Intelligence Engineer

at xpressbees

Posted by Alfiya Khan
Pune, Bengaluru (Bangalore)
6 - 8 yrs
₹15L - ₹25L / yr
Full time
Machine Learning (ML)
Data Science
Natural Language Processing (NLP)
Computer Vision
Artificial Intelligence (AI)
Random boosting
Decision trees
Big Data
recommendation algorithm
Company Profile
XressBees – a logistics company started in 2015 – is amongst the fastest growing companies of its sector. Our
vision to evolve into a strong full-service logistics organization reflects itself in the various lines of business like B2C
logistics 3PL, B2B Xpress, Hyperlocal and Cross border Logistics.
Our strong domain expertise and constant focus on innovation has helped us rapidly evolve as the most trusted
logistics partner of India. XB has progressively carved our way towards best-in-class technology platforms, an
extensive logistics network reach, and a seamless last mile management system.
While on this aggressive growth path, we seek to become the one-stop-shop for end-to-end logistics solutions. Our
big focus areas for the very near future include strengthening our presence as service providers of choice and
leveraging the power of technology to drive supply chain efficiencies.
Job Overview
XpressBees would enrich and scale its end-to-end logistics solutions at a high pace. This is a great opportunity to join
the team working on forming and delivering the operational strategy behind Artificial Intelligence / Machine Learning
and Data Engineering, leading projects and teams of AI Engineers collaborating with Data Scientists. In your role, you
will build high performance AI/ML solutions using groundbreaking AI/ML and BigData technologies. You will need to
understand business requirements and convert them to a solvable data science problem statement. You will be
involved in end to end AI/ML projects, starting from smaller scale POCs all the way to full scale ML pipelines in
Seasoned AI/ML Engineers would own the implementation and productionzation of cutting-edge AI driven algorithmic
components for search, recommendation and insights to improve the efficiencies of the logistics supply chain and
serve the customer better.
You will apply innovative ML tools and concepts to deliver value to our teams and customers and make an impact to
the organization while solving challenging problems in the areas of AI, ML , Data Analytics and Computer Science.
Opportunities for application:
- Route Optimization
- Address / Geo-Coding Engine
- Anomaly detection, Computer Vision (e.g. loading / unloading)
- Fraud Detection (fake delivery attempts)
- Promise Recommendation Engine etc.
- Customer & Tech support solutions, e.g. chat bots.
- Breach detection / prediction
An Artificial Intelligence Engineer would apply himself/herself in the areas of -
- Deep Learning, NLP, Reinforcement Learning
- Machine Learning - Logistic Regression, Decision Trees, Random Forests, XGBoost, etc..
- Driving Optimization via LPs, MILPs, Stochastic Programs, and MDPs
- Operations Research, Supply Chain Optimization, and Data Analytics/Visualization
- Computer Vision and OCR technologies
The AI Engineering team enables internal teams to add AI capabilities to their Apps and Workflows easily via APIs
without needing to build AI expertise in each team – Decision Support, NLP, Computer Vision, for Public Clouds and
Enterprise in NLU, Vision and Conversational AI.Candidate is adept at working with large data sets to find
opportunities for product and process optimization and using models to test the effectiveness of different courses of
action. They must have knowledge using a variety of data mining/data analysis methods, using a variety of data tools,
building, and implementing models, using/creating algorithms, and creating/running simulations. They must be
comfortable working with a wide range of stakeholders and functional teams. The right candidate will have a passion
for discovering solutions hidden in large data sets and working with stakeholders to improve business outcomes.

Roles & Responsibilities
● Develop scalable infrastructure, including microservices and backend, that automates training and
deployment of ML models.
● Building cloud services in Decision Support (Anomaly Detection, Time series forecasting, Fraud detection,
Risk prevention, Predictive analytics), computer vision, natural language processing (NLP) and speech that
work out of the box.
● Brainstorm and Design various POCs using ML/DL/NLP solutions for new or existing enterprise problems.
● Work with fellow data scientists/SW engineers to build out other parts of the infrastructure, effectively
communicating your needs and understanding theirs and address external and internal shareholder's
product challenges.
● Build core of Artificial Intelligence and AI Services such as Decision Support, Vision, Speech, Text, NLP, NLU,
and others.
● Leverage Cloud technology –AWS, GCP, Azure
● Experiment with ML models in Python using machine learning libraries (Pytorch, Tensorflow), Big Data,
Hadoop, HBase, Spark, etc
● Work with stakeholders throughout the organization to identify opportunities for leveraging company data to
drive business solutions.
● Mine and analyze data from company databases to drive optimization and improvement of product
development, marketing techniques and business strategies.
● Assess the effectiveness and accuracy of new data sources and data gathering techniques.
● Develop custom data models and algorithms to apply to data sets.
● Use predictive modeling to increase and optimize customer experiences, supply chain metric and other
business outcomes.
● Develop company A/B testing framework and test model quality.
● Coordinate with different functional teams to implement models and monitor outcomes.
● Develop processes and tools to monitor and analyze model performance and data accuracy.
● Develop scalable infrastructure, including microservices and backend, that automates training and
deployment of ML models.
● Brainstorm and Design various POCs using ML/DL/NLP solutions for new or existing enterprise problems.
● Work with fellow data scientists/SW engineers to build out other parts of the infrastructure, effectively
communicating your needs and understanding theirs and address external and internal shareholder's
product challenges.
● Deliver machine learning and data science projects with data science techniques and associated libraries
such as AI/ ML or equivalent NLP (Natural Language Processing) packages. Such techniques include a good
to phenomenal understanding of statistical models, probabilistic algorithms, classification, clustering, deep
learning or related approaches as it applies to financial applications.
● The role will encourage you to learn a wide array of capabilities, toolsets and architectural patterns for
successful delivery.
What is required of you?
You will get an opportunity to build and operate a suite of massive scale, integrated data/ML platforms in a broadly
distributed, multi-tenant cloud environment.
● B.S., M.S., or Ph.D. in Computer Science, Computer Engineering
● Coding knowledge and experience with several languages: C, C++, Java,JavaScript, etc.
● Experience with building high-performance, resilient, scalable, and well-engineered systems
● Experience in CI/CD and development best practices, instrumentation, logging systems
● Experience using statistical computer languages (R, Python, SLQ, etc.) to manipulate data and draw insights
from large data sets.
● Experience working with and creating data architectures.
● Good understanding of various machine learning and natural language processing technologies, such as
classification, information retrieval, clustering, knowledge graph, semi-supervised learning and ranking.

● Knowledge and experience in statistical and data mining techniques: GLM/Regression, Random Forest,
Boosting, Trees, text mining, social network analysis, etc.
● Knowledge on using web services: Redshift, S3, Spark, Digital Ocean, etc.
● Knowledge on creating and using advanced machine learning algorithms and statistics: regression,
simulation, scenario analysis, modeling, clustering, decision trees, neural networks, etc.
● Knowledge on analyzing data from 3rd party providers: Google Analytics, Site Catalyst, Core metrics,
AdWords, Crimson Hexagon, Facebook Insights, etc.
● Knowledge on distributed data/computing tools: Map/Reduce, Hadoop, Hive, Spark, MySQL, Kafka etc.
● Knowledge on visualizing/presenting data for stakeholders using: Quicksight, Periscope, Business Objects,
D3, ggplot, Tableau etc.
● Knowledge of a variety of machine learning techniques (clustering, decision tree learning, artificial neural
networks, etc.) and their real-world advantages/drawbacks.
● Knowledge of advanced statistical techniques and concepts (regression, properties of distributions,
statistical tests, and proper usage, etc.) and experience with applications.
● Experience building data pipelines that prep data for Machine learning and complete feedback loops.
● Knowledge of Machine Learning lifecycle and experience working with data scientists
● Experience with Relational databases and NoSQL databases
● Experience with workflow scheduling / orchestration such as Airflow or Oozie
● Working knowledge of current techniques and approaches in machine learning and statistical or
mathematical models
● Strong Data Engineering & ETL skills to build scalable data pipelines. Exposure to data streaming stack (e.g.
● Relevant experience in fine tuning and optimizing ML (especially Deep Learning) models to bring down
serving latency.
● Exposure to ML model productionzation stack (e.g. MLFlow, Docker)
● Excellent exploratory data analysis skills to slice & dice data at scale using SQL in Redshift/BigQuery.
Read more

About xpressbees

Company social profiles
Why apply to jobs via Cutshort
Personalized job matches
Stop wasting time. Get matched with jobs that meet your skills, aspirations and preferences.
Verified hiring teams
See actual hiring teams, find common social connections or connect with them directly. No 3rd party agencies here.
Move faster with AI
We use AI to get you faster responses, recommendations and unmatched user experience.
Matches delivered
Network size
Companies hiring

Similar jobs

2 - 10 yrs
₹10L - ₹50L / yr
Artificial Intelligence (AI)
Machine Learning (ML)
  • You're proficient in AI/Machine learning latest technologies  
  • You're proficient in GPT-3 based algorithms 
  • You have a passion for writing code as well as understanding and crafting the ways systems interact
  • You believe in the benefits of agile processes and shipping code often
  • You are pragmatic and work to coalesce requirements into reasonable solutions that provide value

  • Deploy well-tested, maintainable and scalable software solutions
  • Take end-to-end ownership of the technology stack and product 
  • Collaborate with other engineers to architect scalable technical solutions
  • Embrace and improve our standards and processes to reduce friction and unlock efficiency

Current Ecosystem : 
ShibaSwap :
Metaverse :
NFTs :
Game : Shiba Eternity on iOS and Android
Read more
A large software MNC with over 20k employees in India
Agency job
via RS Consultants by Rahul Inamdar
5 - 12 yrs
₹15L - ₹22L / yr
Data engineering
Data Engineer
Apache Kafka
Apache Spark
+6 more

As a Senior Engineer - Big Data Analytics, you will help the architectural design and development for Healthcare Platforms, Products, Services, and Tools to deliver the vision of the Company. You will significantly contribute to engineering, technology, and platform architecture. This will be done through innovation and collaboration with engineering teams and related business functions. This is a critical, highly visible role within the company that has the potential to drive significant business impact. 

The scope of this role will include strong technical contribution in the development and delivery of Big Data Analytics Cloud Platform, Products and Services in collaboration with execution and strategic partners. 



  • Design & develop, operate, and drive scalable, resilient, and cloud native Big Data Analytics platform to address the business requirements
  • Help drive technology transformation to achieve business transformation, through the creation of the Healthcare Analytics Data Cloud that will help Change establish a leadership position in healthcare data & analytics in the industry
  • Help in successful implementation of Analytics as a Service 
  • Ensure Platforms and Services meet SLA requirements
  • Be a significant contributor and partner in the development and execution of the Enterprise Technology Strategy



  • At least 2 years of experience software development for big data analytics, and cloud. At least 5 years of experience in software development
  • Experience working with High Performance Distributed Computing Systems in public and private cloud environments
  • Understands big data open-source eco-systems and its players. Contribution to open source is a strong plus
  • Experience with Spark, Spark Streaming, Hadoop, AWS/Azure, NoSQL Databases, In-Memory caches, distributed computing, Kafka, OLAP stores, etc.
  • Have successful track record of creating working Big Data stack that aligned with business needs, and delivered timely enterprise class products
  • Experience with delivering and managing scale of Operating Environment
  • Experience with Big Data/Micro Service based Systems, SaaS, PaaS, and Architectures
  • Experience Developing Systems in Java, Python, Unix
  • BSCS, BSEE or equivalent, MSCS preferred
Read more
at Klubworks
4 recruiters
Posted by Anupam Arya
Bengaluru (Bangalore)
3 - 6 yrs
Best in industry
Data Warehouse (DWH)
Big Data
+3 more
About the role
You will be building data pipelines that transform raw, unstructured data into formats data scientists can use for analysis. You will be responsible for creating and maintaining the analytics infrastructure that enables almost every other data function. This includes architectures such as databases, servers, and large-scale processing systems.
  • Responsible for setting up a  scalable Data warehouse
  • Building data pipeline mechanisms  to integrate the data from various sources for all of Klub’s data.  
  • Setup data as a service to expose the needed data as part of APIs. 
  • Have a good understanding on how the finance data works.
  • Standardize and optimize design thinking across the technology team.
  • Collaborate with stakeholders across engineering teams to come up with short and long-term architecture decisions.
  • Build robust data models that will help to support various reporting requirements for the business , ops and the leadership team. 
  • Participate in peer reviews , provide code/design comments.
  • Own the problem and deliver to success.

  • Overall 3+ years of industry experience
  • Prior experience on Backend and Data Engineering systems
  • Should have at least 1 + years of working experience in distributed systems
  • Deep understanding on python tech stack with the libraries like Flask, scipy, numpy, pytest frameworks.
  • Good understanding of Apache Airflow or similar orchestration tools. 
  • Good knowledge on data warehouse technologies like Apache Hive or similar. Good knowledge on Apache PySpark or similar. 
  • Good knowledge on how to build analytics services on the data for different reporting and BI needs. 
  • Good knowledge on data pipeline/ETL tools Hevo data or similar. Good knowledge on Trino / graphQL or similar query engine technologies. 
  • Deep understanding of concepts on Dimensional Data Models. Familiarity with RDBMS (MySQL/ PostgreSQL) , NoSQL (MongoDB/DynamoDB) databases & caching(redis or similar).
  • Should be proficient in writing SQL queries. 
  • Good knowledge on kafka. Be able to write clean, maintainable code.
Nice to have
  • Built a Data Warehouse from the scratch and set up a scalable data infrastructure.
  • Prior experience in fintech would be a plus.
  • Prior experience on data modelling.
Read more
5 candid answers
1 video
Posted by Kajal Jain
Remote only
1 - 4 yrs
₹5L - ₹15L / yr
Big Data
+2 more

Big Data Engineer/Data Engineer

What we are solving
Welcome to today’s business data world where:
• Unification of all customer data into one platform is a challenge

• Extraction is expensive
• Business users do not have the time/skill to write queries
• High dependency on tech team for written queries

These facts may look scary but there are solutions with real-time self-serve analytics:
• Fully automated data integration from any kind of a data source into a universal schema
• Analytics database that streamlines data indexing, query and analysis into a single platform.
• Start generating value from Day 1 through deep dives, root cause analysis and micro segmentation

At, this is what we do.
• We help our clients reduce effort and increase effectiveness quickly
• By clearly defining the scope of Projects
• Using Dependable, scalable, future proof technology solution like Big Data Solutions and Cloud Platforms
• Engaging with Data Scientists and Data Engineers to provide End to End Solutions leading to industrialisation of Data Science Model Development and Deployment

What we have achieved so far
Since we started in 2016,
• We have worked across 9 countries with 25+ global brands and 75+ projects
• We have 50+ clients, 100+ Data Sources and 20TB+ data processed daily

Work culture at
We are a small, remote team that believes in
• Working with a few, but only with highest quality team members who want to become the very best in their fields.
• With each member's belief and faith in what we are solving, we collectively see the Big Picture
• No hierarchy leads us to believe in reaching the decision maker without any hesitation so that our actions can have fruitful and aligned outcomes.
• Each one is a CEO of their domain.So, the criteria while making a choice is so our employees and clients can succeed together!

To read more about us click here:

About the role
We are building an exceptional team of Data engineers who are passionate developers and wants to push the boundaries to solve complex business problems using the latest tech stack. As a Big Data Engineer, you will work with various Technology and Business teams to deliver our Data Engineering offerings to our clients across the globe.

Role Description

• The role would involve big data pre-processing & reporting workflows including collecting, parsing, managing, analysing, and visualizing large sets of data to turn information into business insights
• Develop the software and systems needed for end-to-end execution on large projects
• Work across all phases of SDLC, and use Software Engineering principles to build scalable solutions
• Build the knowledge base required to deliver increasingly complex technology projects
• The role would also involve testing various machine learning models on Big Data and deploying learned models for ongoing scoring and prediction.

Education & Experience
• B.Tech. or Equivalent degree in CS/CE/IT/ECE/EEE 3+ years of experience designing technological solutions to complex data problems, developing & testing modular, reusable, efficient and scalable code to implement those solutions.

Must have (hands-on) experience
• Python and SQL expertise
• Distributed computing frameworks (Hadoop Ecosystem & Spark components)
• Must be proficient in any Cloud computing platforms (AWS/Azure/GCP)  • Experience in in any cloud platform would be preferred - GCP (Big Query/Bigtable, Pub sub, Data Flow, App engine )/ AWS/ Azure

• Linux environment, SQL and Shell scripting Desirable
• Statistical or machine learning DSL like R
• Distributed and low latency (streaming) application architecture
• Row store distributed DBMSs such as Cassandra, CouchDB, MongoDB, etc
. • Familiarity with API design

Hiring Process:
1. One phone screening round to gauge your interest and knowledge of fundamentals
2. An assignment to test your skills and ability to come up with solutions in a certain time
3. Interview 1 with our Data Engineer lead
4. Final Interview with our Data Engineer Lead and the Business Teams

Preferred Immediate Joiners

Read more
at RandomTrees
1 recruiter
Posted by HariPrasad Jonnalagadda
Remote only
5 - 10 yrs
₹15L - ₹30L / yr
Data Science
Machine Learning (ML)
Linear regression
Natural Language Processing (NLP)
Deep Learning
+5 more

Work Timings:4:00PM to 11:30PM
Fulltime WFH
6+ Yrs in Data science
Strong Experience ML Regression, Classification, Anomaly detection, NLP, Deep learning, Predictive analytics, Predictive maintenance ,Python, Added advantage Data visualization
Read more
Our Client company is into Telecommunications. (SY1)
Agency job
via Multi Recruit by Sukanya J
Remote, Bengaluru (Bangalore)
4 - 8 yrs
₹21L - ₹23L / yr
Machine Learning (ML)
Deep Learning
ML Tools
NLP Tools
+2 more
  • Participate in full machine learning Lifecycle including data collection, cleaning, preprocessing to training models, and deploying them to Production.
  • Discover data sources, get access to them, ingest them, clean them up, and make them “machine learning ready”. 
  • Work with data scientists to create and refine features from the underlying data and build pipelines to train and deploy models. 
  • Partner with data scientists to understand and implement machine learning algorithms. 
  • Support A/B tests, gather data, perform analysis, draw conclusions on the impact of your models. 
  • Work cross-functionally with product managers, data scientists, and product engineers, and communicate results to peers and leaders. 
  • Mentor junior team members 


Who we have in mind:

  • Graduate in Computer Science or related field, or equivalent practical experience. 
  • 4+ years of experience in software engineering with 2+ years of direct experience in the machine learning field.
  • Proficiency with SQL,  Python, Spark, and basic libraries such as Scikit-learn, NumPy, Pandas.
  • Familiarity with deep learning frameworks such as TensorFlow or Keras
  • Experience with Computer Vision (OpenCV),  NLP frameworks (NLTK, SpaCY, BERT).
  • Basic knowledge of machine learning techniques (i.e. classification, regression, and clustering). 
  • Understand machine learning principles (training, validation, etc.) 
  • Strong hands-on knowledge of data query and data processing tools (i.e. SQL) 
  • Software engineering fundamentals: version control systems (i.e. Git, Github) and workflows, and ability to write production-ready code. 
  • Experience deploying highly scalable software supporting millions or more users 
  • Experience building applications on cloud  (AWS or Azure) 
  • Experience working in scrum teams with Agile tools like JIRA
  • Strong oral and written communication skills. Ability to explain complex concepts and technical material to non-technical users





Read more
A Telecom Industry
Agency job
via Multi Recruit by Sukanya J
Bengaluru (Bangalore)
6 - 10 yrs
₹16L - ₹18L / yr
Big Data
Apache Spark
Spring Boot
  • Expert software implementation and automated testing
  • Promoting development standards, code reviews, mentoring, knowledge sharing
  • Improving our Agile methodology maturity
  • Product and feature design, scrum story writing
  • Build, release, and deployment automation
  • Product support & troubleshooting


Who we have in mind: 

  • Demonstrated experience as a Java
  • Should have a deep understanding of Enterprise/Distributed Architecture patterns and should be able to demonstrate the relevant usage of the same
  • Turn high-level project requirements into application-level architecture and collaborate with the team members to implement the solution
  • Strong experience and knowledge in Spring boot framework and microservice architecture
  • Experience in working with Apache Spark
  • Solid demonstrated object-oriented software development experience with Java, SQL, Maven, relational/NoSQL databases and testing frameworks 
  • Strong working experience with developing RESTful services
  • Should have experience working on Application frameworks such as Spring, Spring Boot, AOP
  • Exposure to tools – Jira, Bamboo, Git, Confluence would be an added advantage
  • Excellent grasp of the current technology landscape, trends and emerging technologies
Read more
at Busigence Technologies
1 video
1 recruiter
Posted by Seema Verma
Bengaluru (Bangalore)
1 - 5 yrs
₹10L - ₹50L / yr
Deep Learning
Artificial Intelligence (AI)
Machine Learning (ML)
Data Science
Artificial Neural Network (ANN)
+1 more
APPLY LINK: Go through the entire job post thoroughly before pressing Apply. There is an eleven characters french word v*n*i*r*t*e mentioned somewhere in the whole text which is irrelevant to the context. You shall be required to enter this word while applying else application won't be considered submitted. ````````````````````````````````````````````````````````````````````````````````````````````````````` Researcher - Deep Learning & AI Team: Research Full-Time, Permanent Bangaluru, India Relevant Exp: 6 months - 5 Years Background: Top Tier institute + Startup exposure Compensation: INR 10-50 LPA . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Busigence is a Decision Intelligence Company. We create decision intelligence products for real people by combining data, technology, business, and behavior enabling strengthened decisions. Scaling established startup by IIT alumni innovating & disrupting marketing domain through artificial intelligence. We bring those people onboard who are dedicated to deliver wisdom to humanity by solving the world’s most pressing problems differently thereby significantly impacting thousands of souls, everyday. We are a deep rooted organization with six years of success story having worked with folks from top tier background (IIT, NSIT, DCE, BITS, IIITs, NITs, IIMs, ISI etc.) maintaining an awesome culture with a common vision to build great data products. In past we have served fifty five customers and presently developing our second product, Robonate. First was emmoQ - an emotion intelligence platform. Third offering, H2HData, an innovation lab where we solve hard problems through data, science, & design. We work extensively & intensely on big data, data science, machine learning, deep learning, reinforcement learning, data analytics, natural language processing, cognitive computing, and business intelligence. First-and-Foremost Before you dive-in exploring this opportunity and press Apply, we wish you to evaluate yourself - We are looking for right candidate, not the best candidate. We love to work with someone who can mandatorily gel with our vision, beliefs, thoughts, methods, and values --- which are aligned with what can be expected in a true startup with ambitious goals. Skills are always secondary to us. Primarily, you must be someone who is not essentially looking for a job or career, rather starving for a challenge, you yourself probably don't know since when. A book can be written on what an applicant must have before joining a . For brevity, in nutshell, we need these three in you: 1. You must be [super sharp] (Just an analogue, but Irodov, Mensa, Feynman, Polya, ACM, NIPS, ICAAC, BattleCode, DOTA etc should have been your Done stuff. Can you relate solution 1 to problem 2? or Do you get confused even when solved similar problem in past? Are you able to grasp problem statement in one go? or get hanged?) 2. You must be [extremely energetic] (Do you raise eyebrows when asked to stretch your limits, both in terms of complexity or extra hours to put in? What comes first in your mind, let's finish it today or this can be done tomorrow too? Its Friday 10 PM at work -Tired?) 3. You must be [honourably honest] (Do you tell others what you think, or what they want to hear? Later is good for sales team for their customers, not for this role. Are you honest with your work? intrinsically with yourself first?) You know yourself the best. If not ask your loved ones and then decide. We clearly need exceedingly motivated people with entrepreneurial traits, not employee mindset - not at all. This is an immediate requirement. We shall have an accelerated interview process for fast closure - you would be required to be proactive and responsive. Real ROLE We are looking for researchers with real passion for deep learning. You would be required to work with our research team on several deep learning applications. Mandatory R1. Must know in-and-out of functional programming in Python with strong flair for data structures, linear algebra, & algorithms implementation R2. Must be able to apply state-of-art deep learning models to our existing research and develop entirely new methods, which are advanced & modern R3. Must have designed multiple deep learning techniques solving complex business scenarios preferably from marketing domain across industries R4. Must have designed proven pre-processing techniques on unstructured (image, audio, video, text, logs) and time-series data R5. Must have designed CNN, RNN (atleast three variants including LSTM), Auto Encoders, DBN, MLP, RBM R6. Must have tweaked & trained on deep learning architectures mainly AlexNet, ResNet, RCNN, GAN R7. Must have worked on classification, clustering, computer vision, speech recognition, text summarization, time-series analysis R8. Must have firm understanding of the underlying mathematics will be needed to adapt modelling techniques to fit the problem space with large data (10M+ entities) R9. Must be extremely strong working knowledge on Python scientific stack (NumPy, SciPy, Matplotlib, Scikit-learn, Pandas, Stats model, Seaborn, PyBrain) and atleast basic working knowledge on TensorFlow. R/ SAS/ SPSS etc are not acceptable. Preferred R10. Good to have knowledge and practical experience in application of Supervised Learning algorithms, Bayesian methods, Gaussian processes, Combinatorial optimization, Meta Learning, Transfer Learning, Convex optimization, Sparse methods, Kernel methods, Stochastic process, Statistical learning theory, and Approximate inference R11. Good to have working knowledge of first-class, high order, & pure functions, recurisons, lazy evaluations, and immutable data structures R12. Good to have worked on Tensorboard, Saving & Restoring models, Continue Training R13. Good to have worked on Keras on top of TensorFlow R14. Good to have worked on Reinforcement Learning and Generative Adverbial Networks R15. Great problem solver with extremely high curiosity on data processing & patterns Ideal YOU Y1. Degree in engineering, or any other data-heavy field at Bachelors/Masters/Ph.D level or above from a top tier institute Y2. Relevant experience of 0.5 - 5 years working on real-world problems in a research lab or a reputed company or a proven startup Y3. You are a fanatical implementer who love to spend time with findings & innovations, more than your loved ones Y4. You are true believer that human intelligence can be augmented through computer science & mathematics and your survival vinaigrette depends on getting the most from the data Y5. You are an entrepreneur mindset with ownership, intellectuality, & creativity as way to work. These are not fancy words, we mean it Actual WE W1. Real startup with Meaningful products W2. Revolutionary not just disruptive W3. Rules creators not followers W4. Small teams with real brains not herd of blockheads W5. Completely trust us and should be trusted back Why Us In addition to the regular stuff which every good startup offers – Lots of learning, Food, Parties, Open culture, Flexible working hours, and what not…. We offer you: You shall be working on our revolutionary products which are pioneer in their respective categories. This is a fact. We try real hard to hire fun loving crazy folks who are driven by more than a paycheck. You shall be working with creamiest talent on extremely challenging problems at most happening workplace. How to Apply You should apply online by clicking "Apply Now". For queries regarding an open position, please write to [email protected] For more information, visit Careers: Research: Jobs: Feel right fit for the position, mandatorily attach PDF resume highlighting your A. Key Skills B. Knowledge Inputs C. Major Accomplishments D. Problems Solved E. Submissions – Github/ StackOverflow/ Kaggle/ Euler Project etc. (if applicable) If you don't see this open position that interests you, join our Talent Pool and let us know how you can make a difference here. Referrals are more than welcome. Keep us in loop.
Read more
Posted by Vipul Tiwari
3 - 8 yrs
₹5L - ₹30L / yr
Data Science
Machine Learning (ML)
Big Data
+3 more
Job Description for Data Scientist/ NLP Engineer

Responsibilities for Data Scientist/ NLP Engineer

Work with customers to identify opportunities for leveraging their data to drive business
• Develop custom data models and algorithms to apply to data sets.
• Basic data cleaning and annotation for any incoming raw data.
• Use predictive modeling to increase and optimize customer experiences, revenue
generation, ad targeting and other business outcomes.
• Develop company A/B testing framework and test model quality.
• Deployment of ML model in production.
Qualifications for Junior Data Scientist/ NLP Engineer

• BS, MS in Computer Science, Engineering, or related discipline.
• 3+ Years of experience in Data Science/Machine Learning.
• Experience with programming language Python.
• Familiar with at least one database query language, such as SQL
• Knowledge of Text Classification & Clustering, Question Answering & Query Understanding,
Search Indexing & Fuzzy Matching.
• Excellent written and verbal communication skills for coordinating acrossteams.
• Willing to learn and master new technologies and techniques.
• Knowledge and experience in statistical and data mining techniques:
GLM/Regression, Random Forest, Boosting, Trees, text mining, NLP, etc.
• Experience with chatbots would be bonus but not required
Read more
at Yulu Bikes
1 video
3 recruiters
Posted by Keerthana k
Bengaluru (Bangalore)
1 - 2 yrs
₹7L - ₹12L / yr
Data Science
Data Analytics
+2 more
Skill Set 
SQL, Python, Numpy,Pandas,Knowledge of Hive and Data warehousing concept will be a plus point.


- Strong analytical skills with the ability to collect, organise, analyse and interpret trends or patterns in complex data sets and provide reports & visualisations.

- Work with management to prioritise business KPIs and information needs Locate and define new process improvement opportunities.

- Technical expertise with data models, database design and development, data mining and segmentation techniques

- Proven success in a collaborative, team-oriented environment

- Working experience with geospatial data will be a plus.
Read more
Did not find a job you were looking for?
Search for relevant jobs from 10000+ companies such as Google, Amazon & Uber actively hiring on Cutshort.
Get to hear about interesting companies hiring right now
iconFollow Cutshort
Want to apply to this role at xpressbees?
Why apply via Cutshort?
Connect with actual hiring teams and get their fast response. No spam.
Learn more
Get to hear about interesting companies hiring right now
iconFollow Cutshort