Bigdata Lead

at Saama Technologies

DP
Posted by Sandeep Chaudhary
icon
Pune
icon
2 - 5 yrs
icon
₹1L - ₹18L / yr
icon
Full time
Skills
Hadoop
Spark
Apache Hive
Apache Flume
Java
Python
Scala
MySQL
Game Design
Technical Writing
Description Deep experience and understanding of Apache Hadoop and surrounding technologies required; Experience with Spark, Impala, Hive, Flume, Parquet and MapReduce. Strong understanding of development languages to include: Java, Python, Scala, Shell Scripting Expertise in Apache Spark 2. x framework principals and usages. Should be proficient in developing Spark Batch and Streaming job in Python, Scala or Java. Should have proven experience in performance tuning of Spark applications both from application code and configuration perspective. Should be proficient in Kafka and integration with Spark. Should be proficient in Spark SQL and data warehousing techniques using Hive. Should be very proficient in Unix shell scripting and in operating on Linux. Should have knowledge about any cloud based infrastructure. Good experience in tuning Spark applications and performance improvements. Strong understanding of data profiling concepts and ability to operationalize analyses into design and development activities Experience with best practices of software development; Version control systems, automated builds, etc. Experienced in and able to lead the following phases of the Software Development Life Cycle on any project (feasibility planning, analysis, development, integration, test and implementation) Capable of working within the team or as an individual Experience to create technical documentation

About Saama Technologies

Saama is a US headquartered, global leader in Big Data and Analytics space. Our unique and focused solutions are helping clients in the Insurance, Life Sciences, Healthcare and CPG industry verticals to achieve faster business results through data driven insights.
Founded
1997
Type
Products & Services
Size
100-1000 employees
Stage
Profitable
View full company details
Why apply to jobs via Cutshort
Personalized job matches
Stop wasting time. Get matched with jobs that meet your skills, aspirations and preferences.
Verified hiring teams
See actual hiring teams, find common social connections or connect with them directly. No 3rd party agencies here.
Move faster with AI
We use AI to get you faster responses, recommendations and unmatched user experience.
2101133
Matches delivered
3712187
Network size
15000
Companies hiring

Similar jobs

DevOps Engineer

at Panamax InfoTech Ltd.

Founded 2004  •  Products & Services  •  100-1000 employees  •  Profitable
Data Warehouse (DWH)
Informatica
ETL
DevOps
Python
Perl
Java
.NET
Shell Scripting
Bash
Terraform
SVN
Maven
Git
Docker
Kubernetes
Chef
Ansible
Puppet
Splunk
Gradle
Software deployment
helm
icon
Remote only
icon
3 - 4 yrs
icon
₹5L - ₹10L / yr
1. Excellent understanding of at least any one of the programming language .NET ,Python, Perl, and Java
2. Good understanding and hands on experience in Shell/Bash scripting, sonarqube, Terraform,
3. Experience with Continuous Integration and Continuous Deployment Pipelines
4. Experience in SVN, Maven, Git and Git workflows
5. Should be able to develop overall strategy for Build & Release management
6. Experience in working with container orchestration tools such as Docker and Kubernetes
7. Good knowledge in Devops Automation Tools like Chef, Ansible, Puppet, helm, splunk, maven, gradle & XL Deploy.etc
8. Managing stakeholders and external interfaces and Setting up tools and required infrastructure
9. Encouraging and building automated processes wherever possible
10. Awareness of critical concepts in DevOps and Agile principles
11. Experience in Cloud infrastructure like AWS, GCP or Azure. In AWS understanding on EC2, S3 & cloud
12. Strong knowledge and hands on experience in unix OS
13.Experience in network, server, application status monitoring and troubleshooting, Security.
14.Design, develop automation suite and integrate with continuous integration process through Jenkins
15. Possess good problem solving and debugging skills. Troubleshoot issues and coordinate with development team to streamline code deployment to generate build
Job posted by
Bhavani P

Senior Software Engineer - Analytics

at Enterprise Artificial Intelligence

Agency job
via Purple Hirez
Analytics
Kubernetes
Apache Kafka
Data Analytics
Python
Spring Boot
Play Framework
ELK
icon
Hyderabad
icon
5 - 12 yrs
icon
₹10L - ₹35L / yr
  • 3+ years of industry experience in administering (including setting up, managing, monitoring) data processing pipelines (both streaming and batch) using frameworks such as Kafka, ELK Stack, Fluentd and streaming databases like druid
  • Strong industry expertise with containerization technologies including kubernetes, docker-compose
  • 2+ years of industry in experience in developing scalable data ingestion processes and ETLs
  • Experience with cloud platform services such as AWS, Azure or GCP especially with EKS, Managed Kafka
  • Experience with scripting languages. Python experience highly desirable.
  • 2+ Industry experience in python
  • Experience with popular modern web frameworks such as Spring boot, Play framework, or Django
  • Demonstrated expertise of building cloud native applications
  • Experience in administering (including setting up, managing, monitoring) data processing pipelines (both streaming and batch) using frameworks such as Kafka, ELK Stack, Fluentd
  • Experience in API development using Swagger
  • Strong expertise with containerization technologies including kubernetes, docker-compose
  • Experience with cloud platform services such as AWS, Azure or GCP.
  • Implementing automated testing platforms and unit tests
  • Proficient understanding of code versioning tools, such as Git
  • Familiarity with continuous integration, Jenkins
Responsibilities
  • Design and Implement Large scale data processing pipelines using Kafka, Fluentd and Druid
  • Assist in dev ops operations
  • Develop data ingestion processes and ETLs
  • Design and Implement APIs
  • Assist in dev ops operations
  • Identify performance bottlenecks and bugs, and devise solutions to these problems
  • Help maintain code quality, organization, and documentation
  • Communicate with stakeholders regarding various aspects of solution.
  • Mentor team members on best practices
Job posted by
Aditya K

Data Engineer_Scala

at Ganit Business Solutions

Founded 2017  •  Products & Services  •  100-1000 employees  •  Bootstrapped
ETL
Informatica
Data Warehouse (DWH)
Big Data
Scala
Hadoop
Apache Hive
PySpark
Spark
icon
Remote only
icon
4 - 7 yrs
icon
₹10L - ₹30L / yr

Job Description:

We are looking for a Big Data Engineer who have worked across the entire ETL stack. Someone who has ingested data in a batch and live stream format, transformed large volumes of daily and built Data-warehouse to store the transformed data and has integrated different visualization dashboards and applications with the data stores.    The primary focus will be on choosing optimal solutions to use for these purposes, then maintaining, implementing, and monitoring them.

Responsibilities:

  • Develop, test, and implement data solutions based on functional / non-functional business requirements.
  • You would be required to code in Scala and PySpark daily on Cloud as well as on-prem infrastructure
  • Build Data Models to store the data in a most optimized manner
  • Identify, design, and implement process improvements: automating manual processes, optimizing data delivery, re-designing infrastructure for greater scalability, etc.
  • Implementing the ETL process and optimal data pipeline architecture
  • Monitoring performance and advising any necessary infrastructure changes.
  • Create data tools for analytics and data scientist team members that assist them in building and optimizing our product into an innovative industry leader.
  • Work with data and analytics experts to strive for greater functionality in our data systems.
  • Proactively identify potential production issues and recommend and implement solutions
  • Must be able to write quality code and build secure, highly available systems.
  • Create design documents that describe the functionality, capacity, architecture, and process.
  • Review peer-codes and pipelines before deploying to Production for optimization issues and code standards

Skill Sets:

  • Good understanding of optimal extraction, transformation, and loading of data from a wide variety of data sources using SQL and ‘big data’ technologies.
  • Proficient understanding of distributed computing principles
  • Experience in working with batch processing/ real-time systems using various open-source technologies like NoSQL, Spark, Pig, Hive, Apache Airflow.
  • Implemented complex projects dealing with the considerable data size (PB).
  • Optimization techniques (performance, scalability, monitoring, etc.)
  • Experience with integration of data from multiple data sources
  • Experience with NoSQL databases, such as HBase, Cassandra, MongoDB, etc.,
  • Knowledge of various ETL techniques and frameworks, such as Flume
  • Experience with various messaging systems, such as Kafka or RabbitMQ
  • Creation of DAGs for data engineering
  • Expert at Python /Scala programming, especially for data engineering/ ETL purposes

 

 

 

Job posted by
Vijitha VS

Business Analyst

at GreedyGame

Founded 2013  •  Product  •  20-100 employees  •  Raised funding
MS-Excel
SQL
Data Analytics
Python
R Language
Business Analysis
icon
Bengaluru (Bangalore)
icon
1 - 2 yrs
icon
₹4L - ₹12L / yr

About Us:

GreedyGame is looking for a Business Analyst to join its clan. We are looking to get an enthusiastic Business Analyst who likes to play with Data. You'll be building insights from Data, creating analytical dashboard and monitoring KPI values. Also you will coordinate with teams working on different layers of the infrastructure.

 

Job details:

 

Seniority Level: Associate

Level Industry: Marketing & Advertising

Employment Type: Full Time

Job Location: Bangalore

Experience: 1-2 years

 

WHAT ARE WE LOOKING FOR?

 

  • Excellent planning, organizational, and time management skills.
  • Exceptional analytical and conceptual thinking skills.
  • A previous experience of working closely with Operations and Product Teams.
  • Competency in Excel and SQL is a must.
  • Experience with a programming language like Python is required.
  • Knowledge of Marketing Tools is preferable.

 

 

WHAT WILL BE YOUR RESPONSIBILITIES?

 

  • Evaluating business processes, anticipating requirements, uncovering areas for improvement, developing and implementing solutions.
  • Should be able to generate meaningful insights to help the marketing team and product team in enhancing the user experience for Mobile and Web Apps.
  • Leading ongoing reviews of business processes and developing optimization strategies.
  • Performing requirements analysis from a user and business point of view
  • Combining data from multiple sources like SQL tables, Google Analytics, Inhouse Analytical signals etc and driving relevant insights
  • Deciding the success metrics and KPIs for different Products and features and making sure they are achieved.
  • Act as quality assurance liaison prior to the release of new data analysis or application.

 

Skills and Abilities:

  • Python
  • SQL
  • Business Analytics
  • BigQuery

 

WHAT'S IN IT FOR YOU?

  • An opportunity to be a part of a fast scaling start-up in the AdTech space that offers unmatched services and products.
  • To work with a team of young enthusiasts who are always upbeat and self-driven to achieve bigger milestones in shorter time spans.
  • A workspace that is wide open as per the open door policy at the company, located in the most happening center of Bangalore.
  • A well-fed stomach makes the mind work better and therefore we provide - free lunch with a wide variety on all days of the week, a stocked-up pantry to satiate your want for munchies, a Foosball table to burst stress and above all a great working environment.
  • We believe that we grow as you grow. Once you are a part of our team, your growth also becomes essential to us, and in order to make sure that happens, there are timely formal and informal feedbacks given
Job posted by
Shreyoshi Ghosh
Apache Hadoop
Hadoop
Apache Hive
HDFS
SSL
Hortonworks
icon
Bengaluru (Bangalore)
icon
5 - 9 yrs
icon
₹16L - ₹20L / yr
  • Responsibilities
         - Responsible for implementation and ongoing administration of Hadoop
    infrastructure.
         - Aligning with the systems engineering team to propose and deploy new
    hardware and software environments required for Hadoop and to expand existing
    environments.
         - Working with data delivery teams to setup new Hadoop users. This job includes
    setting up Linux users, setting up Kerberos principals and testing HDFS, Hive, Pig
    and MapReduce access for the new users.
         - Cluster maintenance as well as creation and removal of nodes using tools like
    Ganglia, Nagios, Cloudera Manager Enterprise, Dell Open Manage and other tools
         - Performance tuning of Hadoop clusters and Hadoop MapReduce routines
         - Screen Hadoop cluster job performances and capacity planning
         - Monitor Hadoop cluster connectivity and security
         - Manage and review Hadoop log files.
         - File system management and monitoring.
         - Diligently teaming with the infrastructure, network, database, application and
    business intelligence teams to guarantee high data quality and availability
         - Collaboration with application teams to install operating system and Hadoop
    updates, patches, version upgrades when required.
        
    READ MORE OF THE JOB DESCRIPTION 
    Qualifications
    Qualifications
         - Bachelors Degree in Information Technology, Computer Science or other
    relevant fields
         - General operational expertise such as good troubleshooting skills,
    understanding of systems capacity, bottlenecks, basics of memory, CPU, OS,
    storage, and networks.
         - Hadoop skills like HBase, Hive, Pig, Mahout
         - Ability to deploy Hadoop cluster, add and remove nodes, keep track of jobs,
    monitor critical parts of the cluster, configure name node high availability, schedule
    and configure it and take backups.
         - Good knowledge of Linux as Hadoop runs on Linux.
         - Familiarity with open source configuration management and deployment tools
    such as Puppet or Chef and Linux scripting.
         Nice to Have
         - Knowledge of Troubleshooting Core Java Applications is a plus.

Job posted by
Harpreet kour

Machine Learning Engineer

at SmartJoules

Founded 2015  •  Product  •  100-500 employees  •  Profitable
Machine Learning (ML)
Python
Big Data
Apache Spark
Deep Learning
icon
Remote, NCR (Delhi | Gurgaon | Noida)
icon
3 - 5 yrs
icon
₹8L - ₹12L / yr

Responsibilities:

  • Exploring and visualizing data to gain an understanding of it, then identifying differences in data distribution that could affect performance when deploying the model in the real world.
  • Verifying data quality, and/or ensuring it via data cleaning.
  • Able to adapt and work fast in producing the output which upgrades the decision making of stakeholders using ML.
  • To design and develop Machine Learning systems and schemes. 
  • To perform statistical analysis and fine-tune models using test results.
  • To train and retrain ML systems and models as and when necessary. 
  • To deploy ML models in production and maintain the cost of cloud infrastructure.
  • To develop Machine Learning apps according to client and data scientist requirements.
  • To analyze the problem-solving capabilities and use-cases of ML algorithms and rank them by how successful they are in meeting the objective.


Technical Knowledge:


  • Worked with real time problems, solved them using ML and deep learning models deployed in real time and should have some awesome projects under his belt to showcase. 
  • Proficiency in Python and experience with working with Jupyter Framework, Google collab and cloud hosted notebooks such as AWS sagemaker, DataBricks etc.
  • Proficiency in working with libraries Sklearn, Tensorflow, Open CV2, Pyspark,  Pandas, Numpy and related libraries.
  • Expert in visualising and manipulating complex datasets.
  • Proficiency in working with visualisation libraries such as seaborn, plotly, matplotlib etc.
  • Proficiency in Linear Algebra, statistics and probability required for Machine Learning.
  • Proficiency in ML Based algorithms for example, Gradient boosting, stacked Machine learning, classification algorithms and deep learning algorithms. Need to have experience in hypertuning various models and comparing the results of algorithm performance.
  • Big data Technologies such as Hadoop stack and Spark. 
  • Basic use of clouds (VM’s example EC2).
  • Brownie points for Kubernetes and Task Queues.      
  • Strong written and verbal communications.
  • Experience working in an Agile environment.
Job posted by
Saksham Dutta

Lead Data Scientist

at Spotmentor Technologies

Founded 2018  •  Product  •  20-100 employees  •  Raised funding
Python
Machine Learning (ML)
Natural Language Processing (NLP)
NOSQL Databases
icon
NCR (Delhi | Gurgaon | Noida)
icon
2 - 5 yrs
icon
₹20L - ₹30L / yr
Spotmentor is focussed on using the Intelligence-age tools and technologies like AI and Text analytics to create HR technology products which go beyond compliance and ERPs to give HR the power to become strategic, improve business results and increase the competitiveness. The HR and People departments have long sought to become strategic partners with businesses. We are focussed on taking this concept out of the board room meetings and making it a reality and you can be a part of this journey. At the end of it, you would be able to claim that there was an inflection point in History, which changed how business was transacted and you made that happen. Our first product is a Learning and Skill development platform which helps the organisations to acquire capabilities critical for them by helping employees attain their best potential through learning opportunities. Spotmentor was started by 4 IIT Kharagpur alumni with experiences in creating Technology products and Management consulting. We are looking for a Data Scientist who will help discover the information hidden in vast amounts of data, and help us make smarter decisions that benefit the employees of our customer organisations. Your primary focus will be on applying data mining techniques, doing statistical analysis, and building high quality prediction systems using structured and unstructured data. Technical Responsibilities: - Selecting features, building and optimizing classifiers using machine learning techniques - Data mining using state-of-the-art methods - Extending the existing data sets with third party sources of information - Processing, cleansing, and verifying the integrity of data used for analysis - Build recommendation systems - Automate scoring of documents using machine learning techniques Salary: This is a founding team member role with a salary of 20 Lacs to 30 Lacs per year and a meaningful ESOP component. Location: Gurgaon We believe in making Spotmentor the best place for the pursuit of excellence and diversity of opinions is an important tool to achieve that. Although as a startup our primary objective is growth, Spotmentor is focussed on creating a diverse and inclusive workplace where everyone can attain their best potential and we welcome female, minority and specially abled candidates to apply.
Job posted by
Deepak Singh

Data Science (R / Python)

at Saama Technologies

Founded 1997  •  Products & Services  •  100-1000 employees  •  Profitable
Data Analytics
Data Science
Product Management
Machine Learning (ML)
Python
SAS
icon
Pune
icon
7 - 11 yrs
icon
₹6L - ₹22L / yr
Description Does solving complex business problems and real world challenges interest you Do you enjoy seeing the impact your contributions make on a daily basis Are you passionate about using data analytics to provide game changing solutions to the Global 2000 clients Do you thrive in a dynamic work environment that constantly pushes you to be the best you can be and more Are you ready to work with smart colleagues who drive for excellence in everything they do If you possess a solutions mindset, strong analytical skills, and commitment to be part of a tremendous journey, come join our growing, global team. See what Saama can do for your career and for your journey. Position: Data Scientist (2255) Location: Hinjewadi Phase 1, Pune Type: Permanent Full time Responsibilities: Work on small and large data sets of structured, semi-structured, and unstructured data to discover hidden knowledge about the client s business and develop methods to leverage that knowledge for their business. Identify and solve business challenges working closely with cross-functional teams, such as Delivery, Business Consulting, Engineering and Product Management. Develop prescriptive and predictive statistical, behavioral or other models via machine learning and/ or traditional statistical modeling techniques, and understand which type of model applies best in a given business situation. Drive the collection of new data and the refinement of existing data sources. Analyze and interpret the results of product experiments. Provide input for engineering and product teams as they develop and support our internal data platform to support ongoing statistical analyses. Requirements: Candidates should demonstrate following expertize- Must have Direct Hands-on, 7 years of experience, building complex systems using any Statistical Programming Language (R / Python / SAS) Must have fundamental knowledge of Inferential Statistics Should have worked on Predictive Modelling Experience should include the following, o File I/ O, Data Harmonization, Data Exploration o Multi-Dimensional Array Processing o Simulation & Optimization Techinques o Machine Learning Techniques (Supervised, Unsupervised) o Artificial Intelligence and Deep Learning o Natural Language Processing o Model Ensembling Techniques o Documenting Reproducible Research o Building Interactive Applications to demonstrate Data Science Use Cases Prior experience in Healthcare Domain, is a plus Experience using Big Data, is a plus. Exposure to SPARK is desirable Should have Excellent Analytical, Problem Solving ability. Should be able to grasp new concepts quickly Should be well familiar with Agile Project Management Methodology Should have experience of managing multiple simultaneous projects Should have played a team lead role Should have excellent written and verbal communication skills Should be a team player with open mind Impact on the business: Plays an important role in making Saama s Solutions game changers for our strategic partners by using data science to solve core, complex business challenges. Key relationships: Sales & pre-sales Product management Engineering Client organization: account management & delivery Saama Competencies: INTEGRITY: we do the right things. INNOVATION: we change the game. TRANSPARENCY: we communicate openly COLLABORATION: we work as one team PROBLEM-SOLVING: we solve core, complex business challenges ENJOY & CELEBRATE: we have fun Competencies: Self-starter who gets results with minimal support and direction in a fast-paced environment. Takes initiative; challenges the status quo to drive change. Learns quickly; takes smart risks to experiment and learn. Works well with others; builds trust and maintains credibility. Planful: identifies and confirms key requirements in dynamic environments; anticipates tasks and contingencies. Communicates effectively; productive communication with clients and all key stakeholders communication in both verbal and written communication. Stays the course despite challenges & setbacks. Works well under pressure. Strong analytical skills; able to apply inductive and deductive thinking to generate solutions for complex problems
Job posted by
Sandeep Chaudhary

Data Scientist

at TintED

Founded 2018  •  Services  •  0-20 employees  •  Bootstrapped
Data Science
Python
R Programming
icon
Remote, Kolkata
icon
0 - 4 yrs
icon
₹3L - ₹7L / yr
We aim to transform recruiting industry.
Job posted by
Kumar Aniket

Artificial Intelligence Developers

at Precily Private Limited

Founded 2016  •  Product  •  20-100 employees  •  Raised funding
Data Science
Artificial Neural Network (ANN)
Artificial Intelligence (AI)
Machine Learning (ML)
Python
TensorFlow
Natural Language Processing (NLP)
Big Data
icon
NCR (Delhi | Gurgaon | Noida)
icon
1 - 3 yrs
icon
₹3L - ₹9L / yr
-Precily AI: Automatic summarization, shortening a business document, book with our AI. Create a summary of the major points of the original document. AI can make a coherent summary taking into account variables such as length, writing style, and syntax. We're also working in the legal domain to reduce the high number of pending cases in India. We use Artificial Intelligence and Machine Learning capabilities such as NLP, Neural Networks in Processing the data to provide solutions for various industries such as Enterprise, Healthcare, Legal.
Job posted by
Bharath Rao
Did not find a job you were looking for?
icon
Search for relevant jobs from 10000+ companies such as Google, Amazon & Uber actively hiring on Cutshort.
Get to hear about interesting companies hiring right now
iconFollow Cutshort
Want to apply to this role at Saama Technologies?
Why apply via Cutshort?
Connect with actual hiring teams and get their fast response. No spam.
Learn more
Get to hear about interesting companies hiring right now
iconFollow Cutshort