Machine Learning Engineer
We are looking for a Machine Learning engineer for on of our premium client.
Experience: 2-9 years
Location: Gurgaon/Bangalore
Tech Stack:
Python, PySpark, the Python Scientific Stack; MLFlow, Grafana, Prometheus for machine learning pipeline management and monitoring; SQL, Airflow, Databricks, our own open-source data pipelining framework called Kedro, Dask/RAPIDS; Django, GraphQL and ReactJS for horizontal product development; container technologies such as Docker and Kubernetes, CircleCI/Jenkins for CI/CD, cloud solutions such as AWS, GCP, and Azure as well as Terraform and Cloudformation for deployment
About Top Management Consulting Company
Similar jobs
- Experience with Cloud native Data tools/Services such as AWS Athena, AWS Glue, Redshift Spectrum, AWS EMR, AWS Aurora, Big Query, Big Table, S3, etc.
- Strong programming skills in at least one of the following languages: Java, Scala, C++.
- Familiarity with a scripting language like Python as well as Unix/Linux shells.
- Comfortable with multiple AWS components including RDS, AWS Lambda, AWS Glue, AWS Athena, EMR. Equivalent tools in the GCP stack will also suffice.
- Strong analytical skills and advanced SQL knowledge, indexing, query optimization techniques.
- Experience implementing software around data processing, metadata management, and ETL pipeline tools like Airflow.
Experience with the following software/tools is highly desired:
- Apache Spark, Kafka, Hive, etc.
- SQL and NoSQL databases like MySQL, Postgres, DynamoDB.
- Workflow management tools like Airflow.
- AWS cloud services: RDS, AWS Lambda, AWS Glue, AWS Athena, EMR.
- Familiarity with Spark programming paradigms (batch and stream-processing).
- RESTful API services.
Essential requirements
- 5+ years experience working as a Data Scientist or ML Engineer in industry (preferably in advertising) is a must, where you have gained hands-on experience delivering data science products,
- Experience in taking data science projects from concept to production in an industry setting,
- Fluent in statistical analysis, data mining, and machine learning,
- Strong application and ML development experience in Python, functional as well as object oriented programming,
- A postgraduate degree in a relevant quantitative field (e.g. applied mathematics, statistics, engineering, computer science, physics, operations research, economics, behavioral sciences) is a plus,
- Proficient in SQL, PySpark, TensorFlow, Sklearn,
- Experience in GCP or AWS
- Experience with developing or supporting high performance APIs,
- Enthusiasm for solving interesting problems,
- Willingness to learn and work in a team-oriented and constantly changing environment.
Domain:
- In this role your main focus would be on our Ad-Exchange Optimization projects
- Bid Price Optimization - Optimize to improve the likelihood of winning auctions with supply-side partners,
- Floor Price Optimization - Optimize to improve demand-side valuation of supply,
- Supply Path Optimization - Optimize auctions to improve efficient utilization of finite bandwidth with demand-side partners,
- Auction Models - Optimizing different auction setups - 1st price, 2nd price, Waterfall.
Research and Development
- Our Machine Learning Engineer role includes the following responsibilities:
- Designing, developing, and researching cutting edge Machine Learning systems, models, and schemes in many different areas of Adtech,
- Developing real-time algorithms for campaign and bidding optimization,
- Discovering insights/patterns in exchange data, and developing methods to leverage/extract/process/analyze complex, high volume, high-dimensional datasets,
- Designing experiments, overseeing A/B testing, evaluating the quality of derived assets and developing dashboards to continuously monitor model performance,
- Studying, transforming, and converting data science prototypes,
- Searching and selecting appropriate data sets,
- Performing statistical analysis and using results to improve models,
Training and retraining ML systems and models as needed,
- Identifying differences in data distribution that could affect model performance in real-world situations.
- Visualizing data for deeper insights,
- Analyzing the use cases of ML algorithms and ranking them by their success probability,
- Understanding when your findings can be applied to business decisions,
- Reducing business problems into optimization problems,
- Supporting, maintaining and enriching existing ML frameworks and libraries,
- Verifying data quality and/or ensuring it via data cleaning.
What we offer
- Working on data science products which play an integral role in the company, and which are widely understood and highly valued,
- Unique opportunity to work on a wide variety of technically challenging and interesting data science projects, with hands-on exposure to a large number of current data science and engineering technologies and being able to solve complex productionising issues,
- Chance to work with truly big data: we are currently processing 120 billion auctions a day and counting,
- Join one of the top mobile advertising companies in the world, which is growing a phenomenal pace in terms of profits and market share,
- Be part of a friendly, supportive and multicultural team.
Professional experience in Python – Mandatory experience
Basic knowledge of any BI Tool (Microsoft Power BI, Tableau etc.) and experience in R
will be an added advantage
Proficient in Excel
Good verbal and written communication skills
Key Responsibilities:
Analyze data trends and provide intelligent business insights, monitor operational and
business metrics
Complete ownership of business excellence dashboard and preparation of reports for
senior management stating trends, patterns, and predictions using relevant data
Review, validate and analyse data points and implement new data analysis
methodologies
Perform data profiling to identify and understand anomalies
Perform analysis to assess quality and meaning of data
Develop policies and procedures for the collection and analysis of data
Analyse existing process with the help of data and propose process change and/or lead
process re-engineering initiatives
Use BI Tools (Microsoft Power BI/Tableau) and develop and manage BI solutions
- 3+ years experience in practical implementation and deployment of ML based systems preferred.
- BE/B Tech or M Tech (preferred) in CS/Engineering with strong mathematical/statistical background
- Strong mathematical and analytical skills, especially statistical and ML techniques, with familiarity with different supervised and unsupervised learning algorithms
- Implementation experiences and deep knowledge of Classification, Time Series Analysis, Pattern Recognition, Reinforcement Learning, Deep Learning, Dynamic Programming and Optimisation
- Experience in working on modeling graph structures related to spatiotemporal systems
- Programming skills in Python
- Experience in developing and deploying on cloud (AWS or Google or Azure)
- Good verbal and written communication skills
- Familiarity with well-known ML frameworks such as Pandas, Keras, TensorFlow
A Business Transformation Organization that partners with businesses to co–create customer-centric hyper-personalized solutions to achieve exponential growth. Invente offers platforms and services that enable businesses to provide human-free customer experience, Business Process Automation.
Location: Hyderabad (WFO)
Budget: Open
Position: Azure Data Engineer
Experience: 5+ years of commercial experience
Responsibilities
● Design and implement Azure data solutions using ADLS Gen 2.0, Azure Data Factory, Synapse, Databricks, SQL, and Power BI
● Build and maintain data pipelines and ETL processes to ensure efficient data ingestion and processing
● Develop and manage data warehouses and data lakes
● Ensure data quality, integrity, and security
● Implement from existing use cases required by the AI and analytics teams.
● Collaborate with other teams to integrate data solutions with other systems and applications
● Stay up-to-date with emerging data technologies and recommend new solutions to improve our data infrastructure
● Proficient in Python and using packages like NLTK, Numpy, Pandas
● Should have worked on deep learning frameworks (like Tensorflow, Keras, PyTorch, etc)
● Hands-on experience in Natural Language Processing, Sequence, and RNN Based models
● Mathematical intuition of ML and DL algorithms
● Should be able to perform thorough model evaluation by creating hypotheses on the basis of statistical
analyses
● Should be comfortable in going through open-source code and reading research papers.
Job description
Role : Lead Architecture (Spark, Scala, Big Data/Hadoop, Java)
Primary Location : India-Pune, Hyderabad
Experience : 7 - 12 Years
Management Level: 7
Joining Time: Immediate Joiners are preferred
- Attend requirements gathering workshops, estimation discussions, design meetings and status review meetings
- Experience of Solution Design and Solution Architecture for the data engineer model to build and implement Big Data Projects on-premises and on cloud.
- Align architecture with business requirements and stabilizing the developed solution
- Ability to build prototypes to demonstrate the technical feasibility of your vision
- Professional experience facilitating and leading solution design, architecture and delivery planning activities for data intensive and high throughput platforms and applications
- To be able to benchmark systems, analyses system bottlenecks and propose solutions to eliminate them
- Able to help programmers and project managers in the design, planning and governance of implementing projects of any kind.
- Develop, construct, test and maintain architectures and run Sprints for development and rollout of functionalities
- Data Analysis, Code development experience, ideally in Big Data Spark, Hive, Hadoop, Java, Python, PySpark,
- Execute projects of various types i.e. Design, development, Implementation and migration of functional analytics Models/Business logic across architecture approaches
- Work closely with Business Analysts to understand the core business problems and deliver efficient IT solutions of the product
- Deployment sophisticated analytics program of code using any of cloud application.
Perks and Benefits we Provide!
- Working with Highly Technical and Passionate, mission-driven people
- Subsidized Meals & Snacks
- Flexible Schedule
- Approachable leadership
- Access to various learning tools and programs
- Pet Friendly
- Certification Reimbursement Policy
- Check out more about us on our website below!
www.datametica.com
The candidate,
1. Must have a very good hands-on technical experience of 3+ years with JAVA or Python
2. Working experience and good understanding of AWS Cloud; Advanced experience with IAM policy and role management
3. Infrastructure Operations: 5+ years supporting systems infrastructure operations, upgrades, deployments using Terraform, and monitoring
4. Hadoop: Experience with Hadoop (Hive, Spark, Sqoop) and / or AWS EMR
5. Knowledge on PostgreSQL/MySQL/Dynamo DB backend operations
6. DevOps: Experience with DevOps automation - Orchestration/Configuration Management and CI/CD tools (Jenkins)
7. Version Control: Working experience with one or more version control platforms like GitHub or GitLab
8. Knowledge on AWS Quick sight reporting
9. Monitoring: Hands on experience with monitoring tools such as AWS CloudWatch, AWS CloudTrail, Datadog and Elastic Search
10. Networking: Working knowledge of TCP/IP networking, SMTP, HTTP, load-balancers (ELB) and high availability architecture
11. Security: Experience implementing role-based security, including AD integration, security policies, and auditing in a Linux/Hadoop/AWS environment. Familiar with penetration testing and scan tools for remediation of security vulnerabilities.
12. Demonstrated successful experience learning new technologies quickly
WHAT WILL BE THE ROLES AND RESPONSIBILITIES?
1. Create procedures/run books for operational and security aspects of AWS platform
2. Improve AWS infrastructure by developing and enhancing automation methods
3. Provide advanced business and engineering support services to end users
4. Lead other admins and platform engineers through design and implementation decisions to achieve balance between strategic design and tactical needs
5. Research and deploy new tools and frameworks to build a sustainable big data platform
6. Assist with creating programs for training and onboarding for new end users
7. Lead Agile/Kanban workflows and team process work
8. Troubleshoot issues to resolve problems
9. Provide status updates to Operations product owner and stakeholders
10. Track all details in the issue tracking system (JIRA)
11. Provide issue review and triage problems for new service/support requests
12. Use DevOps automation tools, including Jenkins build jobs
13. Fulfil any ad-hoc data or report request queries from different functional groups