11+ ATPG Jobs in Bangalore (Bengaluru) | ATPG Job openings in Bangalore (Bengaluru)
Apply to 11+ ATPG Jobs in Bangalore (Bengaluru) on CutShort.io. Explore the latest ATPG Job opportunities across top companies like Google, Amazon & Adobe.
Job Specification
- Experience in JTAG, MBIST , Scan Compression, ATPG, Fault Simulation.
- Experience with industry ATPG tools Synopsys Tetra MAX, Cadence Encounter Test or
- Mentor Fast Scan ATPG tools Synopsys DFT scan insertion.
- Experience with industry simulation tools such as VCS, ModelSim, NC Verilog etc.
- Pattern Simulation with and without timing annotation & debugging
JOB DETAILS:
- Job Title: Lead DevOps Engineer
- Industry: Ride-hailing
- Experience: 6-9 years
- Working Days: 5 days/week
- Work Mode: ONSITE
- Job Location: Bangalore
- CTC Range: Best in Industry
Required Skills: Cloud & Infrastructure Operations, Kubernetes & Container Orchestration, Monitoring, Reliability & Observability, Proficiency with Terraform, Ansible etc., Strong problem-solving skills with scripting (Python/Go/Shell)
Criteria:
1. Candidate must be from a product-based or scalable app-based start-ups company with experience handling large-scale production traffic.
2. Minimum 6 yrs of experience working as a DevOps/Infrastructure Consultant
3. Candidate must have 2 years of experience as an lead (handling team of 3 to 4 members at least)
4. Own end-to-end infrastructure right from non-prod to prod environment including self-managed
5. Candidate must have Self experience in database migration from scratch
6. Must have a firm hold on the container orchestration tool Kubernetes
7. Should have expertise in configuration management tools like Ansible, Terraform, Chef / Puppet
8. Understanding programming languages like GO/Python, and Java
9. Working on databases like Mongo/Redis/Cassandra/Elasticsearch/Kafka.
10. Working experience on Cloud platform -AWS
11. Candidate should have Minimum 1.5 years stability per organization, and a clear reason for relocation.
Description
Job Summary:
As a DevOps Engineer at company, you will be working on building and operating infrastructure at scale, designing and implementing a variety of tools to enable product teams to build and deploy their services independently, improving observability across the board, and designing for security, resiliency, availability, and stability. If the prospect of ensuring system reliability at scale and exploring cutting-edge technology to solve problems, excites you, then this is your fit.
Job Responsibilities:
● Own end-to-end infrastructure right from non-prod to prod environment including self-managed DBs
● Codify our infrastructure
● Do what it takes to keep the uptime above 99.99%
● Understand the bigger picture and sail through the ambiguities
● Scale technology considering cost and observability and manage end-to-end processes
● Understand DevOps philosophy and evangelize the principles across the organization
● Strong communication and collaboration skills to break down the silos
Job Requirements:
● B.Tech. / B.E. degree in Computer Science or equivalent software engineering degree/experience
● Minimum 6 yrs of experience working as a DevOps/Infrastructure Consultant
● Must have a firm hold on the container orchestration tool Kubernetes
● Must have expertise in configuration management tools like Ansible, Terraform, Chef / Puppet
● Strong problem-solving skills, and ability to write scripts using any scripting language
● Understanding programming languages like GO/Python, and Java
● Comfortable working on databases like Mongo/Redis/Cassandra/Elasticsearch/Kafka.
What’s there for you?
Company’s team handles everything – infra, tooling, and self-manages a bunch of databases, such as
● 150+ microservices with event-driven architecture across different tech stacks Golang/ java/ node
● More than 100,000 Request per second on our edge gateways
● ~20,000 events per second on self-managed Kafka
● 100s of TB of data on self-managed databases
● 100s of real-time continuous deployment to production
● Self-managed infra supporting
● 100% OSS
Job Title : Senior Salesforce Marketing Cloud Developer
Experience : 5–8 years
Location : Bangalore (On-site)
Notice Period : Immediate to 15 Days Only
Job Description :
We are looking for a skilled Senior Salesforce Marketing Cloud Developer with 5 to 8 years of hands-on experience.
The ideal candidate will be responsible for designing and implementing personalized marketing journeys, automation workflows, and multi-channel campaigns using Salesforce Marketing Cloud (SFMC).
This role requires strong expertise in APIs, scripting, and data handling to build scalable, personalized customer experiences.
Must-Have Skills :
- Salesforce Marketing Cloud (Automation Studio, Journey Builder, Email Studio, Mobile Studio)
- API Integration
- Strong SQL skills
- AMPscript & JavaScript
Good-to-Have Skills :
- HTML & CSS
- Interaction Studio / Marketing Cloud Personalization
Hands-on experience in leading R2R function for US based Company, Restaurant Accounting. · Should have experience in R365, US GAAP accounting and Financial Statement preparation. · Should be able to lead client meetings, Business review calls. · Work with US counterparts in driving key process initiatives. · Manage and publish daily, weekly, and monthly performance scorecards. · Manage and own the process SLAs agree with client. · Able to interpret Financial Statements to help Executive team to make decisions. · Conduct, monthly, quarterly, and annual one-on-one with team and perform the year end appraisal and performance management. · Should be an Individual contributor. · Managing client experience. · Team management skills. · Subject Matter Expert.
B Com Graduate with 7-8 years of work experience having 3-5 years of team and client management experience. • Strong analytical skills and problem-solving skills. • Proactive, takes initiative, self-motivated, team player. • Strong stakeholder management and interpersonal skills. • Extensive understanding of financial trends both within the company and general market patterns. • Business acumen, Analytical approach, understanding of general business development and operations. • Prior experience in similar BPO/Shared Service Function of MNC. • Should have experience in US Accounting. • Should have excellent communication skills and client interaction experience. • Maintain general ledger, post transactions in R365. • Strong willingness to learn and grow. • Should be willing to work in US Shifts. • Can join immediately. • Should work from Office.
- Good understanding of NODE JS back-end development
- Strong understanding of Node.js and frameworks available for it
- Understanding of front-end technologies, such as React/Angular, HTML and CSS3
-Solid understanding of object-oriented development and design patterns.
- Good experience with AJAX, REST and JSON
- Experience with development practices such as Test-Driven Development and Continuous Integration
-Experience in Version Control Systems like GIT, Bitbucket
- Must have good problem solving and analytical skills.
- Must be a team player, having good communication skills, self-starter
- Ability to pick up new technology is needed.
Job description
Company Description
Honeycomb Creative Support is an integrated marketing communication agency, with
nearly 14 years of experience marketing in diverse sectors such as lifestyles, IT,
pharmaceuticals, healthcare, education, manufacturing, and automotive. The firm
provides creative design and marketing services to local and international clients
using creative vision, information analysis, strategic thinking, and execution to sell
ideas and unique products to the target audience. Honeycomb provides services such
as Web Design & Development, Graphic Design, Corporate and Product Video
Production, Digital Marketing, Branding, Fine-Art Printing and On-site Creative
Studio at Client Location.
Role Description
If you are looking for an exciting opportunity to work with an established marketing
communications agency, you have come to the right place. We are looking for a full-
time On-Site Back End developer Laravel | Exp: 2 to 3 Years | 3 - 5 LPA, based in
Greater Bengaluru Area. As a Back End Engineer at Honeycomb Creative Support,
you will be a part of our development team, responsible for building server
applications and APIs.
Qualifications
• PHP and PHP Frameworks
• Laravel and CodeIgniter must
• Good Knowledge in any MVC Frameworks.
• Basic knowledge of AWS
• MySQL
• Basic knowledge in Node.js
• Experience with building RESTful APIs and API development
• Knowledge of GIT-based source control
• Understanding of how to test and debug web applications
• Understanding of modern PHP development practices, design patterns, and SOLID principles
• Bachelor's or Master's degree in Computer Science or equivalent
• Good problem-solving skills and attention to detail.
XressBees – a logistics company started in 2015 – is amongst the fastest growing companies of its sector. Our
vision to evolve into a strong full-service logistics organization reflects itself in the various lines of business like B2C
logistics 3PL, B2B Xpress, Hyperlocal and Cross border Logistics.
Our strong domain expertise and constant focus on innovation has helped us rapidly evolve as the most trusted
logistics partner of India. XB has progressively carved our way towards best-in-class technology platforms, an
extensive logistics network reach, and a seamless last mile management system.
While on this aggressive growth path, we seek to become the one-stop-shop for end-to-end logistics solutions. Our
big focus areas for the very near future include strengthening our presence as service providers of choice and
leveraging the power of technology to drive supply chain efficiencies.
Job Overview
XpressBees would enrich and scale its end-to-end logistics solutions at a high pace. This is a great opportunity to join
the team working on forming and delivering the operational strategy behind Artificial Intelligence / Machine Learning
and Data Engineering, leading projects and teams of AI Engineers collaborating with Data Scientists. In your role, you
will build high performance AI/ML solutions using groundbreaking AI/ML and BigData technologies. You will need to
understand business requirements and convert them to a solvable data science problem statement. You will be
involved in end to end AI/ML projects, starting from smaller scale POCs all the way to full scale ML pipelines in
production.
Seasoned AI/ML Engineers would own the implementation and productionzation of cutting-edge AI driven algorithmic
components for search, recommendation and insights to improve the efficiencies of the logistics supply chain and
serve the customer better.
You will apply innovative ML tools and concepts to deliver value to our teams and customers and make an impact to
the organization while solving challenging problems in the areas of AI, ML , Data Analytics and Computer Science.
Opportunities for application:
- Route Optimization
- Address / Geo-Coding Engine
- Anomaly detection, Computer Vision (e.g. loading / unloading)
- Fraud Detection (fake delivery attempts)
- Promise Recommendation Engine etc.
- Customer & Tech support solutions, e.g. chat bots.
- Breach detection / prediction
An Artificial Intelligence Engineer would apply himself/herself in the areas of -
- Deep Learning, NLP, Reinforcement Learning
- Machine Learning - Logistic Regression, Decision Trees, Random Forests, XGBoost, etc..
- Driving Optimization via LPs, MILPs, Stochastic Programs, and MDPs
- Operations Research, Supply Chain Optimization, and Data Analytics/Visualization
- Computer Vision and OCR technologies
The AI Engineering team enables internal teams to add AI capabilities to their Apps and Workflows easily via APIs
without needing to build AI expertise in each team – Decision Support, NLP, Computer Vision, for Public Clouds and
Enterprise in NLU, Vision and Conversational AI.Candidate is adept at working with large data sets to find
opportunities for product and process optimization and using models to test the effectiveness of different courses of
action. They must have knowledge using a variety of data mining/data analysis methods, using a variety of data tools,
building, and implementing models, using/creating algorithms, and creating/running simulations. They must be
comfortable working with a wide range of stakeholders and functional teams. The right candidate will have a passion
for discovering solutions hidden in large data sets and working with stakeholders to improve business outcomes.
Roles & Responsibilities
● Develop scalable infrastructure, including microservices and backend, that automates training and
deployment of ML models.
● Building cloud services in Decision Support (Anomaly Detection, Time series forecasting, Fraud detection,
Risk prevention, Predictive analytics), computer vision, natural language processing (NLP) and speech that
work out of the box.
● Brainstorm and Design various POCs using ML/DL/NLP solutions for new or existing enterprise problems.
● Work with fellow data scientists/SW engineers to build out other parts of the infrastructure, effectively
communicating your needs and understanding theirs and address external and internal shareholder's
product challenges.
● Build core of Artificial Intelligence and AI Services such as Decision Support, Vision, Speech, Text, NLP, NLU,
and others.
● Leverage Cloud technology –AWS, GCP, Azure
● Experiment with ML models in Python using machine learning libraries (Pytorch, Tensorflow), Big Data,
Hadoop, HBase, Spark, etc
● Work with stakeholders throughout the organization to identify opportunities for leveraging company data to
drive business solutions.
● Mine and analyze data from company databases to drive optimization and improvement of product
development, marketing techniques and business strategies.
● Assess the effectiveness and accuracy of new data sources and data gathering techniques.
● Develop custom data models and algorithms to apply to data sets.
● Use predictive modeling to increase and optimize customer experiences, supply chain metric and other
business outcomes.
● Develop company A/B testing framework and test model quality.
● Coordinate with different functional teams to implement models and monitor outcomes.
● Develop processes and tools to monitor and analyze model performance and data accuracy.
● Develop scalable infrastructure, including microservices and backend, that automates training and
deployment of ML models.
● Brainstorm and Design various POCs using ML/DL/NLP solutions for new or existing enterprise problems.
● Work with fellow data scientists/SW engineers to build out other parts of the infrastructure, effectively
communicating your needs and understanding theirs and address external and internal shareholder's
product challenges.
● Deliver machine learning and data science projects with data science techniques and associated libraries
such as AI/ ML or equivalent NLP (Natural Language Processing) packages. Such techniques include a good
to phenomenal understanding of statistical models, probabilistic algorithms, classification, clustering, deep
learning or related approaches as it applies to financial applications.
● The role will encourage you to learn a wide array of capabilities, toolsets and architectural patterns for
successful delivery.
What is required of you?
You will get an opportunity to build and operate a suite of massive scale, integrated data/ML platforms in a broadly
distributed, multi-tenant cloud environment.
● B.S., M.S., or Ph.D. in Computer Science, Computer Engineering
● Coding knowledge and experience with several languages: C, C++, Java,JavaScript, etc.
● Experience with building high-performance, resilient, scalable, and well-engineered systems
● Experience in CI/CD and development best practices, instrumentation, logging systems
● Experience using statistical computer languages (R, Python, SLQ, etc.) to manipulate data and draw insights
from large data sets.
● Experience working with and creating data architectures.
● Good understanding of various machine learning and natural language processing technologies, such as
classification, information retrieval, clustering, knowledge graph, semi-supervised learning and ranking.
● Knowledge and experience in statistical and data mining techniques: GLM/Regression, Random Forest,
Boosting, Trees, text mining, social network analysis, etc.
● Knowledge on using web services: Redshift, S3, Spark, Digital Ocean, etc.
● Knowledge on creating and using advanced machine learning algorithms and statistics: regression,
simulation, scenario analysis, modeling, clustering, decision trees, neural networks, etc.
● Knowledge on analyzing data from 3rd party providers: Google Analytics, Site Catalyst, Core metrics,
AdWords, Crimson Hexagon, Facebook Insights, etc.
● Knowledge on distributed data/computing tools: Map/Reduce, Hadoop, Hive, Spark, MySQL, Kafka etc.
● Knowledge on visualizing/presenting data for stakeholders using: Quicksight, Periscope, Business Objects,
D3, ggplot, Tableau etc.
● Knowledge of a variety of machine learning techniques (clustering, decision tree learning, artificial neural
networks, etc.) and their real-world advantages/drawbacks.
● Knowledge of advanced statistical techniques and concepts (regression, properties of distributions,
statistical tests, and proper usage, etc.) and experience with applications.
● Experience building data pipelines that prep data for Machine learning and complete feedback loops.
● Knowledge of Machine Learning lifecycle and experience working with data scientists
● Experience with Relational databases and NoSQL databases
● Experience with workflow scheduling / orchestration such as Airflow or Oozie
● Working knowledge of current techniques and approaches in machine learning and statistical or
mathematical models
● Strong Data Engineering & ETL skills to build scalable data pipelines. Exposure to data streaming stack (e.g.
Kafka)
● Relevant experience in fine tuning and optimizing ML (especially Deep Learning) models to bring down
serving latency.
● Exposure to ML model productionzation stack (e.g. MLFlow, Docker)
● Excellent exploratory data analysis skills to slice & dice data at scale using SQL in Redshift/BigQuery.
1. To handle customers on their enquiries/complaints via multiple channels i.e. call/mail
2. To understand the product features for better problem solving
3. To keep records of customer interactions, transactions, comments, complaints
4. To coordinate with internal teams for solving any customer issues
5. To ensure customer satisfaction and provide empathetic experience to every user
6. To own every interaction with the customer and ensure every customer is a happy customer
Requirements-
1. Graduation is mandatory
2. 1+ years of experience in customer service function (e-com / hospitality industry preferred)
3. Excellent verbal and oral comms in English preferred





