MLOps Lead Engineer
at IT solutions specialized in Apps Lifecycle management. (MG1)
- Automate and maintain ML and Data pipelines at scale
- Collaborate with Data Scientists and Data Engineers on feature development teams to containerize and build out deployment pipelines for new modules
- Maintain and expand our on-prem deployments with spark clusters
- Design, build and optimize applications containerization and orchestration with Docker and Kubernetes and AWS or Azure
- 5 years of IT experience in data-driven or AI technology products
- Understanding of ML Model Deployment and Lifecycle
- Extensive experience in Apache airflow for MLOps workflow automation
- Experience is building and automating data pipelines
- Experience in working on Spark Cluster architecture
- Extensive experience with Unix/Linux environments
- Experience with standard concepts and technologies used in CI/CD build, deployment pipelines using Jenkins
- Strong experience in Python and PySpark and building required automation (using standard technologies such as Docker, Jenkins, and Ansible).
- Experience with Kubernetes or Docker Swarm
- Working technical knowledge of current systems software, protocols, and standards, including firewalls, Active Directory, etc.
- Basic knowledge of Multi-tier architectures: load balancers, caching, web servers, application servers, and databases.
- Experience with various virtualization technologies and multi-tenant, private and hybrid cloud environments.
- Hands-on software and hardware troubleshooting experience.
- Experience documenting and maintaining configuration and process information.
- Basic Knowledge of machine learning frameworks: Tensorflow, Caffe/Caffe2, Pytorch
Similar jobs
DevOps Engineer/ Tech Lead/ Architect-Ahmedabad
at Intuitive Technology Partners
Intuitive is the fastest growing top-tier Cloud Solutions and Services company supporting Global Enterprise Customer across Americas, Europe and Middle East.
Intuitive is looking for highly talented hands-on Cloud Infrastructure Architects to help accelerate our growing Professional Services consulting Cloud & DevOps practice. This is an excellent opportunity to join Intuitive’ s global world class technology teams, working with some of the best and brightest engineers while also developing your skills and furthering your career working with some of the largest customers.
Excellent communication skills
Open to work on EST time zone(6pm to 3am)
- Integrate gates into CI/CD pipeline and push all flaws/issues to developers IDE (as far left as possible) - ideally in code repo but required by the time code is in the artifact repository.
- Demonstrable experience in Containerization-Docker and orchestration (Kubernetes)
- Experience withsetting up self-managed Kubernetes clusters without using any managed cloud offerings like EKS
- Experience working withAWS - Managing AWS services - EC2, S3, Cloudfront, VPC, SNS, Lambda, AWS Autoscaling, AWS IAM, RDS, EBS, Kinesis, SQS, DynamoDB, Elastic Cache, Redshift, Cloudwatch, Amazon Inspector.
- Familiarity withLinux and UNIX systems (e.g. CentOS, RedHat) and command line system administration such as Bash, VIM, SSH.
- Hands on experience in configuration management of server farms (using tools such asPuppet, Chef, Ansible, etc.,).
- Demonstrated understanding of ITIL methodologies, ITIL v3 or v4 certification
- Kubernetes CKA or CKAD certification nice to have
Site Reliability Engineer
at A startup company providing AI based software platforms
Who You Are
- Creative thinker and strong problem solver with meticulous attention to detail
- Highly organized, creative, motivated, and passionate about achieving results
- Able to balance multiple tasks and projects effectively and quickly adapt to new situations and technologies
- Able to work both independently and as part of a team
- Systematic problem-solver, coupled with a strong sense of ownership and drive
What you need
- 3-7 years of experience as a Site Reliability Engineer or a mix of a software engineer and DevOps.
- Strong hands-on knowledge of Linux fundamentals, System administration scripting, performance tuning/scalability, troubleshooting.
- Write great quality code using SOLID principles including unit and integration tests.
- Hands-on development experience in an object-orientated programming language like Python.
- Hands-on experience developing task automations
- Experience using tools to create and manage CI (continuous integration) and CD (continuous delivery) pipelines.
- Familiarity with software development tools: source code management (SCM systems), code review systems, issue tracking tools, build tools, test frameworks, code quality tools.
- Experience implementing open-source observability and alerting tools, like Prometheus, Grafana, Cortex, Thanos, Alertmanager etc
- Have decent knowledge on networking (VPC, VNet, DNS etc) and of the TCP/IP stack, internet routing and load balancing.
- Worked with log and configuration management tool
- Prior experience of working with AWS, Azure, GCP is a plus
- Prior experience of working with Kubernetes, Docker and containers is plus
- Strong interpersonal communication skills (including listening, speaking, and writing) and ability to work well in a diverse, team-focused environment with other SREs, Engineers, Product Managers, etc.
- Documenting your work should be in your DNA
What you get
- A chance to develop and build something (probably from scratch) which you can be proud of
- Build and Implement modern systems observability solutions including monitoring, alerting, metrics, logging, and APM & distributed tracing.
- Scale systems sustainably through automation and evolve systems by pushing for changes that improve reliability and velocity.
- Maintain business continuity by identifying and driving opportunities to make systems highly resilient and human-free.
- Closely work with the software engineering team to ensure accurate monitoring and metrics are being built into applications before going to production.
- Develop and maintain software modules for use and re-use in cloud and on-premise systems automation.
- Identify process gaps and implement process improvements to increase operational reliability
- Drive standardization efforts across the services, infrastructure, systems, and practices
- Develop Systems & Tools to help with Development team to uphold the Reliability principles
- Minimum 4 Years of experience
- Ansible with 3+ years hands on
- Any 1 container – Preferably Kubernetes
- Any 1 Cloud | Preferably Azure
- Overall, 3 to 5 years’ experience
Job Brief:
We are looking for candidates that have experience in development and have performed CI/CD based projects. Should have a good hands-on Jenkins Master-Slave architecture, used AWS native services like CodeCommit, CodeBuild, CodeDeploy and CodePipeline. Should have experience in setting up cross platform CI/CD pipelines which can be across different cloud platforms or on-premise and cloud platform.
Job Location:
Pune.
Job Description:
- Hands on with AWS (Amazon Web Services) Cloud with DevOps services and CloudFormation.
- Experience interacting with customer.
- Excellent communication.
- Hands-on in creating and managing Jenkins job, Groovy scripting.
- Experience in setting up Cloud Agnostic and Cloud Native CI/CD Pipelines.
- Experience in Maven.
- Experience in scripting languages like Bash, Powershell, Python.
- Experience in automation tools like Terraform, Ansible, Chef, Puppet.
- Excellent troubleshooting skills.
- Experience in Docker and Kuberneties with creating docker files.
- Hands on with version control systems like GitHub, Gitlab, TFS, BitBucket, etc.
Assistant Manager-Business Analaytics
Job Title:- Assistant Manager - Business Analytics
Age: Max. 35years.
Working Days:- 6 days a week
Location:- Ahmedabad, Gujarat
Monthly CTC:- Salary will commensurate with experience.
Educational Qualification:- The candidate should have bachelor’s degree in
IT/Engineering from any recognized university..
Experience:- 2+ years of work experience in AI/ML/business analytics with Institute of
repute or corporate.
Required Technical Skills:-
A fair bit of understanding of Business Analytics, Data Science, Visualization/Big Data
etc.
Basic knowledge of different analytical tools such as R Programming, Python etc.
Hands on experience in Moodle Development (desirable).
Good knowledge in customizing Moodle functionalities and developing custom themes
for Moodle (desirable).
An analytical mind-set who enjoy helping participants solving problems and turning data
into useful actionable information
Key Responsibilities include:-
Understand the tools and technologies specific to e-learning and blended learning
development and delivery.
Provide academic as well as technical assistance to the faculty members teaching the
analytics courses.
Working closely with the Instructors and assisting them in programming, coding, testing
etc.
Preparing the lab study material in coordination with the Instructors and assisting
students in programming lab and solving their doubts.
Works on assignments dealing with the routine and daily operation, use, and
configuration of the Learning Management System (LMS).
Administers learning technology platforms including the creation of courses,
certifications and other e-learning programs on the platforms.
Responsible to provide support within the eLearning department, provide technical
support to our external clients, and administrate the Learning Management System.
Creates user groups and assigns content and assessments to the right target audience,
runs reports and creates learning evens in the LMS system.
Performs regular maintenance of LMS database, including adding or removing courses.
Uploads, tests, deploys and maintains all training materials/learning assets hosted in the
LMS.
Ability to Multi-task.
Ability to demonstrate accuracy on detailed oriented and repetitive job assignments.
Responsible and reliable
Position: DevOps Engineer
Job Description
The candidate should have the following Skills:
- Hands-on experience with DevOps & CICD open source tools (Jenkins, ), including AWS DevOps services (CodePipeline, CloudFormation, etc).
- Experience in building and deploying using Java/Python/Node.js on Cloud infrastructure (Docker or Kubernetes containers or Lambda.)
- Exposure to Cloud operations, releases, and configuration management
- Experience in implementing Non-functional requirements for microservices, including performance, security, compliance, HA and Disaster Recovery.
- Good soft skills, great attitude, and passion for working in a product startup environment
Total Experience of 2-5 years post BE or BTech or MCA in Computer Science Engineering.
We are hiring an experienced DevOps Engineer, who can take end - to - end ownership, enjoys working on products and work in a fast-moving environment. This is Hyderabad based role.
Responsibilities
- Containerize existing workloads
- Migrate Docker Compose setups to Kubernetes
- Take complete ownership over the CI/CD pipeline which includes developing, maintaining and monitoring
- Administer in-house datacentre's Linux Servers and the occasional cloud workloads
- Work with the team to setup Development, Production, and Test environments
Skills Required
- Hands-on experience in working with Kubernetes, (Kubernetes with GPUs is a plus)
- Excellent understanding of Docker, Docker Compose, and containers
- Hands-on knowledge of Unix/Linux
- Experience working with Python, and NodeJS (needed to dockerize workloads)
- Experience writing build scripts using Python, Unix Shell (bash,ksh)
- Understanding of networking with Linux
- Experience with AWS and/or GCP
- Experience with continuous delivery and integration tools: Jenkins, Github
- Understanding of version control systems
Good to have
- Previous experience in handling self-hosted/on-prem setups
- Deep programming experience with Python and NodeJS
- Previous experience in working with GPU-powered containers on Kubernetes
- Understanding of the ML/DL pipeline
- Have 3+ years of experience in Python development
- Be familiar with common database access patterns
- Have experience with designing systems and monitoring metrics, looking at graphs.
- Have knowledge of AWS, Kubernetes and Docker.
- Be able to work well in a remote development environment.
- Be able to communicate in English at a native speaking and writing level.
- Be responsible to your fellow remote team members.
- Be highly communicative and go out of your way to contribute to the team and help others