17+ Amazon S3 Jobs in Mumbai | Amazon S3 Job openings in Mumbai
Apply to 17+ Amazon S3 Jobs in Mumbai on CutShort.io. Explore the latest Amazon S3 Job opportunities across top companies like Google, Amazon & Adobe.
Technical Skills:
- Ability to understand and translate business requirements into design.
- Proficient in AWS infrastructure components such as S3, IAM, VPC, EC2, and Redshift.
- Experience in creating ETL jobs using Python/PySpark.
- Proficiency in creating AWS Lambda functions for event-based jobs.
- Knowledge of automating ETL processes using AWS Step Functions.
- Competence in building data warehouses and loading data into them.
Responsibilities:
- Understand business requirements and translate them into design.
- Assess AWS infrastructure needs for development work.
- Develop ETL jobs using Python/PySpark to meet requirements.
- Implement AWS Lambda for event-based tasks.
- Automate ETL processes using AWS Step Functions.
- Build data warehouses and manage data loading.
- Engage with customers and stakeholders to articulate the benefits of proposed solutions and frameworks.
- Python knowledge: object-oriented programming: inheritance, abstract classes, dataclass, dependency injection, design patterns: comand-query, repository, adapter, hexagonal architecture, swagger/Open API, flask, connexion
- Experience on AWS services: lambda, ecs, sqs, s3, dynamodb, auroradb
- Experience with following libraries boto3, behave, pytest, moto, localstack, docker
- Basic knowledge about terraform, gitlab ci
- Experience with SQL DB
Python Developer
6-8 Years
Mumbai
N.p only immediate or who is serving LwD is 1st week of july.
- Python knowledge: object-oriented programming: inheritance, abstract classes, dataclass, dependency injection, design patterns: comand-query, repository, adapter, hexagonal architecture, swagger/Open API, flask, connexion
- Experience on AWS services: lambda, ecs, sqs, s3, dynamodb, auroradb
- Experience with following libraries boto3, behave, pytest, moto, localstack, docker
- Basic knowledge about terraform, gitlab ci
- Experience with SQL DB
Role : Principal Devops Engineer
About the Client
It is a Product base company that has to build a platform using AI and ML technology for their transportation and logiticsThey also have a presence in the global market
Responsibilities and Requirements
• Experience in designing and maintaining high volume and scalable micro-services architecture on cloud infrastructure
• Knowledge in Linux/Unix Administration and Python/Shell Scripting
• Experience working with cloud platforms like AWS (EC2, ELB, S3, Auto-scaling, VPC, Lambda), GCP, Azure
• Knowledge in deployment automation, Continuous Integration and Continuous Deployment (Jenkins, Maven, Puppet, Chef, GitLab) and monitoring tools like Zabbix, Cloud Watch Monitoring, Nagios
• Knowledge of Java Virtual Machines, Apache Tomcat, Nginx, Apache Kafka, Microservices architecture, Caching mechanisms
• Experience in enterprise application development, maintenance and operations
• Knowledge of best practices and IT operations in an always-up, always-available service
• Excellent written and oral communication skills, judgment and decision-making skill
Roles & Responsibilities:
- Bachelor’s degree in Computer Science, Information Technology or a related field
- Experience in designing and maintaining high volume and scalable micro-services architecture on cloud infrastructure
- Knowledge in Linux/Unix Administration and Python/Shell Scripting
- Experience working with cloud platforms like AWS (EC2, ELB, S3, Auto-scaling, VPC, Lambda), GCP, Azure
- Knowledge in deployment automation, Continuous Integration and Continuous Deployment (Jenkins, Maven, Puppet, Chef, GitLab) and monitoring tools like Zabbix, Cloud Watch Monitoring, Nagios Knowledge of Java Virtual Machines, Apache Tomcat, Nginx, Apache Kafka, Microservices architecture, Caching mechanisms
- Experience in enterprise application development, maintenance and operations
- Knowledge of best practices and IT operations in an always-up, always-available service
- Excellent written and oral communication skills, judgment and decision-making skills
DATA ENGINEER
Overview
They started with a singular belief - what is beautiful cannot and should not be defined in marketing meetings. It's defined by the regular people like us, our sisters, our next-door neighbours, and the friends we make on the playground and in lecture halls. That's why we stand for people-proving everything we do. From the inception of a product idea to testing the final formulations before launch, our consumers are a part of each and every process. They guide and inspire us by sharing their stories with us. They tell us not only about the product they need and the skincare issues they face but also the tales of their struggles, dreams and triumphs. Skincare goes deeper than skin. It's a form of self-care for many. Wherever someone is on this journey, we want to cheer them on through the products we make, the content we create and the conversations we have. What we wish to build is more than a brand. We want to build a community that grows and glows together - cheering each other on, sharing knowledge, and ensuring people always have access to skincare that really works.
Job Description:
We are seeking a skilled and motivated Data Engineer to join our team. As a Data Engineer, you will be responsible for designing, developing, and maintaining the data infrastructure and systems that enable efficient data collection, storage, processing, and analysis. You will collaborate with cross-functional teams, including data scientists, analysts, and software engineers, to implement data pipelines and ensure the availability, reliability, and scalability of our data platform.
Responsibilities:
Design and implement scalable and robust data pipelines to collect, process, and store data from various sources.
Develop and maintain data warehouse and ETL (Extract, Transform, Load) processes for data integration and transformation.
Optimize and tune the performance of data systems to ensure efficient data processing and analysis.
Collaborate with data scientists and analysts to understand data requirements and implement solutions for data modeling and analysis.
Identify and resolve data quality issues, ensuring data accuracy, consistency, and completeness.
Implement and maintain data governance and security measures to protect sensitive data.
Monitor and troubleshoot data infrastructure, perform root cause analysis, and implement necessary fixes.
Stay up-to-date with emerging technologies and industry trends in data engineering and recommend their adoption when appropriate.
Qualifications:
Bachelor’s or higher degree in Computer Science, Information Systems, or a related field.
Proven experience as a Data Engineer or similar role, working with large-scale data processing and storage systems.
Strong programming skills in languages such as Python, Java, or Scala.
Experience with big data technologies and frameworks like Hadoop, Spark, or Kafka.
Proficiency in SQL and database management systems (e.g., MySQL, PostgreSQL, or Oracle).
Familiarity with cloud platforms like AWS, Azure, or GCP, and their data services (e.g., S3, Redshift, BigQuery).
Solid understanding of data modeling, data warehousing, and ETL principles.
Knowledge of data integration techniques and tools (e.g., Apache Nifi, Talend, or Informatica).
Strong problem-solving and analytical skills, with the ability to handle complex data challenges.
Excellent communication and collaboration skills to work effectively in a team environment.
Preferred Qualifications:
Advanced knowledge of distributed computing and parallel processing.
Experience with real-time data processing and streaming technologies (e.g., Apache Kafka, Apache Flink).
Familiarity with machine learning concepts and frameworks (e.g., TensorFlow, PyTorch).
Knowledge of containerization and orchestration technologies (e.g., Docker, Kubernetes).
Experience with data visualization and reporting tools (e.g., Tableau, Power BI).
Certification in relevant technologies or data engineering disciplines.
Responsibilities
- Implement various development, testing, automation tools, and IT infrastructure
- Design, build and automate the AWS infrastructure (VPC, EC2, Networking, EMR, RDS, S3, ALB, Cloud Front, etc.) using Terraform
- Manage end-to-end production workloads hosted on Docker and AWS
- Automate CI pipeline using Groovy DSL
- Deploy and configure Kubernetes clusters (EKS)
- Design and build a CI/CD Pipeline to deploy applications using Jenkins and Docker
Eligibility
- At least 8 years of proven experience in AWS-based DevOps/cloud engineering and implementations
- Expertise in all common AWS Cloud services like EC2, EKS, S3, VPC, Lambda, API Gateway, ALB, Redis, etc.
- Experience in deploying and managing production environments in Amazon AWS
- Strong experience in continuous integration and continuous deployment
- Knowledge of application build, deployment, and configuration using one of the tools: Jenkins
Job DescriptionPosition: Sr Data Engineer – Databricks & AWS
Experience: 4 - 5 Years
Company Profile:
Exponentia.ai is an AI tech organization with a presence across India, Singapore, the Middle East, and the UK. We are an innovative and disruptive organization, working on cutting-edge technology to help our clients transform into the enterprises of the future. We provide artificial intelligence-based products/platforms capable of automated cognitive decision-making to improve productivity, quality, and economics of the underlying business processes. Currently, we are transforming ourselves and rapidly expanding our business.
Exponentia.ai has developed long-term relationships with world-class clients such as PayPal, PayU, SBI Group, HDFC Life, Kotak Securities, Wockhardt and Adani Group amongst others.
One of the top partners of Cloudera (leading analytics player) and Qlik (leader in BI technologies), Exponentia.ai has recently been awarded the ‘Innovation Partner Award’ by Qlik in 2017.
Get to know more about us on our website: http://www.exponentia.ai/ and Life @Exponentia.
Role Overview:
· A Data Engineer understands the client requirements and develops and delivers the data engineering solutions as per the scope.
· The role requires good skills in the development of solutions using various services required for data architecture on Databricks Delta Lake, streaming, AWS, ETL Development, and data modeling.
Job Responsibilities
• Design of data solutions on Databricks including delta lake, data warehouse, data marts and other data solutions to support the analytics needs of the organization.
• Apply best practices during design in data modeling (logical, physical) and ETL pipelines (streaming and batch) using cloud-based services.
• Design, develop and manage the pipelining (collection, storage, access), data engineering (data quality, ETL, Data Modelling) and understanding (documentation, exploration) of the data.
• Interact with stakeholders regarding data landscape understanding, conducting discovery exercises, developing proof of concepts and demonstrating it to stakeholders.
Technical Skills
• Has more than 2 Years of experience in developing data lakes, and datamarts on the Databricks platform.
• Proven skill sets in AWS Data Lake services such as - AWS Glue, S3, Lambda, SNS, IAM, and skills in Spark, Python, and SQL.
• Experience in Pentaho
• Good understanding of developing a data warehouse, data marts etc.
• Has a good understanding of system architectures, and design patterns and should be able to design and develop applications using these principles.
Personality Traits
• Good collaboration and communication skills
• Excellent problem-solving skills to be able to structure the right analytical solutions.
• Strong sense of teamwork, ownership, and accountability
• Analytical and conceptual thinking
• Ability to work in a fast-paced environment with tight schedules.
• Good presentation skills with the ability to convey complex ideas to peers and management.
Education:
BE / ME / MS/MCA.
JD / Skills Sets
1. Good knowledge on Python
2. Good knowledge on My-Sql, mongodb
3. Design Pattern
4. OOPs
5. Automation
6. Web scraping
7. Redis queue
8. Basic idea of Finance Domain will be beneficial.
9. Git10. AWS (EC2, RDS, S3)
Product based startup to help DTC brands succeed!
Responsibilities
- Develop new user-facing features using React.js and RESTful APIs using Node.js and MongoDB
- Build reusable code and libraries for future use
- Optimize applications for maximum speed and scalability
- Collaborate with team members, e.g designer, product and other stakeholders to ensure quality in product.
- Ensure the technical feasibility of UI/UX designs
- Manage and maintain cloud infrastructure on AWS
Qualifications
- At least 3-6 years of experience in MERN stack
- Proficiency with React.js, Node.js, MongoDB, and Express.js
- Familiarity with AWS services such as EC2, S3, and RDS, SQS, Lambda
- Understanding of RESTful API design principles
- Understanding of Agile software development methodologies
- Strong problem-solving and analytical skills
Hello,
Greetings from CodersBrain!
Coders Brain is a global leader in its services, digital, and business solutions that partners with its clients to simplify, strengthen, and transform their businesses. We ensure the highest levels of certainty and satisfaction through a deep-set commitment to our clients, comprehensive industry expertise, and a global network of innovation and delivery centers.
This is regarding the urgent opening for the'' ROR Developer role. We found your profile in the Cutshort database and it seems like a good fit for the organization. If you are interested, do revert back with your updated CV along with the details:
Permanent Payroll : CodersBrain
Location:- Mumbai/ Kolkata
Notice Period:- imm./15 daysJob Description
Scope of work This role is an exciting opportunity to build highly interactive
customer facing applications and products.
Candidate will help transform vast collections of data into
actionable insights with intuitive and easy to use interfaces and
visualizations.
Candidate will leverage the power of JavaScript and Ruby on
Rail to build novel features and improvements to our current suite
of tools.
This team is responsible for customer-facing applications that
deliver SEO data insights.
Through the applications customers are able to access insights,
workflows, and aggregations of information above and beyond
core data offerings
Responsibility
Build and maintain the core frontend application
Work collaboratively with the engineers on the Frontend team to
ensure quality and performance of the systems through code
reviews, documentation, analysis and employing engineering
best practices to ensure high-quality software
Contribute to org devops culture by maintaining our systems,
including creating documentation, run books, monitoring, alerting,
and integration tests, etc.
Participate in architecture design and development for new
features and capabilities, and for migration of legacy systems, to
meet business and customer needs.
Take turns in the on-call rotation, handling systems and
operations issues as they arise including responding to off-hours
alerts
Collaborate with other teams on dependent work and integrations
as well as be vigilant for activities happening outside the team
that would have an impact on the work your team is doing.
Work with Product Managers and UX Designers to deliver new
features and capabilities
Use good security practices to protect code and systems
Pitch in where needed during major efforts or when critical issues
arise
Give constructive, critical feedback to other team members
through pull requests, design reviews, and other methods
Seek out opportunities and work to grow skills and expertise.
C) Skills Required
Essential Skills JavaScript (preferred ExtJS framework)
Ruby
Ruby on Rails
MySQL
Docker
Desired Skills Terraform
Basic Unix/Linux administration
Redis
Resque
TravisCI
AWS ECS
AWS RDS
AWS EMR
AWS S3
AWS Step Functions
AWS Lambda Functions
Experience working remotely with a distributed team
Great problem-solving skills
D) Other Information
Educational Qualifications Bachelor's degree/MCA
Experience 5– 8 years
Please confirm the mail with your updated CV if you are interested in this position and also please share the below-mentioned details :
Current CTC:
Expected CTC:
Current Company:
Notice Period:
Notice Period:Are you okay with 1 week Notice period if not then comfortable as freelancing work with us till joining:-
Current Location:
Preferred Location:
Total-experience:
Relevant experience:
Highest qualification:
DOJ(If Offer in Hand from Other company):
offer in hand:
Alternate number:
Interview Availability
Digital Marketing Company
Urgently looking for our client company as Senior Backend Developer in Mumbai
Apply with updated CV
Designation-Sr Backend Developer
Exp-Min 2 Years
Sal-Nego
Location-Mumbai
Technical Skills
• 2+ years of hand on experience in a related field
• Proficiency in Python with hands-on experience in developing APIs with Flask or Django
• Comfortable with Linux development environment
• Familiarity with frontend platforms such as HTML/CSS would be a plus
• Experience with: MySQL, Postgres, AWS Services (Redshift, EC2 & S3)
With a global provider of Business Process Management.
Full Stack senior developer with ReactJS & Redux
Qualifications:
3 - 4 years of experience in application development and maintenance
Bachelor’s Degree/B.Tech/B.E in Computer Science or a related discipline
Advanced knowledge of HTML and CSS
Experience with using and building automated tests
Experience with CI/CD tooling (at least Git and Jenkins)
Experience with the Agile/Scrum/Kanban way of working
Proficient understanding of code versioning tools, such as Git
Experience with a modern frontend framework (e.g. Polymer, Redux, React, Angular, Vue.js or other component-based frameworks)
Experience with web paradigms such as REST, Responsive Web Design, Test-driven Development (TDD), Dependency Injection, etc.
Experience with common front-end development tools such as Babel, Webpack, NPM, etc.
Knowledge of performance testing frameworks including Mocha and Jest.
Knowledge of RESTful APIs.
Understanding of AWS, S3, API Gateway, DynamoDB, RDS, Lambda.
QuickSell is a mobile commerce platform.
- 2-5 years of experience in building API services using NodeJS Express and related
frameworks
- Expert level understanding of NodeJS asynchronous runtime
- Expert level understanding of Javascript concepts on callbacks and closures
- Experience with Postgres, NoSQL, Redis, and Firebase real-time database
- Experience with AWS services like Elastic Beanstalk, Cloudfront, S3, EC2, Lambda,
API Gateway, SQS, etc
- Understanding of patterns and techniques for building scalable back-end
infrastructure including caching, rate limiting, authentication, and authorization
schemes.
- Experience in building highly scalable and high throughput services with
millisecond response times
- Experience working in a collaborative team environment
- Excellent communication & interpersonal skills
- Willingness to learn and pick up new technology along with patience to mentor
Bonus skills -
- Experience with ElasticSearch, Puppeteer
- Experience writing unit tests
at VAYUZ Technologies
Roles and Responsibilities
1. Ability to work on diverse backend stack such as Node JS, Java, Express JS
2. Ability to work on diverse frontend stack such as React JS, Angular 6/7/8/9, HTML5, CSS3
3. Ability to deliver quick POC’s using cutting edge technologies.
4. Preparing reports, manuals and other documentation on the status, operation and maintenance of software.
5. Design, develop, and unit test applications in accordance with established standards
6. Developing, refining, and tuning integrations between applications. Analysing and resolving technical and application problems.
7. Ability to debug application.
8. Should have complete knowledge on developing RESTful Services.
9. Should be able to also work in agile development methodology.
10. Work with designated JavaScript framework to design, develop, and debug web applications
11. Can work on Angular and Integrate backend services
12. Work with the team to manage, optimize, and customize multiple web applications
13. Manage end to end module lifecycle management of the product
14. Push and pull codes via Git repository
Competency Requirements
- Experience in NodeJS, Java and development using AngularJS / ReactJS
- Experience in front end frameworks such as Angular.js, React.js, Bootstrap, Foundation etc
- Experience in client/server application development
- Knowledge of agile development methodologies
- Knowledge of unit testing theory
- Knowledge of AWS cloud
- Experience in Java, Python and Go will be added advantage
making api and wokring on aws and s3 is added advantage.
Job Description
We are looking for a Node.js Developer responsible for managing the interchange of data between the server and the users. Your primary focus will be the development of all server-side logic, definition and maintenance of the central database, and ensuring high performance and responsiveness to requests from the front-end. You will also be responsible for integrating the front-end elements built by your co-workers into the application. Therefore, a basic understanding of front-end technologies is necessary as well.
What we are asking
- A fast-paced environment is something you feel really comfortable in
- Integration of user-facing elements developed by front-end developers with server side logic
- Professional working experience in Node.js
- Design and implementation of low-latency, high-availability, and performant applications
- Implementation of security and data protection
- Integration of data storage solutions with MongoDB, file stores like AWS S3 or Heroku/AWS.
- Creating API for consumption using ExpressJS.
- Good Knowledge of ExpressJS middle-wares.
- Working experience with CI/CD, test- and deployment automation, and containers
Skills And Qualifications
- Strong proficiency with JavaScript.
- Strong knowledge of Node.js and ExpressJS and other related tech/middle-wares/options.
- Understanding the nature of asynchronous programming and its quirks and workarounds
- Basic understanding of front-end technologies, such as HTML5, and CSS3
- Understanding accessibility and security compliance.
- User authentication and authorization between multiple systems, servers, and environments.
- Integration of multiple data sources and databases into one system
- Understanding fundamental design principles behind a scalable application
- Understanding differences between multiple delivery platforms, such as mobile vs. desktop, and optimizing output to match the specific platform
- Creating database schemas that represent and support business processes
- Implementing automated testing platforms and unit tests
- Proficient understanding of code versioning tools, such as Git
- Good to have deployment experience of NodeJS app on Heroku / AWS / Azure etc.