12+ Amazon S3 Jobs in Hyderabad | Amazon S3 Job openings in Hyderabad
Apply to 12+ Amazon S3 Jobs in Hyderabad on CutShort.io. Explore the latest Amazon S3 Job opportunities across top companies like Google, Amazon & Adobe.
Qualifications:*
1. 10+ years of experience, with 3+ years as Database Architect or related role
2. Technical expertise in data schemas, Amazon Redshift, Amazon S3, and Data Lakes
3. Analytical skills in data warehouse design and business intelligence
4. Strong problem-solving and strategic thinking abilities
5. Excellent communication skills
6. Bachelor's degree in Computer Science or related field; Master's degree preferred
*Skills Required:*
1. Database architecture and design
2. Data warehousing and business intelligence
3. Cloud-based data infrastructure (Amazon Redshift, S3, Data Lakes)
4. Data governance and security
5. Analytical and problem-solving skills
6. Strategic thinking and communication
7. Collaboration and team management
main objectives of the Senior Developer role
- Delivery of working software according to specifications
- Providing clarity on the progress of development work
- Assisting team members by sharing knowledge and coaching
- Suggesting process improvements with regard to team collaboration
The key requirements include the following:
- Bachelor's degree in software engineering or a related field
- 7+ years of development experience with Python programming
- Experience in setting up CI/CD workflows
- Experience in drafting solution architecture
- Excellent written and verbal communication skills in English
Key expertise and experience:
- Understanding of application architecture built on AWS/Python/React/MongoDB technological stack
- Proven Python skills and experience with modules such as Flask and FastAPI
- Experience in building high-load applications in AWS Cloud, preferably in a micro-service architecture
- Experience in developing commercial software, preferably in the financial domain
- Proficiency in working with legacy code and acquiring domain knowledge
- Experience with MongoDB and/or other NoSQL databases
- Experience in creating and automating tests (pytest) and using containers (Docker)
- Proficiency in using Linux-based development environments with GitHub and CI/CD
- Familiarity with the Atlassian stack (JIRA/Confluence)
- Nice to have – experience in integration with ERP, CRM, and SAP
- Nice to have – experience in building financial systems/knowledge of enterprise economics
Knowledge of key processes:
- Scrum / Agile way of working
- TDD/BDD
- CI/CD
Technical Skills:
- Ability to understand and translate business requirements into design.
- Proficient in AWS infrastructure components such as S3, IAM, VPC, EC2, and Redshift.
- Experience in creating ETL jobs using Python/PySpark.
- Proficiency in creating AWS Lambda functions for event-based jobs.
- Knowledge of automating ETL processes using AWS Step Functions.
- Competence in building data warehouses and loading data into them.
Responsibilities:
- Understand business requirements and translate them into design.
- Assess AWS infrastructure needs for development work.
- Develop ETL jobs using Python/PySpark to meet requirements.
- Implement AWS Lambda for event-based tasks.
- Automate ETL processes using AWS Step Functions.
- Build data warehouses and manage data loading.
- Engage with customers and stakeholders to articulate the benefits of proposed solutions and frameworks.
Multinational Company providing energy & Automation digital
Skills
Interfaces with other processes and/or business functions to ensure they can leverage the
benefits provided by the AWS Platform process
Responsible for managing the configuration of all IaaS assets across the platforms
Hands-on python experience
Manages the entire AWS platform(Python, Flask, RESTAPI, serverless framework) and
recommend those that best meet the organization's requirements
Has a good understanding of the various AWS services, particularly: S3, Athena, Python code,
Glue, Lambda, Cloud Formation, and other AWS serverless resources.
AWS Certification is Plus
Knowledge of best practices for IT operations in an always-on, always-available service model
Responsible for the execution of the process controls, ensuring that staff comply with process
and data standards
Qualifications
Bachelor’s degree in Computer Science, Business Information Systems or relevant experience and
accomplishments
3 to 6 years of experience in the IT field
AWS Python developer
AWS, Serverless/Lambda, Middleware.
Strong AWS skills including Data Pipeline, S3, RDS, Redshift with familiarity with other components
like - Lambda, Glue, Step functions, CloudWatch
Must have created REST API with AWS Lambda.
Python relevant exp 3 years
Good to have Experience working on projects and problem solving with large scale multivendor
teams.
Good to have knowledge on Agile Development
Good knowledge on SDLC.
Hands on AWS Databases, (RDS, etc)
Good to have Unit testing exp.
Good to have CICD working knowledge.
Decent communication, as there will be client interaction and documentation.
Education (degree): Bachelor’s degree in Computer Science, Business Information Systems or relevant
experience and accomplishments
Years of Experience: 3-6 years
Technical Skills
Linux/Unix system administration
Continuous Integration/Continuous Delivery tools like Jenkins
Cloud provisioning and management – Azure, AWS, GCP
Ansible, Chef, or Puppet
Python, PowerShell & BASH
Job Details
JOB TITLE/JOB CODE: AWS Python Develop[er, III-Sr. Analyst
RC: TBD
PREFERRED LOCATION: HYDERABAD, IND
POSITION REPORTS TO: Manager USI T&I Cloud Managed Platform
CAREER LEVEL: 3
Work Location:
Hyderabad
We are looking for a focussed, energetic and motivated Senior Java Developer. You must have sufficient experience in Java Development, Spring, Spring Boot, AWS and SQS, S3. You will be challenged with complex and robust product development.
As a Senior Java Developer, you will be the most senior developer within the company working alongside other Java Developers, the Engineering Manager, and the CTO.
Senior Java Developer requirements
Solid development experience using Java, Springboot, SQS, S3 and Microservices
Experience working in Swing Java framework
Revises, updates, refactors, and debugs code
Write clear, concise and efficient code and ensure it is fully tested
Service and develop features using Java with AWS, JUnit, SQL and Spring framework
You have a good knowledge of Cloud technologies (ideally AWS)
Participate in peer coding reviews and work in a collaborative, agile environment
Demonstrate technical knowledge and expertise with a clear understanding on the products' technical composition
Senior Java Developer responsibilities
7-9 yrs minimum work experience
Excellent analytical, troubleshooting, and communication skills
Excellent attention to detail and time-management skills
Serves as an expert on developed applications
Ability to adapt and learn new product/technology is key
Collaborates with internal teams to produce software design and architecture
Knowledge of Agile frameworks (i.e. Scrum) and has worked with Jira or Spira
Attend daily stand-up meetings alongside the Scrum Master
A self-starter, hardworking and committed individual
Work UK hours - ideally 9.00am - 6pm with the expectation to work out of ours when needed
Join our team
We're looking for an experienced and passionate Data Engineer to join our team. Our vision is to empower Genesys to leverage data to drive better customer and business outcomes. Our batch and streaming solutions turn vast amounts of data into useful insights. If you’re interested in working with the latest big data technologies, using industry leading BI analytics and visualization tools, and bringing the power of data to our customers’ fingertips then this position is for you!
Our ideal candidate thrives in a fast-paced environment, enjoys the challenge of highly complex business contexts (that are typically being defined in real-time), and, above all, is a passionate about data and analytics.
What you'll get to do
- Work in an agile development environment, constantly shipping and iterating.
- Develop high quality batch and streaming big data pipelines.
- Interface with our Data Consumers, gathering requirements, and delivering complete data solutions.
- Own the design, development, and maintenance of datasets that drive key business decisions.
- Support, monitor and maintain the data models
- Adopt and define the standards and best practices in data engineering including data integrity, performance optimization, validation, reliability, and documentation.
- Keep up-to-date with advances in big data technologies and run pilots to design the data architecture to scale with the increased data volume using cloud services.
- Triage many possible courses of action in a high-ambiguity environment, making use of both quantitative analysis and business judgment.
Your experience should include
- Bachelor’s degree in CS or related technical field.
- 5+ years of experience in data modelling, data development, and data warehousing.
- Experience working with Big Data technologies (Hadoop, Hive, Spark, Kafka, Kinesis).
- Experience with large scale data processing systems for both batch and streaming technologies (Hadoop, Spark, Kinesis, Flink).
- Experience in programming using Python, Java or Scala.
- Experience with data orchestration tools (Airflow, Oozie, Step Functions).
- Solid understanding of database technologies including NoSQL and SQL.
- Strong in SQL queries (experience with Snowflake Cloud Datawarehouse is a plus)
- Work experience in Talend is a plus
- Track record of delivering reliable data pipelines with solid test infrastructure, CICD, data quality checks, monitoring, and alerting.
- Strong organizational and multitasking skills with ability to balance competing priorities.
- Excellent communication (verbal and written) and interpersonal skills and an ability to effectively communicate with both business and technical teams.
- An ability to work in a fast-paced environment where continuous innovation is occurring, and ambiguity is the norm.
Good to have
- Experience with AWS big data technologies - S3, EMR, Kinesis, Redshift, Glue
- Collaborate with Dev, QA and Data Science teams on environment maintenance, monitoring (ELK, Prometheus or equivalent), deployments and diagnostics
- Administer a hybrid datacenter, including AWS and EC2 cloud assets
- Administer, automate and troubleshoot container based solutions deployed on AWS ECS
- Be able to troubleshoot problems and provide feedback to engineering on issues
- Automate deployment (Ansible, Python), build (Git, Maven. Make, or equivalent) and integration (Jenkins, Nexus) processes
- Learn and administer technologies such as ELK, Hadoop etc.
- A self-starter and enthusiasm to learn and pick up new technologies in a fast-paced environment.
Need to have
- Hands-on Experience in Cloud based DevOps
- Experience working in AWS (EC2, S3, CloudFront, ECR, ECS etc)
- Experience with any programming language.
- Experience using Ansible, Docker, Jenkins, Kubernetes
- Experience in Python.
- Should be very comfortable working in Linux/Unix environment.
- Exposure to Shell Scripting.
- Solid troubleshooting skills
About the Role
The Dremio India team owns the DataLake Engine along with Cloud Infrastructure and services that power it. With focus on next generation data analytics supporting modern table formats like Iceberg, Deltalake, and open source initiatives such as Apache Arrow, Project Nessie and hybrid-cloud infrastructure, this team provides various opportunities to learn, deliver, and grow in career. We are looking for technical leaders with passion and experience in architecting and delivering high-quality distributed systems at massive scale.
Responsibilities & ownership
- Lead end-to-end delivery and customer success of next-generation features related to scalability, reliability, robustness, usability, security, and performance of the product
- Lead and mentor others about concurrency, parallelization to deliver scalability, performance and resource optimization in a multithreaded and distributed environment
- Propose and promote strategic company-wide tech investments taking care of business goals, customer requirements, and industry standards
- Lead the team to solve complex, unknown and ambiguous problems, and customer issues cutting across team and module boundaries with technical expertise, and influence others
- Review and influence designs of other team members
- Design and deliver architectures that run optimally on public clouds like GCP, AWS, and Azure
- Partner with other leaders to nurture innovation and engineering excellence in the team
- Drive priorities with others to facilitate timely accomplishments of business objectives
- Perform RCA of customer issues and drive investments to avoid similar issues in future
- Collaborate with Product Management, Support, and field teams to ensure that customers are successful with Dremio
- Proactively suggest learning opportunities about new technology and skills, and be a role model for constant learning and growth
Requirements
- B.S./M.S/Equivalent in Computer Science or a related technical field or equivalent experience
- Fluency in Java/C++ with 15+ years of experience developing production-level software
- Strong foundation in data structures, algorithms, multi-threaded and asynchronous programming models and their use in developing distributed and scalable systems
- 8+ years experience in developing complex and scalable distributed systems and delivering, deploying, and managing microservices successfully
- Subject Matter Expert in one or more of query processing or optimization, distributed systems, concurrency, micro service based architectures, data replication, networking, storage systems
- Experience in taking company-wide initiatives, convincing stakeholders, and delivering them
- Expert in solving complex, unknown and ambiguous problems spanning across teams and taking initiative in planning and delivering them with high quality
- Ability to anticipate and propose plan/design changes based on changing requirements
- Passion for quality, zero downtime upgrades, availability, resiliency, and uptime of the platform
- Passion for learning and delivering using latest technologies
- Hands-on experience of working projects on AWS, Azure, and GCP
- Experience with containers and Kubernetes for orchestration and container management in private and public clouds (AWS, Azure, and GCP)
- Understanding of distributed file systems such as S3, ADLS or HDFS
- Excellent communication skills and affinity for collaboration and teamwork
at VAYUZ Technologies
Roles and Responsibilities
1. Ability to work on diverse backend stack such as Node JS, Java, Express JS
2. Ability to work on diverse frontend stack such as React JS, Angular 6/7/8/9, HTML5, CSS3
3. Ability to deliver quick POC’s using cutting edge technologies.
4. Preparing reports, manuals and other documentation on the status, operation and maintenance of software.
5. Design, develop, and unit test applications in accordance with established standards
6. Developing, refining, and tuning integrations between applications. Analysing and resolving technical and application problems.
7. Ability to debug application.
8. Should have complete knowledge on developing RESTful Services.
9. Should be able to also work in agile development methodology.
10. Work with designated JavaScript framework to design, develop, and debug web applications
11. Can work on Angular and Integrate backend services
12. Work with the team to manage, optimize, and customize multiple web applications
13. Manage end to end module lifecycle management of the product
14. Push and pull codes via Git repository
Competency Requirements
- Experience in NodeJS, Java and development using AngularJS / ReactJS
- Experience in front end frameworks such as Angular.js, React.js, Bootstrap, Foundation etc
- Experience in client/server application development
- Knowledge of agile development methodologies
- Knowledge of unit testing theory
- Knowledge of AWS cloud
- Experience in Java, Python and Go will be added advantage
Great place to work certified organisation among top 50.
Responsibilities:
- Working directly with clients to understand requirements for a green field development
- Designing the technology and cloud infrastructure architecture to achieve the functional and nonfunctional requirements
- The product you will be working on needs to scale up to support millions of users an efficient micro service distributed architecture
- Solution needs to be easy to deploy and manage on multiple cloud providers (AWS, GCP or Azure)
- Mentoring, guiding and training the team on various technology, quality and security aspects
- Guiding the team to implement automated CI/CD processes
- Strong analytical, problem solving and data analysis
- Excellent communication, presentation and interpersonal skills are a must
- Micro service frameworks such as Java SpringBoot
- Design and implement automated unit and integration tests
- Writing scalable, robust, testable, efficient, and easily maintainable code
- Familiarity with most AWS Services - EC2, ECS, RDS, ECR, S3, SNS, SQS, and more
- Experience with Docker and Kubernetes
- Deploying and Scaling microservices to AWS
- Hands-on experience with AWS cloud platform in evaluation and cost estimation
15 years US based Product Company
- Should have good hands-on experience in Informatica MDM Customer 360, Data Integration(ETL) using PowerCenter, Data Quality.
- Must have strong skills in Data Analysis, Data Mapping for ETL processes, and Data Modeling.
- Experience with the SIF framework including real-time integration
- Should have experience in building C360 Insights using Informatica
- Should have good experience in creating performant design using Mapplets, Mappings, Workflows for Data Quality(cleansing), ETL.
- Should have experience in building different data warehouse architecture like Enterprise,
- Federated, and Multi-Tier architecture.
- Should have experience in configuring Informatica Data Director in reference to the Data
- Governance of users, IT Managers, and Data Stewards.
- Should have good knowledge in developing complex PL/SQL queries.
- Should have working experience on UNIX and shell scripting to run the Informatica workflows and to control the ETL flow.
- Should know about Informatica Server installation and knowledge on the Administration console.
- Working experience with Developer with Administration is added knowledge.
- Working experience in Amazon Web Services (AWS) is an added advantage. Particularly on AWS S3, Data pipeline, Lambda, Kinesis, DynamoDB, and EMR.
- Should be responsible for the creation of automated BI solutions, including requirements, design,development, testing, and deployment