

Job Description:-
Design, develop IoT/Cloud-based Typescript/ JavaScript/ Node.JS applications using
Amazon Cloud Computing Services.
Work closely with onsite, offshore, and cross functional teams, Product Management, UI/UX developers, Web and Mobile developers, SQA teams to effectively use technologies to build and deliver high quality and on-time delivery of IoT applications
Bug and issue resolution
Proactively Identify risks and failure modes early in the development lifecycle and develop.
POCs to mitigate the risks early in the program.
Assertive communication and team skills
Primary Skills:
Hands on experience (3+ years) in AWS cloud native environment with work experience in
AWS Lambda, Kinesis, DynamoDB
3+ years’ experience in working with NodeJS, Python, Unit Testing and Git
3+ years in work experience with document, relational or timeseries databases
2+ years in work experience with typescript.
1+ years in IaaS framework like Serverless or CDK with CloudFormation knowledge

Similar jobs


Job description
Key Responsibilities
- Design, develop, and maintain serverless applications using AWS services such as Lambda, API Gateway, DynamoDB, and S3.
- Collaborate with front-end developers to integrate user-facing elements with server-side logic.
- Build and maintain RESTful APIs to support web and mobile applications.
- Implement security best practices for AWS services and manage IAM roles and policies.
- Optimize application performance, scalability, and reliability through monitoring and testing.
- Write clean, maintainable, and efficient code following best practices and design patterns.
- Participate in code reviews, providing constructive feedback to peers.
- Troubleshoot and debug applications, identifying performance bottlenecks and areas for improvement.
- Stay updated with emerging technologies and industry trends related to serverless architectures and Python development.
Qualifications
- Bachelors degree in Computer Science, Engineering, or related field, or equivalent experience.
- Proven experience as a Python backend developer, with a strong portfolio of serverless applications.
- Proficiency in AWS services, particularly in serverless architectures (Lambda, API Gateway, DynamoDB, etc.).
- Solid understanding of RESTful API design principles and best practices.
- Familiarity with CI/CD practices and tools (e.g., AWS CodePipeline, Jenkins).
- Experience with containerization technologies (Docker, Kubernetes) is a plus.
- Strong problem-solving skills and the ability to work independently and collaboratively.
- Excellent communication skills, both verbal and written.
Preferred Skills
- Experience with frontend technologies (JavaScript, React, Angular) is a plus.
- Knowledge of data storage solutions (SQL and NoSQL databases).
- AWS certifications (e.g., AWS Certified Developer Associate) are a plus.


Location Delhi NCR
Opening Immediate
Search Context
Over 1.8 billion non-salaried informal sector workers globally, and roughly 700m Indians are
not eligible for pension or other social protection benefits. Without an urgent and effective
response to pension exclusion, they face the grim prospect of extreme poverty for over 20
years once they are too old to work.
pinBox is the only global pensionTech committed exclusively to mass-scale digital micropension inclusion among self-employed women and youth. We deploy our white-labelled,
API-enabled pension administration and delivery platform, our unique deployment model and
a simple and intuitive UI/UX to make access to regulated pension, savings and insurance
products easy and simple for non-salaried informal sector workers. We're working actively
with governments, regulators, multilateral aid agencies and leading financial inclusion
stakeholders in Asia and Africa. The pinBox model is already operating in Rwanda, Kenya
and India. We will expand to Bangladesh, Uganda, Chile, Indonesia and Nigeria by 2023.
Governments and pension regulators use the our pensionTech to jumpstart digital micropension and insurance inclusion among informal sector workers. Pension funds and insurers
use our pensionTech to build a mass market for their products beyond their traditional agentled customer base. Banks, MNOs, cooperatives, MFIs, fintech firms and gig-platforms use
our plug-and-play pensionTech to instantly offer an integrated social protection solution to
their clients, members and employees without any new investments in IT or capacity
enhancement.
We’ve recently completed our first equity fundraise to enhance our engineering, business
and delivery capacity and embark on the next stage of pinBox pensionTech development
and expansion. By 2025, we aim to enable and assist 100 million excluded individuals to
start saving for their old age in a secure, affordable and well-regulated environment.
pinBox is looking for senior software engineers who are deeply passionate about using IT to
solve difficult, real-life problems at scale across multiple countries.
The Senior Software Engineer will be expected to
1. Design, code, test, deploy and maintain applications to satisfy business requirements,
2. Plan and implement technical efforts through all phases of the software development
process,
3. Collaborate cross-functionally to make continuous improvements to the pinBox
pensionTech platform,
4. Help drive engineering plans through a broad approach to engineering quality
(consistent and thoughtful patterns, improved observability, unit and integration testing,
etc.),
5. Adhere to national and global architecture standards, risk management and security
policies,
6. Monitor the performance of applications and work with developers to continuously
improve and optimize performance.
The ideal candidate processes
1. An undergraduate degree in engineering,
2. At least 6 years’ experience as a software engineer or in a similar role,
3. Experience working with distributed version control systems such as Git / Mercurial
4. Frontend: Experience with HTML, CSS, bootstrap, Javascript, Jquery is necessary.
Experience with React / Angular will be an advantage,
5. Backend: Experience with Django/Python, PostgreSQL or any other RDBMS is
mandatory. Experience with Redis will be an advantage,
6. Experience in working with AWS / Azure / Google Cloud,
7. As our applications use a number of third party micro-services, experience with REST
APIs, as also with the Indian digital finance ecosystem (UPI, e-KYC) will be both
necessary and an advantage,
8. Critical thinking and problem-solving skills, and
9. Excellent teamwork and interpersonal skills, a keen eye for detail and the ability to
function effectively and proactively under tight deadlines.





Baetho is a solution focused company aimed at creating and democratizing
the development of customer experience applications through our
proprietary no code platform.
At Baetho, we offer a fun environment and the chance to work with a highly
skilled and motivated team. Our culture is focused on employee happiness,
customer satisfaction and high-quality execution. If you have the right vibe
and believe in fairness and freedom, you’re a great fit for us.
Working Days: 5 days a week, Monday to Friday (some weekend work
required)

We are looking for Interns or 0-1 Year experienced person for our IoT Division based out of Bangalore. Option for Work from Home is also available.
Qualification:
Any Graduate in Computer Science / Electronics / Information Science / Instrumentation from reputed institutions
Required Skills Experience:
- 1+ years of experience in Python with FastAPI must.
- 1+ years of experience in working with a database like PostgreSQL.
- Experience of working on any Linux based OS
- Experience of working on AWS
- Understanding of micro-services.
- Should be able to meet SLAs as per the priorities.
- Good communication skills.
- Ability to prioritize and execute tasks in a high-pressure environment.
Desired Skills Experience:
- Knowledge of Python, Django Framework or REST API.
- Knowledge of Docker and Redis.
- AWS Certified Solutions Architect Associate is a plus.
- Experience with Fast API.
- Exposure to CI/CD practices.
- Awareness of Shell scripting
Keywords: #AWS #EC2 #Python #Psql #Django #Docker #Redis #Kafka.

Hi All,
We are hiring!!
Company: SpringML India Pvt Ltd.
Role:Lead Data Engineer
Location: Hyderabad
Website: https://springml.com/">https://springml.com/
About Company:
At SpringML, we are all about empowering the 'doers' in companies to make smarter decisions with their data. Our predictive analytics products and solutions apply machine learning to today's most pressing business problems so customers get insights they can trust to drive business growth.
We are a tight-knit, friendly team of passionate and driven people who are dedicated to learning, get excited to solve tough problems and like seeing results, fast. Our core values include placing our customers first, empathy and transparency, and innovation. We are a team with a focus on individual responsibility, rapid personal growth, and execution. If you share similar traits, we want you on our team.
What's the opportunity?
SpringML is looking to hire a top-notch Lead Data Engineer who is passionate about working with data and using the latest distributed framework to process large dataset.
As a Lead Data Engineer, your primary role will be to design and build data pipelines. You will be focused on helping client projects on data integration, data prep and implementing machine learning on datasets.
In this role, you will work on some of the latest technologies, collaborate with partners on early win, consultative approach with clients, interact daily with executive leadership, and help build a great company. Chosen team members will be part of the core team and play a critical role in scaling up our emerging practice.
Responsibilities:
- Ability to work as a member of a team assigned to design and implement data integration solutions.
- Build Data pipelines using standard frameworks in Hadoop, Apache Beam and other open-source solutions.
- Learn quickly – ability to understand and rapidly comprehend new areas – functional and technical – and apply detailed and critical thinking to customer solutions.
- Propose design solutions and recommend best practices for large scale data analysis
Skills:
- B.tech degree in computer science, mathematics or other relevant fields.
- 6+years of experience in ETL, Data Warehouse, Visualization and building data pipelines.
- Strong Programming skills – experience and expertise in one of the following: Java, Python, Scala, C.
- Proficient in big data/distributed computing frameworks such as Apache Spark, Kafka,
- Experience with Agile implementation methodology








