

CoffeeBeans
https://coffeebeans.ioAbout
CoffeeBeans Consulting is a technology partner dedicated to driving business transformation. With deep expertise in Cloud, Data, MLOPs, AI, Infrastructure services, Application modernization services, Blockchain, and Big Data, we help organizations tackle complex challenges and seize growth opportunities in today’s fast-paced digital landscape. We’re more than just a tech service provider; we're a catalyst for meaningful change
Tech stack
Candid answers by the company
CoffeeBeans Consulting, founded in 2017, is a high-end technology consulting firm that helps businesses build better products and improve delivery quality through a mix of engineering, product, and process expertise. They work across domains to deliver scalable backend systems, data engineering pipelines, and AI-driven solutions, often using modern stacks like Java, Spring Boot, Python, Spark, Snowflake, Azure, and AWS. With a strong focus on clean architecture, performance optimization, and practical problem-solving, CoffeeBeans partners with clients for both internal and external projects—driving meaningful business outcomes through tech excellence.
Jobs at CoffeeBeans
Key Responsibilities
- Design, develop, and implement backend services using Java (latest version), Spring Boot, and Microservices architecture.
- Participate in the end-to-end development lifecycle, from requirement analysis to deployment and support.
- Collaborate with cross-functional teams (UI/UX, DevOps, Product) to deliver high-quality, scalable software solutions.
- Integrate APIs and manage data flow between services and front-end systems.
- Work on cloud-based deployment using AWS or GCP environments.
- Ensure performance, security, and scalability of services in production.
- Contribute to technical documentation, code reviews, and best practice implementations.
Required Skills:
- Strong hands-on experience with Core Java (latest versions), Spring Boot, and Microservices.
- Solid understanding of RESTful APIs, JSON, and distributed systems.
- Basic knowledge of Kubernetes (K8s) for containerization and orchestration.
- Working experience or strong conceptual understanding of cloud platforms (AWS / GCP).
- Exposure to CI/CD pipelines, version control (Git), and deployment automation.
- Familiarity with security best practices, logging, and monitoring tools.
Preferred Skills:
- Experience with end-to-end deployment on AWS or GCP.
- Familiarity with payment gateway integrations or fintech applications.
- Understanding of DevOps concepts and infrastructure-as-code tools (Added advantage).
We are looking for experienced Data Engineers who can independently build, optimize, and manage scalable data pipelines and platforms.
In this role, you’ll:
- Work closely with clients and internal teams to deliver robust data solutions powering analytics, AI/ML, and operational systems.
- Mentor junior engineers and bring engineering discipline into our data engagements.
Key Responsibilities
- Design, build, and optimize large-scale, distributed data pipelines for both batch and streaming use cases.
- Implement scalable data models, warehouses/lakehouses, and data lakes to support analytics and decision-making.
- Collaborate with stakeholders to translate business requirements into technical solutions.
- Drive performance tuning, monitoring, and reliability of data pipelines.
- Write clean, modular, production-ready code with proper documentation and testing.
- Contribute to architectural discussions, tool evaluations, and platform setup.
- Mentor junior engineers and participate in code/design reviews.
Must-Have Skills
- Strong programming skills in Python and advanced SQL expertise.
- Deep understanding of ETL/ELT, data modeling (OLTP & OLAP), warehousing, and stream processing.
- Hands-on with distributed data processing frameworks (Apache Spark, Flink, or similar).
- Experience with orchestration tools like Airflow (or similar).
- Familiarity with CI/CD pipelines and Git.
- Ability to debug, optimize, and scale data pipelines in production.
Good to Have
- Experience with cloud platforms (AWS preferred; GCP/Azure also welcome).
- Exposure to Databricks, dbt, or similar platforms.
- Understanding of data governance, quality frameworks, and observability.
- Certifications (e.g., AWS Data Analytics, Solutions Architect, or Databricks).
Other Expectations
- Comfortable working in fast-paced, client-facing environments.
- Strong analytical and problem-solving skills with attention to detail.
- Ability to adapt across tools, stacks, and business domains.
- Willingness to travel within India for short/medium-term client engagements, as needed.
As an L3 Data Scientist, you’ll work alongside experienced engineers and data scientists to solve real-world problems using machine learning (ML) and generative AI (GenAI). Beyond classical data science tasks, you’ll contribute to building and fine-tuning large language model (LLM)– based applications, such as chatbots, copilots, and automation workflows.
Key Responsibilities
- Collaborate with business stakeholders to translate problem statements into data science tasks.
- Perform data collection, cleaning, feature engineering, and exploratory data analysis (EDA).
- Build and evaluate ML models using Python and libraries such as scikit-learn and XGBoost.
- Support the development of LLM-powered workflows like RAG (Retrieval-Augmented Generation), prompt engineering, and fine-tuning for use cases including summarization, Q&A, and task automation.
- Contribute to GenAI application development using frameworks like LangChain, OpenAI APIs, or similar ecosystems.
- Work with engineers to integrate models into applications, build/test APIs, and monitor performance post-deployment.
- Maintain reproducible notebooks, pipelines, and documentation for ML and LLM experiments.
- Stay updated on advancements in ML, NLP, and GenAI, and share insights with the team.
Required Skills & Qualifications
- Bachelor’s or Master’s degree in Computer Science, Engineering, Mathematics, Statistics, or a related field.
- 6–9 years of experience in data science, ML, or AI (projects and internships included).
- Proficiency in Python with experience in libraries like pandas, NumPy, scikit-learn, and matplotlib.
- Basic exposure to LLMs (e.g., OpenAI, Cohere, Mistral, Hugging Face) or a strong interest with the ability to learn quickly.
- Familiarity with SQL and structured data handling.
- Understanding of NLP fundamentals and vector-based retrieval techniques (a plus).
- Strong communication, problem-solving skills, and a proactive attitude.
Nice-to-Have (Not Mandatory)
- Experience with GenAI prototyping using LangChain, LlamaIndex, or similar frameworks.
- Knowledge of REST APIs and model integration into backend systems.
- Familiarity with cloud platforms (AWS/GCP/Azure), Docker, or Git.
As an L1/L2 Data Scientist, you’ll work alongside experienced engineers and data scientists to solve real-world problems using machine learning (ML) and generative AI (GenAI). Beyond classical data science tasks, you’ll contribute to building and fine-tuning large language model (LLM)– based applications, such as chatbots, copilots, and automation workflows.
Key Responsibilities
- Collaborate with business stakeholders to translate problem statements into data science tasks.
- Perform data collection, cleaning, feature engineering, and exploratory data analysis (EDA).
- Build and evaluate ML models using Python and libraries such as scikit-learn and XGBoost.
- Support the development of LLM-powered workflows like RAG (Retrieval-Augmented Generation), prompt engineering, and fine-tuning for use cases including summarization, Q&A, and task automation.
- Contribute to GenAI application development using frameworks like LangChain, OpenAI APIs, or similar ecosystems.
- Work with engineers to integrate models into applications, build/test APIs, and monitor performance post-deployment.
- Maintain reproducible notebooks, pipelines, and documentation for ML and LLM experiments.
- Stay updated on advancements in ML, NLP, and GenAI, and share insights with the team.
Required Skills & Qualifications
- Bachelor’s or Master’s degree in Computer Science, Engineering, Mathematics, Statistics, or a related field.
- 2.5–5 years of experience in data science, ML, or AI (projects and internships included).
- Proficiency in Python with experience in libraries like pandas, NumPy, scikit-learn, and matplotlib.
- Basic exposure to LLMs (e.g., OpenAI, Cohere, Mistral, Hugging Face) or strong interest with the ability to learn quickly.
- Familiarity with SQL and structured data handling.
- Understanding of NLP fundamentals and vector-based retrieval techniques (a plus).
- Strong communication, problem-solving skills, and a proactive attitude.
Nice-to-Have (Not Mandatory)
- Experience with GenAI prototyping using LangChain, LlamaIndex, or similar frameworks.
- Knowledge of REST APIs and model integration into backend systems.
- Familiarity with cloud platforms (AWS/GCP/Azure), Docker, or Git.
Focus Areas:
- Build applications and solutions that process and analyze large-scale data.
- Develop data-driven applications and analytical tools.
- Implement business logic, algorithms, and backend services.
- Design and build APIs for secure and efficient data exchange.
Key Responsibilities:
- Develop and maintain data processing applications using Apache Spark and Hadoop.
- Write MapReduce jobs and complex data transformation logic.
- Implement machine learning models and analytics solutions for business use cases.
- Optimize code for performance and scalability; perform debugging and troubleshooting.
- Work hands-on with Databricks for data engineering and analysis.
- Design and manage Airflow DAGs for orchestration and automation.
- Integrate and maintain CI/CD pipelines (preferably using Jenkins).
Primary Skills & Qualifications:
- Strong programming skills in Scala and Python.
- Expertise in Apache Spark for large-scale data processing.
- Solid understanding of data structures and algorithms.
- Proven experience in application development and software engineering best practices.
- Experience working in agile and collaborative environments.
We are seeking an experienced Lead DevOps Engineer with deep expertise in Kubernetes infrastructure design and implementation. This role requires someone who can architect, build, and manage enterprise-grade Kubernetes clusters from the ground up. You’ll lead modernization initiatives, shape infrastructure strategy, and work with cutting-edge cloud-native technologies.
🚀 Key Responsibilities
Infrastructure Design & Implementation
- Architect and design enterprise-grade Kubernetes clusters across AWS, Azure, and GCP.
- Build production-ready Kubernetes infrastructure with HA, scalability, and security best practices.
- Implement Infrastructure as Code with Terraform, Helm, and GitOps workflows.
- Set up monitoring, logging, and observability for Kubernetes workloads.
- Design and execute backup and disaster recovery strategies for containerized applications.
Leadership & Team Management
- Lead a team of 3–4 DevOps engineers, providing technical mentorship.
- Drive best practices in containerization, orchestration, and cloud-native development.
- Collaborate with development teams to optimize deployment strategies.
- Conduct code reviews and maintain infrastructure quality standards.
- Build knowledge-sharing culture with documentation and training.
Operational Excellence
- Manage and scale CI/CD pipelines integrated with Kubernetes.
- Implement security policies (RBAC, network policies, container scanning).
- Optimize cluster performance and cost-efficiency.
- Automate operations to minimize manual interventions.
- Ensure 99.9% uptime for production workloads.
Strategic Planning
- Define the infrastructure roadmap aligned with business needs.
- Evaluate and adopt new cloud-native technologies.
- Perform capacity planning and cloud cost optimization.
- Drive risk assessment and mitigation strategies.
🛠 Must-Have Technical Skills
Kubernetes Expertise
- 6+ years of hands-on Kubernetes experience in production.
- Deep knowledge of Kubernetes architecture (etcd, API server, scheduler, kubelet).
- Advanced Kubernetes networking (CNI, Ingress, Service mesh).
- Strong grasp of Kubernetes storage (CSI, PVs, StorageClasses).
- Experience with Operators and Custom Resource Definitions (CRDs).
Infrastructure as Code
- Terraform (advanced proficiency).
- Helm (developing and managing complex charts).
- Config management tools (Ansible, Chef, Puppet).
- GitOps workflows (ArgoCD, Flux).
Cloud Platforms
- Hands-on experience with at least 2 of the following:
- AWS: EKS, EC2, VPC, IAM, CloudFormation
- Azure: AKS, VNets, ARM templates
- GCP: GKE, Compute Engine, Deployment Manager
CI/CD & DevOps Tools
- Jenkins, GitLab CI, GitHub Actions, Azure DevOps
- Docker (advanced optimization and security practices)
- Container registries (ECR, ACR, GCR, Docker Hub)
- Strong Git workflows and branching strategies
Monitoring & Observability
- Prometheus & Grafana (metrics and dashboards)
- ELK/EFK stack (centralized logging)
- Jaeger/Zipkin (tracing)
- AlertManager (intelligent alerting)
💡 Good-to-Have Skills
- Service Mesh (Istio, Linkerd, Consul)
- Serverless (Knative, OpenFaaS, AWS Lambda)
- Running databases in Kubernetes (Postgres, MongoDB operators)
- ML pipelines (Kubeflow, MLflow)
- Security tools (Aqua, Twistlock, Falco, OPA)
- Compliance (SOC2, PCI-DSS, GDPR)
- Python/Go for automation
- Advanced Shell scripting (Bash/PowerShell)
🎓 Qualifications
- Bachelor’s in Computer Science, Engineering, or related field.
- Certifications (preferred):
- Certified Kubernetes Administrator (CKA)
- Certified Kubernetes Application Developer (CKAD)
- Cloud provider certifications (AWS/Azure/GCP).
Experience
- 6–7 years of DevOps/Infrastructure engineering.
- 4+ years of Kubernetes in production.
- 2+ years in a lead role managing teams.
- Experience with large-scale distributed systems and microservices.
Role Overview
We're looking for experienced Data Engineers who can independently design, build, and manage scalable data platforms. You'll work directly with clients and internal teams to develop robust data pipelines that support analytics, AI/ML, and operational systems.
You’ll also play a mentorship role and help establish strong engineering practices across our data projects.
Key Responsibilities
- Design and develop large-scale, distributed data pipelines (batch and streaming)
- Implement scalable data models, warehouses/lakehouses, and data lakes
- Translate business requirements into technical data solutions
- Optimize data pipelines for performance and reliability
- Ensure code is clean, modular, tested, and documented
- Contribute to architecture, tooling decisions, and platform setup
- Review code/design and mentor junior engineers
Must-Have Skills
- Strong programming skills in Python and advanced SQL
- Solid grasp of ETL/ELT, data modeling (OLTP & OLAP), and stream processing
- Hands-on experience with frameworks like Apache Spark, Flink, etc.
- Experience with orchestration tools like Airflow
- Familiarity with CI/CD pipelines and Git
- Ability to debug and scale data pipelines in production
Preferred Skills
- Experience with cloud platforms (AWS preferred, GCP or Azure also fine)
- Exposure to Databricks, dbt, or similar tools
- Understanding of data governance, quality frameworks, and observability
- Certifications (e.g., AWS Data Analytics, Solutions Architect, Databricks) are a bonus
What We’re Looking For
- Problem-solver with strong analytical skills and attention to detail
- Fast learner who can adapt across tools, tech stacks, and domains
- Comfortable working in fast-paced, client-facing environments
- Willingness to travel within India when required
Similar companies
About the company
We are Proximity - a global team of coders, designers, product managers, geeks and experts. We solve hard, long-term engineering problems and build cutting edge tech products.
About us
Born in 2019, Proximity Works is a global, fully distributed tech firm headquartered in San Francisco - with hubs across Mumbai, Dubai, Toronto, Stockholm, and Bengaluru. We’re in the business of solving high-stakes engineering challenges with AI-powered solutions tailored for industries like sports, media & entertainment, fintech, and enterprise platforms. From real-time game analytics and ticketing workflows to creative content generation, we help build software that serves millions every day.
About the Founders
At the helm is Hardik Jagda, CEO - a technologist with a startup DNA who brings clarity to complexity and a passion for building delightful experiences.
Milestones & Impact
- Trusted by some of the world’s biggest players - from major media & entertainment giants to one of the world’s largest cricket websites and the second-largest stock exchange in the world.
- Delivered game-changing tech: slashing content creation by 90%, doubling performance metrics for NASDAQ clients, and accelerating speed/performance wins for platforms like Dream11.
Culture & Why It Matters
- Fully distributed and flexible: work 100% remotely, design your own schedule, build habits that work for you and not the other way around.
- People-first culture: Community events, “Proxonaut battles,” monthly off-sites, and a liberal referral policy keep us connected even when we’re apart.
- High-trust environment: autonomy is encouraged. You’re empowered to act, learn fast, and iterate boldly. We know great work comes when talented people have space to think and create.
Jobs
5
About the company
Jobs
20
About the company
]eShipz: Simplifying Global Shipping for Businesses: At eShipz, we are revolutionizing how businesses manage their shipping processes. Our platform is designed to offer seamless multi-carrier integration, enabling businesses of all sizes to ship effortlessly across the globe. Whether you're an e-commerce brand, a manufacturer, or a logistics provider, eShipz helps streamline your supply chain with real-time tracking, automated shipping labels, cost-effective shipping rates, and comprehensive reporting.
Our goal is to empower businesses by simplifying logistics, reducing shipping costs, and improving operational efficiency. With an easy-to-use dashboard and a dedicated support team, eShipz ensures that you focus on scaling your business while we handle your shipping needs.
Jobs
5
About the company
Jobs
12
About the company
At LearnTube, we're reimagining how the world learns making education accessible, affordable, and outcome-driven using Generative AI. Our platform turns scattered internet content into structured, personalised learning journeys using:
- AI-powered tutors that teach live, solve doubts instantly, and give real-time feedback
- Frictionless delivery via WhatsApp, mobile, and web
- Trusted by 2.2 million learners across 64 countries
Founded by Shronit Ladhani and Gargi Ruparelia, both second-time entrepreneurs and ed-tech builders:
- Shronit is a TEDx speaker and an outspoken advocate for disrupting traditional learning systems.
- Gargi is one of the Top Women in AI in India, recognised by the government, and leads our AI and scalability roadmap.
Together, they bring deep product thinking, bold storytelling, and executional clarity to LearnTube’s vision. LearnTube is proudly backed by Google as part of their 2024 AI First Accelerator, giving us access to cutting-edge tech, mentorship, and cloud credits.
Jobs
9
About the company
Founded in 2016 by ex-Product & IT leaders from BCG, KPMG, RBS & Microsoft, Grey Chain is an AI and mobile-first product and services firm that focuses on design-led solutions.
We are trusted by Global companies including the likes of Accenture, UNICEF, BOSE, WHO and many other Fortune 500 Companies
We offer end-to-end engineering and development services for digital journeys, including mobile apps, CRMs, ERPs, and enterprise-grade solutions. We also provide consulting services and emphasize our expertise in Generative AI.
Jobs
6
About the company
Highfly Sourcing, one of the best immigration consultants in delhi having expertise of providing quality solutions in immigration services to individuals, families and corporate clients those who are seeking to Settle, Work, Study, Visit or move temporarily or permanently in Canada, Australia, New Zealand , Denmark, Germany, UK, Hong Kong,Singapore and others countries globally.
We are one of the best immigration consulting services of the country. It is a matter of great pride that till date our success rate has been hundred percent. There is no doubt about the fact, that any assignment that is taken up becomes successful because of not just one but many reasons. From planning to execution, everything has to be done with utmost perfection.
Here at Highfly Sourcing, your personal and professional needs are kept into consideration before recommending a visa for you. Our consulting professionals are there to study your profile thoroughly and counsel you as per your future aspirations. We assure that once you meet our consulting professionals all your doubts and queries will take a back seat and you will just want to be proactive enough to complete the process at the earliest and take a flight to your dream country.
Our Post Landing services will be an added advantage in your new country as we will provide you with job search services, pick up assistance and accommodation assistance. We will be there for you till you are permanently settled in the country.
Jobs
21
About the company
We're a passionate team united by a clear mission: helping merchants grow and succeed. Coming from diverse professional backgrounds, we bring a unique blend of skills to the table, rooted in deep e-commerce expertise, strong tech capabilities, and a solid understanding of what drives merchant success.
Jobs
1
About the company
Jobs
1
About the company
Albert Invent is a cutting-edge R&D software company that’s built by scientists, for scientists. Their cloud-based platform unifies lab data, digitises workflows and uses AI/ML to help materials and chemical R&D teams invent faster and smarter. With thousands of researchers in 30+ countries already using the platform, Albert Invent is helping transform how chemistry-led companies go from idea to product.
What sets them apart: built specifically for chemistry and materials science (not generic SaaS), with deep integrations (ELN, LIMS, AI/ML) and enterprise-grade security and compliance.
Jobs
4








