
- Key responsibility is to design, develop & maintain efficient Data models for the organization maintained to ensure optimal query performance by the consumption layer.
- Developing, Deploying & maintaining a repository of UDXs written in Java / Python.
- Develop optimal Data Model design, analyzing complex distributed data deployments, and making recommendations to optimize performance basis data consumption patterns, performance expectations, the query is executed on the tables/databases, etc.
- Periodic Database health check and maintenance
- Designing collections in a no-SQL Database for efficient performance
- Document & maintain data dictionary from various sources to enable data governance
- Coordination with Business teams, IT, and other stakeholders to provide best-in-class data pipeline solutions, exposing data via APIs, loading in down streams, No-SQL Databases, etc
- Data Governance Process Implementation and ensuring data security
Requirements
- Extensive working experience in Designing & Implementing Data models in OLAP Data Warehousing solutions (Redshift, Synapse, Snowflake, Teradata, Vertica, etc).
- Programming experience using Python / Java.
- Working knowledge in developing & deploying User-defined Functions (UDXs) using Java / Python.
- Strong understanding & extensive working experience in OLAP Data Warehousing (Redshift, Synapse, Snowflake, Teradata, Vertica, etc) architecture and cloud-native Data Lake (S3, ADLS, BigQuery, etc) Architecture.
- Strong knowledge in Design, Development & Performance tuning of 3NF/Flat/Hybrid Data Model.
- Extensive technical experience in SQL including code optimization techniques.
- Strung knowledge of database performance and tuning, troubleshooting, and tuning.
- Knowledge of collection design in any No-SQL DB (DynamoDB, MongoDB, CosmosDB, etc), along with implementation of best practices.
- Ability to understand business functionality, processes, and flows.
- Good combination of technical and interpersonal skills with strong written and verbal communication; detail-oriented with the ability to work independently.
- Any OLAP DWH DBA Experience and User Management will be added advantage.
- Knowledge in financial industry-specific Data models such as FSLDM, IBM Financial Data Model, etc will be added advantage.
- Experience in Snowflake will be added advantage.
- Working experience in BFSI/NBFC & data understanding of Loan/Mortgage data will be added advantage.
Functional knowledge
- Data Governance & Quality Assurance
- Modern OLAP Database Architecture & Design
- Linux
- Data structures, algorithm & data modeling techniques
- No-SQL database architecture
- Data Security

Similar jobs
Job Title : Java Backend Developer
Experience : 3 – 8 Years
Location : Pune (Onsite) (Pune candidates Only)
Notice Period : Immediate to 15 Days (or serving NP whose LWD is near)
About the Role :
We are seeking an experienced Java Backend Developer with strong hands-on skills in backend microservices development, API design, cloud platforms, observability, and CI/CD.
The ideal candidate will contribute to building scalable, secure, and reliable applications while working closely with cross-functional teams.
Mandatory Skills : Java 8 / Java 17, Spring Boot 3.x, REST APIs, Hibernate / JPA, MySQL, MongoDB, Prometheus / Grafana / Spring Actuators, AWS, Docker, Jenkins / GitHub Actions, GitHub, Windows 7 / Linux.
Key Responsibilities :
- Design, develop, and maintain backend microservices and REST APIs
- Implement data persistence using relational and NoSQL databases
- Ensure performance, scalability, and security of backend systems
- Integrate observability and monitoring tools for production environments
- Work within CI/CD pipelines and containerized deployments
- Collaborate with DevOps, QA, and product teams for feature delivery
- Troubleshoot, optimize, and improve existing modules and services
Mandatory Skills :
- Languages & Frameworks : Java 8, Java 17, Spring Boot 3.x, REST APIs, Hibernate, JPA
- Databases : MySQL, MongoDB
- Observability : Prometheus, Grafana, Spring Actuators
- Cloud Technologies : AWS
- Containerization Tools : Docker
- CI/CD Tools : Jenkins, GitHub Actions
- Version Control : GitHub
- Operating Systems : Windows 7, Linux
Nice to Have :
- Strong analytical and debugging abilities
- Experience working in Agile/Scrum environments
- Good communication and collaborative skills
About FrontM
At FrontM, we are on a mission to transform the lives of frontline workforces, particularly in the maritime industry. We believe in creating a more connected, empowered, and engaged workforce by building cutting-edge solutions that merge the power of technology with human-centric needs. Our vision is to develop the world’s leading digital toolbox platform for maritime operations —a platform that brings everything for frontline workforces from digital wallets, recruitment, onboarding, healthcare, and learning to welfare and human capital management under one seamless umbrella.
Role Summary
As a JavaScript Developer at FrontM, you will be at the forefront of developing our pioneering digital toolbox platform and the low-code developer framework that powers it. You will have the opportunity to work with the latest JavaScript frameworks, integrating advanced technologies such as Large Language Models (LLMs), AI, and the latest GPT models. You’ll also be part of our exciting roadmap to evolve our low-code platform into a no-code solution, making app development accessible to everyone. Your contributions will be pivotal in the creation and enhancement of the Maritime App Store, where innovation meets practicality, offering solutions that make a tangible difference in the lives of seafarers and other frontline workers.
Key Responsibilities
Application Development (≈60%)
- Build micro-apps using the frontm.ai framework
- Implement intent-based architectures, context and state management
- Develop responsive UIs, forms, collections, filters, and workflows
- Integrate AWS services (Lambda, S3, DynamoDB, Bedrock)
- Build conversational AI features and real-time capabilities (messaging, video, notifications)
Framework Development (≈25%)
- Enhance and extend the frontm.ai core framework
- Build reusable components, patterns, and accelerators
- Improve performance for low-bandwidth environments
- Contribute to documentation, examples, and design reviews
- Support migration towards TypeScript and future Rust components
AI-Assisted Development (≈15%)
- Use Claude Code for efficient development
- Write and refine prompts for code generation
- Review, validate, and harden AI-generated code
- Implement LLM integrations via AWS Bedrock / OpenAI
- Build AI assistants using the skills layer
Required Technical Skills
JavaScript / TypeScript
- 5+ years professional JavaScript experience
- Strong TypeScript, async patterns, modular design
- Clean code practices and modern tooling
Architecture & Cloud
- Microservices and event-driven systems
- Serverless AWS (Lambda, API Gateway, DynamoDB, S3)
- REST APIs, WebSockets, CI/CD
- Infrastructure as Code experience preferred
AI & LLMs
- Hands-on use of Claude Code or similar tools
- Prompt engineering and hallucination mitigation
- Conversational AI and NLP experience
Data
- MongoDB / MongoDB Atlas
- Caching, indexing, and multi-tenant data patterns
Desired skills
- Experience with low-bandwidth or offline-first systems
- Understanding of secure, distributed deployments
- Exposure to healthcare, logistics, or maritime systems
Experience & Education
- 5+ years software development
- 2+ years AWS serverless
- 1+ year AI-assisted development
- Degree in Computer Science or equivalent experience
Personal Attributes
- Strong problem-solving and critical thinking
- Comfortable reviewing AI-generated code
- Clear communicator and reliable team contributor
- Self-driven, detail-oriented, and adaptable
Why join FrontM?
Above-Market Compensation: We believe in rewarding talent, offering a salary package that reflects your skills and potential.
Long-Term Career Growth: As FrontM expands, so will your opportunities. We are committed to helping our team members develop their careers, offering mentorship, learning opportunities, and the chance to take on more responsibility.
Cutting-Edge Technology: Work with the latest in JavaScript frameworks, AI, LLMs, and GPT models, contributing to a platform that’s at the forefront of technological innovation.
Make a Real Impact: This is your chance to work on something that matters—to build solutions that directly improve the quality of life for thousands of people worldwide.
What you will do:
- Understand the process of CaaStle business teams, KPIs, and pain points
- Build scalable data products, self-service tools, data cubes to analyze and present data associated with acquisition, retention, product performance, operations, client services, etc.
- Closely partner with data engineering, product, and business teams and participate in requirements capture, research design, data collection, dashboard generation, and translation of results into actionable insights that can add value for business stakeholders
- Leverage advanced analytics to drive key success metrics for business and revenue generation
- Operationalize, implement, and automate changes to drive data-driven decisions
- Attend and play an active role in answering questions from the executive and/or business teams through data mining and analysis
We would love for you to have:
- Education: Advanced degree in Computer Science, Statistics, Mathematics, Engineering, Economics, Business Analytics or related field is required
- Experience: 2-4 years of professional experience
- Proficiency in data visualization/reporting tools (i.e. Tableau, Qlikview, etc.)
- Experience in A/B testing and measure performance of experiments
- Strong proficiency with SQL-based languages. Experience with large scale data analytics technologies (i.e., Hadoop and Spark)
- Strong analytical skills and business mindset with the ability to translate complex concepts and analysis into clear and concise takeaways to drive insights and strategies
- Excellent communication, social, and presentation skills with meticulous attention to detail
- Programming experience in Python, R, or other languages
- Knowledge of Data mining, statistical modeling approaches, and techniques
CaaStle is committed to equality of opportunity in employment. It has been and will continue to be the policy of CaaStle to provide full and equal employment opportunities to all employees and candidates for employment without regard to race, color, religion, national or ethnic origin, veteran status, age, sexual orientation, gender identity, or physical or mental disability. This policy applies to all terms, conditions and privileges of employment, such as those pertaining to training, transfer, promotion, compensation and recreational programs.
- 8+Years of Exp in C++ Developement
Software: C++, Jenkins, Visual Studio, Linux, CheckMarx, GitHub, Google test framework.
Application Architecture: Scalable, Resilient, Cloud deployable, high performance based.
DB: Oracle
Libraries: CPLEX (knowledge of CPLEX is a plus)

The Business Development Manager will be responsible for independently prospecting, identifying, developing, managing, and closing new business opportunities and achieve assigned revenue targets and profitability for the Technical Consulting/ Managed Services in the designated territory.
Job Responsibilities:
- Develop and execute effective sales strategies and activities for the defined territory.
- Identifying, qualifying, developing, and closing potential sales opportunities.
- Consistently achieve assigned revenue targets and other key result areas.
- Develop and maintain a long-term business relationship with new and existing accounts.
- Build and maintain an effective relationship with multi-level decision-making contacts.
- Develop and maintain knowledge of assigned solutions, target market segment, etc.
- Work collaboratively with internal and external resources to develop and grow the business.
- Maintain an up-to-date record of all opportunities and activities in the CRM system.
Required Qualifications and Experience:
- At least 4-8 years of experience selling Software Products/Services to enterprise customers.
- Bachelor’s or Master’s degree in Engineering, Business, Marketing or appropriate field.
- Must have a deep understanding of Technical Consulting/ Managed Services and market trends.
- Familiar with solution selling techniques and strong account management skills.
- Proven track record of generating new business and achieving revenue targets.
- Ability to communicate and influence people at all levels in an organization.
- Must have a good industry network and existing relationship with C-level contacts in Bangalore/South.
- Proficient in Word, Excel, PowerPoint, Outlook, and other related applications.
Personal Characteristics:
- Must be able to work in a fast-paced, goal-oriented, and collaborative work environment.
- Ability to multi-task and prioritize work independently with minimal supervision
- Must have excellent communication, organizational and relationship-building skills.
- Possess a high degree of self-motivation, initiative, integrity, discipline, and commitment.
- Must be highly adaptable, process-oriented, result-driven, quick learner, and team player.
2 Areas of Expertise:
- New Business Development B2B Sales Solution Selling Revenue Generation Account Management Relationship Management Strategic Planning Prospecting Demand Generation Pipeline Management Team-Work Negotiation & Closing Solution
- Knowledge: Managed Services Technical Consulting Staffing Services Application Development Business Process Outsourcing.
About the Role
The Dremio India team owns the DataLake Engine along with Cloud Infrastructure and services that power it. With focus on next generation data analytics supporting modern table formats like Iceberg, Deltalake, and open source initiatives such as Apache Arrow, Project Nessie and hybrid-cloud infrastructure, this team provides various opportunities to learn, deliver, and grow in career. We are looking for technical leaders with passion and experience in architecting and delivering high-quality distributed systems at massive scale.
Responsibilities & ownership
- Lead end-to-end delivery and customer success of next-generation features related to scalability, reliability, robustness, usability, security, and performance of the product
- Lead and mentor others about concurrency, parallelization to deliver scalability, performance and resource optimization in a multithreaded and distributed environment
- Propose and promote strategic company-wide tech investments taking care of business goals, customer requirements, and industry standards
- Lead the team to solve complex, unknown and ambiguous problems, and customer issues cutting across team and module boundaries with technical expertise, and influence others
- Review and influence designs of other team members
- Design and deliver architectures that run optimally on public clouds like GCP, AWS, and Azure
- Partner with other leaders to nurture innovation and engineering excellence in the team
- Drive priorities with others to facilitate timely accomplishments of business objectives
- Perform RCA of customer issues and drive investments to avoid similar issues in future
- Collaborate with Product Management, Support, and field teams to ensure that customers are successful with Dremio
- Proactively suggest learning opportunities about new technology and skills, and be a role model for constant learning and growth
Requirements
- B.S./M.S/Equivalent in Computer Science or a related technical field or equivalent experience
- Fluency in Java/C++ with 15+ years of experience developing production-level software
- Strong foundation in data structures, algorithms, multi-threaded and asynchronous programming models and their use in developing distributed and scalable systems
- 8+ years experience in developing complex and scalable distributed systems and delivering, deploying, and managing microservices successfully
- Subject Matter Expert in one or more of query processing or optimization, distributed systems, concurrency, micro service based architectures, data replication, networking, storage systems
- Experience in taking company-wide initiatives, convincing stakeholders, and delivering them
- Expert in solving complex, unknown and ambiguous problems spanning across teams and taking initiative in planning and delivering them with high quality
- Ability to anticipate and propose plan/design changes based on changing requirements
- Passion for quality, zero downtime upgrades, availability, resiliency, and uptime of the platform
- Passion for learning and delivering using latest technologies
- Hands-on experience of working projects on AWS, Azure, and GCP
- Experience with containers and Kubernetes for orchestration and container management in private and public clouds (AWS, Azure, and GCP)
- Understanding of distributed file systems such as S3, ADLS or HDFS
- Excellent communication skills and affinity for collaboration and teamwork
Qualification -
a. Minimum Bachelor's Degree
b. Minimum 3 yr experience
c. Basic English Proficiency
d. International Clients
Location: Jaipur
Interview will be online on Skype/Zoom.
Scala/Akka, Microservices, Akka streams, Play, Kafka, NoSQL. Java/J2EE, Spring, Architecture & Design, Enterprise integration ML/Data analytics, BigData technologies, Python/R programming, ML Algorithms/data models, ML tools(H2O/Tensorflow/Prediction.io), Realtime dashboards & batch, cluster compute (Spark/Akka), Distributed Data platforms (HDFS, Elastic search, Splunk, Casandra), Kafka with cross DC replication, NoSQL & In-memory DB







