
2) Design and Create AAS/OLAP/ Tabular cubes and automate processes for analytical needs. Expertise in developing OLAP cubes and developing complex calculations, Aggregations, implementing a dynamic security model using DAX functions in - Azure Analysis service
3) Writing optimized SQL queries for integration with other applications, Maintain data quality and overseeing database security, Partitions and Index. Extensively used performance monitor/SQL profiler/DMVs to solve deadlocks, to monitor long-running queries and trouble-shoot cubes SQL and T-SQL.
4) Designing database tables and structures, creating views functions and stored procedures.
5) Exposure to PowerBI will be an added advantage.

Similar jobs
Key Qualifications :
- At least 2 years of hands-on experience with cloud infrastructure on AWS or GCP
- Exposure to configuration management and orchestration tools at scale (e.g. Terraform, Ansible, Packer)
- Knowledge in DevOps tools (e.g. Jenkins, Groovy, and Gradle)
- Familiarity with monitoring and alerting tools(e.g. CloudWatch, ELK stack, Prometheus)
- Proven ability to work independently or as an integral member of a team
Preferable Skills :
- Familiarity with standard IT security practices such as encryption, credentials and key management
- Proven ability to acquire various coding languages (Java, Python- ) to support DevOps operation and cloud transformation
- Familiarity in web standards (e.g. REST APIs, web security mechanisms)
- Multi-cloud management experience with GCP / Azure
- Experience in performance tuning, services outage management and troubleshooting
- Delivering a complete front end application
- Ensuring high performance on mobile and desktop
- Writing tested, idiomatic, and documented JavaScript, HTML and CSS
- Coordinating the workflow between the graphic designer, the HTML coder, and yourself
- Cooperating with the back-end developer in the process of building the RESTful API
- Communicating with external web services
Skills and Competencies:
- Proficiency with JavaScript, TypeScript, HTML and CSS
- Fundamental Knowledge of an SPA (Single page application)
- Deep knowledge of AngularJS practices and commonly used modules based on extensive work experience
- Ability to provide SEO solutions for single page apps
- Thorough understanding of the responsibilities of the platform, database, API, caching layer, proxies, and other web services used in the system
- Validating user actions on the client side and providing responsive feedback
- Writing non-blocking, optimized, reusable and scalable code
- Creating custom, general use modules and components which extend the elements and modules of core Angular
- Experience with building the infrastructure for serving the front-end app and assets
- Architecting and automating the build process for production, using task runners or scripts
- Creating configuration, build, and test scripts for Continuous Integration environments
- Proficient understanding of code versioning tools such as git is required
- Excellent presentation, problem-solving and team working skills
Qualification and Experience:
- B.E. / B. Tech. / MCA / PGDCA / M.Sc. (CS) /M.E. / M. Tech
- At least 2 years of experience in Web Application Development with Angular/React
- Proven expertise in testing of application software to assure accuracy, integrity, completeness to achieve desired results
- Experience with AngularJS (v1.x.x) would be a value-add
Technical qualifications
- 2+ years of working experience in delivering web applications.
- Strong experience in Ruby on Rails, API design and architecture.
- Strong experience in databases like MongoDB/PostgreSQL/MySQL.
- Strong understanding of object-oriented programming(OOPs) concepts, data structures, algorithms, APIs, web services, SQL concepts.
- Experience in any JS frameworks like Express. js/Node. js/AngularJS/ReactJS.
- Understanding of version control tools - Git/SVN.
- Supportive in providing clean, maintainable code.
- Empathize with client needs and architect custom solutions.
- Brownie points for knowledge in Golang.
As a Data Engineer, your role will encompass:
- Designing and building production data pipelines from ingestion to consumption within a hybrid big data architecture using Scala, Python, Talend etc.
- Gather and address technical and design requirements.
- Refactor existing applications to optimize its performance through setting the appropriate architecture and integrating the best practices and standards.
- Participate in the entire data life-cycle mainly focusing on coding, debugging, and testing.
- Troubleshoot and debug ETL Pipelines.
- Documentation of each process.
Technical Requirements: -
- BSc degree in Computer Science/Computer Engineering. (Masters is a plus.)
- 2+ years of experience as a Data Engineer.
- In-depth understanding of core ETL concepts, Data Modelling, Data Lineage, Data Governance, Data Catalog, etc.
- 2+ years of work experience in Scala, Python, Java.
- Good Knowledge on Big Data Tools such as Spark/HDFS/Hive/Flume, etc.
- Hands on experience on ETL tools like Talend/Informatica is a plus.
- Good knowledge in Kafka and spark streaming is a big plus.
- 2+ years of experience in using Azure cloud and its resources/services (like Azure Data factory, Azure Databricks, SQL Synapse, Azure Devops, Logic Apps, Power Bi, Azure Event Hubs, etc).
- Strong experience in Relational Databases (MySQL, SQL Server)
- Exposure on data visualization tools like Power BI / Qlik sense / MicroStrategy
- 2+ years of experience in developing APIs (REST & SOAP protocols).
- Strong knowledge in Continuous Integration & Continuous Deployment (CI/CD) utilizing Docker containers, Jenkins, etc.
- Strong competencies in algorithms and software architecture.
- Excellent analytical and teamwork skills.
Good to have: -
- Previous on-prem working experience is a plus.
- In-depth understanding of the entire web development process (design, development, and deployment)
- Previous experience in automated testing including unit testing & UI testing.
Our client is more than 2 decades old NBFC registered with the RBI. The company was set up with an aim to empower rural entrepreneurs, SMEs, and small institutions by micro-financing. The company has come a long way since its inception in 1992 and has evolved with time with the introduction of its App in the small loans sector. At present, the company has more than 5000 employees and has serviced millions of borrowers to accomplish their dreams.
The group has diversified into the education segment and is running two schools. The schools are part of one of the largest private school chains in India. The schools have a flourishing family of 350 teachers and 6500 students. The leadership team at the helm of the company has demonstrable experience in financial services and has built the company by upholding the trust of its customers.
What you will do:
- Understanding the customer’s requirements and suggesting appropriate products (savings account, loan account etc.)
- Assessing prospective customer’s financial status (income, dependents etc.) and current trend of cash flows
- Suggesting appropriate products according to life-cycle needs and income
- Informing the prospective customer regarding application process, product terms and conditions etc
- Explaining to the prospective customer, the terms and conditions of product, application procedure, documents required and timelines for processing the application
- Responding to all queries and concerns regarding products and application process
- Discussing and setting revenue targets with the senior management
- Generating reports on targets achieved
Desired Candidate Profile
What you need to have:- 5 to 6 years of experience in sales and marketing
- MBA (Finance) or equivalent from Tier 1 college
- Prior experience in banking or other financial institutions
- Hands on experience in business correspondent operation
- Problem solver
Database management
Php
Plug-in customisation
API integration
Divi Builder
About us
DataWeave provides Retailers and Brands with “Competitive Intelligence as a Service” that enables them to take key decisions that impact their revenue. Powered by AI, we provide easily consumable and actionable competitive intelligence by aggregating and analyzing billions of publicly available data points on the Web to help businesses develop data-driven strategies and make smarter decisions.
Data Science@DataWeave
We the Data Science team at DataWeave (called Semantics internally) build the core machine learning backend and structured domain knowledge needed to deliver insights through our data products. Our underpinnings are: innovation, business awareness, long term thinking, and pushing the envelope. We are a fast paced labs within the org applying the latest research in Computer Vision, Natural Language Processing, and Deep Learning to hard problems in different domains.
How we work?
It's hard to tell what we love more, problems or solutions! Every day, we choose to address some of the hardest data problems that there are. We are in the business of making sense of messy public data on the web. At serious scale!
What do we offer?
● Some of the most challenging research problems in NLP and Computer Vision. Huge text and image
datasets that you can play with!
● Ability to see the impact of your work and the value you're adding to our customers almost immediately.
● Opportunity to work on different problems and explore a wide variety of tools to figure out what really
excites you.
● A culture of openness. Fun work environment. A flat hierarchy. Organization wide visibility. Flexible
working hours.
● Learning opportunities with courses and tech conferences. Mentorship from seniors in the team.
● Last but not the least, competitive salary packages and fast paced growth opportunities.
Who are we looking for?
The ideal candidate is a strong software developer or a researcher with experience building and shipping production grade data science applications at scale. Such a candidate has keen interest in liaising with the business and product teams to understand a business problem, and translate that into a data science problem.
You are also expected to develop capabilities that open up new business productization opportunities.
We are looking for someone with a Master's degree and 1+ years of experience working on problems in NLP or Computer Vision.
If you have 4+ years of relevant experience with a Master's degree (PhD preferred), you will be considered for a senior role.
Key problem areas
● Preprocessing and feature extraction noisy and unstructured data -- both text as well as images.
● Keyphrase extraction, sequence labeling, entity relationship mining from texts in different domains.
● Document clustering, attribute tagging, data normalization, classification, summarization, sentiment
analysis.
● Image based clustering and classification, segmentation, object detection, extracting text from images,
generative models, recommender systems.
● Ensemble approaches for all the above problems using multiple text and image based techniques.
Relevant set of skills
● Have a strong grasp of concepts in computer science, probability and statistics, linear algebra, calculus,
optimization, algorithms and complexity.
● Background in one or more of information retrieval, data mining, statistical techniques, natural language
processing, and computer vision.
● Excellent coding skills on multiple programming languages with experience building production grade
systems. Prior experience with Python is a bonus.
● Experience building and shipping machine learning models that solve real world engineering problems.
Prior experience with deep learning is a bonus.
● Experience building robust clustering and classification models on unstructured data (text, images, etc).
Experience working with Retail domain data is a bonus.
● Ability to process noisy and unstructured data to enrich it and extract meaningful relationships.
● Experience working with a variety of tools and libraries for machine learning and visualization, including
numpy, matplotlib, scikit-learn, Keras, PyTorch, Tensorflow.
● Use the command line like a pro. Be proficient in Git and other essential software development tools.
● Working knowledge of large-scale computational models such as MapReduce and Spark is a bonus.
● Be a self-starter—someone who thrives in fast paced environments with minimal ‘management’.
● It's a huge bonus if you have some personal projects (including open source contributions) that you work
on during your spare time. Show off some of your projects you have hosted on GitHub.
Role and responsibilities
● Understand the business problems we are solving. Build data science capability that align with our product strategy.
● Conduct research. Do experiments. Quickly build throw away prototypes to solve problems pertaining to the Retail domain.
● Build robust clustering and classification models in an iterative manner that can be used in production.
● Constantly think scale, think automation. Measure everything. Optimize proactively.
● Take end to end ownership of the projects you are working on. Work with minimal supervision.
● Help scale our delivery, customer success, and data quality teams with constant algorithmic improvements and automation.
● Take initiatives to build new capabilities. Develop business awareness. Explore productization opportunities.
● Be a tech thought leader. Add passion and vibrance to the team. Push the envelope. Be a mentor to junior members of the team.
● Stay on top of latest research in deep learning, NLP, Computer Vision, and other relevant areas.









