Ab>initio, Big Data, Informatica, Tableau, Data Architect, Cognos, Microstrategy, Healther Business Analysts, Cloud etc.
at Exusia

About Exusia
About
Connect with the team
Similar jobs
Good communication skills
Good to have Hadoop and spark
Hands on with Apache kafka and confluence kafka
We are looking for a Business Intelligence (BI)/Data Analyst to create and manage Power Bl and analytics solutions that turn data into knowledge. In this role, you should have a background in data and business analysis. If you are self-directed, passionate about data,
and have business acumen and problem-solving aptitude, we'd like to meet you. Ultimately, you will enhance our business intelligence system to help us make better decisions.
Requirements and Qualifications
- BSc/BA in Computer Science, Engineering, or relevant field.
- Financial experience and Marketing background is a plus
- Strong Power BI development skills including Migration of existing deliverables to PowerBl.
- Ability to work autonomously
- Data modelling, Calculations, Conversions, Scheduling Data refreshes in Power-BI.
- Proven experience as a Power BI Developer is a must.
- Industry experience is preferred. Familiarity with other BI tools (Tableau, QlikView).
- Analytical mind with a problem-solving aptitude.
Responsibilities
- Design, develop and maintain business intelligence solutions
- Craft and execute queries upon request for data
- Present information through reports and visualization based on requirements gathered from stakeholders
- Interact with the team to gain an understanding of the business environment, technical context, and organizational strategic direction
- Design, build and deploy new, and extend existing dashboards and reports that synthesize distributed data sources
- Ensure data accuracy, performance, usability, and functionality requirements of BI platform
- Manage data through MS Excel, Google sheets, and SQL applications, as required and support other analytics platforms
- Develop and execute database queries and conduct analyses
- Develop and update technical documentation requirements
- Communicate insights to both technical and non-technical audiences.
make an impact by enabling innovation and growth; someone with passion for what they do and a vision for the future.
Responsibilities:
- Be the analytical expert in Kaleidofin, managing ambiguous problems by using data to execute sophisticated quantitative modeling and deliver actionable insights.
- Develop comprehensive skills including project management, business judgment, analytical problem solving and technical depth.
- Become an expert on data and trends, both internal and external to Kaleidofin.
- Communicate key state of the business metrics and develop dashboards to enable teams to understand business metrics independently.
- Collaborate with stakeholders across teams to drive data analysis for key business questions, communicate insights and drive the planning process with company executives.
- Automate scheduling and distribution of reports and support auditing and value realization.
- Partner with enterprise architects to define and ensure proposed.
- Business Intelligence solutions adhere to an enterprise reference architecture.
- Design robust data-centric solutions and architecture that incorporates technology and strong BI solutions to scale up and eliminate repetitive tasks
Requirements:
- Experience leading development efforts through all phases of SDLC.
- 5+ years "hands-on" experience designing Analytics and Business Intelligence solutions.
- Experience with Quicksight, PowerBI, Tableau and Qlik is a plus.
- Hands on experience in SQL, data management, and scripting (preferably Python).
- Strong data visualisation design skills, data modeling and inference skills.
- Hands-on and experience in managing small teams.
- Financial services experience preferred, but not mandatory.
- Strong knowledge of architectural principles, tools, frameworks, and best practices.
- Excellent communication and presentation skills to communicate and collaborate with all levels of the organisation.
- Team handling preferred for 5+yrs experience candidates.
- Notice period less than 30 days.
Job Responsibilities:
- Identify valuable data sources and automate collection processes
- Undertake preprocessing of structured and unstructured data.
- Analyze large amounts of information to discover trends and patterns
- Helping develop reports and analysis.
- Present information using data visualization techniques.
- Assessing tests and implementing new or upgraded software and assisting with strategic decisions on new systems.
- Evaluating changes and updates to source production systems.
- Develop, implement, and maintain leading-edge analytic systems, taking complicated problems and building simple frameworks
- Providing technical expertise in data storage structures, data mining, and data cleansing.
- Propose solutions and strategies to business challenges
Desired Skills and Experience:
- At least 1 year of experience in Data Analysis
- Complete understanding of Operations Research, Data Modelling, ML, and AI concepts.
- Knowledge of Python is mandatory, familiarity with MySQL, SQL, Scala, Java or C++ is an asset
- Experience using visualization tools (e.g. Jupyter Notebook) and data frameworks (e.g. Hadoop)
- Analytical mind and business acumen
- Strong math skills (e.g. statistics, algebra)
- Problem-solving aptitude
- Excellent communication and presentation skills.
- Bachelor’s / Master's Degree in Computer Science, Engineering, Data Science or other quantitative or relevant field is preferred
Mid / Senior Big Data Engineer
Job Description:
Role: Big Data EngineerNumber of open positions: 5Location: PuneAt Clairvoyant, we're building a thriving big data practice to help enterprises enable and accelerate the adoption of Big data and cloud services. In the big data space, we lead and serve as innovators, troubleshooters, and enablers. Big data practice at Clairvoyant, focuses on solving our customer's business problems by delivering products designed with best in class engineering practices and a commitment to keep the total cost of ownership to a minimum.
Must Have:
- 4-10 years of experience in software development.
- At least 2 years of relevant work experience on large scale Data applications.
- Strong coding experience in Java is mandatory
- Good aptitude, strong problem solving abilities, and analytical skills, ability to take ownership as appropriate
- Should be able to do coding, debugging, performance tuning and deploying the apps to Prod.
- Should have good working experience on
- o Hadoop ecosystem (HDFS, Hive, Yarn, File formats like Avro/Parquet)
- o Kafka
- o J2EE Frameworks (Spring/Hibernate/REST)
- o Spark Streaming or any other streaming technology.
- Strong coding experience in Java is mandatory
- Ability to work on the sprint stories to completion along with Unit test case coverage.
- Experience working in Agile Methodology
- Excellent communication and coordination skills
- Knowledgeable (and preferred hands on) - UNIX environments, different continuous integration tools.
- Must be able to integrate quickly into the team and work independently towards team goals
- Take the complete responsibility of the sprint stories' execution
- Be accountable for the delivery of the tasks in the defined timelines with good quality.
- Follow the processes for project execution and delivery.
- Follow agile methodology
- Work with the team lead closely and contribute to the smooth delivery of the project.
- Understand/define the architecture and discuss the pros-cons of the same with the team
- Involve in the brainstorming sessions and suggest improvements in the architecture/design.
- Work with other team leads to get the architecture/design reviewed.
- Work with the clients and counter-parts (in US) of the project.
- Keep all the stakeholders updated about the project/task status/risks/issues if there are any.
Experience: 4 to 9 years
Keywords: java, scala, spark, software development, hadoop, hive
Locations: Pune
Roles and Responsibilities:
- Review training content and assessments. The review parameters include (but are not limited to) the following:
○ Rewrite the training content wherever necessary to simplify it or to make it more age
appropriate.
○ Create engaging learning activities and provide exercises and activities that enhance
the learning process.
○ Devise modes of assessment, such as tests or quizzes, to measure the effectiveness
of course.
○ Analyze and apply trends and best practices in learning technologies and instructional
design.
○ Suggest instructional enhancements to the content and assessments so the learning
experience is superior.
○ Check if all the training assets follow the instructional guidelines.
○ Work with subject matter experts to identify target audience’s training needs
○ Do a thorough language review.
○ Do a thorough formatting check and fixes.
- Brainstorm with the content developer to storyboard the concepts introduced in a lesson/chapter.
- Create guidelines and templates as and when needed.
- Conduct audits of the live classes to check how the content is getting translated into the actual class experience.
- Give inputs for process improvements.
Skill Set:
- General:
- In-depth understanding of activity based teaching-learning methods with a strong knowledge of pedagogies to teach 21st Century Skills.
- At least 3-4 years of experience in instructional design and/or content writing.
- Experience in editing documents with an eye for details.
- Demonstrated ability to meet deadlines, handle and prioritize simultaneous requests, and manage laterally and upwards.
- Creative and analytical thinker with strong problem-solving skills.
- Excellent verbal and written communication skills.
○ Must demonstrate the ability to communicate effectively at all levels of the
organization.
- Technical:
- Understanding of Object-Oriented concepts.
- Basic understanding of core programming concepts with teaching skills will be an asset.
- Good exposure on-
- Object-Oriented Programming language(s) (preferably JavaScript/Python)
- Web Technology (HTML/CSS/JavaScript)- Front End Development
- Block-Based Coding environment like Code.org or Scratch will be a plus
- Knowledge of RDBMS (MySQL, Airtable, Firebase etc.) will be an added advantage
- We are looking for a Data Engineer to build the next-generation mobile applications for our world-class fintech product.
- The candidate will be responsible for expanding and optimising our data and data pipeline architecture, as well as optimising data flow and collection for cross-functional teams.
- The ideal candidate is an experienced data pipeline builder and data wrangler who enjoys optimising data systems and building them from the ground up.
- Looking for a person with a strong ability to analyse and provide valuable insights to the product and business team to solve daily business problems.
- You should be able to work in a high-volume environment, have outstanding planning and organisational skills.
Qualifications for Data Engineer
- Working SQL knowledge and experience working with relational databases, query authoring (SQL) as well as working familiarity with a variety of databases.
- Experience building and optimising ‘big data’ data pipelines, architectures, and data sets.
- Experience performing root cause analysis on internal and external data and processes to answer specific business questions and identify opportunities for improvement.
- Strong analytic skills related to working with unstructured datasets. Build processes supporting data transformation, data structures, metadata, dependency and workload management.
- Experience supporting and working with cross-functional teams in a dynamic environment.
- Looking for a candidate with 2-3 years of experience in a Data Engineer role, who is a CS graduate or has an equivalent experience.
What we're looking for?
- Experience with big data tools: Hadoop, Spark, Kafka and other alternate tools.
- Experience with relational SQL and NoSQL databases, including MySql/Postgres and Mongodb.
- Experience with data pipeline and workflow management tools: Luigi, Airflow.
- Experience with AWS cloud services: EC2, EMR, RDS, Redshift.
- Experience with stream-processing systems: Storm, Spark-Streaming.
- Experience with object-oriented/object function scripting languages: Python, Java, Scala.
• Participation in the requirements analysis, design, development and testing of applications.
• The candidate is expected to write code himself/herself.
• The candidate is expected to write high level code, code review, unit testing and deployment.
• Practical application of design principles with a focus on the user experience, usability, template
designs, cross browser issues and client server concepts.
• Contributes to the development of project estimates, scheduling, and deliverables.
• Works closely with QA team to determine testing requirements to ensure full coverage and best
quality of product.
• There is also the opportunity to mentor and guide junior team members in excelling their jobs.
Job Specifications
• BE/B. Tech. Computer Science or MCA from a reputed University.
• 6+ Years of experience in software development, with emphasis on JAVA/J2EE Server side
programming.
• Hands on experience in Core Java, Multithreading, RMI, Socket programing, JDBC, NIO,
webservices and Design patterns.
• Should have Knowledge of distributed system, distributed caching, messaging frameworks, ESB
etc.
• Knowledge of Linux operating system and PostgreSQL/MySQL/MongoDB/Cassandra database is
essential.
• Additionally, knowledge of HBase, Hadoop and Hive are desirable.
• Familiarity with message queue systems and AMQP and Kafka is desirable.
• Should have experience as a participant in Agile methodologies.
• Should have excellent written and verbal communication skills and presentation skills.
• This is not a Fullstack requirement, we are purely looking out for Backend resources
|
· Advanced Spark Programming Skills · Advanced Python Skills · Data Engineering ETL and ELT Skills · Expertise on Streaming data · Experience in Hadoop eco system · Basic understanding of Cloud Platforms · Technical Design Skills, Alternative approaches |
· Hands on expertise on writing UDF’s · Hands on expertise on streaming data ingestion · Be able to independently tune spark scripts · Advanced Debugging skills & Large Volume data handling. · Independently breakdown and plan technical Tasks |









