

Must Have Skills:
• Good experience in Pyspark - Including Dataframe core functions and Spark SQL
• Good experience in SQL DBs - Be able to write queries including fair complexity.
• Should have excellent experience in Big Data programming for data transformation and aggregations
• Good at ELT architecture. Business rules processing and data extraction from Data Lake into data streams for business consumption.
• Good customer communication.
• Good Analytical skills

Similar jobs
Job Description:
- Extensive experience in Appian BPM application development
- Knowledge of Appian architecture and its objects best practices
- Participate in analysis, design, and new development of Appian based applications
- Team leadership and provide technical leadership to Scrum teams
- Must be able to multi-task, work in a fast-paced environment and ability to resolve problems faced by team
- Build applications: interfaces, process flows, expressions, data types, sites, integrations,
- Proficient with SQL queries and with accessing data present in DB tables and views
- Experience in Analysis, Designing process models, Records, Reports, SAIL, forms, gateways, smart services, integration services and web services
- Experience working with different Appian Object types, query rules, constant rules and expression rules
Qualifications
At least 6 years of experience in Implementing BPM solutions using Appian 19.x or higher
- Over 8 years in Implementing IT solutions using BPM or integration technologies
- Certification Mandatory- L1 and L2 a
- Experience in Scrum/Agile methodologies with Enterprise level application development projects
- Good understanding of database concepts and strong working knowledge any one of the major databases e g Oracle SQL Server MySQL
- Appian BPM application development on version 19.x or higher
- Experience of integrations using web services e g XML REST WSDL SOAP API JDBC JMS
- Good leadership skills and the ability to lead a team of software engineers technically
- Experience working in Agile Scrum teams
- Good Communication skills
1.Experience with CRMNext or similar CRM platforms
2.Familiarity with SDLC and Agile methodologies
3.Ability to analyze business requirements and deliver effective solutions
4.Proficiency in Java or .NET
5.Experience with API integrations, ESB, and Node.js
6.Strong knowledge of SQL and Oracle databases
7.Understanding of system architecture, operating systems, and data structures
8.Excellent communication skills and a collaborative approach
9.Accountability for end-to-end project delivery Preferred Profile
Candidates with a Java development background and experience in CRM applications will be given preference.
- Maintaining cleanliness and organization within office premises
- Delivering and retrieving documents as required by staff
- Handling office maintenance tasks such as restocking supplies
- Assisting with office clerical duties including filing and photocopying
- Preparing and serving beverages to office staff and visitors
- Managing the disposal of waste and recycling practices
- Supporting staff with various tasks and errands as needed


We are seeking a highly skilled and experienced MERN (MongoDB, Express.js, React, Node.js) Stack Developer to join our dynamic development team. As a MERN Stack Developer, you will be responsible for designing, developing, and maintaining scalable web applications using the MERN stack framework.
- Strong proficiency in JavaScript, including DOM manipulation and the JavaScript object model.
- Expertise in backend programming with Node.js and MongoDB.
- Experience with React.js and redux.
- Material UI and 3rd party libraries.
- Experience with clean code writing practices like avoiding callback hell like promises, async.
- Thorough understanding of Node.js and its core principles.
- Experience with popular React.js workflows (such as Flux or Redux).
- Familiarity with newer specifications of ECMAScript.
- Experience with data structure libraries (e.g., Immutable.js).
- Familiarity with RESTful APIs.
- Knowledge of modern authorization mechanisms, such as JSON Web Token.
- Familiarity with modern front-end builds pipelines and tools.
- A knack for benchmarking and optimization.
- Familiarity with code versioning tools (such as Git, SVN, and Mercurial).
Responsibilities:
- Build Node.js APIs using microservices.
- Rewriting backend code with microservices architecture & Unit tests.
- Developing new user-facing features using React.js.
- Building reusable components and front-end libraries for future use.
- Translating designs and wireframes into high-quality code.
- Optimizing components for maximum performance across a vast array of web-capable devices and browsers.
If you are a passionate MERN Stack Developer looking for an exciting opportunity to work on challenging projects and be part of a talented team, we would love to hear from you. Please submit your application, including your resume and portfolio, highlighting your relevant experience and projects.
Role: Project Manager
Experience: 8-10 Years
Location: Mumbai
Company Profile:
Exponentia.ai is an AI tech organization with a presence across India, Singapore, the Middle East, and the UK. We are an innovative and disruptive organization, working on cutting-edge technology to help our clients transform into the enterprises of the future. We provide artificial intelligence-based products/platforms capable of automated cognitive decision-making to improve productivity, quality, and economics of the underlying business processes. Currently, we are rapidly expanding across machine learning, Data Engineering and Analytics functions. Exponentia.ai has developed long-term relationships with world-class clients such as PayPal, PayU, SBI Group, HDFC Life, Kotak Securities, Wockhardt and Adani Group amongst others.
One of the top partners of Data bricks, Azure, Cloudera (leading analytics player) and Qlik (leader in BI technologies), Exponentia.ai has recently been awarded the ‘Innovation Partner Award’ by Qlik and "Excellence in Business Process Automation Award" (IMEA) by Automation Anywhere.
Get to know more about us at http://www.exponentia.ai and https://in.linkedin.com/company/exponentiaai
Role Overview:
· Project manager shall be responsible to oversee and take responsibility for the successful delivery of a range of projects in Business Intelligence, Data warehousing, and Analytics/AI-ML.
· Project manager is expected to manage the project and lead the teams of BI engineers, data engineers, data scientists and application developers.
Job Responsibilities:
· Efforts estimation, creating a project plan, planning milestones, activities and tracking the progress.
· Identify risks and issues. Come up with a mitigation plan.
· Status reporting to both internal and external stakeholders.
· Communicate with all stakeholders.
· Manage end-to-end project lifecycle - requirements gathering, design, development, testing and go-live.
· Manage end-to-end BI or data warehouse projects.
· Must have experience in running Agile-based project development.
Technical skills
· Experience in Business Intelligence Data warehousing or Analytics projects.
· Understand data lake and data warehouse solutions including ETL pipelines.
· Good to have - Knowledge of Azure blob storage, azure data factory and Synapse analytics.
· Good to have - Knowledge of Qlik Sense or Power BI
· Good to have - Certified in PMP/Prince 2 / Agile Project management.
· Excellent written and verbal communication skills.
Education:
MBA, B.E. or B. Tech. or MCA degree
- Study design briefs and determine requirements
- Conceptualizing visuals based on requirements
- Prepare rough drafts and present ideas
- Develop illustrations, logos and other designs using software
- Use the appropriate colors and layouts for each graphic
- Work with copywriters and the creative director to produce the final design
- Test graphics across various media
- Amend designs after feedback
- Ensure final graphics and layouts are visually appealing and on-brand
- Knowledge of basic animation principles
Requirements and skills:
- Proven graphic designing experience
- A strong portfolio of illustrations or other graphics
- Familiarity with design software and technologies - Illustrator, Photoshop & After Effects
- A keen eye for aesthetics and details
- Excellent communication skills
- Ability to work methodically and meet deadlines
- Degree in Design, Fine Arts or a related field is a plus
Skin and hair problems give many people stress, make them self-conscious and impact their confidence. Re’equil offers reliable and effective personal care formulations so that people look good in their own skin and feel even better.
Re'equil is a direct-to-consumer (D2C) beauty and personal care brand. Founded in 2018, we are a self-funded profitable company scaling new heights. We are currently a team of 30 based out of Chandigarh.
Joining: Immediate - 1 month
Qualification: BCA/MCA/B. Tech (CS/ECE)
- Developing eCommerce sites on the Shopify platform. Shopify theme development and customization, private apps, Shopify Liquid templating language, Shopify JS APIs (storefront, AJAX Cart, Sections etc.), HTML5, CSS3, JavaScript, jQuery & Rest API's)
- Responsible for the development of the website on Shopify as per the design handed off by the UI Designers. Knowledge of Installing & Customizing the new Shopify theme.
- Deliver optimized, fast loading, and super responsive website. Understand the requirements from briefs, do thorough research, and based on previous experiences -propose the most suitable solution platform with an ETA.
- Extending Shopify functionalities to the next level using storefront APIs, liquid programming, meta fields, etc.
- Implementing SEO/CRO best practices in Shopify sites Creating responsive website designs
- Integrating third-party and platform-supported apps in the sites.

Your mission is to help lead team towards creating solutions that improve the way our business is run. Your knowledge of design, development, coding, testing and application programming will help your team raise their game, meeting your standards, as well as satisfying both business and functional requirements. Your expertise in various technology domains will be counted on to set strategic direction and solve complex and mission critical problems, internally and externally. Your quest to embracing leading-edge technologies and methodologies inspires your team to follow suit.
Responsibilities and Duties :
- As a Data Engineer you will be responsible for the development of data pipelines for numerous applications handling all kinds of data like structured, semi-structured &
unstructured. Having big data knowledge specially in Spark & Hive is highly preferred.
- Work in team and provide proactive technical oversight, advice development teams fostering re-use, design for scale, stability, and operational efficiency of data/analytical solutions
Education level :
- Bachelor's degree in Computer Science or equivalent
Experience :
- Minimum 3+ years relevant experience working on production grade projects experience in hands on, end to end software development
- Expertise in application, data and infrastructure architecture disciplines
- Expert designing data integrations using ETL and other data integration patterns
- Advanced knowledge of architecture, design and business processes
Proficiency in :
- Modern programming languages like Java, Python, Scala
- Big Data technologies Hadoop, Spark, HIVE, Kafka
- Writing decently optimized SQL queries
- Orchestration and deployment tools like Airflow & Jenkins for CI/CD (Optional)
- Responsible for design and development of integration solutions with Hadoop/HDFS, Real-Time Systems, Data Warehouses, and Analytics solutions
- Knowledge of system development lifecycle methodologies, such as waterfall and AGILE.
- An understanding of data architecture and modeling practices and concepts including entity-relationship diagrams, normalization, abstraction, denormalization, dimensional
modeling, and Meta data modeling practices.
- Experience generating physical data models and the associated DDL from logical data models.
- Experience developing data models for operational, transactional, and operational reporting, including the development of or interfacing with data analysis, data mapping,
and data rationalization artifacts.
- Experience enforcing data modeling standards and procedures.
- Knowledge of web technologies, application programming languages, OLTP/OLAP technologies, data strategy disciplines, relational databases, data warehouse development and Big Data solutions.
- Ability to work collaboratively in teams and develop meaningful relationships to achieve common goals
Skills :
Must Know :
- Core big-data concepts
- Spark - PySpark/Scala
- Data integration tool like Pentaho, Nifi, SSIS, etc (at least 1)
- Handling of various file formats
- Cloud platform - AWS/Azure/GCP
- Orchestration tool - Airflow



