http://www.synapsica.com/">Synapsica is a https://yourstory.com/2021/06/funding-alert-synapsica-healthcare-ivycap-ventures-endiya-partners/">series-A funded HealthTech startup founded by alumni from IIT Kharagpur, AIIMS New Delhi, and IIM Ahmedabad. We believe healthcare needs to be transparent and objective while being affordable. Every patient has the right to know exactly what is happening in their bodies and they don't have to rely on cryptic 2 liners given to them as a diagnosis.
Towards this aim, we are building an artificial intelligence enabled cloud based platform to analyse medical images and create v2.0 of advanced radiology reporting. We are backed by IvyCap, Endia Partners, YCombinator and other investors from India, US, and Japan. We are proud to have GE and The Spinal Kinetics as our partners. Here’s a small sample of what we’re building: https://www.youtube.com/watch?v=FR6a94Tqqls">https://www.youtube.com/watch?v=FR6a94Tqqls
Your Roles and Responsibilities
We are looking for an experienced MLOps Engineer to join our engineering team and help us create dynamic software applications for our clients. In this role, you will be a key member of a team in decision making, implementations, development and advancement of ML operations of the core AI platform.
Roles and Responsibilities:
Work closely with a cross functional team to serve business goals and objectives.
Develop, Implement and Manage MLOps in cloud infrastructure for data preparation,deployment, monitoring and retraining models
Design and build application containerisation and orchestrate with Docker and Kubernetes in AWS platform.
Build and maintain code, tools, packages in cloud
Requirements:
At Least 2+ years of experience in Data engineering
At Least 3+ yr experience in Python with familiarity in popular ML libraries.
At Least 2+ years experience in model serving and pipelines
Working knowledge of containers like kubernetes , dockers, in AWS
Design distributed systems deployment at scale
Hands-on experience in coding and scripting
Ability to write effective scalable and modular code.
Familiarity with Git workflows, CI CD and NoSQL Mongodb
Familiarity with Airflow, DVC and MLflow is a plus
We are looking for an experienced Data Scientists to join our engineering team and
help us enhance our mobile application with data. In this role, we're looking for
people who are passionate about developing ML/AI in various domains that solves
enterprise problems. We are keen on hiring someone who loves working in fast paced start-up environment and looking to solve some challenging engineering
problems.
As one of the earliest members in engineering, you will have the flexibility to design
the models and architecture from ground up. As any early-stage start-up, we expect
you to be comfortable wearing various hats, and be proactive contributor in building
something truly remarkable.
Responsibilities
Researches, develops and maintains machine learning and statistical models for
business requirements
Work across the spectrum of statistical modelling including supervised,
unsupervised, & deep learning techniques to apply the right level of solution to
the right problem Coordinate with different functional teams to monitor outcomes and refine/
improve the machine learning models Implements models to uncover patterns and predictions creating business value and innovation
Identify unexplored data opportunities for the business to unlock and maximize
the potential of digital data within the organization
Develop NLP concepts and algorithms to classify and summarize structured/unstructured text data
Qualifications
3+ years of experience solving complex business problems using machine
learning.
Fluency in programming languages such as Python, NLP and Bert, is a must
Strong analytical and critical thinking skills
Experience in building production quality models using state-of-the-art technologies
Familiarity with databases like MySQL, Oracle, SQL Server, NoSQL, etc. is
desirable Ability to collaborate on projects and work independently when required.
Previous experience in Fintech/payments domain is a bonus
You should have Bachelor’s or Master’s degree in Computer Science, Statistics
or Mathematics or another quantitative field from a top tier Institute
Responsible to lead a team of analysts to build and deploy predictive models to infuse core business functions with deep analytical insights. The Senior Data Scientist will also work closely with the Kinara management team to investigate strategically important business questions.
Lead a team through the entire analytical and machine learning model life cycle: Define the problem statement Build and clean datasets Exploratory data analysis Feature engineering Apply ML algorithms and assess the performance Code for deployment Code testing and troubleshooting Communicate Analysis to Stakeholders Manage Data Analysts and Data Scientists
Job Title: Data Architect Job Location: Chennai Job Summary
The Engineering team is seeking a Data Architect. As a Data Architect, you will drive a Data Architecture strategy across various Data Lake platforms. You will help develop reference architecture and roadmaps to build highly available, scalable and distributed data platforms using cloud based solutions to process high volume, high velocity and wide variety of structured and unstructured data. This role is also responsible for driving innovation, prototyping, and recommending solutions. Above all, you will influence how users interact with Conde Nast’s industry-leading journalism. Primary Responsibilities Data Architect is responsible for • Demonstrated technology and personal leadership experience in architecting, designing, and building highly scalable solutions and products. • Enterprise scale expertise in data management best practices such as data integration, data security, data warehousing, metadata management and data quality. • Extensive knowledge and experience in architecting modern data integration frameworks, highly scalable distributed systems using open source and emerging data architecture designs/patterns. • Experience building external cloud (e.g. GCP, AWS) data applications and capabilities is highly desirable. • Expert ability to evaluate, prototype and recommend data solutions and vendor technologies and platforms. • Proven experience in relational, NoSQL, ELT/ETL technologies and in-memory databases. • Experience with DevOps, Continuous Integration and Continuous Delivery technologies is desirable. • This role requires 15+ years of data solution architecture, design and development delivery experience. • Solid experience in Agile methodologies (Kanban and SCRUM) Required Skills • Very Strong Experience in building Large Scale High Performance Data Platforms. • Passionate about technology and delivering solutions for difficult and intricate problems. Current on Relational Databases and No sql databases on cloud. • Proven leadership skills, demonstrated ability to mentor, influence and partner with cross teams to deliver scalable robust solutions.. • Mastery of relational database, NoSQL, ETL (such as Informatica, Datastage etc) /ELT and data integration technologies. • Experience in any one of Object Oriented Programming (Java, Scala, Python) and Spark. • Creative view of markets and technologies combined with a passion to create the future. • Knowledge on cloud based Distributed/Hybrid data-warehousing solutions and Data Lake knowledge is mandate. • Good understanding of emerging technologies and its applications. • Understanding of code versioning tools such as GitHub, SVN, CVS etc. • Understanding of Hadoop Architecture and Hive SQL • Knowledge in any one of the workflow orchestration • Understanding of Agile framework and delivery • Preferred Skills: ● Experience in AWS and EMR would be a plus ● Exposure in Workflow Orchestration like Airflow is a plus ● Exposure in any one of the NoSQL database would be a plus ● Experience in Databricks along with PySpark/Spark SQL would be a plus ● Experience with the Digital Media and Publishing domain would be a plus ● Understanding of Digital web events, ad streams, context models About Condé Nast CONDÉ NAST INDIA (DATA) Over the years, Condé Nast successfully expanded and diversified into digital, TV, and social platforms - in other words, a staggering amount of user data. Condé Nast made the right move to invest heavily in understanding this data and formed a whole new Data team entirely dedicated to data processing, engineering, analytics, and visualization. This team helps drive engagement, fuel process innovation, further content enrichment, and increase market revenue. The Data team aimed to create a company culture where data was the common language and facilitate an environment where insights shared in real-time could improve performance. The Global Data team operates out of Los Angeles, New York, Chennai, and London. The team at Condé Nast Chennai works extensively with data to amplify its brands' digital capabilities and boost online revenue. We are broadly divided into four groups, Data Intelligence, Data Engineering, Data Science, and Operations (including Product and Marketing Ops, Client Services) along with Data Strategy and monetization. The teams built capabilities and products to create data-driven solutions for better audience engagement. What we look forward to: We want to welcome bright, new minds into our midst and work together to create diverse forms of self-expression. At Condé Nast, we encourage the imaginative and celebrate the extraordinary. We are a media company for the future, with a remarkable past. We are Condé Nast, and It Starts Here.
Job Title: Senior Data Engineer/Analyst Location: Bengaluru Department: - Engineering
Bidgely is looking for extraordinary and dynamic Senior Data Analyst to be part of its core team in Bangalore. You must have delivered exceptionally high quality robust products dealing with large data. Be part of a highly energetic and innovative team that believes nothing is impossible with some creativity and hard work.
Responsibilities ● Design and implement a high volume data analytics pipeline in Looker for Bidgely flagship product. ● Implement data pipeline in Bidgely Data Lake ● Collaborate with product management and engineering teams to elicit & understand their requirements & challenges and develop potential solutions ● Stay current with the latest tools, technology ideas and methodologies; share knowledge by clearly articulating results and ideas to key decision makers.
Requirements ● 3-5 years of strong experience in data analytics and in developing data pipelines. ● Very good expertise in Looker ● Strong in data modeling, developing SQL queries and optimizing queries. ● Good knowledge of data warehouse (Amazon Redshift, BigQuery, Snowflake, Hive). ● Good understanding of Big data applications (Hadoop, Spark, Hive, Airflow, S3, Cloudera) ● Attention to details. Strong communication and collaboration skills. ● BS/MS in Computer Science or equivalent from premier institutes.
About Company Helical Insight an open source Business Intelligence tool from Helical IT Solutions Pvt. Ltd, based out of Hyderabad, is looking for fresher’s having strong knowledge on SQL. Helical Insight has more than 50+ clients from various sectors. It has been awarded the most promising company in the Business Intelligence space. We are looking for rockstar team mate to join our company. Job Brief We are looking for a Business Intelligence (BI) Developer to create and manage BI and analytics solutions that turn data into knowledge. In this role, you should have a background in data and business analysis. You should be analytical and an excellent communicator. If you also have a business acumen and problemsolving aptitude, we’d like to meet you. Excellent knowledge on SQLQuery is required. Basic knowledge on HTML CSS and JS is required. You would be working closely with customers of various domain to understand their data, understand their business requirement and deliver the required analytics in form of varous reports dashboards etc. Excellent client interfacing role with opportunity to work across various sectors and geographies as well as varioud kind of DB including NoSQL, RDBMS, graph db, Columnar DB etc Skill set and Qualification required Responsibilities Attending client calls to get requriement, show progress Translate business needs to technical specifications Design, build and deploy BI solutions (e.g. reporting tools) Maintain and support data analytics platforms) Conduct unit testing and troubleshooting Evaluate and improve existing BI systems Collaborate with teams to integrate systems Develop and execute database queries and conduct analyses Create visualizations and reports for requested projects Develop and update technical documentation Requirements Excellent expertise on SQLQueries Proven experience as a BI Developer or Data Scientist Background in data warehouse design (e.g. dimensional modeling) and data mining In-depth understanding of database management systems, online analytical processing (OLAP) and ETL (Extract, transform, load) framework Familiarity with BI technologies Proven abilities to take initiative and be innovative Analytical mind with a problem-solving aptitude BE in Computer Science/IT Education: BE/ BTech/ MCA/BCA/ MTech/ MS, or equivalent preferred. Interested candidates call us on +91 7569 765 162
Graphene is a Singapore Head quartered AI company which has been recognized as Singapore’s Best
Start Up By Switzerland’s Seedstarsworld, and also been awarded as best AI platform for healthcare in Vivatech Paris. Graphene India is also a member of the exclusive NASSCOM Deeptech club. We are developing an AI plaform which is disrupting and replacing traditional Market Research with unbiased insights with a focus on healthcare, consumer goods and financial services.
Graphene was founded by Corporate leaders from Microsoft and P&G, and works closely with the Singapore Government & Universities in creating cutting edge technology which is gaining traction with many Fortune 500 companies in India, Asia and USA.
Graphene’s culture is grounded in delivering customer delight by recruiting high potential talent and providing an intense learning and collaborative atmosphere, with many ex-employees now hired by large companies across the world.
Graphene has a 6-year track record of delivering financially sustainable growth and is one of the rare start-ups which is self-funded and is yet profitable and debt free. We have already created a strong bench strength of Singaporean leaders and are recruiting and grooming more talent with a focus on our US expansion.
Job title: - Data Analyst
Job Description
Data Analyst responsible for storage, data enrichment, data transformation, data gathering based on data requests, testing and maintaining data pipelines.
Responsibilities and Duties
Managing end to end data pipeline from data source to visualization layer
Ensure data integrity; Ability to pre-empt data errors
Organized managing and storage of data
Provide quality assurance of data, working with quality assurance analysts if necessary.
Commissioning and decommissioning of data sets.
Processing confidential data and information according to guidelines.
Helping develop reports and analysis.
Troubleshooting the reporting database environment and reports.
Managing and designing the reporting environment, including data sources, security, and metadata.
Supporting the data warehouse in identifying and revising reporting requirements.
Supporting initiatives for data integrity and normalization.
Evaluating changes and updates to source production systems.
Training end-users on new reports and dashboards.
Initiate data gathering based on data requirements
Analyse the raw data to check if the requirement is satisfied
Be an integral part of large scale client business development and delivery engagements
Develop the software and systems needed for end-to-end execution on large projects
Work across all phases of SDLC, and use Software Engineering principles to build scaled solutions
Build the knowledge base required to deliver increasingly complex technology projects
Object-oriented languages (e.g. Python, PySpark, Java, C#, C++ ) and frameworks (e.g. J2EE or .NET)
Database programming using any flavours of SQL
Expertise in relational and dimensional modelling, including big data technologies
Exposure across all the SDLC process, including testing and deployment
Expertise in Microsoft Azure is mandatory including components like Azure Data Factory, Azure Data Lake Storage, Azure SQL, Azure DataBricks, HD Insights, ML Service etc.
Good knowledge of Python and Spark are required
Good understanding of how to enable analytics using cloud technology and ML Ops
Experience in Azure Infrastructure and Azure Dev Ops will be a strong plus
Responsibilities: ● Ability to do exploratory analysis: Fetch data from systems and analyze trends. ● Developing customer segmentation models to improve the efficiency of marketing and product campaigns. ● Establishing mechanisms for cross functional teams to consume customer insights to improve engagement along the customer life cycle. ● Gather requirements for dashboards from business, marketing and operations stakeholders. ● Preparing internal reports for executive leadership and supporting their decision making. ● Analyse data, derive insights and embed it into Business actions. ● Work with cross functional teams. Skills Required • Data Analytics Visionary. • Strong in SQL & Excel and good to have experience in Tableau. • Experience in the field of Data Analysis, Data Visualization. • Strong in analysing the Data and creating dashboards. • Strong in communication, presentation and business intelligence. • Multi-Dimensional, "Growth Hacker" Skill Set with strong sense of ownership for work. • Aggressive “Take no prisoners” approach.
Data Scientist :
Pluto7 is a services and solutions company focused on building ML, Ai, Analytics, and IoT tailored solutions to accelerate business transformation.We are a Premier Google Cloud Partner, servicing Retail, Manufacturing, Healthcare, and Hi-Tech industries. We are a Google premium partner in AI & ML, which means you'll have the opportunity to work and collaborate with folks from Google.
Are you an innovator, have a passion to work with data and find insights, have the inquisitive mind with the constant yearning to learn new ideas; then we are looking for you.As a Pluto7 Data Scientist engineer, you will be one of the key members of our innovative artificial intelligence and machine learning team. You are expected to be unfazed with large volumes of data, love to apply various models, use technology to process and filter data for analysis.
Responsibilities:
Build and Optimize Machine Learning models.
Work with large/complex datasets to solve difficult and non-routine analysis problems, applying advanced analytical methods as needed.
Build and prototype data pipelines for analysis at scale.
Work cross-functionally with Business Analysts and Data Engineers to help develop cutting edge and innovative artificial intelligence and machine learning models.
Make recommendations for selections on machine learning models.
Drive accuracy levels to the next stage of the given ML models.
Experience in developing visualisation and User
Good exposure in exploratory data analysis
Strong experience in Statistics and ML algorithms.
Minimum qualifications:
2+ years of relevant work experience in ML and advanced data analytics(e.g., as a Machine Learning Specialist / Data scientist ).
Strong Experience using machine learning and artificial intelligence frameworks such as Tensorflow, sci-kit learn, Keras using python.
Good in Python/R/SAS programming.
Understanding of Cloud platforms like GCP, AWS, or other.
Preferred qualifications:
Work experience in building data pipelines to ingest, cleanse and transform data.
Applied experience with machine learning on large datasets and experience translating analysis results into business recommendations.
Demonstrated skills in selecting the right statistical tools given a data analysis problem.
Demonstrated effective written and verbal communication skills.
Demonstrated willingness to both teach others and learn new techniques
Work location : Bangalore
Read more
Get to hear about interesting companies hiring right now