Cutshort logo

11+ QGIS Jobs in India

Apply to 11+ QGIS Jobs on CutShort.io. Find your next job, effortlessly. Browse QGIS Jobs and apply today!

icon
Blue Sky Analytics

at Blue Sky Analytics

3 recruiters
Balahun Khonglanoh
Posted by Balahun Khonglanoh
Remote only
1 - 5 yrs
Best in industry
NumPy
SciPy
skill iconData Science
skill iconPython
pandas
+8 more

About the Company

Blue Sky Analytics is a Climate Tech startup that combines the power of AI & Satellite data to aid in the creation of a global environmental data stack. Our funders include Beenext and Rainmatter. Over the next 12 months, we aim to expand to 10 environmental data-sets spanning water, land, heat, and more!


We are looking for a data scientist to join its growing team. This position will require you to think and act on the geospatial architecture and data needs (specifically geospatial data) of the company. This position is strategic and will also require you to collaborate closely with data engineers, data scientists, software developers and even colleagues from other business functions. Come save the planet with us!


Your Role

Manage: It goes without saying that you will be handling large amounts of image and location datasets. You will develop dataframes and automated pipelines of data from multiple sources. You are expected to know how to visualize them and use machine learning algorithms to be able to make predictions. You will be working across teams to get the job done.

Analyze: You will curate and analyze vast amounts of geospatial datasets like satellite imagery, elevation data, meteorological datasets, openstreetmaps, demographic data, socio-econometric data and topography to extract useful insights about the events happening on our planet.

Develop: You will be required to develop processes and tools to monitor and analyze data and its accuracy. You will develop innovative algorithms which will be useful in tracking global environmental problems like depleting water levels, illegal tree logging, and even tracking of oil-spills.

Demonstrate: A familiarity with working in geospatial libraries such as GDAL/Rasterio for reading/writing of data, and use of QGIS in making visualizations. This will also extend to using advanced statistical techniques and applying concepts like regression, properties of distribution, and conduct other statistical tests.

Produce: With all the hard work being put into data creation and management, it has to be used! You will be able to produce maps showing (but not limited to) spatial distribution of various kinds of data, including emission statistics and pollution hotspots. In addition, you will produce reports that contain maps, visualizations and other resources developed over the course of managing these datasets.

Requirements

These are must have skill-sets that we are looking for:

  • Excellent coding skills in Python (including deep familiarity with NumPy, SciPy, pandas).
  • Significant experience with git, GitHub, SQL, AWS (S3 and EC2).
  • Worked on GIS and is familiar with geospatial libraries such as GDAL and rasterio to read/write the data, a GIS software such as QGIS for visualisation and query, and basic machine learning algorithms to make predictions.
  • Demonstrable experience implementing efficient neural network models and deploying them in a production environment.
  • Knowledge of advanced statistical techniques and concepts (regression, properties of distributions, statistical tests and proper usage, etc.) and experience with applications.
  • Capable of writing clear and lucid reports and demystifying data for the rest of us.
  • Be curious and care about the planet!
  • Minimum 2 years of demonstrable industry experience working with large and noisy datasets.

Benefits

  • Work from anywhere: Work by the beach or from the mountains.
  • Open source at heart: We are building a community where you can use, contribute and collaborate on.
  • Own a slice of the pie: Possibility of becoming an owner by investing in ESOPs.
  • Flexible timings: Fit your work around your lifestyle.
  • Comprehensive health cover: Health cover for you and your dependents to keep you tension free.
  • Work Machine of choice: Buy a device and own it after completing a year at BSA.
  • Quarterly Retreats: Yes there's work-but then there's all the non-work+fun aspect aka the retreat!
  • Yearly vacations: Take time off to rest and get ready for the next big assignment by availing the paid leaves.
Read more
My client is a leading media agency
Gurugram
1 - 3 yrs
Best in industry
SQL
Microsoft Excel
skill iconPython
Tableau
PowerBI

About Company

Our client is the worlds largest media investment company and a global digital transformation agency with 1200 employees across 21 nations. They support clients in programmatic, social, paid search, analytics, technology, organic search, affiliate marketing, e-commerce and across traditional channels

 

About The Role

 

· Provide support on Campaign Measurement, Brand Lift studies, Reporting and other data needs (ingestion, cleaning, delivery etc.)

· Code in SQL, Python to create reports and build dashboards.

· Use Bigquery, Python, Datalab (GCP) and automate recurring workloads to improve efficiency

· Maintain underlying data infrastructure and troubleshoot data issues

· Generate meaningful insights from data using advanced analytics/statistical techniques and create dashboards and presentations for wider teams.

· Liaise closely with clients (internally and externally) on projects and work with key stakeholders to assess project needs and ensure best-in-class work is being delivered.

 

 

A bit about yourself

 

· 0-2 years of experience required

· Programming skills in SQL, Python/R

· Logical thinking and problem solving skills

· Strong communication skills (verbal & written)

· Adherence to quality & ability to plan & prioritize the work

· Passionate about data analysis, ability to troubleshoot data anomalies, pulling data from several sources, integrating it to create data layers

· Experience in applying big data, statistics, analytical techniques and technology to solve client problems will be plus

· Experience in creating dashboards, reports, presentations and finding insights, and strong business acumen to provide actionable strategic recommendations will be plus.

 

 


Read more
Expand My Business
Bengaluru (Bangalore), Mumbai, Chennai
7 - 15 yrs
₹10L - ₹28L / yr
Spark
Hadoop
Big Data
Data engineering
PySpark
+2 more

Design, implement, and improve the analytics platform

Implement and simplify self-service data query and analysis capabilities of the BI platform

Develop and improve the current BI architecture, emphasizing data security, data quality

and timeliness, scalability, and extensibility

Deploy and use various big data technologies and run pilots to design low latency

data architectures at scale

Collaborate with business analysts, data scientists, product managers, software development engineers,

and other BI teams to develop, implement, and validate KPIs, statistical analyses, data profiling, prediction,

forecasting, clustering, and machine learning algorithms


Educational

At Ganit we are building an elite team, ergo we are seeking candidates who possess the

following backgrounds:

7+ years relevant experience

Expert level skills writing and optimizing complex SQL

Knowledge of data warehousing concepts

Experience in data mining, profiling, and analysis

Experience with complex data modelling, ETL design, and using large databases

in a business environment

Proficiency with Linux command line and systems administration

Experience with languages like Python/Java/Scala

Experience with Big Data technologies such as Hive/Spark

Proven ability to develop unconventional solutions, sees opportunities to

innovate and leads the way

Good experience of working in cloud platforms like AWS, GCP & Azure. Having worked on

projects involving creation of data lake or data warehouse

Excellent verbal and written communication.

Proven interpersonal skills and ability to convey key insights from complex analyses in

summarized business terms. Ability to effectively communicate with multiple teams


Good to have

AWS/GCP/Azure Data Engineer Certification

Read more
Impact Guru

at Impact Guru

15 recruiters
shubham samel
Posted by shubham samel
Mumbai
3 - 7 yrs
₹5L - ₹9L / yr
Spotfire
Qlikview
Tableau
PowerBI
Data Visualization
+5 more

Job Responsibilities:


Excellent problem solving and analytical skills - ability to develop hypotheses,

understand and interpret data within the context of the product / business -

solve problems and distill data into actionable recommendations.


 Strong communication skills with the ability to confidently work with cross-

functional teams across the globe and to present information to all levels of the

organization.

 Intellectual and analytical curiosity - initiative to dig into the why, what & how.

 Strong number crunching and quantitative skills.

 Advanced knowledge of MS Excel and PowerPoint.

 Good hands on SQL

 Experience with in Google Analytics, Optimize, Tag Manager and other Google Suite tools

 Understanding of Business analytics tools & statistical programming languages - R, SAS, SPSS, Tableau is a plus

 Inherent interest in e-commerce & marketplace technology platforms and

broadly in the consumer Internet & mobile space.

 Previous experience of 1+ years working in a product company in

a product analytics role

 Strong understanding of building and interpreting product funnels.


Perquisites & Benefits:

 Opportunity to work with India's no.1 crowdfunding platform

 Be a part of a young, smart and rapidly growing team with management from Ivy League and Premier colleges

 Competitive compensation and incentives

 Fun, casual, relaxed and flexible work environment


Read more
Mobile Programming LLC

at Mobile Programming LLC

1 video
34 recruiters
Sukhdeep Singh
Posted by Sukhdeep Singh
Bengaluru (Bangalore)
4 - 6 yrs
₹10L - ₹15L / yr
ETL
Informatica
Data Warehouse (DWH)
Snow flake schema
Snowflake
+5 more

Job Title: AWS-Azure Data Engineer with Snowflake

Location: Bangalore, India

Experience: 4+ years

Budget: 15 to 20 LPA

Notice Period: Immediate joiners or less than 15 days

Job Description:

We are seeking an experienced AWS-Azure Data Engineer with expertise in Snowflake to join our team in Bangalore. As a Data Engineer, you will be responsible for designing, implementing, and maintaining data infrastructure and systems using AWS, Azure, and Snowflake. Your primary focus will be on developing scalable and efficient data pipelines, optimizing data storage and processing, and ensuring the availability and reliability of data for analysis and reporting.

Responsibilities:

  1. Design, develop, and maintain data pipelines on AWS and Azure to ingest, process, and transform data from various sources.
  2. Optimize data storage and processing using cloud-native services and technologies such as AWS S3, AWS Glue, Azure Data Lake Storage, Azure Data Factory, etc.
  3. Implement and manage data warehouse solutions using Snowflake, including schema design, query optimization, and performance tuning.
  4. Collaborate with cross-functional teams to understand data requirements and translate them into scalable and efficient technical solutions.
  5. Ensure data quality and integrity by implementing data validation, cleansing, and transformation processes.
  6. Develop and maintain ETL processes for data integration and migration between different data sources and platforms.
  7. Implement and enforce data governance and security practices, including access control, encryption, and compliance with regulations.
  8. Collaborate with data scientists and analysts to support their data needs and enable advanced analytics and machine learning initiatives.
  9. Monitor and troubleshoot data pipelines and systems to identify and resolve performance issues or data inconsistencies.
  10. Stay updated with the latest advancements in cloud technologies, data engineering best practices, and emerging trends in the industry.

Requirements:

  1. Bachelor's or Master's degree in Computer Science, Information Systems, or a related field.
  2. Minimum of 4 years of experience as a Data Engineer, with a focus on AWS, Azure, and Snowflake.
  3. Strong proficiency in data modelling, ETL development, and data integration.
  4. Expertise in cloud platforms such as AWS and Azure, including hands-on experience with data storage and processing services.
  5. In-depth knowledge of Snowflake, including schema design, SQL optimization, and performance tuning.
  6. Experience with scripting languages such as Python or Java for data manipulation and automation tasks.
  7. Familiarity with data governance principles and security best practices.
  8. Strong problem-solving skills and ability to work independently in a fast-paced environment.
  9. Excellent communication and interpersonal skills to collaborate effectively with cross-functional teams and stakeholders.
  10. Immediate joiner or notice period less than 15 days preferred.

If you possess the required skills and are passionate about leveraging AWS, Azure, and Snowflake to build scalable data solutions, we invite you to apply. Please submit your resume and a cover letter highlighting your relevant experience and achievements in the AWS, Azure, and Snowflake domains.

Read more
TransForm Solution

at TransForm Solution

2 recruiters
sowmiya Venkat
Posted by sowmiya Venkat
Surat
3 - 5 yrs
₹3L - ₹7L / yr
skill iconJavascript
skill iconPython
Selenium
Web Scraping
skill iconHTML/CSS
+1 more
  • Design, build web crawlers to scrape data and URLs.
  • Integrate the data crawled and scraped into our databases
  • Create more/better ways to crawl relevant information
  • Strong knowledge of web technologies (HTML, CSS, Javascript, XPath, Regex)
  • Understanding of data privacy policies (esp. GDPR) and personally identifiable information
  • Develop automated and reusable routines for extracting information from various data sources
  • Prepare requirement summary and re-confirm with Operation team
  • Translate business requirements into specific solutions
  • Ability to relay technical information to non-technical users
  • Demonstrate Effective problem solving and analytical skill
  • Ability to pay attention to detail, pro-active, critical thinking and accuracy is essential
  • Ability to work to deadlines and give realistic estimates

    

 

Skills & Expertise

  • 2+ years of web scraping experience
  • Experience with two or more of the following web scraping frameworks and tools: Selenium, Scrapy, Import.io, Webhose.io, ScrapingHub, ParseHub, Phantombuster, Octoparse, Puppeter, etc.
  • Basic knowledge of data engineering (database ingestion, ETL, etc.) 
  • Solution orientation and "can do" attitude - with a desire to tackle complex problems.
Read more
Saviance Technologies

at Saviance Technologies

1 recruiter
Shipra Agrawal
Posted by Shipra Agrawal
NCR (Delhi | Gurgaon | Noida)
3 - 5 yrs
₹7L - ₹9L / yr
PowerBI
power bi
Business Intelligence (BI)
DAX
Data modeling
+3 more

 

Job Title: Power BI Developer(Onsite)

Location: Park Centra, Sec 30, Gurgaon

CTC:        8 LPA

Time:       1:00 PM - 10:00 PM

  

Must Have Skills: 

  • Power BI Desktop Software
  • Dax Queries
  • Data modeling
  • Row-level security
  • Visualizations
  • Data Transformations and filtering
  • SSAS and SQL

 

Job description:

 

We are looking for a PBI Analytics Lead responsible for efficient Data Visualization/ DAX Queries and Data Modeling. The candidate will work on creating complex Power BI reports. He will be involved in creating complex M, Dax Queries and working on data modeling, Row-level security, Visualizations, Data Transformations, and filtering. He will be closely working with the client team to provide solutions and suggestions on Power BI.

 

Roles and Responsibilities:

 

  • Accurate, intuitive, and aesthetic Visual Display of Quantitative Information: We generate data, information, and insights through our business, product, brand, research, and talent teams. You would assist in transforming this data into visualizations that represent easy-to-consume visual summaries, dashboards and storyboards. Every graph tells a story.
  • Understanding Data: You would be performing and documenting data analysis, data validation, and data mapping/design. You would be mining large datasets to determine its characteristics and select appropriate visualizations.
  • Project Owner: You would develop, maintain, and manage advanced reporting, analytics, dashboards and other BI solutions, and would be continuously reviewing and improving existing systems and collaborating with teams to integrate new systems. You would also contribute to the overall data analytics strategy by knowledge sharing and mentoring end users.
  • Perform ongoing maintenance & production of Management dashboards, data flows, and automated reporting.
  • Manage upstream and downstream impact of all changes on automated reporting/dashboards
  • Independently apply problem-solving ability to identify meaningful insights to business
  • Identify automation opportunities and work with a wide range of stakeholders to implement the same.
  • The ability and self-confidence to work independently and increase the scope of the service line

 

Requirements: 

  • 3+ years of work experience as an Analytics Lead / Senior Analyst / Sr. PBI Developer.
  • Sound understanding and knowledge of PBI Visualization and Data Modeling with DAX queries
  • Experience in leading and mentoring a small team.

 

 

 

Read more
Alien Brains

at Alien Brains

5 recruiters
Praveen Baheti
Posted by Praveen Baheti
Kolkata
2 - 7 yrs
₹5L - ₹8L / yr
skill iconJava
skill iconPython
skill iconJavascript
skill iconMachine Learning (ML)
skill iconDeep Learning
+2 more
A research lab with roots laid in innovation, we are looking for someone who can take reins of our AI based development think tank. Given the work ethics and results, the salary can be re-negotiated in 5 months.
Read more
Japan Based Leading Company
Bengaluru (Bangalore)
3 - 10 yrs
₹0L - ₹20L / yr
Big Data
skill iconAmazon Web Services (AWS)
skill iconJava
skill iconPython
MySQL
+2 more
A data engineer with AWS Cloud infrastructure experience to join our Big Data Operations team. This role will provide advanced operations support, contribute to automation and system improvements, and work directly with enterprise customers to provide excellent customer service.
The candidate,
1. Must have a very good hands-on technical experience of 3+ years with JAVA or Python
2. Working experience and good understanding of AWS Cloud; Advanced experience with IAM policy and role management
3. Infrastructure Operations: 5+ years supporting systems infrastructure operations, upgrades, deployments using Terraform, and monitoring
4. Hadoop: Experience with Hadoop (Hive, Spark, Sqoop) and / or AWS EMR
5. Knowledge on PostgreSQL/MySQL/Dynamo DB backend operations
6. DevOps: Experience with DevOps automation - Orchestration/Configuration Management and CI/CD tools (Jenkins)
7. Version Control: Working experience with one or more version control platforms like GitHub or GitLab
8. Knowledge on AWS Quick sight reporting
9. Monitoring: Hands on experience with monitoring tools such as AWS CloudWatch, AWS CloudTrail, Datadog and Elastic Search
10. Networking: Working knowledge of TCP/IP networking, SMTP, HTTP, load-balancers (ELB) and high availability architecture
11. Security: Experience implementing role-based security, including AD integration, security policies, and auditing in a Linux/Hadoop/AWS environment. Familiar with penetration testing and scan tools for remediation of security vulnerabilities.
12. Demonstrated successful experience learning new technologies quickly
WHAT WILL BE THE ROLES AND RESPONSIBILITIES?
1. Create procedures/run books for operational and security aspects of AWS platform
2. Improve AWS infrastructure by developing and enhancing automation methods
3. Provide advanced business and engineering support services to end users
4. Lead other admins and platform engineers through design and implementation decisions to achieve balance between strategic design and tactical needs
5. Research and deploy new tools and frameworks to build a sustainable big data platform
6. Assist with creating programs for training and onboarding for new end users
7. Lead Agile/Kanban workflows and team process work
8. Troubleshoot issues to resolve problems
9. Provide status updates to Operations product owner and stakeholders
10. Track all details in the issue tracking system (JIRA)
11. Provide issue review and triage problems for new service/support requests
12. Use DevOps automation tools, including Jenkins build jobs
13. Fulfil any ad-hoc data or report request queries from different functional groups
Read more
Anaxee Technologies
Indore
0 - 1 yrs
₹1L - ₹2L / yr
skill iconData Science
Internship
skill iconData Analytics
Data Structures

Job role:

As a data analyst, you will be responsible for compiling actionable insights from data and assisting program, sales and marketing managers build data-driven processes. Your role will involve driving initiatives to optimize for operational excellence and revenue.


Job Location: Indore | Full-Time Internship  | Stipend - Performance Based  | 


About the company:

Anaxee Digital Runners is building India's largest last-mile verification & data collection network of Digital Runners (shared feet-on-street, tech-enabled) to help Businesses & Consumers reach remotest parts of India, on-demand.  KYC | Field Verification | Data Collection | eSign | Tier-2, 3 & 4

Sounds like a moonshot? It is. We want to make REACH across India (remotest places), as easy as ordering pizza, on-demand. Already serving 11000 pin codes (57% of India) | Website: www.anaxee.com

Important: https://www.youtube.com/watch?v=7QnyJsKedz8" target="_blank">Check out our company pitch (6 min video) to understand this goal - https://www.youtube.com/watch?v=7QnyJsKedz8" target="_blank">https://www.youtube.com/watch?v=7QnyJsKedz8


Responsibilities:

  • Ensure that data flows smoothly from source to destination so that it can be processed
  • Utilize strong database skills to work with large, complex data sets to extract insights
  • Filter and cleanse unstructured (or ambiguous) data into usable data sets that can be analyzed to extract insights and improve business processes
  • Identify new internal and external data sources to support analytics initiatives and work with appropriate partners to absorb the data into new or existing data infrastructure
  • Build tools for automating repetitive tasks so that bandwidth can be freed for analytics
  • Collaborate with program managers and business analysts  to help them come up with actionable, high-impact insights across product lines and functions
  • Work closely with top management to prioritize information and analytic needs


Requirements:

  • Bachelors or Masters (Pursuing or Graduated) in a quantitative field (such as Engineering, Statistics, Math, Economics, or Computer Science with Modeling/Data Science), preferably with work experience of over [X] years.
  • Ability to program in any high-level language is required. Familiarity with R and statistical packages are preferred.
  • Proven problem solving and debugging skills.
  • Familiar with database technologies and tools (SQL/R/SAS/JMP etc.), data warehousing, transformation, and processing. Work experience with real data for customer insights, business, and market analysis will be advantageous.
  • Experience with text analytics, data mining and social media analytics.
  • Statistical knowledge in standard techniques: Logistic Regression, Classification models, Cluster Analysis, Neural Networks, Random Forests, Ensembles, etc.
Read more
Woodcutter Film Technologies Pvt. Ltd.
Athul Krishnan
Posted by Athul Krishnan
Hyderabad
1 - 5 yrs
₹3L - ₹6L / yr
skill iconData Science
skill iconR Programming
skill iconPython
We're an early stage film-tech startup with a mission to empower filmmakers and independent content creators with data-driven decision-making tools. We're looking for a data person to join the core team. Please get in touch if you would be excited to join us on this super exciting journey of disrupting the film production and distribution business. We are currently collaborating with Rana Daggubatt's Suresh Productions, and work out of their studio in Hyderabad - so exposure and opportunities to work on real issues faced by the media industry will be in plenty.
Read more
Get to hear about interesting companies hiring right now
Company logo
Company logo
Company logo
Company logo
Company logo
Linkedin iconFollow Cutshort
Why apply via Cutshort?
Connect with actual hiring teams and get their fast response. No spam.
Find more jobs
Get to hear about interesting companies hiring right now
Company logo
Company logo
Company logo
Company logo
Company logo
Linkedin iconFollow Cutshort