
About Teradata
About
Connect with the team
Similar jobs
Company name: PulseData labs Pvt Ltd (captive Unit for URUS, USA)
About URUS
We are the URUS family (US), a global leader in products and services for Agritech.
SENIOR DATA ENGINEER
This role is responsible for the design, development, and maintenance of data integration and reporting solutions. The ideal candidate will possess expertise in Databricks and strong skills in SQL Server, SSIS and SSRS, and experience with other modern data engineering tools such as Azure Data Factory. This position requires a proactive and results-oriented individual with a passion for data and a strong understanding of data warehousing principles.
Responsibilities
Data Integration
- Design, develop, and maintain robust and efficient ETL pipelines and processes on Databricks.
- Troubleshoot and resolve Databricks pipeline errors and performance issues.
- Maintain legacy SSIS packages for ETL processes.
- Troubleshoot and resolve SSIS package errors and performance issues.
- Optimize data flow performance and minimize data latency.
- Implement data quality checks and validations within ETL processes.
Databricks Development
- Develop and maintain Databricks pipelines and datasets using Python, Spark and SQL.
- Migrate legacy SSIS packages to Databricks pipelines.
- Optimize Databricks jobs for performance and cost-effectiveness.
- Integrate Databricks with other data sources and systems.
- Participate in the design and implementation of data lake architectures.
Data Warehousing
- Participate in the design and implementation of data warehousing solutions.
- Support data quality initiatives and implement data cleansing procedures.
Reporting and Analytics
- Collaborate with business users to understand data requirements for department driven reporting needs.
- Maintain existing library of complex SSRS reports, dashboards, and visualizations.
- Troubleshoot and resolve SSRS report issues, including performance bottlenecks and data inconsistencies.
Collaboration and Communication
- Comfortable in entrepreneurial, self-starting, and fast-paced environment, working both independently and with our highly skilled teams.
- Collaborate effectively with business users, data analysts, and other IT teams.
- Communicate technical information clearly and concisely, both verbally and in writing.
- Document all development work and procedures thoroughly.
Continuous Growth
- Keep abreast of the latest advancements in data integration, reporting, and data engineering technologies.
- Continuously improve skills and knowledge through training and self-learning.
This job description reflects managements assignment of essential functions; it does not prescribe or restrict the tasks that may be assigned.
Requirements
- Bachelor's degree in computer science, Information Systems, or a related field.
- 7+ years of experience in data integration and reporting.
- Extensive experience with Databricks, including Python, Spark, and Delta Lake.
- Strong proficiency in SQL Server, including T-SQL, stored procedures, and functions.
- Experience with SSIS (SQL Server Integration Services) development and maintenance.
- Experience with SSRS (SQL Server Reporting Services) report design and development.
- Experience with data warehousing concepts and best practices.
- Experience with Microsoft Azure cloud platform and Microsoft Fabric desirable.
- Strong analytical and problem-solving skills.
- Excellent communication and interpersonal skills.
- Ability to work independently and as part of a team.
- Experience with Agile methodologies.
Key Responsibilities:
- Work with distributed systems and implement asynchronous programming patterns
- Design and develop scalable backend applications using Python
- Build and integrate applications leveraging LLMs or traditional Machine Learning techniques
- Develop and maintain microservices-based architectures
- Work with databases and caching systems to optimize application performance
- Participate in code reviews and maintain high code quality standards
- Write clean, maintainable, and well-documented code following best practices
Required Skills:
- 3+ years of relevant experience
- Strong understanding of distributed systems and asynchronous programming in Python
- Experience building scalable applications using LLMs or traditional ML techniques
- Hands-on experience with databases, caching mechanisms, and microservices architecture
- Good problem-solving and debugging skills
We are a boutique IT services & solutions firm headquartered in the Bay Area with offices in India. Our offering includes custom-configured hybrid cloud solutions backed by our managed services. We combine best in class DevOps and IT infrastructure management practices, to manage our clients Hybrid Cloud Environments.
In addition, we build and deploy our private cloud solutions using Open stack to provide our clients with a secure, cost effective and scale able Hybrid Cloud solution. We work with start-ups as well as enterprise clients.
This is an exciting opportunity for an experienced Cloud Engineer to work on exciting projects and have an opportunity to expand their knowledge working on adjacent technologies as well.
Must have skills
• Provisioning skills on IaaS cloud computing for platforms such as AWS, Azure, GCP.
• Strong working experience in AWS space with various AWS services and implementations (i.e. VPCs, SES, EC2, S3, Route 53, Cloud Front, etc.)
• Ability to design solutions based on client requirements.
• Some experience with various network LAN/WAN appliances like (Cisco routers and ASA systems, Barracuda, Meraki, SilverPeak, Palo Alto, Fortinet, etc.)
• Understanding of networked storage like (NFS / SMB / iSCSI / Storage GW / Windows Offline)
• Linux / Windows server installation, maintenance, monitoring, data backup and recovery, security, and administration.
• Good knowledge of TCP/IP protocol & internet technologies.
• Passion for innovation and problem solving, in a start-up environment.
• Good communication skills.
Good to have
• Remote Monitoring & Management.
• Familiarity with Kubernetes and Containers.
• Exposure to DevOps automation scripts & experience with tools like Git, bash scripting, PowerShell, AWS Cloud Formation, Ansible, Chef or Puppet will be a plus.
• Architect / Practitioner certification from OEM with hands-on capabilities.
What you will be working on
• Trouble shoot and handle L2/ L3 tickets.
• Design and architect Enterprise Cloud systems and services.
• Design, Build and Maintain environments primarily in AWS using EC2, S3/Storage, CloudFront, VPC, ELB, Auto Scaling, Direct Connect, Route53, Firewall, etc.
• Build and deploy in GCP/ Azure as needed.
• Architect cloud solutions keeping performance, cost and BCP considerations in mind.
• Plan cloud migration projects as needed.
• Collaborate & work as part of a cohesive team.
• Help build our private cloud offering on Open stack.
Objectives :
- Building and setting up new development tools and infrastructure
- Working on ways to automate and improve development and release processes
- Testing code written by others and analyzing results
- Ensuring that systems are safe and secure against cybersecurity threats
- Identifying technical problems and developing software updates and ‘fixes’
- Working with software developers and software engineers to ensure that development follows established processes and works as intended
- Planning out projects and being involved in project management decisions
Daily and Monthly Responsibilities :
- Deploy updates and fixes
- Build tools to reduce occurrences of errors and improve customer experience
- Develop software to integrate with internal back-end systems
- Perform root cause analysis for production errors
- Investigate and resolve technical issues
- Develop scripts to automate visualization
- Design procedures for system troubleshooting and maintenance
Skills and Qualifications :
- Degree in Computer Science or Software Engineering or BSc in Computer Science, Engineering or relevant field
- 3+ years of experience as a DevOps Engineer or similar software engineering role
- Proficient with git and git workflows
- Good logical skills and knowledge of programming concepts(OOPS,Data Structures)
- Working knowledge of databases and SQL
- Problem-solving attitude
- Collaborative team spirit
Company Overview:
At Codvo, software and people transformations go hand-in-hand. We are a global empathy-led technology services company. Product innovation and mature software engineering are part of our core DNA. Respect, Fairness, Growth, Agility, and Inclusiveness are the core values that we aspire to live by each day.
We continue to expand our digital strategy, design, architecture, and product management capabilities to offer expertise, outside-the-box thinking, and measurable results.
Roles & Responsibilities :
Familiar with the software development life cycle (SDLC) from analysis to deployment & coding standards.
Believes in systematic approach to developing the system through clear documentation (flowcharts, layouts, & etc) of functionality, address every use case through creative solutions.
Ability to integrate the developed functionality and/or component into a fully functional system.
Ensure unit and integration level verification plan are in place
Actively participate in troubleshooting, debugging and updating current live system.
Work closely with analysts, designers and other peer developers.
Preparing technical training documents for onboarding new engineers.
Requirements :
Must have good understanding of Object Oriented Programming
5 years or more of professional software development experience with 3 or more years of experience developing .NET/C# web applications using Microsoft Development Products
Define, design and build varied components of web-based and/or mobile applications
Experience using WPF (Windows Presentation Foundation) to build Windows desktop applications is a plus
Experience using JavaScript frameworks/libraries (Angular, React, JQuery, D3) is a plus
Experience with SQL Server (relational databases) and no-SQL Databases (Apache Cassandra or similar)
Strong analytical skills – be a problem solver and be able to develop complex algorithms
Strong familiarity of software architecture, databases and design patterns
Exposure or understanding of deploying applications on cloud platforms like Azure/AWS
Experience : 5+Years
Job Location : Remote
Work Timings : 3pm- 12 am
Job ID: RP100
Work Location: Remote
Required Experience: 4 to 7 years
Job Description
- Must have Google Cloud Big Query experience
- Strong experience with data analysis, data modeling and governance, with excellent analytical and problem-solving abilities
- Good knowledge of Data Warehouses, data flow ETL pipelines
- Design, configuration/administration of database software in Cloud platform.
- Monitoring, Troubleshooting, and Performance tuning the DB objects.
- Experience on Table Partition, Clustered Table, Materialized View, External Tables etc.
Anyone RDBMS technologies
- Good experience in DB design with knowledge of ER Diagram, PK/FK, Stored procedure, Function, Triggers, and Indexes.
- Understanding the requirement of the App team and creating the necessary DB objects by following the best practices.
- Managing logins and database users, as well as database roles, application roles, and other security principles within the database.
- Deep knowledge about Indexes, Performance tuning, and Complex SQL Query patterns.
- Monitoring, Tuning, and Troubleshooting the database-related issues.
About Us:
Mobile programming LLC is a US-based digital transformation company. We help enterprises transform ideas into innovative and intelligent solutions, governing the Internet of Things, Digital Commerce, Business Intelligence Analytics, and Cloud Programming. Bring your challenges to us, we will give you the smartest solutions. From conceptualizing and engineering to advanced manufacturing, we help customers build and scale products fit for the global marketplace.
Mobile programming LLC has offices located in Los Angeles, San Jose, Glendale, San Diego, Phoenix, Plano, New York, Fort Lauderdale, and Boston. Mobile programming is SAP Preferred Vendor, Apple Adjunct Partner, Google Empaneled Mobile Vendor, and Microsoft Gold Certified Partner.
We are looking for a Tech with a proficiency in backend technologies to design, develop a new state-of-the-art, paradigm changing product in software development. Responsibilities include evaluating technologies, developing proof-of concept prototypes, designing, and developing complete end-to-end products.
What would you be doing?
- should be able to mentor, guide, and train other engineers.
- should also be able to work with different stakeholders such as business owners, leadership, and customers.
- able to evaluate the business requirements, performance constraints, trade-offs and system limitations when developing software.
- should be well versed with new technologies and software development practices and be able to judge the trade-offs to make well thought out decisions, design choices and build the system for extensibility and robustness.
- The Django developer will be part of the core engineering team, entrusted with the complete architecture, design and development of the product.
- Work on a great product which empowers next generation entrepreneurs from its early phase
What makes you a suitable candidate?
- 1+ years of experience in designing and developing large scale, complex software applications
- Strong analytical, problem solving and design skills
- Strong knowledge of data structure, algorithms, and coding
- Proven experience in any high-level programming languages like Python, Django, DRF, Elasticsearch etc
- Knowledge of MVC architecture and hands-on experience with Spring, Spring security, Spring boot, and Hibernate framework
- Proven experience in database technologies (SQL, NoSQL, MongoDB).
- Working knowledge of developing Restful Micro services, ORM technologies etc
- Experience in container-based deployments
- Working experience in Amazon DynamoDB, Docker, AWS-Lambda
- Experience with Messaging technologies like Kakfa, RabbitMQ, redis etc
- Ability to multitask
- Good communication, organizational and leadership skills
- Sr. Data Engineer:
Core Skills – Data Engineering, Big Data, Pyspark, Spark SQL and Python
Candidate with prior Palantir Cloud Foundry OR Clinical Trial Data Model background is preferred
Major accountabilities:
- Responsible for Data Engineering, Foundry Data Pipeline Creation, Foundry Analysis & Reporting, Slate Application development, re-usable code development & management and Integrating Internal or External System with Foundry for data ingestion with high quality.
- Have good understanding on Foundry Platform landscape and it’s capabilities
- Performs data analysis required to troubleshoot data related issues and assist in the resolution of data issues.
- Defines company data assets (data models), Pyspark, spark SQL, jobs to populate data models.
- Designs data integrations and data quality framework.
- Design & Implement integration with Internal, External Systems, F1 AWS platform using Foundry Data Connector or Magritte Agent
- Collaboration with data scientists, data analyst and technology teams to document and leverage their understanding of the Foundry integration with different data sources - Actively participate in agile work practices
- Coordinating with Quality Engineer to ensure the all quality controls, naming convention & best practices have been followed
Desired Candidate Profile :
- Strong data engineering background
- Experience with Clinical Data Model is preferred
- Experience in
- SQL Server ,Postgres, Cassandra, Hadoop, and Spark for distributed data storage and parallel computing
- Java and Groovy for our back-end applications and data integration tools
- Python for data processing and analysis
- Cloud infrastructure based on AWS EC2 and S3
- 7+ years IT experience, 2+ years’ experience in Palantir Foundry Platform, 4+ years’ experience in Big Data platform
- 5+ years of Python and Pyspark development experience
- Strong troubleshooting and problem solving skills
- BTech or master's degree in computer science or a related technical field
- Experience designing, building, and maintaining big data pipelines systems
- Hands-on experience on Palantir Foundry Platform and Foundry custom Apps development
- Able to design and implement data integration between Palantir Foundry and external Apps based on Foundry data connector framework
- Hands-on in programming languages primarily Python, R, Java, Unix shell scripts
- Hand-on experience in AWS / Azure cloud platform and stack
- Strong in API based architecture and concept, able to do quick PoC using API integration and development
- Knowledge of machine learning and AI
- Skill and comfort working in a rapidly changing environment with dynamic objectives and iteration with users.
Demonstrated ability to continuously learn, work independently, and make decisions with minimal supervision
About the company
Anaxee Digital Runners is building India's largest last-mile verification & data collection network of Digital Runners (shared feet-on-street, tech-enabled) to help Businesses & Consumers reach remotest parts of India, on-demand. We are already serving 11000 pin codes (57% of India) | Website: www.anaxee.com
Check out our company pitch (6 min video) to understand the goal-https://www.youtube.com/watch?v=7QnyJsKedz8
Job Role
As a software developer, you’ll be the brain behind crafting, developing, testing, going live and maintaining the system. You are passionate about understanding the business context for features built to drive better customer experience and adoption.
Job Location: Indore | Full Time | Experience: 2 yrs & above | Salary ∝ Expertise | Rs.5L ~ Rs.12L |
Responsibilities:
- Familiar with the software development life cycle (SDLC) from analysis to deployment.
- Comply with coding standards and technical design.
- Believes in a systematic approach to developing the system through clear documentation (flowcharts, layouts, &, etc) of functionality, address every use case through creative solutions.
- Adapts structured coding styles for easy review, testing, and maintainability of the code.
- Integrate the developed functionality and/or component into a fully functional system.
- Ensure unit and integration level verification plan are in place and adheres to a great quality of code at all time.
- Active participate in troubleshooting, debugging and updating current live system.
- Verify user feedback in making system more stable and easy.
- Work closely with the Product manager, designers, and other peer developers.
- Preparing technical training documents for onboarding new engineers.
Requirements
- Bachelor’s degree in computer science or equivalent practical experience.
- 2+ years of experience as a Software Engineer or Software Developer or in a relevant role.
- Smart Hands-on coder who can independently architect solutions and mentor team members.
- Solid experience in PHP, Node, Mongo, and MySQL
- Should have good Web service and Web application fundamentals
- Understanding of OOPS concepts, Persistence, Threading.
- Should have worked with HTML/ CSS/ Jquery/ Js / API.
- Knowledgeable in data structures and algorithms.
- Experience with open-source projects.
- Experience in designing interactive applications.
- A background in Engineering with sound oral and written communication skills.
- Prior experience with AWS, Azure, Google or OpenStack will be a plus.
- Hands-on a google cloud platform and Experience in Suite CRM is an added advantage.
Your application should include
- Details about applications you have previously worked on
- A GitHub or stack overflow profile if you have one.
What we offer
- Competitive salary, Startup Flexibility
- Great working atmosphere in a comfortable office.
- And an opportunity to get associated with a fast-growing VC funded startup.









