Data Engineer, Fivetran, Sqlserver, Python 3+ Yrs Experience
Description
Location: Mumbai
Job Description:
As part of the Astellas commitment to delivering value for our patients, our organization is currently undergoing transformation to achieve this critical goal. This is an opportunity to work on digital transformation and make a real impact within a company dedicated to improving lives.
DigitalX our new information technology function is spearheading this value driven transformation across Astellas. We are looking for people who excel in embracing change, manage technical challenges and have exceptional communication skills.
We are seeking committed and talented MDM Engineers to join our new FoundationX team, which lies at the heart of DigitalX. As a member within FoundationX, you will be playing a critical role in ensuring our MDM systems are operational, scalable and continue to contain the right data to drive business value. You will play a pivotal role in building maintaining and enhancing our MDM systems.
This position is based in India and may require on-site work from time to time.
Hybrid work from certain locations may be permitted in accordance with Astellas’ Responsible Flexibility Guidelines.
Purpose and Scope:
As part of the Astellas commitment to delivering value for our patients, our organization is currently undergoing transformation to achieve this critical goal. This is an opportunity to work on digital transformation and make a real impact within a company dedicated to improving lives.
DigitalX our new information technology function is spearheading this value driven transformation across Astellas. We are looking for people who excel in embracing change, manage technical challenges and have exceptional communication skills.
We are seeking committed and talented Senior Data Engineers, to join our new FoundationX team- which lies at the heart of DigitalX. As a member of our team within FoundationX, you will be responsible for ensuring our data driven systems are operational, scalable and continue to contain the right data to drive business value.
As a Data Engineer, you will play a crucial role in designing, building, and maintaining our data infrastructure. Working with FoundationX Data Engineers and key stakeholders, your expertise in Databricks, PySpark, SQL and other related technologies, will be instrumental in driving data-driven decision-making across the organization. You will play a pivotal role in building maintaining and enhancing our systems across the organization. This is a fantastic global opportunity to use your proven agile delivery skills across a diverse range of initiatives, utilize your development skills, and contribute to the continuous improvement/delivery of critical IT solutions.
Essential Skills & Knowledge:
- Subject Matter Expertise: possess a strong understanding of data architecture/ engineering/operations/data model (e.g. data vault, dimensional data model).
- Experience within Life Sciences/ Pharma /Manufacturing industry is preferred.
- Proven experience in building robust data pipeline; experience in (near) real-time processing is preferred.
- Technical Proficiency: Strong coding skills for example, Python, PySpark, SQL.
- Experience in QLIK, Databricks or Change Data Capture (eg. Fivetran) is preferred. Other technologies include Oracle and APEX.
- Proven experience with data analytics practices and techniques and organizations
- Agile Practices: Experience working in Agile development environments, participating in sprint planning, stand-ups, and retrospectives.
- Cloud Data Solutions: Familiarity with other cloud platforms (AWS, Azure, Google Cloud) and their data services
- Analytical Thinking: Demonstrated ability to lead ad hoc analyses, identify performance gaps, and foster a culture of continuous improvement
- Agile Champion: Adherence to DevOps principles, automation, and a proven track record with CI/CD pipelines for continuous delivery
- Understand and Interpret business requirements and can term them into technical requirements.
Key Responsibilities Include:
Data Modelling and Integration:
- Collaborate with cross-functional teams to analyze requirements and create technical designs, data models, and migration strategies.
- Design, build, and maintain physical databases, dimensional data models, and ETL processes specific to pharmaceutical data.
Cloud Expertise:
- Evaluate and influence the selection of cloud-based technologies such as Azure, AWS, or Google Cloud.
- Implement data warehousing solutions in a cloud environment, ensuring scalability and security.
Qlik Expertise:
- Leverage Qlik for data visualization, dashboards, and self-service analytics.
- Create Qlik applications to empower business users with insights.
- Data Pipeline Development: Design, build, and optimize data pipelines using Databricks and PySpark. Ensure data quality, reliability, and scalability.
Application Transition: Support the migration of internal applications to Databricks (or equivalent) based solutions. Collaborate with application teams to ensure a seamless transition.
Mentorship and Leadership: Lead and mentor junior data engineers. Share best practices, provide technical guidance, and foster a culture of continuous learning.
Data Strategy Contribution: Contribute to the organization’s data strategy by identifying opportunities for data-driven insights and improvements.
- Participate in smaller focused mission teams to deliver value driven solutions aligned to our global and bold move priority initiatives and beyond.
- Design, develop and implement robust and scalable data analytics using modern technologies.
- Collaborate with cross functional teams and practices across the organization including Commercial, Manufacturing, Medical, DataX, GrowthX and support other X (transformation) Hubs and Practices as appropriate, to understand user needs and translate them into technical solutions.
- Provide Technical Support to internal users troubleshooting complex issues and ensuring system uptime as soon as possible.
- Champion continuous improvement initiatives identifying opportunities to optimize performance security and maintainability of existing data and platform architecture and other technology investments.
- Participate in the continuous delivery pipeline. Adhering to DevOps best practices for version control automation and deployment. Ensuring effective management of the FoundationX backlog.
- Leverage your knowledge of data engineering principles to integrate with existing data pipelines and explore new possibilities for data utilization.
- Stay-up-to-date on the latest trends and technologies in data engineering and cloud platforms.
- Requirements
Qualifications:
Required:
- Bachelor's degree in computer science, Information Technology, or related field (Master’s preferred) or equivalent experience.
- Any relevant cloud-based integration certification at foundational level. For example:
- Any QLIK or BI certification
- AWS certified DevOps engineer,
- AWS Certified Developer
- Any Microsoft Certified Azure qualification
- Proficient in RESTful APIs
- AWS, CDMP, MDM, DBA, SQL, SAP, TOGAF, API, CISSP, VCP (any relevant certification)
MuleSoft
- Understanding of MuleSoft's Anypoint Platform and its components (desirable)
- Experience with designing and managing API-led connectivity solutions
- Knowledge of integration patterns and best practices
- Proficiency in network architecture and security concepts
AWS
- Experience provisioning, operating, and managing AWS environments
- Experience developing code in at least one high-level programming language
- Understanding of modern development and operations processes and methodologies
- Ability to automate the deployment and configuration of infrastructure using AWS services and tools
- Experience with continuous integration and continuous delivery (CI/CD) methodologies and tools
Microsoft Azure
- Fundamental understanding of Microsoft Azure and its services
- Experience with Azure services related to computing, networking, storage, and security
- Knowledge of general IT security principles and best practices
- Understanding of cloud integration patterns and Azure integration services such as Logic Apps, Service Bus, and API Management
Preferred Qualifications:
- Subject Matter Expertise: possess a strong understanding of data architecture/ engineering/operations/ reporting within Life Sciences/ Pharma industry across Commercial, Manufacturing and Medical domains.
- Other complex and highly regulated industry experience will be considered across diverse areas like Commercial, Manufacturing and Medical.
- Data Analysis and Automation Skills: Proficient in identifying, standardizing, and automating critical reporting metrics and modelling tools.
- Analytical Thinking: Demonstrated ability to lead ad hoc analyses, identify performance gaps, and foster a culture of continuous improvement.
- Technical Proficiency: Strong coding skills in SQL, R, and/or Python, coupled with expertise in machine learning techniques, statistical analysis, and data visualization.
- Agile Champion: Adherence to DevOps principles and a proven track record with CI/CD pipelines for continuous delivery.
Other critical skills required
Cross-Cultural Experience: Work experience across multiple cultures and regions, facilitating effective collaboration in diverse environments.
Innovation and Creativity: Ability to think innovatively and propose creative solutions to complex technical challenges.
Global Perspective: Demonstrated understanding of global pharmaceutical or healthcare technical delivery, providing exceptional customer service and enabling strategic insights and decision-making.
- 3 years + experience in data engineering with a strong understanding of PySpark and SQL, building data pipelines and optimization.
- 3 years + experience in data engineering and integration tools (e.g., Databricks, Change Data Capture)
- 3 years + experience of utilizing cloud platforms (AWS, Azure, GCP). A deeper understanding/certification of AWS and Azure is considered a plus.
- Experience with relational and non-relational databases.