Note : This role may be filled as a Data Engineer or Sr. Data Engineer. The position description is written at the Data Engineer level.
As a member of the PTT Data Engineering and Architecture Team, you will work in multi-disciplinary environments harnessing data to provide real-world impact for our processes and products.
You’ll work closely with a team of data scientists, subject matter experts and other data engineers in order to curate, transform and construct features which feed directly into our modeling approach.
Must collaborate, listen and learn from colleagues, challenge thoughtfully and prioritize impact of work. Responsible for finding improvements in advanced data analytic models and data architecture.
Looking for someone who can quickly adapt to ever changing environment, has a passion to learn and can be trusted to work in the best interests of colleagues.
Job Responsibilities Hybrid technical role interfacing with process SMEs using state of the art technologies, whilst also being able to communicate complex intractable ideas to non-technical audiences.
Collect clear requirements from SMEs and process experts. Work with our process experts to model their data landscape, obtain data extracts and define secure data exchange approaches Acquire, ingest, and process data from multiple sources and systems into Big Data platforms Understanding, assessing and mapping the data landscape Collaborate with our data scientists to map data fields to hypotheses and curate, wrangle, and prepare data for use in their advanced analytical models Building modular pipeline to construct features and modeling tables Supplementary Responsibilities Maintaining our Information Security standards on the engagement Defining the technology stack to be provisioned by our infrastructure team Use new and creative techniques to deliver impact for our clients as well as internal R&D projects Job Requirements Education and Experience Bachelor’s, Master’s, or Ph.
D. degree in computer science / electrical engineering or relevant field 8+ years of experience in bio-pharmaceutical industry specializing in bioprocess manufacturing sciences and technology.
Knowledge, Skills and Abilities Proven and significant experience building and managing data pipelines. Ability to build and manage data pipelines with Python, Scala or Java.
Knowledge of software engineering best practices such code reviews, testing frameworks, maintainability and readability.
Commercial client-facing project experience is helpful, including working in close-knit teams Ability to work across structured, semi-structured, and unstructured data, extract information and identify linkages across disparate data sets.
Meaningful experience in multiple database technologies (such as Hadoop, MS SQL Server, Oracle, MySQL, Teradata). Confirmed ability in clearly communicating complex solutions Deep understanding of Information Security principles to ensure compliant handling and management of process data Experience and interest in Cloud platforms such as : AWS, Azure, or Google Platform Familiarity with data warehousing and deploying ETL processes with Python.
Extraordinary attention to detail. Strong organizational and interpersonal skills : can get things done in a way that optimizes results, strengthens internal and external relationships, and with consideration of resources.
Knowledge of cGMP’s, Health Authority regulations, and Quality Systems. Work Environment / Physical Demands / Safety Considerations Ability to work in international / global environment.
10-30% travel anticipated. May work in the clean room environment that requires gowning in the form of hospital scrubs, Ability to sit, stand and move within work space for extended periods Able to work on co-located or remote cross-functional teams.