T

Scientist III, Data Sciences

Thermo Fisher Scientific
Full-time
On-site
Pittsburgh, Pennsylvania, United States

Work Schedule

Standard (Mon-Fri)

Environmental Conditions

Office

Job Description

When you join us at Thermo Fisher Scientific, you’ll be part of an inquisitive team that shares your passion for exploration and discovery. With revenues of more than $40 billion and the largest investment in R&D in the industry, we give our people the resources and chances to create significant contributions to the world.

Job Title: Data Engineer– KL6

When your part of the team at ThermoFisher Scientific, you’ll do important work, like helping customers in finding cures for cancer, protecting the environment, or making sure our food is safe. Your work will have real-world impact, and you’ll be supported in achieving your career goals.

Location/Division Specific Information

As a Data Scientist I, based in Pittsburgh, PA team dedicated to increasing business efficiency of a Fortune 200 Company. This new position provides an exciting way for an experienced Data Scientist to employ a cross section of machine learning and analytic processes to develop solutions that will improve efficiency and drive revenue.

How will you make an impact:

Being an organization that provides data engineering and analytics engineering data engineering for all businesses across ThermoFisher Scientific, instrumental in helping our business partners and customers with their data and analytics needs.

How will you get here?

  • Experience in Data Lake, data analytics & Oracle, SQL Server or AWS Redshift type databases
  • Bachelor’s degree in computer science, Information Technology, Information Security or other technical field is desired, though 5years of direct and applicable professional experience working as a Network Engineer will be considered for a substitution.
  • Hands on experience in using S3, AWS Glue jobs, S3 Copy, Lambda and API Gateway.
  • Working SQL experience to solve SQL code. Redshift knowledge is an added advantage.
  • Solid experience in DevOps and CI/CD using Git and Jenkins, experience in cloud native scripting such as CloudFormation and ARM templates.
  • Experience working with Python, Python ML libraries for data analysis, wrangling and insights generation.
  • Experience in Python and common python libraries. 
  • Demonstrates a strong working experience with database availability, monitoring, backups and recovery, database security, high-availability, and disaster recovery for MySQL and PostgreSQL databases.
  • Experience with source control systems such as Git, Bitbucket, and Jenkins build and continuous integration tools. 
  • Support the execution and data integrity of strategic projects.
  • Exposure to Kafka, Redshift, Sage Maker would be added advantage.
  • Exposure to data visualization tools like Power BI, Tableau etc.
  • Functional Knowledge in the areas of Sales & Distribution, Material Management, Finance and Production Planning.
  • Having Certifications like AWS Certified Data Analytics, CCA Spark and Hadoop Developer or CCP Data Engineer will have more understanding about the tools.

Education:

  • Undergraduate degree in Statistics, Computer Science, Data Science or a related field preferred; an MBA or equivalent consulting / working experience is strongly preferred, services background preferred
  • demonstrated ability with five years of AWS Cloud on data integration with Apache Spark, EMR, Glue, Kafka, Kinesis, and Lambda in S3, Redshift, RDS, MongoDB/DynamoDB ecosystems

Experience:

  • Total 8+ years of knowledge in the IT, Leading and developing BI and DW applications.
  • 4+ Years of Experience in Data Lake, Data Analytics & Business Intelligence problems.
  • Experience with relational databases, ETL (Extract-Transform-Load), and ETL and DB scripting language (Databricks, Oracle preferably)
  • Solid experience in Data Lake using AWS Databricks, Apache Spark & Python

Knowledge, Skills, Abilities

  • Strong real-life experience in python development especially in pySpark in AWS Cloud environment.
  • Led life cycle of ETL Pipelines and other cloud platform tools, including GitHub, Jenkins, Terraform, Jira, and Confluence.
  • Highly hard-working, execution-focused, with a willingness to do "what it takes” to deliver results as you will be expected to rapidly cover a considerable amount of demands on data integration
  • Ability to analyze trends associated with huge datasets.
  • Excellent prioritization and problem-solving skills.
  • Takes a broad view when approaching issues; using a global lens.
  • Ability to learn from and train other team members

Education

  • Bachelor’s degree in Computer Science or equivalent with 5+ years of experience in a data engineering role with a solid grasp of technical, business, and operational process requirements.

Our Mission is to enable our customers to make the world healthier, cleaner and safer. Watch as our colleagues explain 5 reasons to work with us. As one team of 100,000+ colleagues, we share a common set of values - Integrity, Intensity, Innovation and Involvement.

Apply today! http://jobs.thermofisher.com 

Thermo Fisher Scientific is an EEO/Affirmative Action Employer and does not discriminate on the basis of race, color, religion, sex, sexual orientation, gender identity, national origin, protected veteran status, disability or any other legally protected status. 

We will ensure that individuals with disabilities are provided reasonable accommodation to participate in the job application or interview process, to perform essential job functions, and to receive other benefits and privileges of employment. Please contact us to request accommodation.