Work Schedule
Other
Environmental Conditions
Office
Job Description
COMPANY: Thermo Fisher Scientific Inc.
LOCATION: 168 Third Avenue, Waltham, MA 02451
TITLE: Scientist III, Data Sciences
HOURS: Monday to Friday, 8:00 am to 5:00 pm
DUTIES: • Develop scalable data pipelines and builds out new API integrations to support continuing increases in data volume and complexity.
• Own and deliver Projects/Enhancements associated with Data platform solutions.
• Develop solutions using Pyspark/EMR, SQL and databases, AWS Athena, S3, Redshift, AWS API Gateway, Lambda, Glue and other Data Engineering technologies.
• Implement solutions using AWS and other cloud platform tools, including GitHub, Jenkins, Terraform, Jira, and Confluence.
Want more jobs like this?
Get jobs in Weston, MA delivered to your inbox every week.
• Follow agile development methodologies to deliver solutions and product features by following DevOps, Data Ops and Dev Sec Ops practices.
• Propose Data load optimizations and continuously implement to improve the performance of the Data loads.
• Identify, design, and implement internal process improvements, including automating manual processes, optimizing data delivery, and re-designing infrastructure for greater scalability.
• Keep data separated and secure across through multiple data centers and AWS regions.
• Be available and participate in on-call schedule to address critical operational incidents and business requests.
• Meet and exceed BI Operational SLAs for Service Now incidents, Major Incidents, xMatters alerts, Employee Experience Metrics and BI application/process availability metrics.
Travel: Up to 10% domestic travel required. Telecommuting Employee: Reports to company headquarters in Waltham, MA. Can work remotely or telecommute.
REQUIREMENTS: MINIMUM Education Requirement: Master's degree in Computer Science, Mechanical Engineering, or related field of study. MINIMUM Experience Requirement: 1 year of IT experience as a Data Engineer, or related experience. Alternative Education and Experience Requirement: Bachelor's degree in Computer Science, Mechanical Engineering, or related field of study plus 3 years of IT experience as a Data Engineer, or related experience.
Required knowledge or experience with:
• Informatica Power center 10.4, Oracle R12, TOAD, SFDC, Data Warehouse Administration, UNIX, Win SCP, Windows 7, Linux, Informatica Power Exchange, Oracle 11g, and Flat Files;
• SQL,PL/SQL, No SQL, and Bigdata;
• Databricks, Data/Delta lake, Oracle, or AWS Redshift type relational databases;
• Databricks/Spark-based Data Engineering Pipeline development;
• Python-based data integration and pipeline development;
• Data lake and Delta lake experience with AWS Glue and Athena;
• AWS Cloud on data integration with Apache Spark, Glue, Kafka, Elastic Search, Lambda, S3, Redshift, RDS, MongoDB/DynamoDB ecosystems;
• Python development, including in pySpark in AWS Cloud environment;
• Python and common python libraries;
• Analytical experience with database in writing complex queries, query optimization, debugging, user defined functions, views, and indexes;
• Source control systems, including Git and Jenkins build and continuous integration tools;
• Development methodology;
• Writing functional and technical design specifications; and
• Resolving complex data integration problems.
#LI-DNI