Tufts University

Sr Data Analytics Engineer

Medford, MA, US

Remote
Full-time
$114.2k–$171.3k/year
2 days ago
Save Job

Summary

Overview Tufts Technology Services (TTS) is a university-wide service organization committed to delivering adaptable, results driven technology solutions in support of Tufts' mission of teaching, learning, research, innovation, and sustainability. With staff working remotely, hybrid and on campus across Tufts University, as well as a 24x7 IT Service Desk, we collaborate with schools and divisions to meet the demands of a global, mobile, and diverse community. We promote a collaborative, forward-thinking, flexible work environment, embrace diversity and inclusion, and encourage personal and professional development. Fostering a culture of organizational citizenship and making others successful, demonstrating integrity, ethical conduct and optimism, active contribution and continuous learning enables staff to serve the goals and values of the University and creates a fulfilling and positive work experience for all. What You'll Do As a member of the Data and Analytic Services team, the Senior Data Analytics Engineer will lead key data and analytics engineering initiatives, including end to end ownership for designing and deploying data solutions. The ideal candidate will have 5 years of technical experience and 1 to 3 years of experience building and maintaining complex data pipelines. The individual filling this highly visible position will be key to helping Tufts enhance our data platform in a way that enables data driven decisions, processes and insight to areas across the University. This position is fully remote. * Meet with clients and/or supervisor to gather requirements and understand project requirements. Develop technical implementation plans to meet project requirements. Designs and develops complex data structures and pipelines to organize, collect and standardize data to generate insights and addresses reporting and analytics needs. Uses advanced programming skills in Python to build robust data pipelines. Writes ETL (Extract / Transform / Load) processes, designs data repositories and develops tools for analytic processing. Provides technical advice to team members. Collaborates with Analytics team to transform data and integrate algorithms and models into highly available, production systems. * Maintain integration projects. Apply updates and security fixes when needed. Troubleshoot any technical issues as they arise. Create and maintain documentation regarding all created/supported integration projects. Partner with application, infrastructure, and information architects to develop a cohesive and scalable future-state integration program. * Collaborate with analysts, designers, and system owners in the testing of newly-integrated software programs and applications. Analyze issues and develop fixes and enhancements to existing integrations as needed. Experiment with available tools and advise on new tools in order to determine optimal solution given the requirements dictated by the model/use case. What We're Looking For Basic Requirements: * Skills and knowledge typically acquired through a Bachelor's degree in Computer Science or a related field and 5+ years of related work experience. * Experienced using the Python programming language as an ETL tool in bringing in data from Enterprise Systems, API's, logs and other sources. * Proficient with SQL. * Experience with version control systems, preferably Git. * A demonstrated ability to produce highly reliable, production grade systems. * A strong desire to learn. Preferred Qualifications: * Familiarity with at least one relational database such as Oracle, SQL Server etc. * Experience working in a Linux environment. * Experience with cloud services, preferably AWS. * Experience with an event-based integration platform, especially Kafka, strongly preferred. * Experience with data virtualization platforms, especially Denodo, strongly preferred. Pay Range Minimum $114,200.00, Midpoint $142,750.00, Maximum $171,350.00 Salary is based on related experience, expertise, and internal equity; generally, new hires can expect pay between the minimum and midpoint of the range. Responsibilities As a member of the Data and Analytic Services team, the Senior Data Analytics Engineer will lead key data and analytics engineering initiatives, including end to end ownership for designing and deploying data solutions. The ideal candidate will have 5 years of technical experience and 1 to 3 years of experience building and maintaining complex data pipelines. The individual filling this highly visible position will be key to helping Tufts enhance our data platform in a way that enables data driven decisions, processes and insight to areas across the University. This position is fully remote. * Meet with clients and/or supervisor to gather requirements and understand project requirements. Develop technical implementation plans to meet project requirements. Designs and develops complex data structures and pipelines to organize, collect and standardize data to generate insights and addresses reporting and analytics needs. Uses advanced programming skills in Python to build robust data pipelines. Writes ETL (Extract / Transform / Load) processes, designs data repositories and develops tools for analytic processing. Provides technical advice to team members. Collaborates with Analytics team to transform data and integrate algorithms and models into highly available, production systems. * Maintain integration projects. Apply updates and security fixes when needed. Troubleshoot any technical issues as they arise. Create and maintain documentation regarding all created/supported integration projects. Partner with application, infrastructure, and information architects to develop a cohesive and scalable future-state integration program. * Collaborate with analysts, designers, and system owners in the testing of newly-integrated software programs and applications. Analyze issues and develop fixes and enhancements to existing integrations as needed. Experiment with available tools and advise on new tools in order to determine optimal solution given the requirements dictated by the model/use case. Qualifications Basic Requirements: * Skills and knowledge typically acquired through a Bachelor's degree in Computer Science or a related field and 5+ years of related work experience. * Experienced using the Python programming language as an ETL tool in bringing in data from Enterprise Systems, API's, logs and other sources. * Proficient with SQL. * Experience with version control systems, preferably Git. * A demonstrated ability to produce highly reliable, production grade systems. * A strong desire to learn. Preferred Qualifications: * Familiarity with at least one relational database such as Oracle, SQL Server etc. * Experience working in a Linux environment. * Experience with cloud services, preferably AWS. * Experience with an event-based integration platform, especially Kafka, strongly preferred. * Experience with data virtualization platforms, especially Denodo, strongly preferred.

How strong is your resume?

Upload your resume and get feedback from our expert to help land this job