Senior Data Engineer
Shutterfly - San Jose, CA
Apply NowJob Description
Senior Data Engineer (#RS-1029)HQ- SAN JOSE, CAFULL-TIMEJob Description:Develop ETL (extract, transform, load) scripts using Oracle, Teradata, Informatica, Ab Initio, PL/SQL, and Unix Shell scripts. Orchestrate ETL workloads using Automic. Follow agile development methodologies in the development and maintenance of the data warehouse with daily scrums for updates and progress reporting. Work on setting up airflow and data bricks instances and build and deploy code changes, continuous integration, and continuous delivery pipelines. Configure tableau (data visualization software) on AWS cloud instances and perform load tests to estimate peak performance capabilities using TabJolt (an open-source server load testing tool). Work with a team of data warehouse engineers for monitoring and maintaining the Extract, Transform and Load jobs on premise and cloud systems while participating in call rotations. Coordinate with Information Security, Networks Engineering, System Engineering, and IT teams to review new architecture setups and deployments. Gather user requirements and create onboarding and best practice documents. Train analysts on using Amazon Redshift and Microsoft Excel Power Query. Configure Amazon Redshift Clusters using Cloud Formation templates for data warehouses migration onto the cloud from on premise systems. Work with Amazon and Microsoft Azure development teams on a plugin to evaluate Single Sign-on authentication for Amazon Redshift using Microsoft Azure Active Directory.Build analytics tools using Amazon Cloud Watch Alarms and dashboards and integrate with Pager Duty and Slack to provide actionable insights into operational efficiency and performance for monitoring all infrastructure and Extract, Transform and Load pipeline progress.Minimum Requirements: Masters degree or equivalent in Computer Science or a related field. 2 years of progressive experience building and maintaining distributed production environments. Employer will accept pre- and post- Masters degree experience. All experience must be post-Bachelors degree.Employer will accept a Bachelors degree or equivalent in Computer Science or a related field and 4 years of progressive experience building and maintaining distributed production environments in lieu of a Masters degree and 2 years of experience. Work experience to include: 1) 2 years of experience with processors and machines that produce multi-dimensional datasets (big data) in consumer enterprises. Employer requires 4 years of experience with a Bachelors degree. 2) Development, deployment, and monitoring experience with ETL jobs and bug fixes. 3) Work in production support and participate in on-call rotations. 4) Development experience with at least one of the following: Oracle, Teradata, or MS SQL databases. 5) Designing and developing pipelines using Informatica or Oracle PL/SQL scripts and Perl/Unix Shell Scripts. 6) Utilizing version control software, including Concurrent Versions Systems (CVS) and Apache Subversion (SVN). 7) Utilizing software programming languages including Java or Python in developing ETL scripts. 8) Writing pipelines for automation, metrics, data collection, and system administration purposes. 9) Utilizing at least one of the following automation and orchestration tools, including Control-M or Automic. Any and all experience can be gained concurrently. Position requires weekly on-call rotations. Hours adjusted accordingly. 40 hrs./wk. Salary: $162,427 to $167,427 per year. Address of employment: 100% telecommute position from anywhere in the U.S. reporting to Shutterfly, LLC located at 10 Almaden Blvd., Suite 900, San Jose, CA 95113.Individuals interested in applying for the position can email resumes to talent@ referencing Job Code: #RS-1029.PDN-9e4f17c6-9bb4-46c4-ab91-f858c9316386
Created: 2025-03-06