Principal Data Engineer

Ref#: 33583
CBS Business Unit: CBS Interactive
Job Type: Full-Time Staff
Job Location: San Francisco, CA, US

About Us:

CBS Interactive is the premier online content network for information and online operations of CBS Corporation as well as some of the top native digital brands in the entertainment industry. Our brands dive deep into the things people care about across entertainment, technology, news, games, business and sports. With over 1 billion users visiting our properties every quarter, we are a global top 10 web property and one of the largest premium content networks online.

Check us out on The Muse, Instagram and YouTube for an inside look into 'Life At CBSi' through employee testimonials, office photos and company updates.
Role Details:
The Principal Data Engineer should strive to design highly reliable and scalable data engineering platforms to deliver high quality data for downstream reporting by other groups such as BI, Data Engineering, Machine Learning, Finance, Yield and Audience Acquisition. The Principal Data Engineer will work with all data engineering teams across the organization to develop and promote best practices and standards company wide.

Your Day-to-Day:
  • Design and develop highly scalable and reliable data engineering pipelines to process large volumes of data across disparate data sources in the cloud
  • Design and develop web applications for use by business stakeholders
  • Manage complex projects by simplifying tough problems
  • Mentor other data engineers and assist with code reviews during design and development phase
  • Design and develop data quality system to identify data issues proactively, rather than reactively
  • Develop and promote best practices in data engineering
  • Collaborate with data engineering and data science teams to understand data challenges and provide scalable and flexible solutions
  • Stay ahead of the technology by researching and learning new technologies 

Key Projects: 
  • Design and implement data pipeline framework for cloud
  • Migrate of on-prem data pipelines to cloud
  • Design and implement data quality system for important data pipelines
What you bring to the team:
You have -
  • Masters degree in computer science or equivalent experience in the related field
  • 10+ years experience doing data engineering work with at least 3+ years working in either GCP or AWS
  • Experience in building operational ETL data pipelines from a number of disparate sources in a Data Warehouse environment
  • Excellent understanding and hands on experience with advanced SQL
  • Experience building ML algorithms, pipelines, predictive analytics, deep learning
  • Experience with Python web framework such as Django or Flask and data analysis tools such as iPython Notebook, Pandas & matplotlib
  • Experience developing ETL, streaming frameworks, and data processing applications
  • Experience as a full stack developer
  • Experience working with large data sets with tens of millions of records every day
  • Experience with Agile development methodologies 
  • Ability to work with others from diverse skill-sets and backgrounds
  • Proven ability to clearly communicate complex solutions
  • Experience working on large scale data migration projects 
  • Exceptional written and verbal communication skills

You might also have-
  • Experience working with enterprise BI tools is a plus
  • Familiarity with Atlassian products such as Jira, Confluence, etc.
  • Experience with Dataflow, Composer/Airflow,Terraform, Apache NiFi, Pytorch, Keras, Tensorflow, Kafka is a plus

EEO Statement:

Equal Opportunity Employer Minorities/Women/Veterans/Disabled

< Back to job list