- Interesting Data Gigs by Marcos Ortiz
- Posts
- Interesting Data Gigs # 6: Data Engineer I at ZoomInfo (NASDAQ: ZI)
Interesting Data Gigs # 6: Data Engineer I at ZoomInfo (NASDAQ: ZI)
Why you must follow Julia Silge and Lex Fridman
If you are not a subscriber of Interesting Data Gigs, join 1200+ other Data geeks that receive it directly in their inbox each week — it’s free.
This edition is sponsored by Ramp
Ramp is the finance automation platform uniquely designed to save businesses their most valuable resources—time and money. By integrating corporate cards with refreshingly modern software to manage all non-payroll spending, Ramp turns typically painful finance processes into an effortless and delightful experience.
Companies like Webflow, Applied Intuition, and Stanley Steemer use Ramp to automate expenses, payments, accounting, and reporting. By closing their books 8 days faster each month and saving their businesses millions of dollars each year, leading companies are using Ramp to build a world where finance teams become more strategic and lean. As a result, the entire business moves faster.
Hi Data Geeks.
Today, I will be talking about ZoomInfo (NASDAQ: ZI), well known as the “Go-To Market Platform“ and a very interesting public company with an amazing business model, looking for incredible Data-Driven people to join the team.
Why you should join ZoomInfo
The position I will be talking about is Data Engineer I, and the POC here is Joseph Russo.
Let's take the profitability test first:
is ZoomInfo profitable? YES. According to its last financial results in Q1 2022, the company generated $241 Million in revenue (an increase of 58%), an Unlevered Free Cash Flow of $125.9 million, and a Cash and Cash Equivalents amount of $394 Million. So, it’s highly profitable and with a good cash reserve on hand. ✅
Is the company inserted in a growing or lagging industry? Just check out this interesting article from Jason M. Lemkin about the company and this incredible conversation with Jesse Pujji and Henry Schuck (CEO and founder at ZoomInfo) on the Business Breakdowns podcast and make your own conclusions.
Has ZoomInfo made a layoff? NO
Has the role you are applying for an actual growth trajectory in this company? ZoomInfo's business is based on data, so I assume that your job will be needed in the long term. If you have any doubts, just ask Henry, but make sure you read this first.
ZoomInfo’s Data Tech Stack
If you read carefully again the job post, you will see that the tech stack is very interesting as well:
MySQL
Apache Airflow for orchestration
Python is the main language for anything related to Data Analytics
Snowflake as the Cloud-based data warehouse
Amazon Web Services (AWS) for some things
Google Cloud for others, especially after the partnership between both companies
Apache Spark
dbt
Fivetran, and more.
So, if you are working right now with all these tools, I encourage you to apply for the position or at least have a conversation with Joseph about it.
Let’s dissect the job here and discuss some ideas on how to approach this job application (THE REAL MEAT)
First, don’t forget to read these interesting resources in the ZoomInfo’s Engineering blog:
Second, if you read the role, you should know at this point that the company is actively using Amazon Web Services and Google Cloud Platform.
Why?
Because they are where the data of their clients is needed, and some clients use AWS and others use GCP, some clients use Snowflake for its data efforts, others use BigQuery. You can see all the integrations that ZoomInfo has when you visit the ZoomInfo’s Apps Market:
ZoomInfo’s Apps Market
So, if are a Data Engineer who has worked with at least two major cloud providers, you are in a very good position to do this application. Even better if you have used different Big Data or Data Analytics related products in both of them.
In one of the items of the job description, they put the “Nice to Have“ points:
Hands-on experience in Python or equivalent programming language
Experience with Airflow and DBT
Experience with AWS/GCP cloud services such as GCS/S3, Lambda/Cloud Function, EMR/Dataproc, Glue/Dataflow, Athena.
Experience with data visualization tools and infrastructures (like Tableau/Sisense/Looker/other)
Experience with development practices – Agile, CI/CD, TDD..
So, you get the idea.
My advice here? Pick a cloud provider and become an expert in the tools they are using right now.
For example: let’s say you want to specialize in Google Cloud Platform, so pick the tools from GCP like Google Cloud Storage, GCP Dataflow, Dataproc, and of course BigQuery, and become “the expert“, even before applying for it.
And especially how to combine all these services to create Data-driven solutions.
How to do that?
The Google Cloud Professional Data Engineer Certification is perfect for that, and the best resource I know in order to do this correctly, it’s the amazing course available at Coursera for it. As a side reading, make sure to check out this insightful post from Chenming Yong about how to prepare for the exam
This book called Building Big Data Pipelines with Apache Beam: Use a single programming model for both batch and stream data processing, by Jan Lukavsky; is an incredible resource as well
The second piece of advice here is simple but powerful: make sure that your solutions are very efficient related to costs vs value and impact.
ZoomInfo started its journey as a bootstrapped business, so you should know that they are very sensitive to costs, and if you can think about this even from the start of the design of your Data Engineering pipelines, do it.
How do to do that with Google Cloud?
Make sure to read the work of Pathik Sharma, the Cloud FinOps Cost Optimization Practice Lead at Google Cloud, especially these resources:
Using these tips, you could combine the knowledge you already have in the Data Engineering front, with a FinOps approach for everything you will be doing at ZoomInfo.
This could give you an edge, at least from my perspective.
So, again if you are interested in the position, chat with Joseph Russo.
Other featured jobs of the Interesting Data Gigs Jobs Board
People to follow: Julia Silge and Lex Fridman
The content of both of them on LinkedIn related to Artificial Intelligence, Data Science, R, Data Visualization, and more is simply amazing
For example, Julia has an interesting hashtag called #TidyTuesday where she shares interesting research, insights, or visualization for a particular problem using the Tidy framework from R.
The last one was focused on how to use the spatialsample package to understand drought in Texas, and how it is related to other characteristics of TX counties:
You can check that code on her blog as well here.
In the case of Lex, the real jewel is his podcast, where he invites incredible people to chat about different topics.
One of my favorite episodes was the conversation with Tony Fadell, the co-creator of the iPod, iPhone, and the Nest thermostat; where they chatted about the new book written by Tony called “Build: An Unorthodox Guide to Making Things Worth Making”.
You can watch the entire conversation here:
If you are not a subscriber of Interesting Data Gigs, join 1200+ other Data geeks that receive it directly in their inbox each week — it’s free.
Reply