What to Expect
We are seeking a highly skilled and collaborative Staff Data Engineer (Big Data) to join our team and drive the development of data pipelines and analytics tooling to support the growth of our energy business, including Industrial, Residential, Supercharger, and Solar products.
As a key member of our team, you will design, build, and maintain large-scale data systems to extract insights from various data sources, including deployed fleets, internal applications, and other data warehouses. Your work will contribute to our understanding of key business performance indicators, informing data-driven decisions to optimize our strategies and drive business growth.
You will be responsible for leading the development of data efforts to identify and resolve fleet-wide performance, operation, and maintenance challenges across our growing fleet of energy products. You will work closely with cross-functional teams, including Industrial Energy, Residential Energy, and Solar Product and Service Engineering, to maximize product health of our fleet.
What You’ll Do
Derive an overall strategy of data management, within an established information architecture that allows us to scale data processing effectively
Gather and processes raw, structured, semi-structured, and unstructured data using batch and real-time data processing frameworks
Develop aggregate and summary tables for various engineering teams that span across multiple product lines and geographies
Mentor and lead data engineers providing technical guidance and oversight
Design and develop new systems and tools to enable cross functional teams to consume and understand data faster
Research and incorporate emerging software infrastructures, tools, and technologies, especially pertaining to big data processing
What You’ll Bring
Degree in Computer Science, Engineering, Physics, or related quantitative field or equivalent experience
Real work experience in building, deploying, and maintaining Py spark-based data pipelines
Experience with SQL and familiarity with database performance concepts like indices, segmentation, projections, and partitions
Knowledge of various data communication protocols (e.g., REST API, GRPC)
Strong dev-ops capabilities and bias for automation
Proficiency in Python or Scala
Experience with Linux
Experience in working with Presto/Kafka is a plus
Experience with power generation/storage devices is a plus
Experience with Airflow/Workflow orchestration tools and data visualization techniques and tools is a plus
Palo Alto, California
Full time