Description
*IN SCHOOL OR GRADUATED WITHIN THE LAST 12 MONTHS? PLEASE VISIT FUTURE FORCE FOR OPPORTUNITIES*
We are open to hiring a Senior, Lead, or Principal Engineer for this role.
About us
We are a technology-driven team focused on enabling scalable, high-performance data processing and analytics. As part of the Big Data Infrastructure team, you’ll use cutting-edge tools like Spark, Trino, Airflow, and Kafka to build resilient, scalable infrastructure for a variety of big data applications.
Position Overview
We are seeking an experienced Distributed Systems Software Engineer to lead our Big Data infrastructure projects. You will play a critical role in developing reliable and efficient data infrastructure that supports scalable data processing and analytics for both internal and external customers.
Your Impact
Build Data Processing and Analytics Services: Develop scalable data processing and analytics services utilizing our big data stack, including Spark, Trino, Airflow, and Kafka, to support real-time and batch data workflows.
Architect Distributed Systems: Design, develop, and operate resilient distributed systems that manage thousands of compute nodes across multiple data centers, ensuring scalability and high availability.
Troubleshoot and Innovate: Resolve complex technical challenges and drive innovations that enhance system resilience, availability, and performance.
Service Ownership and Live-Site Management: Manage the full lifecycle of services, balancing live-site reliability, feature development, and technical debt retirement. Maintain a strong focus on service reliability and performance.
On-Call Support: Participate in the team’s on-call rotation to address complex, real-time issues, keeping critical services operational and highly available.
Mentor and Guide Team Members: Provide mentorship and technical guidance to junior engineers, fostering growth, collaboration, and knowledge-sharing within the team.
Required Qualifications
Education and Experience: Bachelor’s or Master’s in Computer Science, Engineering, or a related field with 5+ years of experience in distributed systems, big data, or similar roles.
Experience with Cloud Environments: Proficiency in cloud platforms (AWS, GCP, or Azure), containerization (Docker, Kubernetes), and infrastructure-as-code (Terraform, Ansible).
Expertise in Big Data Technologies: Hands-on experience with Hadoop, Spark, Trino (or similar SQL query engines), Airflow, Kafka, and related ecosystems.
Proficiency in Programming: Strong skills in Python, Java, Scala, or other programming languages relevant to distributed systems.
Deep Understanding of Distributed Systems: Solid knowledge of distributed computing principles, data partitioning, fault tolerance, and performance tuning.
Analytical and Problem-Solving Skills: Proven ability to troubleshoot complex system issues, optimizing for speed, efficiency, and scale.
For roles in San Francisco and Los Angeles: Pursuant to the San Francisco Fair Chance Ordinance and the Los Angeles Fair Chance Initiative for Hiring, Salesforce will consider for employment qualified applicants with arrest and conviction records.
