About the Role
Β
We are seeking a Staff Engineer with a minimum of 9 years of experience in distributed systems or data platform development. In this role, you will be a technical leader responsible for architecting and driving the technical direction of critical components within the Acceldata Open Data Platform (ODP). You will work on solving the most challenging distributed systems problems, contribute to open-source communities, and mentor senior engineers while shaping the future of large-scale data infrastructure.This is a full-time, on-site position, open to candidates with valid work authorization.
Β
Why Join Us - At Acceldata, you won't just be writing code, you'll be defining the architecture and technical vision for modern data platforms used by some of the world's largest enterprises.
Β
As a Staff Engineer, you'll lead high-impact initiatives that directly influence how petabyte-scale data is managed, observed, and optimized across hybrid and cloud ecosystems.You'll collaborate with industry experts, Apache committers, and seasoned engineers who share your passion for distributed systems and open-source innovation.
Β
Your technical decisions will shape mission-critical systems, and your contributions will be visible across the global data ecosystem. If you're ready to take ownership of complex technical challenges and drive innovation at scale, Acceldata is where you belong.
Responsibilities
β’ Own the end-to-end architecture of major platform components, driving design decisions that impact scalability, reliability, and performance.
β’ Architect and implement next-generation distributed data storage, processing, and query optimization systems.
β’ Partner with product, engineering leadership, and stakeholders to define technical roadmaps and translate business requirements into scalable solutions.
β’ Lead contributions to Apache projects and open-source communities, establishing Acceldata's presence and technical reputation.
β’ Identify and resolve complex performance bottlenecks across distributed systems, optimizing for cost, latency, and throughput.
β’ Mentor and guide Senior Software Engineers, fostering technical excellence and a culture of continuous learning.
β’ Establish and enforce coding standards, perform thorough code reviews, and ensure architectural consistency across the platform.
β’ Evaluate and recommend emerging technologies, tools, and frameworks that align with platform goals.
β’ Lead root cause analysis for critical production issues and drive systemic improvements to prevent recurrence.
β’ Work across diverse environments: Bare Metals, VM, Kubernetes, and multi-cloud architectures.
Mandatory Skills & Qualifications
β’ 9+ years of hands-on software development experience with at least 5 years focused on distributed systems or big data platforms.
β’ Expert-level proficiency in Java or Scala, with strong Python skills. Experience with systems languages is a plus.
β’ Deep understanding of distributed computing principles including consensus algorithms, data partitioning, replication, fault tolerance, and consistency models.
β’ Extensive hands-on experience with multiple components of the big data stack (Hadoop, Spark, Hive, Trino, Kafka, Flink, NiFi or similar).
β’ Expert-level Linux proficiency including kernel tuning, performance analysis, and debugging at the system level.
β’ Strong experience with Kubernetes, Docker, and cloud platforms (AWS, GCP, or Azure).
β’ Deep expertise with Maven, Gradle, or SBT; experienced in CI/CD pipelines, GitHub, and artifact management (Nexus).
β’ Proven ability to diagnose and resolve complex issues including memory leaks, deadlocks, GC tuning, and distributed system failures.
β’ Strong understanding of architectural patterns for scalable, fault-tolerant distributed systems.
β’ Excellent written and verbal communication skills; ability to articulate complex technical concepts to diverse audiences including senior leadership.
β’ Demonstrated ability to work effectively with cross-functional teams, external contributors, and the open-source community.
Desired Skills (Bonus)
β’ History of significant contributions or committer status in Apache projects (e.g., Spark, Kafka, Hive, Iceberg, Flink etc.).
β’ Experience with modern table formats (Apache Iceberg, Delta Lake, Hudi) and data lake architectures.
β’ Experience developing or optimizing query engines, query planners, or execution frameworks.
β’ Familiarity with data governance, security, lineage, and compliance frameworks.
β’ Experience leading large-scale data platform migrations or Hadoop distribution upgrades.
β’ Published technical blogs, papers, or conference presentations on distributed systems topics.
β’ Experience leading or maintaining open-source projects.
β’
To Apply
Please submit your resume and cover letter to Person at File by the date. We look forward to reviewing your application!