
ETL Developer | Data Acquisition and Distribution | Experienced Hire
at Susquehanna
Posted 9 hours ago
No clicks
- Compensation
- Not specified USD
- City
- Bala Cynwyd
- Country
- United States
Currency: $ (USD)
Susquehanna is looking for an experienced ETL Developer to join our technology team at our Bala Cynwyd, PA headquarters. You will design, develop, and optimize PySpark data processing workflows and build real-time and batch data pipelines using Apache Kafka. You will write clean Python code for data transformation and automation, develop shell scripts, and collaborate with data engineers, analysts, and platform teams to deliver end-to-end data solutions. You will troubleshoot pipelines and follow best practices in version control, CI/CD, documentation, and code reviews.
Susquehanna is looking for an experienced ETL Developer to join our technology team at our headquarters in Bala Cynwyd, PA, just outside of Center City Philadelphia. As a member of our ETL team, you will participate the design and development of our ETL infrastructure. You will work directly with technical and business teams to analyze their data needs and develop solutions for both batch and continuous data movements using enterprise class tools and custom scripting methods.
This role will give you exposure to cutting-edge, distributed database environments that support critical business activities. We are looking for someone who is excited to join our efforts as we evolve, evaluate and promote newer technologies within this space across our firm globally.
In this role, you will:
- Design, develop, and optimize PySpark data processing workflows for large-scale datasets.
- Build and maintain real-time and batch data pipelines leveraging Apache Kafka.
- Write clean, efficient, and maintainable Python code for data transformation, ETL, and automation.
- Develop shell scripts and other automation scripts to support data workflows and operational tasks.
- Work with relational databases (e.g., PostgreSQL, MySQL, SQL Server, Oracle) to write efficient SQL queries, manage schemas, and optimize performance.
- Collaborate closely with data engineers, analysts, and platform teams to deliver end‑to‑end data solutions.
- Troubleshoot and improve existing pipelines, ensuring performance, reliability, and scalability.
- Follow best practices in version control, CI/CD, documentation, and code reviews.
What we're looking for
- Bachelor's degree in Computer Science or a related technical discipline is required
- 5+ years of hands‑on development experience in data engineering or backend engineering roles.
- Strong proficiency in:
- PySpark (RDDs, DataFrames, Spark SQL, performance tuning)
- Apache Kafka (producers/consumers, topics, partitions, schema management)
- Python (data manipulation, modular code, error handling)
- Scripting (Bash, Shell, or similar)
- Experience working with relational databases and writing optimized SQL queries.
- Solid understanding of distributed systems and large‑scale data processing concepts.
- Familiarity with ETL best practices, data modeling, and pipeline orchestration
- Basic understanding of LLM operations (LLMOps), including prompt logging, model monitoring, and evaluation
- Knowledge of Spark cluster management, resource optimization, and tuning strategies is preferred.
- Experience with workflow orchestration tools such as Apache Airflow is a plus
SIG does not accept unsolicited resumes from recruiters or search firms. Any resume or referral submitted in the absence of a signed agreement will become the property of SIG and no fee will be paid.

