Junior ETL Developer – SQL & Python Data Pipeline Role

Remotely
Full-time

Kick-start your data career as a Junior ETL Developer. Build SQL & Python pipelines, optimize warehouses, learn Informatica & Talend in a flexible remote setup.


Junior ETL Developer – SQL & Python Data Pipeline Role

Launching your journey as a Junior ETL Developer starts here. This role blends hands-on data extraction, transformation, and loading with modern SQL, Python, and top ETL tools—empowering you to turn raw information into analytical gold for decision-makers across multiple industries.


About the Team

A multidisciplinary data group—engineers, analysts, and architects—drives enterprise intelligence initiatives. You’ll collaborate with mentors dedicated to clean code, repeatable processes, and measurable value. Distributed across time zones, we rely on asynchronous communication and strong documentation to keep momentum high.


Your Mission

Transform fragmented data into trustworthy insights. You will design, test, and refine ETL pipelines that feed cloud warehouses powering dashboards, machine-learning models, and regulatory reporting.


Core Responsibilities

- Develop and maintain repeatable ETL workflows using Informatica or Talend.  

- Write performant SQL queries to extract and aggregate data from relational and columnar stores.  

- Script auxiliary tasks in Python (3.11+) for file manipulation, API ingestion, and automated validations.  

- Debug data discrepancies—trace lineage, identify root causes, and craft durable fixes.  

- Document mappings, transformation logic, and operational run-books for peers and auditors.  

- Monitor daily loads; tune indexing, partitioning, and parallelism to keep SLAs green.  

- Collaborate with BI developers, data scientists, and product owners to refine requirements.  

- Contribute code to Git repositories, open pull requests, and engage in peer reviews.  


Required Skills

- Bachelor’s degree in Computer Science, Information Systems, or related field.  

- 0-2 years of experience building ETL pipelines in academic projects, internships, or full-time roles.  

- Solid understanding of SQL fundamentals (joins, window functions, query optimization).  

- Working knowledge of Python for data manipulation and scripting.  

- Familiarity with at least one enterprise ETL platform such as Informatica PowerCenter, Talend Open Studio, AWS Glue, or Azure Data Factory.  

- Comfort using Git for version control and collaborative development.  

- Analytical mindset and knack for troubleshooting performance bottlenecks.  

- Clear written and verbal communication suitable for technical and non-technical audiences.  


Bonus Points

- Exposure to cloud warehouses (Snowflake, BigQuery, Redshift).  

- Experience with data modeling techniques—Kimball, Data Vault, or third-normal form.  

- Understanding of CI/CD tools (GitHub Actions, Jenkins) for data deployment.  

- Familiarity with scheduling/orchestration platforms like Airflow or Control-M.  

- Knowledge of data privacy regulations (HIPAA, GDPR, CCPA) and how they impact ETL logic.  


Growth Path

Excel here and step into roles such as ETL Engineer II, Data Quality Analyst, or Analytics Engineer. You will gain certifications, attend workshops, and rotate through advanced projects that deepen your skill set—from real-time streaming ingestion to orchestration of ML feature stores.


Why Join

You are empowered to shape scalable, resilient data foundations without bureaucratic drag. Expect fast feedback, mentorship sessions, and the freedom to propose fresh tooling. Remote work flexibility ensures you can thrive while balancing life and learning.