ETL / Data Pipeline Specialist
Confidential
Posted: March 31, 2026
Interested in this position?
Create a free account to apply with AI-powered matching
Quick Summary
We are seeking an ETL / Data Pipeline Specialist to join our data infrastructure team, focusing on building and maintaining data pipelines for our clients.
Required Skills
Job Description
At TechBiz Global, we are providing recruitment service to our TOP clients from our portfolio. We are currently seeking an ETL / Data Pipeline Specialist to join one of our clients' teams. If you're looking for an exciting opportunity to grow in a innovative environment, this could be the perfect fit for you.
Location: Remote
Reports To: Senior ETL Specialist
Employment Type: Full-Time Employee
Working Hours: Overnight shift aligned with US Central Time
Role Overview
We are looking for an ETL / Data Pipeline Engineer to join our data infrastructure team. This role operates on an overnight schedule aligned with US Central Time, providing critical coverage for automated scraping and ingestion systems that run during US business hours and overnight windows. You will help manage and monitor our existing fleet of web scrapers, build new data collection pipelines for regulatory and energy market sources, and contribute to the broader evolution of our ETL architecture.
This is a hands-on engineering role. You will work closely with our Houston-based data and analytics teams to ensure data flows reliably from source to production databases, and that our clients receive accurate, timely intelligence every morning.
Key Responsibilities
Overnight Operations & Data Quality
• Monitor and manage overnight scraper and ingestion runs, triaging failures and applying fixes in real time to minimize data gaps before US market open
• Verify data completeness and quality across all automated feeds, flagging anomalies and coordinating with the Houston team on persistent issues
• Maintain run logs, error documentation, and escalation notes for seamless async handoffs
New & Expanded Data Collection
Build and maintain scrapers, parsers, and ingestion pipelines across a growing set of energy market data domains, including but not limited to:
• Pipeline operator portals, electronic bulletin boards, and related filings (notices, maintenance, capacity, gas quality, customer indices)
• Government and regulatory agency databases at the federal, state, and provincial level across North America
• International energy data sources covering European, Canadian, and Mexican supply, demand, and power markets
• Emissions and environmental reporting systems
• Financial and corporate filings, including public company disclosures and production reporting
• Geospatial and mapping data related to production, infrastructure, and market geography
ETL Architecture & Enhancement
• Contribute to the design and build-out of our broader ETL infrastructure, including scheduling, orchestration, and error handling
• Write transformation logic to clean, normalize, and load raw data into PostgreSQL staging and production tables
• Optimize existing pipelines for performance, reliability, and cost efficiency
• Help build monitoring dashboards and alerting for pipeline health and data freshness
• Document data lineage, schema changes, and pipeline dependencies
Technical Environment
• Languages: Python (primary), SQL, Bash scripting
• Database: PostgreSQL, Mongo and Snowflake
• Infrastructure: AWS (EC2, S3, Lambda), Docker
• Scraping: Selenium, Playwright, BeautifulSoup, Scrapy, or similar frameworks
• Orchestration: cron, Airflow, or equivalent scheduling tools
• Version Control: Git / GitHub
• Communication: Slack, with async handoffs to US-based team