Jose Acosta Data Engineer Logo
HomeAbout MeServicesPortfolioBlog
ResumeLet's Talk

Data Engineer | Open to Full-Time Opportunities

10+yearsexperience.
Infrastructurethatsurvivesfailure.
High-availabilitywithautomaticfailover
10+yearsprofessionalexperience:ConstructionPM→QuantTrader→DataEngineer.
I'veseenbridgesfailandtradingsystemscrash.
NowIbuilddatainfrastructurethatsurvivesboth.
High-availabilitysystemswithautomaticfailoverandproduction-scaleprocessing
Apply for Full-Time
View Portfolio
Scroll to explore

What I Build

Data infrastructure engineered for real-world conditions—not just the happy path. I specialize in building systems that handle failures gracefully, recover automatically, and provide the observability needed to debug issues when they inevitably occur.

16.5 MB/s Peak Throughput | Production-Grade

SEC Financial Data Platform

Production-grade parser processing SEC filings with automatic recovery when parsing fails mid-document. In financial data, partial results are worse than no results—built with fault-tolerant design and data quality validation.

Architecture & Impact

Multi-engine parsing: 3 parsing engines (SGML, XBRL, HTML) with automatic fallback for maximum coverage

High throughput: 16.5 MB/s peak processing speed with parallel document handling

Fault-tolerant parsing: Automatic recovery from mid-document failures with comprehensive error handling

Data quality validation: Built-in validation ensuring financial data accuracy and completeness

PythonPostgreSQLSQLAlchemyDockerApache AirflowRedis

Key Features

Python, PostgreSQL, 3 parsing engines
Fault-tolerant parsing with auto-recovery
Data quality validation & monitoring
16.5
MB/s Speed
Prod
Grade
From Construction to Cloud

Data Architecture Principles

Four years managing construction projects taught me: bad architecture is expensive to fix later. I design data systems the same way—thinking about failure modes, maintenance costs, and what happens at 3x scale.

Modern stack (dbt, Snowflake)
Cost-optimization strategies
Reliability by design
Real-Money Trading Systems

Trading Data Infrastructure

Built end-to-end data platform powering algorithmic trading with real capital. When your pipeline fails at market open, you lose money every second—taught me to build systems that stay up.

WebSocket, Kafka, TimeScaleDB
Low-latency data ingestion
Production-tested with real capital

Why Work With Me?

I build data systems where failure has consequences:

Reliability First

Systems that recover automatically, not manually

Business-Aligned

Data quality tied to actual business impact

Cost-Conscious

Right-sizing and optimization from day one

The difference: I've lost real money to bad pipelines. Now I build so you don't have to.

Explore Data SolutionsSchedule Introduction Call

Ready to build data systems that work when business decisions depend on them?

About Me

Data Engineer with a non-traditional path that makes me better at the job.

I spent four years in construction project management learning how systems fail under pressure. Four years as a quantitative trader where bad data meant real money lost. Now three years building production data infrastructure where those lessons matter every day.

The pattern is clear: I've always worked where reliability isn't optional and data drives decisions. Construction taught me to design for failure modes. Trading taught me that data quality is non-negotiable. Data engineering is where both disciplines converge.

I specialize in high-availability systems, real-time pipelines, and cost-conscious architecture—because I've seen what happens when any of those fail.

Currently seeking full-time Data Engineering roles where complex data challenges need someone who thinks like an engineer, plans like a project manager, and measures impact like a trader.

I also take on select consulting engagements helping startups build data foundations that won't collapse at scale.

Available for Full-TimeConsulting Services
Data Engineer Profile

Data Engineering Built on Real-World Experience

My data engineering expertise comes from 10+ years across high-stakes environments. Each role taught me critical skills I now apply to building production data systems.

Currently Seeking Full-Time Data Engineering Roles
Interested in fintech, e-commerce, logistics, or SaaS companies with real-time data challenges

How My Background Strengthens My Data Engineering

Each career phase developed specific skills that make me a better data engineer today

4 YEARS

Construction PM

Data Architecture

Managing construction projects taught me to design for failure modes and scale. I now apply this to data architecture: planning for 3x growth, calculating resource constraints, and building systems that don't collapse under load.

Data Engineering Skill: Scalable architecture design
Data Engineering Skill: Capacity planning & cost optimization
4 YEARS

Quant Trading

Real-Time Pipelines
17.89% CAGR
2.34 Sharpe

Built algorithmic trading systems where bad data meant real money lost. This taught me to build data pipelines with obsessive data quality checks, sub-second latency requirements, and automatic failover.

Data Engineering Skill: Real-time streaming pipelines
Data Engineering Skill: Data quality validation & monitoring
3 YEARS • CURRENT

Data Engineering in Production

Currently building production data systems: SEC financial parser (16.5 MB/s throughput), ETL pipelines, data quality frameworks. Combining construction discipline with trading urgency to deliver reliable data infrastructure.

Core Skills: Python, SQL, Kafka, Airflow, dbt
Specialization: Fault-tolerant ETL & data quality

Bottom line: 10+ years of experience building systems where failure isn't an option. Now applying that to data engineering.

See Detailed Data Engineering Projects

Free Data Engineering Resources

Get battle-tested tools and templates that have saved companies $100K+ in development costs

Data Engineering Production Checklist

Complete 47-point checklist to ensure your data pipelines are production-ready

  • Avoid common pipeline failures
  • Reduce deployment time significantly
  • Improve system reliability
  • Save debugging time and costs

Data Pipeline ROI Calculator

Calculate the exact ROI of your data engineering investments in minutes

  • Estimate cost savings potential
  • Help justify data engineering budget
  • Compare different solution approaches
  • Support stakeholder presentations

Modern Data Stack Templates

5 proven architecture templates for different use cases and scales

  • Accelerate architecture planning
  • Proven patterns and approaches
  • Cost-conscious configurations
  • Implementation guidance included
Popular resources
Used by data teams
No spam guarantee

Calculate Your Data Engineering ROI

See exactly how much you could save with optimized data infrastructure

Your Current Situation

$
$

Your Potential Savings

Monthly Cost Savings
$0
25% infrastructure cost reduction
Time Savings Value
$0
50% reduction in maintenance time
Downtime Cost Avoided
$0
80% downtime reduction
TOTAL ANNUAL ROI
$0
Payback in 0.0 months
Schedule ROI Discussion

* Results are estimates based on typical improvements seen in similar projects. Actual results may vary depending on your specific infrastructure and requirements.

Ready to Build Reliable Data Systems?

Let's discuss your specific needs and create a custom solution

Efficient Delivery

Most projects completed in 6-10 weeks

Quality Focus

High uptime and reliable systems

Ongoing Support

Post-launch support and documentation

Start Your ProjectView Case Studies
10+ satisfied clients
Professional service
Usually responds within 4 hours

Stay Ahead in Data Engineering

Get weekly insights, case studies, and industry trends delivered to your inbox

Industry Insights
Technical Tips
Case Studies

No spam, unsubscribe at any time.•Privacy Policy

Technologies I Work With

Ready to Transform Your Data?

Let's discuss how I can help you build scalable data systems that drive real business value.

Usually responds within 24 hours
Free initial consultation
10+ years experience

Schedule Call

Free 30-min consultation

View Portfolio

See my work

Connect on LinkedIn

Professional network

Want to learn more about my services?

Explore Services

Your Data Solutions Partner

Data Engineer focused on building robust data pipelines, scalable architectures, and automated workflows. Enabling teams to make smarter, data-driven decisions through reliable systems and practical engineering skills.

Useful Links

  • Portfolio
  • About Me
  • LinkedIn
  • GitHub
  • Contact

Additional Pages

  • Trading Strategies
  • Privacy Policy
  • Terms of Service

Contact

Ready to Connect?

For full-time Data Engineering opportunities or consulting projects, let's discuss how I can help build reliable data infrastructure.

Schedule CallView Services
© 2025 Jose Acosta. All rights reserved.
Design & Development by
Jose Acosta
Git
Jupyter
pandas
AWS
Azure
Airflow
dbt
PySpark
Snowflake
PostgreSQL
Git
Jupyter
pandas
AWS
Azure
Airflow
dbt
PySpark
Snowflake
PostgreSQL
Python
University of Michigan
scikit-learn
Pytest
PySpark
AWS
Azure
Airflow
dbt
Snowflake
PostgreSQL
Python
University of Michigan
scikit-learn
Pytest
PySpark
AWS
Azure
Airflow
dbt
Snowflake
PostgreSQL