Flexport helps more than 10,000 clients and suppliers lead all aspects of their supply chain operations. Started in 2013, we've raised over $1.3B from investors that include SoftBank Vision Fund, Founders Fund, GV, First Round Capital and Y Combinator.
With offices on three continents, our team is as global as our client base and we’re excited to continue building a product and service they love. Wherever you are, whichever role you play, you’re guaranteed to share your day with committed, encouraging, and resourceful team members.
What you’ll do:
You will develop resilient pipelines from a variety of data sources, both internal and external
You will implement Comprehensive Testing and Continuous Integration frameworks for schema, data, and functional processes/pipelines
You will design the scheduling framework (Airflow) to automatically recognize and set-up object dependencies for robust and timely data delivery
You will be part of a close-knit data engineering team that ships new code every day
What you’ll need:
You have 5+ years experience developing data frameworks using Python
You have 5+ years of using SQL for data manipulation in a fast-paced work environment
You have experience with Airflow is strongly preferred
Data Warehousing development and fundamentals preferred
You have a passion for business-oriented data development
Where you’ll work:
Our flagship office is on Market Street. We’re dog-friendly. We do a lot of 1:1s at the beautiful Yerba Buena Gardens and, as a truly global company, we don’t let geography slow us down. We rotate hack-a-thons and move Flexporters all around the world.
Investing your time with Flexport means having immediate impact, all over the world. You’re empowered to do what’s best for everyone and trusted to make the right decisions when and where you need them. Join our collective of entrepreneurs and improve the world’s experience in global trade.
We are an equal opportunity employer and value diversity at our company. We do not discriminate on the basis of race, religion, color, national origin, gender, sexual orientation, age, marital status, veteran status, or disability status.
Our pipelines are built using Python and SQL, scheduled in Apache Airflow, glued with bash. Near Real-Time Data Warehouse in Postgres and Data Lake in Redshift. All of our infrastructure is based on AWS. Our BI data is only 15 minutes behind production servers.