The financial industry is growing at a record pace, but our data providers are still stuck in the past — with cumbersome onboarding processes, complicated APIs, slow infrastructure, and expensive licensing costs.

Databento is the next-generation market data provider — with the radical idea that you should only pay for the data that you use. We power the world's largest finance and fintech institutions and lower the barrier of entry for small startups.

Since starting in 2019, we've raised over $27.8M in funding and have over 2,000 companies signed up pre-launch. Our team consists of former data users from firms like Two Sigma, Belvedere, Pico, Bloomberg, Pinterest, and Google.

We offer health, dental, disability, and life insurance benefits, as well as 401(k) matching and visa sponsorships. We accommodate 100% remote work, with teammates living around the globe and paid in their local currency.

Responsibilities

  • Architecting and building high performance data processing applications.
  • Implementing data pipelines and ETL workflows.
  • Writing market data feed handlers and parsers.
  • Working with product team and other engineers to design and implement the backend, APIs and wire protocols behind our core product.
  • Acting as an inhouse technical reference on areas of domain expertise and setting standards for best practices for software development.
  • Improving release cycle, testing and internal documentation.

Required experience

  • Experience developing in at least two of the following languages: C, C++, Python, Rust.
  • 5+ years designing and writing applications in at least one of the following domains: electronic trading, information retrieval, data storage, database design, machine learning, distributed computing, security, natural language processing, enterprise SaaS.
  • Degree in computer science, applied mathematics, or a related field.

Desired qualifications

  • Experience working in a financial institution, financial data vendor, or any significant provider or consumer of data analytics.
  • Experience with financial data (multicast feeds, time series, security references, machine-readable news) or analytics.
  • Experience implementing backends for web APIs or wire protocols with heavy traffic.
  • Experience with socket programming.
  • Familiarity with CI/CD, deployment, containerization and orchestration (Jenkins, TeamCity, GitLab CI, Docker, Kubernetes).
  • Familiarity with ETL workflows or automated data pipelines (Apache Airflow).
  • Broad knowledge of performance optimizations (memory hierarchy and I/O bounds, CPU architecture, kernel bypass networking, lock-free algorithms, compiler behavior, zero copy, instruction pipelining and prefetching).
  • Familiarity with clustered databases and messaging platforms (Vitess, ClickHouse, kdb, Vertica, Pulsar).
  • Experience working within a Linux environment.

Our recruiting data suggests that underrepresented applicants often downplay their skills. Even if your experience doesn’t exactly match the qualifications listed, we still want to hear from you. Please apply! 

Apply for this Job

* Required
resume chosen  
(File types: pdf, doc, docx, txt, rtf)
cover_letter chosen  
(File types: pdf, doc, docx, txt, rtf)