Companies you'll love to work for

Everywhere Ventures
Everywhere Ventures

Senior Data Engineer (Europe)

The Block

The Block

This job is no longer accepting applications

See open jobs at The Block.
Data Science
Remote · Europe
Posted on Saturday, January 13, 2024
Company Overview: Founded in 2018 and headquartered in New York City, The Block has rapidly emerged as a key player in the digital asset domain, offering research, analysis, and news. Our global team provides round-the-clock coverage of the cryptocurrency landscape, merging insights from the digital assets sector with traditional finance, technology, government, and markets. Our reach extends to a diverse audience via our website, The Block Pro, podcasts, and events.
Job Description:
You will be responsible for managing, securing, and scaling our ETL infrastructure in AWS and GCP. You will also provide development support for various backend systems such as our CMS/API and ETL data pipeline.
An effective Data Engineer should be able to effectively communicate with the Product & Engineering team while developing/maintaining ETL pipelines that support our data and analytics product efforts. The ideal candidate must be able to prioritize tasks and complete them in a timely manner.

Key responsibilities

  • Collaborate with our engineers, researchers, executives, and designers to develop new data capabilities using database systems (SQL, PostgreSQL) and data warehousing solutions within AWS and GCP.
  • Process unstructured data into structured and clean data.
  • Collaborate with our DevOps and Research teams to leverage ETL systems to help users analyze data relevant to a specific business problems.
  • Utilize data APIs to enable data scientists and analysts to query important data.
  • Maintain the security and health of our data pipelines.
  • Conduct code reviews and engage in pair programming within the Agile Development lifecycle.
  • Contribute to CMS development and other backend systems using PHP and/or Python.
  • Work with MWAA (Airflow) and Redash.


  • 7 or more years experience in data pipeline engineering.
  • Familiarity with programming languages such as Python and Scala.
  • Knowledge about source control software such as git.
  • Knowledge of the workings of Redis, Redshift, and BigQuery.
  • Experience in Business Intelligence products such as Google Data Studio, Looker, Domo, or Tableau.
  • Experience using development environments such as Docker, Vagrant, Virtual Box or similar platform.
  • Knowledge of algorithms and data structures to support data filtering and data optimization.
  • Strong familiarity with Unix / Linux operating systems and general command-line knowledge.
  • Knowledge of digital assets, crypto, and blockchain technologies.

Nice to have

  • Experience with CI/CD tools to support developer lifecycle.
  • Collaborate with our engineers to build and maintain infrastructure required for feature development.
  • Provide development support for CMS and other backend systems.
  • Familiarity managing data pipelines in cloud infrastructure in AWS or other cloud providers (Azure, GCP.)
  • Familiarity with AWS infrastructure suite (ECS, Elasticache, RDS, ALB, EC2, IAM, Cloudsearch) or other distributed systems.
  • Familiarity managing resources with IaC tools such as Terraform, Ansible, Puppet, or Chef.
  • Knowledge of building Docker images.
  • Industry-standard certifications.
  • Experience with the Atlassian suite.
  • Proficiency in scripting.

This job is no longer accepting applications

See open jobs at The Block.