Job Description

    Key Responsibilities:
    Design, build, and maintain scalable ETL/ELT pipelines to collect and process large datasets.

    Develop and manage data architecture across data lakes, warehouses, and cloud platforms.

    Ensure data quality, consistency, and governance throughout the pipeline.

    Work closely with data analysts, scientists, and product teams to understand data needs and deliver clean, reliable data.

    Monitor and optimize the performance of data systems.

    Develop and maintain automated data workflows using tools like Airflow or dbt.

    Document data processes, definitions, and pipeline workflows.

    Requirements

    Required Skills & Experience:
    Proficiency in SQL and at least one programming language (Python, Java, or Scala).

    Experience with cloud platforms like AWS (e.g., S3, Redshift, Glue), GCP (e.g., BigQuery, Dataflow), or Azure.

    Hands-on experience with data pipeline tools (e.g., Apache Airflow, dbt, Kafka, Spark).

    Strong understanding of data modeling, data warehousing, and performance tuning.

    Familiarity with both structured and unstructured data systems.

    Version control with Git and experience working in Agile environments.

    What we offer

    Good Work environment

    Salary

    The salary range for this position is between $ 50.000-$75.000


    Contact person



    Cityjobs.info connects job seekers and employers in various fields, making the job market accessible and inclusive for everyone.