management of all data pipelines, including 3rd party integrations, internal ETL procedures, DWH management and modeling.
Operate in a real big data environment (ingestion of billions of daily events) into multiple endpoints.
Communicate with stakeholders and responsible for all data related developments.
Is an independent, self-learner who understands business processes and able to translate business needs into data models.
Has extensive experience in writing data pipelines - spark, hive, pig and similar.
Can quickly learn new technologies and have a good level of Python.
Is highly proficient in SQL, including performance tuning.
Has proficiency with database modeling in both relational and data warehouse environments.
Has vast experience in cloud environments (GCP/AWS/Azure) and modern data warehouse solutions (Cloudera Impala, BigQuery etc.).
Is highly skilled in data visualization (Tableau//Sisense/similar) – bonus
In each category we need him to have good knowledge with at least one technology: