Experience as a Data Engineer with strong track record of designing and implementing data solutions.
• Experience in programming language as Python or Java, with experience in building data pipelines and workflows.
• Experience with cloud data warehousing technologies, such as Snowflake and Redshift. ...
• Experience with distributed computing frameworks such as Cloudera, Apache Hadoop and Spark.
• Experience with Cloud platforms such as AWS, Azure or Google Cloud Platform.
• Experience with AWS cloud services, such as S3, EC2, EMR, Glue, CloudWatch, Athena, Lambda.
• Experience with containerization and orchestration technologies, such as Docker and Kubernetes.
• Experience with building CI/CD pipeline using tools, such has GitLab and Bitbucket.
• Experience with data pipeline orchestration tools, such as Airflow and Jenkins.
• Knowledge of database concepts, data modelling, schemas and query languages, SQL and Hive.
• Knowledge of data visualization and reporting tools, such as MicroStrategy, Tableau and Power BI.
• Knowledge of data quality and monitoring techniques and tools, such as Great Expectations or similar.
• Knowledge of data governance processes, lineage, cataloging, dictionaries using tools, such as DataHub or similar.
• Knowledge of streaming data processing and real-time analytics technologies, such as Kafka.
experience
8show more Experience as a Data Engineer with strong track record of designing and implementing data solutions.
• Experience in programming language as Python or Java, with experience in building data pipelines and workflows.
• Experience with cloud data warehousing technologies, such as Snowflake and Redshift.
• Experience with distributed computing frameworks such as Cloudera, Apache Hadoop and Spark.
• Experience with Cloud platforms such as AWS, Azure or Google Cloud Platform.
• Experience with AWS cloud services, such as S3, EC2, EMR, Glue, CloudWatch, Athena, Lambda.
• Experience with containerization and orchestration technologies, such as Docker and Kubernetes.
• Experience with building CI/CD pipeline using tools, such has GitLab and Bitbucket.
• Experience with data pipeline orchestration tools, such as Airflow and Jenkins.
• Knowledge of database concepts, data modelling, schemas and query languages, SQL and Hive.
• Knowledge of data visualization and reporting tools, such as MicroStrategy, Tableau and Power BI.
• Knowledge of data quality and monitoring techniques and tools, such as Great Expectations or similar. ...
• Knowledge of data governance processes, lineage, cataloging, dictionaries using tools, such as DataHub or similar.
• Knowledge of streaming data processing and real-time analytics technologies, such as Kafka.
experience
8show more