- Hands-on experience with Pyspark/Java Spark/Scala Spark
- Proficient understanding of distributed computing principles
- Proficiency with Data Processing: HDFS, Hive, Spark, Scala/Python
- Independent thinker, willing to engage, challenge and learn new technologies.
- Understanding of the benefits of data warehousing, data architecture, data quality processes, data warehousing design, and implementation,
- table structure, fact and dimension tables, logical and physical database design, data modeling, reporting process metadata, and ETL processes.
Requirements
- Client-facing skills: Solid experience working with clients directly, to be able to build trusted relationships with stakeholders.
- In-depth understanding of Data Warehouse, ETL concept and modeling structure principles
- Expertise in AWS cloud native services
- Hand-on experience in developing data processing task using Spark on cloud native services like Glue/EMR.
- Good to have experience Snowflake SQL queries against Snowflake Developing scripts using java scripts to do Extract, Load, and Transform data
...
Good to have experience with Snowflake utilities such as SnowSQL, SnowPipe, Python, Tasks, Streams, Time travel, Optimizer, Metadata Manager, data sharing, and stored procedures.
Excellent verbal and written communications skills
Ability to collaborate effectively across global teamsshow more