The main cloud ecosystem being used is AWS.
Skill
- Implementation experience on building large scale data applications from scratch (initial stages)
- Programming experience in Python or Java
- Good experience of deploying applications on AWS and usage of its services
- Must have experience with Hadoop Distributed File System (HDFS), Amazon Simple Storage Service (S3)
- Must have good experience on SQL
- Data organization in Data Lake (experience in Delta Lake or Databricks is added advantage)
- Detailed understanding of Data pipeline creation
- Detailed experience of Data ingestion
- Techno functional experience working with technical team (Data Engineering / data science) and the business (functional) teams