Location India, Remote (Work from Anywhere in India)
Minimum Qualification B. Tech. (or equivalent) from an accredited institution
Indicative Experience 8+ years
Domain Preferably Life-Sciences/Pharma.
Customer Profile Captive research and development pods for a $ 500 million group of pharma data
research companies that help patients gain access to life-saving therapies. We help our clients navigate the complexities at each step of the drug development life cycle, from pipeline to patient.
Other benefits Health Insurance, Provident Fund, Life Insurance, Reimbursement of Certification
Expenses, Gratuity, 24x7 Health Desk
About the Company
We are headquartered in Pittsburgh, USA,, with locations across the globe we are a team of thoughtful experts driven by the power of our client's unique ideas. We also have micro-offices in Hyderabad, Chennai, Bengaluru, and Delhi NCR in
India. While technical expertise is ingrained into Agilite's DNA, we are more than just engineers and developers we are trusted product strategists. We pride ourselves on being a ready resource for critical market insights with the knowledge and experience required to design, build, and scale big ideas to serve our growing list of customers in the USA and Europe.
Our preferred working model is "Work from Anywhere" (WFA). In addition, you can also decide on your work schedule: All
we need is the outcome. Our people-centric culture is built on the belief that extraordinary employees create amazing things. Work with us and attain your Ikigai in a place where your aspirations and business objectives intersect!
Role Summary
In this role as a Data Engineer, you will report to the Lead Data Engineer and will be based on our Data Science team. Throughout 2021, we have been investing in developing our own internal Data Science team. We have developed a number of processes and pipelines, and plan on making many more as the team continues to grow and expand its scope.
Responsibilities
- Design and implement data pipelines on high-volume datasets in AWS using Airflow for orchestration.
- Design suitable data models for new data sources and structures in Snowflake
- Assist the SREs with deployment to production AWS accounts for completed projects.
- Work with data scientists and implement predictive and NLP models in the cloud using task-appropriate tooling.
- Work with the Lead Data Engineer, define and size new projects in Jira. Assist other Data Engineers in the team with technical challenges.
Mandatory skill sets:
- You have proven experience as a data architect and can implement solutions for complex Data Engineering tasks.
- You have a bachelor's degree in computer science, a similar technical field of study or equivalent with 8 years of experience
- You have great knowledge of AWS computer options, including Glue, EC2, ECS, EKS, etc.
- You have excellent knowledge of advanced data technologies, including Snowflake and Airflow.
- You have experience with Python and Py Spark for Data Engineering tasks.
- Data Modelling: Snowflake, Data Pipelines, and Star Schema
- Cloud Compute and DB Technologies: Preference: AWS Services (RDS), AWS Lambda, and S3 as part of building custom data pipelines for Data Lakes. Cloud Compute services- Preference: You have experience using Apache Airflow as an orchestration engine for data flow management.
Nice to have:
- Experience with automated deployment on AWS infrastructure with Terraform and GitHub Actions in a multi-environment CICD pipeline.
- You have experience in orchestrating automated data pipelines with Apache, Airflow, Preference: AWS
- Familiarity with biological or pharmaceutical data and integration with complex data science models
- Snap Knowledge (Data Orchestration)
- Microservices and Enterprise Cloud APIs
- Knowledge of Tableau or any reporting tool, especially efficient usage with large datasets.