As a Data Engineer at Praava Health, you are in charge of transforming data into a format that can be easily accessed and analyzed. You develop and maintain data pipelines to read data from external data sources, format them in a way that is easy to analyze, and store them in data stores that are scalable, stable, and accessible by other services in the organization. You are also responsible for writing, running, and maintaining business reports. In addition, you also have to make the data available to other applications by developing APIs and maintaining them.
Key Responsibility Areas:
- Integrate internal and external data sources into the Data Warehouse (DWH).
- Perform Extract, Transform, Load (ETL) on various data to the data warehouse.
- Develop data pipelines with batch and stream processing capabilities.
- Create optimized data model(e.g. Star-Schema/Hybrid), normalized database tables, relationships, views, procedures, events, and triggers for data warehouse and other in-house application databases.
- Ensure data integrity, usability, and standards.
- Create queries and dashboards in analytical tools connected to the data warehouse.
- Plan and execute regular data backup.
- Manage database roles, users, and access.
- Develop APIs to make data available to other applications.
Qualifications:
- B.Sc/M.Sc in Computer Science/Computer Engineering/Electrical Engineering.
Experience:
- Minimum 3 years of experience in Data Engineering roles.
Technical Skills/Knowledge:
- Solid knowledge in SQL databases (e.g. Postgres and Oracle), should be able to write complex queries.
- Knowledge in PL/SQL; writing procedures, functions, triggers, etc.
- Performance Optimization, Database Design, Database Administration, and Reporting in both OLTP and Data Warehouse systems.
- Solid experience in design, development, and maintenance data loading and manipulation framework (ETL) with complex transformations for Big Data aggregation
- Strong analytical skills and knowledge of data software development.
- Experience in working with BI tools e.g. Metabase and schedulers e.g. Apache Airflow is mandatory.
- Passion for solving intricate business problems using data science.
- Experienced in large-scale Database Analysis Languages (e.g. SQL, Python(Pandas, Numpy)), and Data Visualization tools.
- Strong understanding of working with Linux.
Behavioral Competencies:
- Honest
- Service Orientation
- Quantitative Analysis
- Teamwork