Data Engineer
Job details
Job SummaryWe are seeking a skilled and motivated Data Engineer with a strong background in data integration, ETL processes, and data warehousing. The ideal candidate will have over 5 years of experience in data engineering, with a good understanding of data architecture, pipeline creation, and big data technologies. This role requires a proactive individual who can design, build, and maintain data systems, collaborate with cross-functional teams, and ensure the highest standards of data quality and performance.Job ResponsibilitiesDesign, develop, and maintain scalable data pipelines for data ingestion, processing, and storage.Build and optimize data architectures and data models for efficient data storage and retrieval.Develop ETL processes to transform and load data from various sources into data warehouses and data lakes.Ensure data integrity, quality, and security across all data systems.Collaborate with data scientists, analysts, and other stakeholders to understand data requirements and deliver solutions that meet business needs.Monitor and troubleshoot data pipelines and workflows to ensure high availability and performance.Document data processes, architectures, and data flow diagrams.RequirementsCandidate Required SkillsBachelor's degree in Computer Science, Engineering, or a related field.5+ years of experience in data engineering and data architecture.Proficiency in SQL and at least one programming language (e.g., Python, Java, Scala).Experience with cloud data platforms (e.g., AWS, Azure, GCP) and their data services.Knowledge of ETL tools and frameworks (e.g., Apache NiFi, Talend, Informatica).Understanding of data modeling, data structures, and database design.Strong analytical and problem-solving skills.Strong communication and collaboration skills.Ability to work independently and as part of a team.Experience with data warehousing solutions (e.g., Redshift, BigQuery, Snowflake).Implement and maintain data integration solutions using industry-standard tools and technologies (e.g., Apache Spark, Kafka, Airflow, etc.).Hands-on experience with big data technologies (e.g., Hadoop, Spark, Kafka).Knowledge of data governance and best practices in data management.Familiarity with containerization and orchestration tools (e.g., Docker, Kubernetes).Experience with data visualization tools (e.g., Tableau, Power BI).BenefitsThis role offers the flexibility of working remotely in India.Sahiba8296043355 PRB
Apply safely
To stay safe in your job search, information on common scams and to get free expert advice, we recommend that you visit SAFERjobs, a non-profit, joint industry and law enforcement organization working to combat job scams.