Lead AWS Data Engineer
Job details
Title : Lead AWS Data Engineer Mandatory skills : Python, PySpark, SQL, and AWS services Years of experience : 10+ years Mode of Work : Remote AWS Data Engineer We are seeking a skilled AWS Data Engineer with expertise in various AWS services. The ideal candidate will have hands-on experience with Lambda, Glue, SNS, SQS, Step Functions, PySpark, Python, Athena, CloudWatch, S3, and more. The successful candidate should also have working experience with various data file formats such as JSON, XML, CSV, Parquet. Proficiency in SQL, and visualization tool experience with Looker or Power BI. Responsibilities : 1. Develop and maintain robust data pipelines using AWS Glue for efficient ETL processes. 2. Implement serverless computing solutions with AWS Lambda to automate tasks and processes. 3. Utilize SNS and SQS for efficient messaging and event-driven architecture. 4. Design and orchestrate data workflows using AWS Step Functions. 5. Leverage PySpark, Python, and SQL for data processing, analysis, and transformation. 6. Implement and optimize queries using AWS Athena for efficient querying of large datasets. 7. Monitor and manage resources and applications using AWS CloudWatch. 8. Manage data storage and retrieval using AWS S3. 9. Work with various data file formats, including JSON, XML, CSV, TSV, Parquet, and execute SQL queries as needed. 10. Utilize visualization tools such as Looker or Power BI for effective data representation. 11. Build end-to-end data pipelines, from conception to implementation, ensuring scalability and efficiency. 12. Hands-on experience with CI/CD tools such as Jenkins, GitLab/Github, Jira, Confluence, and other related tools. 13. Experience working with Delta Lake for efficient version control and data management. Qualifications : 7+ years experience as a Data Engineer in consumer finance or equivalent industry (consumer loans, collections, servicing, optional product, and insurance sales). Proven experience as a Data Engineer with a strong focus on AWS services. Proficiency in Python, PySpark, SQL, and AWS services for data processing and analysis. Hands-on experience with AWS Lambda, Glue, SNS, SQS, Step Functions, Athena, CloudWatch, and S3. Practical experience working with JSON, XML, CSV, TSV, Parquet file formats. Experience with visualization tools such as Looker or Power BI is a significant plus. Good understanding of serverless architecture and event-driven design. Hands-on experience with CI/CD tools, including Jenkins, GitLab/Github, Jira, Confluence, and other related tools. Comfortable learning about and deploying new technologies and tools. Organizational skills and the ability to handle multiple projects and priorities simultaneously and meet established deadlines. Good written and oral communication skills and the ability to present results to non- technical audiences. Knowledge of business intelligence and analytical tools, technologies, and techniques. Experience with Terraform is a plus. If Intrested. Please submit your CV to Khushboo@Sourcebae.com or share it via WhatsApp at 8827565832 Stay updated with our latest job opportunities and company news by following us on LinkedIn: :
Apply safely
To stay safe in your job search, information on common scams and to get free expert advice, we recommend that you visit SAFERjobs, a non-profit, joint industry and law enforcement organization working to combat job scams.