Data EngineerSenior Associate/Manager
Experience
3-6 years
Location
Mumbai/Bangalore
Type
Full Time
Role Summary
AuxoAI is seeking a skilled and experienced Data Engineer to join our dynamic team. The ideal candidate will have 3-6 years of prior experience in data engineering, with a strong background in AWS (Amazon Web Services) technologies. This role offers an exciting opportunity to work on diverse projects, collaborating with cross-functional teams to design, build, and optimize data pipelines and infrastructure.
Additionally, you are required to stay up to date on AI trends and best practices, sharing knowledge with the team to foster continuous learning and improvement.
If you are proactive, self-motivated, and passionate about AI, this role offers an exciting opportunity to make a meaningful impact and drive business transformation.
Responsibilities
- Design, develop, and maintain scalable data pipelines and ETL processes leveraging AWS services such as S3, Glue, EMR, Lambda, Aurora, RDS, Lake formation, Athena, DMS and Redshift
- Collaborate with data scientists and analysts to understand data requirements and implement solutions that support analytics and machine learning initiatives
- Optimize data storage and retrieval mechanisms to ensure performance, reliability, and cost-effectiveness
- Implement data governance and security best practices to ensure compliance and data integrity
- Troubleshoot and debug data pipeline issues, providing timely resolution and proactive monitoring
- Stay abreast of emerging technologies and industry trends, recommending innovative solutions to enhance data engineering capabilities
- Develop and maintain strong relationships with key clients, serving as a trusted advisor and strategic partner while identifying opportunities for upselling and cross-selling additional services to drive revenue growth
Qualifications
- Bachelor's or Master's degree in Computer Science, Engineering, or a related field
- 3-6 years of prior experience in data engineering, with a focus on designing and building data pipelines
- Proficiency in AWS services, particularly S3, Glue, EMR, Lambda, Aurora, RDS, MWAA, Lake formation, Athena, DMS and Redshift
- Strong programming skills in languages such as Python, Java, or Scala
- Proficient in Spark, Databricks and messaging queues like RabbitMQ and Kafka
- Experience with SQL and NoSQL databases, data warehousing concepts, and big data technologies
- Familiarity with containerization technologies (e.g., Docker, Kubernetes) and orchestration tools (e.g., Apache Airflow)