This role involves joining our dynamic data engineering team where you will focus on developing data pipelines, transforming data, participating in proof of concept for new data patterns, optimizing existing data feeds, and enhancing overall data processes.
Must Have Skills
- Strong knowledge of SQL and Python.
- Experience in Snowflake is preferred.
- Knowledge of AWS services such as S3, Lambda, IAM, Step Functions, SNS, SQS, ECS, DynamoDB.
- Proficiency in data movement technologies - ETL/ELT processes.
Good to Have
- Familiarity with DevOps practices, Continuous Integration and Continuous Delivery (CI/CD) tools like Maven, Jenkins, Stash, Control-M, Docker.
- Automation skills and understanding of REST APIs.
Primary Skills Needed
- Proficiency in RDBMS systems (Snowflake, Postgres) and Cloud platforms and Services (AWS - IAM, EC2, S3, Lambda, RDS, Timestream, Glue, etc.).
- Experience with data streaming tools such as Kafka.
- Hands-on experience with ETL/ELT tools and processes.
- Understanding of data modeling techniques would be advantageous.
Additional Details
This position requires a proactive individual who can thrive in a collaborative environment. You will be instrumental in driving data engineering initiatives, ensuring high-quality standards in data operations and contributing to the adoption of best practices in data management and processing.
About Us
Join our team and be part of a company that values innovation, collaboration, and continuous learning. We offer competitive compensation and a stimulating work environment where your expertise will make a significant impact.
Apply Now
If you are passionate about data engineering and possess the required skills and experience, we invite you to apply immediately. Take the next step in your career with us and contribute to shaping the future of our data infrastructure.