3+ years of experience with aws services including sqs, s3, step functions, efs, lambda, and opensearch.
strong experience in api integrations, including experience working with large-scale api endpoints.
proficiency in pyspark for data processing and parallelism in large-scale ingestion pipelines.
experience with aws opensearch apis for managing search indices.
terraform expertise for automating and managing cloud infrastructure.
hands-on experience with aws sagemaker, including working with machine learning models and endpoints.
strong understanding of data flow architectures, document stores, and journal-based systems.
experience in parallelizing data processing workflows to meet strict performance and sla requirements.
familiarity with aws tools like cloudwatch for monitoring pipeline performance.
additional preferred qualifications:
strong problem-solving and debugging skills in distributed systems.
prior experience in optimizing ingestion pipelines with a focus on cost-efficiency and scalability.
solid understanding of distributed data processing and workflow orchestration in aws environments.
soft skills:
strong communication and collaboration skills, with the ability to work effectively with cross-functional teams.
ability to work in a fast-paced environment and deliver high-quality results under tight deadlines.
analytical mindset, with a focus on performance optimization and continuous improvement.