SYSTUM Inc. - US
2019 - 2020
This project involved developing and maintaining data infrastructure for a company, implementing automated workflows and cloud solutions to enhance data processing and operational efficiency.
Engineered and maintained robust data pipelines, implementing Apache Airflow to orchestrate complex data integration workflows, improving data reliability and reducing manual intervention.
Designed and deployed scalable cloud infrastructure on AWS using CloudFormation, leveraging Lambda functions, EC2 instances, Fargate containers, Aurora databases, Athena for serverless querying, and Redshift for data warehousing.
Implemented CI/CD pipelines with GitHub, Jenkins, Docker Compose, and Docker to automate code testing, building, and deployment, ensuring rapid and reliable software delivery.
Integrated monitoring and logging tools (DataDog, Sentry) to proactively identify and resolve performance bottlenecks and system errors, enhancing overall system stability.
Developed backend components of data processing applications using the Python Django framework, contributing to efficient data management and analysis.
Utilized NEO4J graph database and Cypher/OpenCypher query language to model complex relationships within datasets, enabling deeper insights and analysis.
Technologies: Apache Airflow, AWS, CloudFormation, Lambda, EC2, Fargate, Aurora, Athena, Redshift, CI/CD, GitHub, Jenkins, Docker Compose, Docker, DataDog, Sentry, Python Django, NEO4J, Cypher, OpenCypher