Data Engineer with solid experience in designing, building, and optimizing scalable batch and streaming data pipelines, supporting Business Intelligence, Analytics, Machine Learning, and Artificial Intelligence initiatives. Proven track record working in Big Data and cloud environments (AWS, Azure, and GCP), with strong expertise in modern tools such as Apache Spark, Kafka, Databricks, Snowflake, MongoDB, PostgreSQL, and Elasticsearch.
With experience collaborating in agile, cross-functional teams, and holding a background in Engineering along with an MBA in Project Management and BIM, I combine strong technical expertise with a strategic mindset, delivering data solutions that drive efficiency, innovation, and business value.
Summarily, my curiosity extends to the domains of Machine Learning, Artificial Intelligence, Data Engineering, and new technologies. I am always enthusiastic to embrace innovations.
-Cloud Computing (AWS, Azure, GCP) - Big Data - Data Warehouse - Data Lake - Data Lakehouse - IaC (Terraform) - Docker, Kubernetes - PostgreSQL - pgAdmin - ETL x ELT pipeline - Python -PySpark - SQL - Google BigQuery - Amazon Redshift - AWS Lake Formation - AWS Cloud Formation - Amazon S3 - Amazon EMR - Amazon Athena - AWS Glue - Data Modeling - Data Quality - Data Lineage - Apache Hadoop HDFS - Apache Hadoop YARN - Apache Spark - Apache NiFi - Apache Kafka - Apache Airflow - Apache Zookeeper - Airbyte - Batch and streaming data acquisition - Data Lake On-premises - Data Lake AWS - Databricks - Snowflake - Dremio - Metabase - Looker Studio - Kerberos Security Protocol - PENTEST - Kali Linux - Linux
