As a Big Data Engineer, I leverage my 10+ years of experience to deliver robust, scalable, and innovative data solutions that drive business value and performance. I am adept at transforming raw, unstructured data into strategic insights that inform business decisions and optimize processes. I am passionate about using data as a catalyst for enterprise-wide innovation and growth.
My core competencies include architecting and developing data systems that capably handle, process, and analyze colossal data volumes using distributed computing technologies such as Hadoop, Spark, Flink, and Kafka.I also orchestrate seamless data pipelines using Python or Scala, and construct intricate SQL queries.
Additionally, I am proficient in cloud platforms such as AWS, Google Cloud Platform, and Azure, where I deploy solutions, manage resources, and use data-centric tools like Kinesis, AWS Lambda, EMR, SQS, SNS, Redshift, and Snowflake.
Furthermore, I am skilled in data visualization and storytelling, using tools like Google Dataflow/Apache Beam, Google Pub/Sub, Bigquery, Druid, HBase, Hive, and others to present complex data findings in a compelling and accessible way for stakeholders.
Beyond my technical arsenal, I offer a keen understanding of business dynamics and data-driven growth. I align data strategies with business goals and deliver results that not only meet but exceed expectations.
Projekthistorie
03/2021
-
bis jetzt
Big Data Consultant
Klarna
(Banken und Finanzdienstleistungen, 5000-10.000 Mitarbeiter)
Merchant Product Data Catalog: Designed and delivered an up-to-date, standardized catalog optimized for downstream service consumption.
Real-Time Behavioral Data Ingestion: Integrated real-time behavioral data into a modern Lakehouse framework, enabling instant insights and decision-making.
End-to-End Leadership Across Projects:
Architectural Oversight: Led design strategies to ensure scalable and seamless data integration across platforms.
ETL & Real-Time Processing: Developed Spark-based ETL pipelines for the catalog and enabled real-time data processing for behavioral ingestion.
Lifecycle Management: Managed all phases from conceptualization, design, and implementation to DevOps-related tasks, ensuring system efficiency and resilience.
Enhanced data accessibility and system responsiveness across the company's data ecosystem.
Established blueprints for future data initiatives, emphasizing comprehensive architectural and engineering practices.
04/2022
-
06/2022
Apache Flink Consultant
Metro Markt
Optimization: Fine-tuned streaming applications to achieve superior performance while ensuring consistency and fault tolerance.
Consultation & Education: Collaborated with engineering teams, imparting knowledge and expertise in crafting new Flink jobs. Advocated for and instructed on best practices to ensure optimal design and implementation.
Key Technologies & Stack:
Streaming & Processing: Apache Flink, optimized for high-throughput and low-latency performance.
Event Streaming & Change Data Capture: Kafka and Debezium, ensuring real-time data handling and integrity.
Programming: Java, tailored for streamlined streaming applications.
Cloud & Infrastructure: GCP, providing a scalable and robust environment for deployment and operation.