Running a Spark Batch Job on Google Cloud Dataproc As Data Engineers, one of the most powerful capabilities we often use is running batch Spark jobs on cloud clusters. Google Cloud Dataproc makes this seamless by letting us submit jobs directly to a managed Spark cluster. Here’s how I recently submitted a batch Spark job […]
Read MoreProblem StatementTruecaller deals with millions of user settings change events daily.Each event looks like this: id (long)name (string)value (string)timestamp (long)The goal: Group events by id. Convert (name, value) pairs into a Map. Always pick the value for each key that has the latest timestamp. Output a partitioned table for faster downstream queries. Example: id name […]
Read MoreIntroductionRetail today is not just about selling products – it’s about instant insights. Customers expect personalized offers, faster checkouts, and always-available inventory. For that, retailers need real-time data processing. In this tutorial, we’ll build a real-time data streaming pipeline for a retail company using Google Cloud Pub/Sub. Use CaseA retail chain with 500+ stores wants […]
Read MoreCatalyst Optimizer in Spark 3.0: Types, Use Cases, and Real-World CommandsIf you’re working with big data, speed matters. Spark 3.0 introduced major internal improvements—none more powerful than the Catalyst Optimizer and Adaptive Query Execution (AQE). 🧠 What is Catalyst Optimizer?It’s Spark’s rule-based and cost-based optimizer. Every time you write PySpark code, it: Converts it to […]
Read MorePySpark vs Pandas – When to Use What for Big Data “Should I use Pandas or PySpark for my data processing?” Let’s break this down across key dimensions with real examples: 1️⃣ Performance & ScalabilityFeature Pandas PySparkExecution Single-threaded Distributed (multi-node)In-memory limit Limited to RAM Designed for TBs+File handling Local files only HDFS, S3, GCS, JDBC, […]
Read More“If you have 100GB of data, how do you process it efficiently in PySpark?” It’s a classic interview question — but also a real challenge every data engineer faces when working with big data in production. Let’s break down the best practices step by step, using practical commands you can reuse in your projects. ✅ […]
Read MoreSource: Real-time CSV files with stock prices Target: JSON format files consumable by Data Analysts Goal: Automate the transformation and cataloging for query-ready analytics 🛠️ Step-by-Step Pipeline with AWS Services🔹 1. CSV Files Drop into S3Incoming files: stock data like stock_data_2025-06-16.csv S3 Source Bucket: s3://reedx-stock-raw/ These files were pushed by upstream providers or batched ingestion […]
Read More✅ 15 PySpark Interview Q&As for Data Engineers: pythonCopyEditfrom pyspark.sql.functions import udffrom pyspark.sql.types import StringType def convert_upper(text):return text.upper() upper_udf = udf(convert_upper, StringType())df.withColumn(“upper_name”, upper_udf(df[“name”]))
Read More🔹 Looking to replicate your on-prem Oracle database to Google Cloud with real-time changes? Oracle GoldenGate (OGG) provides a seamless solution for heterogeneous replication with minimal latency. 👉 In this post, I’ll walk you through a step-by-step process to configure Oracle GoldenGate replication from an On-Premises Oracle database to Google Cloud (Cloud SQL / Bare […]
Read MoreEssential PostgreSQL Queries Every Data Engineer Should Know 🚀 As a Data Engineer, mastering PostgreSQL queries can help you optimize database performance and troubleshoot issues efficiently. Here are some essential queries to keep in your toolkit! 🛠️ 1️⃣ Check Tablespace Size Monitor the disk space used by your tablespaces: SELECT spcname AS tablespace, pg_size_pretty(pg_tablespace_size(spcname))FROM pg_tablespace; […]
Read More