PySpark

PySpark

This page contains PySpark tutorials. Dive into the world of PySpark, the powerful Python API for Apache Spark, designed for big data processing and analytics. Our hands-on tutorials equip you with the skills to handle large-scale data and perform distributed computing with ease. Learn step-by-step how to leverage PySpark's rich ecosystem to build data pipelines, execute complex transformations, and perform machine learning on big datasets. Our hands-on tutorials will help you master PySpark.

42 posts
PySpark - Convert Column from String to Date Format
Academy Membership PySparkPython

PySpark - Convert Column from String to Date Format

Introduction In data processing, it's common to find date fields as strings. Converting these string representations into proper date formats is crucial for accurate data analysis and processing. In this tutorial, we will explore how to convert a string to a date column in a PySpark DataFrame. Import...

PySpark - Extract a Substring from a DataFrame Column
Academy Membership PySparkPython

PySpark - Extract a Substring from a DataFrame Column

Introduction When dealing with large datasets in PySpark, it's common to encounter situations where you need to manipulate string data within your DataFrame columns. One such common operation is extracting a portion of a string—also known as a substring—from a column. In this tutorial, we will...

PySpark - Count Rows and Columns of a DataFrame
Academy Membership PySparkPython

PySpark - Count Rows and Columns of a DataFrame

Introduction In data processing and analysis with PySpark, it's often important to know the structure of your data, such as the number of rows and columns in a DataFrame. This is crucial for various operations, including data validation, transformations, and general exploration. In this tutorial, we'll...

PySpark - Count Distinct Values of a DataFrame Column
Academy Membership PySparkPython

PySpark - Count Distinct Values of a DataFrame Column

Introduction In this tutorial, we want to count the distinct values of a PySpark DataFrame column. In order to do this, we use the distinct().count() method and the  countDistinct() function of PySpark. Import Libraries First, we import the following python modules: from pyspark.sql import SparkSession from pyspark.sql....

PySpark - How to create and use Broadcast Variables
Academy Membership PySparkPython

PySpark - How to create and use Broadcast Variables

Introduction In distributed computing environments like Apache Spark, efficient data handling is critical for performance. One useful feature for optimizing computations is broadcast variables. Broadcast variables allow you to share large read-only data across all nodes in a Spark cluster without duplicating the data for each task. In this tutorial,...

PySpark - Create Embedding Vectors with Sentence-Transformers
Academy Membership PySparkPython

PySpark - Create Embedding Vectors with Sentence-Transformers

Introduction In today's data-driven world, understanding text data is crucial across various domains, from data analysis to engineering and architecture. However, dealing with text data often requires converting it into numerical representations for machine learning models to process efficiently. This is where embedding vectors come into play, offering...

You’ve successfully subscribed to Deep Learning Nerds | The ultimate Learning Platform for AI and Data Science
Welcome back! You’ve successfully signed in.
Great! You’ve successfully signed up.
Success! Your email is updated.
Your link has expired
Success! Check your email for magic link to sign-in.