Creating RDD from Text Files:

# Create RDD from a text file
rdd = spark.sparkContext.textFile("path/to/textfile.txt")

Replace “path/to/textfile.txt” with the actual path to your text file. Each line in the text file will become an element in the RDD.

Creating RDD from CSV Files:

To create an RDD from CSV files, you can use PySpark’s csv module or read the file as a text file and apply transformations to parse the CSV data.

Using csv module:

from pyspark.sql import SparkSession

# Create a SparkSession
spark = SparkSession.builder.getOrCreate()

# Read CSV file using SparkSession
df = spark.read.csv("path/to/csv_file.csv", header=True)

# Convert DataFrame to RDD
rdd = df.rdd

Replace “path/to/csv_file.csv” with the actual path to your CSV file.

Parsing CSV data from text file:

# Create RDD from a CSV file
rdd = spark.sparkContext.textFile("path/to/csv_file.csv") \
    .map(lambda line: line.split(","))  # Split each line into a list of values

Replace “path/to/csv_file.csv” with the actual path to your CSV file. The map transformation splits each line into a list of values.

Creating RDD from In-Memory Data:

You can create RDDs directly from in-memory data using the parallelize method or by converting a list or collection into an RDD.

# Create RDD from a list
data = [1, 2, 3, 4, 5]
rdd = spark.sparkContext.parallelize(data)

In this example, the list data is converted into an RDD using the parallelize method.

These examples demonstrate how to create RDDs from different data sources in PySpark. You can leverage these techniques to work with various data formats and sources, enabling distributed processing and analysis of large-scale datasets.