from pyspark.sql.types import *userSchema = StructType([ StructField("lsoa_code", StringType(), True), StructField("borough", StringType(), True), StructField("major_category", StringType(), True), StructField("minor_category", StringType(), True), StructField("value", StringType(), True), StructField("year", StringType(), True), StructField("month", StringType(), True)])# Read data with StructurefileStreamDF = spark.readStream.schema(userSchema).csv("file:///home/ak/datasets/droplocation")# maxFilesPerTrigger specifies the number of files to be read at a timefileStreamDF = spark.readStream.schema(userSchema).option("maxFilesPerTrigger", 2).csv("file:///home/ak/datasets/droplocation")