Home > Workload Solutions > Data Analytics > Guides > Design Guide—Modern Data Stack with Symcloud Platform and Delta Lake with Intel > Component validation
kubectl exec -it kakfa7-broker-01 /bin/bash
kubectl exec -it kakfa7-broker-01 -- /bin/bash -c "/usr/bin/Kafka-console-producer --broker-list kakfa7-broker-01:9092 --topic test < /home/appuser/TestDataOne.csv"
/spark/bin/pyspark --packages \
io.delta:delta-core_2.12:2.3.0,org.apache.hadoop:hadoop-aws:3.3.4,\
org.apache.spark:spark-sql-kafka-0-10_2.12:3.3.2 --conf
"spark.sql.extensions=io.delta.sql.DeltaSparkSessionExtension" --conf
"spark.sql.catalog.spark_catalog=org.apache.spark.sql.delta.catalog.DeltaCatalog"
spark container :import sysimport osimport pyspark.sql.functions as Ffrom pyspark.sql import SparkSessionfrom pyspark.context import SparkContextfrom pyspark.sql.functions import *from pyspark.sql.functions import *from delta.tables import *hadoopConf = spark._jsc.hadoopConfiguration()hadoopConf.set("fs.s3a.access.key", "<access_key>")hadoopConf.set("fs.s3a.secret.key", "<secret_key>")hadoopConf.set("fs.s3a.endpoint", "http://<s3_api_address>")hadoopConf.set("fs.s3a.impl", "org.apache.hadoop.fs.s3a.S3AFileSystem")hadoopConf.set("fs.s3a.aws.credentials.provider", "org.apache.hadoop.fs.s3a.SimpleAWSCredentialsProvider")df = (spark\ .readStream \ .format("kafka") \ .option("kafka.bootstrap.servers", "kakfa7-broker-01:9092") \ .option("subscribe", "kafkaspark") \ .option("startingOffsets", "latest") \ .load()) df.writeStream \ .format("delta") \ .outputMode("append") \ .option("mergeSchema", "true") \ .option("checkpointLocation", "/tmp/kafaktest") \ .start("s3a://sparkdelta/kafkaspark2/")df.selectExpr("CAST(value AS STRING)").show(20)