Advertisement
Not a member of Pastebin yet?
Sign Up,
it unlocks many cool features!
- from pyspark.sql import SparkSession
- spark_jars_packages = ",".join(
- [
- "org.apache.spark:spark-sql-kafka-0-10_2.12:3.3.0"
- ]
- )
- spark = (
- SparkSession.builder
- .master("local")
- .appName('test connect to kafka')
- .config("spark.jars.packages", spark_jars_packages)
- .getOrCreate()
- )
- df = (spark.read
- .format('kafka')
- .option('kafka.bootstrap.servers', 'rc1b-2erh7b35n4j4v869.mdb.yandexcloud.net:9091')
- .option('kafka.security.protocol', 'SASL_SSL')
- .option('kafka.sasl.mechanism', 'SCRAM-SHA-512')
- .option('kafka.sasl.jaas.config','org.apache.kafka.common.security.scram.ScramLoginModule required username=\"de-student\" password=\"ltcneltyn\";')
- .option("subscribe", "persist_topic")
- .load())
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement