Tuesday, April 13, 2021

JDBC connection with Spark using pyspark

 from pyspark.sql.session import SparkSession

from lib2to3.tests.support import driver



def orcl_conn():

    spark =SparkSession.builder.getOrCreate()

    

    dbconn = spark.read \

    .format("jdbc").option("url","jdbc:oracle:thin:scott/tiger@//LAPTOP-K8VAJ401:1521/xe")\

    .option("dbtable","emp")\

    .option("user", "scott") \

    .option("password", "tiger") \

    .option("driver", "oracle.jdbc.driver.OracleDriver").load()

    

    empdf=spark.read.format('csv').option('header','True').load('D:/projects/data/test/emp.csv')

    

    empdf.write \

    .format("jdbc").option("url","jdbc:oracle:thin:scott/tiger@//LAPTOP-K8VAJ401:1521/xe")\

    .option("dbtable","empdf")\

    .option("user", "scott") \

    .option("password", "tiger") \

    .option("driver", "oracle.jdbc.driver.OracleDriver").save()

    

    empdf.show()

    

    #dbconn.printSchema()

    

orcl_conn()

No comments:

Post a Comment