1.安装启动检查Mysql服务。
netstat -tunlp (3306)
2.spark 连接mysql驱动程序。
–cp /usr/local/hive/lib/mysql-connector-java-5.1.40-bin.jar /usr/local/spark/jars
3.启动 Mysql shell,新建数据库spark,表student。
select * from student;
create database spark; use spark; create table student (id int(4), name char(20), gender char(4), age int(4)); alter table student change id id int auto_increment primary key; insert into student values(1,‘Xueqian‘,‘F‘,23); insert into student values(2,‘Weiliang‘,‘M‘,24); insert into student values(3,‘Rongcheng‘,‘M‘,26); insert into student values(4,‘Guanhua‘,‘M‘,27); select * from student;
4.spark读取MySQL数据库中的数据
spark.read.format("jdbc").option("url", "jdbc:mysql://localhost:3306/spark?useSSL=false") ... .load()
spark.read.format("jdbc")\
.option("url", "jdbc:mysql://localhost:3306/spark")\
.option("driver","com.mysql.jdbc.Driver").option("dbtable", "student")\
.option("user", "root")\
.option("password", "1")\
.load()\
.show()
5.spark向MySQL数据库写入数据
studentDF.write.format(‘jdbc’).option(…).mode(‘append’).save()
from pyspark.sql.types import Row,StructField,StringType,StructType,IntegerType #下面设置两条数据,表示两个学生的信息 studentRDD = spark.sparkContext.parallelize(["5 zhao M 26","6 wang M 27"]).map(lambda line:line.split(" ")) #下面创建Row对象,每个Row对象都是rowRDD中的一行 rowRDD = studentRDD.map(lambda p:Row(p[1].strip(),p[2].strip(),int(p[3]))) # 下面设置模式信息 schema = StructType([StructField("name",StringType(),True),StructField("gender",StringType(),True),StructField("age",IntegerType(),True)]) #建立起Row对象和模式之间的对应关系,也就是把数据和模式对应起来 studentDF = spark.createDataFrame(rowRDD,schema) #写入数据库 prop = {} prop[‘user‘] = ‘root‘ prop[‘password‘] = ‘123456‘ prop[‘driver‘] = ‘com.mysql.jdbc.Driver‘ studentDF.write.jdbc("jdbc:mysql://localhost:3306/spark?useSSL=false",‘student‘,‘append‘,prop)