Code Examples
MongoDB
import com.mongodb.spark._
import com.mongodb.spark.sql._
import com.mongodb.spark.config._
val dfFA = spark.read.option("database", "foxconn-analytics")
.option("collection", "collect_pageview")
// option uri can be omitted if it has been defined in Spark interpreter
.option("uri", "mongodb://mongo201,mongo202,mongo203/?replicaSet=bigdata&readPreference=secondaryPreferred")
.mongo()case class EHAVE_TBL(
CNAME: String, CCOVEY: String, ISCORE: String,
IFPASS: String, FACTORY: String, ISEX: String
)
val dsEHAVE = spark.read
.option("database", "IEDB")
.option("collection", "EHAVE_TBL_2016")
.mongo[EHAVE_TBL]()MySQL/Maridb
Orc File
Parquet
CSV
Oracle
MS SQL Server
Use isin
Concatenate all columns
Hash an array of columns
UDF
converts a Scala list to its Java equivalent
Window Function
Rename Dataset Column Name
Regular Expression to extract values
Transpose rows to column
Last updated