I have to read a whole directory of xlsx files, and I need to load all the directory with Apache Spark using Scala.
Actually I'm using this dependency : "com.crealytics" %% "spark-excel" % "0.12.3"
, and I don't know how to load all.
I have to read a whole directory of xlsx files, and I need to load all the directory with Apache Spark using Scala.
Actually I'm using this dependency : "com.crealytics" %% "spark-excel" % "0.12.3"
, and I don't know how to load all.
Copyright © 2021 Jogjafile Inc.
There doesnt seem a shortcut option to be put into the path through option method. So I have created a workaround as below(assuming each excel file has same number of columns). Created a method to get all the paths of every file in the source directory and ran a loop over those file paths creating new dataframe and appending to the previous one.