This shows you the differences between two versions of the page.
Both sides previous revision Previous revision Next revision | Previous revision | ||
en:site:recherche:logiciels:sparqlwithspark:datasetwatdiv [16/09/2016 23:00] hubert |
en:site:recherche:logiciels:sparqlwithspark:datasetwatdiv [16/09/2016 23:01] (current) hubert [Load VP's] |
||
---|---|---|---|
Line 88: | Line 88: | ||
Create one dataset per property. | Create one dataset per property. | ||
- | <code> | + | <code scala> |
/* | /* | ||
val df = num. | val df = num. | ||
Line 119: | Line 119: | ||
===== Load VP's ===== | ===== Load VP's ===== | ||
- | <code> | + | <code scala> |
// S2RDF VP | // S2RDF VP | ||
Line 147: | Line 147: | ||
//val dictSO = sqlContext.read.parquet(dictSOFile).repartition(NB_FRAGMENTS, col("so")) | //val dictSO = sqlContext.read.parquet(dictSOFile).repartition(NB_FRAGMENTS, col("so")) | ||
dictSO.persist().count | dictSO.persist().count | ||
- | //dictSO.unpersist() | ||
// VP Dataset | // VP Dataset | ||
// ------- | // ------- | ||
- | //val encodedFile = dir + "/frame" + scale | ||
val vpDir = dir + "/vp" + scale | val vpDir = dir + "/vp" + scale | ||
- | // CHRONO | + | // TIMER |
def queryTimeDFIter(q: DataFrame, nbIter: Int): Unit = { | def queryTimeDFIter(q: DataFrame, nbIter: Int): Unit = { | ||
var l = new scala.collection.mutable.ArrayBuffer[Double](nbIter) | var l = new scala.collection.mutable.ArrayBuffer[Double](nbIter) | ||
Line 172: | Line 170: | ||
- | // define VPs to be loaded | + | // Define the VPs to be loaded |
//------------------------- | //------------------------- | ||
val nbP = dictP.count.toInt | val nbP = dictP.count.toInt |