ИК
https://stackoverflow.com/questions/48677883/using-predicates-in-spark-jdbc-read-method
Size: a a a
ИК
ИК
А
АА
A
FT
RI
K
FT
N
S
GP
ME
FT
GP
V
S
FT
val inputPath="alice.txt"
//val outputPath="alice-wordpunctcount"
val textFile = sc.textFile(inputPath)
val sorted = textFile.flatMap(line => line.split(" ")).
filter(x=> x.contains(",")||x.contains(".")||x.contains("?")||x.contains("!")||x.contains("-")||x.contains("`")||x.contains("\'")||x.contains(";")||x.contains(":")||x.contains("\"")).
map(word => (word, 1)).
reduceByKey(_ + _).
map(x => x.swap).
sortByKey(false).
map(x => x.swap)
val key=sorted.keys
key.foreach(s=>s.filterNot(c=>c=='.'))
key.foreach(x=>println(x))
FT
val inputPath="alice.txt"
//val outputPath="alice-wordpunctcount"
val textFile = sc.textFile(inputPath)
val sorted = textFile.flatMap(line => line.split(" ")).
filter(x=> x.contains(",")||x.contains(".")||x.contains("?")||x.contains("!")||x.contains("-")||x.contains("`")||x.contains("\'")||x.contains(";")||x.contains(":")||x.contains("\"")).
map(word => (word, 1)).
reduceByKey(_ + _).
map(x => x.swap).
sortByKey(false).
map(x => x.swap)
val key=sorted.keys
key.foreach(s=>s.filterNot(c=>c=='.'))
key.foreach(x=>println(x))
key.foreach(s=>s.filterNot(c=>c=='.'))
АР
key.foreach(s=>s.filterNot(c=>c=='.'))