代码如下:
package day01
import java.io.PrintWriter
import java.io.File
import org.apache.spark.{SparkConf, SparkContext}
object countWord_2 {
def main(args: Array[String]): Unit = {
val conf = new SparkConf().setAppName("countWord_2").setMaster("local")
val sc = new SparkContext(conf)
System.setProperty("hadoop.home.dir", "E:\school_Study\spark\hadoop")
val input = "F:\score.txt"
val count = sc.textFile(input).flatMap(x => x.split(" ")).map(x => (x, 1)).reduceByKey((x, y) => x + y)
//文件写入
val writer = new PrintWriter(new File("res.txt"))
count.foreach(x => println(String.valueOf(x._1 + " " + x._2)))
val tuples = count.toArray()
for (elem <- tuples) {
writer.write(elem._1 + " " + elem._2 + "n")
}
writer.close()
}
}
使用的Scala版本为:scala-2.10.5



