import org.apache.spark.rdd.RDD
import org.apache.spark.{SparkConf, SparkContext}
object SparkWordCount {
def main(args: Array[String]): Unit = {
val sc = new SparkContext(new SparkConf().setAppName("SparkWordCount").setMaster("local"))
//args(0):文件路径
val value: RDD[String] = sc.textFile(args(0))
value.flatMap(_.split(" ")).map((_, 1)).reduceByKey(_ + _).foreach(println)
//回收资源停止sc,结束任务
sc.stop()
}
}
运行结果



