python from pyspark import SparkContext sc = SparkContext("local", "WordCountExample") data = sc.parallelize(["Hello", "World", "Hello", "Spark"]) word_counts = data.flatMap(lambda line: line.split(" ")).map(lambda x: (x, 1)).reduceByKey(lambda a, b: a + b) word_counts.collect() python from dpark import DparkContext dc = DparkContext() data = dc.parallelize(["Hello", "World", "Hello", "Dpark"]) word_counts = data.flatMap(lambda line: line.split(" ")).map(lambda x: (x, 1)).reduceByKey(lambda a, b: a + b) word_counts.collect()


上一篇:
下一篇:
切换中文