python
from pyspark import SparkContext
sc = SparkContext("local", "WordCountExample")
data = sc.parallelize(["Hello", "World", "Hello", "Spark"])
word_counts = data.flatMap(lambda line: line.split(" ")).map(lambda x: (x, 1)).reduceByKey(lambda a, b: a + b)
word_counts.collect()
python
from dpark import DparkContext
dc = DparkContext()
data = dc.parallelize(["Hello", "World", "Hello", "Dpark"])
word_counts = data.flatMap(lambda line: line.split(" ")).map(lambda x: (x, 1)).reduceByKey(lambda a, b: a + b)
word_counts.collect()