Spark3000门徒第10课Java开发Spark实战总结

时间:2022-04-08 03:25:55

今晚听了王家林老师的第10课Java开发Spark实战,课后作业是:用Java方式采用Maven开发Spark的WordCount并运行在集群中

先配置pom.xml

<groupId>com.dt.spark</groupId>
<artifactId>SparkApps</artifactId>
<version>0.0.1-SNAPSHOT</version>
<packaging>jar</packaging>

<dependency>
<groupId>org.apache.spark</groupId>
<artifactId>spark-core_2.10</artifactId>
<version>1.6.0</version>
</dependency>
<dependency>
<groupId>org.apache.spark</groupId>
<artifactId>spark-sql_2.10</artifactId>
<version>1.6.0</version>
</dependency>
<dependency>
<groupId>org.apache.hadoop</groupId>
<artifactId>hadoop-client</artifactId>
<version>2.6.0</version>
</dependency>

然后写程序:

public class WordCount {

	public static void main(String[] args) {
		
		SparkConf conf = new SparkConf().setAppName("Spark WordCount by java");   
		JavaSparkContext sc = new JavaSparkContext(conf);
		JavaRDD<String> lines = sc.textFile(args(0));

		JavaRDD<String> words = lines.flatMap(new FlatMapFunction<String,String>(){   
			public Iterable<String> call(String line) throws Exception{
				return Arrays.asList(line.split(" "));
			}
		});  

		JavaPairRDD<String,Integer> pairs = words.mapToPair(new PairFunction<String,String,Integer>(){
			public Tuple2<String,Integer> call(String word) throws Exception {
				return new Tuple2<String,Integer>(word,1);
			}    
		});

		JavaPairRDD<String,Integer> wordsCount = pairs.reduceByKey(new Function2<Integer,Integer,Integer>(){			
			public Integer call(Integer v1,Integer v2){
				return v1+v2;
			}
		});

		wordsCount.foreach(new VoidFunction<Tuple2<String,Integer>>(){
			public void call(Tuple2<String,Integer> pairs) throws Exception{
				System.out.println(pairs._1+":"+pairs._2);
			}
		});
		
		sc.close();
	}

}

  

打包成jar文件放服务器上执行:

/usr/lib/spark/bin/spark-submit --master  yarn-client  --class com.dt.spark.WordCount --executor-memory 2G --executor-cores  4 ~/spark/wc.jar  ./mydir/tmp.txt

可以看到结果跟用scala写的一致。

后续课程可以参照新浪微博 王家林_DT大数据梦工厂:http://weibo.com/ilovepains

王家林  中国Spark第一人,微信公共号DT_Spark

博客:http://bolg.sina.com.cn/ilovepains

 

转载请说明出处!