spark 2.2.0 accumulator使用方法 java版 python版__python

來源:互聯網
上載者:User

java版

package cn.spark.study.core;import org.apache.spark.Accumulator;import org.apache.spark.SparkConf;import org.apache.spark.api.java.JavaRDD;import org.apache.spark.api.java.JavaSparkContext;import org.apache.spark.api.java.function.VoidFunction;import java.util.*;public class AccumulatorVairable {public static void main(String[] args) {SparkConf conf =new SparkConf().setAppName("persist").setMaster("local");JavaSparkContext sc =new JavaSparkContext(conf);final Accumulator<Integer> accumulator = sc.accumulator(0);List<Integer> numberlist =Arrays.asList(1,2,3,4,5);JavaRDD<Integer> num =sc.parallelize(numberlist);num.foreach(new VoidFunction<Integer>() {@Overridepublic void call(Integer arg0) throws Exception {accumulator.add(arg0);}});System.out.println(accumulator.value());sc.close();}}
python版

import osimport sysos.environ['SPARK_HOME'] = '/opt/spark'sys.path.append("/opt/spark/python")from pyspark import SparkContextfrom pyspark import SparkConfif __name__ == '__main__':    sc = SparkContext("local", "reduce")    sumact = sc.accumulator(0)    list = [1, 2, 3, 4, 5]    listRdd = sc.parallelize(list)    def f(num):        global sumact        sumact +=num    listRdd.foreach(f)    print sumact


相關文章

聯繫我們

該頁面正文內容均來源於網絡整理,並不代表阿里雲官方的觀點,該頁面所提到的產品和服務也與阿里云無關,如果該頁面內容對您造成了困擾,歡迎寫郵件給我們,收到郵件我們將在5個工作日內處理。

如果您發現本社區中有涉嫌抄襲的內容,歡迎發送郵件至: info-contact@alibabacloud.com 進行舉報並提供相關證據,工作人員會在 5 個工作天內聯絡您,一經查實,本站將立刻刪除涉嫌侵權內容。

A Free Trial That Lets You Build Big!

Start building with 50+ products and up to 12 months usage for Elastic Compute Service

  • Sales Support

    1 on 1 presale consultation

  • After-Sales Support

    24/7 Technical Support 6 Free Tickets per Quarter Faster Response

  • Alibaba Cloud offers highly flexible support services tailored to meet your exact needs.