學習隨筆--JavaSparkJDBC操作Oracle

來源:互聯網
上載者:User

標籤:++   edr   for   contex   acl   context   rdd   apach   call   

package stuSpark.com;import java.io.IOException;import java.util.ArrayList;import java.util.Arrays;import java.util.HashMap;import java.util.List;import java.util.Map;import org.apache.spark.SparkConf;import org.apache.spark.api.java.JavaPairRDD;import org.apache.spark.api.java.JavaRDD;import org.apache.spark.api.java.JavaSparkContext;import org.apache.spark.api.java.function.FlatMapFunction;import org.apache.spark.api.java.function.Function;import org.apache.spark.api.java.function.Function2;import org.apache.spark.api.java.function.PairFunction;import org.apache.spark.sql.DataFrame;import org.apache.spark.sql.DataFrameReader;import org.apache.spark.sql.Row;import org.apache.spark.sql.SQLContext;import scala.Tuple2;public class JavaSparkJDBCSQL {public static void main(String[] args)throws IOException{System.out.println("begin");SparkConf sparkConf = new SparkConf().setAppName("JavaSparkJDBCSQL").setMaster("local[2]");JavaSparkContext sc = new JavaSparkContext(sparkConf);SQLContext sqlContext = new SQLContext(sc);//設定資料庫連接參數Map<String,String> dBConOption = new HashMap<String,String>();dBConOption.put("url", "jdbc:oracle:thin:@127.0.0.1:1521:ORCL");dBConOption.put("user", "Xho");dBConOption.put("password", "sys");dBConOption.put("driver", "oracle.jdbc.driver.OracleDriver");dBConOption.put("dbtable", "NUMB");DataFrameReader dfRead = sqlContext.read().format("jdbc").options(dBConOption);DataFrame df=dfRead.load();//註冊為表,然後在SQL語句中使用df.registerTempTable("lk");// SQL可以在登入為表的RDDS上運行DataFrame df2 = sqlContext.sql("select * from lk");df2.show();/*+---+---+-----+----+|ONE|TWO|THREE|FOUR|+---+---+-----+----+|  a|  b|    c|   d||  a|  a|    b|   b||  c|  c|    a|   d||  a|  a|    c|   s||  m|  s|    b|   j||  a|  l|    o|   k|+---+---+-----+----+*/List<String> list = df2.toJavaRDD().map(new Function<Row, String>(){public String call(Row row){return row.getString(0);}}).collect();JavaRDD<String> words = df2.toJavaRDD().flatMap(new FlatMapFunction<Row,String>(){public Iterable<String> call(Row row){List<String> ll = new ArrayList<String>();for (int i = 0; i < row.length(); i++) {ll.add(row.getString(i));}return ll;}});//maptopair 將集合資料存為key valueJavaPairRDD<String, Integer> ones = words.mapToPair(new PairFunction<String, String, Integer>() {public Tuple2<String, Integer> call(String s) {return new Tuple2<String, Integer>(s, 1);}});//reduceBykey 根據key聚集,對value進行操作JavaPairRDD<String, Integer> counts = ones.reduceByKey(new Function2<Integer, Integer, Integer>() {public Integer call(Integer i1, Integer i2) {return i1 + i2;}});//collect封裝返回一個數組List<Tuple2<String, Integer>> output = counts.collect();for (Tuple2<?, ?> tuple : output) {System.out.println(tuple._1() + ": " + tuple._2());}/*d: 2s: 2a: 7k: 1b: 4o: 1j: 1l: 1m: 1c: 4*/sc.stop();System.out.println("end");}}

  

學習隨筆--JavaSparkJDBC操作Oracle

聯繫我們

該頁面正文內容均來源於網絡整理,並不代表阿里雲官方的觀點,該頁面所提到的產品和服務也與阿里云無關,如果該頁面內容對您造成了困擾,歡迎寫郵件給我們,收到郵件我們將在5個工作日內處理。

如果您發現本社區中有涉嫌抄襲的內容,歡迎發送郵件至: info-contact@alibabacloud.com 進行舉報並提供相關證據,工作人員會在 5 個工作天內聯絡您,一經查實,本站將立刻刪除涉嫌侵權內容。

A Free Trial That Lets You Build Big!

Start building with 50+ products and up to 12 months usage for Elastic Compute Service

  • Sales Support

    1 on 1 presale consultation

  • After-Sales Support

    24/7 Technical Support 6 Free Tickets per Quarter Faster Response

  • Alibaba Cloud offers highly flexible support services tailored to meet your exact needs.