hadoop shell命令遠程提交

來源:互聯網
上載者:User
hadoop shell命令遠程提交 一,hadoop shell命令遠程提交原理    hadoop shell命令執行目前很多情境下面主要通過 Linux shell來互動操作,無論對於遠程操作還是習慣於windows/web操作的開發人員而言,也是非常痛苦的事情。    在hadoop安裝包中的src\test\org\apache\hadoop\cli\util 目錄中,CommandExecutor.java地實現方式或許對大家有一定啟發。    如下是一段hadoop dfsadmin 命令執行的過程。 public
static int executeDFSAdminCommand( final String cmd,
final String namenode) {
      exitCode = 0;

      ByteArrayOutputStream bao =
new ByteArrayOutputStream();
      PrintStream origOut = System.out;
      PrintStream origErr = System.err;

      System.setOut(new PrintStream(bao));
      System.setErr(new PrintStream(bao));

      DFSAdmin shell =
new DFSAdmin();
      String[] args = getCommandAsArgs(cmd,
"NAMENODE", namenode);
      cmdExecuted = cmd;

      try {
        ToolRunner.run(shell, args);
      } catch (Exception e) {
        e.printStackTrace();
        lastException = e;
        exitCode =
-1;
      } finally {
        System.setOut(origOut);
        System.setErr(origErr);
      }

      commandOutput = bao.toString();

      return exitCode;
  }        在開始階段,通過System.setOut和System.setErr來設定當前應用程式的標準輸出和錯誤輸出資料流方式是ByteArrayOutputStream。    初始化DFSAdmin shell之後,調用ToolRunner.run方法運行 args的命令參數。    當調用完成之後,重新設定標準輸出和標準錯誤輸出的方式為預設。    DFSAdmin對象類似於hadoop shell命令的 "hadoop dfsadmin "    MRAdmin對象類似於hadoop shell命令的 "hadoop mradmin "    FsShell對象類似於hadoop shell命令的 "hadoop fs "二,利用內建jetty方式,開發jetty servlet來實現一個基於web遠程方式提交hadoop shell命令的基本操作。    1, 設計一個html頁面向servlet提交命令參數,如:    2,servlet程式編寫,如下:           PrintWriter writer=response.getWriter();

        response.setContentType("text/html");
        if(request.getParameter("select_type")==null){
            writer.write("select is null");
            return;
        }
        if(request.getParameter("txt_command")==null){
            writer.write("command is null");
            return;
        }
        String type=request.getParameter("select_type");
        String command=request.getParameter("txt_command");
        ByteArrayOutputStream bao =
new ByteArrayOutputStream();
        PrintStream origOut = System.out;
        PrintStream origErr = System.err;

        System.setOut(new PrintStream(bao));
        System.setErr(new PrintStream(bao));
        if(type.equals("1")){
            DFSAdmin shell =
new DFSAdmin();
            String[] items=command.trim().split(" ");
            try{
                ToolRunner.run(shell,items);
            }
            catch (Exception e) {
                e.printStackTrace();
            }
            finally{
                System.setOut(origOut);
                System.setErr(origErr);
            }
            writer.write(bao.toString().replaceAll("\n",
"<br>"));
        }
        else
if(type.equals("2")){
            MRAdmin shell =
new MRAdmin();
            String[] items=command.trim().split(" ");
            try{
                ToolRunner.run(shell,items);
            }
            catch (Exception e) {
                e.printStackTrace();
            }
            finally{
                System.setOut(origOut);
                System.setErr(origErr);
            }
            writer.write(bao.toString().replaceAll("\n",
"<br>"));
        }
        else
if(type.equals("3")){
            FsShell shell =
new FsShell();
            String[] items=command.trim().split(" ");
            try{
                ToolRunner.run(shell,items);
            }
            catch (Exception e) {
                e.printStackTrace();
            }
            finally{
                System.setOut(origOut);
                System.setErr(origErr);
            }
            writer.write(bao.toString().replaceAll("\n",
"<br>"));
        }    上述程式主要用於簡單處理dfsadmin,mradmin,fs等hadoop shell,並且最終以字串列印輸出到用戶端        簡單測試 -report的結果,截取部分圖片如下:Configured Capacity:
7633977958400 (6.94 TB)
Present Capacity: 7216439562240 (6.56 TB)
DFS Remaining: 6889407496192 (6.27 TB)
DFS Used: 327032066048 (304.57 GB)
DFS Used%:
4.53%
Under replicated blocks:
42
Blocks with corrupt replicas:
0
Missing blocks: 0

-------------------------------------------------
Datanodes available:
4 (4 total, 0 dead)

Name: 10.16.45.226:50010
Decommission Status : Normal
Configured Capacity:
1909535137792 (1.74 TB)
DFS Used: 103113867264 (96.03 GB)
Non DFS Used: 97985679360 (91.26 GB)
DFS Remaining: 1708435591168(1.55 TB)
DFS Used%:
5.4%
DFS Remaining%:
89.47%
Last contact: Wed Mar
21 14:37:24 CST
2012       

上述代碼利用jetty內嵌方式開發,運行時候還需要載入hadoop 相關依賴jar以及hadoop config檔案,如所示意:

#!/bin/sh
CLASSPATH="/usr/local/hadoop/conf"
for f
in $HADOOP_HOME/hadoop-core-*.jar;
do
  CLASSPATH=${CLASSPATH}:$f;
done
# add libs to CLASSPATH
for f
in $HADOOP_HOME/lib/*.jar;
do
  CLASSPATH=${CLASSPATH}:$f;
done
for f
in $HADOOP_HOME/lib/jsp-2.1/*.jar;
do
  CLASSPATH=${CLASSPATH}:$f;
done

echo $CLASSPATH
java -cp
"$CLASSPATH:executor.jar" RunServer

相關文章

聯繫我們

該頁面正文內容均來源於網絡整理,並不代表阿里雲官方的觀點,該頁面所提到的產品和服務也與阿里云無關,如果該頁面內容對您造成了困擾,歡迎寫郵件給我們,收到郵件我們將在5個工作日內處理。

如果您發現本社區中有涉嫌抄襲的內容,歡迎發送郵件至: info-contact@alibabacloud.com 進行舉報並提供相關證據,工作人員會在 5 個工作天內聯絡您,一經查實,本站將立刻刪除涉嫌侵權內容。

A Free Trial That Lets You Build Big!

Start building with 50+ products and up to 12 months usage for Elastic Compute Service

  • Sales Support

    1 on 1 presale consultation

  • After-Sales Support

    24/7 Technical Support 6 Free Tickets per Quarter Faster Response

  • Alibaba Cloud offers highly flexible support services tailored to meet your exact needs.