1, download the source code, according to their own environment compiled, I download here is the spark1.3 versionI use SBT to compile,spark_hadoop_version=2.5.2 spark_yarn=ture SBT/SBT AssemblyThis code has two parameters, the first refers to the native version of the Hadoop environment, the second parameter refers to whether to run on yarn,2. make-distribution.s
Since I have been using idea to develop a Java project, Scala can now also use idea development: http://www.jetbrains.com/idea/Community Edition free and Ultimate Edition free 30-day trial all support Scala development, I use the Ultimate EditionWhen we download idea after installing and booting, we need to install a Scala Plugin, the following steps:The following interface appears when you enter plugins:Click Install jetbrains Plugin. ButtonEnter Scala back to the following interface:Clicking t
representation as a Kafka Cluster, and the above architecture diagram is relatively detailed;Kafka version: 0.8.0Kafka download and Documentation: HTTP://KAFKA.APACHE.ORG/KAFKA installation:
> Tar xzf kafka-
> CD kafka-
>./SBT Update
>./SBT Package
>./SBT assembly-package-dependency
Copy CodeStart and Test commands: (1) Start server
the architecture is just the Kafka concise representation as a Kafka Cluster, and the above architecture diagram is relatively detailed;Kafka version: 0.8.0Kafka download and Documentation: HTTP://KAFKA.APACHE.ORG/KAFKA installation:
> Tar xzf kafka-
> CD kafka-
>./SBT Update
>./SBT Package
>./SBT assembly-package-dependency
Copy Co
Scalajs is a compiler that compiles Scala into JS to build a stable and extensible JS application using many of Scala's class libraries and strongly typed features.build.sbtThe build file is as follows:enablePlugins(ScalaJSPlugin)name := """scalajs"""version := "1.0"scalaVersion := "2.12.1"libraryDependencies += "org.scala-js" %%% "scalajs-dom" % "0.9.1"libraryDependencies += "be.doeraene" %%% "scalajs-jquery" % "0.9.1"libraryDependencies += "com.lihaoyi" %%% "scalatags" % "0.6.2"libraryDependen
Environment: CentOS 6.3, Hadoop 1.1.2, JDK 1.6, Spark 1.0.0, Scala 2.10.3, Eclipse (Hadoop can not be installed)
First of all understand that Spark is the Scala project. In other words, Spark was developed in Scala.
1. Please confirm that you have properly installed spark and run, if not, please install the following link to install.
http://blog.csdn.net/zlcd1988/article/details/21177187
2. New Scala Project
Create a Scala project based on your blog below
http://blog.csdn.net/zlcd1988/article/de
Intelij idea compiling the spark source processUnder the spark source package, unzip IntelliJ to install Scala plugin using IntelliJ Ieda Open Project feature opens the source folder
After that, idea automatically downloads all of the dependencies, and after the download is finished, make Project compiles.
It is recommended to use the new version of Ieda, it is recommended to pre-install the SBT environment, to download the
http://www.lydsy.com/JudgeOnline/problem.php?id=3236
Bit set SBT
Treap hopeless, no matter how to optimize the constant is not over, so casually changed to SBT ...
Start SBT sz I use the S maintenance results mad re ...
The last 90s card over ..... Slower than the MO team ....
Abused 2h ...
I knew I'd write a mo team ...
Mo team is clearly O (nsqrt (n) logn) My
Project Address: Https://github.com/quantifind/wisp
Wisp is a real-time, interactive charting tool.
Install SBT, then download the WISP project locally, then CD to the WISP project root directory, such as mine (CD D:\spark\wisp), then execute the SBT "project Wisp" console, which will enter the SBT console after the compilation is successful.
Then you can start w
, because now IntelliJ idea and Eclipse do the Scala plugin is not good enough, there will be a lot of problems in development, we recommend the creation of Typesafe Magic Board.2. Open the BULID.SBT document of the project; add this sentence:3. Then add the MySQL driverAfter adding a new dependency in SBT, be sure to refresh it so that SBT can add something new to it.4. Next, we open the conf/application.c
set policies that do not retain any dataRman> Configure retention policy to none;Backup optimizationBackup optimization in Rman means that, in the backup process, if certain conditions are met, RmanSome files are automatically skipped and not included in the backup set to save time and space. Usually meet the followingconditions, the ability to enable backup optimizations is enabled:Configure Backup optimization parameter is set to on;The backup database or backup Archivelog command is executed
The first attempt to back up an Oracle database to the AWS Cloud was typically backed up to a tape library or disk, and now it's easier to have a cloud. This is primarily backed up using AWS's S3 storage as an SBT device. As for what is AWS, what is S3, please refer to: http://www.amazonaws.cn/products/specific steps are as follows:Prerequisites: You will have an AWS Accunt (AWS access key ID and AWS secret key) and you will need to have Oracle's OTN
Due to the licene restrictions, not put into the default build, so on the official website downloaded binary files do not contain the Gangla module, if needed, need to compile themselves. When using MAVEN to compile spark, we can add the -Pspark-ganglia-lgpl option to package the ganglia related classes into the spark-assembly-x.x.x-hadoopx.x.x.jar command as follows:./make-distribution.sh--tgz-phadoop-2.4 -pyarn-dskiptests dhadoop.version=2.4. 0 -PSPARK-GANGLIA-LGPLYou can also compile with SBT
create the Java program.The above command creates a new directory getbookmarks, creating files and directories.
The app directory contains program-specific code such as controllers, views, and modules. The controller package contains Java code to respond to URL routing. The views directory contains the server-side template, models directory contains the program domain module, in this program, the domain is a story class.
The Conf directory contains the program configurat
First, environmental windows_x64 system Java1.8
Scala2.10.6 spark1.6.0 hadoop2.7.5
Idea IntelliJ 2017.2 nmap tool (NCAT command in which corresponds to NC commands in Linux)
Second, local application set up 2.1 environment variable setting method: System Parameter--"Add variable-" form: xxx_home, then copy the root directory of the corresponding installation package as the variable value; add:%xxx_home%\bin in Path variable;
1,hadoop need to set environment variables; 2,scala best to download an
-top actor, the child actor, in the existing actor by calling Context.actorof (). The method signature of Context.actorof () is the same as system.actorof ().
The simplest way to view the actor hierarchy is to print an instance of Actorref. In this little experiment, we created an actor, printed its reference, created a child actor for him, and printed a reference to its child actor. We start with the Hello World project and if you haven't downloaded it, please download the QuickStart project fr
Reference:
http://spark.incubator.apache.org/docs/latest/
Http://spark.incubator.apache.org/docs/latest/spark-standalone.html
http://www.yanjiuyanjiu.com/blog/20130617/
1. Installing the JDK
2. Install Scala 2.9.3
Spark 0.7.2 relies on Scala 2.9.3 and we have to install Scala 2.9.3.
Download the scala-2.9.3.tgz and save it to your home directory (already on sg206).$ TAR-ZXF scala-2.9.3.tgz$ sudo mv Scala-2.9.3/usr/lib$ sudo vim/etc/profile# Add the following lines at the endExport scala_home=/us
= Sqlcontext.jsonfile (path)//inferred pattern can be explicitly people.printschema ()//root//|--by using the Printschema () method : integertype// |--name:stringtype//to register Schemardd as a table people.registerastable ("people")// The SQL state can be run by using the SQL method provided by the SqlContext val teenagers = sqlcontext.sql ("Select name from people WHERE age >= 19 In addition, a schemardd can also generate Val Anotherpeoplerdd = Sc.parallelize ("" "{" name ") by storing a s
The content source of this page is from Internet, which doesn't represent Alibaba Cloud's opinion;
products and services mentioned on that page don't have any relationship with Alibaba Cloud. If the
content of the page makes you feel confusing, please write us an email, we will handle the problem
within 5 days after receiving your email.
If you find any instances of plagiarism from the community, please send an email to:
info-contact@alibabacloud.com
and provide relevant evidence. A staff member will contact you within 5 working days.