Bootstrap IntroductionBootstrap vs. css, JSJust like ppt template relative to PPTPlainly, the predecessors have done a lot of templates (pre-build), you can directly take doctrineBootstarp 4Ways to use it1. DownloadYou'll see a bunch of filesBootstarp.min.css means they is minified2. CDN (Content Delivery Network)We can directly copy this code into index.htmlBecause JS would is read after CSS, it is put in the bottom of body part123456 789Ten $ - - theThen beautiful Navbar is shownUse the same s
Build Your First WebsiteInstall a subline textHTML Default RuleTags with opening and closing1 2 3 4 5 6 7 Hello WORLD!!!! 8 9 DEVELOPER FUNDAMENTALS:IIIDOCTYPE says, hey just a heads up a file was going to use HTML5The senior developers get good salaries and good jobs is the ones fully understand all the meaning behind the things they DoHTML tags10-15 tags is used in 99% timeIt acutally goes all the the-the-to 6Self Closing HTML tagsSrc:attribute which had special properties to the s
changes.At the top left, click on the logo with PoniterCan see the various modules of the Web page, you can change the moduleIt's like we changed the HTML, CSS, JS file that Google server had previously passed.Once we reques Google Server again,Then the interface will revert to the HTML, CSS, and JS files from Google Server as shown (e.g.)The Internet BackboneWe have Wifi from routerRouter is connected-to-modem (If you are an up-to-the-Internet, you are someone come to your house and install it
WWW vs InternetFor the begining, Internet is there. It is for the academics among universitiesThen they need a-share documents in the whole worldLee created WWW which is the common langauge that computers can speakYou can think Internet as your phoneWWW as an app the runs on your phoneLee created the first browser and the first server and the first Web siteThis is the first page in the world he wrote in 1991.What a cold wind!HTML, CSS, JavescriptHtml:text and links (for example, the whole web is
This course focuses onSpark, the hottest, most popular and promising technology in the big Data world today. In this course, from shallow to deep, based on a large number of case studies, in-depth analysis and explanation of Spark, and will contain completely from the enterprise real complex business needs to extract the actual case. The course will cover Scala programming, spark core programming,
"Note" This series of articles, as well as the use of the installation package/test data can be in the "big gift –spark Getting Started Combat series" get1 Spark Streaming Introduction1.1 OverviewSpark Streaming is an extension of the Spark core API that enables the processing of high-throughput, fault-tolerant real-time streaming data. Support for obtaining data
"Note" This series of articles and the use of the installation package/test data can be in the "big gift--spark Getting Started Combat series" Get 1, compile sparkSpark can be compiled in SBT and maven two ways, and then the deployment package is generated through the make-distribution.sh script. SBT compilation requires the installation of Git tools, and MAVEN installation requires MAVEN tools, both of which need to be carried out under the network,
"Note" This series of articles and the use of the installation package/test data can be in the "big gift--spark Getting Started Combat series" Get 1, compile sparkSpark can be compiled in SBT and maven two ways, and then the deployment package is generated through the make-distribution.sh script. SBT compilation requires the installation of Git tools, and MAVEN installation requires MAVEN tools, both of which need to be carried out under the network,
Three, in-depth rddThe Rdd itself is an abstract class with many specific implementations of subclasses:
The RDD will be calculated based on partition:
The default partitioner is as follows:
The documentation for Hashpartitioner is described below:
Another common type of partitioner is Rangepartitioner:
The RDD needs to consider the memory policy in the persistence:
Spark offers many storagelevel
1. Introduction
The Spark-submit script in the Spark Bin directory is used to start the application on the cluster. You can use the Spark for all supported cluster managers through a unified interface, so you do not have to specifically configure your application for each cluster Manager (It can using all Spark ' s su
The main contents of this section
Hadoop Eco-Circle
Spark Eco-Circle
1. Hadoop Eco-CircleOriginal address: http://os.51cto.com/art/201508/487936_all.htm#rd?sukey= a805c0b270074a064cd1c1c9a73c1dcc953928bfe4a56cc94d6f67793fa02b3b983df6df92dc418df5a1083411b53325The key products in the Hadoop ecosystem are given:Image source: http://www.36dsj.com/archives/26942The following is a brief introduction to the products1 HadoopApache's Hadoop p
1, first download the image to local. https://hub.docker.com/r/gettyimages/spark/~$ Docker Pull Gettyimages/spark2, download from https://github.com/gettyimages/docker-spark/blob/master/docker-compose.yml to support the spark cluster DOCKER-COMPOSE.YML fileStart it$ docker-compose Up$ docker-compose UpCreating spark_master_1Creating spark_worker_1Attaching to Sp
Step 1: Test spark through spark Shell
Step 1:Start the spark cluster. This is very detailed in the third part. After the spark cluster is started, webui is as follows:
Step 2: Start spark shell:
In this case, you can view the shell in the following Web console:
S
Install spark
Spark must be installed on the master, slave1, and slave2 machines.
First, install spark on the master. The specific steps are as follows:
Step 1: Decompress spark on the master:
Decompress the package directly to the current directory:
In this case, create the spa
Step 1: Test spark through spark Shell
Step 1:Start the spark cluster. This is very detailed in the third part. After the spark cluster is started, webui is as follows:
Step 2:Start spark shell:
In this case, you can view the shell in the following Web console:
Step 3:Co
Install spark
Spark must be installed on the master, slave1, and slave2 machines.
First, install spark on the master. The specific steps are as follows:
Step 1: Decompress spark on the master:
Decompress the package directly to the current directory:
In this case, create the
Spark Communication Module
1, Spark Cluster Manager can have local, standalone, mesos, yarn and other deployment methods, in order to
Centralized communication mode
1, RPC remote produce call
Spark Communication mechanism:
The advantages and characteristics of Akka are as follows:
1, parallel and distributed: Akka in design with asynchronous communication and dis
The content source of this page is from Internet, which doesn't represent Alibaba Cloud's opinion;
products and services mentioned on that page don't have any relationship with Alibaba Cloud. If the
content of the page makes you feel confusing, please write us an email, we will handle the problem
within 5 days after receiving your email.
If you find any instances of plagiarism from the community, please send an email to:
info-contact@alibabacloud.com
and provide relevant evidence. A staff member will contact you within 5 working days.