Original article: http://blog.csdn.net/michzel/article/details/6219903
Run SOLR example on Windows
1. Download the compressed package apache-solr-1.2.0.zip and decompress it.
2. Rename/apache-solr-1.2.0/Dist/apache-solr-1.2.0.war to SOLR. War and copy it to Tomcat's web
SOLR multicore is a new feature of SOLR 1.3. It is a SOLR instance and can have multiple search applications.Here's a start to run out of a example given by SOLR. This article is based on the "Use SOLR to build enterprise search platform-run
1. Obtain solrserver 1.1 commonshttpsolrserver
Commonshttpsolrserver uses httpclient to communicate with the SOLR server.Java code
String url = "http: // localhost: 8983/SOLR ";
Solrserver Server = new commonshttpsolrserver (URL );
Setting xmlresponseparser
Sorlr J currently uses the binary format as the default format. Solr1.2 users can use the XML format only through the display settings.Java code
Se
I don't know why the search war file that comes with nutch is gone in the Nutch-1.3, and in the Nutch-1.3, After grabbing the file, the generated directory is only crawldb, linkdb, segments, and no indexs and INDEX DIRECTORY
Check the official site wiki, the above is the nutch index ing to SOLR, with SOLR to provide the search function, see the official site wiki Description: http://wiki.apache.org/nutch/Ru
In this paper, we simply tell SOLR to build a full-text index, import data from a database, generate an index file, this article is built in the case of the SOLR application has been set up, if you want to know how to deploy SOLR services can first look at the article: SOLR installationEnvironment: JDK 7,solr4.10.1, To
Today, I want to use DIH to import CSV files, so the data source is roughly implemented using filedatasource + custom converter.
Package COM. besttone. transformer; import Java. util. map; public class csvtransformer {// reference http://wiki.apache.org/solr/DIHCustomTransformerpublic object transformrow (Map
Many problems have been found, such as the comma in the field, and so on. This rough converter cannot be implemented, so I continued to find th
Faceted search has become a critical feature for enhancing findability and the user search experience for all types of search applications. In this article, SOLR creator yonik Seeley gives an introduction to Faceted search with SOLR.
By yonik SeeleyWhat is Faceted search?Faceted search is the dynamic clustering of items or search results into categories that let users drill into search results (or
We talked about the basic usage and configuration file of SOLR several times before, and then we started our real code journey.
1) start with a simple program:
Public static void main (string [] ARGs) throws solrserverexception, ioexception, parserconfigurationexception, saxexception {// Set SOLR. Home. Note that the environment variable SOLR.
1 Overview
SOLR is an independent enterprise-class search application server that provides an API interface similar to Web-service. The user can submit a certain format XML file to the Search engine server through HTTP request, generate the index, or can make the lookup request through the HTTP GET operation, and obtain the return result of the XML format. This is mainly explained in this way through HTTP GET requests.
First, we have to go through H
Mongo-connector integrates MongoDB to Solr for incremental Indexing
Configure the MongoDB replica set
Deploy a replica set for testing and development
Deployment replica set
1. create necessary data folders for each node:
1 mkdir-p/srv/mongodb/rs0-0/srv/mongodb/rs0-1/srv/mongodb/rs0-2
2. Run the following command to start the mongod instance:
First node:
Mongod -- port 27017 -- dbpath/srv/mongodb/rs0-0 -- replSet rs0 -- smallfiles -- oplogSize 128 --
1Create a collection
SSH connects remotely to the CDH node that has SOLR installed.
Running the solrctl instancedir--generate/solr/test/gx_sh_tl_tgryxx_2015 command generates the default configuration for the Gx_sh_tl_tgryxx_2015 collection.
Enter the/solr/test/gx_sh_tl_tgryxx_2015/conf directory, first edit the Schema.xml configuration field information,
Prerequisites:Installing SOLR Version: 4.8.0 deployment SOLR Path:/data/solr-4.8.01. Upload some configuration information via zookeeper:Upload configuration information to the ZK environment via the ZK command/data/solr-4.8.0/node/scripts/cloud-scripts/zkcli.sh-zkhost solr1-cmd upconfig-confdir/data/
The previous article built the zookeeper clusterWell, today we can build a cluster of SOLR search service, which is different from Redis cluster, it needs ZK management, as an agent layerInstall four Tomcat and modify its port number to not conflict. 8080~8083If it is a formal environment, use 4 Linux nodes respectivelyModify the Server.xml file to modify the port number by a total of 3The above steps are repeated on the tomcat03,tomcat04, but 3 ports
Next year work mainly and search related, in their own learning process to share to let more people benefit is my usual practice, so I will take the study of SOLR here to share with you, if you are also interested in search, but also novice, then you and I start from scratch.Today we are going to talk about the Quick start SOLR, let's play in the window environment, ready, we're about to start.about SOLRSOL
With the wide application of SOLR, SOLR version updates have also accelerated scripts, easynet. SOLR (http://easynet.codeplex.com/) has been in the maintenance of intermittent, today also released the latest version, fully support SOLR 3.5.1. Compared with SOLR client solrsh
Zookeeper can manage the configuration files of SOLR and other software. The configuration file is stored on the disk of the Linux server, but it does not change the location where SOLR reads the SOLR/home configuration file.
Currently, the SOLR/home configuration file is located in two places:
E: \ apache-Tomcat-7.0.2
Reprinted from: http://www.cnblogs.com/ezhangliang/archive/2012/04/11/2441945.html
Scheduler mainly solves two problems:
1. Update indexes on a regular basis.
2. Redo the index regularly.
After testing, scheduler has been able to implement completely configuration-based, without the need for development features, without manual intervention to implement the above two features (combined with SOLR data import request handler ).
To facilitate later use,
I just deployed SOLR under tomcat, so I ran out of it. I'm so excited. Call ~
In fact, as long as the following conditions are met, the deployment will not fail:
1. Realize that SOLR is a webapp role;
2. Download the war package under the SOLR directory DIST and put it under the Tomcat \ webapps directory.
3. start Tomcat;
4. after Tomcat is started,
want to crawl the data of a watercress movie, which can be set:#注释掉这一行# skip URLs containing certain characters as probable queries, etc.#-[?*[emailprotected]=]# accept anything else#注释掉这行#+.+^http:\/\/movie\.douban\.com\/subject\/[0-9]+\/(\?.+)?$5 Setting the agent nameConf/nutch-site.xml:property> name>http.agent.namename> value>My Nutch Spidervalue>property>This step is seen from this book, Web crawling and Data Mining with Apache Nutch, page 14th.6 Installing SOLRBecause
I tried the following three open-source Chinese Word divider in SOLR, two of which were not available because the SOLR version was too high. I finally decompiled the jar package and found the reason, the following briefly describes three open source Chinese Word splitters.
Ding jieniu: The last code submission on Google Code was 2008.6 months. It was not very active, but many people were using it.
Mmseg4j:
The content source of this page is from Internet, which doesn't represent Alibaba Cloud's opinion;
products and services mentioned on that page don't have any relationship with Alibaba Cloud. If the
content of the page makes you feel confusing, please write us an email, we will handle the problem
within 5 days after receiving your email.
If you find any instances of plagiarism from the community, please send an email to:
info-contact@alibabacloud.com
and provide relevant evidence. A staff member will contact you within 5 working days.