See information:Official websitehttp://www.pentaho.com/Chinese enthusiasts Communityhttp://www.pentahochina.com/portal.phpSourceForge Community Edition:https://sourceforge.net/projects/pentaho/files/Business Intelligence Server:https://sourceforge.net/projects/pentaho/files/Business%20Intelligence%20Server/Report Designer:https://sourceforge.net/projects/pentaho/
I. Extracting data from HDFS to an RDBMS1. Download the sample file from the address below.Http://wiki.pentaho.com/download/attachments/23530622/weblogs_aggregate.txt.zip?version=1modificationDate =13270678580002. Use the following command to place the extracted Weblogs_aggregate.txt file in the/user/grid/aggregate_mr/directory of HDFs.Hadoop fs-put weblogs_aggregate.txt/user/grid/aggregate_mr/3. Open PDI, create a new transformation, 1.Figure 14. Edi
Pentaho report designer: pentaho data statistics Tool 589 people read comments (0) collect reports
1. Set the publish password on the Bi server.
Pentaho release PasswordLocated in pentaho-Solutions/SystemPublisher_config.xml
After setting the release password, you can directly copy the report-designProgramPubli
The Pentaho project is divided into three main parts:
Üpentaho engine (This section after basic rarely changes)
Üpentaho-solution (solution, which is the part that is built on the basis of different needs)
Üpentaho-style (This is a standalone application, full-time responsible for displaying the style)
Pentaho Home Research Note (home.jsp)
The template of the homepage is ${solution-path}/system/custom/temp
Pentaho Report Designer Connecting HSQLDB issuesPentahoreport Designer IntroductionThe latest version of Pentaho Report Designer is 5.1, a very easy-to-use, Java-Lightweight, open source reporting software. Reports can be integrated into a Java project (b/S or C/s) or deployed as a separate reporting platform. You can use Pentaho report to design operational repo
1. Create a new dashboard
After logging on to pentaho, click File-> New-> new dashboard to create a new dashboard.
Ii. new dashboardThe new dashboard is shown in the following figure: layout structure, component panel, and dashboard panel.
3. Layout StructureLayout structure is used to manage the layout of the entire dashboard. Generally, a dashboard is composed of tables, which are divided into row and column. You can click an element to edit it
Tags: 1 Download the JDBC driver for SQL Server first. See the following link address: [1] http://msdn.microsoft.com/en-us/data/aa937724.aspx[2] Google input into SQL Server JDBC is also available. [3] Here Sqljdbc4.jar is the jar package we need 2 download Pentaho of the multidimensional data server mondrian and [1] http://sourceforge.net/→ input Mondrian Download { As of press time, the latest version of mondrian3.5.0} backup address is as follows:
Pentaho biserver Community edtion 6.1 is a kettle component that can run kettle program scripts. However, since Kettle does not publish directly to the BISERVER-CE service, the ETL scripts (. KTR . KJB) that are developed in the local (Windows environment) through a graphical interface need to be uploaded to the Biserver-ce managed repository. Can be run and dispatched by Biserver-ce.Focus: Kettle Repository and BISERVER-CE resource pool establish a c
First, Introduction:Pentaho BI Server is divided into two versions of enterprise and Community editions. Where the Community edition CE (community edtion) is the free version.Second, download the CE version (CentOS):Background Download command:Nohup wget-b-c-q HTTP://JAIST.DL.SOURCEFORGE.NET/PROJECT/PENTAHO/BUSINESS%20INTELLIGENCE%20SERVER/6.1/ Biserver-ce-6.1.0.1-196.zip Third, installationUnzip biserver-ce-6.1.0.1-196.zip-d/opt/ptools/, automaticall
multiple) under Properties: Locale (Specify the country language code, such as: EN_US,ZH_CN Value: the corresponding text (5) Localized_tooltip/tooltip (plugin hint text, can be multiple) under Properties: Locale (Specify the country language code, such as: EN_US,ZH_CN Value: the corresponding text C. Second way: Scan All of the jar packages in these three directories have type-corresponding declared classes (this method needs to be done through the definition file) type of interface
website Link: http://wiki.pentaho.com/display/EAI/Call+DB+Procedure DescriptionCalling the database stored procedure step allows the user to execute a database stored procedure and obtain the results. Stored procedures or methods can only return
1. It can only have one primary query result set.
2. Place the $ {parameter name} parameter in the SQL statement at the bottom of the DATA page in the upper-right corner. The parameter must have a default value. Otherwise, no DATA is displayed in
This example is simple and the difficulty lies in the installation of your Hadoop2.20 plugin (my previous blog post). The steps to implement are as follows:
1. Create a job
Create a kettle job to achieve the following effects.
2. Configure
Kettle requires a JDK environment. You can download it on the Oracle official website. In addition, JDBC or ODBC is required to use kettle. I prefer JDBC. I hate to understand the concept and knowledge of JDBC.
"What is JDBC?
JDBC (Java Data Base
The layout of the dashboard is very troublesome. This CSS layout is a virtue. This mainly involves layout and compontens layout size settings.
Layout layout includes row objects and column objects, as well as images and HTML objects.
A row object
Http://www.aboutyun.com/thread-7450-1-1.html
There is a very large table: Trlog The table is about 2T.Trlog:CREATE TABLE Trlog(PLATFORM string,user_id int,Click_time String,Click_url string)Row format delimitedFields terminated by ' t ';
Ls-a to display all objects, because.) The object of XX is hidden by default)
Execute again./spoon.sh
[Cognos@bitic data-integration]$./spoon.sh
/home/cognos/pdi-ce-4.2.0-stable/data-integrationINFO 11-11 14:56:34,164-spoon-logging goes to File:///tmp/spoon_66d28e63-4a9e-11e3-a301-7b09d1d32e5b.logINFO 11-11 14:56:35,641-logging to Org.slf4j.impl.JCLLoggerAdapter (Org.mortbay.log) via Org.mortbay.log.Slf4jLogINFO 11-11 14:56:35,646-class org.pentaho.
How to use a PDI job to move a file into HDFS.PrerequisitesIn order to follow along with this how-to guide you'll need the following:
Hadoop
Pentaho Data Integration
Sample FilesThe sample data file needed is:
File Name
Content
Weblogs_rebuild.
1 Introduction:The project recently introduced Big Data technology, using it to process the processing day on-line data, need to kettle the source system text data load into the Hadoop environment2 Preparatory work:1 FirstTo understand the Kettle version of the support Hadoop , because the kettle data online less, so it is best to go to the official website, the URL:Http://wiki.pentaho.com/display/BAD/Configuring+Pentaho+for+your+Hadoop+Distro+and+Ve
1 Introduction:The project recently introduced Big Data technology, using its processing day-to-date data on the internet, requiring kettle to load raw text data into the Hadoop environment2 Preparatory work:1 FirstTo understand the Kettle version of the support Hadoop , because the kettle data online less, so it is best to go to the official website, the URL:Http://wiki.pentaho.com/display/BAD/Configuring+Pentaho+for+your+Hadoop+Distro+and+VersionOp
The content source of this page is from Internet, which doesn't represent Alibaba Cloud's opinion;
products and services mentioned on that page don't have any relationship with Alibaba Cloud. If the
content of the page makes you feel confusing, please write us an email, we will handle the problem
within 5 days after receiving your email.
If you find any instances of plagiarism from the community, please send an email to:
info-contact@alibabacloud.com
and provide relevant evidence. A staff member will contact you within 5 working days.