In the past, we used the underlying C-API of each database as wrapping to realize the function of data import and export between several heterogeneous databases. However, the code is complex and it is inconvenient to open source.
In the afternoon, a simple data extraction program was written in Java to port the MySQL database to Sybase ASE. Put it open-source, put it on: http://code.google.com/p/jmyetl/ top. I originally named myetl, and someone applied for it on sf.net. Then I added a J to it.
ETL scheduling development (4) -- file subroutine loading through FTP
The most basic function of the ETL tool is to load files on the remote server. The following applet obtains files on the remote server in binary mode:
#! /Usr/bin/bash # created by lubinsu #2014 source ~ /. Bash_profilefilename = $6 srcdir = $4 descdir = $5 ftpip = $1 ftpusr = $2 ftppwd = $3 # get filesftp-I-in
The input parameters
In this section, we mainly talk about my game transaction Data Analysis Project ETL (data extraction, loading, conversion) exactly how to do.
First of all, the next source system, because our main trading station server is not in the company, so can not directly from the source system directly extracted data. In fact, we already have a simple data analysis system, but this is the previous people do, not using sqlserver2005 bi platform to do, but dire
The kettle of ETL tools extracts data from one database into another database:
1. Open the ETL folder, double-click Spoon.bat start Kettle
2. Resource pool selection, Connaught no choice to cancel
3. Select Close
4. Create a new transformation
5. Configure the required database
6. The data table that needs to be extracted, with the table input to get
7. Select the database and table
Tags: ETL kettle jdbc Oracle RAC1 problem Phenomena:Previously done Kettle connect an Oracle database for table extractionThe table input information for the script is as follows:Error message in the table input report when executing (script uploaded to Linux machine with sh command) :But in the machine with the Sqlplus command login can be successful:2 resolution process:After the problem, the first contact with the source data system manufacturers t
BI Development process and ETL introduction
BI Development process1. Building Dimension-Fact model2. Build data warehouses (dimensions, facts) based on dimension-fact model3. Data extraction (ETL)4. Analysis model topics for building sales information5. Build report analysis, instrument panel
The BI Business intelligence system, according to the enterprise needs to solve the problem, helps the enterpris
Label:The collation of SQL Server is roughly divided into Windows collation and SQL Servers collation. When the data is installed, defaults to Sql_latin1_general_cp1_ci_ai are not set by default. When the database is created, if you do not set a collation that uses the default data, you can also set the collation for the columns in the table.Here are just a few things to keep in mind when you have recently encountered such problems.First Sql_latin1_general_cp1_ci_ai corresponds to 1252, while ch
This document describes the ETL testing process and general project conditions to describe the ETL testing method.
ETL test Flowchart
Test phase
1,Requirement Analysis
Familiar with business processes and business rules, analyze the ing relationship between the source table and the target table as required, and parse the business data flow diagram:
1,Test Ana
ETL is responsible for the scattered, heterogeneous data sources such as relational data, flat data files, such as the extraction of the temporary middle layer after the cleaning, transformation, integration, and finally loaded into the data warehouse or data mart, as the basis for online analysis processing, data mining. The term ETL often appears in the Data warehouse, but its object is not confined to th
During the three-day holiday on May Day, some ETL logic problems occurred, resulting in the daily incremental data to be loaded into DW is not loaded as designed. Therefore, you need to check the generated incremental data after ETL to avoid the problem of passive processing when the incremental data is lost one day.
Requirement: if there is a problem with the incremental data of
What does the ETL data conversion system bring to customers?With the development of society and computer technology, people began to reprocess the data in the original database to form a comprehensive and analysis-oriented environment to support the emergence of scientific decision-making. As a result, the ideas, technologies, and products of data warehouses are gradually formed. The purpose of building a data warehouse is to establish a systematic da
Label:Data integration has become the enterprise in the pursuit of market share of the key technology components, and rely on manual coding in different ways, more and more enterprises choose a complete data integration solution to support its IT strategy, from big data analysis to cloud platform integration.A recent study by Dao Research compares the differences between several of the world's leading data integration solutions and the impact of these product technologies on the real business. T
; mysql. sqlecho "userId"> mysql. sqlecho "case"> mysql. SQL sed-I-e '1d 'm.txt cat m.txt | while read line do par1 =1 (echo "$ {line}" | awk-f''' {print $1} ') par2 = $ (echo "$ {line}" | awk-F ''' {print $2 }') id = $ (echo "$ {line}" | awk-F ''' {print $3} ') echo "par1 :1 {par1}" echo "par2: $ {par2} "echo" when hour_time >=$ {par1} and hour_time
3) All scripts are stored in the database, and parameters are parsed and called and executed by the program.
Refer to kettle design:
Each
What is ETL?
SDE: Source Dependent Extract
SDE mappings -- extracts the data from the transactional Source System and loads into the data warehouse staging tables.
SDE mappings are designed with respect to the source's unique data model.
SDE _ * workflows have only the staging table, the workflow will load the data into the staging area tables.
In the staging the tables will not have index.
It always truncates the data and loads the data into staging
ETL Application scenario, if the interface file is not provided, the task will be in the loop wait until the peer to provide, the method greatly consumes the system resources. To this end think of a method, one time to obtain a platform file, the realization of the following ideas:1, the first time to obtain the peer platform to provide the directory under the given date all the interface files, and save the file list;2, the subsequent restart every n
First, the purposeMerge tables on different servers onto another server. For example, merge table B on server 1 on table A and server 2 to table C on server 3Requirements: Table A needs to be cropped (removing unnecessary fields), table B needs to add some fieldsIi. Methods of Use(1) Create a new Table C (field that conforms to the actual system design) in the database on server 3(2) Create a new table input, connect to server 1, select the table you want to use by getting the SQL statement, or
Introduction to the outsetToday, when you load data using QV, you run into some state in the column, and the information is separated by a symbol, which is not conducive to data analysis because the content in the string is itself a dimension. Search the Internet to find a solution to the method, record.For example, in the first picture, s200,m250,r35 are all Invoice types, which need to be taken out as the dimension DIMENSION of the analysis.You can use the following code to achieve the separat
identifier for tombstone dataCreate multiset volatile table Del5. Inserting data into the temporary table in accordance with certain loading rulesINSERT INTO new6, using the data of the temporary table and the warehouse table data as a comparison of the newly changed data into the Delta tableINSERT INTO Inc Select ... from new7, the source table data has a special identification (generally end_dt=min_date) into the delete tableInsert INTO del Select. From New where end_dt=min_date8, to all in t
default, Python uses ASCII encoding as follows:Python-c "Import sys; Print sys.getdefaultencoding () "ASCIIand when Python converts between encodings, Unicode is used as an "intermediate encoding", but Unicode is the largest So long, so here when trying to put ASCII The encoded string is converted into " Intermediate Encoding the Unicode due to exceeding its range, the above error has been reported. 2. Solutions1) First: Here we will change the python default encoding mode to utf-8, we can
The content source of this page is from Internet, which doesn't represent Alibaba Cloud's opinion;
products and services mentioned on that page don't have any relationship with Alibaba Cloud. If the
content of the page makes you feel confusing, please write us an email, we will handle the problem
within 5 days after receiving your email.
If you find any instances of plagiarism from the community, please send an email to:
info-contact@alibabacloud.com
and provide relevant evidence. A staff member will contact you within 5 working days.