Äú¿ÉÒÔ¾èÖú£¬Ö§³ÖÎÒÃǵĹ«ÒæÊÂÒµ¡£

1Ôª 10Ôª 50Ôª





ÈÏÖ¤Â룺  ÑéÖ¤Âë,¿´²»Çå³þ?Çëµã»÷Ë¢ÐÂÑéÖ¤Âë ±ØÌî



  ÇóÖª ÎÄÕ ÎÄ¿â Lib ÊÓÆµ iPerson ¿Î³Ì ÈÏÖ¤ ×Éѯ ¹¤¾ß ½²×ù Modeler   Code  
»áÔ±   
 
   
 
 
     
   
 ¶©ÔÄ
  ¾èÖú
ÓÃMaven¹¹½¨HadoopÏîÄ¿
 
×÷Õß Õŵ¤(Conan)£¬»ðÁú¹ûÈí¼þ    ·¢²¼ÓÚ 2014-07-30
  5101  次浏览      27
 

ǰÑÔ

HadoopµÄMapReduce»·¾³ÊÇÒ»¸ö¸´Ôӵıà³Ì»·¾³£¬ËùÒÔÎÒÃÇÒª¾¡¿ÉÄܵؼò»¯¹¹½¨MapReduceÏîÄ¿µÄ¹ý³Ì¡£MavenÊÇÒ»¸öºÜ²»´íµÄ×Ô¶¯»¯ÏîÄ¿¹¹½¨¹¤¾ß£¬Í¨¹ýMavenÀ´°ïÖúÎÒÃÇ´Ó¸´ÔӵĻ·¾³ÅäÖÃÖнâÍѳöÀ´£¬´Ó¶ø±ê×¼»¯¿ª·¢¹ý³Ì¡£ËùÒÔ£¬Ð´MapReduce֮ǰ£¬ÈÃÎÒÃÇÏÈ»¨µãʱ¼ä°Ñµ¶Ä¥¿ì£¡£¡µ±È»£¬³ýÁËMaven»¹ÓÐÆäËûµÄÑ¡ÔñGradle(ÍÆ¼ö), Ivy¡­.

ºóÃæ½«»áÓнéÉܼ¸ÆªMapReduce¿ª·¢µÄÎÄÕ£¬¶¼ÒªÒÀÀµÓÚ±¾ÎÄÖÐMavenµÄ¹¹½¨µÄMapReduce»·¾³¡£

1. Maven½éÉÜ

Apache Maven£¬ÊÇÒ»¸öJavaµÄÏîÄ¿¹ÜÀí¼°×Ô¶¯¹¹½¨¹¤¾ß£¬ÓÉApacheÈí¼þ»ù½ð»áËùÌṩ¡£»ùÓÚÏîÄ¿¶ÔÏóÄ£ÐÍ£¨Ëõд£ºPOM£©¸ÅÄMavenÀûÓÃÒ»¸öÖÐÑëÐÅϢƬ¶ÏÄܹÜÀíÒ»¸öÏîÄ¿µÄ¹¹½¨¡¢±¨¸æºÍÎĵµµÈ²½Öè¡£ÔøÊÇJakartaÏîÄ¿µÄ×ÓÏîÄ¿£¬ÏÖΪ¶ÀÁ¢ApacheÏîÄ¿¡£

mavenµÄ¿ª·¢ÕßÔÚËûÃÇ¿ª·¢ÍøÕ¾ÉÏÖ¸³ö£¬mavenµÄÄ¿±êÊÇҪʹµÃÏîÄ¿µÄ¹¹½¨¸ü¼ÓÈÝÒ×£¬Ëü°Ñ±àÒë¡¢´ò°ü¡¢²âÊÔ¡¢·¢²¼µÈ¿ª·¢¹ý³ÌÖеIJ»Í¬»·½ÚÓлúµÄ´®ÁªÁËÆðÀ´£¬²¢²úÉúÒ»Öµġ¢¸ßÖÊÁ¿µÄÏîÄ¿ÐÅÏ¢£¬Ê¹µÃÏîÄ¿³ÉÔ±Äܹ»¼°Ê±µØµÃµ½·´À¡¡£mavenÓÐЧµØÖ§³ÖÁ˲âÊÔÓÅÏÈ¡¢³ÖÐø¼¯³É£¬ÌåÏÖÁ˹ÄÀø¹µÍ¨£¬¼°Ê±·´À¡µÄÈí¼þ¿ª·¢ÀíÄî¡£Èç¹û˵AntµÄ¸´ÓÃÊǽ¨Á¢ÔÚ¡±¿½±´¨CÕ³Ìù¡±µÄ»ù´¡Éϵģ¬ÄÇôMavenͨ¹ý²å¼þµÄ»úÖÆÊµÏÖÁËÏîÄ¿¹¹½¨Âß¼­µÄÕæÕý¸´Óá£

2. Maven°²×°(win)

ÏÂÔØMaven£ºhttp://maven.apache.org/download.cgi

ÏÂÔØ×îеÄxxx-bin.zipÎļþ£¬ÔÚwinÉϽâѹµ½ D:\toolkit\maven3

²¢°Ñmaven/binĿ¼ÉèÖÃÔÚ»·¾³±äÁ¿PATH£º

È»ºó£¬´ò¿ªÃüÁîÐÐÊäÈëmvn£¬ÎÒÃǻῴµ½mvnÃüÁîµÄÔËÐÐЧ¹û

~ C:\Users\Administrator>mvn
[INFO] Scanning for projects...
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 0.086s
[INFO] Finished at: Mon Sep 30 18:26:58 CST 2013
[INFO] Final Memory: 2M/179M
[INFO] ------------------------------------------------------------------------
[ERROR] No goals have been specified for this build. You must specify a valid lifecycle phase or a goal in the format : or :[:]:. Available lifecycle phases are: validate, initialize, generate-sources, process-sources, generate-resources, process-resources, compile, process-class
es, generate-test-sources, process-test-sources, generate-test-resources, process-test-resources, test-compile, process-test-classes, test, prepare-package, package, pre-integration-test, integration-test, post-integration-test, verify, install, deploy, pre-clean, clean, post-clean, pre-site, site, post-site, site-deploy. -> [Help 1]
[ERROR]
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR]
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/NoGoalSpecifiedException

°²×°EclipseµÄMaven²å¼þ£ºMaven Integration for Eclipse

MavenµÄEclipse²å¼þÅäÖÃ

3. Hadoop¿ª·¢»·¾³½éÉÜ

ÈçÉÏͼËùʾ£¬ÎÒÃÇ¿ÉÒÔÑ¡ÔñÔÚwinÖпª·¢£¬Ò²¿ÉÒÔÔÚlinuxÖпª·¢£¬±¾µØÆô¶¯Hadoop»òÕßÔ¶³Ìµ÷ÓÃHadoop£¬±êÅäµÄ¹¤¾ß¶¼ÊÇMavenºÍEclipse¡£

Hadoop¼¯ÈºÏµÍ³»·¾³£º

Linux: Ubuntu 12.04.2 LTS 64bit Server

Java: 1.6.0_29

Hadoop: hadoop-1.0.3£¬µ¥½Úµã£¬IP:192.168.1.210

4. ÓÃMaven¹¹½¨Hadoop»·¾³

1. ÓÃMaven´´½¨Ò»¸ö±ê×¼»¯µÄJavaÏîÄ¿

2. µ¼ÈëÏîÄ¿µ½eclipse

3. Ôö¼ÓhadoopÒÀÀµ£¬ÐÞ¸Äpom.xml

4. ÏÂÔØÒÀÀµ

5. ´ÓHadoop¼¯Èº»·¾³ÏÂÔØhadoopÅäÖÃÎļþ

6. ÅäÖñ¾µØhost

1). ÓÃMaven´´½¨Ò»¸ö±ê×¼»¯µÄJavaÏîÄ¿

~ D:\workspace\java>mvn archetype:generate -DarchetypeGroupId=
org.apache.maven.archetypes -DgroupId=org.conan.myhadoop.mr
-DartifactId=myHadoop -DpackageName=org.conan.myhadoop.mr -Dversion= 1.0-SNAPSHOT -DinteractiveMode=false
[INFO] Scanning for projects...
[INFO]
[INFO] ---------------------------------------------
[INFO] Building Maven Stub Project (No POM) 1
[INFO] -----------------------------------------------
[INFO]
[INFO] >>> maven-archetype-plugin:2.2:generate (default-cli) @ standalone-pom >>>
[INFO]
[INFO] <<< maven-archetype-plugin:2.2:generate (default-cli) @ standalone-pom <<<
[INFO]
[INFO] --- maven-archetype-plugin:2.2:generate (default-cli) @ standalone-pom ---
[INFO] Generating project in Batch mode
[INFO] No archetype defined. Using maven-archetype-quickstart (org.apache.maven.archetypes:maven-archetype-quickstart:1.0)
Downloading: http://repo.maven.apache.org/maven2/org/apache/maven/archetypes/ maven-archetype-quickstart/1.0/maven-archet
ype-quickstart-1.0.jar
Downloaded: http://repo.maven.apache.org/maven2/org/apache/maven/archetypes/ maven-archetype-quickstart/1.0/maven-archety
pe-quickstart-1.0.jar (5 KB at 4.3 KB/sec)
Downloading: http://repo.maven.apache.org/maven2/org/apache/maven/archetypes/ maven-archetype-quickstart/1.0/maven-archet
ype-quickstart-1.0.pom
Downloaded: http://repo.maven.apache.org/maven2/org/apache/maven/archetypes /maven-archetype-quickstart/1.0/maven-archety
pe-quickstart-1.0.pom (703 B at 1.6 KB/sec)
[INFO] --------------------------------------------------
[INFO] Using following parameters for creating project from Old (1.x) Archetype: maven-archetype-quickstart:1.0
[INFO] ----------------------------------------------------
[INFO] Parameter: groupId, Value: org.conan.myhadoop.mr
[INFO] Parameter: packageName, Value: org.conan.myhadoop.mr
[INFO] Parameter: package, Value: org.conan.myhadoop.mr
[INFO] Parameter: artifactId, Value: myHadoop
[INFO] Parameter: basedir, Value: D:\workspace\java
[INFO] Parameter: version, Value: 1.0-SNAPSHOT
[INFO] project created from Old (1.x) Archetype in dir: D:\workspace\java\myHadoop
[INFO] ----------------------------------------------------
[INFO] BUILD SUCCESS
[INFO] ----------------------------------------------------
[INFO] Total time: 8.896s
[INFO] Finished at: Sun Sep 29 20:57:07 CST 2013
[INFO] Final Memory: 9M/179M
[INFO] ----------------------------------------------------

½øÈëÏîÄ¿£¬Ö´ÐÐmvnÃüÁî

~ D:\workspace\java>cd myHadoop
~ D:\workspace\java\myHadoop>mvn clean install
[INFO]
[INFO] --- maven-jar-plugin:2.3.2:jar (default-jar) @ myHadoop ---
[INFO] Building jar: D:\workspace\java\myHadoop\target\myHadoop-1.0-SNAPSHOT.jar
[INFO]
[INFO] --- maven-install-plugin:2.3.1:install (default-install) @ myHadoop ---
[INFO] Installing D:\workspace\java\myHadoop\target\myHadoop-1.0-SNAPSHOT.jar to C:\Users\Administrator\.m2\repository\o
rg\conan\myhadoop\mr\myHadoop\1.0-SNAPSHOT\myHadoop-1.0-SNAPSHOT.jar
[INFO] Installing D:\workspace\java\myHadoop\pom.xml to C:\Users\Administrator\.m2\repository\org\conan\myhadoop\mr\myHa
doop\1.0-SNAPSHOT\myHadoop-1.0-SNAPSHOT.pom
[INFO] -------------------------------------------------------
[INFO] BUILD SUCCESS
[INFO] -------------------------------------------------------
[INFO] Total time: 4.348s
[INFO] Finished at: Sun Sep 29 20:58:43 CST 2013
[INFO] Final Memory: 11M/179M
[INFO] -------------------------------------------------------

2). µ¼ÈëÏîÄ¿µ½eclipse

ÎÒÃÇ´´½¨ºÃÁËÒ»¸ö»ù±¾µÄmavenÏîÄ¿£¬È»ºóµ¼Èëµ½eclipseÖС£ ÕâÀïÎÒÃÇ×îºÃÒѰ²×°ºÃÁËMavenµÄ²å¼þ¡£

3). Ôö¼ÓhadoopÒÀÀµ

ÕâÀïÎÒʹÓÃhadoop-1.0.3°æ±¾£¬ÐÞ¸ÄÎļþ£ºpom.xml

~ vi pom.xml

<project xmlns="http://maven.apache.org/POM/4.0.0" xmlns:xsi=

"http://www.w3.org/2001/XMLSchema-instance"
xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 http://maven.apache.org/maven-v4_0_0.xsd">
<modelVersion>4.0.0</modelVersion>
<groupId>org.conan.myhadoop.mr</groupId>
<artifactId>myHadoop</artifactId>
<packaging>jar</packaging>
<version>1.0-SNAPSHOT</version>
<name>myHadoop</name>
<url>http://maven.apache.org</url>
<dependencies>
<dependency>
<groupId>org.apache.hadoop</groupId>
<artifactId>hadoop-core</artifactId>
<version>1.0.3</version>
</dependency>

<dependency>
<groupId>junit</groupId>
<artifactId>junit</artifactId>
<version>4.4</version>
<scope>test</scope>
</dependency>
</dependencies>
</project>

4). ÏÂÔØÒÀÀµ

ÏÂÔØÒÀÀµ£º

~ mvn clean install

ÔÚeclipseÖÐË¢ÐÂÏîÄ¿£º

ÏîÄ¿µÄÒÀÀµ³ÌÐò£¬±»×Ô¶¯¼ÓÔØµÄ¿â·¾¶ÏÂÃæ¡£

5). ´ÓHadoop¼¯Èº»·¾³ÏÂÔØhadoopÅäÖÃÎļþ

core-site.xml
hdfs-site.xml
mapred-site.xml

²é¿´core-site.xml

<?xml version="1.0"?>
<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>

<configuration>
<property>
<name>fs.default.name</name>
<value>hdfs://master:9000</value>
</property>
<property>
<name>hadoop.tmp.dir</name>
<value>/home/conan/hadoop/tmp</value>
</property>
<property>
<name>io.sort.mb</name>
<value>256</value>
</property>
</configuration>

²é¿´hdfs-site.xml

<?xml version="1.0"?>
<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>

<configuration>
<property>
<name>dfs.data.dir</name>
<value>/home/conan/hadoop/data</value>
</property>
<property>
<name>dfs.replication</name>
<value>1</value>
</property>
<property>
<name>dfs.permissions</name>
<value>false</value>
</property>
</configuration>

²é¿´mapred-site.xml

<?xml version="1.0"?>
<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>

<configuration>
<property>
<name>mapred.job.tracker</name>
<value>hdfs://master:9001</value>
</property>
</configuration>

±£´æÔÚsrc/main/resources/hadoopĿ¼ÏÂÃæ

ɾ³ýÔ­×Ô¶¯Éú³ÉµÄÎļþ£ºApp.javaºÍAppTest.java

6).ÅäÖñ¾µØhost£¬Ôö¼ÓmasterµÄÓòÃûÖ¸Ïò

~ vi c:/Windows/System32/drivers/etc/hosts

192.168.1.210 master

5. MapReduce³ÌÐò¿ª·¢

±àдһ¸ö¼òµ¥µÄMapReduce³ÌÐò£¬ÊµÏÖwordcount¹¦ÄÜ¡£

ÐÂÒ»¸öJavaÎļþ£ºWordCount.java

package org.conan.myhadoop.mr;

import java.io.IOException;
import java.util.Iterator;
import java.util.StringTokenizer;

import org.apache.hadoop.fs.Path;
import org.apache.hadoop.io.IntWritable;
import org.apache.hadoop.io.Text;
import org.apache.hadoop.mapred.FileInputFormat;
import org.apache.hadoop.mapred.FileOutputFormat;
import org.apache.hadoop.mapred.JobClient;
import org.apache.hadoop.mapred.JobConf;
import org.apache.hadoop.mapred.MapReduceBase;
import org.apache.hadoop.mapred.Mapper;
import org.apache.hadoop.mapred.OutputCollector;
import org.apache.hadoop.mapred.Reducer;
import org.apache.hadoop.mapred.Reporter;
import org.apache.hadoop.mapred.TextInputFormat;
import org.apache.hadoop.mapred.TextOutputFormat;

public class WordCount {

public static class WordCountMapper extends MapReduceBase implements Mapper<Object, Text, Text, IntWritable> {
private final static IntWritable one = new IntWritable(1);
private Text word = new Text();

@Override
public void map(Object key, Text value, OutputCollector<Text, IntWritable> output, Reporter reporter) throws IOException {
StringTokenizer itr = new StringTokenizer(value.toString());
while (itr.hasMoreTokens()) {
word.set(itr.nextToken());
output.collect(word, one);
}

}
}

public static class WordCountReducer extends MapReduceBase implements Reducer<Text, IntWritable, Text, IntWritable> {
private IntWritable result = new IntWritable();

@Override
public void reduce(Text key, Iterator values, OutputCollector<Text, IntWritable> output, Reporter reporter) throws IOException {
int sum = 0;
while (values.hasNext()) {
sum += values.next().get();
}
result.set(sum);
output.collect(key, result);
}
}

public static void main(String[] args) throws Exception {
String input = "hdfs://192.168.1.210:9000/user/hdfs/o_t_account";
String output = "hdfs://192.168.1.210:9000/user/hdfs/o_t_account/result";

JobConf conf = new JobConf(WordCount.class);
conf.setJobName("WordCount");
conf.addResource("classpath:/hadoop/core-site.xml");
conf.addResource("classpath:/hadoop/hdfs-site.xml");
conf.addResource("classpath:/hadoop/mapred-site.xml");

conf.setOutputKeyClass(Text.class);
conf.setOutputValueClass(IntWritable.class);

conf.setMapperClass(WordCountMapper.class);
conf.setCombinerClass(WordCountReducer.class);
conf.setReducerClass(WordCountReducer.class);

conf.setInputFormat(TextInputFormat.class);
conf.setOutputFormat(TextOutputFormat.class);

FileInputFormat.setInputPaths(conf, new Path(input));
FileOutputFormat.setOutputPath(conf, new Path(output));

JobClient.runJob(conf);
System.exit(0);
}

}

Æô¶¯Java APP.

¿ØÖÆÌ¨´íÎó

2013-9-30 19:25:02 org.apache.hadoop.util.NativeCodeLoader 
¾¯¸æ: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
2013-9-30 19:25:02 org.apache.hadoop.security.UserGroupInformation doAs
ÑÏÖØ: PriviledgedActionException as:Administrator cause:java.io.IOException: Failed to set permissions of path: \tmp\hadoop-Administrator\mapred\staging\Administrator1702422322\.staging to 0700
Exception in thread "main" java.io.IOException: Failed to set permissions of path: \tmp\hadoop-Administrator\mapred\staging\Administrator1702422322\.staging to 0700
at org.apache.hadoop.fs.FileUtil.checkReturnValue(FileUtil.java:689)
at org.apache.hadoop.fs.FileUtil.setPermission(FileUtil.java:662)
at org.apache.hadoop.fs.RawLocalFileSystem.setPermission(RawLocalFileSystem.java:509)
at org.apache.hadoop.fs.RawLocalFileSystem.mkdirs(RawLocalFileSystem.java:344)
at org.apache.hadoop.fs.FilterFileSystem.mkdirs(FilterFileSystem.java:189)
at org.apache.hadoop.mapreduce.JobSubmissionFiles.getStagingDir(JobSubmissionFiles.java:116)
at org.apache.hadoop.mapred.JobClient$2.run(JobClient.java:856)
at org.apache.hadoop.mapred.JobClient$2.run(JobClient.java:850)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:396)
at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1121)
at org.apache.hadoop.mapred.JobClient.submitJobInternal(JobClient.java:850)
at org.apache.hadoop.mapred.JobClient.submitJob(JobClient.java:824)
at org.apache.hadoop.mapred.JobClient.runJob(JobClient.java:1261)
at org.conan.myhadoop.mr.WordCount.main(WordCount.java:78)

Õâ¸ö´íÎóÊÇwinÖпª·¢ÌØÓеĴíÎó£¬ÎļþȨÏÞÎÊÌ⣬ÔÚLinuxÏ¿ÉÒÔÕý³£ÔËÐС£

½â¾ö·½·¨ÊÇ£¬ÐÞ¸Ä/hadoop-1.0.3/src/core/org/apache/hadoop/fs/FileUtil.javaÎļþ

688-692ÐÐ×¢ÊÍ£¬È»ºóÖØÐ±àÒëÔ´´úÂë£¬ÖØÐ´òÒ»¸öhadoop.jarµÄ°ü¡£

685 private static void checkReturnValue(boolean rv, File p,
686 FsPermission permission
687 ) throws IOException {
688 /*if (!rv) {
689 throw new IOException("Failed to set permissions of path: " + p +
690 " to " +
691 String.format("%04o", permission.toShort()));
692 }*/
693 }

ÎÒÕâÀï×Ô¼º´òÁËÒ»¸öhadoop-core-1.0.3.jar°ü£¬·Åµ½ÁËlibÏÂÃæ¡£

ÎÒÃÇ»¹ÒªÌæ»»mavenÖеÄhadoopÀà¿â¡£

~ cp lib/hadoop-core-1.0.3.jar C:\Users\Administrator
\.m2\repository\org\apache\hadoop\hadoop-core\1.0.3\hadoop-core-1.0.3.jar

ÔÙ´ÎÆô¶¯Java APP£¬¿ØÖÆÌ¨Êä³ö£º

2013-9-30 19:50:49 org.apache.hadoop.util.NativeCodeLoader 
¾¯¸æ: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
2013-9-30 19:50:49 org.apache.hadoop.mapred.JobClient copyAndConfigureFiles
¾¯¸æ: Use GenericOptionsParser for parsing the arguments. Applications should implement Tool for the same.
2013-9-30 19:50:49 org.apache.hadoop.mapred.JobClient copyAndConfigureFiles
¾¯¸æ: No job jar file set.  User classes may not be found. See JobConf(Class) or JobConf#setJar(String).
2013-9-30 19:50:49 org.apache.hadoop.io.compress.snappy.LoadSnappy 
¾¯¸æ: Snappy native library not loaded
2013-9-30 19:50:49 org.apache.hadoop.mapred.FileInputFormat listStatus
ÐÅÏ¢: Total input paths to process : 4
2013-9-30 19:50:50 org.apache.hadoop.mapred.JobClient monitorAndPrintJob
ÐÅÏ¢: Running job: job_local_0001
2013-9-30 19:50:50 org.apache.hadoop.mapred.Task initialize
ÐÅÏ¢:  Using ResourceCalculatorPlugin : null
2013-9-30 19:50:50 org.apache.hadoop.mapred.MapTask runOldMapper
ÐÅÏ¢: numReduceTasks: 1
2013-9-30 19:50:50 org.apache.hadoop.mapred.MapTask$MapOutputBuffer 
ÐÅÏ¢: io.sort.mb = 100
2013-9-30 19:50:50 org.apache.hadoop.mapred.MapTask$MapOutputBuffer 
ÐÅÏ¢: data buffer = 79691776/99614720
2013-9-30 19:50:50 org.apache.hadoop.mapred.MapTask$MapOutputBuffer 
ÐÅÏ¢: record buffer = 262144/327680
2013-9-30 19:50:50 org.apache.hadoop.mapred.MapTask$MapOutputBuffer flush
ÐÅÏ¢: Starting flush of map output
2013-9-30 19:50:50 org.apache.hadoop.mapred.MapTask$MapOutputBuffer sortAndSpill
ÐÅÏ¢: Finished spill 0
2013-9-30 19:50:50 org.apache.hadoop.mapred.Task done
ÐÅÏ¢: Task:attempt_local_0001_m_000000_0 is done. And is in the process of commiting
2013-9-30 19:50:51 org.apache.hadoop.mapred.JobClient monitorAndPrintJob
ÐÅÏ¢:  map 0% reduce 0%
2013-9-30 19:50:53 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate
ÐÅÏ¢: hdfs://192.168.1.210:9000/user/hdfs/o_t_account/part-m-00003:0+119
2013-9-30 19:50:53 org.apache.hadoop.mapred.Task sendDone
ÐÅÏ¢: Task 'attempt_local_0001_m_000000_0' done.
2013-9-30 19:50:53 org.apache.hadoop.mapred.Task initialize
ÐÅÏ¢:  Using ResourceCalculatorPlugin : null
2013-9-30 19:50:53 org.apache.hadoop.mapred.MapTask runOldMapper
ÐÅÏ¢: numReduceTasks: 1
2013-9-30 19:50:53 org.apache.hadoop.mapred.MapTask$MapOutputBuffer 
ÐÅÏ¢: io.sort.mb = 100
2013-9-30 19:50:53 org.apache.hadoop.mapred.MapTask$MapOutputBuffer 
ÐÅÏ¢: data buffer = 79691776/99614720
2013-9-30 19:50:53 org.apache.hadoop.mapred.MapTask$MapOutputBuffer 
ÐÅÏ¢: record buffer = 262144/327680
2013-9-30 19:50:53 org.apache.hadoop.mapred.MapTask$MapOutputBuffer flush
ÐÅÏ¢: Starting flush of map output
2013-9-30 19:50:53 org.apache.hadoop.mapred.MapTask$MapOutputBuffer sortAndSpill
ÐÅÏ¢: Finished spill 0
2013-9-30 19:50:53 org.apache.hadoop.mapred.Task done
ÐÅÏ¢: Task:attempt_local_0001_m_000001_0 is done. And is in the process of commiting
2013-9-30 19:50:54 org.apache.hadoop.mapred.JobClient monitorAndPrintJob
ÐÅÏ¢:  map 100% reduce 0%
2013-9-30 19:50:56 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate
ÐÅÏ¢: hdfs://192.168.1.210:9000/user/hdfs/o_t_account/part-m-00000:0+113
2013-9-30 19:50:56 org.apache.hadoop.mapred.Task sendDone
ÐÅÏ¢: Task 'attempt_local_0001_m_000001_0' done.
2013-9-30 19:50:56 org.apache.hadoop.mapred.Task initialize
ÐÅÏ¢:  Using ResourceCalculatorPlugin : null
2013-9-30 19:50:56 org.apache.hadoop.mapred.MapTask runOldMapper
ÐÅÏ¢: numReduceTasks: 1
2013-9-30 19:50:56 org.apache.hadoop.mapred.MapTask$MapOutputBuffer 
ÐÅÏ¢: io.sort.mb = 100
2013-9-30 19:50:56 org.apache.hadoop.mapred.MapTask$MapOutputBuffer 
ÐÅÏ¢: data buffer = 79691776/99614720
2013-9-30 19:50:56 org.apache.hadoop.mapred.MapTask$MapOutputBuffer 
ÐÅÏ¢: record buffer = 262144/327680
2013-9-30 19:50:56 org.apache.hadoop.mapred.MapTask$MapOutputBuffer flush
ÐÅÏ¢: Starting flush of map output
2013-9-30 19:50:56 org.apache.hadoop.mapred.MapTask$MapOutputBuffer sortAndSpill
ÐÅÏ¢: Finished spill 0
2013-9-30 19:50:56 org.apache.hadoop.mapred.Task done
ÐÅÏ¢: Task:attempt_local_0001_m_000002_0 is done. And is in the process of commiting
2013-9-30 19:50:59 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate
ÐÅÏ¢: hdfs://192.168.1.210:9000/user/hdfs/o_t_account/part-m-00001:0+110
2013-9-30 19:50:59 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate
ÐÅÏ¢: hdfs://192.168.1.210:9000/user/hdfs/o_t_account/part-m-00001:0+110
2013-9-30 19:50:59 org.apache.hadoop.mapred.Task sendDone
ÐÅÏ¢: Task 'attempt_local_0001_m_000002_0' done.
2013-9-30 19:50:59 org.apache.hadoop.mapred.Task initialize
ÐÅÏ¢:  Using ResourceCalculatorPlugin : null
2013-9-30 19:50:59 org.apache.hadoop.mapred.MapTask runOldMapper
ÐÅÏ¢: numReduceTasks: 1
2013-9-30 19:50:59 org.apache.hadoop.mapred.MapTask$MapOutputBuffer 
ÐÅÏ¢: io.sort.mb = 100
2013-9-30 19:50:59 org.apache.hadoop.mapred.MapTask$MapOutputBuffer 
ÐÅÏ¢: data buffer = 79691776/99614720
2013-9-30 19:50:59 org.apache.hadoop.mapred.MapTask$MapOutputBuffer 
ÐÅÏ¢: record buffer = 262144/327680
2013-9-30 19:50:59 org.apache.hadoop.mapred.MapTask$MapOutputBuffer flush
ÐÅÏ¢: Starting flush of map output
2013-9-30 19:50:59 org.apache.hadoop.mapred.MapTask$MapOutputBuffer sortAndSpill
ÐÅÏ¢: Finished spill 0
2013-9-30 19:50:59 org.apache.hadoop.mapred.Task done
ÐÅÏ¢: Task:attempt_local_0001_m_000003_0 is done. And is in the process of commiting
2013-9-30 19:51:02 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate
ÐÅÏ¢: hdfs://192.168.1.210:9000/user/hdfs/o_t_account/part-m-00002:0+79
2013-9-30 19:51:02 org.apache.hadoop.mapred.Task sendDone
ÐÅÏ¢: Task 'attempt_local_0001_m_000003_0' done.
2013-9-30 19:51:02 org.apache.hadoop.mapred.Task initialize
ÐÅÏ¢:  Using ResourceCalculatorPlugin : null
2013-9-30 19:51:02 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate
ÐÅÏ¢: 
2013-9-30 19:51:02 org.apache.hadoop.mapred.Merger$MergeQueue merge
ÐÅÏ¢: Merging 4 sorted segments
2013-9-30 19:51:02 org.apache.hadoop.mapred.Merger$MergeQueue merge
ÐÅÏ¢: Down to the last merge-pass, with 4 segments left of total size: 442 bytes
2013-9-30 19:51:02 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate
ÐÅÏ¢: 
2013-9-30 19:51:02 org.apache.hadoop.mapred.Task done
ÐÅÏ¢: Task:attempt_local_0001_r_000000_0 is done. And is in the process of commiting
2013-9-30 19:51:02 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate
ÐÅÏ¢: 
2013-9-30 19:51:02 org.apache.hadoop.mapred.Task commit
ÐÅÏ¢: Task attempt_local_0001_r_000000_0 is allowed to commit now
2013-9-30 19:51:02 org.apache.hadoop.mapred.FileOutputCommitter commitTask
ÐÅÏ¢: Saved output of task 'attempt_local_0001_r_000000_0' to hdfs://192.168.1.210:9000/user/hdfs/o_t_account/result
2013-9-30 19:51:05 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate
ÐÅÏ¢: reduce > reduce
2013-9-30 19:51:05 org.apache.hadoop.mapred.Task sendDone
ÐÅÏ¢: Task 'attempt_local_0001_r_000000_0' done.
2013-9-30 19:51:06 org.apache.hadoop.mapred.JobClient monitorAndPrintJob
ÐÅÏ¢:  map 100% reduce 100%
2013-9-30 19:51:06 org.apache.hadoop.mapred.JobClient monitorAndPrintJob
ÐÅÏ¢: Job complete: job_local_0001
2013-9-30 19:51:06 org.apache.hadoop.mapred.Counters log
ÐÅÏ¢: Counters: 20
2013-9-30 19:51:06 org.apache.hadoop.mapred.Counters log
ÐÅÏ¢:   File Input Format Counters 
2013-9-30 19:51:06 org.apache.hadoop.mapred.Counters log
ÐÅÏ¢:     Bytes Read=421
2013-9-30 19:51:06 org.apache.hadoop.mapred.Counters log
ÐÅÏ¢:   File Output Format Counters 
2013-9-30 19:51:06 org.apache.hadoop.mapred.Counters log
ÐÅÏ¢:     Bytes Written=348
2013-9-30 19:51:06 org.apache.hadoop.mapred.Counters log
ÐÅÏ¢:   FileSystemCounters
2013-9-30 19:51:06 org.apache.hadoop.mapred.Counters log
ÐÅÏ¢:     FILE_BYTES_READ=7377
2013-9-30 19:51:06 org.apache.hadoop.mapred.Counters log
ÐÅÏ¢:     HDFS_BYTES_READ=1535
2013-9-30 19:51:06 org.apache.hadoop.mapred.Counters log
ÐÅÏ¢:     FILE_BYTES_WRITTEN=209510
2013-9-30 19:51:06 org.apache.hadoop.mapred.Counters log
ÐÅÏ¢:     HDFS_BYTES_WRITTEN=348
2013-9-30 19:51:06 org.apache.hadoop.mapred.Counters log
ÐÅÏ¢:   Map-Reduce Framework
2013-9-30 19:51:06 org.apache.hadoop.mapred.Counters log
ÐÅÏ¢:     Map output materialized bytes=458
2013-9-30 19:51:06 org.apache.hadoop.mapred.Counters log
ÐÅÏ¢:     Map input records=11
2013-9-30 19:51:06 org.apache.hadoop.mapred.Counters log
ÐÅÏ¢:     Reduce shuffle bytes=0
2013-9-30 19:51:06 org.apache.hadoop.mapred.Counters log
ÐÅÏ¢:     Spilled Records=30
2013-9-30 19:51:06 org.apache.hadoop.mapred.Counters log
ÐÅÏ¢:     Map output bytes=509
2013-9-30 19:51:06 org.apache.hadoop.mapred.Counters log
ÐÅÏ¢:     Total committed heap usage (bytes)=1838546944
2013-9-30 19:51:06 org.apache.hadoop.mapred.Counters log
ÐÅÏ¢:     Map input bytes=421
2013-9-30 19:51:06 org.apache.hadoop.mapred.Counters log
ÐÅÏ¢:     SPLIT_RAW_BYTES=452
2013-9-30 19:51:06 org.apache.hadoop.mapred.Counters log
ÐÅÏ¢:     Combine input records=22
2013-9-30 19:51:06 org.apache.hadoop.mapred.Counters log
ÐÅÏ¢:     Reduce input records=15
2013-9-30 19:51:06 org.apache.hadoop.mapred.Counters log
ÐÅÏ¢:     Reduce input groups=13
2013-9-30 19:51:06 org.apache.hadoop.mapred.Counters log
ÐÅÏ¢:     Combine output records=15
2013-9-30 19:51:06 org.apache.hadoop.mapred.Counters log
ÐÅÏ¢:     Reduce output records=13
2013-9-30 19:51:06 org.apache.hadoop.mapred.Counters log
ÐÅÏ¢:     Map output records=22

³É¹¦ÔËÐÐÁËwordcount³ÌÐò£¬Í¨¹ýÃüÁîÎÒÃDz鿴Êä³ö½á¹û

~ hadoop fs -ls hdfs://192.168.1.210:9000/user/hdfs/o_t_account/result

Found 2 items
-rw-r--r--   3 Administrator supergroup          0 2013-09-30 19:51 /user/hdfs/o_t_account/result/_SUCCESS
-rw-r--r--   3 Administrator supergroup        348 2013-09-30 19:51 /user/hdfs/o_t_account/result/part-00000

~ hadoop fs -cat hdfs://192.168.1.210:9000/user/hdfs/o_t_account/result/part-00000

1,abc@163.com,2013-04-22        1
10,ade121@sohu.com,2013-04-23   1
11,addde@sohu.com,2013-04-23    1
17:21:24.0      5
2,dedac@163.com,2013-04-22      1
20:21:39.0      6
3,qq8fed@163.com,2013-04-22     1
4,qw1@163.com,2013-04-22        1
5,af3d@163.com,2013-04-22       1
6,ab34@163.com,2013-04-22       1
7,q8d1@gmail.com,2013-04-23     1
8,conan@gmail.com,2013-04-23    1
9,adeg@sohu.com,2013-04-23      1

ÕâÑù£¬ÎÒÃǾÍʵÏÖÁËÔÚwin7ÖеĿª·¢£¬Í¨¹ýMaven¹¹½¨HadoopÒÀÀµ»·¾³£¬ÔÚEclipseÖпª·¢MapReduceµÄ³ÌÐò£¬È»ºóÔËÐÐJavaAPP¡£HadoopÓ¦Óûá×Ô¶¯°ÑÎÒÃǵÄMR³ÌÐò´ò³Éjar°ü£¬ÔÙÉÏ´«µÄÔ¶³ÌµÄhadoop»·¾³ÖÐÔËÐУ¬·µ»ØÈÕÖ¾ÔÚEclipse¿ØÖÆÌ¨Êä³ö¡£

6. Ä£°åÏîÄ¿ÉÏ´«github

https://github.com/bsspirit/maven_hadoop_template

´ó¼Ò¿ÉÒÔÏÂÔØÕâ¸öÏîÄ¿£¬×öΪ¿ª·¢µÄÆðµã¡£

~ git clone https://github.com/bsspirit/maven_hadoop_template.git

ÎÒÃÇÍê³ÉµÚÒ»²½£¬ÏÂÃæ¾Í½«Õýʽ½øÈëMapReduce¿ª·¢Êµ¼ù¡£

   
5101 ´Îä¯ÀÀ       27
Ïà¹ØÎÄÕÂ

»ùÓÚEAµÄÊý¾Ý¿â½¨Ä£
Êý¾ÝÁ÷½¨Ä££¨EAÖ¸ÄÏ£©
¡°Êý¾Ýºþ¡±£º¸ÅÄî¡¢ÌØÕ÷¡¢¼Ü¹¹Óë°¸Àý
ÔÚÏßÉ̳ÇÊý¾Ý¿âϵͳÉè¼Æ ˼·+Ч¹û
 
Ïà¹ØÎĵµ

GreenplumÊý¾Ý¿â»ù´¡Åàѵ
MySQL5.1ÐÔÄÜÓÅ»¯·½°¸
ijµçÉÌÊý¾ÝÖÐ̨¼Ü¹¹Êµ¼ù
MySQL¸ßÀ©Õ¹¼Ü¹¹Éè¼Æ
Ïà¹Ø¿Î³Ì

Êý¾ÝÖÎÀí¡¢Êý¾Ý¼Ü¹¹¼°Êý¾Ý±ê×¼
MongoDBʵս¿Î³Ì
²¢·¢¡¢´óÈÝÁ¿¡¢¸ßÐÔÄÜÊý¾Ý¿âÉè¼ÆÓëÓÅ»¯
PostgreSQLÊý¾Ý¿âʵսÅàѵ
×îл¼Æ»®
DeepSeekÔÚÈí¼þ²âÊÔÓ¦ÓÃʵ¼ù 4-12[ÔÚÏß]
DeepSeek´óÄ£ÐÍÓ¦Óÿª·¢Êµ¼ù 4-19[ÔÚÏß]
UAF¼Ü¹¹ÌåϵÓëʵ¼ù 4-11[±±¾©]
AIÖÇÄÜ»¯Èí¼þ²âÊÔ·½·¨Óëʵ¼ù 5-23[ÉϺ£]
»ùÓÚ UML ºÍEA½øÐзÖÎöÉè¼Æ 4-26[±±¾©]
ÒµÎñ¼Ü¹¹Éè¼ÆÓ뽨ģ 4-18[±±¾©]

MySQLË÷Òý±³ºóµÄÊý¾Ý½á¹¹
MySQLÐÔÄܵ÷ÓÅÓë¼Ü¹¹Éè¼Æ
SQL ServerÊý¾Ý¿â±¸·ÝÓë»Ö¸´
ÈÃÊý¾Ý¿â·ÉÆðÀ´ 10´óDB2ÓÅ»¯
oracleµÄÁÙʱ±í¿Õ¼äдÂú´ÅÅÌ
Êý¾Ý¿âµÄ¿çƽ̨Éè¼Æ


²¢·¢¡¢´óÈÝÁ¿¡¢¸ßÐÔÄÜÊý¾Ý¿â
¸ß¼¶Êý¾Ý¿â¼Ü¹¹Éè¼ÆÊ¦
HadoopÔ­ÀíÓëʵ¼ù
Oracle Êý¾Ý²Ö¿â
Êý¾Ý²Ö¿âºÍÊý¾ÝÍÚ¾ò
OracleÊý¾Ý¿â¿ª·¢Óë¹ÜÀí


GE Çø¿éÁ´¼¼ÊõÓëʵÏÖÅàѵ
º½Ìì¿Æ¹¤Ä³×Ó¹«Ë¾ Nodejs¸ß¼¶Ó¦Óÿª·¢
ÖÐÊ¢Òæ»ª ׿Խ¹ÜÀíÕß±ØÐë¾ß±¸µÄÎåÏîÄÜÁ¦
ijÐÅÏ¢¼¼Êõ¹«Ë¾ PythonÅàѵ
ij²©²ÊITϵͳ³§ÉÌ Ò×ÓÃÐÔ²âÊÔÓëÆÀ¹À
ÖйúÓÊ´¢ÒøÐÐ ²âÊÔ³ÉÊì¶ÈÄ£Ðͼ¯³É(TMMI)
ÖÐÎïÔº ²úÆ·¾­ÀíÓë²úÆ·¹ÜÀí