±à¼ÍƼö: |
±¾ÎÄÀ´×ÔÓÚ51cto£¬Apache
Hadoop YARN £¨Yet Another Resource Negotiator£¬ÁíÒ»ÖÖ×ÊԴе÷Õߣ©ÊÇÒ»ÖÖеÄ
Hadoop ×ÊÔ´¹ÜÀíÆ÷£¬ËüÊÇÒ»¸öͨÓÃ×ÊÔ´¹ÜÀíϵͳ£¬¿ÉΪÉϲãÓ¦ÓÃÌṩͳһµÄ×ÊÔ´¹ÜÀíºÍµ÷¶È.
|
|
YARN²úÉú±³¾°
YARNÊÇHadoop2.x²ÅÓеģ¬ËùÒÔÔÚ½éÉÜYARN֮ǰ£¬ÎÒÃÇÏÈ¿´Ò»ÏÂMapReduce1.xʱËù´æÔÚµÄÎÊÌ⣺
µ¥µã¹ÊÕÏ
½ÚµãѹÁ¦´ó
²»Ò×À©Õ¹
MapReduce1.xʱµÄ¼Ü¹¹ÈçÏ£º

·Ö²¼Ê½×ÊÔ´µ÷¶È¡ª¡ªYARN¿ò¼Ü
¿ÉÒÔ¿´µ½£¬1.xʱҲÊÇMaster/SlaveÕâÖÖÖ÷´Ó½á¹¹£¬ÔÚ¼¯ÈºÉϵıíÏÖ¾ÍÊÇÒ»¸öJobTracker´ø¶à¸öTaskTracker¡£
JobTracker£º¸ºÔð×ÊÔ´¹ÜÀíºÍ×÷Òµµ÷¶È
TaskTracker£º¶¨ÆÚÏòJobTracker»ã±¨±¾½ÚµãµÄ½¡¿µ×´¿ö¡¢×ÊԴʹÓÃÇé¿öÒÔ¼°×÷ÒµÖ´ÐÐÇé¿ö¡£»¹¿ÉÒÔ½ÓÊÕÀ´×ÔJobTrackerµÄÃüÁÀýÈçÆô¶¯ÈÎÎñ»ò½áÊøÈÎÎñµÈ¡£
ÄÇôÕâÖּܹ¹´æÔÚÄÄЩÎÊÌâÄØ£º
1.Õû¸ö¼¯ÈºÖÐÖ»ÓÐÒ»¸öJobTracker£¬¾Í´ú±í×Å»á´æÔÚµ¥µã¹ÊÕϵÄÇé¿ö
2.JobTracker½ÚµãµÄѹÁ¦ºÜ´ó£¬²»½öÒª½ÓÊÕÀ´×Ô¿Í»§¶ËµÄÇëÇ󣬻¹Òª½ÓÊÕ´óÁ¿TaskTracker½ÚµãµÄÇëÇó
3.ÓÉÓÚJobTrackerÊǵ¥½Úµã£¬ËùÒÔÈÝÒ׳ÉΪ¼¯ÈºÖÐµÄÆ¿¾±£¬¶øÇÒÒ²²»Ò×ÓòÀ©Õ¹
4.JobTracker³ÐÔØµÄÖ°Ôð¹ý¶à£¬»ù±¾Õû¸ö¼¯ÈºÖеÄÊÂÇé¶¼ÊÇJobTrackerÀ´¹ÜÀí
1.x°æ±¾µÄÕû¸ö¼¯ÈºÖ»Ö§³ÖMapReduce×÷Òµ£¬ÆäËûÀýÈçSparkµÄ×÷Òµ¾Í²»Ö§³ÖÁË
ÓÉÓÚ1.x°æ±¾²»Ö§³ÖÆäËû¿ò¼ÜµÄ×÷Òµ£¬ËùÒÔµ¼ÖÂÎÒÃÇÐèÒª¸ù¾Ý²»Í¬µÄ¿ò¼ÜÈ¥´î½¨¶à¸ö¼¯Èº¡£ÕâÑù¾Í»áµ¼ÖÂ×ÊÔ´ÀûÓÃÂʱȽϵÍÒÔ¼°ÔËά³É±¾¹ý¸ß£¬ÒòΪ¶à¸ö¼¯Èº»áµ¼Ö·þÎñ»·¾³±È½Ï¸´ÔÓ¡£ÈçÏÂͼ£º

ÔÚÉÏͼÖÐÎÒÃÇ¿ÉÒÔ¿´µ½£¬²»Í¬µÄ¿ò¼ÜÎÒ²»½öÐèÒª´î½¨²»Í¬µÄ¼¯Èº¡£¶øÇÒÕâЩ¼¯ÈººÜ¶àʱºò²¢²»ÊÇ×ÜÊÇÔÚ¹¤×÷£¬ÈçÉÏͼ¿ÉÒÔ¿´µ½£¬Hadoop¼¯ÈºÔÚæµÄʱºòSpark¾Í±È½ÏÏУ¬Spark¼¯Èº±È½ÏæµÄʱºòHadoop¼¯Èº¾Í±È½ÏÏУ¬¶øMPI¼¯ÈºÔòÊÇÕûÌå²¢²»ÊǺÜæ¡£ÕâÑù¾ÍÎÞ·¨¸ßЧµÄÀûÓÃ×ÊÔ´£¬ÒòΪÕâЩ²»Í¬µÄ¼¯ÈºÎÞ·¨»¥ÏàʹÓÃ×ÊÔ´¡£³ý´ËÖ®Í⣬ÎÒÃÇ»¹µÃÔËάÕâЩ¸ö²»Í¬µÄ¼¯Èº£¬¶øÇÒÎļþϵͳÊÇÎÞ·¨¹²ÏíµÄ¡£Èç¹ûµ±ÐèÒª½«Hadoop¼¯ÈºÉϵÄHDFSÀï´æ´¢µÄÊý¾Ý´«Êäµ½Spark¼¯ÈºÉϽøÐмÆËãʱ£¬»¹»áºÄ·ÑÏ൱´óµÄÍøÂçIOÁ÷Á¿¡£
ËùÒÔÎÒÃǾÍÏë×ÅÒª°ÑÕâЩ¼¯Èº¶¼ºÏ²¢ÔÚÒ»Æð£¬ÈÃÕâЩ²»Í¬µÄ¿ò¼ÜÄܹ»ÔËÐÐÔÚͬһ¸ö¼¯ÈºÉÏ£¬ÕâÑù¾ÍÄܽâ¾öÕâ¸÷ÖÖ¸÷ÑùµÄÎÊÌâÁË¡£ÈçÏÂͼ£º

ÕýÊÇÒòΪÔÚ1.xÖУ¬Óи÷ÖÖ¸÷ÑùµÄÎÊÌ⣬²ÅʹµÃYARNµÃÒÔµ®Éú£¬¶øYARN¾Í¿ÉÒÔÁîÕâЩ²»Í¬µÄ¿ò¼ÜÔËÐÐÔÚͬһ¸ö¼¯ÈºÉÏ£¬²¢ÎªËüÃǵ÷¶È×ÊÔ´¡£ÎÒÃÇÀ´¿´¿´Hadoop2.xµÄ¼Ü¹¹Í¼£º

ÔÚÉÏͼÖУ¬ÎÒÃÇ¿ÉÒÔ¿´µ½£¬¼¯Èº×îµ×²ãµÄÊÇHDFS£¬ÔÚÆäÖ®ÉϵľÍÊÇYARN²ã£¬¶øÔÚYARN²ãÉÏÔòÊǸ÷ÖÖ²»Í¬µÄ¼ÆËã¿ò¼Ü¡£ËùÒÔ²»Í¬¼ÆËã¿ò¼Ü¿ÉÒÔ¹²Ïíͬһ¸öHDFS¼¯ÈºÉϵÄÊý¾Ý£¬ÏíÊÜÕûÌåµÄ×ÊÔ´µ÷¶È£¬½ø¶øÌá¸ß¼¯Èº×ÊÔ´µÄÀûÓÃÂÊ£¬ÕâÒ²¾ÍÊÇËùνµÄ
xxx on YARN¡£
YARN¼Ü¹¹
YARN¸ÅÊö£º
YARNÊÇ×ÊÔ´µ÷¶È¿ò¼Ü
ͨÓõÄ×ÊÔ´¹ÜÀíϵͳ
ΪÉϲãÓ¦ÓÃÌṩͳһµÄ×ÊÔ´¹ÜÀíºÍµ÷¶È
YARN¼Ü¹¹Í¼£¬Ò²ÊÇMaster/Slave½á¹¹µÄ£º

´ÓÉÏͼÖУ¬ÎÒÃÇ¿ÉÒÔ¿´µ½YARNÖ÷ÒªÓÉÒÔϼ¸¸öºËÐÄ×é¼þ¹¹³É£º
1. ResourceManager£¬ ¼ò³ÆRM£¬Õû¸ö¼¯ÈºÍ¬Ò»Ê±¼äÌṩ·þÎñµÄRMÖ»ÓÐÒ»¸ö£¬Ëü¸ºÔð¼¯Èº×ÊÔ´µÄͳһ¹ÜÀíºÍµ÷¶È¡£ÒÔ¼°»¹ÐèÒª´¦Àí¿Í»§¶ËµÄÇëÇó£¬ÀýÈ磺Ìá½»×÷Òµ»ò½áÊø×÷ÒµµÈ¡£²¢ÇÒ¼à¿Ø¼¯ÈºÖеÄNM£¬Ò»µ©Ä³¸öNM¹ÒÁË£¬ÄÇô¾ÍÐèÒª½«¸ÃNMÉÏÔËÐеÄÈÎÎñ¸æËßAMÀ´ÈçºÎ½øÐд¦Àí¡£
2. NodeManager£¬ ¼ò³ÆNM£¬Õû¸ö¼¯ÈºÖлáÓжà¸öNM£¬ËüÖ÷Òª¸ºÔð×Ô¼º±¾Éí½ÚµãµÄ×ÊÔ´¹ÜÀíºÍʹÓã¬ÒÔ¼°¶¨Ê±ÏòRM»ã±¨±¾½ÚµãµÄ×ÊԴʹÓÃÇé¿ö¡£½ÓÊÕ²¢´¦ÀíÀ´×ÔRMµÄ¸÷ÖÖÃüÁÀýÈ磺Æô¶¯Container¡£NM»¹ÐèÒª´¦ÀíÀ´×ÔAMµÄÃüÁÀýÈ磺AM»á¸æËßNMÐèÒªÆô¶¯¶àÉÙ¸öContainerÀ´ÅÜtask¡£
3. ApplicationMaster£¬ ¼ò³ÆAM£¬Ã¿¸öÓ¦ÓóÌÐò¶¼¶ÔÓ¦×ÅÒ»¸öAM¡£ÀýÈ磺MapReduce»á¶ÔÓ¦Ò»¸ö¡¢Spark»á¶ÔÓ¦Ò»¸ö¡£ËüÖ÷Òª¸ºÔðÓ¦ÓóÌÐòµÄ¹ÜÀí£¬ÎªÓ¦ÓóÌÐòÏòRMÉêÇë×ÊÔ´£¨Core¡¢Memory£©£¬½«×ÊÔ´·ÖÅ䏸ÄÚ²¿µÄtask¡£AMÐèÒªÓëNMͨÐÅ£¬ÒÔ´ËÀ´Æô¶¯»òÍ£Ö¹task¡£taskÊÇÔËÐÐÔÚContainerÀïÃæµÄ£¬ËùÒÔAMÒ²ÊÇÔËÐÐÔÚContainerÀïÃæ¡£
4. Container£¬ ·â×°ÁËCPU¡¢MemoryµÈ×ÊÔ´µÄÒ»¸öÈÝÆ÷£¬Ï൱ÓÚÊÇÒ»¸öÈÎÎñÔËÐл·¾³µÄ³éÏó¡£
5. Client£¬ ¿Í»§¶Ë£¬Ëü¿ÉÒÔÌá½»×÷Òµ¡¢²éѯ×÷ÒµµÄÔËÐнø¶ÈÒÔ¼°½áÊø×÷Òµ¡£
YARN¹Ù·½ÎĵµµØÖ·
YARNÖ´ÐÐÁ÷³Ì
¼ÙÉè¿Í»§¶ËÏòResourceManagerÌá½»Ò»¸ö×÷Òµ£¬ResourceManagerÔò»áΪÕâ¸ö×÷Òµ·ÖÅäÒ»¸öContainer¡£ËùÒÔResourceManager»áÓëNodeManager½øÐÐͨÐÅ£¬ÒªÇóÕâ¸öNodeManagerÆô¶¯Ò»¸öContainer¡£¶øÕâ¸öContainerÊÇÓÃÀ´Æô¶¯ApplicationMasterµÄ£¬ApplicationMasterÆô¶¯ÍêÖ®ºó»áÓëResourceManager½øÐÐÒ»¸ö×¢²á¡£Õâʱºò¿Í»§¶Ë¾Í¿ÉÒÔͨ¹ýResourceManager²éѯ×÷ÒµµÄÔËÐÐÇé¿öÁË¡£È»ºóApplicationMaster»¹»áµ½ResourceManagerÉÏÉêÇë×÷ÒµËùÐèÒªµÄ×ÊÔ´£¬ÉêÇëµ½ÒÔºó¾Í»áµ½¶ÔÓ¦µÄNodeManagerÖ®ÉÏÔËÐпͻ§¶ËËùÌá½»µÄ×÷Òµ£¬È»ºóNodeManager¾Í»á°ÑtaskÔËÐÐÔÚÆô¶¯µÄContainerÀï¡£
ÈçÏÂͼ£º

YARN»·¾³´î½¨
½éÉÜÍê»ù±¾µÄÀíÂÛ²¿·ÖÖ®ºó£¬ÎÒÃÇÀ´´î½¨Ò»¸öα·Ö²¼Ê½µÄµ¥½ÚµãYARN»·¾³£¬Ê¹ÓõÄhadoop°æ±¾ÈçÏ£º
hadoop-2.6.0-cdh5.7.0
¹Ù·½µÄ°²×°ÎĵµµØÖ·
1.ÏÂÔØ²¢½âѹºÃhadoop-2.6.0-cdh5.7.0£¬ÕâÒ»²½¿ÉÒԲο¼ÎÒ֮ǰдµÄһƪ¹ØÓÚHDFSα·Ö²¼Ê½»·¾³´î½¨µÄÎÄÕ£¬ÎÒÕâÀï¾Í²»ÔÙ׸ÊöÁË¡£
È·±£HDFSÊÇÕý³£Æô¶¯×´Ì¬£º
[root@localhost
~]# jps
3827 Jps
3383 NameNode
3500 DataNode
3709 SecondaryNameNode
[root@localhost ~]# |
2.±à¼mapred-site.xmlÅäÖÃÎļþ£¬ÔÚÎļþÖÐÔö¼ÓÈçÏÂÄÚÈÝ£º
[root@localhost
~] # cd /usr/local/hadoop-2.6.0 -cdh5.7.0/etc/hadoop
[root@localhost /usr/local/hadoop -2.6.0-cdh5.7.0/etc/hadoop] #
cp mapred-site.xml.template mapred-site.xml #
¿½±´Ä£°åÎļþ
[root@localhost /usr/local/hadoop -2.6.0-cdh5.7.0/etc/hadoop]
# vim mapred-site.xml # Ôö¼ÓÈçÏÂÄÚÈÝ
<property> <name>mapreduce.framework.name</name>
<value>yarn</value>
</property> |
3.±à¼yarn-site.xmlÅäÖÃÎļþ£¬ÔÚÎļþÖÐÔö¼ÓÈçÏÂÄÚÈÝ£º
[root@localhost
/usr/local/hadoop-2.6.0 -cdh5.7.0/etc/hadoop]
# vim yarn-site.xml # Ôö¼ÓÈçÏÂÄÚÈÝ
<property>
<name>yarn.nodemanager.aux-services </name>
<value>mapreduce_shuffle</value>
</property> |
4.Æô¶¯ResourceManager½ø³ÌÒÔ¼°NodeManager½ø³Ì£º
[root@localhost
/usr/local/hadoop-2.6.0 -cdh5.7.0/etc/hadoop]#
cd ../../sbin/
[root@localhost /usr/local/hadoop-2.6.0 -cdh5.7.0/sbin]#
./start-yarn.sh
starting yarn daemons
starting resourcemanager, logging to /usr/local/hadoop-2.6.0-cdh5.7.0/logs/ yarn-root-resourcemanager-localhost.out
localhost: starting nodemanager, logging to /usr/local/hadoop-2.6.0-cdh5.7.0/logs/yarn -root-nodemanager-localhost.out
[root@localhost /usr/local/hadoop -2.6.0-cdh5.7.0/sbin]#
jps
3984 NodeManager # Æô¶¯³É¹¦ºó¿ÉÒÔ¿´µ½¶à³öÁËNodeManager
4947 DataNode
5252 Jps
5126 SecondaryNameNode
3884 ResourceManager # ºÍResourceManager½ø³Ì£¬ÕâÑù²ÅÊÇÕý³£µÄ¡£
4813 NameNode
[root@localhost /usr/local/hadoop-2.6.0 -cdh5.7.0/sbin]#
netstat -lntp |grep java
tcp 0 0 0.0.0.0:50090 0.0.0.0:* LISTEN 5126/java
tcp 0 0 127.0.0.1:42602
0.0.0.0:*
LISTEN 4947/java
tcp 0 0 192.168.77.130:8020 0.0.0.0:* LISTEN 4813/java
tcp 0 0 0.0.0.0:50070 0.0.0.0:* LISTEN 4813/java
tcp 0 0 0.0.0.0:50010 0.0.0.0:* LISTEN 4947/java
tcp 0 0 0.0.0.0:50075 0.0.0.0:* LISTEN 4947/java
tcp 0 0 0.0.0.0:50020 0.0.0.0:* LISTEN 4947/java
tcp6 0 0 :::8040
:::*
LISTEN 5566/java
tcp6 0 0 :::8042
:::*
LISTEN
5566/java
tcp6 0 0 :::8088 :::*
LISTEN
5457/java
tcp6 0 0 :::13562
:::*
LISTEN 5566/java
tcp6 0 0 :::8030 :::*
LISTEN
5457/java
tcp6 0 0 :::8031
;:::*
LISTEN 5457/java
tcp6 0 0 :::8032 :::*
LISTEN
5457/java
tcp6 0 0 :::48929 :::*
LISTEN 5566/java
tcp6 0 0 :::8033 :::*
LISTEN
5457/java
[root@localhost /usr/local/hadoop-2.6.0-cdh5.7.0/sbin]#
|
5.ͨ¹ýä¯ÀÀÆ÷À´·ÃÎÊResourceManager£¬Ä¬È϶˿ÚÊÇ8088£¬ÀýÈç192.168.77.130:8088£¬¾Í»á·ÃÎʵ½ÕâÑùµÄÒ»¸öÒ³ÃæÉÏ£º

´íÎó½â¾ö£º
´ÓÉÏͼÖУ¬¿ÉÒÔ¿´µ½ÓÐÒ»¸ö²»½¡¿µµÄ½Úµã£¬Ò²¾ÍÊÇ˵ÎÒÃǵĵ¥½Úµã»·¾³ÓÐÎÊÌ⣬µã»÷ºìÉ«¿ò¿òÖбê¼ÇµÄÊý×Ö¿ÉÒÔ½øÈëµ½ÏêϸµÄÐÅÏ¢Ò³Ãæ£¬ÔÚ¸ÃÒ³ÃæÖп´µ½ÁËÈçÏÂÐÅÏ¢£º

ÓÚÊDz鿴yarnµÄÈÕÖ¾Îļþ£ºyarn-root-nodemanager-localhost.log£¬·¢ÏÖÈçϾ¯¸æÓëÒì³££º

ºÜÃ÷ÏÔÊÇÒòΪ´ÅÅ̵ÄʹÓÿռä´ïµ½ÁË90%£¬ËùÒÔÎÒÃÇÐèҪɾ³ýһЩûÓеÄÊý¾Ý£¬»òÕßÀ©ÈÝ´ÅÅ̿ռä²ÅÐС£ÓÚÊÇɾ³ýÁËÒ»¶Ñ°²×°°ü£¬ÈôÅÅ̿ռ併µÍµ½90%ÒÔÏÂÁË£º

ÕâʱÔÙ´ÎË¢ÐÂÒ³Ãæ£¬¿ÉÒÔ·¢ÏÖÕâ¸ö½Úµã¾ÍÕý³£ÁË£º

µ½´ËΪֹ£¬ÎÒÃǵÄyarn»·¾³¾Í´î½¨Íê³ÉÁË¡£
Èç¹ûÐèÒª¹Ø±Õ½ø³ÌÔòʹÓÃÒÔÏÂÃüÁ
[root@localhost
/usr/local/hadoop -2.6.0-cdh5.7.0/sbin] # stop-yarn.sh |
³õʶÌá½»PIµÄMapReduce×÷Òµµ½YARNÉÏÖ´ÐÐ
ËäÈ»ÎÒÃÇûÓдMapReduceµÄ»·¾³£¬µ«ÊÇÎÒÃÇ¿ÉÒÔʹÓÃHadoop×Ô´øµÄһЩ²âÊÔÀý×ÓÀ´ÑÝʾһÏÂÈçºÎÌá½»×÷Òµµ½YARNÉÏÖ´ÐС£Hadoop°ÑexampleµÄ°ü·ÅÔÚÁËÈçÏ·¾¶£¬¿ÉÒÔ¿´µ½Óкü¸¸öjar°ü£º
[root@localhost
~]# cd /usr/local/hadoop-2.6.0 -cdh5.7.0/share/hadoop/mapreduce/
[root@localhost /usr/local/hadoop -2.6.0-cdh5.7.0/share/hadoop/mapreduce]#
ls
hadoop-mapreduce -client-app-2.6.0-cdh5.7.0.jar
hadoop-mapreduce -client-common-2.6.0-cdh5.7.0.jar
hadoop-mapreduce -client-core-2.6.0-cdh5.7.0.jar
hadoop-mapreduce -client-hs-2.6.0-cdh5.7.0.jar
hadoop-mapreduce -client-hs-plugins-2.6.0-cdh5.7.0.jar
hadoop-mapreduce -client-jobclient-2.6.0-cdh5.7.0.jar
hadoop-mapreduce -client-jobclient-2.6.0-cdh5.7.0-tests.jar
hadoop-mapreduce -client-nativetask-2.6.0-cdh5.7.0.jar
hadoop-mapreduce -client-shuffle-2.6.0-cdh5.7.0.jar
hadoop-mapreduce -examples-2.6.0-cdh5.7.0.jar
lib
lib-examples
sources
[root@localhost /usr/local/hadoop -2.6.0-cdh5.7.0/share/hadoop/mapreduce]#
|
ÔÚÕâÀïÎÒÃÇʹÓÃhadoop-mapreduce-examples-2.6.0-cdh5.7.0.jarÕâ¸öjar°üÀ´½øÐÐÑÝʾ£º
[root@localhost
/usr/local/hadoop-2.6.0 -cdh5.7.0/share/hadoop/mapreduce]#
hadoop jar hadoop -mapreduce -examples-2.6.0-cdh5.7.0.jar
pi 2 3 |
ÃüÁî˵Ã÷£º
hadoop jar Ö´ÐÐÒ»¸öjar°ü×÷ÒµµÄÃüÁî
hadoop-mapreduce-examples-2.6.0-cdh5.7.0.jar ÐèÒª±»Ö´ÐеÄjar°ü·¾¶
pi ±íʾ¼ÆËãÔ²ÖÜÂÊ£¬¿ÉÒÔдÆäËûµÄ
ĩβµÄÁ½¸öÊý¾Ý·Ö±ð±íʾָ¶¨ÔËÐÐ2´Îmap£¬ ÒÔ¼°Ö¸¶¨Ã¿¸ömapÈÎÎñÈ¡Ñù3´Î£¬Á½ÊýÏà³Ë¼´Îª×ܵÄÈ¡ÑùÊý¡£
ÔËÐÐÒÔÉÏÃüÁîºó£¬µ½ä¯ÀÀÆ÷Ò³ÃæÉϽøÐв鿴£¬»áÓÐÒÔÏÂÈý¸ö½×¶Î£º
1.½ÓÊÕ×ÊÔ´£¬Õâ¸ö½×¶Î¾ÍÊÇApplicationMasterµ½ResourceManagerÉÏÉêÇë×÷ÒµËùÐèÒªµÄ×ÊÔ´£º

2.ÔËÐÐ×÷Òµ£¬ÕâʱºòNodeManager¾Í»á°ÑtaskÔËÐÐÔÚÆô¶¯µÄContainerÀ

3.×÷ÒµÍê³É£º

ÖÕ¶ËÊä³öÐÅÏ¢ÈçÏ£º
[root@localhost
/usr/local/hadoop-2.6.0 -cdh5.7.0/share/hadoop/mapreduce] #
hadoop jar hadoop-mapreduce -examples-2.6.0-cdh5.7.0.jar
pi 2 3
Number of Maps = 2
Samples per Map = 3
Wrote input for Map #0
Wrote input for Map #1
Starting Job
18/03/27 23:00:01 INFO client.RMProxy: Connecting
to ResourceManager at /0.0.0.0:8032
18/03/27 23:00:01 INFO input.FileInputFormat:
Total input paths to process : 2
18/03/27 23:00:01 INFO mapreduce.JobSubmitter:
number of splits:2
18/03/27 23:00:02 INFO mapreduce.JobSubmitter:
Submitting tokens for job: job_1522162696272_0001
18/03/27 23:00:02 INFO impl.YarnClientImpl: Submitted
application application_1522162696272_0001
18/03/27 23:00:02 INFO mapreduce.Job: The url
to track the job: http://localhost:8088/proxy/application _1522162696272_0001/
18/03/27 23:00:02 INFO mapreduce.Job: Running
job: job_1522162696272_0001
18/03/27 23:00:10 INFO mapreduce.Job: Job job_1522162696272_0001
running in uber mode : false
18/03/27 23:00:10 INFO mapreduce.Job: map 0% reduce
0%
18/03/27 23:00:15 INFO mapreduce.Job: map 50%
reduce 0%
18/03/27 23:00:16 INFO mapreduce.Job: map 100%
reduce 0%
18/03/27 23:00:19 INFO mapreduce.Job: map 100%
reduce 100%
18/03/27 23:00:20 INFO mapreduce.Job: Job job_1522162696272_0001
completed successfully
18/03/27 23:00:20 INFO mapreduce.Job: Counters:
49
File System Counters
FILE: Number of bytes read=50
FILE: Number of bytes written=335298
FILE: Number of read operations=0
FILE: Number of large read operations=0
FILE: Number of write operations=0
HDFS: Number of bytes read=536
HDFS: Number of bytes written=215
HDFS: Number of read operations=11
HDFS: Number of large read operations=0
HDFS: Number of write operations=3
Job Counters
Launched map tasks=2
Launched reduce tasks=1
Data-local map tasks=2
Total time spent by all maps in occupied slots
(ms)=7108
Total time spent by all reduces in occupied slots
(ms)=2066
Total time spent by all map tasks (ms)=7108
Total time spent by all reduce tasks (ms)=2066
Total vcore-seconds taken by all map tasks=7108
Total vcore-seconds taken by all reduce tasks=2066
Total megabyte-seconds taken by all map tasks=7278592
Total megabyte-seconds taken by all reduce tasks=2115584
Map-Reduce Framework
Map input records=2
Map output records=4
Map output bytes=36
Map output materialized bytes=56
Input split bytes=300
Combine input records=0
Combine output records=0
Reduce input groups=2
Reduce shuffle bytes=56
Reduce input records=4
Reduce output records=0
Spilled Records=8
Shuffled Maps =2
Failed Shuffles=0
Merged Map outputs=2
GC time elapsed (ms)=172
CPU time spent (ms)=2990
Physical memory (bytes) snapshot=803618816
Virtual memory (bytes) snapshot=8354324480
Total committed heap usage (bytes)=760217600
Shuffle Errors
BAD_ID=0
CONNECTION=0
IO_ERROR=0
WRONG_LENGTH=0
WRONG_MAP=0
WRONG_REDUCE=0
File Input Format Counters
Bytes Read=236
File Output Format Counters
Bytes Written=97
Job Finished in 19.96 seconds
Estimated value of Pi is 4.00000000000000000000
[root@localhost /usr/local/hadoop -2.6.0-cdh5.7.0/share/hadoop/mapreduce]# |
ÒÔÉÏÕâ¸öÀý×Ó¼ÆËãÁËÒ»¸öPIÖµ£¬ÏÂÃæÎÒÃÇÔÙÀ´ÑÝʾһ¸öhadoopÖбȽϾµäµÄÀý×Ó£ºwordcount
£¬ÕâÊÇÒ»¸ö¾µäµÄ´ÊƵͳ¼ÆµÄÀý×Ó¡£Ê×ÏÈ´´½¨ºÃÓÃÓÚ²âÊÔµÄÎļþ£º
[root@localhost
~]# mkdir /tmp/input
[root@localhost ~]# cd /tmp/input/
[root@localhost /tmp/input] # echo "hello
word" > file1.txt
[root@localhost /tmp/input] # echo "hello
hadoop" > file2.txt
[root@localhost /tmp/input] # echo "hello
mapreduce" >> file2.txt
[root@localhost /tmp/input] # hdfs dfs -mkdir /wc_input
[root@localhost /tmp/input] # hdfs dfs -put ./file*
/wc_input
[root@localhost /tmp/input] # hdfs dfs -ls /wc_input
Found 2 items
-rw-r--r-- 1 root supergroup
11 2018-03-27 23:11 /wc_input/file1.txt
-rw-r--r-- 1 root supergroup 29
2018-03-27 23:11 /wc_input/file2.txt
[root@localhost /tmp/input]# |
È»ºóÖ´ÐÐÒÔÏÂÃüÁ
[root@localhost
/tmp/input] # cd /usr/local/hadoop-2.6.0-cdh5.7.0/share/hadoop/mapreduce
[root@localhost /usr/local/hadoop -2.6.0-cdh5.7.0/share/hadoop/mapreduce]#
hadoop jar ./hadoop-mapreduce -examples-2.6.0-cdh5.7.0.jar
wordcount /wc_input /wc_output |
ÔÚyarnÒ³ÃæÉÏÏÔʾµÄ½×¶ÎÐÅÏ¢£º



ÖÕ¶ËÊä³öÐÅÏ¢ÈçÏ£º
[root@localhost
/usr/local/hadoop-2.6.0 -cdh5.7.0/share/hadoop/mapreduce] #
hadoop jar ./hadoop-mapreduce -examples-2.6.0-cdh5.7.0.jar
wordcount /wc_input /wc_output
18/03/27 23:12:54 INFO client.RMProxy: Connecting
to ResourceManager at /0.0.0.0:8032
18/03/27 23:12:55 INFO input.FileInputFormat:
Total input paths to process : 2
18/03/27 23:12:55 INFO mapreduce.JobSubmitter:
number of splits:2
18/03/27 23:12:55 INFO mapreduce.JobSubmitter:
Submitting tokens for job: job_1522162696272_0002
18/03/27 23:12:56 INFO impl.YarnClientImpl: Submitted
application application_1522162696272_0002
18/03/27 23:12:56 INFO mapreduce.Job: The url
to track the job: http://localhost:8088/proxy/application_1522162696272_0002/
18/03/27 23:12:56 INFO mapreduce.Job: Running
job: job_1522162696272_0002
18/03/27 23:13:02 INFO mapreduce.Job: Job job_1522162696272_0002
running in uber mode : false
18/03/27 23:13:02 INFO mapreduce.Job: map 0% reduce
0%
18/03/27 23:13:06 INFO mapreduce.Job: map 50%
reduce 0%
18/03/27 23:13:07 INFO mapreduce.Job: map 100%
reduce 0%
18/03/27 23:13:11 INFO mapreduce.Job: map 100%
reduce 100%
18/03/27 23:13:12 INFO mapreduce.Job: Job job_1522162696272_0002
completed successfully
18/03/27 23:13:12 INFO mapreduce.Job: Counters:
49
File System Counters
FILE: Number of bytes read=70
FILE: Number of bytes written=334375
FILE: Number of read operations=0
FILE: Number of large read operations=0
FILE: Number of write operations=0
HDFS: Number of bytes read=260
HDFS: Number of bytes written=36
HDFS: Number of read operations=9
HDFS: Number of large read operations=0
HDFS: Number of write operations=2
Job Counters
Launched map tasks=2
Launched reduce tasks=1
Data-local map tasks=2
Total time spent by all maps in occupied slots
(ms)=5822
Total time spent by all reduces in occupied slots
(ms)=1992
Total time spent by all map tasks (ms)=5822
Total time spent by all reduce tasks (ms)=1992
Total vcore-seconds taken by all map tasks=5822
Total vcore-seconds taken by all reduce tasks=1992
Total megabyte-seconds taken by all map tasks=5961728
Total megabyte-seconds taken by all reduce tasks=2039808
Map-Reduce Framework
Map input records=3
Map output records=6
Map output bytes=64
Map output materialized bytes=76
Input split bytes=220
Combine input records=6
Combine output records=5
Reduce input groups=4
Reduce shuffle bytes=76
Reduce input records=5
Reduce output records=4
Spilled Records=10
Shuffled Maps =2
Failed Shuffles=0
Merged Map outputs=2
GC time elapsed (ms)=157
CPU time spent (ms)=2290
Physical memory (bytes) snapshot=800239616
Virtual memory (bytes) snapshot=8352272384
Total committed heap usage (bytes)=762314752
Shuffle Errors
BAD_ID=0
CONNECTION=0
IO_ERROR=0
WRONG_LENGTH=0
WRONG_MAP=0
WRONG_REDUCE=0
File Input Format Counters
Bytes Read=40
File Output Format Counters
Bytes Written=36
[root@localhost /usr/local/hadoop -2.6.0-cdh5.7.0/share/hadoop/mapreduce] #
|
²é¿´Êä³öµÄ½á¹ûÎļþ£º
[root@localhost
/usr/local/hadoop -2.6.0- cdh5.7.0/share/hadoop/mapreduce]#
hdfs dfs -ls /wc_output
Found 2 items
-rw-r--r-- 1 root supergroup 0 2018-03-27 23:13
/wc_output/_SUCCESS
-rw-r--r-- 1 root supergroup
36 2018-03-27 23:13
/wc_output/part-r-00000
[root@localhost /usr/local/hadoop-2.6.0 -cdh5.7.0/share/hadoop/mapreduce]#
hdfs dfs -cat /wc_output/part -r-00000 # ʵ¼ÊÊä³ö½á¹ûÔÚpart-r-00000ÖÐ
hadoop 1
hello 3
mapreduce 1
word 1
[root@localhost /usr/local/hadoop -2.6.0-cdh5.7.0/share/hadoop/mapreduce]# |
ΪÁ˸üºÃµÄ²é¿´´úÂëÇë²é¿´µçÄÔ°æ |