日韩性视频-久久久蜜桃-www中文字幕-在线中文字幕av-亚洲欧美一区二区三区四区-撸久久-香蕉视频一区-久久无码精品丰满人妻-国产高潮av-激情福利社-日韩av网址大全-国产精品久久999-日本五十路在线-性欧美在线-久久99精品波多结衣一区-男女午夜免费视频-黑人极品ⅴideos精品欧美棵-人人妻人人澡人人爽精品欧美一区-日韩一区在线看-欧美a级在线免费观看

歡迎訪問 生活随笔!

生活随笔

當前位置: 首頁 > 编程资源 > 编程问答 >内容正文

编程问答

Mahout分步式程序开发 聚类Kmeans

發布時間:2025/3/21 编程问答 24 豆豆
生活随笔 收集整理的這篇文章主要介紹了 Mahout分步式程序开发 聚类Kmeans 小編覺得挺不錯的,現在分享給大家,幫大家做個參考.

Hadoop家族系列文章,主要介紹Hadoop家族產品,常用的項目包括Hadoop, Hive, Pig, HBase, Sqoop, Mahout, Zookeeper, Avro, Ambari, Chukwa,新增加的項目包括,YARN, Hcatalog, Oozie, Cassandra, Hama, Whirr, Flume, Bigtop, Crunch, Hue等。

從2011年開始,中國進入大數據風起云涌的時代,以Hadoop為代表的家族軟件,占據了大數據處理的廣闊地盤。開源界及廠商,所有數據軟件,無一不向Hadoop靠攏。Hadoop也從小眾的高富帥領域,變成了大數據開發的標準。在Hadoop原有技術基礎之上,出現了Hadoop家族產品,通過“大數據”概念不斷創新,推出科技進步。

作為IT界的開發人員,我們也要跟上節奏,抓住機遇,跟著Hadoop一起雄起!

關于作者:

  • 張丹(Conan), 程序員Java,R,PHP,Javascript
  • weibo:@Conan_Z
  • blog:?http://blog.fens.me
  • email: bsspirit@gmail.com

轉載請注明出處:
http://blog.fens.me/hadoop-mahout-kmeans/

前言

Mahout是基于Hadoop用于機器學習的程序開發框架,Mahout封裝了3大類的機器學習算法,其中包括聚類算法。kmeans是我們經常會提到用到的聚類算法之一,特別處理未知數據集的時,都會先聚類一下,看看數據集會有一些什么樣的規則。

本文主要講解,基于Mahout程序開發,實現分步式的kmeans算法。

目錄

  • 聚類算法kmeans
  • Mahout開發環境介紹
  • 用Mahout實現聚類算法kmeans
  • 用R語言可視化結果
  • 模板項目上傳github
  • 1. 聚類算法kmeans

    聚類分析是數據挖掘及機器學習領域內的重點問題之一,在數據挖掘、模式識別、決策支持、機器學習及圖像分割等領域有廣泛的應用,是最重要的數據分析方法之一。聚類是在給定的數據集合中尋找同類的數據子集合,每一個子集合形成一個類簇,同類簇中的數據具有更大的相似性。聚類算法大體上可分為基于劃分的方法、基于層次的方法、基于密度的方法、基于網格的方法以及基于模型的方法。

    k-means algorithm算法是一種得到最廣泛使用的基于劃分的聚類算法,把n個對象分為k個簇,以使簇內具有較高的相似度。相似度的計算根據一個簇中對象的平均值來進行。它與處理混合正態分布的最大期望算法很相似,因為他們都試圖找到數據中自然聚類的中心。

    算法首先隨機地選擇k個對象,每個對象初始地代表了一個簇的平均值或中心。對剩余的每個對象根據其與各個簇中心的距離,將它賦給最近的簇,然后重新計算每個簇的平均值。這個過程不斷重復,直到準則函數收斂。

    kmeans介紹摘自:http://zh.wikipedia.org/wiki/K平均算法

    2. Mahout開發環境介紹

    接上一篇文章:Mahout分步式程序開發 基于物品的協同過濾ItemCF

    所有環境變量 和 系統配置 與上文一致!

    3. 用Mahout實現聚類算法kmeans

    實現步驟:

    • 1. 準備數據文件: randomData.csv
    • 2. Java程序:KmeansHadoop.java
    • 3. 運行程序
    • 4. 聚類結果解讀
    • 5. HDFS產生的目錄

    1). 準備數據文件: randomData.csv
    數據文件randomData.csv,由R語言通過“隨機正太分布函數”程序生成,單機內存實驗請參考文章:
    用Maven構建Mahout項目

    原始數據文件:這里只截取了一部分數據。

    ~ vi datafile/randomData.csv-0.883033363823402 -3.31967192630249 -2.39312626419456 3.34726861118871 2.66976353341256 1.85144276077058 -1.09922906899594 -6.06261735207489 -4.36361936997216 1.90509905380532 -0.00351835125495037 -0.610105996559153 -2.9962958796338 -3.60959839525735 -3.27529418132066 0.0230099799641799 2.17665594420569 6.77290756817957 -2.47862038335637 2.53431833167278 5.53654901906814 2.65089785582474 5.66257474538338 6.86783609641077 -0.558946883114376 1.22332819416237 5.11728525486132 3.74663871584768 1.91240516693351 2.95874731384062 -2.49747101306535 2.05006504756875 3.98781883213459 1.00780938946366 5.47470532716682 5.35084411045171

    注:由于Mahout中kmeans算法,默認的分融符是” “(空格),因些我把逗號分隔的數據文件,改成以空格分隔。

    2). Java程序:KmeansHadoop.java

    kmeans的算法實現,請查看Mahout in Action。

    package org.conan.mymahout.cluster08;import org.apache.hadoop.fs.Path; import org.apache.hadoop.mapred.JobConf; import org.apache.mahout.clustering.conversion.InputDriver; import org.apache.mahout.clustering.kmeans.KMeansDriver; import org.apache.mahout.clustering.kmeans.RandomSeedGenerator; import org.apache.mahout.common.distance.DistanceMeasure; import org.apache.mahout.common.distance.EuclideanDistanceMeasure; import org.apache.mahout.utils.clustering.ClusterDumper; import org.conan.mymahout.hdfs.HdfsDAO; import org.conan.mymahout.recommendation.ItemCFHadoop;public class KmeansHadoop {private static final String HDFS = "hdfs://192.168.1.210:9000";public static void main(String[] args) throws Exception {String localFile = "datafile/randomData.csv";String inPath = HDFS + "/user/hdfs/mix_data";String seqFile = inPath + "/seqfile";String seeds = inPath + "/seeds";String outPath = inPath + "/result/";String clusteredPoints = outPath + "/clusteredPoints";JobConf conf = config();HdfsDAO hdfs = new HdfsDAO(HDFS, conf);hdfs.rmr(inPath);hdfs.mkdirs(inPath);hdfs.copyFile(localFile, inPath);hdfs.ls(inPath);InputDriver.runJob(new Path(inPath), new Path(seqFile), "org.apache.mahout.math.RandomAccessSparseVector");int k = 3;Path seqFilePath = new Path(seqFile);Path clustersSeeds = new Path(seeds);DistanceMeasure measure = new EuclideanDistanceMeasure();clustersSeeds = RandomSeedGenerator.buildRandom(conf, seqFilePath, clustersSeeds, k, measure);KMeansDriver.run(conf, seqFilePath, clustersSeeds, new Path(outPath), measure, 0.01, 10, true, 0.01, false);Path outGlobPath = new Path(outPath, "clusters-*-final");Path clusteredPointsPath = new Path(clusteredPoints);System.out.printf("Dumping out clusters from clusters: %s and clusteredPoints: %s\n", outGlobPath, clusteredPointsPath);ClusterDumper clusterDumper = new ClusterDumper(outGlobPath, clusteredPointsPath);clusterDumper.printClusters(null);}public static JobConf config() {JobConf conf = new JobConf(ItemCFHadoop.class);conf.setJobName("ItemCFHadoop");conf.addResource("classpath:/hadoop/core-site.xml");conf.addResource("classpath:/hadoop/hdfs-site.xml");conf.addResource("classpath:/hadoop/mapred-site.xml");return conf;}}

    3). 運行程序
    控制臺輸出:

    Delete: hdfs://192.168.1.210:9000/user/hdfs/mix_data Create: hdfs://192.168.1.210:9000/user/hdfs/mix_data copy from: datafile/randomData.csv to hdfs://192.168.1.210:9000/user/hdfs/mix_data ls: hdfs://192.168.1.210:9000/user/hdfs/mix_data ========================================================== name: hdfs://192.168.1.210:9000/user/hdfs/mix_data/randomData.csv, folder: false, size: 36655 ========================================================== SLF4J: Failed to load class "org.slf4j.impl.StaticLoggerBinder". SLF4J: Defaulting to no-operation (NOP) logger implementation SLF4J: See http://www.slf4j.org/codes.html#StaticLoggerBinder for further details. 2013-10-14 15:39:31 org.apache.hadoop.util.NativeCodeLoader 警告: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable 2013-10-14 15:39:31 org.apache.hadoop.mapred.JobClient copyAndConfigureFiles 警告: Use GenericOptionsParser for parsing the arguments. Applications should implement Tool for the same. 2013-10-14 15:39:31 org.apache.hadoop.mapreduce.lib.input.FileInputFormat listStatus 信息: Total input paths to process : 1 2013-10-14 15:39:31 org.apache.hadoop.io.compress.snappy.LoadSnappy 警告: Snappy native library not loaded 2013-10-14 15:39:31 org.apache.hadoop.mapred.JobClient monitorAndPrintJob 信息: Running job: job_local_0001 2013-10-14 15:39:31 org.apache.hadoop.mapred.Task initialize 信息: Using ResourceCalculatorPlugin : null 2013-10-14 15:39:31 org.apache.hadoop.mapred.Task done 信息: Task:attempt_local_0001_m_000000_0 is done. And is in the process of commiting 2013-10-14 15:39:31 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate 信息: 2013-10-14 15:39:31 org.apache.hadoop.mapred.Task commit 信息: Task attempt_local_0001_m_000000_0 is allowed to commit now 2013-10-14 15:39:31 org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter commitTask 信息: Saved output of task 'attempt_local_0001_m_000000_0' to hdfs://192.168.1.210:9000/user/hdfs/mix_data/seqfile 2013-10-14 15:39:31 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate 信息: 2013-10-14 15:39:31 org.apache.hadoop.mapred.Task sendDone 信息: Task 'attempt_local_0001_m_000000_0' done. 2013-10-14 15:39:32 org.apache.hadoop.mapred.JobClient monitorAndPrintJob 信息: map 100% reduce 0% 2013-10-14 15:39:32 org.apache.hadoop.mapred.JobClient monitorAndPrintJob 信息: Job complete: job_local_0001 2013-10-14 15:39:32 org.apache.hadoop.mapred.Counters log 信息: Counters: 11 2013-10-14 15:39:32 org.apache.hadoop.mapred.Counters log 信息: File Output Format Counters 2013-10-14 15:39:32 org.apache.hadoop.mapred.Counters log 信息: Bytes Written=31390 2013-10-14 15:39:32 org.apache.hadoop.mapred.Counters log 信息: File Input Format Counters 2013-10-14 15:39:32 org.apache.hadoop.mapred.Counters log 信息: Bytes Read=36655 2013-10-14 15:39:32 org.apache.hadoop.mapred.Counters log 信息: FileSystemCounters 2013-10-14 15:39:32 org.apache.hadoop.mapred.Counters log 信息: FILE_BYTES_READ=475910 2013-10-14 15:39:32 org.apache.hadoop.mapred.Counters log 信息: HDFS_BYTES_READ=36655 2013-10-14 15:39:32 org.apache.hadoop.mapred.Counters log 信息: FILE_BYTES_WRITTEN=506350 2013-10-14 15:39:32 org.apache.hadoop.mapred.Counters log 信息: HDFS_BYTES_WRITTEN=68045 2013-10-14 15:39:32 org.apache.hadoop.mapred.Counters log 信息: Map-Reduce Framework 2013-10-14 15:39:32 org.apache.hadoop.mapred.Counters log 信息: Map input records=1000 2013-10-14 15:39:32 org.apache.hadoop.mapred.Counters log 信息: Spilled Records=0 2013-10-14 15:39:32 org.apache.hadoop.mapred.Counters log 信息: Total committed heap usage (bytes)=188284928 2013-10-14 15:39:32 org.apache.hadoop.mapred.Counters log 信息: SPLIT_RAW_BYTES=124 2013-10-14 15:39:32 org.apache.hadoop.mapred.Counters log 信息: Map output records=1000 2013-10-14 15:39:32 org.apache.hadoop.io.compress.CodecPool getCompressor 信息: Got brand-new compressor 2013-10-14 15:39:32 org.apache.hadoop.io.compress.CodecPool getDecompressor 信息: Got brand-new decompressor 2013-10-14 15:39:32 org.apache.hadoop.mapred.JobClient copyAndConfigureFiles 警告: Use GenericOptionsParser for parsing the arguments. Applications should implement Tool for the same. 2013-10-14 15:39:32 org.apache.hadoop.mapreduce.lib.input.FileInputFormat listStatus 信息: Total input paths to process : 1 2013-10-14 15:39:32 org.apache.hadoop.mapred.JobClient monitorAndPrintJob 信息: Running job: job_local_0002 2013-10-14 15:39:32 org.apache.hadoop.mapred.Task initialize 信息: Using ResourceCalculatorPlugin : null 2013-10-14 15:39:32 org.apache.hadoop.mapred.MapTask$MapOutputBuffer 信息: io.sort.mb = 100 2013-10-14 15:39:32 org.apache.hadoop.mapred.MapTask$MapOutputBuffer 信息: data buffer = 79691776/99614720 2013-10-14 15:39:32 org.apache.hadoop.mapred.MapTask$MapOutputBuffer 信息: record buffer = 262144/327680 2013-10-14 15:39:33 org.apache.hadoop.mapred.MapTask$MapOutputBuffer flush 信息: Starting flush of map output 2013-10-14 15:39:33 org.apache.hadoop.mapred.MapTask$MapOutputBuffer sortAndSpill 信息: Finished spill 0 2013-10-14 15:39:33 org.apache.hadoop.mapred.Task done 信息: Task:attempt_local_0002_m_000000_0 is done. And is in the process of commiting 2013-10-14 15:39:33 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate 信息: 2013-10-14 15:39:33 org.apache.hadoop.mapred.Task sendDone 信息: Task 'attempt_local_0002_m_000000_0' done. 2013-10-14 15:39:33 org.apache.hadoop.mapred.Task initialize 信息: Using ResourceCalculatorPlugin : null 2013-10-14 15:39:33 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate 信息: 2013-10-14 15:39:33 org.apache.hadoop.mapred.Merger$MergeQueue merge 信息: Merging 1 sorted segments 2013-10-14 15:39:33 org.apache.hadoop.mapred.Merger$MergeQueue merge 信息: Down to the last merge-pass, with 1 segments left of total size: 623 bytes 2013-10-14 15:39:33 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate 信息: 2013-10-14 15:39:33 org.apache.hadoop.mapred.Task done 信息: Task:attempt_local_0002_r_000000_0 is done. And is in the process of commiting 2013-10-14 15:39:33 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate 信息: 2013-10-14 15:39:33 org.apache.hadoop.mapred.Task commit 信息: Task attempt_local_0002_r_000000_0 is allowed to commit now 2013-10-14 15:39:33 org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter commitTask 信息: Saved output of task 'attempt_local_0002_r_000000_0' to hdfs://192.168.1.210:9000/user/hdfs/mix_data/result/clusters-1 2013-10-14 15:39:33 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate 信息: reduce > reduce 2013-10-14 15:39:33 org.apache.hadoop.mapred.Task sendDone 信息: Task 'attempt_local_0002_r_000000_0' done. 2013-10-14 15:39:33 org.apache.hadoop.mapred.JobClient monitorAndPrintJob 信息: map 100% reduce 100% 2013-10-14 15:39:33 org.apache.hadoop.mapred.JobClient monitorAndPrintJob 信息: Job complete: job_local_0002 2013-10-14 15:39:33 org.apache.hadoop.mapred.Counters log 信息: Counters: 19 2013-10-14 15:39:33 org.apache.hadoop.mapred.Counters log 信息: File Output Format Counters 2013-10-14 15:39:33 org.apache.hadoop.mapred.Counters log 信息: Bytes Written=695 2013-10-14 15:39:33 org.apache.hadoop.mapred.Counters log 信息: FileSystemCounters 2013-10-14 15:39:33 org.apache.hadoop.mapred.Counters log 信息: FILE_BYTES_READ=4239303 2013-10-14 15:39:33 org.apache.hadoop.mapred.Counters log 信息: HDFS_BYTES_READ=203963 2013-10-14 15:39:33 org.apache.hadoop.mapred.Counters log 信息: FILE_BYTES_WRITTEN=4457168 2013-10-14 15:39:33 org.apache.hadoop.mapred.Counters log 信息: HDFS_BYTES_WRITTEN=140321 2013-10-14 15:39:33 org.apache.hadoop.mapred.Counters log 信息: File Input Format Counters 2013-10-14 15:39:33 org.apache.hadoop.mapred.Counters log 信息: Bytes Read=31390 2013-10-14 15:39:33 org.apache.hadoop.mapred.Counters log 信息: Map-Reduce Framework 2013-10-14 15:39:33 org.apache.hadoop.mapred.Counters log 信息: Map output materialized bytes=627 2013-10-14 15:39:33 org.apache.hadoop.mapred.Counters log 信息: Map input records=1000 2013-10-14 15:39:33 org.apache.hadoop.mapred.Counters log 信息: Reduce shuffle bytes=0 2013-10-14 15:39:33 org.apache.hadoop.mapred.Counters log 信息: Spilled Records=6 2013-10-14 15:39:33 org.apache.hadoop.mapred.Counters log 信息: Map output bytes=612 2013-10-14 15:39:33 org.apache.hadoop.mapred.Counters log 信息: Total committed heap usage (bytes)=376569856 2013-10-14 15:39:33 org.apache.hadoop.mapred.Counters log 信息: SPLIT_RAW_BYTES=130 2013-10-14 15:39:33 org.apache.hadoop.mapred.Counters log 信息: Combine input records=0 2013-10-14 15:39:33 org.apache.hadoop.mapred.Counters log 信息: Reduce input records=3 2013-10-14 15:39:33 org.apache.hadoop.mapred.Counters log 信息: Reduce input groups=3 2013-10-14 15:39:33 org.apache.hadoop.mapred.Counters log 信息: Combine output records=0 2013-10-14 15:39:33 org.apache.hadoop.mapred.Counters log 信息: Reduce output records=3 2013-10-14 15:39:33 org.apache.hadoop.mapred.Counters log 信息: Map output records=3 2013-10-14 15:39:34 org.apache.hadoop.mapred.JobClient copyAndConfigureFiles 警告: Use GenericOptionsParser for parsing the arguments. Applications should implement Tool for the same. 2013-10-14 15:39:34 org.apache.hadoop.mapreduce.lib.input.FileInputFormat listStatus 信息: Total input paths to process : 1 2013-10-14 15:39:34 org.apache.hadoop.mapred.JobClient monitorAndPrintJob 信息: Running job: job_local_0003 2013-10-14 15:39:34 org.apache.hadoop.mapred.Task initialize 信息: Using ResourceCalculatorPlugin : null 2013-10-14 15:39:34 org.apache.hadoop.mapred.MapTask$MapOutputBuffer 信息: io.sort.mb = 100 2013-10-14 15:39:34 org.apache.hadoop.mapred.MapTask$MapOutputBuffer 信息: data buffer = 79691776/99614720 2013-10-14 15:39:34 org.apache.hadoop.mapred.MapTask$MapOutputBuffer 信息: record buffer = 262144/327680 2013-10-14 15:39:34 org.apache.hadoop.mapred.MapTask$MapOutputBuffer flush 信息: Starting flush of map output 2013-10-14 15:39:34 org.apache.hadoop.mapred.MapTask$MapOutputBuffer sortAndSpill 信息: Finished spill 0 2013-10-14 15:39:34 org.apache.hadoop.mapred.Task done 信息: Task:attempt_local_0003_m_000000_0 is done. And is in the process of commiting 2013-10-14 15:39:34 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate 信息: 2013-10-14 15:39:34 org.apache.hadoop.mapred.Task sendDone 信息: Task 'attempt_local_0003_m_000000_0' done. 2013-10-14 15:39:34 org.apache.hadoop.mapred.Task initialize 信息: Using ResourceCalculatorPlugin : null 2013-10-14 15:39:34 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate 信息: 2013-10-14 15:39:34 org.apache.hadoop.mapred.Merger$MergeQueue merge 信息: Merging 1 sorted segments 2013-10-14 15:39:34 org.apache.hadoop.mapred.Merger$MergeQueue merge 信息: Down to the last merge-pass, with 1 segments left of total size: 677 bytes 2013-10-14 15:39:34 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate 信息: 2013-10-14 15:39:34 org.apache.hadoop.mapred.Task done 信息: Task:attempt_local_0003_r_000000_0 is done. And is in the process of commiting 2013-10-14 15:39:34 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate 信息: 2013-10-14 15:39:34 org.apache.hadoop.mapred.Task commit 信息: Task attempt_local_0003_r_000000_0 is allowed to commit now 2013-10-14 15:39:34 org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter commitTask 信息: Saved output of task 'attempt_local_0003_r_000000_0' to hdfs://192.168.1.210:9000/user/hdfs/mix_data/result/clusters-2 2013-10-14 15:39:34 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate 信息: reduce > reduce 2013-10-14 15:39:34 org.apache.hadoop.mapred.Task sendDone 信息: Task 'attempt_local_0003_r_000000_0' done. 2013-10-14 15:39:35 org.apache.hadoop.mapred.JobClient monitorAndPrintJob 信息: map 100% reduce 100% 2013-10-14 15:39:35 org.apache.hadoop.mapred.JobClient monitorAndPrintJob 信息: Job complete: job_local_0003 2013-10-14 15:39:35 org.apache.hadoop.mapred.Counters log 信息: Counters: 19 2013-10-14 15:39:35 org.apache.hadoop.mapred.Counters log 信息: File Output Format Counters 2013-10-14 15:39:35 org.apache.hadoop.mapred.Counters log 信息: Bytes Written=695 2013-10-14 15:39:35 org.apache.hadoop.mapred.Counters log 信息: FileSystemCounters 2013-10-14 15:39:35 org.apache.hadoop.mapred.Counters log 信息: FILE_BYTES_READ=7527467 2013-10-14 15:39:35 org.apache.hadoop.mapred.Counters log 信息: HDFS_BYTES_READ=271193 2013-10-14 15:39:35 org.apache.hadoop.mapred.Counters log 信息: FILE_BYTES_WRITTEN=7901744 2013-10-14 15:39:35 org.apache.hadoop.mapred.Counters log 信息: HDFS_BYTES_WRITTEN=142099 2013-10-14 15:39:35 org.apache.hadoop.mapred.Counters log 信息: File Input Format Counters 2013-10-14 15:39:35 org.apache.hadoop.mapred.Counters log 信息: Bytes Read=31390 2013-10-14 15:39:35 org.apache.hadoop.mapred.Counters log 信息: Map-Reduce Framework 2013-10-14 15:39:35 org.apache.hadoop.mapred.Counters log 信息: Map output materialized bytes=681 2013-10-14 15:39:35 org.apache.hadoop.mapred.Counters log 信息: Map input records=1000 2013-10-14 15:39:35 org.apache.hadoop.mapred.Counters log 信息: Reduce shuffle bytes=0 2013-10-14 15:39:35 org.apache.hadoop.mapred.Counters log 信息: Spilled Records=6 2013-10-14 15:39:35 org.apache.hadoop.mapred.Counters log 信息: Map output bytes=666 2013-10-14 15:39:35 org.apache.hadoop.mapred.Counters log 信息: Total committed heap usage (bytes)=575930368 2013-10-14 15:39:35 org.apache.hadoop.mapred.Counters log 信息: SPLIT_RAW_BYTES=130 2013-10-14 15:39:35 org.apache.hadoop.mapred.Counters log 信息: Combine input records=0 2013-10-14 15:39:35 org.apache.hadoop.mapred.Counters log 信息: Reduce input records=3 2013-10-14 15:39:35 org.apache.hadoop.mapred.Counters log 信息: Reduce input groups=3 2013-10-14 15:39:35 org.apache.hadoop.mapred.Counters log 信息: Combine output records=0 2013-10-14 15:39:35 org.apache.hadoop.mapred.Counters log 信息: Reduce output records=3 2013-10-14 15:39:35 org.apache.hadoop.mapred.Counters log 信息: Map output records=3 2013-10-14 15:39:35 org.apache.hadoop.mapred.JobClient copyAndConfigureFiles 警告: Use GenericOptionsParser for parsing the arguments. Applications should implement Tool for the same. 2013-10-14 15:39:35 org.apache.hadoop.mapreduce.lib.input.FileInputFormat listStatus 信息: Total input paths to process : 1 2013-10-14 15:39:35 org.apache.hadoop.mapred.JobClient monitorAndPrintJob 信息: Running job: job_local_0004 2013-10-14 15:39:35 org.apache.hadoop.mapred.Task initialize 信息: Using ResourceCalculatorPlugin : null 2013-10-14 15:39:35 org.apache.hadoop.mapred.MapTask$MapOutputBuffer 信息: io.sort.mb = 100 2013-10-14 15:39:35 org.apache.hadoop.mapred.MapTask$MapOutputBuffer 信息: data buffer = 79691776/99614720 2013-10-14 15:39:35 org.apache.hadoop.mapred.MapTask$MapOutputBuffer 信息: record buffer = 262144/327680 2013-10-14 15:39:35 org.apache.hadoop.mapred.MapTask$MapOutputBuffer flush 信息: Starting flush of map output 2013-10-14 15:39:35 org.apache.hadoop.mapred.MapTask$MapOutputBuffer sortAndSpill 信息: Finished spill 0 2013-10-14 15:39:35 org.apache.hadoop.mapred.Task done 信息: Task:attempt_local_0004_m_000000_0 is done. And is in the process of commiting 2013-10-14 15:39:35 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate 信息: 2013-10-14 15:39:35 org.apache.hadoop.mapred.Task sendDone 信息: Task 'attempt_local_0004_m_000000_0' done. 2013-10-14 15:39:35 org.apache.hadoop.mapred.Task initialize 信息: Using ResourceCalculatorPlugin : null 2013-10-14 15:39:35 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate 信息: 2013-10-14 15:39:35 org.apache.hadoop.mapred.Merger$MergeQueue merge 信息: Merging 1 sorted segments 2013-10-14 15:39:35 org.apache.hadoop.mapred.Merger$MergeQueue merge 信息: Down to the last merge-pass, with 1 segments left of total size: 677 bytes 2013-10-14 15:39:35 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate 信息: 2013-10-14 15:39:35 org.apache.hadoop.mapred.Task done 信息: Task:attempt_local_0004_r_000000_0 is done. And is in the process of commiting 2013-10-14 15:39:35 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate 信息: 2013-10-14 15:39:35 org.apache.hadoop.mapred.Task commit 信息: Task attempt_local_0004_r_000000_0 is allowed to commit now 2013-10-14 15:39:35 org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter commitTask 信息: Saved output of task 'attempt_local_0004_r_000000_0' to hdfs://192.168.1.210:9000/user/hdfs/mix_data/result/clusters-3 2013-10-14 15:39:35 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate 信息: reduce > reduce 2013-10-14 15:39:35 org.apache.hadoop.mapred.Task sendDone 信息: Task 'attempt_local_0004_r_000000_0' done. 2013-10-14 15:39:36 org.apache.hadoop.mapred.JobClient monitorAndPrintJob 信息: map 100% reduce 100% 2013-10-14 15:39:36 org.apache.hadoop.mapred.JobClient monitorAndPrintJob 信息: Job complete: job_local_0004 2013-10-14 15:39:36 org.apache.hadoop.mapred.Counters log 信息: Counters: 19 2013-10-14 15:39:36 org.apache.hadoop.mapred.Counters log 信息: File Output Format Counters 2013-10-14 15:39:36 org.apache.hadoop.mapred.Counters log 信息: Bytes Written=695 2013-10-14 15:39:36 org.apache.hadoop.mapred.Counters log 信息: FileSystemCounters 2013-10-14 15:39:36 org.apache.hadoop.mapred.Counters log 信息: FILE_BYTES_READ=10815685 2013-10-14 15:39:36 org.apache.hadoop.mapred.Counters log 信息: HDFS_BYTES_READ=338143 2013-10-14 15:39:36 org.apache.hadoop.mapred.Counters log 信息: FILE_BYTES_WRITTEN=11346320 2013-10-14 15:39:36 org.apache.hadoop.mapred.Counters log 信息: HDFS_BYTES_WRITTEN=143877 2013-10-14 15:39:36 org.apache.hadoop.mapred.Counters log 信息: File Input Format Counters 2013-10-14 15:39:36 org.apache.hadoop.mapred.Counters log 信息: Bytes Read=31390 2013-10-14 15:39:36 org.apache.hadoop.mapred.Counters log 信息: Map-Reduce Framework 2013-10-14 15:39:36 org.apache.hadoop.mapred.Counters log 信息: Map output materialized bytes=681 2013-10-14 15:39:36 org.apache.hadoop.mapred.Counters log 信息: Map input records=1000 2013-10-14 15:39:36 org.apache.hadoop.mapred.Counters log 信息: Reduce shuffle bytes=0 2013-10-14 15:39:36 org.apache.hadoop.mapred.Counters log 信息: Spilled Records=6 2013-10-14 15:39:36 org.apache.hadoop.mapred.Counters log 信息: Map output bytes=666 2013-10-14 15:39:36 org.apache.hadoop.mapred.Counters log 信息: Total committed heap usage (bytes)=775290880 2013-10-14 15:39:36 org.apache.hadoop.mapred.Counters log 信息: SPLIT_RAW_BYTES=130 2013-10-14 15:39:36 org.apache.hadoop.mapred.Counters log 信息: Combine input records=0 2013-10-14 15:39:36 org.apache.hadoop.mapred.Counters log 信息: Reduce input records=3 2013-10-14 15:39:36 org.apache.hadoop.mapred.Counters log 信息: Reduce input groups=3 2013-10-14 15:39:36 org.apache.hadoop.mapred.Counters log 信息: Combine output records=0 2013-10-14 15:39:36 org.apache.hadoop.mapred.Counters log 信息: Reduce output records=3 2013-10-14 15:39:36 org.apache.hadoop.mapred.Counters log 信息: Map output records=3 2013-10-14 15:39:36 org.apache.hadoop.mapred.JobClient copyAndConfigureFiles 警告: Use GenericOptionsParser for parsing the arguments. Applications should implement Tool for the same. 2013-10-14 15:39:36 org.apache.hadoop.mapreduce.lib.input.FileInputFormat listStatus 信息: Total input paths to process : 1 2013-10-14 15:39:36 org.apache.hadoop.mapred.JobClient monitorAndPrintJob 信息: Running job: job_local_0005 2013-10-14 15:39:36 org.apache.hadoop.mapred.Task initialize 信息: Using ResourceCalculatorPlugin : null 2013-10-14 15:39:36 org.apache.hadoop.mapred.MapTask$MapOutputBuffer 信息: io.sort.mb = 100 2013-10-14 15:39:36 org.apache.hadoop.mapred.MapTask$MapOutputBuffer 信息: data buffer = 79691776/99614720 2013-10-14 15:39:36 org.apache.hadoop.mapred.MapTask$MapOutputBuffer 信息: record buffer = 262144/327680 2013-10-14 15:39:36 org.apache.hadoop.mapred.MapTask$MapOutputBuffer flush 信息: Starting flush of map output 2013-10-14 15:39:36 org.apache.hadoop.mapred.MapTask$MapOutputBuffer sortAndSpill 信息: Finished spill 0 2013-10-14 15:39:36 org.apache.hadoop.mapred.Task done 信息: Task:attempt_local_0005_m_000000_0 is done. And is in the process of commiting 2013-10-14 15:39:36 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate 信息: 2013-10-14 15:39:36 org.apache.hadoop.mapred.Task sendDone 信息: Task 'attempt_local_0005_m_000000_0' done. 2013-10-14 15:39:36 org.apache.hadoop.mapred.Task initialize 信息: Using ResourceCalculatorPlugin : null 2013-10-14 15:39:36 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate 信息: 2013-10-14 15:39:36 org.apache.hadoop.mapred.Merger$MergeQueue merge 信息: Merging 1 sorted segments 2013-10-14 15:39:36 org.apache.hadoop.mapred.Merger$MergeQueue merge 信息: Down to the last merge-pass, with 1 segments left of total size: 677 bytes 2013-10-14 15:39:36 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate 信息: 2013-10-14 15:39:36 org.apache.hadoop.mapred.Task done 信息: Task:attempt_local_0005_r_000000_0 is done. And is in the process of commiting 2013-10-14 15:39:36 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate 信息: 2013-10-14 15:39:36 org.apache.hadoop.mapred.Task commit 信息: Task attempt_local_0005_r_000000_0 is allowed to commit now 2013-10-14 15:39:36 org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter commitTask 信息: Saved output of task 'attempt_local_0005_r_000000_0' to hdfs://192.168.1.210:9000/user/hdfs/mix_data/result/clusters-4 2013-10-14 15:39:36 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate 信息: reduce > reduce 2013-10-14 15:39:36 org.apache.hadoop.mapred.Task sendDone 信息: Task 'attempt_local_0005_r_000000_0' done. 2013-10-14 15:39:37 org.apache.hadoop.mapred.JobClient monitorAndPrintJob 信息: map 100% reduce 100% 2013-10-14 15:39:37 org.apache.hadoop.mapred.JobClient monitorAndPrintJob 信息: Job complete: job_local_0005 2013-10-14 15:39:37 org.apache.hadoop.mapred.Counters log 信息: Counters: 19 2013-10-14 15:39:37 org.apache.hadoop.mapred.Counters log 信息: File Output Format Counters 2013-10-14 15:39:37 org.apache.hadoop.mapred.Counters log 信息: Bytes Written=695 2013-10-14 15:39:37 org.apache.hadoop.mapred.Counters log 信息: FileSystemCounters 2013-10-14 15:39:37 org.apache.hadoop.mapred.Counters log 信息: FILE_BYTES_READ=14103903 2013-10-14 15:39:37 org.apache.hadoop.mapred.Counters log 信息: HDFS_BYTES_READ=405093 2013-10-14 15:39:37 org.apache.hadoop.mapred.Counters log 信息: FILE_BYTES_WRITTEN=14790888 2013-10-14 15:39:37 org.apache.hadoop.mapred.Counters log 信息: HDFS_BYTES_WRITTEN=145655 2013-10-14 15:39:37 org.apache.hadoop.mapred.Counters log 信息: File Input Format Counters 2013-10-14 15:39:37 org.apache.hadoop.mapred.Counters log 信息: Bytes Read=31390 2013-10-14 15:39:37 org.apache.hadoop.mapred.Counters log 信息: Map-Reduce Framework 2013-10-14 15:39:37 org.apache.hadoop.mapred.Counters log 信息: Map output materialized bytes=681 2013-10-14 15:39:37 org.apache.hadoop.mapred.Counters log 信息: Map input records=1000 2013-10-14 15:39:37 org.apache.hadoop.mapred.Counters log 信息: Reduce shuffle bytes=0 2013-10-14 15:39:37 org.apache.hadoop.mapred.Counters log 信息: Spilled Records=6 2013-10-14 15:39:37 org.apache.hadoop.mapred.Counters log 信息: Map output bytes=666 2013-10-14 15:39:37 org.apache.hadoop.mapred.Counters log 信息: Total committed heap usage (bytes)=974651392 2013-10-14 15:39:37 org.apache.hadoop.mapred.Counters log 信息: SPLIT_RAW_BYTES=130 2013-10-14 15:39:37 org.apache.hadoop.mapred.Counters log 信息: Combine input records=0 2013-10-14 15:39:37 org.apache.hadoop.mapred.Counters log 信息: Reduce input records=3 2013-10-14 15:39:37 org.apache.hadoop.mapred.Counters log 信息: Reduce input groups=3 2013-10-14 15:39:37 org.apache.hadoop.mapred.Counters log 信息: Combine output records=0 2013-10-14 15:39:37 org.apache.hadoop.mapred.Counters log 信息: Reduce output records=3 2013-10-14 15:39:37 org.apache.hadoop.mapred.Counters log 信息: Map output records=3 2013-10-14 15:39:37 org.apache.hadoop.mapred.JobClient copyAndConfigureFiles 警告: Use GenericOptionsParser for parsing the arguments. Applications should implement Tool for the same. 2013-10-14 15:39:37 org.apache.hadoop.mapreduce.lib.input.FileInputFormat listStatus 信息: Total input paths to process : 1 2013-10-14 15:39:37 org.apache.hadoop.mapred.JobClient monitorAndPrintJob 信息: Running job: job_local_0006 2013-10-14 15:39:37 org.apache.hadoop.mapred.Task initialize 信息: Using ResourceCalculatorPlugin : null 2013-10-14 15:39:37 org.apache.hadoop.mapred.MapTask$MapOutputBuffer 信息: io.sort.mb = 100 2013-10-14 15:39:37 org.apache.hadoop.mapred.MapTask$MapOutputBuffer 信息: data buffer = 79691776/99614720 2013-10-14 15:39:37 org.apache.hadoop.mapred.MapTask$MapOutputBuffer 信息: record buffer = 262144/327680 2013-10-14 15:39:37 org.apache.hadoop.mapred.MapTask$MapOutputBuffer flush 信息: Starting flush of map output 2013-10-14 15:39:37 org.apache.hadoop.mapred.MapTask$MapOutputBuffer sortAndSpill 信息: Finished spill 0 2013-10-14 15:39:37 org.apache.hadoop.mapred.Task done 信息: Task:attempt_local_0006_m_000000_0 is done. And is in the process of commiting 2013-10-14 15:39:37 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate 信息: 2013-10-14 15:39:37 org.apache.hadoop.mapred.Task sendDone 信息: Task 'attempt_local_0006_m_000000_0' done. 2013-10-14 15:39:37 org.apache.hadoop.mapred.Task initialize 信息: Using ResourceCalculatorPlugin : null 2013-10-14 15:39:37 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate 信息: 2013-10-14 15:39:37 org.apache.hadoop.mapred.Merger$MergeQueue merge 信息: Merging 1 sorted segments 2013-10-14 15:39:37 org.apache.hadoop.mapred.Merger$MergeQueue merge 信息: Down to the last merge-pass, with 1 segments left of total size: 677 bytes 2013-10-14 15:39:37 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate 信息: 2013-10-14 15:39:37 org.apache.hadoop.mapred.Task done 信息: Task:attempt_local_0006_r_000000_0 is done. And is in the process of commiting 2013-10-14 15:39:37 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate 信息: 2013-10-14 15:39:37 org.apache.hadoop.mapred.Task commit 信息: Task attempt_local_0006_r_000000_0 is allowed to commit now 2013-10-14 15:39:37 org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter commitTask 信息: Saved output of task 'attempt_local_0006_r_000000_0' to hdfs://192.168.1.210:9000/user/hdfs/mix_data/result/clusters-5 2013-10-14 15:39:37 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate 信息: reduce > reduce 2013-10-14 15:39:37 org.apache.hadoop.mapred.Task sendDone 信息: Task 'attempt_local_0006_r_000000_0' done. 2013-10-14 15:39:38 org.apache.hadoop.mapred.JobClient monitorAndPrintJob 信息: map 100% reduce 100% 2013-10-14 15:39:38 org.apache.hadoop.mapred.JobClient monitorAndPrintJob 信息: Job complete: job_local_0006 2013-10-14 15:39:38 org.apache.hadoop.mapred.Counters log 信息: Counters: 19 2013-10-14 15:39:38 org.apache.hadoop.mapred.Counters log 信息: File Output Format Counters 2013-10-14 15:39:38 org.apache.hadoop.mapred.Counters log 信息: Bytes Written=695 2013-10-14 15:39:38 org.apache.hadoop.mapred.Counters log 信息: FileSystemCounters 2013-10-14 15:39:38 org.apache.hadoop.mapred.Counters log 信息: FILE_BYTES_READ=17392121 2013-10-14 15:39:38 org.apache.hadoop.mapred.Counters log 信息: HDFS_BYTES_READ=472043 2013-10-14 15:39:38 org.apache.hadoop.mapred.Counters log 信息: FILE_BYTES_WRITTEN=18235456 2013-10-14 15:39:38 org.apache.hadoop.mapred.Counters log 信息: HDFS_BYTES_WRITTEN=147433 2013-10-14 15:39:38 org.apache.hadoop.mapred.Counters log 信息: File Input Format Counters 2013-10-14 15:39:38 org.apache.hadoop.mapred.Counters log 信息: Bytes Read=31390 2013-10-14 15:39:38 org.apache.hadoop.mapred.Counters log 信息: Map-Reduce Framework 2013-10-14 15:39:38 org.apache.hadoop.mapred.Counters log 信息: Map output materialized bytes=681 2013-10-14 15:39:38 org.apache.hadoop.mapred.Counters log 信息: Map input records=1000 2013-10-14 15:39:38 org.apache.hadoop.mapred.Counters log 信息: Reduce shuffle bytes=0 2013-10-14 15:39:38 org.apache.hadoop.mapred.Counters log 信息: Spilled Records=6 2013-10-14 15:39:38 org.apache.hadoop.mapred.Counters log 信息: Map output bytes=666 2013-10-14 15:39:38 org.apache.hadoop.mapred.Counters log 信息: Total committed heap usage (bytes)=1174011904 2013-10-14 15:39:38 org.apache.hadoop.mapred.Counters log 信息: SPLIT_RAW_BYTES=130 2013-10-14 15:39:38 org.apache.hadoop.mapred.Counters log 信息: Combine input records=0 2013-10-14 15:39:38 org.apache.hadoop.mapred.Counters log 信息: Reduce input records=3 2013-10-14 15:39:38 org.apache.hadoop.mapred.Counters log 信息: Reduce input groups=3 2013-10-14 15:39:38 org.apache.hadoop.mapred.Counters log 信息: Combine output records=0 2013-10-14 15:39:38 org.apache.hadoop.mapred.Counters log 信息: Reduce output records=3 2013-10-14 15:39:38 org.apache.hadoop.mapred.Counters log 信息: Map output records=3 2013-10-14 15:39:38 org.apache.hadoop.mapred.JobClient copyAndConfigureFiles 警告: Use GenericOptionsParser for parsing the arguments. Applications should implement Tool for the same. 2013-10-14 15:39:38 org.apache.hadoop.mapreduce.lib.input.FileInputFormat listStatus 信息: Total input paths to process : 1 2013-10-14 15:39:38 org.apache.hadoop.mapred.JobClient monitorAndPrintJob 信息: Running job: job_local_0007 2013-10-14 15:39:38 org.apache.hadoop.mapred.Task initialize 信息: Using ResourceCalculatorPlugin : null 2013-10-14 15:39:38 org.apache.hadoop.mapred.MapTask$MapOutputBuffer 信息: io.sort.mb = 100 2013-10-14 15:39:38 org.apache.hadoop.mapred.MapTask$MapOutputBuffer 信息: data buffer = 79691776/99614720 2013-10-14 15:39:38 org.apache.hadoop.mapred.MapTask$MapOutputBuffer 信息: record buffer = 262144/327680 2013-10-14 15:39:38 org.apache.hadoop.mapred.MapTask$MapOutputBuffer flush 信息: Starting flush of map output 2013-10-14 15:39:38 org.apache.hadoop.mapred.MapTask$MapOutputBuffer sortAndSpill 信息: Finished spill 0 2013-10-14 15:39:38 org.apache.hadoop.mapred.Task done 信息: Task:attempt_local_0007_m_000000_0 is done. And is in the process of commiting 2013-10-14 15:39:38 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate 信息: 2013-10-14 15:39:38 org.apache.hadoop.mapred.Task sendDone 信息: Task 'attempt_local_0007_m_000000_0' done. 2013-10-14 15:39:38 org.apache.hadoop.mapred.Task initialize 信息: Using ResourceCalculatorPlugin : null 2013-10-14 15:39:38 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate 信息: 2013-10-14 15:39:38 org.apache.hadoop.mapred.Merger$MergeQueue merge 信息: Merging 1 sorted segments 2013-10-14 15:39:38 org.apache.hadoop.mapred.Merger$MergeQueue merge 信息: Down to the last merge-pass, with 1 segments left of total size: 677 bytes 2013-10-14 15:39:38 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate 信息: 2013-10-14 15:39:38 org.apache.hadoop.mapred.Task done 信息: Task:attempt_local_0007_r_000000_0 is done. And is in the process of commiting 2013-10-14 15:39:38 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate 信息: 2013-10-14 15:39:38 org.apache.hadoop.mapred.Task commit 信息: Task attempt_local_0007_r_000000_0 is allowed to commit now 2013-10-14 15:39:38 org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter commitTask 信息: Saved output of task 'attempt_local_0007_r_000000_0' to hdfs://192.168.1.210:9000/user/hdfs/mix_data/result/clusters-6 2013-10-14 15:39:38 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate 信息: reduce > reduce 2013-10-14 15:39:38 org.apache.hadoop.mapred.Task sendDone 信息: Task 'attempt_local_0007_r_000000_0' done. 2013-10-14 15:39:39 org.apache.hadoop.mapred.JobClient monitorAndPrintJob 信息: map 100% reduce 100% 2013-10-14 15:39:39 org.apache.hadoop.mapred.JobClient monitorAndPrintJob 信息: Job complete: job_local_0007 2013-10-14 15:39:39 org.apache.hadoop.mapred.Counters log 信息: Counters: 19 2013-10-14 15:39:39 org.apache.hadoop.mapred.Counters log 信息: File Output Format Counters 2013-10-14 15:39:39 org.apache.hadoop.mapred.Counters log 信息: Bytes Written=695 2013-10-14 15:39:39 org.apache.hadoop.mapred.Counters log 信息: FileSystemCounters 2013-10-14 15:39:39 org.apache.hadoop.mapred.Counters log 信息: FILE_BYTES_READ=20680339 2013-10-14 15:39:39 org.apache.hadoop.mapred.Counters log 信息: HDFS_BYTES_READ=538993 2013-10-14 15:39:39 org.apache.hadoop.mapred.Counters log 信息: FILE_BYTES_WRITTEN=21680040 2013-10-14 15:39:39 org.apache.hadoop.mapred.Counters log 信息: HDFS_BYTES_WRITTEN=149211 2013-10-14 15:39:39 org.apache.hadoop.mapred.Counters log 信息: File Input Format Counters 2013-10-14 15:39:39 org.apache.hadoop.mapred.Counters log 信息: Bytes Read=31390 2013-10-14 15:39:39 org.apache.hadoop.mapred.Counters log 信息: Map-Reduce Framework 2013-10-14 15:39:39 org.apache.hadoop.mapred.Counters log 信息: Map output materialized bytes=681 2013-10-14 15:39:39 org.apache.hadoop.mapred.Counters log 信息: Map input records=1000 2013-10-14 15:39:39 org.apache.hadoop.mapred.Counters log 信息: Reduce shuffle bytes=0 2013-10-14 15:39:39 org.apache.hadoop.mapred.Counters log 信息: Spilled Records=6 2013-10-14 15:39:39 org.apache.hadoop.mapred.Counters log 信息: Map output bytes=666 2013-10-14 15:39:39 org.apache.hadoop.mapred.Counters log 信息: Total committed heap usage (bytes)=1373372416 2013-10-14 15:39:39 org.apache.hadoop.mapred.Counters log 信息: SPLIT_RAW_BYTES=130 2013-10-14 15:39:39 org.apache.hadoop.mapred.Counters log 信息: Combine input records=0 2013-10-14 15:39:39 org.apache.hadoop.mapred.Counters log 信息: Reduce input records=3 2013-10-14 15:39:39 org.apache.hadoop.mapred.Counters log 信息: Reduce input groups=3 2013-10-14 15:39:39 org.apache.hadoop.mapred.Counters log 信息: Combine output records=0 2013-10-14 15:39:39 org.apache.hadoop.mapred.Counters log 信息: Reduce output records=3 2013-10-14 15:39:39 org.apache.hadoop.mapred.Counters log 信息: Map output records=3 2013-10-14 15:39:39 org.apache.hadoop.mapred.JobClient copyAndConfigureFiles 警告: Use GenericOptionsParser for parsing the arguments. Applications should implement Tool for the same. 2013-10-14 15:39:39 org.apache.hadoop.mapreduce.lib.input.FileInputFormat listStatus 信息: Total input paths to process : 1 2013-10-14 15:39:39 org.apache.hadoop.mapred.JobClient monitorAndPrintJob 信息: Running job: job_local_0008 2013-10-14 15:39:39 org.apache.hadoop.mapred.Task initialize 信息: Using ResourceCalculatorPlugin : null 2013-10-14 15:39:39 org.apache.hadoop.mapred.MapTask$MapOutputBuffer 信息: io.sort.mb = 100 2013-10-14 15:39:39 org.apache.hadoop.mapred.MapTask$MapOutputBuffer 信息: data buffer = 79691776/99614720 2013-10-14 15:39:39 org.apache.hadoop.mapred.MapTask$MapOutputBuffer 信息: record buffer = 262144/327680 2013-10-14 15:39:39 org.apache.hadoop.mapred.MapTask$MapOutputBuffer flush 信息: Starting flush of map output 2013-10-14 15:39:40 org.apache.hadoop.mapred.MapTask$MapOutputBuffer sortAndSpill 信息: Finished spill 0 2013-10-14 15:39:40 org.apache.hadoop.mapred.Task done 信息: Task:attempt_local_0008_m_000000_0 is done. And is in the process of commiting 2013-10-14 15:39:40 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate 信息: 2013-10-14 15:39:40 org.apache.hadoop.mapred.Task sendDone 信息: Task 'attempt_local_0008_m_000000_0' done. 2013-10-14 15:39:40 org.apache.hadoop.mapred.Task initialize 信息: Using ResourceCalculatorPlugin : null 2013-10-14 15:39:40 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate 信息: 2013-10-14 15:39:40 org.apache.hadoop.mapred.Merger$MergeQueue merge 信息: Merging 1 sorted segments 2013-10-14 15:39:40 org.apache.hadoop.mapred.Merger$MergeQueue merge 信息: Down to the last merge-pass, with 1 segments left of total size: 677 bytes 2013-10-14 15:39:40 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate 信息: 2013-10-14 15:39:40 org.apache.hadoop.mapred.Task done 信息: Task:attempt_local_0008_r_000000_0 is done. And is in the process of commiting 2013-10-14 15:39:40 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate 信息: 2013-10-14 15:39:40 org.apache.hadoop.mapred.Task commit 信息: Task attempt_local_0008_r_000000_0 is allowed to commit now 2013-10-14 15:39:40 org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter commitTask 信息: Saved output of task 'attempt_local_0008_r_000000_0' to hdfs://192.168.1.210:9000/user/hdfs/mix_data/result/clusters-7 2013-10-14 15:39:40 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate 信息: reduce > reduce 2013-10-14 15:39:40 org.apache.hadoop.mapred.Task sendDone 信息: Task 'attempt_local_0008_r_000000_0' done. 2013-10-14 15:39:40 org.apache.hadoop.mapred.JobClient monitorAndPrintJob 信息: map 100% reduce 100% 2013-10-14 15:39:40 org.apache.hadoop.mapred.JobClient monitorAndPrintJob 信息: Job complete: job_local_0008 2013-10-14 15:39:40 org.apache.hadoop.mapred.Counters log 信息: Counters: 19 2013-10-14 15:39:40 org.apache.hadoop.mapred.Counters log 信息: File Output Format Counters 2013-10-14 15:39:40 org.apache.hadoop.mapred.Counters log 信息: Bytes Written=695 2013-10-14 15:39:40 org.apache.hadoop.mapred.Counters log 信息: FileSystemCounters 2013-10-14 15:39:40 org.apache.hadoop.mapred.Counters log 信息: FILE_BYTES_READ=23968557 2013-10-14 15:39:40 org.apache.hadoop.mapred.Counters log 信息: HDFS_BYTES_READ=605943 2013-10-14 15:39:40 org.apache.hadoop.mapred.Counters log 信息: FILE_BYTES_WRITTEN=25124624 2013-10-14 15:39:40 org.apache.hadoop.mapred.Counters log 信息: HDFS_BYTES_WRITTEN=150989 2013-10-14 15:39:40 org.apache.hadoop.mapred.Counters log 信息: File Input Format Counters 2013-10-14 15:39:40 org.apache.hadoop.mapred.Counters log 信息: Bytes Read=31390 2013-10-14 15:39:40 org.apache.hadoop.mapred.Counters log 信息: Map-Reduce Framework 2013-10-14 15:39:40 org.apache.hadoop.mapred.Counters log 信息: Map output materialized bytes=681 2013-10-14 15:39:40 org.apache.hadoop.mapred.Counters log 信息: Map input records=1000 2013-10-14 15:39:40 org.apache.hadoop.mapred.Counters log 信息: Reduce shuffle bytes=0 2013-10-14 15:39:40 org.apache.hadoop.mapred.Counters log 信息: Spilled Records=6 2013-10-14 15:39:40 org.apache.hadoop.mapred.Counters log 信息: Map output bytes=666 2013-10-14 15:39:40 org.apache.hadoop.mapred.Counters log 信息: Total committed heap usage (bytes)=1572732928 2013-10-14 15:39:40 org.apache.hadoop.mapred.Counters log 信息: SPLIT_RAW_BYTES=130 2013-10-14 15:39:40 org.apache.hadoop.mapred.Counters log 信息: Combine input records=0 2013-10-14 15:39:40 org.apache.hadoop.mapred.Counters log 信息: Reduce input records=3 2013-10-14 15:39:40 org.apache.hadoop.mapred.Counters log 信息: Reduce input groups=3 2013-10-14 15:39:40 org.apache.hadoop.mapred.Counters log 信息: Combine output records=0 2013-10-14 15:39:40 org.apache.hadoop.mapred.Counters log 信息: Reduce output records=3 2013-10-14 15:39:40 org.apache.hadoop.mapred.Counters log 信息: Map output records=3 2013-10-14 15:39:41 org.apache.hadoop.mapred.JobClient copyAndConfigureFiles 警告: Use GenericOptionsParser for parsing the arguments. Applications should implement Tool for the same. 2013-10-14 15:39:41 org.apache.hadoop.mapreduce.lib.input.FileInputFormat listStatus 信息: Total input paths to process : 1 2013-10-14 15:39:41 org.apache.hadoop.mapred.JobClient monitorAndPrintJob 信息: Running job: job_local_0009 2013-10-14 15:39:41 org.apache.hadoop.mapred.Task initialize 信息: Using ResourceCalculatorPlugin : null 2013-10-14 15:39:41 org.apache.hadoop.mapred.MapTask$MapOutputBuffer 信息: io.sort.mb = 100 2013-10-14 15:39:41 org.apache.hadoop.mapred.MapTask$MapOutputBuffer 信息: data buffer = 79691776/99614720 2013-10-14 15:39:41 org.apache.hadoop.mapred.MapTask$MapOutputBuffer 信息: record buffer = 262144/327680 2013-10-14 15:39:41 org.apache.hadoop.mapred.MapTask$MapOutputBuffer flush 信息: Starting flush of map output 2013-10-14 15:39:41 org.apache.hadoop.mapred.MapTask$MapOutputBuffer sortAndSpill 信息: Finished spill 0 2013-10-14 15:39:41 org.apache.hadoop.mapred.Task done 信息: Task:attempt_local_0009_m_000000_0 is done. And is in the process of commiting 2013-10-14 15:39:41 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate 信息: 2013-10-14 15:39:41 org.apache.hadoop.mapred.Task sendDone 信息: Task 'attempt_local_0009_m_000000_0' done. 2013-10-14 15:39:41 org.apache.hadoop.mapred.Task initialize 信息: Using ResourceCalculatorPlugin : null 2013-10-14 15:39:41 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate 信息: 2013-10-14 15:39:41 org.apache.hadoop.mapred.Merger$MergeQueue merge 信息: Merging 1 sorted segments 2013-10-14 15:39:41 org.apache.hadoop.mapred.Merger$MergeQueue merge 信息: Down to the last merge-pass, with 1 segments left of total size: 677 bytes 2013-10-14 15:39:41 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate 信息: 2013-10-14 15:39:41 org.apache.hadoop.mapred.Task done 信息: Task:attempt_local_0009_r_000000_0 is done. And is in the process of commiting 2013-10-14 15:39:41 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate 信息: 2013-10-14 15:39:41 org.apache.hadoop.mapred.Task commit 信息: Task attempt_local_0009_r_000000_0 is allowed to commit now 2013-10-14 15:39:41 org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter commitTask 信息: Saved output of task 'attempt_local_0009_r_000000_0' to hdfs://192.168.1.210:9000/user/hdfs/mix_data/result/clusters-8 2013-10-14 15:39:41 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate 信息: reduce > reduce 2013-10-14 15:39:41 org.apache.hadoop.mapred.Task sendDone 信息: Task 'attempt_local_0009_r_000000_0' done. 2013-10-14 15:39:42 org.apache.hadoop.mapred.JobClient monitorAndPrintJob 信息: map 100% reduce 100% 2013-10-14 15:39:42 org.apache.hadoop.mapred.JobClient monitorAndPrintJob 信息: Job complete: job_local_0009 2013-10-14 15:39:42 org.apache.hadoop.mapred.Counters log 信息: Counters: 19 2013-10-14 15:39:42 org.apache.hadoop.mapred.Counters log 信息: File Output Format Counters 2013-10-14 15:39:42 org.apache.hadoop.mapred.Counters log 信息: Bytes Written=695 2013-10-14 15:39:42 org.apache.hadoop.mapred.Counters log 信息: FileSystemCounters 2013-10-14 15:39:42 org.apache.hadoop.mapred.Counters log 信息: FILE_BYTES_READ=27256775 2013-10-14 15:39:42 org.apache.hadoop.mapred.Counters log 信息: HDFS_BYTES_READ=673669 2013-10-14 15:39:42 org.apache.hadoop.mapred.Counters log 信息: FILE_BYTES_WRITTEN=28569192 2013-10-14 15:39:42 org.apache.hadoop.mapred.Counters log 信息: HDFS_BYTES_WRITTEN=152767 2013-10-14 15:39:42 org.apache.hadoop.mapred.Counters log 信息: File Input Format Counters 2013-10-14 15:39:42 org.apache.hadoop.mapred.Counters log 信息: Bytes Read=31390 2013-10-14 15:39:42 org.apache.hadoop.mapred.Counters log 信息: Map-Reduce Framework 2013-10-14 15:39:42 org.apache.hadoop.mapred.Counters log 信息: Map output materialized bytes=681 2013-10-14 15:39:42 org.apache.hadoop.mapred.Counters log 信息: Map input records=1000 2013-10-14 15:39:42 org.apache.hadoop.mapred.Counters log 信息: Reduce shuffle bytes=0 2013-10-14 15:39:42 org.apache.hadoop.mapred.Counters log 信息: Spilled Records=6 2013-10-14 15:39:42 org.apache.hadoop.mapred.Counters log 信息: Map output bytes=666 2013-10-14 15:39:42 org.apache.hadoop.mapred.Counters log 信息: Total committed heap usage (bytes)=1772093440 2013-10-14 15:39:42 org.apache.hadoop.mapred.Counters log 信息: SPLIT_RAW_BYTES=130 2013-10-14 15:39:42 org.apache.hadoop.mapred.Counters log 信息: Combine input records=0 2013-10-14 15:39:42 org.apache.hadoop.mapred.Counters log 信息: Reduce input records=3 2013-10-14 15:39:42 org.apache.hadoop.mapred.Counters log 信息: Reduce input groups=3 2013-10-14 15:39:42 org.apache.hadoop.mapred.Counters log 信息: Combine output records=0 2013-10-14 15:39:42 org.apache.hadoop.mapred.Counters log 信息: Reduce output records=3 2013-10-14 15:39:42 org.apache.hadoop.mapred.Counters log 信息: Map output records=3 2013-10-14 15:39:42 org.apache.hadoop.mapred.JobClient copyAndConfigureFiles 警告: Use GenericOptionsParser for parsing the arguments. Applications should implement Tool for the same. 2013-10-14 15:39:42 org.apache.hadoop.mapreduce.lib.input.FileInputFormat listStatus 信息: Total input paths to process : 1 2013-10-14 15:39:42 org.apache.hadoop.mapred.JobClient monitorAndPrintJob 信息: Running job: job_local_0010 2013-10-14 15:39:42 org.apache.hadoop.mapred.Task initialize 信息: Using ResourceCalculatorPlugin : null 2013-10-14 15:39:42 org.apache.hadoop.mapred.MapTask$MapOutputBuffer 信息: io.sort.mb = 100 2013-10-14 15:39:42 org.apache.hadoop.mapred.MapTask$MapOutputBuffer 信息: data buffer = 79691776/99614720 2013-10-14 15:39:42 org.apache.hadoop.mapred.MapTask$MapOutputBuffer 信息: record buffer = 262144/327680 2013-10-14 15:39:42 org.apache.hadoop.mapred.MapTask$MapOutputBuffer flush 信息: Starting flush of map output 2013-10-14 15:39:42 org.apache.hadoop.mapred.MapTask$MapOutputBuffer sortAndSpill 信息: Finished spill 0 2013-10-14 15:39:42 org.apache.hadoop.mapred.Task done 信息: Task:attempt_local_0010_m_000000_0 is done. And is in the process of commiting 2013-10-14 15:39:42 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate 信息: 2013-10-14 15:39:42 org.apache.hadoop.mapred.Task sendDone 信息: Task 'attempt_local_0010_m_000000_0' done. 2013-10-14 15:39:42 org.apache.hadoop.mapred.Task initialize 信息: Using ResourceCalculatorPlugin : null 2013-10-14 15:39:42 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate 信息: 2013-10-14 15:39:42 org.apache.hadoop.mapred.Merger$MergeQueue merge 信息: Merging 1 sorted segments 2013-10-14 15:39:42 org.apache.hadoop.mapred.Merger$MergeQueue merge 信息: Down to the last merge-pass, with 1 segments left of total size: 677 bytes 2013-10-14 15:39:42 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate 信息: 2013-10-14 15:39:42 org.apache.hadoop.mapred.Task done 信息: Task:attempt_local_0010_r_000000_0 is done. And is in the process of commiting 2013-10-14 15:39:42 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate 信息: 2013-10-14 15:39:42 org.apache.hadoop.mapred.Task commit 信息: Task attempt_local_0010_r_000000_0 is allowed to commit now 2013-10-14 15:39:42 org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter commitTask 信息: Saved output of task 'attempt_local_0010_r_000000_0' to hdfs://192.168.1.210:9000/user/hdfs/mix_data/result/clusters-9 2013-10-14 15:39:42 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate 信息: reduce > reduce 2013-10-14 15:39:42 org.apache.hadoop.mapred.Task sendDone 信息: Task 'attempt_local_0010_r_000000_0' done. 2013-10-14 15:39:43 org.apache.hadoop.mapred.JobClient monitorAndPrintJob 信息: map 100% reduce 100% 2013-10-14 15:39:43 org.apache.hadoop.mapred.JobClient monitorAndPrintJob 信息: Job complete: job_local_0010 2013-10-14 15:39:43 org.apache.hadoop.mapred.Counters log 信息: Counters: 19 2013-10-14 15:39:43 org.apache.hadoop.mapred.Counters log 信息: File Output Format Counters 2013-10-14 15:39:43 org.apache.hadoop.mapred.Counters log 信息: Bytes Written=695 2013-10-14 15:39:43 org.apache.hadoop.mapred.Counters log 信息: FileSystemCounters 2013-10-14 15:39:43 org.apache.hadoop.mapred.Counters log 信息: FILE_BYTES_READ=30544993 2013-10-14 15:39:43 org.apache.hadoop.mapred.Counters log 信息: HDFS_BYTES_READ=741007 2013-10-14 15:39:43 org.apache.hadoop.mapred.Counters log 信息: FILE_BYTES_WRITTEN=32013760 2013-10-14 15:39:43 org.apache.hadoop.mapred.Counters log 信息: HDFS_BYTES_WRITTEN=154545 2013-10-14 15:39:43 org.apache.hadoop.mapred.Counters log 信息: File Input Format Counters 2013-10-14 15:39:43 org.apache.hadoop.mapred.Counters log 信息: Bytes Read=31390 2013-10-14 15:39:43 org.apache.hadoop.mapred.Counters log 信息: Map-Reduce Framework 2013-10-14 15:39:43 org.apache.hadoop.mapred.Counters log 信息: Map output materialized bytes=681 2013-10-14 15:39:43 org.apache.hadoop.mapred.Counters log 信息: Map input records=1000 2013-10-14 15:39:43 org.apache.hadoop.mapred.Counters log 信息: Reduce shuffle bytes=0 2013-10-14 15:39:43 org.apache.hadoop.mapred.Counters log 信息: Spilled Records=6 2013-10-14 15:39:43 org.apache.hadoop.mapred.Counters log 信息: Map output bytes=666 2013-10-14 15:39:43 org.apache.hadoop.mapred.Counters log 信息: Total committed heap usage (bytes)=1966735360 2013-10-14 15:39:43 org.apache.hadoop.mapred.Counters log 信息: SPLIT_RAW_BYTES=130 2013-10-14 15:39:43 org.apache.hadoop.mapred.Counters log 信息: Combine input records=0 2013-10-14 15:39:43 org.apache.hadoop.mapred.Counters log 信息: Reduce input records=3 2013-10-14 15:39:43 org.apache.hadoop.mapred.Counters log 信息: Reduce input groups=3 2013-10-14 15:39:43 org.apache.hadoop.mapred.Counters log 信息: Combine output records=0 2013-10-14 15:39:43 org.apache.hadoop.mapred.Counters log 信息: Reduce output records=3 2013-10-14 15:39:43 org.apache.hadoop.mapred.Counters log 信息: Map output records=3 2013-10-14 15:39:43 org.apache.hadoop.mapred.JobClient copyAndConfigureFiles 警告: Use GenericOptionsParser for parsing the arguments. Applications should implement Tool for the same. 2013-10-14 15:39:43 org.apache.hadoop.mapreduce.lib.input.FileInputFormat listStatus 信息: Total input paths to process : 1 2013-10-14 15:39:43 org.apache.hadoop.mapred.JobClient monitorAndPrintJob 信息: Running job: job_local_0011 2013-10-14 15:39:43 org.apache.hadoop.mapred.Task initialize 信息: Using ResourceCalculatorPlugin : null 2013-10-14 15:39:43 org.apache.hadoop.mapred.MapTask$MapOutputBuffer 信息: io.sort.mb = 100 2013-10-14 15:39:43 org.apache.hadoop.mapred.MapTask$MapOutputBuffer 信息: data buffer = 79691776/99614720 2013-10-14 15:39:43 org.apache.hadoop.mapred.MapTask$MapOutputBuffer 信息: record buffer = 262144/327680 2013-10-14 15:39:43 org.apache.hadoop.mapred.MapTask$MapOutputBuffer flush 信息: Starting flush of map output 2013-10-14 15:39:43 org.apache.hadoop.mapred.MapTask$MapOutputBuffer sortAndSpill 信息: Finished spill 0 2013-10-14 15:39:43 org.apache.hadoop.mapred.Task done 信息: Task:attempt_local_0011_m_000000_0 is done. And is in the process of commiting 2013-10-14 15:39:43 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate 信息: 2013-10-14 15:39:43 org.apache.hadoop.mapred.Task sendDone 信息: Task 'attempt_local_0011_m_000000_0' done. 2013-10-14 15:39:43 org.apache.hadoop.mapred.Task initialize 信息: Using ResourceCalculatorPlugin : null 2013-10-14 15:39:43 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate 信息: 2013-10-14 15:39:43 org.apache.hadoop.mapred.Merger$MergeQueue merge 信息: Merging 1 sorted segments 2013-10-14 15:39:43 org.apache.hadoop.mapred.Merger$MergeQueue merge 信息: Down to the last merge-pass, with 1 segments left of total size: 677 bytes 2013-10-14 15:39:43 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate 信息: 2013-10-14 15:39:43 org.apache.hadoop.mapred.Task done 信息: Task:attempt_local_0011_r_000000_0 is done. And is in the process of commiting 2013-10-14 15:39:43 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate 信息: 2013-10-14 15:39:43 org.apache.hadoop.mapred.Task commit 信息: Task attempt_local_0011_r_000000_0 is allowed to commit now 2013-10-14 15:39:43 org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter commitTask 信息: Saved output of task 'attempt_local_0011_r_000000_0' to hdfs://192.168.1.210:9000/user/hdfs/mix_data/result/clusters-10 2013-10-14 15:39:43 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate 信息: reduce > reduce 2013-10-14 15:39:43 org.apache.hadoop.mapred.Task sendDone 信息: Task 'attempt_local_0011_r_000000_0' done. 2013-10-14 15:39:44 org.apache.hadoop.mapred.JobClient monitorAndPrintJob 信息: map 100% reduce 100% 2013-10-14 15:39:44 org.apache.hadoop.mapred.JobClient monitorAndPrintJob 信息: Job complete: job_local_0011 2013-10-14 15:39:44 org.apache.hadoop.mapred.Counters log 信息: Counters: 19 2013-10-14 15:39:44 org.apache.hadoop.mapred.Counters log 信息: File Output Format Counters 2013-10-14 15:39:44 org.apache.hadoop.mapred.Counters log 信息: Bytes Written=695 2013-10-14 15:39:44 org.apache.hadoop.mapred.Counters log 信息: FileSystemCounters 2013-10-14 15:39:44 org.apache.hadoop.mapred.Counters log 信息: FILE_BYTES_READ=33833211 2013-10-14 15:39:44 org.apache.hadoop.mapred.Counters log 信息: HDFS_BYTES_READ=808345 2013-10-14 15:39:44 org.apache.hadoop.mapred.Counters log 信息: FILE_BYTES_WRITTEN=35458320 2013-10-14 15:39:44 org.apache.hadoop.mapred.Counters log 信息: HDFS_BYTES_WRITTEN=156323 2013-10-14 15:39:44 org.apache.hadoop.mapred.Counters log 信息: File Input Format Counters 2013-10-14 15:39:44 org.apache.hadoop.mapred.Counters log 信息: Bytes Read=31390 2013-10-14 15:39:44 org.apache.hadoop.mapred.Counters log 信息: Map-Reduce Framework 2013-10-14 15:39:44 org.apache.hadoop.mapred.Counters log 信息: Map output materialized bytes=681 2013-10-14 15:39:44 org.apache.hadoop.mapred.Counters log 信息: Map input records=1000 2013-10-14 15:39:44 org.apache.hadoop.mapred.Counters log 信息: Reduce shuffle bytes=0 2013-10-14 15:39:44 org.apache.hadoop.mapred.Counters log 信息: Spilled Records=6 2013-10-14 15:39:44 org.apache.hadoop.mapred.Counters log 信息: Map output bytes=666 2013-10-14 15:39:44 org.apache.hadoop.mapred.Counters log 信息: Total committed heap usage (bytes)=2166095872 2013-10-14 15:39:44 org.apache.hadoop.mapred.Counters log 信息: SPLIT_RAW_BYTES=130 2013-10-14 15:39:44 org.apache.hadoop.mapred.Counters log 信息: Combine input records=0 2013-10-14 15:39:44 org.apache.hadoop.mapred.Counters log 信息: Reduce input records=3 2013-10-14 15:39:44 org.apache.hadoop.mapred.Counters log 信息: Reduce input groups=3 2013-10-14 15:39:44 org.apache.hadoop.mapred.Counters log 信息: Combine output records=0 2013-10-14 15:39:44 org.apache.hadoop.mapred.Counters log 信息: Reduce output records=3 2013-10-14 15:39:44 org.apache.hadoop.mapred.Counters log 信息: Map output records=3 2013-10-14 15:39:44 org.apache.hadoop.mapred.JobClient copyAndConfigureFiles 警告: Use GenericOptionsParser for parsing the arguments. Applications should implement Tool for the same. 2013-10-14 15:39:44 org.apache.hadoop.mapreduce.lib.input.FileInputFormat listStatus 信息: Total input paths to process : 1 2013-10-14 15:39:44 org.apache.hadoop.mapred.JobClient monitorAndPrintJob 信息: Running job: job_local_0012 2013-10-14 15:39:44 org.apache.hadoop.mapred.Task initialize 信息: Using ResourceCalculatorPlugin : null 2013-10-14 15:39:44 org.apache.hadoop.mapred.Task done 信息: Task:attempt_local_0012_m_000000_0 is done. And is in the process of commiting 2013-10-14 15:39:44 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate 信息: 2013-10-14 15:39:44 org.apache.hadoop.mapred.Task commit 信息: Task attempt_local_0012_m_000000_0 is allowed to commit now 2013-10-14 15:39:44 org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter commitTask 信息: Saved output of task 'attempt_local_0012_m_000000_0' to hdfs://192.168.1.210:9000/user/hdfs/mix_data/result/clusteredPoints 2013-10-14 15:39:44 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate 信息: 2013-10-14 15:39:44 org.apache.hadoop.mapred.Task sendDone 信息: Task 'attempt_local_0012_m_000000_0' done. 2013-10-14 15:39:45 org.apache.hadoop.mapred.JobClient monitorAndPrintJob 信息: map 100% reduce 0% 2013-10-14 15:39:45 org.apache.hadoop.mapred.JobClient monitorAndPrintJob 信息: Job complete: job_local_0012 2013-10-14 15:39:45 org.apache.hadoop.mapred.Counters log 信息: Counters: 11 2013-10-14 15:39:45 org.apache.hadoop.mapred.Counters log 信息: File Output Format Counters 2013-10-14 15:39:45 org.apache.hadoop.mapred.Counters log 信息: Bytes Written=41520 2013-10-14 15:39:45 org.apache.hadoop.mapred.Counters log 信息: File Input Format Counters 2013-10-14 15:39:45 org.apache.hadoop.mapred.Counters log 信息: Bytes Read=31390 2013-10-14 15:39:45 org.apache.hadoop.mapred.Counters log 信息: FileSystemCounters 2013-10-14 15:39:45 org.apache.hadoop.mapred.Counters log 信息: FILE_BYTES_READ=18560374 2013-10-14 15:39:45 org.apache.hadoop.mapred.Counters log 信息: HDFS_BYTES_READ=437203 2013-10-14 15:39:45 org.apache.hadoop.mapred.Counters log 信息: FILE_BYTES_WRITTEN=19450325 2013-10-14 15:39:45 org.apache.hadoop.mapred.Counters log 信息: HDFS_BYTES_WRITTEN=120417 2013-10-14 15:39:45 org.apache.hadoop.mapred.Counters log 信息: Map-Reduce Framework 2013-10-14 15:39:45 org.apache.hadoop.mapred.Counters log 信息: Map input records=1000 2013-10-14 15:39:45 org.apache.hadoop.mapred.Counters log 信息: Spilled Records=0 2013-10-14 15:39:45 org.apache.hadoop.mapred.Counters log 信息: Total committed heap usage (bytes)=1083047936 2013-10-14 15:39:45 org.apache.hadoop.mapred.Counters log 信息: SPLIT_RAW_BYTES=130 2013-10-14 15:39:45 org.apache.hadoop.mapred.Counters log 信息: Map output records=1000 Dumping out clusters from clusters: hdfs://192.168.1.210:9000/user/hdfs/mix_data/result/clusters-*-final and clusteredPoints: hdfs://192.168.1.210:9000/user/hdfs/mix_data/result/clusteredPoints CL-552{n=443 c=[1.631, -0.412] r=[1.563, 1.407]}Weight : [props - optional]: Point:1.0: [-2.393, 3.347]1.0: [-4.364, 1.905]1.0: [-3.275, 0.023]1.0: [-2.479, 2.534]1.0: [-0.559, 1.223]...CL-847{n=77 c=[-2.953, -0.971] r=[1.767, 2.189]}Weight : [props - optional]: Point:1.0: [-0.883, -3.320]1.0: [-1.099, -6.063]1.0: [-0.004, -0.610]1.0: [-2.996, -3.610]1.0: [3.988, 1.008]...CL-823{n=480 c=[0.219, 2.600] r=[1.479, 1.385]}Weight : [props - optional]: Point:1.0: [2.670, 1.851]1.0: [2.177, 6.773]1.0: [5.537, 2.651]1.0: [5.663, 6.868]1.0: [5.117, 3.747]1.0: [1.912, 2.959]...

    4). 聚類結果解讀
    我們可以把上面的日志分解析成3個部分解讀

    • a. 初始化環境
    • b. 算法執行
    • c. 打印聚類結果

    a. 初始化環境
    出初HDFS的數據目錄和工作目錄,并上傳數據文件。

    Delete: hdfs://192.168.1.210:9000/user/hdfs/mix_data Create: hdfs://192.168.1.210:9000/user/hdfs/mix_data copy from: datafile/randomData.csv to hdfs://192.168.1.210:9000/user/hdfs/mix_data ls: hdfs://192.168.1.210:9000/user/hdfs/mix_data ========================================================== name: hdfs://192.168.1.210:9000/user/hdfs/mix_data/randomData.csv, folder: false, size: 36655

    b. 算法執行
    算法執行,有3個步驟。

    • 1):把原始數據randomData.csv,轉成Mahout sequence files of VectorWritable。
    • 2):通過隨機的方法,選中kmeans的3個中心,做為初始集群
    • 3):根據迭代次數的設置,執行MapReduce,進行計算

    1):把原始數據randomData.csv,轉成Mahout sequence files of VectorWritable。

    程序源代碼:

    InputDriver.runJob(new Path(inPath), new Path(seqFile), "org.apache.mahout.math.RandomAccessSparseVector");

    日志輸出:

    Job complete: job_local_0001

    2):通過隨機的方法,選中kmeans的3個中心,做為初始集群

    程序源代碼:

    int k = 3;Path seqFilePath = new Path(seqFile);Path clustersSeeds = new Path(seeds);DistanceMeasure measure = new EuclideanDistanceMeasure();clustersSeeds = RandomSeedGenerator.buildRandom(conf, seqFilePath, clustersSeeds, k, measure);

    日志輸出:

    Job complete: job_local_0002

    3):根據迭代次數的設置,執行MapReduce,進行計算
    程序源代碼:

    KMeansDriver.run(conf, seqFilePath, clustersSeeds, new Path(outPath), measure, 0.01, 10, true, 0.01, false);

    日志輸出:

    Job complete: job_local_0003 Job complete: job_local_0004 Job complete: job_local_0005 Job complete: job_local_0006 Job complete: job_local_0007 Job complete: job_local_0008 Job complete: job_local_0009 Job complete: job_local_0010 Job complete: job_local_0011 Job complete: job_local_0012

    c. 打印聚類結果

    Dumping out clusters from clusters: hdfs://192.168.1.210:9000/user/hdfs/mix_data/result/clusters-*-final and clusteredPoints: hdfs://192.168.1.210:9000/user/hdfs/mix_data/result/clusteredPoints CL-552{n=443 c=[1.631, -0.412] r=[1.563, 1.407]} CL-847{n=77 c=[-2.953, -0.971] r=[1.767, 2.189]} CL-823{n=480 c=[0.219, 2.600] r=[1.479, 1.385]}

    運行結果:有3個中心。

    • Cluster1, 包括443個點,中心坐標[1.631, -0.412]
    • Cluster2, 包括77個點,中心坐標[-2.953, -0.971]
    • Cluster3, 包括480 個點,中心坐標[0.219, 2.600]

    5). HDFS產生的目錄

    # 根目錄 ~ hadoop fs -ls /user/hdfs/mix_data Found 4 items -rw-r--r-- 3 Administrator supergroup 36655 2013-10-04 15:31 /user/hdfs/mix_data/randomData.csv drwxr-xr-x - Administrator supergroup 0 2013-10-04 15:31 /user/hdfs/mix_data/result drwxr-xr-x - Administrator supergroup 0 2013-10-04 15:31 /user/hdfs/mix_data/seeds drwxr-xr-x - Administrator supergroup 0 2013-10-04 15:31 /user/hdfs/mix_data/seqfile# 輸出目錄 ~ hadoop fs -ls /user/hdfs/mix_data/result Found 13 items -rw-r--r-- 3 Administrator supergroup 194 2013-10-04 15:31 /user/hdfs/mix_data/result/_policy drwxr-xr-x - Administrator supergroup 0 2013-10-04 15:31 /user/hdfs/mix_data/result/clusteredPoints drwxr-xr-x - Administrator supergroup 0 2013-10-04 15:31 /user/hdfs/mix_data/result/clusters-0 drwxr-xr-x - Administrator supergroup 0 2013-10-04 15:31 /user/hdfs/mix_data/result/clusters-1 drwxr-xr-x - Administrator supergroup 0 2013-10-04 15:31 /user/hdfs/mix_data/result/clusters-10-final drwxr-xr-x - Administrator supergroup 0 2013-10-04 15:31 /user/hdfs/mix_data/result/clusters-2 drwxr-xr-x - Administrator supergroup 0 2013-10-04 15:31 /user/hdfs/mix_data/result/clusters-3 drwxr-xr-x - Administrator supergroup 0 2013-10-04 15:31 /user/hdfs/mix_data/result/clusters-4 drwxr-xr-x - Administrator supergroup 0 2013-10-04 15:31 /user/hdfs/mix_data/result/clusters-5 drwxr-xr-x - Administrator supergroup 0 2013-10-04 15:31 /user/hdfs/mix_data/result/clusters-6 drwxr-xr-x - Administrator supergroup 0 2013-10-04 15:31 /user/hdfs/mix_data/result/clusters-7 drwxr-xr-x - Administrator supergroup 0 2013-10-04 15:31 /user/hdfs/mix_data/result/clusters-8 drwxr-xr-x - Administrator supergroup 0 2013-10-04 15:31 /user/hdfs/mix_data/result/clusters-9# 產生的隨機中心種子目錄 ~ hadoop fs -ls /user/hdfs/mix_data/seeds Found 1 items -rw-r--r-- 3 Administrator supergroup 599 2013-10-04 15:31 /user/hdfs/mix_data/seeds/part-randomSeed# 輸入文件換成Mahout格式文件的目錄 ~ hadoop fs -ls /user/hdfs/mix_data/seqfile Found 2 items -rw-r--r-- 3 Administrator supergroup 0 2013-10-04 15:31 /user/hdfs/mix_data/seqfile/_SUCCESS -rw-r--r-- 3 Administrator supergroup 31390 2013-10-04 15:31 /user/hdfs/mix_data/seqfile/part-m-00000

    4. 用R語言可視化結果

    分別把聚類后的點,保存到不同的cluster*.csv文件,然后用R語言畫圖。

    c1<-read.csv(file="cluster1.csv",sep=",",header=FALSE) c2<-read.csv(file="cluster2.csv",sep=",",header=FALSE) c3<-read.csv(file="cluster3.csv",sep=",",header=FALSE) y<-rbind(c1,c2,c3) cols<-c(rep(1,nrow(c1)),rep(2,nrow(c2)),rep(3,nrow(c3))) plot(y, col=c("black","blue","green")[cols]) center<-matrix(c(1.631, -0.412,-2.953, -0.971,0.219, 2.600),ncol=2,byrow=TRUE) points(center, col="violetred", pch = 19)

    從上圖中,我們看到有 黑,藍,綠,三種顏色的空心點,這些點就是原始數據。
    3個紫色實點,是Mahout的kmeans后生成的3個中心。

    對比文章中用R語言實現的kmeans的分類和中心,都不太一樣。?用Maven構建Mahout項目

    簡單總結一下,在使用kmeans時,根據距離算法,閾值,出始中心,迭代次數的不同,kmeans計算的結果是不相同的。因此,用kmeans算法,我們一般只能得到一個模糊的分類標準,這個標準對于我們認識未知領域的數據集是很有幫助的。不能做為精確衡量數據的指標。

    5. 模板項目上傳github

    https://github.com/bsspirit/maven_mahout_template/tree/mahout-0.8

    大家可以下載這個項目,做為開發的起點。

    ~ git clone https://github.com/bsspirit/maven_mahout_template ~ git checkout mahout-0.8

    這樣,我們完成了Mahout的聚類算法Kmeans的分步式實現。接下來,我們會繼續做關于Mahout中分類的實驗!

    轉載請注明出處:
    http://blog.fens.me/hadoop-mahout-kmeans/

    總結

    以上是生活随笔為你收集整理的Mahout分步式程序开发 聚类Kmeans的全部內容,希望文章能夠幫你解決所遇到的問題。

    如果覺得生活随笔網站內容還不錯,歡迎將生活随笔推薦給好友。

    主站蜘蛛池模板: 久久精品66| 亚洲精品一区二区三区不卡 | ,国产精品国产三级国产 | 一级片亚洲 | 国产乱妇无码大片在线观看 | 中国亚洲老头同性gay男男… | 国产 日韩 欧美 精品 | 久久依人网 | 可以看的毛片 | www男人天堂 | 欧美毛片在线观看 | 国产夜色视频 | 少妇一级淫片aaaaaaa | 青娱乐青青草 | 国产精品乱码一区二区视频 | 在线不卡中文字幕 | www五月 | 91福利在线免费观看 | 九久久久久 | av2018| 视频一区二区三 | 国产五月婷婷 | 自拍偷拍av | 四虎4hu永久免费网站影院 | 国产偷亚洲偷欧美偷精品 | 男女做爰猛烈高潮描写 | 天天爽视频 | 国产黄色三级 | 久久久男人天堂 | 久草免费在线观看 | 婷婷国产成人精品视频 | 污视频网站在线播放 | 在线a网 | 久久久久久999 | 日本一区二区三区四区在线观看 | 爱操在线| 打白嫩屁屁网站视频短裙 | 人妻久久一区二区 | 今天最新中文字幕mv高清 | 乱色精品无码一区二区国产盗 | 美女被c出白浆 | 这里只有精品视频 | 澳门三级 | 九七影院在线观看免费观看电视 | 久久aⅴ国产欧美74aaa | 日韩一区二区三区四区五区六区 | jizjiz中国少妇高潮水多 | 潮喷失禁大喷水无码 | 奇米四色在线观看 | 国产粉嫩白浆 | 免费黄色看片网站 | 国产午夜精品一区二区三区欧美 | 国产欧美日韩综合 | 日韩美女一区 | 国产精品嫩草69影院 | 欧美日韩毛片 | 不卡av电影在线观看 | 欧美xxxx黑人又粗又长密月 | 熟女毛毛多熟妇人妻aⅴ在线毛片 | 国产喷水在线 | 欧美性狂猛xxxxxbbbbb | 看了下面会湿的视频 | 搞黄视频在线观看 | 国产成人麻豆精品午夜在线 | 成人午夜视频在线 | 成人精品在线播放 | 国产一区二区精品久久 | 亚洲 成人 av | 性欧美最猛 | 性xxxxbbbb| 国产精品福利在线观看 | 亚洲综合一区中 | 日韩av一区在线观看 | 伊人逼逼| 色av性av丰满av| 欧美成人乱码一区二区三区 | 快色视频 | 国产精品毛片一区二区在线看 | 中文字幕第六页 | 台湾av在线播放 | 国产精品xxxx喷水欧美 | 欧美日韩国产黄色 | 亚洲熟妇无码av在线播放 | 少妇伦子伦精品无吗 | 91精品国产综合久久久蜜臀九色 | 自拍偷拍视频在线 | 深夜久久久 | 欧美日韩综合一区二区 | wwww日本60| 免费黄色观看 | 公交顶臀绿裙妇女配视频 | 国产精品久免费的黄网站 | 成人免费看片入口 | 国产色图视频 | 尤物在线免费观看 | 国产欧美日韩精品一区 | 五月天狠狠操 | 一级黄色片在线播放 | 国产亚洲欧美一区二区 |