设为首页 加入收藏

TOP

hive排序特性研究(七)
2014-11-24 07:25:19 来源: 作者: 【 】 浏览:2
Tags:hive 排序 特性 研究
Number of reduce tasks not specified. Defaulting to jobconf value of: 2
In order to change the average load for a reducer (in bytes):
set hive.exec.reducers.bytes.per.reducer=
In order to limit the maximum number of reducers:
set hive.exec.reducers.max=
In order to set a constant number of reducers:
set mapred.reduce.tasks=
Starting Job = job_201307151509_15499, Tracking URL = http://mwtec-50:50030/jobdetails.jsp jobid=job_201307151509_15499
Kill Command = /home/hadoop/hadoop-0.20.2/bin/hadoop job -Dmapred.job.tracker=mwtec-50:9002 -kill job_201307151509_15499
Hadoop job information for Stage-1: number of mappers: 1; number of reducers: 2
2013-08-05 18:37:14,681 Stage-1 map = 0%, reduce = 0%
2013-08-05 18:37:16,691 Stage-1 map = 100%, reduce = 0%, Cumulative CPU 1.33 sec
2013-08-05 18:37:17,697 Stage-1 map = 100%, reduce = 0%, Cumulative CPU 1.33 sec
2013-08-05 18:37:18,703 Stage-1 map = 100%, reduce = 0%, Cumulative CPU 1.33 sec
2013-08-05 18:37:19,710 Stage-1 map = 100%, reduce = 0%, Cumulative CPU 1.33 sec
2013-08-05 18:37:20,717 Stage-1 map = 100%, reduce = 0%, Cumulative CPU 1.33 sec
2013-08-05 18:37:21,727 Stage-1 map = 100%, reduce = 0%, Cumulative CPU 1.33 sec
2013-08-05 18:37:22,733 Stage-1 map = 100%, reduce = 0%, Cumulative CPU 1.33 sec
2013-08-05 18:37:23,739 Stage-1 map = 100%, reduce = 50%, Cumulative CPU 3.1 sec
2013-08-05 18:37:24,745 Stage-1 map = 100%, reduce = 100%, Cumulative CPU 4.89 sec
2013-08-05 18:37:25,751 Stage-1 map = 100%, reduce = 100%, Cumulative CPU 4.89 sec
2013-08-05 18:37:26,757 Stage-1 map = 100%, reduce = 100%, Cumulative CPU 4.89 sec
MapReduce Total cumulative CPU time: 4 seconds 890 msec
Ended Job = job_201307151509_15499
Copying data to local directory /tmp/hivetest/distributeby
Copying data to local directory /tmp/hivetest/distributeby
7 Rows loaded to /tmp/hivetest/distributeby
MapReduce Jobs Launched:
Job 0: Map: 1 Reduce: 2 Cumulative CPU: 4.89 sec HDFS Read: 458 HDFS Write: 112 SUCCESS
Total MapReduce CPU Time Spent: 4 seconds 890 msec
OK
Time taken: 16.785 seconds
查看写入的查询数据:
结果说明:distribute by采用hash算法,将查询的结果写入不同的reduce文件中。数据分配到哪个reduce文件中,是在map端控制的。
--使用job_time进行排序且set mapred.reduce.tasks=2;
hive> set mapred.reduce.tasks=2;
hive> insert overwrite local directory '/tmp/hivetest/distributeby' select id,devid,job_time from tb_in_base where job_time=030729 distribute by job_time;
Total MapReduce jobs = 1
Launching Job 1 out of 1
Number of reduce tasks not specified. Defaulting to jobconf value of: 2
In order to change the average load for a reducer (in bytes):
set hive.exec.reducers.bytes.per.reducer=
In order to limit the maximum number of reducers:
set hive.exec.reducers.max=
In order to set a constant number of reducers:
set mapred.reduce.tasks=
Starting Job = job_201307151509_15500, Tracking URL = http://mwtec-50:50030/jobdetails.jsp jobid=job_201307151509_15500
Kill Command = /home/hadoop/hadoop-0.20.2/bin/hadoop job -Dmapred.job.tracker=mwtec-50:9002 -kill job_201307151509_15500
Hadoop job information for Stage-1: number of mappers: 1; number of reducers: 2
2013-08-05 18:42:07,764 Stage-1 map = 0%, reduce = 0%
2013-08-05 18:42:10,778 Stage-1 map = 100%, reduce = 0%, Cumulative CPU 0.61 sec
201
首页 上一页 4 5 6 7 8 9 10 下一页 尾页 7/10/10
】【打印繁体】【投稿】【收藏】 【推荐】【举报】【评论】 【关闭】 【返回顶部
分享到: 
上一篇Redis批量导入数据 下一篇如何优化单表大批量数据提取插入..

评论

帐  号: 密码: (新用户注册)
验 证 码:
表  情:
内  容:

·Libevent C++ 高并发 (2025-12-26 00:49:30)
·C++ dll 设计接口时 (2025-12-26 00:49:28)
·透彻理解 C 语言指针 (2025-12-26 00:22:52)
·C语言指针详解 (经典 (2025-12-26 00:22:49)
·C 指针 | 菜鸟教程 (2025-12-26 00:22:46)