原hive程序:

INSERT OVERWRITE TABLE MID_RES_LINK_LGC_BT PARTITION (

local_area_id = 20,

dt = '20130831'

) SELECT

LL.LGC_LINE_ID,

LL.RES_SPEC_ID,

T.LINK_ID,

LP2LA.PHY_EQP_ID,

PEA.RES_SPEC_ID,

LL2PA.PHY_PORT_ID,

PTPA.RES_SPEC_ID,

LP2LZ.PHY_EQP_ID,

PEZ.RES_SPEC_ID,

LL2PZ.PHY_PORT_ID,

PTPZ.RES_SPEC_ID,

T.A_LGC_EQP_ID,

LEA.RES_SPEC_ID,

T.A_LGC_PORT_ID,

LPA.RES_SPEC_ID,

T.Z_LGC_EQP_ID,

LEZ.RES_SPEC_ID,

T.Z_LGC_PORT_ID,

LPZ.RES_SPEC_ID

FROM

TC_RES_LNK_LGC_LINK_COMPONENT T

JOIN TC_RES_LGC_LINE LL ON (

T.LINE_ID = LL.LGC_LINE_ID

AND LL.LOCAL_AREA_ID = 20

AND LL.DT = '20130801'

AND LL.RES_SPEC_ID IN (

1744,

1745,

1749,

2929,

2931,

2941,

2943

)

AND T.LOCAL_AREA_ID = 20

AND T.DT = '20130801'

)

JOIN TC_RES_LGC_PHYEQP_2_LGCEQP LP2LA ON (

LP2LA.LGC_EQP_ID = T.A_LGC_EQP_ID

AND LP2LA.LOCAL_AREA_ID = 20

AND LP2LA.DT = '20130801'

AND T.LOCAL_AREA_ID = 20

AND T.DT = '20130801'

)

JOIN TC_RES_PHY_EQUIPMENT PEA ON (

LP2LA.PHY_EQP_ID = PEA.PHY_EQP_ID

AND LP2LA.LOCAL_AREA_ID = 20

AND LP2LA.DT = '20130801'

AND PEA.LOCAL_AREA_ID = 20

AND PEA.DT = '20130801'

)

JOIN TC_RES_LGC_LGCPORT_2_PHYPORT LL2PA ON (

LL2PA.LGC_PORT_ID = T.A_LGC_PORT_ID

AND LL2PA.LOCAL_AREA_ID = 20

AND LL2PA.DT = '20130801'

AND T.LOCAL_AREA_ID = 20

AND T.DT = '20130801'

)

JOIN TC_RES_PHY_TERM_PORT PTPA ON (

PTPA.PHY_PORT_ID = LL2PA.PHY_PORT_ID

AND PTPA.LOCAL_AREA_ID = 20

AND PTPA.DT = '20130801'

AND LL2PA.LOCAL_AREA_ID = 20

AND LL2PA.DT = '20130801'

)

JOIN TC_RES_LGC_PHYEQP_2_LGCEQP LP2LZ ON (

LP2LZ.LGC_EQP_ID = T.Z_LGC_EQP_ID

AND LP2LZ.LOCAL_AREA_ID = 20

AND LP2LZ.DT = '20130801'

AND T.LOCAL_AREA_ID = 20

AND T.DT = '20130801'

)

JOIN TC_RES_PHY_EQUIPMENT PEZ ON (

LP2LZ.PHY_EQP_ID = PEZ.PHY_EQP_ID

AND LP2LZ.LOCAL_AREA_ID = 20

AND LP2LZ.DT = '20130801'

AND PEZ.LOCAL_AREA_ID = 20

AND PEZ.DT = '20130801'

)

JOIN TC_RES_LGC_LGCPORT_2_PHYPORT LL2PZ ON (

LL2PZ.LGC_PORT_ID = T.Z_LGC_PORT_ID

AND LL2PZ.LOCAL_AREA_ID = 20

AND LL2PZ.DT = '20130801'

AND T.LOCAL_AREA_ID = 20

AND T.DT = '20130801'

)

JOIN TC_RES_PHY_TERM_PORT PTPZ ON (

PTPZ.PHY_PORT_ID = LL2PZ.PHY_PORT_ID

AND PTPZ.LOCAL_AREA_ID = 20

AND PTPZ.DT = '20130801'

AND LL2PZ.LOCAL_AREA_ID = 20

AND LL2PZ.DT = '20130801'

)

JOIN TC_RES_LGC_EQUIPMENT LEA ON (

LEA.LGC_EQP_ID = T.A_LGC_EQP_ID

AND LEA.LOCAL_AREA_ID = 20

AND LEA.DT = '20130801'

AND T.LOCAL_AREA_ID = 20

AND T.DT = '20130801'

)

JOIN TC_RES_LGC_PORT LPA ON (

LPA.LGC_PORT_ID = T.A_LGC_PORT_ID

AND LPA.LOCAL_AREA_ID = 20

AND LPA.DT = '20130801'

AND T.LOCAL_AREA_ID = 20

AND T.DT = '20130801'

)

JOIN TC_RES_LGC_EQUIPMENT LEZ ON (

LEZ.LGC_EQP_ID = T.Z_LGC_EQP_ID

AND LEA.LOCAL_AREA_ID = 20

AND LEA.DT = '20130801'

AND T.LOCAL_AREA_ID = 20

AND T.DT = '20130801'

)

JOIN TC_RES_LGC_PORT LPZ ON (

LPZ.LGC_PORT_ID = T.Z_LGC_PORT_ID

AND LPA.LOCAL_AREA_ID = 20

AND LPA.DT = '20130801'

AND T.LOCAL_AREA_ID = 20

AND T.DT = '20130801'

)

生成执行计划

STAGE DEPENDENCIES:

Stage-1 is a root stage

Stage-2 depends on stages: Stage-1

Stage-3 depends on stages: Stage-2

Stage-4 depends on stages: Stage-3

Stage-5 depends on stages: Stage-4

Stage-6 depends on stages: Stage-5

Stage-7 depends on stages: Stage-6

Stage-8 depends on stages: Stage-7

Stage-9 depends on stages: Stage-8

Stage-0 depends on stages: Stage-9

Stage-10 depends on stages: Stage-0

STAGE PLANS:

Stage: Stage-1

Map Reduce

Alias -> Map Operator Tree:

ll

TableScan

alias: ll

Filter Operator

predicate:

expr: (((local_area_id = 20.0) and (dt = '20130801')) and (res_spec_id) IN (1744, 1745, 1749, 2929, 2931, 2941, 2943))

type: boolean

Reduce Output Operator

key expressions:

expr: lgc_line_id

type: bigint

sort order: +

Map-reduce partition columns:

expr: lgc_line_id

type: bigint

tag: 1

value expressions:

expr: lgc_line_id

type: bigint

expr: res_spec_id

type: int

t

TableScan

alias: t

Filter Operator

predicate:

expr: ((local_area_id = 20.0) and (dt = '20130801'))

type: boolean

Reduce Output Operator

key expressions:

expr: line_id

type: bigint

sort order: +

Map-reduce partition columns:

expr: line_id

type: bigint

tag: 0

value expressions:

expr: link_id

type: bigint

expr: a_lgc_eqp_id

type: bigint

expr: a_lgc_port_id

type: bigint

expr: z_lgc_eqp_id

type: bigint

expr: z_lgc_port_id

type: bigint

expr: local_area_id

type: string

expr: dt

type: string

Reduce Operator Tree:

Join Operator

condition map:

Inner Join 0 to 1

condition expressions:

0 {VALUE._col1} {VALUE._col3} {VALUE._col4} {VALUE._col6} {VALUE._col7} {VALUE._col16} {VALUE._col17}

1 {VALUE._col0} {VALUE._col3}

handleSkewJoin: false

outputColumnNames: _col1, _col3, _col4, _col6, _col7, _col16, _col17, _col20, _col23

File Output Operator

compressed: false

GlobalTableId: 0

table:

input format: org.apache.hadoop.mapred.SequenceFileInputFormat

output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat

Stage: Stage-2

Map Reduce

Alias -> Map Operator Tree:

$INTNAME

Reduce Output Operator

key expressions:

expr: _col3

type: bigint

sort order: +

Map-reduce partition columns:

expr: _col3

type: bigint

tag: 0

value expressions:

expr: _col20

type: bigint

expr: _col23

type: int

expr: _col1

type: bigint

expr: _col3

type: bigint

expr: _col4

type: bigint

expr: _col6

type: bigint

expr: _col7

type: bigint

expr: _col16

type: string

expr: _col17

type: string

lea

TableScan

alias: lea

Filter Operator

predicate:

expr: ((local_area_id = 20.0) and (dt = '20130801'))

type: boolean

Reduce Output Operator

key expressions:

expr: lgc_eqp_id

type: bigint

sort order: +

Map-reduce partition columns:

expr: lgc_eqp_id

type: bigint

tag: 2

value expressions:

expr: res_spec_id

type: int

expr: local_area_id

type: string

expr: dt

type: string

lp2la

TableScan

alias: lp2la

Filter Operator

predicate:

expr: ((local_area_id = 20.0) and (dt = '20130801'))

type: boolean

Reduce Output Operator

key expressions:

expr: lgc_eqp_id

type: bigint

sort order: +

Map-reduce partition columns:

expr: lgc_eqp_id

type: bigint

tag: 1

value expressions:

expr: phy_eqp_id

type: bigint

expr: local_area_id

type: string

expr: dt

type: string

Reduce Operator Tree:

Join Operator

condition map:

Inner Join 0 to 1

Inner Join 0 to 2

condition expressions:

0 {VALUE._col21} {VALUE._col23} {VALUE._col24} {VALUE._col26} {VALUE._col27} {VALUE._col36} {VALUE._col37} {VALUE._col0} {VALUE._col3}

1 {VALUE._col1} {VALUE._col8} {VALUE._col9}

2 {VALUE._col4} {VALUE._col19} {VALUE._col20}

handleSkewJoin: false

outputColumnNames: _col1, _col3, _col4, _col6, _col7, _col16, _col17, _col20, _col23, _col41, _col48, _col49, _col56, _col71, _col72

File Output Operator

compressed: false

GlobalTableId: 0

table:

input format: org.apache.hadoop.mapred.SequenceFileInputFormat

output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat

Stage: Stage-3

Map Reduce

Alias -> Map Operator Tree:

$INTNAME

Reduce Output Operator

key expressions:

expr: _col41

type: bigint

sort order: +

Map-reduce partition columns:

expr: _col41

type: bigint

tag: 0

value expressions:

expr: _col20

type: bigint

expr: _col23

type: int

expr: _col1

type: bigint

expr: _col3

type: bigint

expr: _col4

type: bigint

expr: _col6

type: bigint

expr: _col7

type: bigint

expr: _col16

type: string

expr: _col17

type: string

expr: _col56

type: int

expr: _col71

type: string

expr: _col72

type: string

expr: _col41

type: bigint

pea

TableScan

alias: pea

Filter Operator

predicate:

expr: ((local_area_id = 20.0) and (dt = '20130801'))

type: boolean

Reduce Output Operator

key expressions:

expr: phy_eqp_id

type: bigint

sort order: +

Map-reduce partition columns:

expr: phy_eqp_id

type: bigint

tag: 1

value expressions:

expr: res_spec_id

type: int

Reduce Operator Tree:

Join Operator

condition map:

Inner Join 0 to 1

condition expressions:

0 {VALUE._col21} {VALUE._col23} {VALUE._col24} {VALUE._col26} {VALUE._col27} {VALUE._col36} {VALUE._col37} {VALUE._col0} {VALUE._col3} {VALUE._col64} {VALUE._col44} {VALUE._col59} {VALUE._col60}

1 {VALUE._col3}

handleSkewJoin: false

outputColumnNames: _col1, _col3, _col4, _col6, _col7, _col16, _col17, _col20, _col23, _col41, _col56, _col71, _col72, _col78

File Output Operator

compressed: false

GlobalTableId: 0

table:

input format: org.apache.hadoop.mapred.SequenceFileInputFormat

output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat

Stage: Stage-4

Map Reduce

Alias -> Map Operator Tree:

$INTNAME

Reduce Output Operator

key expressions:

expr: _col4

type: bigint

sort order: +

Map-reduce partition columns:

expr: _col4

type: bigint

tag: 0

value expressions:

expr: _col20

type: bigint

expr: _col23

type: int

expr: _col1

type: bigint

expr: _col3

type: bigint

expr: _col4

type: bigint

expr: _col6

type: bigint

expr: _col7

type: bigint

expr: _col16

type: string

expr: _col17

type: string

expr: _col78

type: int

expr: _col56

type: int

expr: _col71

type: string

expr: _col72

type: string

expr: _col41

type: bigint

ll2pa

TableScan

alias: ll2pa

Filter Operator

predicate:

expr: ((local_area_id = 20.0) and (dt = '20130801'))

type: boolean

Reduce Output Operator

key expressions:

expr: lgc_port_id

type: bigint

sort order: +

Map-reduce partition columns:

expr: lgc_port_id

type: bigint

tag: 1

value expressions:

expr: phy_port_id

type: bigint

expr: local_area_id

type: string

expr: dt

type: string

lpa

TableScan

alias: lpa

Filter Operator

predicate:

expr: ((local_area_id = 20.0) and (dt = '20130801'))

type: boolean

Reduce Output Operator

key expressions:

expr: lgc_port_id

type: bigint

sort order: +

Map-reduce partition columns:

expr: lgc_port_id

type: bigint

tag: 2

value expressions:

expr: res_spec_id

type: int

expr: local_area_id

type: string

expr: dt

type: string

Reduce Operator Tree:

Join Operator

condition map:

Inner Join 0 to 1

Inner Join 0 to 2

condition expressions:

0 {VALUE._col21} {VALUE._col23} {VALUE._col24} {VALUE._col26} {VALUE._col27} {VALUE._col36} {VALUE._col37} {VALUE._col0} {VALUE._col3} {VALUE._col43} {VALUE._col114} {VALUE._col94} {VALUE._col109} {VALUE._col110}

1 {VALUE._col1} {VALUE._col8} {VALUE._col9}

2 {VALUE._col4} {VALUE._col19} {VALUE._col20}

handleSkewJoin: false

outputColumnNames: _col1, _col3, _col4, _col6, _col7, _col16, _col17, _col20, _col23, _col43, _col91, _col106, _col121, _col122, _col126, _col133, _col134, _col141, _col156, _col157

File Output Operator

compressed: false

GlobalTableId: 0

table:

input format: org.apache.hadoop.mapred.SequenceFileInputFormat

output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat

Stage: Stage-5

Map Reduce

Alias -> Map Operator Tree:

$INTNAME

Reduce Output Operator

key expressions:

expr: _col126

type: bigint

sort order: +

Map-reduce partition columns:

expr: _col126

type: bigint

tag: 0

value expressions:

expr: _col20

type: bigint

expr: _col23

type: int

expr: _col1

type: bigint

expr: _col3

type: bigint

expr: _col4

type: bigint

expr: _col6

type: bigint

expr: _col7

type: bigint

expr: _col16

type: string

expr: _col17

type: string

expr: _col43

type: int

expr: _col141

type: int

expr: _col156

type: string

expr: _col157

type: string

expr: _col126

type: bigint

expr: _col106

type: int

expr: _col121

type: string

expr: _col122

type: string

expr: _col91

type: bigint

ptpa

TableScan

alias: ptpa

Filter Operator

predicate:

expr: ((local_area_id = 20.0) and (dt = '20130801'))

type: boolean

Reduce Output Operator

key expressions:

expr: phy_port_id

type: bigint

sort order: +

Map-reduce partition columns:

expr: phy_port_id

type: bigint

tag: 1

value expressions:

expr: res_spec_id

type: int

Reduce Operator Tree:

Join Operator

condition map:

Inner Join 0 to 1

condition expressions:

0 {VALUE._col21} {VALUE._col23} {VALUE._col24} {VALUE._col26} {VALUE._col27} {VALUE._col36} {VALUE._col37} {VALUE._col0} {VALUE._col3} {VALUE._col43} {VALUE._col94} {VALUE._col109} {VALUE._col110} {VALUE._col114} {VALUE._col149} {VALUE._col129} {VALUE._col144} {VALUE._col145}

1 {VALUE._col2}

handleSkewJoin: false

outputColumnNames: _col1, _col3, _col4, _col6, _col7, _col16, _col17, _col20, _col23, _col43, _col94, _col109, _col110, _col114, _col126, _col141, _col156, _col157, _col162

File Output Operator

compressed: false

GlobalTableId: 0

table:

input format: org.apache.hadoop.mapred.SequenceFileInputFormat

output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat

Stage: Stage-6

Map Reduce

Alias -> Map Operator Tree:

$INTNAME

Reduce Output Operator

key expressions:

expr: _col6

type: bigint

sort order: +

Map-reduce partition columns:

expr: _col6

type: bigint

tag: 0

value expressions:

expr: _col20

type: bigint

expr: _col23

type: int

expr: _col1

type: bigint

expr: _col3

type: bigint

expr: _col4

type: bigint

expr: _col6

type: bigint

expr: _col7

type: bigint

expr: _col16

type: string

expr: _col17

type: string

expr: _col162

type: int

expr: _col43

type: int

expr: _col94

type: int

expr: _col109

type: string

expr: _col110

type: string

expr: _col114

type: bigint

expr: _col141

type: int

expr: _col126

type: bigint

lez

TableScan

alias: lez

Reduce Output Operator

key expressions:

expr: lgc_eqp_id

type: bigint

sort order: +

Map-reduce partition columns:

expr: lgc_eqp_id

type: bigint

tag: 2

value expressions:

expr: res_spec_id

type: int

lp2lz

TableScan

alias: lp2lz

Filter Operator

predicate:

expr: ((local_area_id = 20.0) and (dt = '20130801'))

type: boolean

Reduce Output Operator

key expressions:

expr: lgc_eqp_id

type: bigint

sort order: +

Map-reduce partition columns:

expr: lgc_eqp_id

type: bigint

tag: 1

value expressions:

expr: phy_eqp_id

type: bigint

expr: local_area_id

type: string

expr: dt

type: string

Reduce Operator Tree:

Join Operator

condition map:

Inner Join 0 to 1

Inner Join 0 to 2

condition expressions:

0 {VALUE._col21} {VALUE._col23} {VALUE._col24} {VALUE._col26} {VALUE._col27} {VALUE._col36} {VALUE._col37} {VALUE._col0} {VALUE._col3} {VALUE._col75} {VALUE._col42} {VALUE._col126} {VALUE._col141} {VALUE._col142} {VALUE._col146} {VALUE._col181} {VALUE._col161}

1 {VALUE._col1} {VALUE._col8} {VALUE._col9}

2 {VALUE._col4}

handleSkewJoin: false

outputColumnNames: _col1, _col3, _col4, _col6, _col7, _col16, _col17, _col20, _col23, _col43, _col92, _col126, _col141, _col142, _col146, _col158, _col173, _col193, _col200, _col201, _col208

File Output Operator

compressed: false

GlobalTableId: 0

table:

input format: org.apache.hadoop.mapred.SequenceFileInputFormat

output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat

Stage: Stage-7

Map Reduce

Alias -> Map Operator Tree:

$INTNAME

Reduce Output Operator

key expressions:

expr: _col193

type: bigint

sort order: +

Map-reduce partition columns:

expr: _col193

type: bigint

tag: 0

value expressions:

expr: _col20

type: bigint

expr: _col23

type: int

expr: _col1

type: bigint

expr: _col3

type: bigint

expr: _col4

type: bigint

expr: _col6

type: bigint

expr: _col7

type: bigint

expr: _col16

type: string

expr: _col17

type: string

expr: _col92

type: int

expr: _col43

type: int

expr: _col126

type: int

expr: _col141

type: string

expr: _col142

type: string

expr: _col146

type: bigint

expr: _col173

type: int

expr: _col158

type: bigint

expr: _col193

type: bigint

expr: _col208

type: int

pez

TableScan

alias: pez

Filter Operator

predicate:

expr: ((local_area_id = 20.0) and (dt = '20130801'))

type: boolean

Reduce Output Operator

key expressions:

expr: phy_eqp_id

type: bigint

sort order: +

Map-reduce partition columns:

expr: phy_eqp_id

type: bigint

tag: 1

value expressions:

expr: res_spec_id

type: int

Reduce Operator Tree:

Join Operator

condition map:

Inner Join 0 to 1

condition expressions:

0 {VALUE._col21} {VALUE._col23} {VALUE._col24} {VALUE._col26} {VALUE._col27} {VALUE._col36} {VALUE._col37} {VALUE._col0} {VALUE._col3} {VALUE._col75} {VALUE._col42} {VALUE._col126} {VALUE._col141} {VALUE._col142} {VALUE._col146} {VALUE._col181} {VALUE._col161} {VALUE._col193} {VALUE._col208}

1 {VALUE._col3}

handleSkewJoin: false

outputColumnNames: _col1, _col3, _col4, _col6, _col7, _col16, _col17, _col20, _col23, _col43, _col92, _col126, _col141, _col142, _col146, _col158, _col173, _col193, _col208, _col230

File Output Operator

compressed: false

GlobalTableId: 0

table:

input format: org.apache.hadoop.mapred.SequenceFileInputFormat

output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat

Stage: Stage-8

Map Reduce

Alias -> Map Operator Tree:

$INTNAME

Reduce Output Operator

key expressions:

expr: _col7

type: bigint

sort order: +

Map-reduce partition columns:

expr: _col7

type: bigint

tag: 0

value expressions:

expr: _col20

type: bigint

expr: _col23

type: int

expr: _col1

type: bigint

expr: _col3

type: bigint

expr: _col4

type: bigint

expr: _col6

type: bigint

expr: _col7

type: bigint

expr: _col92

type: int

expr: _col43

type: int

expr: _col126

type: int

expr: _col146

type: bigint

expr: _col173

type: int

expr: _col158

type: bigint

expr: _col193

type: bigint

expr: _col208

type: int

expr: _col230

type: int

ll2pz

TableScan

alias: ll2pz

Filter Operator

predicate:

expr: ((local_area_id = 20.0) and (dt = '20130801'))

type: boolean

Reduce Output Operator

key expressions:

expr: lgc_port_id

type: bigint

sort order: +

Map-reduce partition columns:

expr: lgc_port_id

type: bigint

tag: 1

value expressions:

expr: phy_port_id

type: bigint

expr: local_area_id

type: string

expr: dt

type: string

lpz

TableScan

alias: lpz

Reduce Output Operator

key expressions:

expr: lgc_port_id

type: bigint

sort order: +

Map-reduce partition columns:

expr: lgc_port_id

type: bigint

tag: 2

value expressions:

expr: res_spec_id

type: int

Reduce Operator Tree:

Join Operator

condition map:

Inner Join 0 to 1

Inner Join 0 to 2

condition expressions:

0 {VALUE._col21} {VALUE._col23} {VALUE._col24} {VALUE._col26} {VALUE._col27} {VALUE._col0} {VALUE._col3} {VALUE._col75} {VALUE._col42} {VALUE._col126} {VALUE._col146} {VALUE._col181} {VALUE._col161} {VALUE._col193} {VALUE._col208} {VALUE._col230}

1 {VALUE._col1} {VALUE._col8} {VALUE._col9}

2 {VALUE._col4}

handleSkewJoin: false

outputColumnNames: _col1, _col3, _col4, _col6, _col7, _col20, _col23, _col43, _col92, _col126, _col146, _col158, _col173, _col193, _col208, _col230, _col278, _col285, _col286, _col293

File Output Operator

compressed: false

GlobalTableId: 0

table:

input format: org.apache.hadoop.mapred.SequenceFileInputFormat

output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat

Stage: Stage-9

Map Reduce

Alias -> Map Operator Tree:

$INTNAME

Reduce Output Operator

key expressions:

expr: _col278

type: bigint

sort order: +

Map-reduce partition columns:

expr: _col278

type: bigint

tag: 0

value expressions:

expr: _col20

type: bigint

expr: _col23

type: int

expr: _col43

type: int

expr: _col146

type: bigint

expr: _col193

type: bigint

expr: _col230

type: int

expr: _col1

type: bigint

expr: _col3

type: bigint

expr: _col4

type: bigint

expr: _col6

type: bigint

expr: _col7

type: bigint

expr: _col92

type: int

expr: _col126

type: int

expr: _col158

type: bigint

expr: _col173

type: int

expr: _col278

type: bigint

expr: _col208

type: int

expr: _col293

type: int

ptpz

TableScan

alias: ptpz

Filter Operator

predicate:

expr: ((local_area_id = 20.0) and (dt = '20130801'))

type: boolean

Reduce Output Operator

key expressions:

expr: phy_port_id

type: bigint

sort order: +

Map-reduce partition columns:

expr: phy_port_id

type: bigint

tag: 1

value expressions:

expr: res_spec_id

type: int

Reduce Operator Tree:

Join Operator

condition map:

Inner Join 0 to 1

condition expressions:

0 {VALUE._col0} {VALUE._col3} {VALUE._col23} {VALUE._col71} {VALUE._col83} {VALUE._col97} {VALUE._col145} {VALUE._col147} {VALUE._col148} {VALUE._col150} {VALUE._col151} {VALUE._col166} {VALUE._col200} {VALUE._col220} {VALUE._col235} {VALUE._col255} {VALUE._col270} {VALUE._col293}

1 {VALUE._col2}

handleSkewJoin: false

outputColumnNames: _col0, _col3, _col23, _col71, _col83, _col97, _col145, _col147, _col148, _col150, _col151, _col166, _col200, _col220, _col235, _col255, _col270, _col293, _col314

Select Operator

expressions:

expr: _col0

type: bigint

expr: _col3

type: int

expr: _col145

type: bigint

expr: _col220

type: bigint

expr: _col23

type: int

expr: _col71

type: bigint

expr: _col166

type: int

expr: _col83

type: bigint

expr: _col97

type: int

expr: _col255

type: bigint

expr: _col314

type: int

expr: _col147

type: bigint

expr: _col235

type: int

expr: _col148

type: bigint

expr: _col200

type: int

expr: _col150

type: bigint

expr: _col270

type: int

expr: _col151

type: bigint

expr: _col293

type: int

outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18

Select Operator

expressions:

expr: _col0

type: bigint

expr: UDFToLong(_col1)

type: bigint

expr: _col2

type: bigint

expr: _col3

type: bigint

expr: UDFToLong(_col4)

type: bigint

expr: _col5

type: bigint

expr: UDFToLong(_col6)

type: bigint

expr: _col7

type: bigint

expr: UDFToLong(_col8)

type: bigint

expr: _col9

type: bigint

expr: UDFToLong(_col10)

type: bigint

expr: _col11

type: bigint

expr: UDFToLong(_col12)

type: bigint

expr: _col13

type: bigint

expr: UDFToLong(_col14)

type: bigint

expr: _col15

type: bigint

expr: UDFToLong(_col16)

type: bigint

expr: _col17

type: bigint

expr: UDFToLong(_col18)

type: bigint

outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18

File Output Operator

compressed: false

GlobalTableId: 1

table:

input format: org.apache.hadoop.mapred.TextInputFormat

output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat

serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

name: default.mid_res_link_lgc_bt

Stage: Stage-0

Move Operator

tables:

partition:

dt 20130831

local_area_id 20

replace: true

table:

input format: org.apache.hadoop.mapred.TextInputFormat

output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat

serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe

name: default.mid_res_link_lgc_bt

Stage: Stage-10

Stats-Aggr Operator

查看任务执行,有Map数2个,1个reduce。错误发生在map阶段,两个map阶段,错误后重新执行4次,均失败。

进入查看详细的每次task运行出错日志

2013-09-11 09:42:44,392 WARN org.apache.hadoop.util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable

2013-09-11 09:42:44,774 WARN mapreduce.Counters: Group org.apache.hadoop.mapred.Task$Counter is deprecated. Use org.apache.hadoop.mapreduce.TaskCounter instead

2013-09-11 09:42:45,257 INFO org.apache.hadoop.mapred.TaskRunner: Creating symlink: /app/tmp/hdfs/tmp/mapred/local/taskTracker/distcache/-1273717043127285981_-158808503_179256813/tascluster/tmp/hive-hadoop/hive_2013-09-11_09-42-32_079_3962951880312773437/-mr-10011/4659e5b9-e5ba-4a26-9a81-c82fbdbf0af3 <- /app/tmp/hdfs/tmp/mapred/local/taskTracker/hadoop/jobcache/job_201309102223_0001/attempt_201309102223_0001_m_000000_0/work/HIVE_PLAN4659e5b9-e5ba-4a26-9a81-c82fbdbf0af3

2013-09-11 09:42:45,279 INFO org.apache.hadoop.filecache.TrackerDistributedCacheManager: Creating symlink: /app/tmp/hdfs/tmp/mapred/local/taskTracker/hadoop/jobcache/job_201309102223_0001/jars/.job.jar.crc <- /app/tmp/hdfs/tmp/mapred/local/taskTracker/hadoop/jobcache/job_201309102223_0001/attempt_201309102223_0001_m_000000_0/work/.job.jar.crc

2013-09-11 09:42:45,292 INFO org.apache.hadoop.filecache.TrackerDistributedCacheManager: Creating symlink: /app/tmp/hdfs/tmp/mapred/local/taskTracker/hadoop/jobcache/job_201309102223_0001/jars/job.jar <- /app/tmp/hdfs/tmp/mapred/local/taskTracker/hadoop/jobcache/job_201309102223_0001/attempt_201309102223_0001_m_000000_0/work/job.jar

2013-09-11 09:42:45,362 WARN org.apache.hadoop.conf.Configuration: session.id is deprecated. Instead, use dfs.metrics.session-id

2013-09-11 09:42:45,364 INFO org.apache.hadoop.metrics.jvm.JvmMetrics: Initializing JVM Metrics with processName=MAP, sessionId=

2013-09-11 09:42:45,731 INFO org.apache.hadoop.util.ProcessTree: setsid exited with exit code 0

2013-09-11 09:42:45,760 INFO org.apache.hadoop.mapred.Task:  Using ResourceCalculatorPlugin : org.apache.hadoop.util.LinuxResourceCalculatorPlugin@5b32627

2013-09-11 09:42:46,050 INFO org.apache.hadoop.mapred.MapTask: Processing split: Paths:/tmp/hive-hadoop/hive_2013-09-11_09-42-32_079_3962951880312773437/-mr-10010/1/emptyFile:0+0InputFormatClass: org.apache.hadoop.mapred.TextInputFormat

2013-09-11 09:42:46,087 WARN org.apache.hadoop.hive.conf.HiveConf: hive-site.xml not found on CLASSPATH

2013-09-11 09:42:46,371 INFO org.apache.hadoop.hive.ql.io.HiveContextAwareRecordReader: Processing file hdfs://tascluster/tmp/hive-hadoop/hive_2013-09-11_09-42-32_079_3962951880312773437/-mr-10010/1/emptyFile

2013-09-11 09:42:46,372 WARN mapreduce.Counters: Counter name MAP_INPUT_BYTES is deprecated. Use FileInputFormatCounters as group name and  BYTES_READ as counter name instead

2013-09-11 09:42:46,375 INFO org.apache.hadoop.mapred.MapTask: numReduceTasks: 1

2013-09-11 09:42:46,380 INFO org.apache.hadoop.mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

2013-09-11 09:42:46,383 INFO org.apache.hadoop.mapred.MapTask: io.sort.mb = 100

2013-09-11 09:42:46,642 INFO org.apache.hadoop.mapred.TaskLogsTruncater: Initializing logs' truncater with mapRetainSize=-1 and reduceRetainSize=-1

2013-09-11 09:42:46,678 FATAL org.apache.hadoop.mapred.Child: Error running child : java.lang.OutOfMemoryError: Java heap space

at org.apache.hadoop.mapred.MapTask$MapOutputBuffer.init(MapTask.java:827)

at org.apache.hadoop.mapred.MapTask.createSortingCollector(MapTask.java:377)

at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:407)

at org.apache.hadoop.mapred.MapTask.run(MapTask.java:333)

at org.apache.hadoop.mapred.Child$4.run(Child.java:268)

at java.security.AccessController.doPrivileged(Native Method)

at javax.security.auth.Subject.doAs(Subject.java:416)

at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1408)

at org.apache.hadoop.mapred.Child.main(Child.java:262)

看到最后的错误日志应该是mapred.Child的内存太小了,所以加大task的内存限制。

修改 mapred.child.java.opts 参数,原来设置的是100M,该为200。问题解决

转载于:https://www.cnblogs.com/banrixian/p/3314498.html

Hive 程序内存溢出错误分析相关推荐

  1. 下面的一堆文字为了说明一件事情---.NET程序,内存溢出,如何控制.

    ========下面的一堆文字为了说明一件事情---.NET程序,内存溢出,如何控制.飞秋2010下载官网主要是堆HEAP大小如何控制以及优化.以减轻GC突发性负担及这个时候服务器当机的可能*. 对于 ...

  2. 利用VC检测程序内存溢出(转)

    VisualC++没有默认启动内存泄露检测,即如果某段代码产生内存溢出也不会在"输出窗口"调试标签下输出内存溢出相关信息 (1)需要手工添加代码检测 #define _CRTDBG ...

  3. 总结Java程序内存溢出原因

    目录 内存溢出和内存泄漏 直接内存溢出 堆溢出 方法区溢出 这篇日志总结下Java程序中的发生内存溢出的一些原因,我们知道JVM堆空间十分重要,大部分对象在创建时都是放在堆中(除了一些逃逸对象是栈上分 ...

  4. 完美收官!java程序内存溢出找原因

    原理讲解前,先看一个最经典的业务场景,如开发一个电商网站,要实现支付订单的功能,流程如下: 创建一个订单之后,如果用户立刻支付了这个订单,我们需要将订单状态更新为"已支付" 扣减相 ...

  5. java调用dll报内存溢出错误或者程序崩溃

    因为业务需要,需要用Java去调用dll动态库来分析图片,结果调用后导致程序内存溢出,或者直接崩溃:再经过反复检查后,发现本身的我写的java代码并没有问题,最终去验证图片的时候: 1.图片访问不到, ...

  6. 为什么Go中有的自定义error会导致内存溢出

    分享一个在go tour上看到的练习题,练习里要求用户自己定义一个错误类型,实现error接口,函数在参数不满足条件的时候返回自定义的错误类型的值.练习中特别提示用户不要在实现的Error方法里直接使 ...

  7. 为什么Go的自定义error有时候会内存溢出

    分享一个在go tour上看到的练习题,练习里要求用户自己定义一个错误类型,实现 error接口,函数在参数不满足条件的时候返回自定义的错误类型的值.练习中特别提示用户不要在实现的 Error方法里直 ...

  8. Hive内存溢出常见问题

    1 MR作业运行过程中内存溢出错误分类 1.1 Mapper/Reducer阶段JVM内存溢出(一般都是堆) 1.1.1 JVM堆(Heap)内存溢出:堆内存不足时,一般会抛出如下异常: 1)-&qu ...

  9. HIVE 内存溢出常见问题

    MapReduce作业运行过程中内存溢出错误分类 1.Mapper/Reducer阶段JVM内存溢出(一般都是堆) 1)JVM堆(Heap)内存溢出:堆内存不足时,一般会抛出如下异常: 第一种:&qu ...

最新文章

  1. datetime 索引_date类型和varchar类型在查询索引效率上比较
  2. QT精彩实例分析第5章-0
  3. 为了OFFER而战,大四那些日子在牛客网和Leetcode刷SQL题目
  4. 互联网产品之百万级高并发技术整体架构
  5. android targetapi23,Android Studio 3.0+上的Kapt问题(TargetApi)
  6. testng_TestNG注释
  7. xp系统做无盘服务器,锐起无盘网吧系统无盘XP系统特点
  8. url重写(urlrewrite.xml)
  9. 消防信号二总线有没电压_消防的电源总线和二总线的区别,二总线是不是信号线......
  10. 实践秘籍:Boosting 与 AdaBoost
  11. THREE实战2_正交投影相机与透视相机
  12. 【Vue】实现出生日期计算年龄
  13. [京东备注插旗接口]-拼多多商家如何做好直播,直播必备小技巧
  14. 蓝桥杯李白打酒php,蓝桥杯:李白打酒
  15. 9个妙招教你玩转微信
  16. 远程桌面优化避坑指南
  17. OpenCV 32F 与 8U Mat数据类型相互转换(C++版)
  18. Windows技能(二):有效清理C盘垃圾的一些使用技巧 | C盘垃圾爆满
  19. 云计算大数据学习路线课程大纲资料:marpreduce快速入门
  20. cd——Linux王国中的导游_莫韵乐的linux王国英雄传

热门文章

  1. WordPress如何调用其他网站的最新文章
  2. 图解:HTTP 范围请求,助力断点续传、多线程下载的核心原理
  3. VR消防灭火模拟训练系统
  4. 自动挡驾照后上路学习总结
  5. python培训实习报告
  6. python - one day
  7. 服务器系统没网灯在闪,电脑没网交换机灯全部一闪一闪的解决方法
  8. ThreadLocalMap中为什么valu是强引用,不是弱引用
  9. LeetCode 342. 4的幂 附带《Effective C++》pdf电子书
  10. 电子计算机应用是啥,第二代电子计算机使用的电子元件是什么