Hello All,
i have one issue with hadoop, every day job runs based on query, the query i am using it is getting 3 years data but we are looking for 6 years data.
the query is:
SELECT account.account,
denorm.account_id,
denorm.incident_number,
denorm.incident_id,
denorm.casenumber,
denorm.incident_type,
denorm.incident_status,
denorm.comm_pref_code,
denorm.complexity,
denorm.current_severity,
denorm.initial_severity,
denorm.max_severity,
denorm.bug_cnt,
denorm.outage,
denorm.initial_portfolio_name,
denorm.entry_channel,
denorm.creation_date,
denorm.closed_date,
denorm.current_serial_number,
denorm.router_node_name,
denorm.summary,
denorm.customer_ticket_number,
denorm.incident_contact_email,
denorm.problem_code,
denorm.resolution_code,
denorm.sr_create_pfg,
denorm.install_at_site_id,
denorm.solution_release,
denorm.nlp_status,
denorm.b2b_flag,
denorm.install_at_site_key,
denorm.portfolio_number,
denorm.portfolio_desc,
denorm.contact_party_name,
denorm.contact_details,
denorm.org_party_name,
denorm.cco_id,
denorm.contract_number,
denorm.contract_service_line,
denorm.contract_line_status,
denorm.coverage_template_desc,
denorm.contract_start_date,
denorm.contract_end_date,
denorm.contract_expire_date,
denorm.tech_name,
denorm.hw_part_number,
denorm.hw_family,
denorm.hw_platform,
denorm.hw_business_unit,
denorm.sw_part_number,
denorm.sw_version,
denorm.sw_part_type,
denorm.sw_business_unit,
denorm.sw_family,
denorm.producttable_item_name,
denorm.producttable_item_description,
denorm.producttable_business_unit,
denorm.producttable_family,
denorm.bl_last_update_date,
denorm.sub_tech_name,
denorm.change_done_by_cco_id
FROM
"csp_tsbi.csp_tss_incidents curated_input account inner join service_request_transformed_tsbi.sr_denorm_incidents denorm on account.contract=denorm.contract_number where coalesce(to_date(closed_date), to_date(from_unixtime(unix_timestamp()))) between date_sub(to_date(from_unixtime( unix_timestamp() ) ) ,1095) and to_date(from_unixtime(unix_timestamp()))";
Now i am trying for 6 years so i have changed to 2190 for 6 years instead of 1095 but when i run the job, it is geting failed when exporting data to 1 server.
please see the error below. I would really appreciate if anyone can guide on this. Thanks in adv
LogType:stderr
Log Upload Time:Tue Sep 03 05:15:51 -0400 2019
LogLength:0
Log Contents:
End of LogType:stderr
LogType:stdout
Log Upload Time:Tue Sep 03 05:15:51 -0400 2019
LogLength:0
Log Contents:
End of LogType:stdout
LogType:syslog
Log Upload Time:Tue Sep 03 05:15:51 -0400 2019
LogLength:10480
Log Contents:
2019-09-03 02:05:22,625 INFO [main] org.apache.hadoop.metrics2.impl.MetricsConfig: loaded properties from hadoop-metrics2.properties
2019-09-03 02:05:22,678 INFO [main] org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Scheduled snapshot period at 10 second(s).
2019-09-03 02:05:22,679 INFO [main] org.apache.hadoop.metrics2.impl.MetricsSystemImpl: MapTask metrics system started
2019-09-03 02:05:22,680 INFO [main] org.apache.hadoop.mapred.YarnChild: Executing with tokens:
2019-09-03 02:05:22,680 INFO [main] org.apache.hadoop.mapred.YarnChild: Kind: mapreduce.job, Service: job_1563651888010_2140784, Ident: (org.apache.hadoop.mapreduce.security.token.JobTokenIdentifier@30bce90b)
2019-09-03 02:05:22,764 INFO [main] org.apache.hadoop.mapred.YarnChild: Sleeping for 0ms before retrying again. Got null now.
2019-09-03 02:05:23,119 INFO [main] org.apache.hadoop.mapred.YarnChild: mapreduce.cluster.local.dir for child: /hdfs/app/local.hdprd-c01-r06-07.cisco.com.logs/usercache/phodisvc/appcache/application_1563651888010_2140784
2019-09-03 02:05:23,245 INFO [main] com.pepperdata.supervisor.agent.resource.O: Set a new configuration for the first time.
2019-09-03 02:05:23,330 INFO [main] com.pepperdata.common.reflect.d: Method not implemented in this version of Hadoop: org.apache.hadoop.fs.FileSystem.getGlobalStorageStatistics
2019-09-03 02:05:23,330 INFO [main] com.pepperdata.common.reflect.d: Method not implemented in this version of Hadoop: org.apache.hadoop.fs.FileSystem$Statistics.getBytesReadLocalHost
2019-09-03 02:05:23,344 INFO [main] com.pepperdata.supervisor.agent.resource.u: Scheduling statistics report every 2000 millisecs
2019-09-03 02:05:23,491 INFO [Pepperdata Statistics Reporter] com.pepperdata.supervisor.protocol.handler.http.Handler: Shuffle URL path prefix: /mapOutput
2019-09-03 02:05:23,491 INFO [Pepperdata Statistics Reporter] com.pepperdata.supervisor.protocol.handler.http.Handler: Initialized shuffle handler, starting uncontrolled.
2019-09-03 02:05:23,519 INFO [main] org.apache.hadoop.mapred.Task: mapOutputFile class: org.apache.hadoop.mapred.MapRFsOutputFile
2019-09-03 02:05:23,519 INFO [main] org.apache.hadoop.conf.Configuration.deprecation: session.id is deprecated. Instead, use dfs.metrics.session-id
2019-09-03 02:05:23,544 INFO [main] org.apache.hadoop.mapred.Task: Using ResourceCalculatorProcessTree : [ ]
2019-09-03 02:05:23,670 INFO [main] org.apache.hadoop.mapred.MapTask: Processing split: Paths:/app/SmartAnalytics/Apps/CSP/hivewarehouse/csp_tsbi.db/sr_passim_data_platform_mood_db_p01/000000_0:0+30277399
2019-09-03 02:05:23,674 INFO [main] org.apache.hadoop.conf.Configuration.deprecation: map.input.file is deprecated. Instead, use mapreduce.map.input.file
2019-09-03 02:05:23,674 INFO [main] org.apache.hadoop.conf.Configuration.deprecation: map.input.start is deprecated. Instead, use mapreduce.map.input.start
2019-09-03 02:05:23,675 INFO [main] org.apache.hadoop.conf.Configuration.deprecation: map.input.length is deprecated. Instead, use mapreduce.map.input.length
2019-09-03 02:05:26,449 WARN [Thread-12] org.apache.sqoop.mapreduce.SQLServerExportDBExecThread: Error executing statement: java.sql.BatchUpdateException: String or binary data would be truncated.
2019-09-03 02:05:26,450 WARN [Thread-12] org.apache.sqoop.mapreduce.SQLServerAsyncDBExecThread: Trying to recover from DB write failure:
java.sql.BatchUpdateException: String or binary data would be truncated.
at com.microsoft.sqlserver.jdbc.SQLServerPreparedStatement.executeBatch(SQLServerPreparedStatement.java:1178)
at org.apache.sqoop.mapreduce.SQLServerExportDBExecThread.executeStatement(SQLServerExportDBExecThread.java:96)
at org.apache.sqoop.mapreduce.SQLServerAsyncDBExecThread.write(SQLServerAsyncDBExecThread.java:272)
at org.apache.sqoop.mapreduce.SQLServerAsyncDBExecThread.run(SQLServerAsyncDBExecThread.java:240)
2019-09-03 02:05:26,451 WARN [Thread-12] org.apache.sqoop.mapreduce.db.SQLServerConnectionFailureHandler: Cannot handle error with SQL State: 22001
2019-09-03 02:05:26,451 ERROR [Thread-12] org.apache.sqoop.mapreduce.SQLServerAsyncDBExecThread: Failed to write records.
java.io.IOException: Registered handler cannot recover error with SQL State: 22001, error code: 8152
at org.apache.sqoop.mapreduce.SQLServerAsyncDBExecThread.write(SQLServerAsyncDBExecThread.java:293)
at org.apache.sqoop.mapreduce.SQLServerAsyncDBExecThread.run(SQLServerAsyncDBExecThread.java:240)
Caused by: java.sql.BatchUpdateException: String or binary data would be truncated.
at com.microsoft.sqlserver.jdbc.SQLServerPreparedStatement.executeBatch(SQLServerPreparedStatement.java:1178)
at org.apache.sqoop.mapreduce.SQLServerExportDBExecThread.executeStatement(SQLServerExportDBExecThread.java:96)
at org.apache.sqoop.mapreduce.SQLServerAsyncDBExecThread.write(SQLServerAsyncDBExecThread.java:272)
... 1 more
2019-09-03 02:05:26,452 ERROR [Thread-12] org.apache.sqoop.mapreduce.SQLServerAsyncDBExecThread: Got exception in update thread: java.io.IOException: Registered handler cannot recover error with SQL State: 22001, error code: 8152
at org.apache.sqoop.mapreduce.SQLServerAsyncDBExecThread.write(SQLServerAsyncDBExecThread.java:293)
at org.apache.sqoop.mapreduce.SQLServerAsyncDBExecThread.run(SQLServerAsyncDBExecThread.java:240)
Caused by: java.sql.BatchUpdateException: String or binary data would be truncated.
at com.microsoft.sqlserver.jdbc.SQLServerPreparedStatement.executeBatch(SQLServerPreparedStatement.java:1178)
at org.apache.sqoop.mapreduce.SQLServerExportDBExecThread.executeStatement(SQLServerExportDBExecThread.java:96)
at org.apache.sqoop.mapreduce.SQLServerAsyncDBExecThread.write(SQLServerAsyncDBExecThread.java:272)
... 1 more
2019-09-03 02:05:26,460 ERROR [main] org.apache.sqoop.mapreduce.SQLServerAsyncDBExecThread: Asynchronous writer thread encountered the following exception: java.io.IOException: Registered handler cannot recover error with SQL State: 22001, error code: 8152
2019-09-03 02:05:26,461 ERROR [main] org.apache.sqoop.mapreduce.TextExportMapper:
2019-09-03 02:05:26,461 ERROR [main] org.apache.sqoop.mapreduce.TextExportMapper: Exception raised during data export
2019-09-03 02:05:26,461 ERROR [main] org.apache.sqoop.mapreduce.TextExportMapper:
2019-09-03 02:05:26,461 ERROR [main] org.apache.sqoop.mapreduce.TextExportMapper: Exception:
java.io.IOException: java.io.IOException: Registered handler cannot recover error with SQL State: 22001, error code: 8152
at org.apache.sqoop.mapreduce.SQLServerAsyncDBExecThread.put(SQLServerAsyncDBExecThread.java:175)
at org.apache.sqoop.mapreduce.SQLServerResilientExportOutputFormat$SQLServerExportRecordWriter.write(SQLServerResilientExportOutputFormat.java:159)
at org.apache.sqoop.mapreduce.SQLServerResilientExportOutputFormat$SQLServerExportRecordWriter.write(SQLServerResilientExportOutputFormat.java:104)
at org.apache.hadoop.mapred.MapTask$NewDirectOutputCollector.write(MapTask.java:667)
at org.apache.hadoop.mapreduce.task.TaskInputOutputContextImpl.write(TaskInputOutputContextImpl.java:89)
at org.apache.hadoop.mapreduce.lib.map.WrappedMapper$Context.write(WrappedMapper.java:112)
at org.apache.sqoop.mapreduce.TextExportMapper.map(TextExportMapper.java:84)
at org.apache.sqoop.mapreduce.TextExportMapper.map(TextExportMapper.java:39)
at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:146)
at org.apache.sqoop.mapreduce.AutoProgressMapper.run(AutoProgressMapper.java:64)
at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:796)
at org.apache.hadoop.mapred.MapTask.run(MapTask.java:346)
at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:163)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:422)
at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1633)
at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:158)
Caused by: java.io.IOException: Registered handler cannot recover error with SQL State: 22001, error code: 8152
at org.apache.sqoop.mapreduce.SQLServerAsyncDBExecThread.write(SQLServerAsyncDBExecThread.java:293)
at org.apache.sqoop.mapreduce.SQLServerAsyncDBExecThread.run(SQLServerAsyncDBExecThread.java:240)
Caused by: java.sql.BatchUpdateException: String or binary data would be truncated.
at com.microsoft.sqlserver.jdbc.SQLServerPreparedStatement.executeBatch(SQLServerPreparedStatement.java:1178)
at org.apache.sqoop.mapreduce.SQLServerExportDBExecThread.executeStatement(SQLServerExportDBExecThread.java:96)
at org.apache.sqoop.mapreduce.SQLServerAsyncDBExecThread.write(SQLServerAsyncDBExecThread.java:272)
... 1 more
2019-09-03 02:05:26,461 ERROR [main] org.apache.sqoop.mapreduce.TextExportMapper: On input: 6862654212019-03-01 19:26:50JP Morgan Chase S1333Spark room / issue not accepate push from cucmClosedCONFIG_ASSISTANCE\NPHONEmichael.j.chiappalone@jpmorgan.com2019-03-01 20:26:05SW_CONFIGTACMichael Chiappalone1-+19179397379- Ext: JP MORGAN CHASE BANKTelepresenceCTSSOL2006241223103554YYY403475303JP MORGAN CHASE BANK770658JEFFERSONVILLE\N47130-3451USNAMJPMC11228303652019-03-02 03:01:27Webex Room Kit (On-Prem/not cloud registered)
2019-09-03 02:05:26,461 ERROR [main] org.apache.sqoop.mapreduce.TextExportMapper: On input file: maprfs:///app/SmartAnalytics/Apps/CSP/hivewarehouse/csp_tsbi.db/sr_passim_data_platform_mood_db_p01/000000_0
2019-09-03 02:05:26,461 ERROR [main] org.apache.sqoop.mapreduce.TextExportMapper: At position 23180227
2019-09-03 02:05:26,462 ERROR [main] org.apache.sqoop.mapreduce.TextExportMapper:
2019-09-03 02:05:26,462 ERROR [main] org.apache.sqoop.mapreduce.TextExportMapper: Currently processing split:
2019-09-03 02:05:26,462 ERROR [main] org.apache.sqoop.mapreduce.TextExportMapper: Paths:/app/SmartAnalytics/Apps/CSP/hivewarehouse/csp_tsbi.db/sr_passim_data_platform_mood_db_p01/000000_0:0+30277399
2019-09-03 02:05:26,462 ERROR [main] org.apache.sqoop.mapreduce.TextExportMapper:
2019-09-03 02:05:26,462 ERROR [main] org.apache.sqoop.mapreduce.TextExportMapper: This issue might not necessarily be caused by current input
2019-09-03 02:05:26,462 ERROR [main] org.apache.sqoop.mapreduce.TextExportMapper: due to the batching nature of export.
2019-09-03 02:05:26,462 ERROR [main] org.apache.sqoop.mapreduce.TextExportMapper:
2019-09-03 02:05:26,462 INFO [Thread-13] org.apache.sqoop.mapreduce.AutoProgressMapper: Auto-progress thread is finished. keepGoing=false
End of LogType:syslog