@@ -56,26 +56,23 @@ let cs_payloadseqid
56
56
await consumer . commitOffset ( { topic, partition, offset : m . offset } ) // Commit offset only on success
57
57
if ( message . payload [ 'retryCount' ] ) retryvar = message . payload . retryCount ;
58
58
auditTrail ( [ cs_payloadseqid , cs_processId , message . payload . table , message . payload . Uniquecolumn ,
59
- message . payload . operation , "Informix-updated" , retryvar , "" , "" , message . payload . data , message . timestamp , message . topic ] , 'consumer' )
59
+ message . payload . operation , "Informix-updated" , retryvar , "" , "" , JSON . stringify ( message ) , message . timestamp , message . topic ] , 'consumer' )
60
60
} catch ( err ) {
61
- const errmsg2 = `Could not process kafka message or informix DB error: "${ err . message } "`
61
+ const errmsg2 = `error-sync: Could not process kafka message or informix DB error: "${ err . message } "`
62
62
logger . error ( errmsg2 )
63
- //await callposttoslack(errmsg2)
64
- //logger.logFullError(err)
65
63
logger . debug ( `error-sync: consumer "${ err . message } "` )
66
64
if ( ! cs_payloadseqid ) {
67
- cs_payloadseqid = 'err-' + ( new Date ( ) ) . getTime ( ) . toString ( 36 ) + Math . random ( ) . toString ( 36 ) . slice ( 2 ) ;
68
- }
69
-
70
- await auditTrail ( [ cs_payloadseqid , 3333 , 'message.payload.table' , 'message.payload.Uniquecolumn' ,
65
+ cs_payloadseqid = 'err-' + ( new Date ( ) ) . getTime ( ) . toString ( 36 ) + Math . random ( ) . toString ( 36 ) . slice ( 2 ) ; }
66
+ /* await auditTrail([cs_payloadseqid,3333,'message.payload.table','message.payload.Uniquecolumn',
71
67
'message.payload.operation',"Error-Consumer","",err.message,"",'message.payload.data',new Date(),'message.topic'],'consumer')
68
+ }else{
69
+ auditTrail([cs_payloadseqid,4444,message.payload.table,message.payload.Uniquecolumn,
70
+ message.payload.operation,"Informix-updated",retryvar,"consumer2","",JSON.stringify(message), message.timestamp,message.topic],'consumer')
71
+ }*/
72
+
72
73
try {
73
- //var retryvar
74
74
if ( message . payload [ 'retryCount' ] ) retryvar = message . payload . retryCount ;
75
75
await consumer . commitOffset ( { topic, partition, offset : m . offset } ) // Commit success as will re-publish
76
- // await auditTrail([cs_payloadseqid,3333,'message.payload.table','message.payload.Uniquecolumn',
77
- // 'message.payload.operation',"Informix-Updated1",retryvar,"","",'message.payload.data',new Date(),'message.topic'],'consumer')
78
- //await callposttoslack(`Retry for Kafka push : retrycount : "${retryvar}"`)
79
76
logger . debug ( `Trying to push same message after adding retryCounter` )
80
77
if ( ! message . payload . retryCount ) {
81
78
message . payload . retryCount = 0
@@ -95,14 +92,14 @@ let cs_payloadseqid
95
92
}
96
93
message . payload [ 'retryCount' ] = message . payload . retryCount + 1 ;
97
94
await pushToKafka ( message )
98
- var errmsg9 = `Retry for Kafka push : retrycount : "${ message . payload . retryCount } " : "${ cs_payloadseqid } "`
95
+ var errmsg9 = `error-sync: Retry for Kafka push : retrycount : "${ message . payload . retryCount } " : "${ cs_payloadseqid } "`
99
96
logger . debug ( errmsg9 )
100
97
//await callposttoslack(errmsg9)
101
98
} catch ( err ) {
102
99
103
100
await auditTrail ( [ cs_payloadseqid , cs_processId , message . payload . table , message . payload . Uniquecolumn ,
104
101
message . payload . operation , "Error-republishing" , message . payload [ 'retryCount' ] , err . message , "" , message . payload . data , message . timestamp , message . topic ] , 'consumer' )
105
- const errmsg1 = `postgres-ifx-processor: consumer : Error-republishing: "${ err . message } "`
102
+ const errmsg1 = `error-sync: postgres-ifx-processor: consumer : Error-republishing: "${ err . message } "`
106
103
logger . error ( errmsg1 )
107
104
logger . debug ( `error-sync: consumer re-publishing "${ err . message } "` )
108
105
// push to slack - alertIt("slack message"
0 commit comments