What data source(s) are you integrating with DataHub? Redshift
error:
java.lang.IllegalStateException: This error handler cannot process 'SerializationException's directly; please consider configuring an 'ErrorHandlingDeserializer' in the value and/or key deserializer
at org.springframework.kafka.listener.SeekUtils.seekOrRecover(SeekUtils.java:194)
at org.springframework.kafka.listener.SeekToCurrentErrorHandler.handle(SeekToCurrentErrorHandler.java:112)
at org.springframework.kafka.listener.KafkaMessageListenerContainer$ListenerConsumer.handleConsumerException(KafkaMessageListenerContainer.java:1604)
at org.springframework.kafka.listener.KafkaMessageListenerContainer$ListenerConsumer.run(KafkaMessageListenerContainer.java:1212)
at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)
at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
at java.base/java.lang.Thread.run(Thread.java:829)
Caused by: org.apache.kafka.common.errors.SerializationException: Error deserializing key/value for partition MetadataChangeLog_Versioned_v1-0 at offset 11102. If needed, please seek past the record to continue consumption.
Caused by: org.apache.kafka.common.errors.SerializationException: Error deserializing Avro message for id 2
Caused by: java.lang.ArrayIndexOutOfBoundsException: null```
The error message you’re encountering indicates a SerializationException in the Kafka consumer within the DataHub GMS pod. This is likely due to an issue with deserializing an Avro message. The error message suggests configuring an ErrorHandlingDeserializer for better error handling.
Check Schema Registry:
Ensure that your schema registry is correctly configured and accessible. If the schema registry is not reachable, it can cause deserialization errors.
Seek Past Faulty Record:
You might need to seek past the faulty record causing the deserialization issue. This can be done manually or programmatically.
Here is an example of how you can configure the ErrorHandlingDeserializer in your Kafka consumer properties:
The error “Failed to create token. Unknown error occurred.” suggests that there might be an issue with the token generation process in DataHub. Here are some steps to troubleshoot and resolve this issue:
Steps to Troubleshoot
Check Console and Network Logs:
Open the browser’s developer tools (usually by pressing F12 or right-clicking and selecting “Inspect”).
Go to the “Console” and “Network” tabs to check for any error messages or failed network requests when you attempt to create a token.
Verify Metadata Service Authentication Configuration:
Ensure that the datahub.metadata_service_authentication.enabled parameter is set to true in your values.yml file.
Ensure that the user attempting to create the token has the necessary permissions. The user must have either Generate Personal Access Tokens or Manage All Access Tokens permissions.
Check the logs of the DataHub services, especially the GMS (Graph Metadata Service) and frontend services, for any errors or warnings related to token creation.
You can use docker logs <container-id> or kubectl logs <pod-name> to inspect the logs.
Example Configuration for Token Generation
Ensure that your application.yml or equivalent configuration file has the following settings:
metadataService:
auth:
enabled: true
Example of Using Personal Access Tokens
Once the token is generated, you can use it in your ingestion recipes as follows: