{'exec_id': '8937ace4-b901-47bc-95b9-9081443d002b',
'infos': ['2024-01-08 17:28:00.994034 INFO: Starting execution for task with name=RUN_INGEST',
'2024-01-08 17:30:14.936525 INFO: Caught exception EXECUTING task_id=8937ace4-b901-47bc-95b9-9081443d002b, name=RUN_INGEST, '
'stacktrace=Traceback (most recent call last):\n'
' File "/usr/local/lib/python3.10/site-packages/datahub/emitter/rest_emitter.py", line 266, in _emit_generic\n'
' response.raise_for_status()\n'
' File "/usr/local/lib/python3.10/site-packages/requests/models.py", line 1021, in raise_for_status\n'
' raise HTTPError(http_error_msg, response=self)\n'
'requests.exceptions.HTTPError: 500 Server Error: Server Error for url: <http://datahub-datahub-gms:8080/aspects?action=ingestProposal>\n'
'\n'
'The above exception was the direct cause of the following exception:\n'
'\n'
'Traceback (most recent call last):\n'
' File "/usr/local/lib/python3.10/site-packages/acryl/executor/execution/sub_process_ingestion_task.py", line 191, in execute\n'
' await tasks.gather(read_output_task, report_progress_task, process_waiter_task)\n'
' File "/usr/local/lib/python3.10/site-packages/acryl/executor/execution/sub_process_ingestion_task.py", line 177, in _report_progress\n'
' ctx.request.progress_callback(report)\n'
' File "/usr/local/lib/python3.10/site-packages/acryl/executor/execution/reporting_executor.py", line 75, in <lambda>\n'
' request.progress_callback = lambda partial_report: self._datahub_graph.emit_mcp(self._build_progress_mcp(request, start_time_ms, '
'partial_report=partial_report))\n'
' File "/usr/local/lib/python3.10/site-packages/datahub/emitter/rest_emitter.py", line 245, in emit_mcp\n'
' self._emit_generic(url, payload)\n'
' File "/usr/local/lib/python3.10/site-packages/datahub/emitter/rest_emitter.py", line 274, in _emit_generic\n'
' raise OperationalError(\n'
"datahub.configuration.common.OperationalError: ('Unable to emit metadata to DataHub GMS: java.lang.RuntimeException: Failed to produce "
"MCLs', {'exceptionClass': 'com.linkedin.restli.server.RestLiServiceException', 'message': 'java.lang.RuntimeException: Failed to produce "
"MCLs', 'status': 500})\n"
'\n'
'The above exception was the direct cause of the following exception:\n'
'\n'
'Traceback (most recent call last):\n'
' File "/usr/local/lib/python3.10/site-packages/acryl/executor/execution/default_executor.py", line 122, in execute_task\n'
' task_event_loop.run_until_complete(task_future)\n'
' File "/usr/local/lib/python3.10/asyncio/base_events.py", line 649, in run_until_complete\n'
' return future.result()\n'
' File "/usr/local/lib/python3.10/site-packages/acryl/executor/execution/sub_process_ingestion_task.py", line 214, in execute\n'
" raise RuntimeError(f'Something went wrong in the subprocess executor: {e}') from e\n"
"RuntimeError: Something went wrong in the subprocess executor: ('Unable to emit metadata to DataHub GMS: java.lang.RuntimeException: "
"Failed to produce MCLs', {'exceptionClass': 'com.linkedin.restli.server.RestLiServiceException', 'message': 'java.lang.RuntimeException: "
"Failed to produce MCLs', 'status': 500})\n"],
'errors': []}```
I get the following error for Vertica and PowerBI resource. I was getting the error also on Snowflake but then I changed CLI to 0.12.1 and it worked. However it maintained the same error on the other resources. Any help here?
Please, can you check the datahub gms server logs for further info about why it returned 500?
<@UV14447EU> attached is the GMS log for a Vertica run. It seemed to fail at a particular table, I then excluded the table but the error came back for a different table. Will also pull out PowerBI logs
fixed this by changing to cli version vertica 0.10.5 and powerbi 0.12.1
awesome, thanks for the heads up
I seem to get same error again for Tableau. Logs are pretty much similar to the ones given before. any help <@UV14447EU> ?
Please, can you check in the GMS logs what error you see?
<@UV14447EU> I switched the tableau cli to 0.12.1, 11.0 and 12.0 and got the following error
Traceback (most recent call last):
File "/tmp/datahub/ingest/venv-tableau-0.11.0/lib/python3.10/site-packages/datahub/entrypoints.py", line 186, in main
sys.exit(datahub(standalone_mode=False, **kwargs))
File "/tmp/datahub/ingest/venv-tableau-0.11.0/lib/python3.10/site-packages/click/core.py", line 1157, in __call__
return self.main(*args, **kwargs)
File "/tmp/datahub/ingest/venv-tableau-0.11.0/lib/python3.10/site-packages/click/core.py", line 1078, in main
rv = self.invoke(ctx)
File "/tmp/datahub/ingest/venv-tableau-0.11.0/lib/python3.10/site-packages/click/core.py", line 1688, in invoke
return _process_result(sub_ctx.command.invoke(sub_ctx))
File "/tmp/datahub/ingest/venv-tableau-0.11.0/lib/python3.10/site-packages/click/core.py", line 1688, in invoke
return _process_result(sub_ctx.command.invoke(sub_ctx))
File "/tmp/datahub/ingest/venv-tableau-0.11.0/lib/python3.10/site-packages/click/core.py", line 1434, in invoke
return ctx.invoke(self.callback, **ctx.params)
File "/tmp/datahub/ingest/venv-tableau-0.11.0/lib/python3.10/site-packages/click/core.py", line 783, in invoke
return __callback(*args, **kwargs)
File "/tmp/datahub/ingest/venv-tableau-0.11.0/lib/python3.10/site-packages/click/decorators.py", line 33, in new_func
return f(get_current_context(), *args, **kwargs)
File "/tmp/datahub/ingest/venv-tableau-0.11.0/lib/python3.10/site-packages/datahub/telemetry/telemetry.py", line 448, in wrapper
raise e
File "/tmp/datahub/ingest/venv-tableau-0.11.0/lib/python3.10/site-packages/datahub/telemetry/telemetry.py", line 397, in wrapper
res = func(*args, **kwargs)
File "/tmp/datahub/ingest/venv-tableau-0.11.0/lib/python3.10/site-packages/datahub/utilities/memory_leak_detector.py", line 95, in wrapper
return func(ctx, *args, **kwargs)
File "/tmp/datahub/ingest/venv-tableau-0.11.0/lib/python3.10/site-packages/datahub/cli/ingest_cli.py", line 198, in run
ret = loop.run_until_complete(run_ingestion_and_check_upgrade())
File "/usr/local/lib/python3.10/asyncio/base_events.py", line 649, in run_until_complete
return future.result()
File "/tmp/datahub/ingest/venv-tableau-0.11.0/lib/python3.10/site-packages/datahub/cli/ingest_cli.py", line 182, in run_ingestion_and_check_upgrade
ret = await ingestion_future
File "/tmp/datahub/ingest/venv-tableau-0.11.0/lib/python3.10/site-packages/datahub/cli/ingest_cli.py", line 140, in run_pipeline_to_completion
raise e
File "/tmp/datahub/ingest/venv-tableau-0.11.0/lib/python3.10/site-packages/datahub/cli/ingest_cli.py", line 132, in run_pipeline_to_completion
pipeline.run()
File "/tmp/datahub/ingest/venv-tableau-0.11.0/lib/python3.10/site-packages/datahub/ingestion/run/pipeline.py", line 367, in run
for wu in itertools.islice(
File "/tmp/datahub/ingest/venv-tableau-0.11.0/lib/python3.10/site-packages/datahub/ingestion/api/source_helpers.py", line 119, in auto_stale_entity_removal
for wu in stream:
File "/tmp/datahub/ingest/venv-tableau-0.11.0/lib/python3.10/site-packages/datahub/ingestion/api/source_helpers.py", line 143, in auto_workunit_reporter
for wu in stream:
File "/tmp/datahub/ingest/venv-tableau-0.11.0/lib/python3.10/site-packages/datahub/ingestion/api/source_helpers.py", line 208, in auto_browse_path_v2
for urn, batch in _batch_workunits_by_urn(stream):
File "/tmp/datahub/ingest/venv-tableau-0.11.0/lib/python3.10/site-packages/datahub/ingestion/api/source_helpers.py", line 346, in _batch_workunits_by_urn
for wu in stream:
File "/tmp/datahub/ingest/venv-tableau-0.11.0/lib/python3.10/site-packages/datahub/ingestion/api/source_helpers.py", line 156, in auto_materialize_referenced_tags
for wu in stream:
File "/tmp/datahub/ingest/venv-tableau-0.11.0/lib/python3.10/site-packages/datahub/ingestion/api/source_helpers.py", line 70, in auto_status_aspect
for wu in stream:
File "/tmp/datahub/ingest/venv-tableau-0.11.0/lib/python3.10/site-packages/datahub/ingestion/source/tableau.py", line 2567, in get_workunits_internal
yield from self.emit_workbooks()
File "/tmp/datahub/ingest/venv-tableau-0.11.0/lib/python3.10/site-packages/datahub/ingestion/source/tableau.py", line 791, in emit_workbooks
for workbook in self.get_connection_objects(
File "/tmp/datahub/ingest/venv-tableau-0.11.0/lib/python3.10/site-packages/datahub/ingestion/source/tableau.py", line 770, in get_connection_objects
) = self.get_connection_object_page(
File "/tmp/datahub/ingest/venv-tableau-0.11.0/lib/python3.10/site-packages/datahub/ingestion/source/tableau.py", line 704, in get_connection_object_page
query_data = query_metadata(
File "/tmp/datahub/ingest/venv-tableau-0.11.0/lib/python3.10/site-packages/datahub/ingestion/source/tableau_common.py", line 761, in query_metadata
return server.metadata.query(query)
File "/tmp/datahub/ingest/venv-tableau-0.11.0/lib/python3.10/site-packages/tableauserverclient/server/endpoint/endpoint.py", line 290, in wrapper
self.parent_srv.assert_at_least_version(version, self.__class__.__name__)
File "/tmp/datahub/ingest/venv-tableau-0.11.0/lib/python3.10/site-packages/tableauserverclient/server/server.py", line 196, in assert_at_least_version
raise EndpointUnavailableError(error)
tableauserverclient.server.exceptions.EndpointUnavailableError: Metadata is not available in API version 2.4. Requires 3.5```
I got the error metadata is not available in API version 2.4 requires 3.5 . I deploy using helm charts on kubernetes how I can I change the default value used?