Error Resolution for Various Resource Ingestion Tasks

Original Slack Thread

{'exec_id': '8937ace4-b901-47bc-95b9-9081443d002b',
 'infos': ['2024-01-08 17:28:00.994034 INFO: Starting execution for task with name=RUN_INGEST',
           '2024-01-08 17:30:14.936525 INFO: Caught exception EXECUTING task_id=8937ace4-b901-47bc-95b9-9081443d002b, name=RUN_INGEST, '
           'stacktrace=Traceback (most recent call last):\n'
           '  File "/usr/local/lib/python3.10/site-packages/datahub/emitter/rest_emitter.py", line 266, in _emit_generic\n'
           '    response.raise_for_status()\n'
           '  File "/usr/local/lib/python3.10/site-packages/requests/models.py", line 1021, in raise_for_status\n'
           '    raise HTTPError(http_error_msg, response=self)\n'
           'requests.exceptions.HTTPError: 500 Server Error: Server Error for url: <http://datahub-datahub-gms:8080/aspects?action=ingestProposal>\n'
           '\n'
           'The above exception was the direct cause of the following exception:\n'
           '\n'
           'Traceback (most recent call last):\n'
           '  File "/usr/local/lib/python3.10/site-packages/acryl/executor/execution/sub_process_ingestion_task.py", line 191, in execute\n'
           '    await tasks.gather(read_output_task, report_progress_task, process_waiter_task)\n'
           '  File "/usr/local/lib/python3.10/site-packages/acryl/executor/execution/sub_process_ingestion_task.py", line 177, in _report_progress\n'
           '    ctx.request.progress_callback(report)\n'
           '  File "/usr/local/lib/python3.10/site-packages/acryl/executor/execution/reporting_executor.py", line 75, in &lt;lambda&gt;\n'
           '    request.progress_callback = lambda partial_report: self._datahub_graph.emit_mcp(self._build_progress_mcp(request, start_time_ms, '
           'partial_report=partial_report))\n'
           '  File "/usr/local/lib/python3.10/site-packages/datahub/emitter/rest_emitter.py", line 245, in emit_mcp\n'
           '    self._emit_generic(url, payload)\n'
           '  File "/usr/local/lib/python3.10/site-packages/datahub/emitter/rest_emitter.py", line 274, in _emit_generic\n'
           '    raise OperationalError(\n'
           "datahub.configuration.common.OperationalError: ('Unable to emit metadata to DataHub GMS: java.lang.RuntimeException: Failed to produce "
           "MCLs', {'exceptionClass': 'com.linkedin.restli.server.RestLiServiceException', 'message': 'java.lang.RuntimeException: Failed to produce "
           "MCLs', 'status': 500})\n"
           '\n'
           'The above exception was the direct cause of the following exception:\n'
           '\n'
           'Traceback (most recent call last):\n'
           '  File "/usr/local/lib/python3.10/site-packages/acryl/executor/execution/default_executor.py", line 122, in execute_task\n'
           '    task_event_loop.run_until_complete(task_future)\n'
           '  File "/usr/local/lib/python3.10/asyncio/base_events.py", line 649, in run_until_complete\n'
           '    return future.result()\n'
           '  File "/usr/local/lib/python3.10/site-packages/acryl/executor/execution/sub_process_ingestion_task.py", line 214, in execute\n'
           "    raise RuntimeError(f'Something went wrong in the subprocess executor: {e}') from e\n"
           "RuntimeError: Something went wrong in the subprocess executor: ('Unable to emit metadata to DataHub GMS: java.lang.RuntimeException: "
           "Failed to produce MCLs', {'exceptionClass': 'com.linkedin.restli.server.RestLiServiceException', 'message': 'java.lang.RuntimeException: "
           "Failed to produce MCLs', 'status': 500})\n"],
 'errors': []}```
I get the following error for Vertica and PowerBI resource. I was getting the error also on Snowflake but then I changed CLI to 0.12.1 and it worked. However it maintained the same error on the other resources. Any help here?

Please, can you check the datahub gms server logs for further info about why it returned 500?

<@UV14447EU> attached is the GMS log for a Vertica run. It seemed to fail at a particular table, I then excluded the table but the error came back for a different table. Will also pull out PowerBI logsattachment

fixed this by changing to cli version vertica 0.10.5 and powerbi 0.12.1

awesome, thanks for the heads up

I seem to get same error again for Tableau. Logs are pretty much similar to the ones given before. any help <@UV14447EU> ?

Please, can you check in the GMS logs what error you see?

<@UV14447EU> I switched the tableau cli to 0.12.1, 11.0 and 12.0 and got the following error

Traceback (most recent call last):
  File "/tmp/datahub/ingest/venv-tableau-0.11.0/lib/python3.10/site-packages/datahub/entrypoints.py", line 186, in main
    sys.exit(datahub(standalone_mode=False, **kwargs))
  File "/tmp/datahub/ingest/venv-tableau-0.11.0/lib/python3.10/site-packages/click/core.py", line 1157, in __call__
    return self.main(*args, **kwargs)
  File "/tmp/datahub/ingest/venv-tableau-0.11.0/lib/python3.10/site-packages/click/core.py", line 1078, in main
    rv = self.invoke(ctx)
  File "/tmp/datahub/ingest/venv-tableau-0.11.0/lib/python3.10/site-packages/click/core.py", line 1688, in invoke
    return _process_result(sub_ctx.command.invoke(sub_ctx))
  File "/tmp/datahub/ingest/venv-tableau-0.11.0/lib/python3.10/site-packages/click/core.py", line 1688, in invoke
    return _process_result(sub_ctx.command.invoke(sub_ctx))
  File "/tmp/datahub/ingest/venv-tableau-0.11.0/lib/python3.10/site-packages/click/core.py", line 1434, in invoke
    return ctx.invoke(self.callback, **ctx.params)
  File "/tmp/datahub/ingest/venv-tableau-0.11.0/lib/python3.10/site-packages/click/core.py", line 783, in invoke
    return __callback(*args, **kwargs)
  File "/tmp/datahub/ingest/venv-tableau-0.11.0/lib/python3.10/site-packages/click/decorators.py", line 33, in new_func
    return f(get_current_context(), *args, **kwargs)
  File "/tmp/datahub/ingest/venv-tableau-0.11.0/lib/python3.10/site-packages/datahub/telemetry/telemetry.py", line 448, in wrapper
    raise e
  File "/tmp/datahub/ingest/venv-tableau-0.11.0/lib/python3.10/site-packages/datahub/telemetry/telemetry.py", line 397, in wrapper
    res = func(*args, **kwargs)
  File "/tmp/datahub/ingest/venv-tableau-0.11.0/lib/python3.10/site-packages/datahub/utilities/memory_leak_detector.py", line 95, in wrapper
    return func(ctx, *args, **kwargs)
  File "/tmp/datahub/ingest/venv-tableau-0.11.0/lib/python3.10/site-packages/datahub/cli/ingest_cli.py", line 198, in run
    ret = loop.run_until_complete(run_ingestion_and_check_upgrade())
  File "/usr/local/lib/python3.10/asyncio/base_events.py", line 649, in run_until_complete
    return future.result()
  File "/tmp/datahub/ingest/venv-tableau-0.11.0/lib/python3.10/site-packages/datahub/cli/ingest_cli.py", line 182, in run_ingestion_and_check_upgrade
    ret = await ingestion_future
  File "/tmp/datahub/ingest/venv-tableau-0.11.0/lib/python3.10/site-packages/datahub/cli/ingest_cli.py", line 140, in run_pipeline_to_completion
    raise e
  File "/tmp/datahub/ingest/venv-tableau-0.11.0/lib/python3.10/site-packages/datahub/cli/ingest_cli.py", line 132, in run_pipeline_to_completion
    pipeline.run()
  File "/tmp/datahub/ingest/venv-tableau-0.11.0/lib/python3.10/site-packages/datahub/ingestion/run/pipeline.py", line 367, in run
    for wu in itertools.islice(
  File "/tmp/datahub/ingest/venv-tableau-0.11.0/lib/python3.10/site-packages/datahub/ingestion/api/source_helpers.py", line 119, in auto_stale_entity_removal
    for wu in stream:
  File "/tmp/datahub/ingest/venv-tableau-0.11.0/lib/python3.10/site-packages/datahub/ingestion/api/source_helpers.py", line 143, in auto_workunit_reporter
    for wu in stream:
  File "/tmp/datahub/ingest/venv-tableau-0.11.0/lib/python3.10/site-packages/datahub/ingestion/api/source_helpers.py", line 208, in auto_browse_path_v2
    for urn, batch in _batch_workunits_by_urn(stream):
  File "/tmp/datahub/ingest/venv-tableau-0.11.0/lib/python3.10/site-packages/datahub/ingestion/api/source_helpers.py", line 346, in _batch_workunits_by_urn
    for wu in stream:
  File "/tmp/datahub/ingest/venv-tableau-0.11.0/lib/python3.10/site-packages/datahub/ingestion/api/source_helpers.py", line 156, in auto_materialize_referenced_tags
    for wu in stream:
  File "/tmp/datahub/ingest/venv-tableau-0.11.0/lib/python3.10/site-packages/datahub/ingestion/api/source_helpers.py", line 70, in auto_status_aspect
    for wu in stream:
  File "/tmp/datahub/ingest/venv-tableau-0.11.0/lib/python3.10/site-packages/datahub/ingestion/source/tableau.py", line 2567, in get_workunits_internal
    yield from self.emit_workbooks()
  File "/tmp/datahub/ingest/venv-tableau-0.11.0/lib/python3.10/site-packages/datahub/ingestion/source/tableau.py", line 791, in emit_workbooks
    for workbook in self.get_connection_objects(
  File "/tmp/datahub/ingest/venv-tableau-0.11.0/lib/python3.10/site-packages/datahub/ingestion/source/tableau.py", line 770, in get_connection_objects
    ) = self.get_connection_object_page(
  File "/tmp/datahub/ingest/venv-tableau-0.11.0/lib/python3.10/site-packages/datahub/ingestion/source/tableau.py", line 704, in get_connection_object_page
    query_data = query_metadata(
  File "/tmp/datahub/ingest/venv-tableau-0.11.0/lib/python3.10/site-packages/datahub/ingestion/source/tableau_common.py", line 761, in query_metadata
    return server.metadata.query(query)
  File "/tmp/datahub/ingest/venv-tableau-0.11.0/lib/python3.10/site-packages/tableauserverclient/server/endpoint/endpoint.py", line 290, in wrapper
    self.parent_srv.assert_at_least_version(version, self.__class__.__name__)
  File "/tmp/datahub/ingest/venv-tableau-0.11.0/lib/python3.10/site-packages/tableauserverclient/server/server.py", line 196, in assert_at_least_version
    raise EndpointUnavailableError(error)
tableauserverclient.server.exceptions.EndpointUnavailableError: Metadata is not available in API version 2.4. Requires 3.5```
I got the error metadata is not available in API version 2.4 requires 3.5 . I deploy using helm charts on kubernetes how I can I change the default value used?