datahub version: 0.12.0.
try to add super data source with recipe:
source:
type: superset
config:
connect_uri: ‘https://my.superset/’
username: xxx
password: xxx
provider: ldap
package install on datahub-actions: pip install ‘acryl-datahub[mysql,datahub-rest,snowflake,superset,dbt]==0.12.0’
How do I resolve the following error?
Error log:
Execution finished with errors.
{'exec_id': '7026aea0-a449-4072-9b07-59485ecb421f',
'infos': ['2023-11-14 20:31:35.431473 INFO: Starting execution for task with name=RUN_INGEST',
"2023-11-14 20:31:37.447644 INFO: Failed to execute 'datahub ingest', exit code 1",
'2023-11-14 20:31:37.447814 INFO: Caught exception EXECUTING task_id=7026aea0-a449-4072-9b07-59485ecb421f, name=RUN_INGEST, '
'stacktrace=Traceback (most recent call last):\n'
' File "/usr/local/lib/python3.10/site-packages/acryl/executor/execution/default_executor.py", line 140, in execute_task\n'
' task_event_loop.run_until_complete(task_future)\n'
' File "/usr/local/lib/python3.10/asyncio/base_events.py", line 649, in run_until_complete\n'
' return future.result()\n'
' File "/usr/local/lib/python3.10/site-packages/acryl/executor/execution/sub_process_ingestion_task.py", line 282, in execute\n'
' raise TaskError("Failed to execute \'datahub ingest\'")\n'
"acryl.executor.execution.task.TaskError: Failed to execute 'datahub ingest'\n"],
'errors': []}
~~~~ Ingestion Logs ~~~~
Obtaining venv creation lock...
Acquired venv creation lock
venv is already set up
venv setup time = 0 sec
This version of datahub supports report-to functionality
+ exec datahub ingest run -c /tmp/datahub/ingest/7026aea0-a449-4072-9b07-59485ecb421f/recipe.yml --report-to /tmp/datahub/ingest/7026aea0-a449-4072-9b07-59485ecb421f/ingestion_report.json
[2023-11-14 20:31:36,916] INFO {datahub.cli.ingest_cli:147} - DataHub CLI version: 0.12.0
[2023-11-14 20:31:36,948] INFO {datahub.ingestion.run.pipeline:213} - Sink configured successfully. DataHubRestEmitter: configured to talk to <http://datahub-gms:8080>
[2023-11-14 20:31:37,144] ERROR {datahub.entrypoints:201} - Command failed: Failed to configure the source (superset): 'access_token'
Traceback (most recent call last):
File "/home/datahub/.local/lib/python3.10/site-packages/datahub/ingestion/run/pipeline.py", line 120, in _add_init_error_context
yield
File "/home/datahub/.local/lib/python3.10/site-packages/datahub/ingestion/run/pipeline.py", line 226, in __init__
self.source = source_class.create(
File "/home/datahub/.local/lib/python3.10/site-packages/datahub/ingestion/source/superset.py", line 198, in create
return cls(ctx, config)
File "/home/datahub/.local/lib/python3.10/site-packages/datahub/ingestion/source/superset.py", line 177, in __init__
self.access_token = login_response.json()["access_token"]
KeyError: 'access_token'
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/home/datahub/.local/lib/python3.10/site-packages/datahub/entrypoints.py", line 188, in main
sys.exit(datahub(standalone_mode=False, **kwargs))
File "/usr/local/lib/python3.10/site-packages/click/core.py", line 1130, in __call__
return self.main(*args, **kwargs)
...
File "/home/datahub/.local/lib/python3.10/site-packages/datahub/cli/ingest_cli.py", line 167, in run_ingestion_and_check_upgrade
pipeline = Pipeline.create(
File "/home/datahub/.local/lib/python3.10/site-packages/datahub/ingestion/run/pipeline.py", line 336, in create
return cls(
File "/home/datahub/.local/lib/python3.10/site-packages/datahub/ingestion/run/pipeline.py", line 225, in __init__
with _add_init_error_context(f"configure the source ({source_type})"):
File "/usr/local/lib/python3.10/contextlib.py", line 153, in __exit__
self.gen.throw(typ, value, traceback)
File "/home/datahub/.local/lib/python3.10/site-packages/datahub/ingestion/run/pipeline.py", line 122, in _add_init_error_context
raise PipelineInitError(f"Failed to {step}: {e}") from e
datahub.ingestion.run.pipeline.PipelineInitError: Failed to configure the source (superset): 'access_token'```