Resolving 'access_token' Error in DataHub Superset Source Configuration

Original Slack Thread

datahub version: 0.12.0.
try to add super data source with recipe:
source:
type: superset
config:
connect_uri: ‘https://my.superset/
username: xxx
password: xxx
provider: ldap

package install on datahub-actions: pip install ‘acryl-datahub[mysql,datahub-rest,snowflake,superset,dbt]==0.12.0’

How do I resolve the following error?
Error log:

Execution finished with errors.
{'exec_id': '7026aea0-a449-4072-9b07-59485ecb421f',
 'infos': ['2023-11-14 20:31:35.431473 INFO: Starting execution for task with name=RUN_INGEST',
           "2023-11-14 20:31:37.447644 INFO: Failed to execute 'datahub ingest', exit code 1",
           '2023-11-14 20:31:37.447814 INFO: Caught exception EXECUTING task_id=7026aea0-a449-4072-9b07-59485ecb421f, name=RUN_INGEST, '
           'stacktrace=Traceback (most recent call last):\n'
           '  File "/usr/local/lib/python3.10/site-packages/acryl/executor/execution/default_executor.py", line 140, in execute_task\n'
           '    task_event_loop.run_until_complete(task_future)\n'
           '  File "/usr/local/lib/python3.10/asyncio/base_events.py", line 649, in run_until_complete\n'
           '    return future.result()\n'
           '  File "/usr/local/lib/python3.10/site-packages/acryl/executor/execution/sub_process_ingestion_task.py", line 282, in execute\n'
           '    raise TaskError("Failed to execute \'datahub ingest\'")\n'
           "acryl.executor.execution.task.TaskError: Failed to execute 'datahub ingest'\n"],
 'errors': []}

~~~~ Ingestion Logs ~~~~
Obtaining venv creation lock...
Acquired venv creation lock
venv is already set up
venv setup time = 0 sec
This version of datahub supports report-to functionality
+ exec datahub ingest run -c /tmp/datahub/ingest/7026aea0-a449-4072-9b07-59485ecb421f/recipe.yml --report-to /tmp/datahub/ingest/7026aea0-a449-4072-9b07-59485ecb421f/ingestion_report.json
[2023-11-14 20:31:36,916] INFO     {datahub.cli.ingest_cli:147} - DataHub CLI version: 0.12.0
[2023-11-14 20:31:36,948] INFO     {datahub.ingestion.run.pipeline:213} - Sink configured successfully. DataHubRestEmitter: configured to talk to <http://datahub-gms:8080>
[2023-11-14 20:31:37,144] ERROR    {datahub.entrypoints:201} - Command failed: Failed to configure the source (superset): 'access_token'
Traceback (most recent call last):
  File "/home/datahub/.local/lib/python3.10/site-packages/datahub/ingestion/run/pipeline.py", line 120, in _add_init_error_context
    yield
  File "/home/datahub/.local/lib/python3.10/site-packages/datahub/ingestion/run/pipeline.py", line 226, in __init__
    self.source = source_class.create(
  File "/home/datahub/.local/lib/python3.10/site-packages/datahub/ingestion/source/superset.py", line 198, in create
    return cls(ctx, config)
  File "/home/datahub/.local/lib/python3.10/site-packages/datahub/ingestion/source/superset.py", line 177, in __init__
    self.access_token = login_response.json()["access_token"]
KeyError: 'access_token'

The above exception was the direct cause of the following exception:

Traceback (most recent call last):
  File "/home/datahub/.local/lib/python3.10/site-packages/datahub/entrypoints.py", line 188, in main
    sys.exit(datahub(standalone_mode=False, **kwargs))
  File "/usr/local/lib/python3.10/site-packages/click/core.py", line 1130, in __call__
    return self.main(*args, **kwargs)
 ...
  File "/home/datahub/.local/lib/python3.10/site-packages/datahub/cli/ingest_cli.py", line 167, in run_ingestion_and_check_upgrade
    pipeline = Pipeline.create(
  File "/home/datahub/.local/lib/python3.10/site-packages/datahub/ingestion/run/pipeline.py", line 336, in create
    return cls(
  File "/home/datahub/.local/lib/python3.10/site-packages/datahub/ingestion/run/pipeline.py", line 225, in __init__
    with _add_init_error_context(f"configure the source ({source_type})"):
  File "/usr/local/lib/python3.10/contextlib.py", line 153, in __exit__
    self.gen.throw(typ, value, traceback)
  File "/home/datahub/.local/lib/python3.10/site-packages/datahub/ingestion/run/pipeline.py", line 122, in _add_init_error_context
    raise PipelineInitError(f"Failed to {step}: {e}") from e
datahub.ingestion.run.pipeline.PipelineInitError: Failed to configure the source (superset): 'access_token'```

Hey there! :wave: Make sure your message includes the following information if relevant, so we can help more effectively!

  1. Which DataHub version are you using? (e.g. 0.12.0)

  2. Please post any relevant error logs on the thread!

Hey Difu! Apologies for the delayed response here - <@U01JUBAPZPS> might have some ideas!

Hey <@U065R5SUB18> i can definitely help here! Are you available to sync today?

Thank you for response <@U01JUBAPZPS> I find a fix: change provider: ldap to provider: db

Nice! How else is your deployment going? need help with anything else?