Currently we are able to use Teradata source plugin to ingest metadata but for some reason the lineage info is not getting synced ,even though we make sure that query logging is enabled and dbqlogtbl access to user is present. Currently we are on version 0.12.0 and we don’t see any error in logs.
Also ,below are the packages version that been installed.
Hi Tamas, apologies for delay response since i was OOO for few days . I ran recipe in debug mode and couldn’t find any error message and couldn’t attach the logs here due to logs containing table details (can verify and attach if possible) . Also yes i have enabled “include_table_lineage=true” statement in recipe .
Hi <@UV14447EU> , any suggestion will be helpful here . Also i observed this warning message popping up in debug logs , but i m not sure how useful it would be.
WARNING {py.warnings:109} - /usr/local/lib/python3.9/dist-packages/datahub/ingestion/source/sql/teradata.py:162: RemovedIn20Warning: \\u001b[31mDeprecated API features detected! These feature(s) are not compatible with SQLAlchemy 2.0. \\u001b[32mTo prevent incompatible upgrades prior to updating applications, ensure requirements files are pinned to \\\"sqlalchemy<2.0\\\". \\u001b[36mSet environment variable SQLALCHEMY_WARN_20=1 to show all deprecation warnings. Set environment variable SQLALCHEMY_SILENCE_UBER_WARNING=1 to silence this message.\\u001b[0m (Background on SQLAlchemy 2.0 at: <https://sqlalche.me/e/b8d9>)\\n for entry in engine.execute
s.QueryID as "query_id",
UserName as "user",
StartTime AT TIME ZONE 'GMT' as "timestamp",
DefaultDatabase as default_database,
s.SqlTextInfo as "query_text",
s.SqlRowNo as "row_no"
FROM "DBC".DBQLogTbl as l
JOIN "DBC".DBQLSqlTbl as s on s.QueryID = l.QueryID
WHERE
l.ErrorCode = 0
AND l.statementtype not in (
'Unrecognized type',
'Create Database/User',
'Help',
'Modify Database',
'Drop Table',
'Show',
'Not Applicable',
'Grant',
'Abort',
'Database',
'Flush Query Logging',
'Null',
'Begin/End DBQL',
'Revoke'
)
and default_database not in ('DEMONOW_MONITOR')
ORDER BY "query_id", "row_no"```
we further filter the resultset based on the start/end time of lineage (it is usually the last 24 hours) and we also add filter based on your table filter
sure , just trying to understand where i need to run this query at ? our setup is based on Kubernetes cluster, where we are triggering ingestion recipe.
i verified the query from the output with the debug logs and it seems like we are not getting those sql queries in it . any suggestion on how we can proceed further