Hi, team! I’m trying to add properties when running a job and they don’t go to the datahub, although it says it’s a dictionary. What am I doing wrong?
Hey there! Make sure your message includes the following information if relevant, so we can help more effectively!
- Which DataHub version are you using? (e.g. 0.12.0)
- Please post any relevant error logs on the thread!

version: 0.12.1
- creating an emitter datahub
- creating DataFlow obj
- creating DataJob obj, add properties
- creating DataProcessInstance from DataFlow
- emmit start/stop
How to add properties for the task?
use acryl-datahub-0.13.1.3
datajob_info = DataJobInfo(
customProperties=custom_properties,
type=AzkabanJobTypeClass.COMMAND,
name=datajob_name
)
# Construct a MetadataChangeProposalWrapper object for dataset
mcpw = MetadataChangeProposalWrapper(
entityUrn=datajob_urn,
aspect=datajob_info,
changeType=ChangeType.UPSERT
)
rest_emitter = DatahubRestEmitter(gms_server=gms)
rest_emitter.emit(mcpw)```
thx, I needed something else, I already found a solution - just assign the dictionary with the necessary data to the properties attribute of the DataJob object
At first I tried to put them in the properties of the DataProcessInstance object, but it didn’t work)
DataProcessInstance represents an instance of a datajob, it will have run , input and output information

properties: Custom properties to set for the DataProcessInstance
Yes, that’s correct it properties for DataProcessInstance not to datajob
there is a separate aspect DataProcessInstanceProperties
for properties DataProcessInstance
https://github.com/datahub-project/datahub/blob/934ab03d16dc52f992a807a2002e9949cc6f95fa/metadata-ingestion/src/datahub/api/entities/dataprocess/dataprocess_instance.py#L234C20-L234C49|https://github.com/datahub-project/datahub/blob/934ab03d16dc52f992a807a2002e9949cc[…]on/src/datahub/api/entities/dataprocess/dataprocess_instance.py
https://datahubproject.io/docs/generated/metamodel/entities/dataprocessinstance/