I'm asking about how DSS monitors issues during datasets processing. I see two kinds of potential issues:
Volume : Inconsistant number of records in a dataset (eg : I expected at least 1k records per day for my "webtraffic" dataset)
Schema / values: One or more rows have fields that don't respect the defined schema or expected values (eg : in webtraffic dataset, IP adresses are not valid or values of a date field are not expected).
Is there a way to monitor / handle those errors in DSS and be notified by email or something ?