Found this content helpful? Log in or sign up to leave a like!
Profiling and/or quality assurance testing Canvas data
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
Has anyone implemented some form of data profiling, quality assurance checks, or similar mechanisms to identify significant changes and/or problems in the data you're bringing in from DAP? If so, would you be willing to share what you're doing as examples?
For context, Instructure delivered a change this summer that resulted in a new record in the pseudonyms table for each user. AFAIK, this wasn't announced in any of the release notes so I found out about it when an end user of the data found issues in a query. My school's configuration previously meant each user only had one pseudonym, so this effectively changed the granularity of the table for us and required making some updates to the data model.
I'm now trying to brainstorm methods to find out about something like this before my end users. I'm considering, for example, recording how many new rows are being inserted to a table during each run and then creating an alert if the table growth exceeds X%. I'd love to hear other suggestions or ideas if anyone is doing something similar. I'm already handling schema version changes, but that didn't apply to the pseudonyms change since the schema itself wasn't updated.
Thanks in advance!
Andy