You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
we discussed datatypes and agreed overall these are well covered by issues raised
Linking References
Every reference must link to an entity : Referential integrity between linked datasets (ex A4DA and A4D, LB and LBO) - ticket in review for further testing (Referential Integrity - Further Tasks config#589)
Entity Lookup
Every row after filter must have an entity ??
Dataset
remove missing start date
every conservation area must have a name
geometry and points should be within the provision boundary
check document matches entity (AI)
Information
Freshness (the information has changed but the data hasn't changed)
tree points are on trees (faculty- supported using satellite map?)
--------------------------------------------------- OTHER DISCUSSIONS---------------------------------------------------
Source and Endpoint URLs
Charts and discussions around these
how good is our provision? Chart looking at coverage (map of England) and graph showing amount of data collected going up over time
ratio of authoritative provisions versus endpoints
sources and endpoints documented over time
every provision should have a count range. We discussed this and further exploration will be required. There may be instances in which we can be confident of a range expected for a provision such as conservation areas, however for some datasets (for example tree) it is unclear if we can be confident around this. Could AI be harnessed to help with this?
chart to show freshness of data on the platform- every provision to have a TTL (time to live) and endpoints could have a 'mute button'. Discussions around this included the idea that a data provider may not have updated data in a particular dataset for years, however in some cases this may not indicate data is not accurate or up to date. The idea of freshness is something which requires more exploration.
provision quality (table to allow visualisation of this is being built)
more row than expected?
reduction in duplicates
reduction in old entities and endpoints
Site Information Architecture
We discussed that the future aim for information architecture is a page per;
reacted with thumbs up emoji reacted with thumbs down emoji reacted with laugh emoji reacted with hooray emoji reacted with confused emoji reacted with heart emoji reacted with rocket emoji reacted with eyes emoji
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
@Ben-Hodgkiss, @paris-dp, @psd and @Swati-Dash met yesterday in Darlington to look at improving tests for assessing our data quality
Provision
We start our data collection process with "provision" (the combination of an organisation and a dataset).
The quality of each provision make available in a provision-quality dataset.
Provisions only currently exist to support our service supporting providers. We need to add other provisions.
Endpoint
Source (is it legit?)
Collection Logs
Resource
--------------------------------------------------- PIPLELINE---------------------------------------------------
Mapping Fields
NOTE: No hardcoding in our pipeline code
Datatypes
Linking References
Entity Lookup
Dataset
Information
--------------------------------------------------- OTHER DISCUSSIONS---------------------------------------------------
Charts and discussions around these
Site Information Architecture
We discussed that the future aim for information architecture is a page per;
How do we find people we haven't funded?
Beta Was this translation helpful? Give feedback.
All reactions