-
Notifications
You must be signed in to change notification settings - Fork 2
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Add queries for validating data against CubiQL's expectations. #145
base: master
Are you sure you want to change the base?
Conversation
Issue #127 - Add validation queries for use with rdf-validator which checks a data source against the requirements CubiQL makes on data cubes. The queries are templates which expect the relevant configuration values to be provided when executed. Add instructions to the README on running these validations.
@zeginis - Please could you try running these validation queries against your data and let me know if you have any problems? |
@lkitching I checked the PR. The queries are ok. I have tested them on some data created by Table2qb and they pass the tests. However I think we need some more tests to cover other CubiQL requirements:
What do you think? |
1. There is a code list that contains ONLY the concepts used at the cube 2.There is a code list for each qb:DimensionProperty
@lkitching I added 2 new SPARQL queries to support the CubiQL requirements I mentioned. When I run them independently at the SPARQL endpoint they return no results -> they succeed The config I use:
I run the validator at: http://195.251.218.39:8893/sparql |
@zeginis - Dimensions no longer need to specify a codelist - any dimensions which do not specify one and which are not ref area, ref period, string or decimal types are mapped to a |
Move the comments for the dimension validation queries to the end of of the file. Comments before the query cause sesame to infer the wrong query type (i.e. a graph query instead of a tuple query) which results in the wrong accept headers being sent to the remote SPARQL endpoint.
@zeginis - I've pushed a fix to the new queries to allow them to run as expected in |
@lkitching what do you meant they are mapped to a Can we use such dimensions without codelist to lock dimensions? e.g.
|
@lkitching I tried using a dimension that has values URIs but there is no codelist defined.
|
You can try at the endpoint: http://195.251.218.39:8893/sparql The dimension The configuration I use:
|
@zeginis - I've pushed a fix for the exception to |
@lkitching yes this works fine. Thank you |
@lkitching I understand that it is not mandatory for the dimensions to have a codelist However, if a codelist for the usedCodes is defined, then it should contain all and only the used codes at the cube. This is a common error we need to catch. The error occur at the transformation of data using Table2qb due to not matching URIs between the |
@lkitching I removed the query that checks if each dimension has a codelist. I left the other query that checks the dimensions that have a codelist if the codelists contain all and onlye the used codes at the cube. |
Issue #127 - Add validation queries for use with rdf-validator which
checks a data source against the requirements CubiQL makes on data cubes.
The queries are templates which expect the relevant configuration values
to be provided when executed.
Add instructions to the README on running these validations.