Skip to content

Commit

Permalink
Merge branch 'develop' of github.com:IQSS/dataverse into 10943-featur…
Browse files Browse the repository at this point in the history
…ed-items
  • Loading branch information
GPortas committed Jan 11, 2025
2 parents b4f5ce9 + 4373753 commit 998cc6c
Show file tree
Hide file tree
Showing 11 changed files with 122 additions and 14 deletions.
Original file line number Diff line number Diff line change
@@ -0,0 +1,8 @@
### Json Printer Bug fix

DatasetFieldTypes in MetadataBlock response that are also a child of another DatasetFieldType were being returned twice. The child DatasetFieldType was included in the "fields" object as well as in the "childFields" of it's parent DatasetFieldType. This fix suppresses the standalone object so only one instance of the DatasetFieldType is returned (in the "childFields" of its parent).
This fix changes the Json output of the API `/api/dataverses/{dataverseAlias}/metadatablocks`

## Backward Incompatible Changes

The Json response of API call `/api/dataverses/{dataverseAlias}/metadatablocks` will no longer include the DatasetFieldTypes in "fields" if they are children of another DatasetFieldType. The child DatasetFieldType will only be included in the "childFields" of it's parent DatasetFieldType.
3 changes: 3 additions & 0 deletions doc/release-notes/11107-fake-to-perma-demo.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,3 @@
### Demo/Eval Container Tutorial

The demo/eval container tutorial has been updated to use the Permalink PID provider instead of the FAKE DOI Provider. See also #11107.
5 changes: 5 additions & 0 deletions doc/release-notes/11113-avoid-orphan-perm-docs.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,5 @@
This release fixes a bug that caused Dataverse to generate unnecessary solr documents for files when a file is added/deleted from a draft dataset. These documents could accumulate and potentially impact performance.

Assuming the upgrade to solr 9.7.0 also occurs in this release, there's nothing else needed for this PR. (Starting with a new solr insures the solr db is empty and that a reindex is already required.)


5 changes: 5 additions & 0 deletions doc/sphinx-guides/source/container/running/demo.rst
Original file line number Diff line number Diff line change
Expand Up @@ -160,6 +160,11 @@ Next, set up the UI toggle between English and French, again using the unblock k

Stop and start the Dataverse container in order for the language toggle to work.

PID Providers
+++++++++++++

Dataverse supports multiple Persistent ID (PID) providers. The ``compose.yml`` file uses the Permalink PID provider. Follow :ref:`pids-configuration` to reconfigure as needed.

Next Steps
----------

Expand Down
12 changes: 6 additions & 6 deletions docker/compose/demo/compose.yml
Original file line number Diff line number Diff line change
Expand Up @@ -20,12 +20,12 @@ services:
-Ddataverse.files.file1.type=file
-Ddataverse.files.file1.label=Filesystem
-Ddataverse.files.file1.directory=${STORAGE_DIR}/store
-Ddataverse.pid.providers=fake
-Ddataverse.pid.default-provider=fake
-Ddataverse.pid.fake.type=FAKE
-Ddataverse.pid.fake.label=FakeDOIProvider
-Ddataverse.pid.fake.authority=10.5072
-Ddataverse.pid.fake.shoulder=FK2/
-Ddataverse.pid.providers=perma1
-Ddataverse.pid.default-provider=perma1
-Ddataverse.pid.perma1.type=perma
-Ddataverse.pid.perma1.label=Perma1
-Ddataverse.pid.perma1.authority=DV
-Ddataverse.pid.perma1.permalink.separator=/
#-Ddataverse.lang.directory=/dv/lang
ports:
- "8080:8080" # HTTP (Dataverse Application)
Expand Down
8 changes: 8 additions & 0 deletions src/main/java/edu/harvard/iq/dataverse/DataFile.java
Original file line number Diff line number Diff line change
Expand Up @@ -1142,4 +1142,12 @@ public boolean isDeaccessioned() {
}
return inDeaccessionedVersions; // since any published version would have already returned
}
public boolean isInDatasetVersion(DatasetVersion version) {
for (FileMetadata fmd : getFileMetadatas()) {
if (fmd.getDatasetVersion().equals(version)) {
return true;
}
}
return false;
}
} // end of class
Original file line number Diff line number Diff line change
Expand Up @@ -155,7 +155,15 @@ private List<DvObjectSolrDoc> constructDatafileSolrDocs(DataFile dataFile, Map<L
Map<DatasetVersion.VersionState, Boolean> desiredCards = searchPermissionsService.getDesiredCards(dataFile.getOwner());
for (DatasetVersion datasetVersionFileIsAttachedTo : datasetVersionsToBuildCardsFor(dataFile.getOwner())) {
boolean cardShouldExist = desiredCards.get(datasetVersionFileIsAttachedTo.getVersionState());
if (cardShouldExist) {
/*
* Since datasetVersionFileIsAttachedTo should be a draft or the most recent
* released one, it could be more efficient to stop the search through
* FileMetadatas after those two (versus continuing through all prior versions
* as in isInDatasetVersion). Alternately, perhaps filesToReIndexPermissionsFor
* should not combine the list of files for the different datsetversions into a
* single list to start with.
*/
if (cardShouldExist && dataFile.isInDatasetVersion(datasetVersionFileIsAttachedTo)) {
String solrIdStart = IndexServiceBean.solrDocIdentifierFile + dataFile.getId();
String solrIdEnd = getDatasetOrDataFileSolrEnding(datasetVersionFileIsAttachedTo.getVersionState());
String solrId = solrIdStart + solrIdEnd;
Expand Down
24 changes: 24 additions & 0 deletions src/main/java/edu/harvard/iq/dataverse/util/json/JsonPrinter.java
Original file line number Diff line number Diff line change
Expand Up @@ -650,6 +650,19 @@ public static JsonObjectBuilder json(MetadataBlock metadataBlock, boolean printO
.add("displayName", metadataBlock.getDisplayName())
.add("displayOnCreate", metadataBlock.isDisplayOnCreate());

List<DatasetFieldType> datasetFieldTypesList;

if (ownerDataverse != null) {
datasetFieldTypesList = datasetFieldService.findAllInMetadataBlockAndDataverse(
metadataBlock, ownerDataverse, printOnlyDisplayedOnCreateDatasetFieldTypes);
} else {
datasetFieldTypesList = printOnlyDisplayedOnCreateDatasetFieldTypes
? datasetFieldService.findAllDisplayedOnCreateInMetadataBlock(metadataBlock)
: metadataBlock.getDatasetFieldTypes();
}

Set<DatasetFieldType> datasetFieldTypes = filterOutDuplicateDatasetFieldTypes(datasetFieldTypesList);

JsonObjectBuilder fieldsBuilder = Json.createObjectBuilder();

Predicate<DatasetFieldType> isNoChild = element -> element.isChild() == false;
Expand Down Expand Up @@ -679,6 +692,17 @@ public static JsonObjectBuilder json(MetadataBlock metadataBlock, boolean printO
return jsonObjectBuilder;
}

// This will remove datasetFieldTypes that are in the list but also a child of another datasetFieldType in the list
// Prevents duplicate datasetFieldType information from being returned twice
// See: https://github.com/IQSS/dataverse/issues/10472
private static Set<DatasetFieldType> filterOutDuplicateDatasetFieldTypes(List<DatasetFieldType> datasetFieldTypesList) {
// making a copy of the list as to not damage the original when we remove items
List<DatasetFieldType> datasetFieldTypes = new ArrayList<>(datasetFieldTypesList);
// exclude/remove datasetFieldTypes if datasetFieldType exists as a child of another datasetFieldType
datasetFieldTypesList.forEach(dsft -> dsft.getChildDatasetFieldTypes().forEach(c -> datasetFieldTypes.remove(c)));
return new TreeSet<>(datasetFieldTypes);
}

public static JsonArrayBuilder jsonDatasetFieldTypes(List<DatasetFieldType> fields) {
JsonArrayBuilder fieldsJson = Json.createArrayBuilder();
for (DatasetFieldType field : fields) {
Expand Down
7 changes: 3 additions & 4 deletions src/test/java/edu/harvard/iq/dataverse/api/DataversesIT.java
Original file line number Diff line number Diff line change
Expand Up @@ -927,7 +927,7 @@ public void testListMetadataBlocks() {
.body("data.size()", equalTo(1))
.body("data[0].name", is("citation"))
.body("data[0].fields.title.displayOnCreate", equalTo(true))
.body("data[0].fields.size()", is(10))
.body("data[0].fields.size()", is(10)) // 28 - 18 child duplicates
.body("data[0].fields.author.childFields.size()", is(4));

Response setMetadataBlocksResponse = UtilIT.setMetadataBlocks(dataverseAlias, Json.createArrayBuilder().add("citation").add("astrophysics"), apiToken);
Expand Down Expand Up @@ -1008,14 +1008,13 @@ public void testListMetadataBlocks() {
// Since the included property of notesText is set to false, we should retrieve the total number of fields minus one
int citationMetadataBlockIndex = geospatialMetadataBlockIndex == 0 ? 1 : 0;
listMetadataBlocksResponse.then().assertThat()
.body(String.format("data[%d].fields.size()", citationMetadataBlockIndex), equalTo(34));
.body(String.format("data[%d].fields.size()", citationMetadataBlockIndex), equalTo(34)); // 79 minus 45 child duplicates

// Since the included property of geographicCoverage is set to false, we should retrieve the total number of fields minus one
listMetadataBlocksResponse.then().assertThat()
.body(String.format("data[%d].fields.size()", geospatialMetadataBlockIndex), equalTo(2));

listMetadataBlocksResponse = UtilIT.getMetadataBlock("geospatial");

listMetadataBlocksResponse = UtilIT.getMetadataBlock("geospatial");
String actualGeospatialMetadataField1 = listMetadataBlocksResponse.then().extract().path(String.format("data.fields['geographicCoverage'].name"));
String actualGeospatialMetadataField2 = listMetadataBlocksResponse.then().extract().path(String.format("data.fields['geographicCoverage'].childFields['country'].name"));
String actualGeospatialMetadataField3 = listMetadataBlocksResponse.then().extract().path(String.format("data.fields['geographicCoverage'].childFields['city'].name"));
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -44,8 +44,7 @@ void testListMetadataBlocks() {

// returnDatasetFieldTypes=true
listMetadataBlocksResponse = UtilIT.listMetadataBlocks(false, true);
int expectedNumberOfMetadataFields = 35;
listMetadataBlocksResponse.prettyPrint();
int expectedNumberOfMetadataFields = 35; // 80 - 45 child duplicates;
listMetadataBlocksResponse.then().assertThat()
.statusCode(OK.getStatusCode())
.body("data[0].fields", not(equalTo(null)))
Expand All @@ -57,7 +56,7 @@ void testListMetadataBlocks() {
// onlyDisplayedOnCreate=true and returnDatasetFieldTypes=true
listMetadataBlocksResponse = UtilIT.listMetadataBlocks(true, true);
listMetadataBlocksResponse.prettyPrint();
expectedNumberOfMetadataFields = 10;
expectedNumberOfMetadataFields = 10; // 28 - 18 child duplicates
listMetadataBlocksResponse.then().assertThat()
.statusCode(OK.getStatusCode())
.body("data[0].fields", not(equalTo(null)))
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -25,6 +25,7 @@
import jakarta.json.JsonString;

import edu.harvard.iq.dataverse.util.BundleUtil;
import org.assertj.core.util.Lists;
import org.junit.jupiter.api.Test;
import org.junit.jupiter.api.BeforeEach;

Expand Down Expand Up @@ -268,6 +269,54 @@ public void testDatasetContactWithPrivacy() {

}

@Test
public void testDatasetFieldTypesWithChildren() {
MetadataBlock block = new MetadataBlock();
block.setId(0L);
block.setName("citation");
long id = 0L;
// create datasetFieldTypes
List<DatasetFieldType> datasetFieldTypes = new ArrayList<>();
for (int i = 0; i < 10; i++) {
DatasetFieldType dft = new DatasetFieldType();
dft.setId(id++);
dft.setDisplayOrder(i);
dft.setMetadataBlock(block);
dft.setFieldType(FieldType.TEXT);
dft.setName("subType" + dft.getId());
dft.setTitle(dft.getName());
dft.setChildDatasetFieldTypes(Lists.emptyList());
datasetFieldTypes.add(dft);
}
// add DatasetFieldType as children to another DatasetFieldType to test the suppression of duplicate data
// adding 3 and 4 as children of 2
datasetFieldTypes.get(3).setParentDatasetFieldType(datasetFieldTypes.get(2));
datasetFieldTypes.get(4).setParentDatasetFieldType(datasetFieldTypes.get(2));
datasetFieldTypes.get(2).setChildDatasetFieldTypes(List.of(datasetFieldTypes.get(3), datasetFieldTypes.get(4)));
// adding 6 as child of 9
datasetFieldTypes.get(6).setParentDatasetFieldType(datasetFieldTypes.get(9));
datasetFieldTypes.get(9).setChildDatasetFieldTypes(List.of(datasetFieldTypes.get(6)));

block.setDatasetFieldTypes(datasetFieldTypes);

DatasetFieldServiceBean nullDFServiceBean = null;
JsonPrinter.injectSettingsService(new MockSettingsSvc(), nullDFServiceBean);

JsonObject jsonObject = JsonPrinter.json(block).build();
assertNotNull(jsonObject);

System.out.println("json: " + JsonUtil.prettyPrint(jsonObject.toString()));
assertEquals("subType2 subType3", jsonObject.getJsonObject("fields").getJsonObject("subType2")
.getJsonObject("childFields").getJsonObject("subType3").getString("displayName"));
assertEquals("subType2 subType4", jsonObject.getJsonObject("fields").getJsonObject("subType2")
.getJsonObject("childFields").getJsonObject("subType4").getString("displayName"));
assertEquals("subType9 subType6", jsonObject.getJsonObject("fields").getJsonObject("subType9")
.getJsonObject("childFields").getJsonObject("subType6").getString("displayName"));
assertNull(jsonObject.getJsonObject("fields").getJsonObject("subType3"));
assertNull(jsonObject.getJsonObject("fields").getJsonObject("subType4"));
assertNull(jsonObject.getJsonObject("fields").getJsonObject("subType6"));
}

@Test
public void testDataversePrinter() {
Dataverse dataverse = new Dataverse();
Expand Down

0 comments on commit 998cc6c

Please sign in to comment.