added logic for selection criteria implementation on datasources
export of link between publication and software (related to #4593)
print json but commented out.Added test to get a proto from a json
added resourcetype and resulttype, new work type mapping
code cleaning
code adapted to new version of mapping utils 6.3.25 which supports journal information also on data sources
this mapper handles orcid publications (without doi) actions
generate actionset for orcid works and unit tests
Avoid nullpointer for publisher
Discard records without a valid author as requested in #4392, #4393, #4395, #4396.If the record has also at least one valid author, the record is kept but the invalid authors are removed.
Workaround for #4362: instances from Unpaywall mapped into licenses and then lost
Addressing quality of the research graph: #4368 and #4360.
add provenance to bulktagging
updated label for a counter that I hope I will never see
removed useless comment
refactoring and update of default variable values
remove from context association to Zenodo Community and refactoring of code
propagation of country from institutional repository and of result projects through semantic link
update for zenodo community bulk tagging
fixed DOIBoost Bug
fixed NPE in OrcidEventFactory, improved serialisation of the ORCID in the OpenAIRE event payload
added first implementation for OrcidEventFactory
Implemented ORCID event generation process and relative configuration profileAdded workflow to orchestrate the event generation for software links
aligning with MASTER branch
migrated classes for the FixRelation job from MASTER branch
replaced CrossRef with Crossref
logging index feed retry number
branch for solr 7.5.0
import form master branch
avoid to import non necessary affiliations from DOIBoost
fixed bug for uncompressed abstract in DOIBoostToAction
changed mapping for compressed abstract in DOIBoostToAction
less verbose logging
using proper import for LogFactory
added Mapper and Reducer class for infoSpace counts workflows
added reducer to produce counts on the infospace
cannot use Guava's Splitter.splitToList, must stick to basic split method. Classpath is messed up
implemented ConfigurableExportMapper
Removed un-used import
fixed test
updated Mapper to return the whole invalid record
export invalid xml records
refactored Action
Map only job that produces [openaireId, doi] pairs of records containing invalid characters
added parameter to filter only organization in DOIBoostToAction
fixed problem of missing name in authors
merged beta branch to master
introduced use of BlockProcessor
fixed issue when country information is not present for datasource
change throwing of exception with counters
change parameter from ImmutableBytesWritable to Text
refactoring and change of counters
rollback wrong commit
fixing and testing propagation implementation
reducer for country propagation that writes on hdfs
cleanup pid types in order to make them valid attributes
added code for propagation of countries from institutional organization
master branch for deployments @ICM
why parse strings as Floats?
reverted to r52985 . Test runs shows we need to rely on the edgeIds produced by the connected components identfication phase instead of the vertexIds
alignment to trunk version
avoid to produce duplicated events by eliminating the roots from the comparison process
introduced mapping resulttype -> portal url
Fixed log class name
avoid collisions when hashing pids by value
cleaned up unused method, using setDurability in put operation
added mapper and hadoop job configuration file for importing Grid.AC organization data
integrating bulktag from trunk to beta branch
rule out invalid dates also on CrossRefToActions
rule out invalid dates on ScholixToActions
cleanup
produce 'supplement' subrel type in case of supplement relationships
simplified connected component application on the graph
adding check to understand the bug of wrong relation generated
do not skip processing datasets in DedupBuildRootsMapper, improved error reporting in DedupBuildRootsReducer
do not push vertex ids in memory, process them on the fly
added jobs for predatory journal analysis
added invisible setup
fixed null element
Created CrossrefImportMapper
add CrossRefToAction
fixed mapping from scholix to openaire model
small fixes
changed key type
implemented mapper writing
added configuration
added Mapper for tranform scholexplorer links into actionsets
deprecation: use setDurability instead of setWriteToWAL
introduced subType in pace wf configuration
adjusted ids export procedure
avoid to emit enrichment events when the similarity score is below the threshold
javadoc and test
indentation