-
Notifications
You must be signed in to change notification settings - Fork 288
Initial code for wikidata-based normalizer, untested #380
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: master
Are you sure you want to change the base?
Conversation
|
Sorry, still got lots of build errors. |
|
@chile12 could you test this out with a few languages? Maybe start up all the mappings langs with nohup. Here's the command I use: Specify your base-dir in the config file. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
add wiki-code in the filename
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
if the script is re-started with another language withouth deleting the file it will produce false results
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Are you talking about the cache file? wikiFinder is instantiated with the baseDir and language - so for enwiki it creates the file enwiki-YYYYMMDD-sameas-mappings.obj in the respective directory.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
perfect, you save them on the lang folder not the wikidata one, I got confused
…into normalize Conflicts: scripts/src/main/scala/org/dbpedia/extraction/scripts/NormalizeDatasets.scala
Haven't tested this on sample data yet, mustn't merge. @chile12 could you check this?