You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
{{ message }}
This repository was archived by the owner on Feb 1, 2022. It is now read-only.
keilexandra edited this page Jul 17, 2012
·
7 revisions
Two separate possibilities exist for how large amounts of data processed through the Curator might be made available to the larger NLP community. We would need:
For pre-processed annotations upon request:
A dynamic web page interface cataloging all annotations presently available for download. (In PHP?)
A way to get specific Records from the Master Curator's database.
For user-submitted data returned with annotations:
A separate downloadable reader, easily configurable by the user to parse documents into Thrift-serializable text files.
A simple web page interface, implemented in PHP, which permits the user to upload a tarball of properly read document records.
A message sent to the Master Curator telling it to launch a MapReduce job(s) via shell scripts.