Scope of Negation with MRS

Packard, W., Bender, E. M., Read, J., Oepen, S., & Dridan, R. (2014).   Simple Negation Scope Resolution Through Deep Parsing: A Semantic Solution to a Semantic Problem.   In Proceedings of the 52nd Meeting of the Association for Computational Linguistics.   Baltimore, MD.

Here you will find the data and code necessary to reproduce the results published in this paper. The system runs under Linux and has dependencies that will make it difficult to run on other systems.
  1. Download the code and data and unpack it.
  2. Decide which experiment you want to reproduce.
  3. To reproduce results reported for the Ranker system, use e.g.:
    $ perl eval.cd-sco.pl -s mrs_experiments/ssd_ranker.txt -g gold/cdd.txt
    
  4. To reproduce results reported for the Crawler system or the combinations, use the following steps. The choice of which system is tested (Crawler, Crawler_N, Crawler_P, oracle) is determined by the combined choices made in steps 2 and 3 below.
    1. Navigate into the uw/scope-predictor/ directory.
    2. Run the system:
      • For Crawler, Crawler_N, or oracle, use e.g.:
        ./cue ../../tsdb/1212/goldcue/ssd.ace.prob
      • For Crawler_P, use e.g.:
        ./cue ../../tsdb/1212/goldcue/ssd.ace.prob 0.5
      This produces a file uw/scope-predictor/connl.txt, which is used by the scripts in the following step, but (contrary to its name) is not exactly CONNL-formatted.

    3. Evaluate the system:
      • For gold cues with Crawler, use e.g.:
        bash scripts/eval-ssd-only.sh
      • For gold cues with Crawler_P or Crawler_N, use e.g.:
        bash scripts/eval-ssd.sh
      • For system cues with Crawler, use e.g.:
        bash scripts/eval-ssd-only-syscue.sh
      • For system cues with Crawler_P or Crawler_N, use e.g.:
        bash scripts/eval-ssd-syscue.sh
      • For oracle comparisons (with gold cues only), use e.g.:
        bash scripts/eval-ssd-oracle.sh
      The scoring script tends to produce a few error messages about uninitialized values, which should be ignored.
Please note that evaluations will not be meaningful if the evaluation script invoked does not match the data profile used in the cue invocation, in terms of the dataset and the cue type (system vs. gold). The parsed profiles in tsdb/1212 were created using the ACE parser (version 0.9.16) with the English Resource Grammar (version 1212), with support for the CONNL token format provided by [incr tsdb()].