You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+9-6Lines changed: 9 additions & 6 deletions
Original file line number
Diff line number
Diff line change
@@ -1,4 +1,4 @@
1
-
# Python interface to Stanford Core NLP tools v1.3.1
1
+
# Python interface to Stanford Core NLP tools v1.3.3
2
2
3
3
This is a Python wrapper for Stanford University's NLP group's Java-based [CoreNLP tools](http://nlp.stanford.edu/software/corenlp.shtml). It can either be imported as a module or run as a JSON-RPC server. Because it uses many large trained models (requiring 3GB RAM on 64-bit machines and usually a few minutes loading time), most applications will probably want to run it as a server.
4
4
@@ -10,7 +10,7 @@ This is a Python wrapper for Stanford University's NLP group's Java-based [CoreN
10
10
11
11
It requires [pexpect](http://www.noah.org/wiki/pexpect) and (optionally) [unidecode](http://pypi.python.org/pypi/Unidecode) to handle non-ASCII text. This script includes and uses code from [jsonrpc](http://www.simple-is-better.org/rpc/) and [python-progressbar](http://code.google.com/p/python-progressbar/).
12
12
13
-
It runs the Stanford CoreNLP jar in a separate process, communicates with the java process using its command-line interface, and makes assumptions about the output of the parser in order to parse it into a Python dict object and transfer it using JSON. The parser will break if the output changes significantly, but it has been tested on **Core NLP tools version 1.3.1** released 2012-04-09.
13
+
It runs the Stanford CoreNLP jar in a separate process, communicates with the java process using its command-line interface, and makes assumptions about the output of the parser in order to parse it into a Python dict object and transfer it using JSON. The parser will break if the output changes significantly, but it has been tested on **Core NLP tools version 1.3.3** released 2012-07-09.
14
14
15
15
## Download and Usage
16
16
@@ -19,10 +19,10 @@ To use this program you must [download](http://nlp.stanford.edu/software/corenlp
19
19
In other words:
20
20
2
FECA
1
21
sudo pip install pexpect unidecode # unidecode is optional
@@ -113,6 +113,7 @@ To use it in a regular script or to edit/debug it (because errors via RPC are op
113
113
corenlp.parse("Parse it")
114
114
115
115
<!--
116
+
116
117
## Adding WordNet
117
118
118
119
Note: wordnet doesn't seem to be supported using this approach. Looks like you'll need Java.
@@ -122,6 +123,7 @@ tar xvfz WNprolog-3.0.tar.gz
122
123
123
124
-->
124
125
126
+
125
127
## Questions
126
128
127
129
**Stanford CoreNLP tools require a large amount of free memory**. Java 5+ uses about 50% more RAM on 64-bit machines than 32-bit machines. 32-bit machine users can lower the memory requirements by changing `-Xmx3g` to `-Xmx2g` or even less.
@@ -131,6 +133,7 @@ If pexpect timesout while loading models, check to make sure you have enough mem
131
133
132
134
You can reach me, Dustin Smith, by sending a message on GitHub or through email (contact information is available [on my webpage](http://web.media.mit.edu/~dustin)).
133
135
136
+
134
137
# Contributors
135
138
136
139
This is free and open source software and has benefited from the contribution and feedback of others. Like Stanford's CoreNLP tools, it is covered under the [GNU General Public License v2 +](http://www.gnu.org/licenses/gpl-2.0.html), which in short means that modifications to this program must maintain the same free and open source distribution policy.
0 commit comments