37,720
社区成员
发帖
与我相关
我的任务
分享
import itertools
#import nltk
from nltk.tokenize import StanfordSegmenter
from nltk.tokenize import StanfordTokenizer
from nltk.tag import StanfordPOSTagger
from nltk.parse.stanford import StanfordParser
from nltk.parse import stanford
from nltk.tag.api import TaggerI
parser = StanfordParser(
path_to_jar='C:\\ProgramData\\Anaconda3\\stanfordparser\\stanford-parser.jar',
path_to_models_jar='C:\\ProgramData\\Anaconda3\\stanfordparser\\stanford-parser-3.6.0-models.jar',
model_path='C:\\ProgramData\\Anaconda3\\stanfordparser\\stanford-parser-3.6.0-models\\edu\\stanford\\nlp\\models\\lexparser\\englishPCFG.ser.gz')
#parser = StanfordParser('C:\\ProgramData\\Anaconda3\\stanfordparser\\stanford-parser.jar','C:\\ProgramData\\Anaconda3\\stanfordparser\\slf4j-api.jar','C:\\ProgramData\\Anaconda3\\stanfordparser\\data\\english-onesent.txt')
lt = parser.parse("the quick brown fox jumps over the lazy dog".split())
print(list(lt))
def __init__(self, path_to_jar=None, path_to_models_jar=None,
model_path='edu/stanford/nlp/models/lexparser/englishPCFG.ser.gz',
encoding='utf8', verbose=False,
java_options='-mx1000m', corenlp_options=''):
# find the most recent code and model jar
stanford_jar = max(
find_jar_iter(
self._JAR, path_to_jar,
env_vars=('STANFORD_PARSER', 'STANFORD_CORENLP'),
searchpath=(), url=_stanford_url,
verbose=verbose, is_regex=True
),
key=lambda model_name: re.match(self._JAR, model_name)
)