Git Personal Project 开源项目分类 人工智能 大模型 企业应用 科学研究 AI写开源项目 孵化优质开源项目 数字人定制 AI工具集合-AI导航
使用标签,可以设置提交历史上的特定点为重要提交
  • v1.0.0
    fb99d0b6 · Update website README.md ·
  • 0.101.0
    2016-05-0 0.101.0: Fixed German model
    
    * Fixed bug that prevented German parses from being deprojectivised.
    * Bug fixes to sentence boundary detection.
    * Add rich comparison methods to the Lexeme class.
    * Add missing Doc.has_vector and Span.has_vector properties.
    * Add missing Span.sent property.
  • 0.100.7
    2016-04-05 v0.100.7: German!
    ----------------------------
    
    spaCy finally supports another language, in addition to English. We're lucky to have Wolfgang Seeker on the team, and the new German model is just the beginning.
    Now that there are multiple languages, you should consider loading spaCy via the load() function. This function also makes it easier to load extra word vector data for English:
    
    .. code:: python
    
        import spacy
        en_nlp = spacy.load('en', vectors='en_glove_cc_300_1m_vectors')
        de_nlp = spacy.load('de')
    
    To support use of the load function, there are also two new helper functions: spacy.get_lang_class and spacy.set_lang_class.
    Once the German model is loaded, you can use it just like the English model:
    
    .. code:: python
    
        doc = nlp(u'''Wikipedia ist ein Projekt zum Aufbau einer Enzyklopädie aus freien Inhalten, zu dem du mit deinem Wissen beitragen kannst. Seit Mai 2001 sind 1.936.257 Artikel in deutscher Sprache entstanden.''')
    
        for sent in doc.sents:
            print(sent.root.text, sent.root.n_lefts, sent.root.n_rights)
    
        # (u'ist', 1, 2)
        # (u'sind', 1, 3)
    
    The German model provides tokenization, POS tagging, sentence boundary detection, syntactic dependency parsing, recognition of organisation, location and person entities, and word vector representations trained on a mix of open subtitles and Wikipedia data. It doesn't yet provide lemmatisation or morphological analysis, and it doesn't yet recognise numeric entities such as numbers and dates.
    
    Bugfixes
    --------
    
    * spaCy < 0.100.7 had a bug in the semantics of the Token.__str__ and Token.__unicode__ built-ins: they included a trailing space.
    * Improve handling of "infixed" hyphens. Previously the tokenizer struggled with multiple hyphens, such as "well-to-do".
    
    * Improve handling of periods after mixed-case tokens
    
    * Improve lemmatization for English special-case tokens
    
    * Fix bug that allowed spaces to be treated as heads in the syntactic parse
    
    * Fix bug that led to inconsistent sentence boundaries before and after serialisation.
    
    * Fix bug from deserialising untagged documents.
  • 0.100.6
    1508528c · * Increment version ·
    Add support for GloVe vectors
    
    This release offers improved support for replacing the word vectors used by spaCy.
    
    To install Stanford's GloVe vectors, trained on the Common Crawl, just run
    
        sputnik --name spacy install en_glove_cc_300_1m_vectors
    
    To reduce memory usage and loading time, we've trimmed the vocabulary down to 1m entries.
    
    This release also integrates all the code necessary for German parsing. A German model will be
    released shortly. To assist in multi-lingual processing, we've added a load() function. To load the English model with the GloVe vectors:
    
        spacy.load('en', vectors='en_glove_cc_300_1m_vectors')
  • 0.100.5
    5d96b3ef · * Increment version ·
    Fix incorrect use of header file, caused from problem with thinc
  • 0.100.4
    Fix OSX problem introduced in 0.100.3
    Small correction to right_edge calculation
  • 0.100.3
    a95974ad · * Fix oov probability ·
    Release 0.100.3
    
    Support multi-threading, via the .pipe method. spaCy now releases the GIL around the
    parser and entity recognizer, so systems that support OpenMP should be able to do
    shared memory parallelism at close to full efficiency.
    
    We've also greatly reduced loading time, and fixed a number of bugs.
  • 0.100.2
    3af84cfd · * Increment version ·
    v0.100.2
    
    Fix data version lock that affected v0.100.1
  • 0.100.1
    7d16f252 · * Update release notes ·
    Fix install for OSX
    
    v0.100 included header files built on Linux that caused installation to fail on OSX.
    This should now be corrected. We also update the default data distribution, to
    include a small fix to the tokenizer.
  • 0.100
    82d011ac · * Fix test for whitespace ·
    Version 0.100: Revise setup.py, better model downloads, bug fixes
    
    * Redo setup.py, and remove ugly headers_workaround hack. Should result in fewer install problems.
    * Update data downloading and installation functionality, by migrating to the Sputnik data-package manager. This will allow us to offer finer grained control of data installation in future.
    * Fix bug when using custom entity types in Matcher. This should work by default when using the
      English.__call__ method of running the pipeline. If invoking Parser.__call__ directly to do NER,
      you should call the Parser.add_label() method to register your entity type.
    * Fix head-finding rules in Span.
    * Fix problem that caused doc.merge() to sometimes hang
    * Fix problems in handling of whitespace
  • 0.99
    Improve span merging, internal refactoring
    
    * Merging multi-word tokens into one, via the doc.merge() and span.merge() methods, no longer invalidates existing Span objects. This makes it much easier to merge multiple spans, e.g. to merge all named entities, or all base noun phrases. Thanks to @andreasgrv for help on this patch.
    * Lots of internal refactoring, especially around the machine learning module, thinc. The thinc API has now been improved, and the spacy._ml wrapper module is no longer necessary.
    * The lemmatizer now lower-cases non-noun, noun-verb and non-adjective words.
    * A new attribute, .rank, is added to Token and Lexeme objects, giving the frequency rank of the word.
  • 0.98
    ee3f9ba5 · * Fix test of serializer ·
    Smaller package, bug fixes
    
    * Remove binary data from PyPi package.
    * Delete archive after downloading data
    * Use updated cymem, preshed and thinc packages
    * Fix information loss in deserialize
    * Fix __str__ methods for Python2
  • 0.97
    9ee1ddab · * Increment version ·
    * Load the StringStore from a json list, instead of a text file
    * Fix bugs in download.py
    * Require --force to over-write the data directory in download.py
    * Fix bugs in Matcher and doc.merge()
  • 0.96
    v0.96: Hotfix to .merge method
    
    * Fix bug that caused text to be lost after .merge
    * Fix bug in Matcher when matched entities overlapped
  • 0.95
    Version v0.95: Bugfixes
    
    * Reform encoding of symbols
    * Fix bugs in Matcher
    * Fix bugs in Span
    * Add tokenizer rule to fix numeric range tokenization
    * Add specific string-length cap in Tokenizaer
    * Fix token.conjuncts
  • 0.94
    a3dfe2b9 · * Increment data version ·
    v0.94
    
    * Fix memory error that caused crashes on 32bit platforms
    * Fix parse errors caused by smart quotes and em-dashes
  • 0.93
    v0.93: Bug fixes to word vectors