doc.noun_chunks is not supported for Chinese language, how to figure this out? · Issue #7436 · explosion/spaCy

By GitHub - 2021-03-16

Description

I want to extract noun chunks from the Chinese text, but it seems that doc.noun_chunks is not supported for Chinese language, what shall I do?

Summary

  • What exactly defines a "noun chunk" is highly language dependent, so each language in spaCy has its own implementation of this.
  • The way you would add it to a language is to add or modify the syntax_iterators.py file for your language and define a noun_chunks function.
  • You can see the English version here: You could use a similar approach for Chinese.

 

Topics

  1. NLP (0.34)

Similar Articles

8 Awesome React Hooks

By DEV Community - 2020-12-30

React.js is currently the most popular JavaScript library for front end developers. Invented by Faceb... Tagged with react, javascript, webdev, beginners.