Fork me on GitHub

Welcome to janome’s documentation! (English)


What’s Janome?


Illustration by janome. All rights reserved.

Janome (蛇の目; ◉) is a Japanese morphological analysis engine (or tokenizer, pos-tagger) written in pure Python including the built-in dictionary and the language model.

We aim to build a library which is easy to install and provides concise, well-designed APIs for various python applications.

Janome uses mecab-ipadic-2.7.0-20070801 as the built-in dictionary. Also Japanese new era “令和” (Reiwa) has been added to the dictionary since v0.3.8.

Source Code

If you like janome, please star the repository! :)


Python 2.7.x or Python 3.5+ interpreter

Latest version

  • janome: 0.3.10



$ pip install janome


This requires 500 to 600 MB RAM for pre-compiling the dictionary data.


Create janome.tokenizer.Tokenizer object and call tokenize() method with the sentences you want to analyze.

The return value is a list of Token objects. Token includes morphologic information such as surface form, part-of-speech. See reference for more details.

>>> from janome.tokenizer import Tokenizer
>>> t = Tokenizer()
>>> for token in t.tokenize(u'すもももももももものうち'):
...     print(token)
すもも 名詞,一般,*,*,*,*,すもも,スモモ,スモモ
も    助詞,係助詞,*,*,*,*,も,モ,モ
もも  名詞,一般,*,*,*,*,もも,モモ,モモ
も    助詞,係助詞,*,*,*,*,も,モ,モ
もも  名詞,一般,*,*,*,*,もも,モモ,モモ
の    助詞,連体化,*,*,*,*,の,ノ,ノ
うち  名詞,非自立,副詞可能,*,*,*,うち,ウチ,ウチ

for Windows users

Use decode('utf8') if the output is garbled.

>>> from janome.tokenizer import Tokenizer
>>> t = Tokenizer()
>>> for token in t.tokenize(u'すもももももももものうち'):
...     print(str(token).decode('utf8'))

How to use with user defined dictionary

MeCab IPADIC format

You can add custom entries besides the built-in dictionary at runtime by using user defined dictionary.

Default dictionary format is equal to MeCab IPADIC format. Create a CSV file as below and pass the file path and the character encoding to Tokenizer’s constructor.

See the MeCab document for more details.


>>> from janome.tokenizer import Tokenizer
>>> t = Tokenizer("userdic.csv", udic_enc="utf8")
>>> for token in t.tokenize(u'東京スカイツリーへのお越しは、東武スカイツリーライン「とうきょうスカイツリー駅」が便利です。'):
...   print(token)

東京スカイツリー         名詞,固有名詞,一般,*,*,*,東京スカイツリー,トウキョウスカイツリー,トウキョウスカイツリー
へ        助詞,格助詞,一般,*,*,*,へ,ヘ,エ
の        助詞,連体化,*,*,*,*,の,ノ,ノ
お越し    名詞,一般,*,*,*,*,お越し,オコシ,オコシ
は        助詞,係助詞,*,*,*,*,は,ハ,ワ
、        記号,読点,*,*,*,*,、,、,、
東武スカイツリーライン    名詞,固有名詞,一般,*,*,*,東武スカイツリーライン,トウブスカイツリーライン,トウブスカイツリーライン
「        記号,括弧開,*,*,*,*,「,「,「
とうきょうスカイツリー駅  名詞,固有名詞,一般,*,*,*,とうきょうスカイツリー駅,トウキョウスカイツリーエキ,トウキョウスカイツリーエキ
」        記号,括弧閉,*,*,*,*,」,」,」
が        助詞,格助詞,一般,*,*,*,が,ガ,ガ
便利      名詞,形容動詞語幹,*,*,*,*,便利,ベンリ,ベンリ
です      助動詞,*,*,*,特殊・デス,基本形,です,デス,デス
。        記号,句点,*,*,*,*,。,。,。

Simplified dictionary format (v0.2.7+)

Janome provides alternative simplified dictionary format similar to Kuromoji user dictionary. This format supports surface form, part-of-speech and reading only.

To use simplified dictionary format, create a CSV file that includes “<surface form>,<part-of-speech>,<reading>” in each line (see the example below) and pass the file path and the dictionary type (udic_type='simpledic') to Tokenizer’s constructor.


>>> from janome.tokenizer import Tokenizer
>>> t = Tokenizer("user_simpledic.csv", udic_type="simpledic", udic_enc="utf8")
>>> for token in t.tokenize(u'東京スカイツリーへのお越しは、東武スカイツリーライン「とうきょうスカイツリー駅」が便 利です。'):
...   print(token)
東京スカイツリー     カスタム名詞,*,*,*,*,*,東京スカイツリー,トウキョウスカイツリー,トウキョウスカイツリー
へ    助詞,格助詞,一般,*,*,*,へ,ヘ,エ
の    助詞,連体化,*,*,*,*,の,ノ,ノ
お越し    名詞,一般,*,*,*,*,お越し,オコシ,オコシ
は    助詞,係助詞,*,*,*,*,は,ハ,ワ
、    記号,読点,*,*,*,*,、,、,、
東武スカイツリーライン   カスタム名詞,*,*,*,*,*,東武スカイツリーライン,トウブスカイツリーライン,トウブスカイツリーライン
「    記号,括弧開,*,*,*,*,「,「,「
とうきょうスカイツリー駅    カスタム名詞,*,*,*,*,*,とうきょうスカイツリー駅,トウキョウスカイツリーエキ,トウキョウスカイツリーエキ
 」   記号,括弧閉,*,*,*,*,」,」,」
が    助詞,格助詞,一般,*,*,*,が,ガ,ガ
便利    名詞,形容動詞語幹,*,*,*,*,便利,ベンリ,ベンリ
です    助動詞,*,*,*,特殊・デス,基本形,です,デス,デス
。    記号,句点,*,*,*,*,。,。,。

Pre-compiled user dictionary

With large user dictionary, it can take much time to convert CSV file to the binary data structure. You can compile the user dictionary in advance and use that at runtime.

For now, there is no tools for compiling user dictionary. Use APIs as below.

How to compile user dictionary (MeCab IPADIC format):

>>> from janome.dic import UserDictionary
>>> import sysdic
>>> user_dict = UserDictionary("userdic.csv", "utf8", "ipadic", sysdic.connections)

How to compile user dictionary (simplified format):

>>> from janome.dic import UserDictionary
>>> import sysdic
>>> user_dict = UserDictionary("user_simpledic.csv", "utf8", "simpledic", sysdic.connections)

Once compiling has been successfully completed, the data is saved in /tmp/userdic directory. Pass the directory path to Tokenizer’s constructor to use it.

>>> t = Tokenizer("/tmp/userdic")


Use same major python version at both compile time and runtime. Forward/backward dictionary data compatibility is not guaranteed.

(experimental) Analyzer framework (v0.3.4+)

Analyzer framework is for pre- and post- processing. Analyzer framework includes following classes.

  • CharFilter class performs pre-processing such as character normalization.

  • TokenFilter class performs post-processing such as lowercase/uppercase conversion, token filtering by POS tags.

  • Analyzer class combines CharFilters, a Tokenizer and TokenFilters to assemble custom analysis chain.

Analyzser usage

>>> from janome.tokenizer import Tokenizer
>>> from janome.analyzer import Analyzer
>>> from janome.charfilter import *
>>> from janome.tokenfilter import *
>>> text = u'蛇の目はPure Pythonな形態素解析器です。'
>>> char_filters = [UnicodeNormalizeCharFilter(), RegexReplaceCharFilter(u'蛇の目', u'janome')]
>>> tokenizer = Tokenizer()
>>> token_filters = [CompoundNounFilter(), POSStopFilter(['記号','助詞']), LowerCaseFilter()]
>>> a = Analyzer(char_filters, tokenizer, token_filters)
>>> for token in a.analyze(text):
...     print(token)
janome  名詞,固有名詞,組織,*,*,*,*,*,*
pure    名詞,固有名詞,組織,*,*,*,*,*,*
python  名詞,一般,*,*,*,*,*,*,*
な       助動詞,*,*,*,特殊・ダ,体言接続,だ,ナ,ナ
形態素解析器  名詞,複合,*,*,*,*,形態素解析器,ケイタイソカイセキキ,ケイタイソカイセキキ
です     助動詞,*,*,*,特殊・デス,基本形,です,デス,デス

Analyzer usage for word count (v0.3.5+)

You can count word frequencies in the input text by using TokenCountFilter.

>>> from janome.tokenizer import Tokenizer
>>> from janome.analyzer import Analyzer
>>> from janome.tokenfilter import *
>>> text = u'すもももももももものうち'
>>> token_filters = [POSKeepFilter(['名詞']), TokenCountFilter()]
>>> a = Analyzer(token_filters=token_filters)
>>> for k, v in a.analyze(text):
...   print('%s: %d' % (k, v))
すもも: 1
もも: 2
うち: 1

See API reference for other built-in CharFilters and TokenFilters. You can implement custom filters by extending CharFilter or TokenFilter.

Streaming mode (v0.3.1+)

When stream = True option is given to tokenize() method, it runs on streaming mode. On streaming mode, partial analyzed results are returned through generator interface.

Use this option when you analyze very large text data.

t = Tokenizer()
with open('very_large_text.txt') as f:
    txt =
    for token in t.tokenize(txt, stream=True):

‘wakati-gaki’ mode (v0.3.1+)

When ‘wakati = True’ option is given to tokenize() method, it runs on ‘wakati-gaki’ (‘分かち書き’) mode. On wakati-gaki mode, tokenize() method returns sufrace forms only. Return type is a list of string, not list of Token.

>>> t = Tokenizer()
>>> tokens = t.tokenize(u'分かち書きモードがつきました!', wakati=True)
>>> tokens
['分かち書き', 'モード', 'が', 'つき', 'まし', 'た', '!']

If you use ‘wakati-gaki’ mode only, it is recommended to give wakati = True option to Tokenizer.__init__(). When Tokenizer object is initialized as below, extra information (detailed part of speech, reading, etc.) for tokens are not loaded from dictionary so the memory usage is reduced.

>>> t = Tokenizer(wakati=True)

When this option is passed to Tokenizer object, tokenize() method always runs in wakati-gaki mode (wakati = False option is ignored.)

‘wakati-gaki’ mode works well with streaming mode. tokenize() method returns generator of string when it is given stream=True and wakati=True options.

t = Tokenizer()
for token in t.tokenize(txt, stream=True, wakati=True):

Memory-mapped file support (v0.3.3+)

If mmap=True option is given to Tokenizer.__init__(), dictionary entries are not loaded to process space but searched through memory-mapped file.

Graphviz file (DOT file) support (v0.3.7+)

When dotfile=<dotfile output path> option is given, Tokenizer.tokenize() method converts the lattice graph to Graphviz DOT file. For performance reasons, this option is ignored when running on streaming mode or analyzing very long text.

janome command (the details are mentioned later) has options to visualize the lattice graph easily.

Command-line interface (Linux/Mac v0.2.6+, Windows v0.3.7+)

Janome has executable built-in script “janome” for command-line usage.

It reads a sentence at a time from standard input and outputs the analyzed results. To see supported options, type “janome -h”.


(env)$ janome
猫    名詞,一般,*,*,*,*,猫,ネコ,ネコ
は    助詞,係助詞,*,*,*,*,は,ハ,ワ
液体  名詞,一般,*,*,*,*,液体,エキタイ,エキタイ
で    助動詞,*,*,*,特殊・ダ,連用形,だ,デ,デ
ある  助動詞,*,*,*,五段・ラ行アル,基本形,ある,アル,アル
(Type Ctrl-C to quit.)


Use -e sjis option if the output is garbled.

>janome -e sjis
ウィンドウズ    名詞,固有名詞,一般,*,*,*,ウィンドウズ,ウィンドウズ,ウィンドウズ
簡単    名詞,形容動詞語幹,*,*,*,*,簡単,カンタン,カンタン
インストール    名詞,一般,*,*,*,*,インストール,インストール,インストール
(Type Ctrl-Z to quit.)

Visualizing lattice graph


You need Graphviz to use this functionality. Please install Graphviz from here.

If given -g option, janome command outputs the visualized lattice image to the current directory after analyzing the text. Default output format is PNG.

$ echo "カレーは飲み物" | janome -g
カレー 名詞,一般,*,*,*,*,カレー,カレー,カレー
は   助詞,係助詞,*,*,*,*,は,ハ,ワ
飲み物 名詞,一般,*,*,*,*,飲み物,ノミモノ,ノミモノ
Graph was successfully output to lattice.gv.png

lattice.gv.png (Click to show the full size image.)


You can change the output file location by using --gv-out option. Also you can specify the output file format by --gv-format option. See Graphviz documentation for all supported output formats.

$ echo "カレーは飲み物" | janome -g --gv-out /tmp/a.gv --gv-format svg
Graph was successfully output to /tmp/a.gv.svg

How to bundle janome with an application by PyInstaller (v0.3.9+)

You can create (and distribute) stand-alone executables which bundle janome by PyInstaller.

(venv) $ janome --version
janome 0.3.9
(venv) $ pyinstaller -v

(venv) $ cat
# -*- utf-8
from janome.tokenizer import Tokenizer
t = Tokenizer()
    for token in t.tokenize('令和元年'):

(venv) $ pyinstaller --onefile
44 INFO: PyInstaller: 3.4
44 INFO: Python: 3.6.6

(venv) $ ls dist/
(venv) $ ./dist/test
令和  名詞,固有名詞,一般,*,*,*,令和,レイワ,レイワ
元年  名詞,一般,*,*,*,*,元年,ガンネン,ガンネン

Note for analyzing large document set


This memory leak problem was solved at v0.3. The analyzed results with version 0.3 or over can be a bit different from ones with version 0.2. You may want to examine streaming and/or wakati-gaki mode to reduce memory usage more.

In older version (< 0.3), Janome can consume large memory when a very large document is passed all at once. Please split large documents (larger than tens of killobytes) into small chunks or sentences.


  1. How is the accuracy of analysis?

  1. Janome uses MeCab IPADIC dictionary, so the accuracy is roughly same to MeCab.

  1. How is the speed of analysis?

  1. Basically depends on the input length. According to my benchmark script, one sentence would take a few milliseconds to a few tens of milliseconds on commodity PCs.

  1. What data structures and algorithms are used?

  1. Janome uses FST (Minimal Acyclic Subsequential Transducer) for internal dictionary data structure. I implemented the automaton by referring to Apache Lucene (written in Java) and kagome (written in Go). And for analysis engine, I implemented basic viterbi algorithm by referring the book 自然言語処理の基礎 .

  1. I found bugs. Or have requests for enhancement.

  1. Bug reports or requests (and of course, patches) are welcome. Create issues in Github repository or leave your comment to Gitter room.




Licensed under Apache License 2.0 and uses the MeCab-IPADIC dictionary/statistical model.

See LICENSE.txt and NOTICE.txt for license details.


  • 2019.11.03 janome Version 0.3.10 released

  • 2019.05.12 janome Version 0.3.9 released

  • 2019.04.03 janome Version 0.3.8 released

  • 2018.12.11 janome Version 0.3.7 released

  • 2017.12.07 janome Version 0.3.6 released

  • 2017.08.06 janome Version 0.3.5 released

  • 2017.07.29 janome Version 0.3.4 released

  • 2017.07.23 janome Version 0.3.3 released

  • 2017.07.05 janome Version 0.3.2 released

  • 2017.07.02 janome Version 0.3.1 released

  • 2017.06.30 janome Version 0.3.0 released

  • 2016.05.07 janome Version 0.2.8 released

  • 2016.03.05 janome Version 0.2.7 released

  • 2015.10.26 janome Version 0.2.6 released

  • 2015.05.11 janome Version 0.2.5 released

  • 2015.05.03 janome Version 0.2.4 released

  • 2015.05.03 janome Version 0.2.3 released

  • 2015.04.24 janome Version 0.2.2 released

  • 2015.04.24 janome Version 0.2.0 released

  • 2015.04.11 janome Version 0.1.4 released

  • 2015.04.08 janome Version 0.1.3 released

Change details: CHANGES


About this badge