This repository explains how to analyze Japanese texts by J-LIWC2015. J-LIWC2015 is a Japanese-translated version of the LIWC2015 dictionary (Pennebaker, Booth, Jordan, & Blackburn, 2015). Psychometric properties of the J-LIWC2015 dictionary are introduced in Igarashi, Okuda, & Sasahara (2022).
Unlike English, there is no word boundary in a sentence in Japanese. It is mandatory for natural language processing to segment a Japanese text document into words before the main analysis. This repository provides basic instructions for using J-LIWC2015 and sample scripts for preprocessing texts for word segmentation.
Make sure you have downloaded one of the J-LIWC2015 dictionary files (valid serial number is required).
- LIWC2015 format:
Japanese_Dictionary.dic
- available from http://www.liwc.net/dictionaries/
- LIWC-22 format:
LIWC2015 Dictionary - Japanese.dicx
- available from https://www.liwc.app/dictionaries/
You can analyze Japanese texts in either of the following ways. Non-commercial users of the LIWC2015/LIWC-22 software can use the dictionary file in other programming languages (Python, R, etc.) with MeCab/IPADIC. This means that the users can seamlessly integrate the findings of preprocessing, main analysis, and postprocessing (optional) (see the example in Colab or R). Note that the developer does not formally support the use of the dictionary outside of the LIWC software (please do it at your own risk). The compatibility of the outputs generated by and outside of the LIWC software is also not guaranteed.
- Main analysis: Use the LIWC-22 software with the J-LIWC2015 dictionary (
LIWC2015 Dictionary - Japanese.dicx
) for category-by-category word frequency analysis. Wword segmentation is conducted internally.
- Python (Colab) (LIWC2015 dictionary format): Preprocessing and postprocessing (by @soramame0518 and @okudashi)
- Preprocessing: Analyze a Japanese text (e.g.,
sample.txt
) by MeCab/IPADIC withuser_dict.dic
(in this repository) in Python. In the example in Colab, two output files are generated:wakachi.txt
: word segmentation outputpos_rate.txt
: POS tagging output
- Main analysis: Run the LIWC2015 software, click
Dictionary
→Load New Dictionary
, and chooseJapanese_Dictionary.dic
. Then analyzewakachi.txt
(word segmentation output). An output file is:LIWC2015 Results (wakachi).txt
: word frequency analysis output
- Postprocessing (optional): Combine
LIWC2015 Results (wakachi).txt
(word frequency analysis output) withpos_rate.txt
(POS tagging output) so that users can obtain more detailed information about POS of the text (see the example in Colab). A merged output file is:result.txt
: word frequency and POS tagging output
- Python (Colab) (LIWC2015 dictionary format): Preprocessing, main analysis, and postprocessing (by @soramame0518 and @okudashi)
- R (LIWC2015 and LIWC-22 dictionary formats): The jliwc package can be used for preprocessing, main analysis, and postprocessing. MeCab/IPADIC and the user dictionary will be automatically installed and a POS tagging output will be included in the results of the main analysis. See Analyzing Japanese texts with R and J-LIWC2015 for a tutorial in Japanese.
Any request for the distribution of the dictionary file is not accepted. Questions about the commercial use of J-LIWC2015 should be directed to Receptiviti.
Igarashi, T., Okuda, S., & Sasahara, K. (2022). Development of the Japanese Version of the Linguistic Inquiry and Word Count Dictionary 2015. Frontiers in Psychology, 13:841534. https://doi.org/10.3389/fpsyg.2022.841534
MIT