- Japanese language and computers
In relation to the Japanese language and computers many adaptation issues arise, some unique to Japanese and others common to
languages which have a very large number of characters. The number of characters needed in order to write English is very small, and thus it is possible to use only one byteto encode one English character. However, the number of characters in Japanese is much more than 256, and hence Japanese cannot be encoded using only one byte, and Japanese is thus encoded using two or more bytes, in a so-called "double byte" or "multi-byte" encoding. Some problems relate to transliterationand romanization, some to character encoding, and some to the inputof Japanese text.
There are several standard methods to encode Japanese characters for use on a computer, including JIS,
Shift-JIS, EUC, and Unicode. While mapping the set of kanais a simple matter, kanjihas proven more difficult. Despite efforts, none of the encoding schemes have become the de facto standard, and multiple encoding standards are still in use today.
For example, most Japanese
JIS encodingand web pages in Shift-JISand yet mobile phones in Japan usually use some form of Extended Unix Code. If a program fails to determine the encoding scheme employed, it can cause " mojibake" (misconverted characters, literally "transformed characters" from the combination of "moji" _ja. 文字 meaning "character" and the stem of "bakeru" _ja. 化ける meaning "to change form") and thus unreadable text on computers.
To understand how this state of affairs has arisen, it is useful to learn a little about the history of the encodings. The first encoding to become widely used was
JIS X 0201, which is a single-byte encoding that only covers standard 7-bit ASCII characters with half-width katakana extensions. This was widely used in systems that were neither powerful enough nor had the storage to handle kanji (including DOS and old embedded equipment such as cash registers). The development of kanji encodings was the beginning of the split. Shift JISwas developed to be completely backward compatible with JIS X 0201, and thus is used in Windows (for backwards compatibility with DOS), and in much embedded electronic equipment.Confusing|date=September 2007 However, Shift JIShas the unfortunate property that it often breaks any parser that is not specifically designed to handle it. EUC, on the other hand, is not backwards compatible with JIS X 0201, but is handled much better by parsers that have been written for 7-bit ASCII (and thus EUCencodings are used on UNIX where much of the file-handling code was historically only written for English encodings). Further complications arise because the original Internet e-mail standards only support 7-bit transfer protocols. Thus JIS encodingwas developed for sending and receiving e-mails.
Not all required characters may be included in a
character setstandard such as JIS, so gaiji( _ja. 外字 "external characters") are sometimes used to supplement the character set. Gaiji may come in the form of external font packs, where normal characters have been replaced with new characters, or the new characters have been added to unused character positions. However, gaiji are not practical in Internetenvironments since the font set must be transferred with text to use the gaiji. As a result, such characters are written with similar or simpler characters in place, or the text may need to be written using a larger character set (such as Unicode) that supports the required character. Unicodeis supposed to solve all encoding problems in all languages of the world. For Japanese, the kanji characters have been unified with Chinese, that is a character considered to be the same in both Japanese and Chinese have been given one and the same code number in Unicode, even if they look a little different. This process, called Han unification, has caused controversy. The previous encodings in Japan, Taiwan, China and Korea have only handled one language and Unicode should handle all. There has been resistance against Unicode in Japan since it is said to be an American invention not Japanese. The handling of Kanji/Chinese have however been designed by a committee of Japanese/Korean/Taiwanese/Chinese people. Unicode is slowly growing because it is better supported by US made software, but still most homepages in Japanese use Shift-JIS. Wikipedia uses Unicode.
Written Japanese uses several different scripts:
kanji(Chinese characters), 2 sets of "kana" (phonetic syllabaries) and roman letters. While kana and roman letters can be typed directly into a computer, entering kanji is a more complicated process as there are far more kanji than there are keys on most keyboards. To input kanji on modern computers, the reading of kanji is usually entered first, then an input method editor(IME), also sometimes known as a front-end processor, shows a list of candidate kanji that are a phonetic match, and allows the user to choose the correct kanji. More-advanced IMEs work not by word but by phrase, thus increasing the likelihood of getting the desired characters as the first option presented. Kanji readings inputs can be either via romanization(" rōmajinyūryoku," Nihongo2|ローマ字入力) or direct kana input ("kana nyūryoku," Nihongo2|かな入力). Direct kana input is not commonly used, but is widely supported.
There are two main systems for the
romanizationof Japanese, known as " Kunrei-shiki" and "Hepburn"; "keyboard romaji" (also known as "wāpuro rōmaji" or "word processor romaji") generally allows a loose combination of both. IME implementations may even handle keys for letters unused in any romanization scheme, such as "L", converting them to the most appropriate equivalent. With kana input, each key on the keyboard directly corresponds to one kana. The JIS keyboard system is national standard, but some people use alternatives like Oyayubi shiftsystem.
Direction of text
Japanese has two directions of writing, called yokogaki and tategaki. The "yokogaki" style is the same as English, but the "tategaki" style involves columns of text written downwards, stacked right to left.
At present, handling of downward text is incomplete. For example,
HTMLhas no support for "tategaki" and Japanese users must use HTML tables to simulate it. However, CSS level 3 includes a property "writing-mode" which can render "tategaki" when given the value "tb-rl" (i.e top to bottom, right to left). Word processors and DTP software have more complete support for it.
Japanese writing system
Korean language and computers
* [http://shittoku.com/default_cat.asp?data_id=32 Japanese Owned computer companies in United States]
* [http://web.archive.org/web/20060527013315/http://www.cs.mcgill.ca/~aelias4/encodings.html A complete introduction to Japanese character encodings]
* [http://examples.oreilly.com/cjkvinfo/doc/cjk.inf Chinese, Japanese, and Korean character set standards and encoding systems]
* [http://lfw.org/text/jp.html Japanese text encoding]
* [http://www.geocities.jp/ep3797/japanese_fonts.html A collection of free Japanese typefaces]
* [http://www.hesjapanese.com How to install japanese font]
Wikimedia Foundation. 2010.