Content deleted Content added
m Dating maintenance tags: {{When}} |
Darklanlan (talk | contribs) +use of half-width kana and gaiji, recent usage share of encodes in Internet |
||
Line 4:
==Character encodings==
There are several standard methods to [[character encoding|encode]] Japanese characters for use on a computer, including [[JIS encoding|JIS]], [[Shift-JIS]], [[Extended Unix Code|EUC]], and [[Unicode]].
▲There are several standard methods to [[character encoding|encode]] Japanese characters for use on a computer, including [[JIS encoding|JIS]], [[Shift-JIS]], [[Extended Unix Code|EUC]], and [[Unicode]]. While mapping the set of [[kana]] is a simple matter, [[kanji]] has proven more difficult. Despite efforts, none of the encoding schemes have become the de facto standard, and multiple encoding standards are still in use today.
For example, most Japanese [[email]]s
[[File:PC-9801F Kanji ROM board.jpg|thumb|Kanji [[Read-only memory|ROM]] card installed in [[PC-9800 series|PC-98]], which stored about 3000 glyphs, and enabled to display them in fast. It also had a [[Random-access memory|RAM]] to store gaiji.]]
[[File:Control panel of public background music system.jpg|thumb|Display of background music system using [[half-width kana]]]]
The first encoding to become widely used was [[JIS X 0201]], which is a [[ISO 646|single-byte encoding]] that only covers standard 7-bit ASCII characters with [[Half-width kana|half-width katakana]] extensions. This was widely used in systems that were neither powerful enough nor had the storage to handle kanji (including old embedded equipment such as cash registers) because kanji input required a complicated process, and kanji display required much memory and high resolution. This means that only
The development of kanji encodings was the beginning of the split.
▲The first encoding to become widely used was [[JIS X 0201]], which is a [[ISO 646|single-byte encoding]] that only covers standard 7-bit ASCII characters with half-width katakana extensions. This was widely used in systems that were neither powerful enough nor had the storage to handle kanji (including old embedded equipment such as cash registers). This means that only [[katakana]], not kanji, was supported using this technique. Some embedded displays still have this limitation.
▲The development of kanji encodings was the beginning of the split. [[Shift JIS]] supports kanji and was developed to be completely backward compatible with [[JIS X 0201]], and thus is in much embedded electronic equipment.
However, [[Shift JIS]] has the unfortunate property that it often breaks any parser (software that reads the coded text) that is not specifically designed to handle it. For example, a text search method can get false hits if it is not designed for Shift JIS. [[Extended Unix Code|EUC]], on the other hand, is handled much better by parsers that have been written for 7-bit ASCII (and thus [[Extended Unix Code|EUC]] encodings are used on UNIX, where much of the file-handling code was historically only written for English encodings). But EUC is not backwards compatible with JIS X 0201, the first main Japanese encoding. Further complications arise because the original Internet e-mail standards only support 7-bit transfer protocols. Thus {{IETF RFC|1468}} ("[[ISO-2022-JP]]", often simply called [[JIS encoding]]) was developed for sending and receiving e-mails.
[[File:Japanese TV closed caption using gaiji.jpg|thumb|[[Gaiji]] is used in closed caption of Japanese TV broadcasting]]
In [[character set]] standards such as [[JIS X 0208|JIS]], not all required characters are included, so [[gaiji]] ({{lang|ja|外字}} "external characters") are sometimes used to supplement the character set. Gaiji may come in the form of external font packs, where normal characters have been replaced with new characters, or the new characters have been added to unused character positions. However, gaiji are not practical in [[Internet]] environments since the font set must be transferred with text to use the gaiji.
[[Unicode]] was intended to solve all encoding problems over all languages. The [[UTF-8]] encoding used to encode Unicode in web pages does not have the disadvantages that Shift-JIS has. Unicode is supported by international software, and it eliminates the need for gaiji. There are still controversies, however. For Japanese, the kanji characters have been [[Han unification|unified]] with Chinese; that is, a character considered to be the same in both Japanese and Chinese is given a single number, even if the appearance is actually somewhat different, with the precise appearance left to the use of a locale-appropriate font. This process, called [[Han unification]], has caused controversy. The previous encodings in Japan, [[Free area of the Republic of China|Taiwan Area]], [[Mainland China]] and [[Korea]] have only handled one language and Unicode should handle all. The handling of Kanji/Chinese have however been designed by a committee composed of representatives from all four countries/areas. As of 2011, Unicode
▲In [[character set]] standards such as [[JIS X 0208|JIS]], not all required characters are included, so [[gaiji]] ({{lang|ja|外字}} "external characters") are sometimes used to supplement the character set. Gaiji may come in the form of external font packs, where normal characters have been replaced with new characters, or the new characters have been added to unused character positions. However, gaiji are not practical in [[Internet]] environments since the font set must be transferred with text to use the gaiji. As a result, such characters are written with similar or simpler characters in place, or the text may need to be encoded using a larger character set (such as Unicode) that supports the required character.
▲[[Unicode]] was intended to solve all encoding problems over all languages. The [[UTF-8]] encoding used to encode Unicode in web pages does not have the disadvantages that Shift-JIS has. Unicode is supported by international software, and it eliminates the need for gaiji. There are still controversies, however. For Japanese, the kanji characters have been [[Han unification|unified]] with Chinese; that is, a character considered to be the same in both Japanese and Chinese is given a single number, even if the appearance is actually somewhat different, with the precise appearance left to the use of a locale-appropriate font. This process, called [[Han unification]], has caused controversy. The previous encodings in Japan, [[Free area of the Republic of China|Taiwan Area]], [[Mainland China]] and [[Korea]] have only handled one language and Unicode should handle all. The handling of Kanji/Chinese have however been designed by a committee composed of representatives from all four countries/areas. Unicode is slowly growing because it is better supported by software from outside Japan, but still (as of 2011) most web pages in Japanese use Shift-JIS. The [[Japanese Wikipedia]] uses Unicode.
== Text input ==
|