Java and Unicode: Difference between revisions
Jump to navigation
Jump to search
Line 11: | Line 11: | ||
=Overview= | =Overview= | ||
Character information is maintained in Java by the primitive type <tt>char</tt>, which was designed based on the original Unicode 1.0 specification that allowed only 2<sup>16</sup> code points, so it was defined as a fixed-with 16-bit/2-byte entity. Since then, [[Character_Encoding#Unicode|the Unicode standard has evolved]] to allow for characters whose representation requires [[Character_Encoding#Unicode_Code_Points|more than 16 bits]] | Character information is maintained in Java by the primitive type <tt>char</tt>, which was designed based on the original Unicode 1.0 specification that allowed only 2<sup>16</sup> code points, so it was defined as a fixed-with 16-bit/2-byte entity. Since then, [[Character_Encoding#Unicode|the Unicode standard has evolved]] to allow for characters whose representation requires [[Character_Encoding#Unicode_Code_Points|more than 16 bits]]. | ||
=U+n Notation Support= | =U+n Notation Support= |
Revision as of 19:00, 26 June 2018
External
- https://docs.oracle.com/javase/10/docs/api/java/lang/Character.html
- http://www.oracle.com/us/technologies/java/supplementary-142654.html
Internal
Overview
Character information is maintained in Java by the primitive type char, which was designed based on the original Unicode 1.0 specification that allowed only 216 code points, so it was defined as a fixed-with 16-bit/2-byte entity. Since then, the Unicode standard has evolved to allow for characters whose representation requires more than 16 bits.
U+n Notation Support
U+n notation is supported in Java as follows: