text-icu-0.8.0.3: Bindings to the ICU library
Copyright(c) 2009 2010 Bryan O'Sullivan
LicenseBSD-style
Maintainerbos@serpentine.com
Stabilityexperimental
PortabilityGHC
Safe HaskellSafe-Inferred
LanguageHaskell98

Data.Text.ICU.Normalize2

Description

Character set normalization functions for Unicode, implemented as bindings to the International Components for Unicode (ICU) libraries. See http://www.unicode.org/reports/tr15/ for a description of Unicode normalization modes and why these are needed.

Synopsis

Unicode normalization API

The normalize function transforms Unicode text into an equivalent composed or decomposed form, allowing for easier sorting and searching of text. normalize supports the standard normalization forms described in http://www.unicode.org/unicode/reports/tr15/, Unicode Standard Annex #15: Unicode Normalization Forms.

Characters with accents or other adornments can be encoded in several different ways in Unicode. For example, take the character A-acute. In Unicode, this can be encoded as a single character (the "composed" form):

     00C1    LATIN CAPITAL LETTER A WITH ACUTE

or as two separate characters (the "decomposed" form):

     0041    LATIN CAPITAL LETTER A
     0301    COMBINING ACUTE ACCENT

To a user of your program, however, both of these sequences should be treated as the same "user-level" character "A with acute accent". When you are searching or comparing text, you must ensure that these two sequences are treated equivalently. In addition, you must handle characters with more than one accent. Sometimes the order of a character's combining accents is significant, while in other cases accent sequences in different orders are really equivalent.

Similarly, the string "ffi" can be encoded as three separate letters:

     0066    LATIN SMALL LETTER F
     0066    LATIN SMALL LETTER F
     0069    LATIN SMALL LETTER I

or as the single character

     FB03    LATIN SMALL LIGATURE FFI

The "ffi" ligature is not a distinct semantic character, and strictly speaking it shouldn't be in Unicode at all, but it was included for compatibility with existing character sets that already provided it. The Unicode standard identifies such characters by giving them "compatibility" decompositions into the corresponding semantic characters. When sorting and searching, you will often want to use these mappings.

normalize helps solve these problems by transforming text into the canonical composed and decomposed forms as shown in the first example above. In addition, you can have it perform compatibility decompositions so that you can treat compatibility characters the same as their equivalents. Finally, normalize rearranges accents into the proper canonical order, so that you do not have to worry about accent rearrangement on your own.

Form FCD, "Fast C or D", is also designed for collation. It allows to work on strings that are not necessarily normalized with an algorithm (like in collation) that works under "canonical closure", i.e., it treats precomposed characters and their decomposed equivalents the same.

It is not a normalization form because it does not provide for uniqueness of representation. Multiple strings may be canonically equivalent (their NFDs are identical) and may all conform to FCD without being identical themselves.

The form is defined such that the "raw decomposition", the recursive canonical decomposition of each character, results in a string that is canonically ordered. This means that precomposed characters are allowed for as long as their decompositions do not need canonical reordering.

Its advantage for a process like collation is that all NFD and most NFC texts - and many unnormalized texts - already conform to FCD and do not need to be normalized (NFD) for such a process. The FCD quickCheck will return Yes for most strings in practice.

normalize FCD may be implemented with NFD.

For more details on FCD see the collation design document: http://source.icu-project.org/repos/icu/icuhtml/trunk/design/collation/ICU_collation_design.htm

ICU collation performs either NFD or FCD normalization automatically if normalization is turned on for the collator object. Beyond collation and string search, normalized strings may be useful for string equivalence comparisons, transliteration/transcription, unique representations, etc.

The W3C generally recommends to exchange texts in NFC. Note also that most legacy character encodings use only precomposed forms and often do not encode any combining marks by themselves. For conversion to such character encodings the Unicode text needs to be normalized to NFC. For more usage examples, see the Unicode Standard Annex.

data NormalizationMode Source #

Normalization modes analog (but not identical) to the ones in the Normalize module.

Constructors

NFD

Canonical decomposition.

NFKD

Compatibility decomposition.

NFC

Canonical decomposition followed by canonical composition.

NFKC

Compatibility decomposition followed by canonical composition.

NFKCCasefold

NFKC with Casefold.

normalizer :: NormalizationMode -> IO Normalizer Source #

Create a normalizer for a given normalization mode. This function is more similar to the interface in the Normalize module.

nfcNormalizer :: IO Normalizer Source #

Create an NFC normalizer.

nfdNormalizer :: IO Normalizer Source #

Create an NFD normalizer.

nfkcNormalizer :: IO Normalizer Source #

Create an NFKC normalizer.

nfkdNormalizer :: IO Normalizer Source #

Create an NFKD normalizer.

nfkcCasefoldNormalizer :: IO Normalizer Source #

Create an NFKCCasefold normalizer.

Normalize unicode strings

nfc :: Text -> Text Source #

Create an NFC normalizer and apply this to the given text.

Let's have a look at a concrete example that contains the letter a with an acute accent twice. First as a combination of two codepoints and second as a canonical composite or precomposed character. Both look exactly the same but one character consists of two and one of only one codepoint. A bytewise comparison does not give equality of these.

>>> import Data.Text
>>> let t = pack "a\x301á"
>>> t
"a\769\225"
>>> putStr t
áá
pack "a\x301" == pack "á"
False

But now lets apply some normalization functions and see how these characters coincide afterwards in two different ways:

>>> nfc t
"\225\225"
>>> nfd t
"a\769a\769"

That is exactly what compareUnicode' does:

>>> pack "a\x301" `compareUnicode'` pack "á"

nfd :: Text -> Text Source #

Create an NFD normalizer and apply this to the given text.

nfkc :: Text -> Text Source #

Create an NFKC normalizer and apply this to the given text.

nfkd :: Text -> Text Source #

Create an NFC normalizer and apply this to the given text.

nfkcCasefold :: Text -> Text Source #

Create an NFKCCasefold normalizer and apply this to the given text.

normalize :: NormalizationMode -> Text -> Text Source #

Normalize a string using the given normalization mode.

normalizeWith :: Normalizer -> Text -> Text Source #

Normalize a string with the given normalizer.

Checks for normalization

isNormalizedWith :: Normalizer -> Text -> Bool Source #

Indicate whether a string is in a given normalization form.

Unlike quickCheck, this function returns a definitive result. For NFD and NFKD normalization forms, both functions work in exactly the same ways. For NFC and NFKC forms, where quickCheck may return Nothing, this function will perform further tests to arrive at a definitive result.

Comparison of unicode strings

compareUnicode :: [CompareOption] -> Text -> Text -> Ordering Source #

Compare two strings for canonical equivalence. Further options include case-insensitive comparison and codepoint order (as opposed to code unit order).

Canonical equivalence between two strings is defined as their normalized forms (NFD or NFC) being identical. This function compares strings incrementally instead of normalizing (and optionally case-folding) both strings entirely, improving performance significantly.

Bulk normalization is only necessary if the strings do not fulfill the FCD conditions. Only in this case, and only if the strings are relatively long, is memory allocated temporarily. For FCD strings and short non-FCD strings there is no memory allocation.

compareUnicode' :: Text -> Text -> Ordering Source #

This is equivalent to `compareUnicode []`.

data CompareOption Source #

Options to compare.

Constructors

InputIsFCD

The caller knows that both strings fulfill the FCD conditions. If not set, compare will quickCheck for FCD and normalize if necessary.

CompareIgnoreCase

Compare strings case-insensitively using case folding, instead of case-sensitively. If set, then the following case folding options are used.

FoldCaseExcludeSpecialI

When case folding, exclude the special I character. For use with Turkic (Turkish/Azerbaijani) text data.