Explore chapters and articles related to this topic
Globalization, Localization, and Cross-Cultural User-Interface Design
Published in Julie A. Jacko, The Human–Computer Interaction Handbook, 2012
Preparing texts in local languages often requires the use of additional or different characters. The American Standard Code for Information (ASCII) system, which uses seven or eight bits to represent characters, supports English, and the single-byte ISO 8859–1 character set supports Western European languages that use the Latin alphabet, such as Spanish, French, and German. Other character encoding systems include EBCDIC, Shift-JIS, UTF-8, and UTF-16. ISO has established specific character sets for languages such as Cyrillic, Modern Greek, Hebrew, Japanese, and so on. However, most companies planning to globalize their products should use Unicode, a double-byte (16 bit) system that can represent 65,536 characters, which is sufficient to display Asian languages like Japanese and Korean and permits easier translation and presentation of character sets.
Character Mapping and Code Sets
Published in Cliff Wootton, Developing Quality Metadata, 2009
These are the implications of using the different Unicode Transformation Formats: A document encoded in UTF-32 could be twice as large as a UTF-16 encoded version, because the UTF-32 scheme requires four bytes to encode any character, while UTF-16 only uses two bytes for the characters inside the basic multilingual plane. Characters outside the Basic Multilingual Plane are rare. You’ll seldom need to encode in UTF-32.The character encoding in UTF-8 is a variable length format and uses between one and four bytes to encode a character. UTF-8 may encode more or less economically than UTF-16, which always encodes in 2 or more bytes. UTF-8 is more economic than UTF-16 when characters having a low (one byte) value predominate.UTF-7 encoding is more compact than other Unicode encodings with quoted-printable or BASE64 encodings when operated in a 7-bit environment but it is difficult to parse.
Data Types and Data Storage
Published in Julio Sanchez, Maria P. Canton, Microcontroller Programming, 2018
Julio Sanchez, Maria P. Canton
One of the limitations of the ASCII code is that eight bits are not enough for representing characters sets in languages such as Japanese or Chinese which use large character sets. This has led to the development of encodings which allow representing large character sets. Unicode has been proposed as a universal character encoding standard that can be used for representation of text for computer processing.
Automatic extraction of materials and properties from superconductors scientific literature
Published in Science and Technology of Advanced Materials: Methods, 2023
Luca Foppiano, Pedro Baptista Castro, Pedro Ortiz Suarez, Kensei Terashima, Yoshihiko Takano, Masashi Ishii
We developed Grobid-superconductors as a Grobid module following principles (multi-step, sentence-based, full-text-based) discussed in a previous preliminary study [19]. Grobid has several advantages: 1) it can be integrated with pdfalto (https://github.com/kermitt2/pdfalto), a specialised tool for converting PDF to XML, which mitigates extraction issues such as the resolution of embedded fonts, invalid character encoding, and the reconstruction of the correct reading order, 2) it allows access to PDF document layout information for both machine learning and document decoration (e.g. coordinates in the PDF document); and, 3) it provides access to a set of high-quality, pre-trained machine learning models for structuring documents. Grobid-superconductors is structured as a three-steps process illustrated in Figure 1 and described in the Sections 2.1, 2.2, and 2.3.
Extending brain-computer interface access with a multilingual language model in the P300 speller
Published in Brain-Computer Interfaces, 2022
P Loizidou, E Rios, A Marttini, O Keluo-Udeke, J Soetedjo, J Belay, K Perifanos, N Pouratian, W Speier
An additional technical complication when using a non-English language is that characters may not be in the ASCII character encoding used by most computer programs (including BCI2000) and programming languages. As an alternative, the 8-bit Unicode Transformation Format (UTF-8) can be used to represent these characters. However, an added difficulty arises because this format has a variable bit-width that can be difficult to represent in a system dependent on string lengths and character locations. We handled this situation by creating a lookup table where each character was represented by a single 8-bit code. These codes were used for the character representations within the language model and during the computation of character probabilities. Lookups were performed at the input and output stages of the system so that it would be able to read in the language model in UTF-8 format and then present the output as Greek characters through the graphical interface.
Optimization of Discrete Anamorphic Stretch Transform and Phase Recovery for ECG Signal Compression
Published in IETE Journal of Research, 2021
R. Thilagavathy, B. Venkataramani
Hybrid data compression is used for IoT wireless sensors in healthcare applications [18]. Adaptive Fourier decomposition and symbol substitution techniques are used for pervasive e-health applications [19]. High performance and dynamic compression schemes are used in wireless Biosensors [20]. The compressed sensing using dictionaries for Telecardiology application ([21], [22]), lossless compression on multichannel ECG ([23], [24]), compression using discrete orthogonal Stockwell transform-DCT [25], and ASCII character encoding of optimum singular values [26] are reported in the literature. In [27], empirical mode of decomposition (EMD) and wavelet transform are combined for ECG data compression. A detailed review of the Electrocardiogram Data Compression Techniques for Cardiac Healthcare Systems reported in the literature are presented in [28]. Recently, Wireless ECG and cardiac monitoring systems have been adopted in commercial devices [29].