Widely used coding scheme used to represent characters in a personal computer

While every effort has been made to follow citation style rules, there may be some discrepancies. Please refer to the appropriate style manual or other sources if you have any questions.

Select Citation Style

Copy Citation

Share

Share

Share to social media

Facebook Twitter

URL

https://www.britannica.com/topic/ASCII

Give Feedback

External Websites

Feedback

Corrections? Updates? Omissions? Let us know if you have suggestions to improve this article (requires login).

Feedback Type

Your Feedback Submit Feedback

Thank you for your feedback

Our editors will review what you’ve submitted and determine whether to revise the article.

External Websites

  • Engineering and Technology History Wiki - ASCII

Britannica Websites

Articles from Britannica Encyclopedias for elementary and high school students.

  • ASCII (American Standard Code for Information Interchange) - Student Encyclopedia (Ages 11 and up)

Print Cite

verifiedCite

While every effort has been made to follow citation style rules, there may be some discrepancies. Please refer to the appropriate style manual or other sources if you have any questions.

Select Citation Style

Copy Citation

Share

Share

Share to social media

Facebook Twitter

URL

https://www.britannica.com/topic/ASCII

Feedback

External Websites

Feedback

Corrections? Updates? Omissions? Let us know if you have suggestions to improve this article (requires login).

Feedback Type

Your Feedback Submit Feedback

Thank you for your feedback

Our editors will review what you’ve submitted and determine whether to revise the article.

External Websites

  • Engineering and Technology History Wiki - ASCII

Britannica Websites

Articles from Britannica Encyclopedias for elementary and high school students.

  • ASCII (American Standard Code for Information Interchange) - Student Encyclopedia (Ages 11 and up)

Alternate titles: American Standard Code for Information Interchange

By The Editors of Encyclopaedia Britannica Article History

Table of Contents

Related Topics:ASCII art extended ASCII coding system...(Show more)

See all related content →

ASCII, in full American Standard Code for Information Interchange, a standard data-encoding format for electronic communication between computers. ASCII assigns standard numeric values to letters, numerals, punctuation marks, and other characters used in computers.

Before ASCII was developed, different makes and models of computers could not communicate with one another. Each computer manufacturer represented alphabets, numerals, and other characters in its own way. IBM (International Business Machines Corporation) alone used nine different character sets. In 1961 Bob Bemer of IBM submitted a proposal to the American National Standards Institute (ANSI) for a common computer code. The X3.4 committee, with representation from key computer manufacturers of the day, was formed to work on the new code. On June 17, 1963, ASCII was approved as the American standard. However, it did not gain wide acceptance, mainly because IBM chose to use EBCDIC (Extended Binary Coded Decimal Interchange Code) in its OS/360 series of computers released in 1964. Nevertheless, ASCII underwent further development, and revisions were issued in 1965 and 1967. On March 11, 1968, U.S. Pres. Lyndon B. Johnson mandated that ASCII be termed a federal standard to minimize incompatibility across federal computer and telecommunications systems. Furthermore, he mandated that all new computers and related equipment purchased by the U.S. government from July 1, 1969, onward should be ASCII-compatible. The code was revised again in 1968, 1977, and 1986.

ASCII was originally developed for teleprinters, or teletypewriters, but it eventually found wide application in personal computers (PCs), beginning with IBM’s first PC, in 1981. ASCII uses seven-digit binary numbers—i.e., numbers consisting of various sequences of 0’s and 1’s. Since there are 128 different possible combinations of seven 0’s and 1’s, the code can represent 128 different characters. The binary sequence 1010000, for example, represents an uppercase P, while the sequence 1110000 represents a lowercase p.

Digital computers use a binary code that is arranged in groups of eight, rather than seven, digits, or bits; each such eight-bit group is called a byte. Consequently, ASCII is commonly embedded in an eight-bit field, which consists of the seven information bits and a parity bit that is used for error checking or for representing special symbols. This eight-bit system increases the number of characters ASCII can represent to 256, and it ensures that all special characters, as well as characters from other languages, can be represented. Extended ASCII, as the eight-bit code is known, was introduced by IBM in 1981 for use in its first PC, and it soon became the industry standard for personal computers. In extended ASCII, 32 code combinations are used for machine and control commands, such as “start of text,” “carriage return,” and “form feed.” Control commands do not represent printable information, but rather they help control devices, such as printers, that may use ASCII. For example, the binary sequence 00001000 represents “backspace.” Another group of 32 combinations is used for numerals and various punctuation marks, another for uppercase letters and a few other punctuation marks, and yet another for lowercase letters.

However, even extended ASCII does not include enough code combinations to support all written languages. Asian languages, for instance, require thousands of characters. This limitation gave rise to new encoding standards—Unicode and UCS (Universal Coded Character Set)—that can support all the principal written languages. Because it incorporates ASCII as its first 128 code combinations, Unicode (specifically UTF-8) is backward-compatible with ASCII while also representing many characters that ASCII cannot. Unicode, which was introduced in 1991, saw its usage jump sharply in the first decade of the 21st century, and it became the most common character-encoding system on the World Wide Web.

Which coding scheme is typically used for personal computers?

Almost all computers now use ASCII or Unicode encoding. The exceptions are some IBM mainframes that use the proprietary 8-bit code called Extended Binary Coded Decimal Interchange Code (EBCDIC).

What are different codes that can be used to represent a character in a computer?

ASCII uses seven-digit binary numbers—i.e., numbers consisting of various sequences of 0's and 1's. Since there are 128 different possible combinations of seven 0's and 1's, the code can represent 128 different characters.

What are the 4 types of coding?

Common styles are imperative, functional, logical, and object-oriented languages. Programmers can choose from these coding language paradigms to best-serve their needs for a specific project.

What is EBCDIC and Unicode?

The term UTF-EBCDIC stands for EBCDIC-friendly Unicode (or UCS) Transformation Format. EBCDIC, IBM's Extended Binary Coded Decimal Interchange Code, is one of the widely used 8-bit industry encodings.