Krakozyabry- What kind of word is this interesting? This word is usually used by Russian users to describe the incorrect/incorrect display (encoding) of characters in programs or the Operating System itself.
Why does this happen? You won't find a definite answer. This may be due to the tricks of our “favorite” viruses, perhaps due to a malfunction of the Windows OS (for example, the electricity went out and the computer turned off), perhaps the program created a conflict with another OS and everything went haywire. IN general reasons there can be many, but the most interesting one is “It just got up and broke.”
Read the article and find out how to fix the problem with encoding in programs and Windows OS, once it has happened.
For those who still don’t understand what I mean, here are a few:
By the way, I also found myself in this situation once and I still have a file on my desktop that helped me cope with it. That's why I decided to write this article.
Several “things” are responsible for displaying the encoding (font) in Windows - the language, the registry, and the files of the OS itself. Now we will check them separately and point by point.
How to remove and correct krakozyabry instead of Russian (Russian letters) in a program or Windows.
1. Checking installed language for programs that do not support Unicode. Maybe it's lost on you.
So, let's follow the path: Control Panel - Regional and Language Options - Advanced tab
There we make sure that the language is Russian.
In Windows XP, in addition to this, there is a list below " Code pages conversion tables" and there is a line in it with the number 20880. It is necessary that there is also Russian
6.
Last point, in which I give you a file that helped me fix everything once and that’s why I left it as a keepsake. Here is the archive:
There are two files inside: krakozbroff.cmd and krakozbroff.reg
They have the same principle - correct hieroglyphs, squares, questions or exclamation marks in programs and Windows OS (in common parlance krakozyabry). I used the first one and it helped me.
And finally, a couple of tips:
1) If you work with the registry, do not forget to make a backup ( backup copy) in case something goes wrong.
2) It is advisable to check the 1st point after each point.
That's all. Now you know how to fix/remove Crackers (squares, hieroglyphs, exclamation and question marks) in a program or Windows.
Hello, dear readers of the blog site. Today we will talk to you about where krakozyabrs come from on a website and in programs, what text encodings exist and which ones should be used. Let's take a closer look at the history of their development, starting from basic ASCII, as well as its extended versions CP866, KOI8-R, Windows 1251 and ending with modern Unicode Consortium encodings UTF 16 and 8.
To some, this information may seem unnecessary, but would you know how many questions I receive specifically regarding the crawling krakozyabrs (unreadable set of characters). Now I will have the opportunity to refer everyone to the text of this article and find my own mistakes. Well, get ready to absorb the information and try to follow the flow of the story.
ASCII - basic text encoding for the Latin alphabet
The development of text encodings occurred simultaneously with the formation of the IT industry, and during this time they managed to undergo quite a lot of changes. Historically, it all started with EBCDIC, which was rather dissonant in Russian pronunciation, which made it possible to encode letters of the Latin alphabet, Arabic numerals and punctuation marks with control characters.
But still, the starting point for the development of modern text encodings should be considered the famous ASCII(American Standard Code for Information Interchange, which in Russian is usually pronounced as “aski”). It describes the first 128 characters most commonly used by English-speaking users - , Arabic numerals and punctuation marks.
These 128 characters described in ASCII also included some service characters like brackets, hash marks, asterisks, etc. In fact, you can see them yourself:
It is these 128 characters from the original version of ASCII that have become the standard, and in any other encoding you will definitely find them and they will appear in this order.
But the fact is that with one byte of information you can encode not 128, but as many as 256 different values (two to the power of eight equals 256), so after the basic version of Asuka a whole series of extended ASCII encodings, in which, in addition to 128 basic characters, it was also possible to encode symbols of the national encoding (for example, Russian).
Here, it’s probably worth saying a little more about the number systems that are used in the description. Firstly, as you all know, a computer only works with numbers in the binary system, namely with zeros and ones (“Boolean algebra”, if anyone took it at an institute or school). , each of which is a two to the power, starting from zero, and up to two to the seventh:
It is not difficult to understand that all possible combinations of zeros and ones in such a design can only be 256. Converting a number from the binary system to the decimal system is quite simple. You just need to add up all the powers of two with ones above them.
In our example, this turns out to be 1 (2 to the power of zero) plus 8 (two to the power of 3), plus 32 (two to the fifth power), plus 64 (to the sixth power), plus 128 (to the seventh power). Total gets 233 in decimal system Reckoning. As you can see, everything is very simple.
But if you look closely at the table with ASCII characters, you will see that they are represented in hexadecimal encoding. For example, "asterisk" corresponds to the hexadecimal number 2A in Aski. You probably know that in the hexadecimal number system, in addition to Arabic numerals, Latin letters from A (means ten) to F (means fifteen) are also used.
Well then, for translation binary number to hexadecimal resort to the following simple and obvious method. Each byte of information is divided into two parts of four bits, as shown in the above screenshot. That. In each half byte, only sixteen values (two to the fourth power) can be encoded in binary, which can easily be represented as a hexadecimal number.
Moreover, in the left half of the byte the degrees will need to be counted again starting from zero, and not as shown in the screenshot. As a result, through simple calculations, we get that the number E9 is encoded in the screenshot. I hope that the course of my reasoning and the solution to this puzzle were clear to you. Well, now let’s continue, in fact, talking about text encodings.
Extended versions of Asuka - CP866 and KOI8-R encodings with pseudographics
So, we started talking about ASCII, which was, as it were, the starting point for the development of all modern encodings (Windows 1251, Unicode, UTF 8).
Initially, it contained only 128 characters of the Latin alphabet, Arabic numerals and something else, but in the extended version it became possible to use all 256 values that can be encoded in one byte of information. Those. It became possible to add symbols of letters of your language to Aski.
Here we will need to digress again to explain - why do we need encodings at all? texts and why it is so important. The characters on your computer screen are formed on the basis of two things - sets of vector forms (representations) of various characters (they are located in files with ) and code that allows you to pull out from this set of vector forms (font file) exactly the character that will need to be inserted into Right place.
It is clear that the fonts themselves are responsible for the vector shapes, but the operating system and the programs used in it are responsible for the encoding. Those. any text on your computer will be a set of bytes, each of which encodes one single character of this very text.
The program that displays this text on the screen (text editor, browser, etc.), when parsing the code, reads the encoding of the next character and looks for the corresponding vector form in the required file font that is connected to display this text document. Everything is simple and banal.
This means that in order to encode any character we need (for example, from the national alphabet), two conditions must be met - the vector form of this character must be in the font used and this character could be encoded in extended ASCII encodings in one byte. Therefore, there are a whole bunch of such options. Just for encoding Russian language characters, there are several varieties of extended Aska.
For example, originally appeared CP866, which had the ability to use characters from the Russian alphabet and was an extended version of ASCII.
Those. her top part completely coincided with the basic version of Aska (128 Latin characters, numbers and other crap), which is presented in the screenshot just above, but the lower part of the table with CP866 encoding had the form indicated in the screenshot just below and allowed you to encode another 128 characters (Russian letters and all sorts of pseudo-graphics):
You see, in the right column the numbers start with 8, because... numbers from 0 to 7 refer to the basic part of ASCII (see first screenshot). That. The Russian letter "M" in CP866 will have the code 9C (it is located at the intersection of the corresponding row with 9 and column with the number C in the hexadecimal number system), which can be written in one byte of information, and if there is a suitable font with Russian characters, this letter without problems will appear in the text.
Where did this amount come from? pseudographics in CP866? The whole point is that this encoding for Russian text was developed back in those shaggy years when graphical operating systems were not as widespread as they are now. And in Dosa and similar text operating systems, pseudographics made it possible to at least somehow diversify the design of texts, and therefore CP866 and all its other peers from the category of extended versions of Asuka abound in it.
CP866 was distributed by IBM, but in addition to this, a number of encodings were developed for Russian language characters, for example, the same type (extended ASCII) can be attributed KOI8-R:
The principle of its operation remains the same as that of the CP866 described a little earlier - each character of text is encoded by one single byte. The screenshot shows the second half of the KOI8-R table, because the first half is completely consistent with the basic Asuka, which is shown in the first screenshot in this article.
Among the features of the KOI8-R encoding, it can be noted that the Russian letters in its table are not in alphabetical order, as, for example, they did it in CP866.
If you look at the very first screenshot (of the basic part, which is included in all extended encodings), you will notice that in KOI8-R Russian letters are located in the same cells of the table as the corresponding letters of the Latin alphabet from the first part of the table. This was done for the convenience of switching from Russian to Latin characters by discarding just one bit (two to the seventh power or 128).
Windows 1251 - the modern version of ASCII and why the cracks come out
The further development of text encodings was due to the fact that graphical operating systems were gaining popularity and the need to use pseudographics in them disappeared over time. As a result, a whole group arose that, in essence, were still extended versions of Asuka (one character of text is encoded with just one byte of information), but without the use of pseudographic symbols.
They belonged to the so-called ANSI encodings, which were developed by the American Standards Institute. In common parlance, the name Cyrillic was also used for the version with Russian language support. An example of this would be.
It differed favorably from the previously used CP866 and KOI8-R in that the place of pseudographic symbols in it was taken by the missing symbols of Russian typography (except for the accent mark), as well as symbols used in Slavic languages close to Russian (Ukrainian, Belarusian, etc.). ):
Due to such an abundance of Russian language encodings, font manufacturers and manufacturers software headaches constantly arose, and you and I, dear readers, often got those same notorious krakozyabry when there was confusion with the version used in the text.
Very often they appeared when sending and receiving messages via e-mail, which entailed the creation of very complex conversion tables, which, in fact, were fundamentally unable to solve this problem, and users often used for correspondence to avoid the notorious gimmicks when using Russian encodings like CP866, KOI8-R or Windows 1251.
In fact, the krakozyabrs appearing instead of the Russian text were the result of incorrect use of the encoding of this language, which did not match the one in which it was encoded text message initially.
For example, if you try to display characters encoded using CP866 using code table Windows 1251, then these same gibberish (a meaningless set of characters) will come out, completely replacing the text of the message.
A similar situation very often arises on forums or blogs, when text with Russian characters is mistakenly saved in the wrong encoding that is used on the site by default, or in the wrong encoding text editor, which adds gags to the code that are not visible to the naked eye.
In the end, many people got tired of this situation with a lot of encodings and constantly creeping out crap, and the prerequisites appeared for the creation of a new universal variation that would replace all the existing ones and would finally solve the problem with the appearance of unreadable texts. In addition, there was the problem of languages like Chinese, where there were much more language characters than 256.
Unicode - universal encodings UTF 8, 16 and 32
These thousands of characters of the Southeast Asian language group could not possibly be described in one byte of information that was allocated for encoding characters in extended versions of ASCII. As a result, a consortium was created called Unicode(Unicode - Unicode Consortium) with the collaboration of many IT industry leaders (those who produce software, who encode hardware, who create fonts), who were interested in the emergence of a universal text encoding.
The first variation released under the auspices of the Unicode Consortium was UTF 32. The number in the encoding name means the number of bits that are used to encode one character. 32 bits equal 4 bytes of information that will be needed to encode one single character in the new universal UTF encoding.
As a result, the same file with text encoded in the extended version of ASCII and in UTF-32, in the latter case, will have a size (weigh) four times larger. This is bad, but now we have the opportunity to encode using YTF a number of characters equal to two to the thirty-second power ( billions of characters, which will cover any really necessary value with a colossal margin).
But many countries with languages of the European group did not need to use such a huge number of characters in encoding at all, however, when using UTF-32, they for no reason received a fourfold increase in weight text documents, and as a result, an increase in the volume of Internet traffic and the amount of stored data. This is a lot, and no one could afford such waste.
As a result of the development of Unicode, UTF-16, which turned out to be so successful that it was adopted by default as the base space for all the characters that we use. It uses two bytes to encode one character. Let's see how this thing looks.
In the Windows operating system, you can follow the path “Start” - “Programs” - “Accessories” - “System Tools” - “Character Table”. As a result, a table will open with the vector shapes of all the fonts installed on your system. If you select in " Additional options» set of Unicode characters, you can see for each font separately the entire range of characters included in it.
By the way, by clicking on any of them, you can see its two-byte code in UTF-16 format, consisting of four hexadecimal digits:
How many characters can be encoded in UTF-16 using 16 bits? 65,536 (two to the power of sixteen), and this is the number that was adopted as the base space in Unicode. In addition, there are ways to encode about two million characters using it, but they were limited to an expanded space of a million characters of text.
But even this successful version of the Unicode encoding did not bring much satisfaction to those who wrote, for example, programs only in English language, because after the transition from the extended version of ASCII to UTF-16, the weight of documents doubled (one byte per character in Aski and two bytes per the same character in UTF-16).
It was precisely to satisfy everyone and everything in the Unicode consortium that it was decided to come up with variable length encoding. It was called UTF-8. Despite the eight in the name, it actually has a variable length, i.e. Each character of text can be encoded into a sequence of one to six bytes in length.
In practice, UTF-8 only uses the range from one to four bytes, because beyond four bytes of code it is no longer even theoretically possible to imagine anything. All Latin characters in it are encoded into one byte, just like in the good old ASCII.
What is noteworthy is that in the case of encoding only the Latin alphabet, even those programs that do not understand Unicode will still read what is encoded in YTF-8. Those. the core part of Asuka was simply transferred to this creation of the Unicode consortium.
Cyrillic characters in UTF-8 are encoded in two bytes, and, for example, Georgian characters are encoded in three bytes. The Unicode Consortium, after creating UTF 16 and 8, solved the main problem - now we have fonts have a single code space. And now their manufacturers can only fill it with vector forms of text characters based on their strengths and capabilities. Now they even come in sets.
In the “Character Table” above you can see that different fonts support different numbers of characters. Some Unicode-rich fonts can be quite heavy. But now they differ not in the fact that they were created for different encodings, but in the fact that the font manufacturer has filled or not completely filled the single code space with certain vector forms.
Crazy words instead of Russian letters - how to fix it
Let's now see how krakozyabrs appear instead of text or, in other words, how the correct encoding for Russian text is selected. Actually, it is set in the program in which you create or edit this very text, or code using text fragments.
For editing and creating text files Personally, I use a very good one, in my opinion, . However, it can highlight the syntax of hundreds of other programming and markup languages, and also has the ability to be extended using plugins. Read detailed review this wonderful program at the link provided.
IN top menu Notepad++ has an “Encodings” item, where you will have the opportunity to convert an existing option to the one used by default on your site:
In the case of a site on Joomla 1.5 and higher, as well as in the case of a blog on WordPress, you should select the option to avoid the appearance of cracks UTF 8 without BOM. What is the BOM prefix?
The fact is that when they were developing the YUTF-16 encoding, for some reason they decided to attach to it such a thing as the ability to write the character code both in direct sequence (for example, 0A15) and in reverse (150A). And in order for programs to understand exactly in what sequence to read the codes, it was invented BOM(Byte Order Mark or, in other words, signature), which was expressed in adding three additional bytes to the very beginning of the documents.
In the UTF-8 encoding, no BOMs were provided for in the Unicode consortium, and therefore adding a signature (those notorious extra three bytes at the beginning of the document) simply prevents some programs from reading the code. Therefore, when saving files in UTF, we must always select the option without BOM (without signature). So you are in advance protect yourself from crawling krakozyabrs.
What is noteworthy is that some programs in Windows cannot do this (they cannot save text in UTF-8 without a BOM), for example, the same notorious Windows Notepad. It saves the document in UTF-8, but still adds the signature (three extra bytes) to the beginning of it. Moreover, these bytes will always be the same - read the code in direct sequence. But on servers, because of this little thing, a problem can arise - crooks will come out.
Therefore, under no circumstances Don't use regular Windows notepad to edit documents on your site if you don’t want any cracks to appear. The best and most simple option I consider the already mentioned Notepad++ editor, which has practically no disadvantages and consists only of advantages.
In Notepad++, when you select an encoding, you will have the option to convert text to UCS-2 encoding, which is very close in nature to the Unicode standard. Also in Notepad it will be possible to encode text in ANSI, i.e. in relation to the Russian language, this will be Windows 1251, which we have already described just above. Where does this information come from?
It is registered in your register operating system Windows - which encoding to choose in the case of ANSI, which to choose in the case of OEM (for the Russian language it will be CP866). If you set another default language on your computer, then these encodings will be replaced with similar ones from the ANSI or OEM category for that same language.
After you save the document in Notepad++ in the encoding you need or open the document from the site for editing, you can see its name in the lower right corner of the editor:
To avoid rednecks, in addition to the actions described above, it will be useful to write in its header source code all pages of the site information about this very encoding, so that there is no confusion on the server or local host.
In general, all hypertext markup languages except Html use a special xml declaration, which specifies the text encoding.
Before parsing the code, the browser knows which version is being used and how exactly it needs to interpret the character codes of that language. But what’s noteworthy is that if you save the document in the default Unicode, then this xml declaration can be omitted (the encoding will be considered UTF-8 if there is no BOM or UTF-16 if there is a BOM).
In the case of an Html language document, the encoding is used to indicate Meta element, which is written between the opening and closing Head tags:
... ...
This entry is quite different from the one adopted in, but is fully consistent with the new one being gradually introduced HTML standard 5, and it will be 100% correctly understood by any currently used browsers.
In theory, it would be better to place a Meta element indicating the Html document encoding as high as possible in the document header so that at the time of encountering the first character in the text not from the basic ANSI (which are always read correctly and in any variation), the browser should already have information on how to interpret the codes of these characters.
Good luck to you! See you soon on the pages of the blog site
You might be interested
What are URL addresses, how do absolute and relative links for a site differ?
OpenServer - modern local server and an example of its use to install WordPress on a computer
What is Chmod, what permissions to assign to files and folders (777, 755, 666) and how to do it via PHP
Yandex search by site and online store
In the early 90s, when the collapse of the USSR occurred and the borders of Russia were open, we began to receive software products Western production. Naturally, they were all English-speaking. At the same time, the Internet began to develop. The problem of Russification of resources and programs has become acute. Then the Russian Windows encoding 1251 was invented. It allows you to correctly display the letters of the Slavic alphabets:
- Russian;
- Ukrainian;
- Belarusian;
- Serbian;
- Bulgarian;
- Macedonian
The development was carried out by the Russian representative office of Microsoft together with the companies Dialog and Paragraph. The basis was taken from self-written developments, which in 1990-91 were in circulation among the few IT ideologists in Russia.
To date, more than universal method encode characters - UTF-8 (Unicode). It represents almost 90% of all software and web resources. Windows 1251 is used in 1.6% of cases. (Information from Web Technology Surveys)
Website encoding utf 8 or windows 1251?
To answer this question, you need to understand a little about what an encoding is and how they differ. Text information, like any other information, is stored in a computer in encoded form. It's easier for us to think of it as numbers. Each character can occupy one or more bytes. Windows 1251 is a single-byte encoding, while UTF-8 is an eight-byte encoding. This means that only 256 characters can be encoded in Windows 1251.
Since everything comes down to the binary system, and a byte is 8 bits (0 and 1), the maximum number of combinations is 28 = 256. Unicode allows you to represent a much larger number of characters, and a larger size can be allocated for each.
This is where the advantages of Unicode follow:
- In the site header you should indicate the encoding that is used. Otherwise, “crazy symbols” will be displayed instead of symbols. And Unicode is standard for all browsers - they catch it “on the fly” as installed by default.
- The site symbols will remain the same no matter in which country the resource is loaded. This does not depend on the geographical location of the servers, but on the software language of the client workstations. A resident of Portugal obviously uses the keyboard and all software, including the operating system, in his native language. His computer most likely does not have Windows 1251 at all. And if this is the case, then sites in Russian will not open correctly. Unicode, in turn, is “hardwired” into any OS in any language.
- UTF-8 allows encoding large quantity characters. On this moment 6 bytes out of 8 are used, and Russian characters are encoded in two bytes.
That is why it is preferable to use a universal encoding rather than a highly specialized one, which is used only in Slavic countries.
Windows 1251 encoding table
For programmers and website developers, it is sometimes necessary to know character numbers. For this purpose, special encoding tables are used. Below is the table for Windows 1251.
What to do if the command line encoding is lost?
Sometimes you may encounter a situation where strange characters are displayed on the command line instead of Russian ones. This means that there is a command encoding problem Windows strings 7. Why 7? Because, starting from version 8, UTF-8 is used, and in version 7, Windows 1251 is also used.
The chcp 866 command can help solve the problem at one time. The current session will work correctly. But to fix the error radically, you will need a registry.
![](https://i2.wp.com/windowstips.ru/wp-content/uploads/2017/10/kdrvk-wndws-10-2.jpg)
Have a great day!
Windows-1251 - character set and encoding, which is the standard 8-bit encoding for all Russian versions Microsoft Windows. This encoding is quite popular in Eastern European countries. Windows-1251 differs favorably from other 8-bit Cyrillic encodings (such as CP866, KOI8-R and ISO 8859-5) by the presence of almost all the characters used in traditional Russian typography for plain text(only the accent mark is missing). Cyrillic characters are in alphabetical order.
Windows-1251 also contains all the characters for languages close to Russian: Belarusian, Ukrainian, Serbian, Macedonian and Bulgarian.
In practice, this turned out to be enough for the Windows-1251 encoding to gain a foothold on the Internet until the spread of UTF-8.
Dec | Hex | Symbol | Dec | Hex | Symbol | |
---|---|---|---|---|---|---|
000 | 00 | NOP | 128 | 80 | Ђ | |
001 | 01 | SOH | 129 | 81 | Ѓ | |
002 | 02 | STX | 130 | 82 | ‚ | |
003 | 03 | ETX | 131 | 83 | ѓ | |
004 | 04 | EOT | 132 | 84 | „ | |
005 | 05 | ENQ | 133 | 85 | … | |
006 | 06 | ACK | 134 | 86 | † | |
007 | 07 | BEL | 135 | 87 | ‡ | |
008 | 08 | B.S. | 136 | 88 | € | |
009 | 09 | TAB | 137 | 89 | ‰ | |
010 | 0A | LF | 138 | 8A | Љ | |
011 | 0B | VT | 139 | 8B | ‹ | |
012 | 0C | FF | 140 | 8C | Њ | |
013 | 0D | CR | 141 | 8D | Ќ | |
014 | 0E | SO | 142 | 8E | Ћ | |
015 | 0F | S.I. | 143 | 8F | Џ | |
016 | 10 | DLE | 144 | 90 | ђ | |
017 | 11 | DC1 | 145 | 91 | ‘ | |
018 | 12 | DC2 | 146 | 92 | ’ | |
019 | 13 | DC3 | 147 | 93 | “ | |
020 | 14 | DC4 | 148 | 94 | ” | |
021 | 15 | N.A.K. | 149 | 95 | ||
022 | 16 | SYN | 150 | 96 | – | |
023 | 17 | ETB | 151 | 97 | — | |
024 | 18 | CAN | 152 | 98 | ||
025 | 19 | E.M. | 153 | 99 | ™ | |
026 | 1A | SUB | 154 | 9A | љ | |
027 | 1B | ESC | 155 | 9B | › | |
028 | 1C | FS | 156 | 9C | њ | |
029 | 1D | G.S. | 157 | 9D | ќ | |
030 | 1E | R.S. | 158 | 9E | ћ | |
031 | 1F | US | 159 | 9F | џ | |
032 | 20 | SP | 160 | A0 | ||
033 | 21 | ! | 161 | A1 | Ў | |
034 | 22 | " | 162 | A2 | ў | |
035 | 23 | # | 163 | A3 | Ћ | |
036 | 24 | $ | 164 | A4 | ¤ | |
037 | 25 | % | 165 | A5 | Ґ | |
038 | 26 | & | 166 | A6 | ¦ | |
039 | 27 | " | 167 | A7 | § | |
040 | 28 | ( | 168 | A8 | Yo | |
041 | 29 | ) | 169 | A9 | © | |
042 | 2A | * | 170 | A.A. | Є | |
043 | 2B | + | 171 | AB | « | |
044 | 2C | , | 172 | A.C. | ¬ | |
045 | 2D | - | 173 | AD | | |
046 | 2E | . | 174 | A.E. | ® | |
047 | 2F | / | 175 | A.F. | Ї | |
048 | 30 | 0 | 176 | B0 | ° | |
049 | 31 | 1 | 177 | B1 | ± | |
050 | 32 | 2 | 178 | B2 | І | |
051 | 33 | 3 | 179 | B3 | і | |
052 | 34 | 4 | 180 | B4 | ґ | |
053 | 35 | 5 | 181 | B5 | µ | |
054 | 36 | 6 | 182 | B6 | ¶ | |
055 | 37 | 7 | 183 | B7 | · | |
056 | 38 | 8 | 184 | B8 | e | |
057 | 39 | 9 | 185 | B9 | № | |
058 | 3A | : | 186 | B.A. | є | |
059 | 3B | ; | 187 | BB | » | |
060 | 3C | < | 188 | B.C. | ј | |
061 | 3D | = | 189 | BD | Ѕ | |
062 | 3E | > | 190 | BE | ѕ | |
063 | 3F | ? | 191 | B.F. | ї | |
064 | 40 | @ | 192 | C0 | A | |
065 | 41 | A | 193 | C1 | B | |
066 | 42 | B | 194 | C2 | IN | |
067 | 43 | C | 195 | C3 | G | |
068 | 44 | D | 196 | C4 | D | |
069 | 45 | E | 197 | C5 | E | |
070 | 46 | F | 198 | C6 | AND | |
071 | 47 | G | 199 | C7 | Z | |
072 | 48 | H | 200 | C8 | AND | |
073 | 49 | I | 201 | C9 | Y | |
074 | 4A | J | 202 | C.A. | TO | |
075 | 4B | K | 203 | C.B. | L | |
076 | 4C | L | 204 | CC | M | |
077 | 4D | M | 205 | CD | N | |
078 | 4E | N | 206 | C.E. | ABOUT | |
079 | 4F | O | 207 | CF | P | |
080 | 50 | P | 208 | D0 | R | |
081 | 51 | Q | 209 | D1 | WITH | |
082 | 52 | R | 210 | D2 | T | |
083 | 53 | S | 211 | D3 | U | |
084 | 54 | T | 212 | D4 | F | |
085 | 55 | U | 213 | D5 | X | |
086 | 56 | V | 214 | D6 | C | |
087 | 57 | W | 215 | D7 | H | |
088 | 58 | X | 216 | D8 | Sh | |
089 | 59 | Y | 217 | D9 | SCH | |
090 | 5A | Z | 218 | D.A. | Kommersant | |
091 | 5B | [ | 219 | D.B. | Y | |
092 | 5C | \ | 220 | DC | b | |
093 | 5D | ] | 221 | DD | E | |
094 | 5E | ^ | 222 | DE | YU | |
095 | 5F | _ | 223 | DF | I | |
096 | 60 | ` | 224 | E0 | A | |
097 | 61 | a | 225 | E1 | b | |
098 | 62 | b | 226 | E2 | V | |
099 | 63 | c | 227 | E3 | G | |
100 | 64 | d | 228 | E4 | d | |
101 | 65 | e | 229 | E5 | e | |
102 | 66 | f | 230 | E6 | and | |
103 | 67 | g | 231 | E7 | h | |
104 | 68 | h | 232 | E8 | And | |
105 | 69 | i | 233 | E9 | th | |
106 | 6A | j | 234 | E.A. | To | |
107 | 6B | k | 235 | E.B. | l | |
108 | 6C | l | 236 | E.C. | m | |
109 | 6D | m | 237 | ED | n | |
110 | 6E | n | 238 | E.E. | O | |
111 | 6F | o | 239 | E.F. | P | |
112 | 70 | p | 240 | F0 | R | |
113 | 71 | q | 241 | F1 | With | |
114 | 72 | r | 242 | F2 | T | |
115 | 73 | s | 243 | F3 | at | |
116 | 74 | t | 244 | F4 | f | |
117 | 75 | u | 245 | F5 | X | |
118 | 76 | v | 246 | F6 | ts | |
119 | 77 | w | 247 | F7 | h | |
120 | 78 | x | 248 | F8 | w | |
121 | 79 | y | 249 | F9 | sch | |
122 | 7A | z | 250 | F.A. | ъ | |
123 | 7B | { | 251 | FB | s | |
124 | 7C | | | 252 | F.C. | b | |
125 | 7D | } | 253 | FD | uh | |
126 | 7E | ~ | 254 | F.E. | Yu | |
127 | 7F | DEL | 255 | FF | I |
Description of special (control) characters
Originally, ASCII table control characters (range 00-31, plus 127) were designed to control hardware devices such as teletypewriters, punched tape data entry, etc.
Control characters (except horizontal tab, line feed, and carriage return) are not used in HTML documents.
Code | Description |
---|---|
NUL, 00 | Null, empty |
SOH, 01 | Start Of Heading |
STX, 02 | Start of TeXt, beginning of the text |
ETX, 03 | End of TeXt, end of text |
EOT, 04 | End of Transmission, end of transmission |
ENQ, 05 | Enquire. Please confirm |
ACK, 06 | Acknowledgment. I confirm |
BEL, 07 | Bell, call |
BS, 08 | Backspace, go back one character |
TAB, 09 | Tab, horizontal tab |
LF, 0A | Line Feed, line feed Nowadays in most programming languages it is denoted as \n |
VT, 0B | Vertical Tab, vertical tabulation |
FF, 0C | Form Feed, page feed, new page |
CR, 0D | Carriage Return Nowadays in most programming languages it is denoted as \r |
SO,0E | Shift Out, change the color of the ink ribbon in the printing device |
SI, 0F | Shift In, return the color of the ink ribbon in the printing device back |
DLE, 10 | Data Link Escape, switching the channel to data transmission |
DC1, 11 DC2, 12 DC3, 13 DC4, 14 | Device Control, device control symbols |
NAK, 15 | Negative Acknowledgment, I do not confirm |
SYN, 16 | Synchronization. Synchronization symbol |
ETB, 17 | End of Text Block, end of the text block |
CAN, 18 | Cancel, cancellation of a previously transmitted |
EM, 19 | End of Medium |
SUB, 1A | Substitute, substitute. Placed in place of a symbol whose meaning was lost or corrupted during transmission |
ESC, 1B | Escape Control Sequence |
FS, 1C | File Separator, file separator |
GS, 1D | Group Separator |
RS, 1E | Record Separator, record separator |
US, 1F | Unit Separator |
DEL, 7F | Delete, erase the last character. |
As is known, Windows 10 supports the installation of language packs, which can be used to change the localization of the operating system interface. The concept of a system language is very closely related to this. The system language is not the one that was added using a language pack, but the one that was selected when installing the system itself. If you are working on a corporate computer whose language has been changed by installing MUI
-packages, you might want to know what language was originally used on the system.
There are several ways to determine the system language.
Let's look at them in order.
Using the DISM console utility
Run as administrator command line and execute in it dism command/online/get-intl. In a few seconds you will get the result.
Pay attention to the parameter "Default UI Language" . Its meaning will be the system language.
Using the PowerShell console
Launch the console as administrator and run the command ::InstalleduICulture .
Parameter value DisplayName will point to the system language. Please also pay attention to the parameter LCID, it contains the language code written in Windows registry(see below).
Using the Registry Editor and Decoding Table
Team regedit open the registry editor and expand the branch
HKEY_LOCAL_MACHINE\SYSTEM\CurrentControlSet\Control\Nls\Language
In the right column of the editor window, find the string parameter (he's at the very bottom) and see its meaning.
It is represented by a four-digit code, to decipher which you should use the decoding table on the website MSDN. In our example the code 0419 corresponds to the Russian language.