Encoding Why are accented characters rendering inconsistently when accessing the same code on the same server at a different URL?

There is a page on our server that's reachable via two different URLs. http://www.spotlight.com/6213-5613-0721 http://www.spotlight.com/interactive/cv/1/M103546.html There's classic ASP behind the scenes, and both of those URLs actually do a Server.Transfer to the same underlying ASP page. The accents in the name at the top of the page are rendering correctly on one URL and incorrectly on the other - but as far as I can tell, the two requests are returning identical responses (same markup, s

translate data file with odd Hebrew encoding

I have a binary data file, in a format used by a relatively ancient program, which I am trying to convert into something sane. With the help of a Hex editor I have basically worked out the file format except that it contains Hebrew characters with an odd encoding. All characters are 8 bits. The "standard" 27 consonants (including "final" consonants) go from hex 80 to 9A. Then there are vowels that seem to start around hex 9B or so (I'm guessing right after the standard consonants end). Then

vlc hardware encoding for real-time screen multicast

We are preparing a solution to multicast the teacher's screen to 40 students' pcs. Teacher and students pcs can be both Ubuntu and Windows. Some solutions were tested : iTalc ... not stable yet. multicast a "vnc -viewonly" ... no solution found capture the screen with VLC and multicast it. That latest seems to work ... except that with a resolution like 1920x1200 it is just too much CPU intensive. One idea would be to capture only the 4th of the screen. CPU is not saturated anymore but eve

Encoding Any way to remove the BOM part when a NSData has many BOM inside it?(Objective-c)

Normally if the file contains a BOM at the beginning of a file or a data, it is useful to identify what encoding the data is. But if you use NSData's dataUsingEncoding method, it would add a BOM if the encoding requires one.So if you use this method several time and append these data together as a output data and write it to file, then the file will contains several BOM.And the BOMs except the first one will be in the data part. For example: NSMutableData *data = [NSMutableData data]; [data a

Encoding Strangely encoded url

Background: I have been troubled by the following problem for the last week. Normally webpages encrypt urls using base64, so that http://stackoverflow.com becomes aHR0cDovL3N0YWNrb3ZlcmZsb3cuY29t and it's used like this: http://www.mysite.com/redir.php?url=aHR0cDovL3N0YWNrb3ZlcmZsb3cuY29t Problem: There are pages that use a different encryption of urls I can't crack. They all start with o5o4m4p4b434s2q43626z3 and I assume that's for http://www. comparing different links of this kind. The url

wiki dump encoding

I'm using WikiPrep to process the latest wiki dump enwiki-20121101-pages-articles.xml.bz2. Instead of "use Parse::MediaWikiDump;" I replaced that by "use MediaWiki::DumpFile::Compat;" and did the proper changes in the code. Then, I ran perl wikiprep.pl -f enwiki-20121101-pages-articles.xml.bz2 I got an error enwiki-20121101-pages-articles.xml.bz2:1: parser error : Document is empty BZh91AY&SY±H¦ÂOÿ~Ð`ÿÿÿ¿ÿÿÿ¿ÿÿÿÿÿÿÿÿÿÿ½ÿýþdß8õEnÞ¶zëJ¨Eà®mEÓP|f÷Ô ^ I guess there are some non-utf8 char

How can get Encoding.GetEncoding("iso-8859-1") Hexadecimal code?

I need to send a SMS via HTTP Post with some special characters like "caiò" So, with my code , I try to convert my string in ISO-8859-1. basically, if my string is converted in ISO-8859-1, and the method Uri.EscapeDataString() is invoked on it, in place of "ò" I should have "F2" Hexadecimal code. But I get %C3%A0 , that is the Hexadecimal code for UTF8 encoding. Encoding iso = Encoding.etEncoding("ISO-8859-1"); string StringBody = iso.GetString(iso.GetBytes(Body));

Encoding of a multipart/form-data in Struts 1

I have a very strange situation with a Struts web application: I have a page with a multipart/form-data enctype form. The page is in ISO-8859-1 and I have 2 different situations depending on the tomcat environment: The development environment is a Windows machine with the windows default encoding. The production environment is a Linux machine with UTF-8 encoding. :$ echo $LANG :$ es_ES.UTF-8 In windows environment, when the form is submited all parameters are correctly encoded. In linux

Use of regex on iso-2022-jp encoding with Python

I have some ISO-2022-JP encoded text. Ex. : まだ 正式 に 決まっ た わけ で は ない の で 。 According to the re library documentation, it can accept both ascii and unicode, so I tried to convert my text to unicode and to cut at the word level: text.decode('iso-2022-jp') print(text) print(re.findall(r"[\w']+", text)) However, here is the kind of output I get: まだ 正式 に 決まっ た わけ で は ない の で 。 ['B', 'B', 'B', 'B', 'B', '5', '0', 'B', 'B', 'K', 'B', 'B7h', 'C', 'B', 'B', 'B', 'B', 'o', '1', 'B', 'B', 'G', 'B

Why doesn't Heroku specify a default character encoding on their virtual machines?

It seems that Heroku doesn't explicitly specify a character encoding on their virtual machines, so the default becomes 8-bit ASCII. Other modern versions of Debian and Ubuntu specify UTF-8 as a default, and this avoids problems when languages fall back to operating-system character encoding for reading input. Why wouldn't Heroku use this default as well? Based on other SO posts, I've found that at least Ruby, SBCL and Haskell application developers have dealt with this problem, and they suggest

Encoding XMLHTTP - Read iso-8859-2 content and write UTF-8

I Need read a content from a page that is iso-8859-2 and write in UTF-8 in my code. Code Example: <%@ language="VBSCRIPT" codepage="65001" %> <% set xmlhttp=Server.CreateObject("Msxml2.XMLHttp.6.0") Set re=New RegExp re.IgnoreCase=True re.Global=True xmlhttp.open "get", link, false xmlhttp.setRequestHeader "Content-type", "application/x-www-form-urlencoded; charset=ISO-8859-2" xmlhttp.send() html=xmlhttp.responsetext re.Pattern="<h1>.*?<

Google Sites api non-latin language encoding issue

I'm programming with Google sites api's. My requests body consist parameters such as site title and category in non Latin language (Hebrew and Arab). After the requests are done I see the site title and category name in ??? characters, which is obviously an encoding problem. Tried to change and request ContentType header to application/atom+xml;charset=UTF-8 from application/atom+xml but it's not helping. Thanks in advance

Encoding MediaCodec Decoder issues with dequeueOutputBuffer call (H.264)

The Question: Why do I su...I mean, what is going on with my decoder? It's always returning timed out (-1) when I try to get it's juicy output data! Let me start by saying I've read through the relative threads, but still can't solve the problem. I really needs some guidance and assistance here. Okay, so what I am doing is passing AVC encoded data to my decoder. The encoder works, pulling data from camera preview, where I do the conversion from NV21 to NV12 before passing to the encoder. The

Encoding wkhtmltopdf pass cyrillic value to --header parameter

I am using wkhtmltopdf.exe command line tool to render HTML into PDF. I try to generate PDF with custom header using cyrillic symbols. So I have cmd command like this: wkhtmltopdf.exe --header-center "Заказ в Австралию — Test" http://localhost/MyPage c:/1.pdf The issue is that in generated PDF the header looks like this: "?????? ? ??????? - Test" I tried to add parameter --encoding utf-8 but this not worked for me Note that my pc is running Windows with the only installed English (US) syst

Encoding Where is my interpretation of ASN1 der wrong?

Here is what my structure looks like: SET OF SEQUENCE: INTEGER: XX INTEGER: YY My encoding looks like this: 11 08 10 06 02 01 XX 02 01 YY 11 08 -- SET OF 10 06 -- SEQUENCE However, when I decode with openssl, I don't see the expected output. It looks like 0:d=0 hl=2 l= 8 prim: SET 0000 - 10 06 02 01 XX 02 01 YY- This is not what I expected to see. (Look at the structure I wanted it to look like) I am not sure what I am missing. Any h

Encoding Correct display Cyrillic symbols in SublimeREPL PowerShell

I can not to make correct display Cyrillic symbols in SublimeREPL PowerShell. Expected behavior In Windows PowerShell I have not this problem. Actual behavior Cyrillic symbols always not displayed for me. I see mojibake instead it. Steps to reproduce The problem is reproduced for me in a version of Sublime Text without plugins and user settings. Preferences → Key Bindings → I change Default (Windows).sublime-keymap - User: [ { "keys": ["super+alt+p"], "comman

Encoding How to decode string of curve data from XML

i have a Thormed spirometry measure file with manoeuvers and each other has curve data, which normally is represented as a chain of numerical values (separated by space, comma, semicolon etc). But this time i wasn't able to decode this string, it is HEX code but i stuck after decoding it from HEX. What should be the next step? String is inside "blob" tag <!--Thormed Software generated XML file!--> <Table name="FVC_CURVE_DATA"> <Field name="ID" type="integer">1</Field>

react native encoding conversion

I'm trying to convert a fetch response from ISO-8859-1 to UTF-8 in react native. I found this answer : Encoding conversion of a fetch response but in react native i'm having some troubles... I didn't manage to use iconv, i tryed : npm install --save iconv-lite // success then i do var iconv = require('iconv-lite') // Error 500 from react native I didn't manage to use Buffer, it is unknown by react native and require('buffer').Buffer // Error 500 from react native So here are my questio

Encoding Heirloom mailx doesn't display UTF-8 characters

I'm using Heirloom mailx 12.5-2. When I'm receiving a mail with characters like "ä", "ö" or "ü" mailx can't display these and just shows "�". The mails charset is "iso-8859-1". I already tried to set the options "ttycharset", "sendcharsets" to "iso-8859-1" or "UTF-8", but it doesn't help. Also the options "print-all-chars", "encoding" or "reply-in-same-charset" don't solve the problem. My locale and terminal are on UTF-8. Does anyone has an other idea? Thanx

Can't encode audio from PCM 16 bits encoding to G711-MLAW using MediaCodec?

I want to encode audio to G711-MLAW using MediaCodec. I listed the supported codecs, the result was as follows: OMX.google.mp3.decoder OMX.google.amrnb.decoder OMX.google.amrwb.decoder OMX.google.aac.decoder OMX.google.g711.alaw.decoder OMX.google.vorbis.decoder OMX.google.opus.decoder OMX.google.raw.decoder OMX.google.aac.encoder OMX.google.amrnb.encoder OMX.google.amrwb.encoder OMX.google.flac.encoder OMX.google.gsm.decoder OMX.google.mpeg4.decoder OMX.google.h263.decoder OMX.google.h264.deco

Encoding Avro serialize and desiaralize List<UUID>

I cannot understand how to serialize List to binary format and deserialize back to List. I have tried to use CustomEncoding for this purpose: public class ListUUIDAsListStringEncoding extends CustomEncoding<List<UUID>> { { schema = Schema.createArray(Schema.createUnion(Schema.create(Schema.Type.STRING))); schema.addProp("CustomEncoding", "com.my.lib.common.schemaregistry.encoding.ListUUIDAsListStringEncoding"); } @Override protected void write(Object

Encoding system for an image?

Like we have ASCII encoding system for characters. Then what is the encoding system used for other things such as an image, audio or video etc?

DataWeave Multipart/form-data input encoding not recognized in Mule4

Is there an option to set encoding by part in multipart/form-data received to listener? I have this kind of multipart to be received: ----------------------------180928595588258919887097 Content-Disposition: form-data; name="qualifyResult"; filename="json1.json" Content-Type: application/json { "json1": "1" } ----------------------------180928595588258919887097 Content-Disposition: form-data; name="raceOneResult"; filename="json2.json&

Working out file encoding: I know the string, know the character, what is the encoding?

I'm adding data from a csv file into a database. If I open the CSV file, some of the entries contain bullet points - I can see them. file says it is encoded as ISO-8859. $ file data_clean.csv data_clean.csv: ISO-8859 English text, with very long lines, with CRLF, LF line terminators I read it in as follows and convert it from ISO-8859-1 to UTF-8, which my database requires. row = [unicode(x.decode("ISO-8859-1").strip()) for x in row] print row[4] description = row[4].encode

UTF8 encoding problem, same results work fine in wordpress

I have a wordpress installation that clients can edit, all characters display ok. On the main homepage I query the same database for the same title and post content, but it doesn't display correctly - just a question mark I have tried sending the utf8 headers manually, through htaccess and through meta tags. I have used SET name UTF8 (which turns the characters into the diamond symbol with a questionmark inside). I genuinely cant figure out what it could be now and I really need these characte

Encoding Find out how a file was encoded

I have an accounting program from the 1990s called ocl, the program no longer runs correctly and I want to extract the data, when opening the files in sublime text it gives me a long file like the following 393d 2020 2020 2020 2020 2020 5072 6576 696f 7573 2054 7261 6e73 2020 2020 2020 2020 2020 2020 2020 2020 2020 2020 2020 2020 2035 3430 322e 3235 0000 0000 0000 0000 0000 0000 0000 0000 0000 0000 0000 My question is does anyone have any idea what this is or how to go about decoding it?

Encoding is there a way to highlight all the special accent characters in sublime text or any other text editor?

I a using the the HTML encode special characters in Sublime text to convert all the special character into their HTML code. I have a lot of accented characters in different parts of the file. So, it would be great if I could select all the special character and then use the plugin to convert all at once! Is there a regex that helps select all special characters only?

Encoding How to decode cp1252 which is in decimal &#147 instead of \x93?

I am getting the source code of a webpage and the encoding is cp1252. Chrome displays the page correctly. Here is my code: import sys from urllib.request import urlopen from bs4 import BeautifulSoup, UnicodeDammit import re import codecs url = "http://www.sec.gov/Archives/edgar/data/1400810/000119312513211026/d515005d10q.htm" page = urlopen(url).read() print(page) # A little preview : # b'...Regulation S-T (&#167;232.405 of this chapter) during the preceding 12 months (or for such short

How to Determine encoding/compression of the string which appears like characters are dancing in gangam style

While analyzing the HTTP Requests OF a website. I found that in one of the POST request it sends three postdata to the server the first one was SAML data first base64 encoded then urlencoded. But I am not able to figure out the value of other two postvars. One thing I am sure about is that it is not using any encryption methods like md5 or sha1 etc. COZ the response text contains my user name value which according to my research is neither stored in session variable or cookies means this encod

video.js how to manage subtitles special characters encoding

Video.js works perfectly with subtitles in English but as for the french or the spanish, there're always some encoding problems with special characters like "é, ñ, ê" etc… which result in the usual encoding difficulties on the web when there's a wrong charset. Yet my pages are encoded in UTF-8… So is there a workaround ?

Encoding httpgetrequest uri encoded to iso-8859-2

Hello I spent ages by searching community forums to find working solution. It looks simple job, I need send HTTP request passing arguments by GET method, but server side service required URI encoded in ISO-8895-2. I use .NET System.net class HttpWebRequest. My code: String myURI = (this._services + "?" + this._request); HttpWebRequest request = HttpWebRequest.Create(myURI) as HttpWebRequest; request.ContentType = "application/x-www-form-urlencoded, charset=iso-8859-2"; request.Method = "GET";

Encoding difference between gzip and gzip, deflate

I want to know what is the difference when I add Accept-Encoding: gzip, deflate in request header and when I only add Accept-Encoding: gzip. Because in response header I could see Content-Encoding : gzip only every time and the compression percentage is also same. And which one is the best practice to use.

Encoding How do I decode xterm responses?

I'm building an application that uses pty.js to open up a pseudo terminal on my computer. I'm getting responses that look like: ]0;ec2-user@ip-172-31-62-237:~[?1034h[ec2-user@ip-172-31-62-237 ~]$ ls ]0;ec2-user@ip-172-31-62-237:~[ec2-user@ip-172-31-62-237 ~]$ pwd /home/ec2-user I'm assuming pty.js is sending back a specific encoding, but I'm not sure what the encoding is and how to decode it. Any help would be appreciated, thanks.

Encoding Log::Log4perl: one appender to the file in UTF-8 and another to the screen in CP866 on win32

I have Log::Log4perl config with two appenders to file and screen, like this: log4perl.logger.bc_log = INFO, bc_log, console log4perl.appender.bc_log = Log::Log4perl::Appender::File log4perl.appender.bc_log.filename = bc.log log4perl.appender.bc_log.layout = Log::Log4perl::Layout::PatternLayout log4perl.appender.bc_log.layout.ConversionPattern = %d{ISO8601}%d{Z} %H %m%n log4perl.appender.console = Log::Log4perl::Appender::Screen log4perl.append

Encoding How can I add JPEG image with non-UTF characters into a docx using python-docx?

I am trying to insert a series of JPEGs into a Word document using python-docx but it seems that some of them may have non-UTF-8 metadata included, which is causing docx to issue a Unicode decoding error message. How can I get around this? Here is the code: #!/usr/bin/env python3 # -*- coding: utf-8 -*- from docx import Document from docx.shared import Inches from docx.enum.table import * from docx.enum.text import WD_ALIGN_PARAGRAPH from PIL import Image from PIL.ExifTags import TAGS documen

Is this a bug in the passlib base64 encoding?

I am trying to decode an re-encode a bytesytring using passlibs base64 encoding: from passlib.utils import binary engine = binary.Base64Engine(binary.HASH64_CHARS) s2 = engine.encode_bytes(engine.decode_bytes(b"1111111111111111111111w")) print(s2) This prints b'1111111111111111111111A' which is of course not what I expected. The last character is different. Where is my mistake? Is this a bug?

JSF 2.0 Encoding

What settings and considerations are necessary for making a JSF 2.0 application (on Tomcat 6) use UTF-8 encoding?

Encoding Batch script that separates that creates folder by month

I am trying to edit this code to when we run the batch file it creates a folder based off the month of our file format. We ran this script a whole lot but forgot to make a folder called April now we have May mixed in with April. The files are formatted like Username-R705-2011.05.04-1601-Disconnected.7z I was wondering if there is anyway it can go off the .04 and make a folder for that month so it will automatically put it in the folder it is needing to go into, so it will be easy to search

Encoding How is it possible to encode black/white picture into ".wav"-file?

How is it possible to encode black/white picture into ".wav"-file? I know that it is possible for sure with help of "stenography". But I don't know it's algorithms. What algorithms exist? And what books/sources are the best for understanding of their principles? Edited: Actually I have stereo wav-file. My task is to decode pictures from it. The task says, that frequencies of the left channel show the X-coordinate, frequencies of the right channel show the Y-coordinate of Cartesian coordinate

Wrong encoding in R presentation

I've struggling a lot with this problem since last week: I am using R presentation (.Rpres file) for the first time and it started alright, meaning that I could build a slide and visualize the result in the Presentation tab in RStudio. However, for reasons I don't understand, after a few hours of working on my presentations the Presentation tab began to show weird symbols for all the french characters in my presentation. The only way so far I could get the presentation back to showing the right

Encoding Accessing chars in utf-8 strings

First, I want to make sure I understand the concept of utf-8 correctly. When a string is stored in utf-8 each character is actually 1-4 bytes depending on the character it is representing. If I had an ascii string like this: string meh = "blah"; then all it has to do to obtain the fourth char is to obtain a pointer to the first char and add an offset of size char to locate the character, like this: meh[3] == *(meh + 3); // not real code, just pseudo c for what the compiler does however

How to correctly detect file encodings with LispWorks?

I am trying to detect file encoding using LispWorks. LispWorks should be capable of such functionality, see External Formats and File Streams. [Note: details based on @rainer-joswig and @svante comments] system:*file-encoding-detection-algorithm* is set to its default, (setf system:*file-encoding-detection-algorithm* '(find-filename-pattern-encoding-match find-encoding-option detect-utf32-bom detect-unicode-bom detect-utf8-bom specific-valid-file-enc

Encoding Diamonds with question marks

I'm getting these little diamonds with question marks in them in my HTML attributes when I present data from my database. I'm using EPiServer and a few custom properties. This is the information I've gathered, I save my data as a XML document, since I use custom EPiServer properties which need more than one defined value. This is saved as UTF8. It's only attributes in element tags which have this problem, such as align=left becomes align=�left�. There is no " character there, but I get the di

Encoding Number of bits calculation for binary Genetic Algorithm

I am busy completing an assignment on genetic algorithms and came across the following problem: In a binary GA, how many bits are needed to represent numbers between 1 and 10 with a resolution of 0.0001? Formula for calculations The answer I get is 16.45 and the answer from the solutions is 17 (with no calculations shown). Is it necessary to round up the bit number? You cannot use a portion of a bit.

  1    2   3   4   5   6  ... 下一页 最后一页 共 7 页