diff options
author | Joey Hess <joey@kodama.kitenet.net> | 2008-11-12 17:19:41 -0500 |
---|---|---|
committer | Joey Hess <joey@kodama.kitenet.net> | 2008-11-12 17:30:54 -0500 |
commit | 716560b7f15b6e15b246c39c11eb8181d91c8662 (patch) | |
tree | 5b7a30dd2f18e02b02f064d0a1ab59fe891b6a71 /debian/NEWS | |
parent | 2c858c9c95e287ebe3740a94f983f6ae9d6fb080 (diff) |
check for invalid utf-8, and toss it back to avoid crashes
Since ikiwiki uses open :utf8, perl assumes that files contain valid utf-8.
If it turns out to be malformed it may later crash while processing strings
read from them, with 'Malformed UTF-8 character (fatal)'.
As at least a quick fix, use utf8::valid as soon as data is read, and if
it's not valid, call encode_utf8 on the string, thus clearing the utf-8
flag. This may cause follow-on encoding problems, but will avoid this
crash, and the input file was broken anyway, so GIGO is a reasonable
response. (I looked at calling decode_utf8 after, but it seemed to cause
more trouble than it was worth. BTW, use open ':encoding(utf8)' avaoids
this problem, but the corrupted data later causes Storable to crash when
writing the index.)
This is a quick fix, clearly imperfect:
- It might be better to explicitly call decode_utf8 when reading files,
rather than using the IO layer.
- Data read other than by readfile() can still sneak in bad utf-8. While
ikiwiki does very little file input not using it, stdin for the CGI
would be one way.
Diffstat (limited to 'debian/NEWS')
0 files changed, 0 insertions, 0 deletions