I guess this isn't really the place for this query, but perhaps someone has had a similar problem.<br><br>I'm using XeTeX and TeXWorks for academic work, like many of us. I cut-n-past bibliographical information from sites like <a href="http://copac.ac.uk" target="_blank">copac.ac.uk</a> and <a href="http://worldcat.org" target="_blank">worldcat.org</a>, into JabRef for use in my documents. What I'm finding, though, is that several of these big online bibliographical databases have their records in un-normalized Unicode. And it doesn't print nicely with XeTeX. <br>
<br>Rather than struggel with XeTeX's accent placement, which seems to be an unattractively per-font problem in any case, it makes better sense to me to normalize the Unicode to NFC form. I.e., <a href="http://en.wiktionary.org/wiki/Appendix:Unicode_normalization" target="_blank">http://en.wiktionary.org/wiki/Appendix:Unicode_normalization</a><br>
What I'd like, ideally, is a little filter to run on my bib files periodically to clean up any char+non-spacing-accent glyphs.<br><br>I've been looking around for an appropriate tool for this job. All I could find is <a href="http://charlint.pl" target="_blank">charlint.pl</a>. But I can't for the life of me get it to work. It halts, throwing up errors along the lines of <br>
\begin{verbatim}<br>Checking duplicates, takes some time.<br>Finished processing character data file(s).<br><br>Line 1: Non-Existing codepoints.<br>Giving up!<br>\end{verbatim}<br>
<br>Has anyone any better suggestions than charlint, or experience getting charlint working?<br><br>Best,<br>Dominik<br>