On 11/5/06, Hans Hagen wrote:
Philipp Reichmuth wrote:
I've been writing a script that sifts through the unic-xxx.tex files to get a readable mapping what Unicode characters are supported using \Amacron-style names.
mtxtools can create such lists using the unicode consotium glyph table, mojca's mapping list and enco/regi files
we use mtxtools to create the tables needed for xetex (used for case mapping) and luatex (more extensive manipulations)
I have mtxtools.bat, but no mtxtools.rb here.
Are the unic-xxx files automatically generated or maintained by hand?
maintained by hand, again, just send me the fixed file, but we need to make sure that the fix is ok (i.e. works as expected)
Although there should be no reason for not generating them automatically. I did that for regime files (I only wrote a script, executed it and Hans included the files, so it's only semi-automatic; it would be polite from me if I managed to incorporate that into existing [whateverthename]tools.rb).
Incidentally, it would be trivial now to put the list of ConTeXt glyphs on the Wiki, if anyone's interested.
there is a file contextnames.txt in the distributions (maintained by mojca), while the not yet distributed char-def.lua has the info for luatex
If you find errors there, please let me know. (Missing letter in Cyrillic was due to missing position in Unicode).
I wanted to use this to work towards better support for the whole range of ConTeXt glyphs with OpenType fonts under XeTeX, by reading what ConTeXt glyphs are available in a font and building a list of "\catcode`ā=\active \def ā {\amacron}"-style list for the rest. (Unfortunately this kind of list would be font-specific, but the generic alternative would be a huge list of active characters with an \ifnum\XeTeXcharglyph"....>0 macro behind it, and that would probable be quite slow.) I wonder if there is a more intelligent way to achieve this goal; since part of the logic for mapping code points into glyph macros exists already, it would be easier if there was a way to reuse that.
best take a look at mtxtools; if needed we can generate the definitions ; concerning speed, it will not be that slow, because tex is quite fast on such tests (unless XeTeXcharglyph is slow due to lib access); the biggest thing is to make sure that things don't expand in unwanted ways.
(i must find time to update my xetex bin ; i must admit that i never tried to use open type fonts in xetex (the mac is broken)
But OpenType fonts also work on Linux & Windows.
The best way out would be if I could enable ConTeXt's UTF-8 regime while running XeTeX in \XeTeXinputencoding=bytes mode, but I haven't gotten that to work yet.
maybe mojca has
You could theoretically comment out \beginXETEX \expandafter \endinput \endXETEX in regi-utf.tex, but that's not the best idea. Mojca