Thanks, Hans. Combined Characters such as should survive through multiple transmissions across different platforms, as long as the applications utilized are fully compatible with Unicode. 

Technically speaking, I believe the problems you have had with your wonderful volume are not really single-glyph characters vs. multi-glyph characters, but Unicode vs. non-Unicode. As you may know, the standard for extended Latin characters predates Unicode. For example, ISO 8859-1 includes many commonly used symbols for Indological transliteration such as  Ā, ā, Ē, ē, Ī, ī, Ō, ō, Ū, ū was first published in March 1985. Many non-Unicode applications support these extended Latin characters, and therefore you find them survive through multiple transmissions (Unicode is compatible with these Latin extensions as well). 

However, many of the new characters added for the Unicode standard, will not survive through the use of non-Unicode applications, whether they are single glyph or multi-glyph. For example, Ḃ (Latin Capital Letter B with dot above) is a single glyph character, but I am pretty sure it will not survive transmission through non-Unicode applications (let's see how many find Ḃ in their emails correctly).

Alas, until the computing world rids itself of all pestiferous non-Unicode applications, perhaps, we will have to deal with the headaches of diacritics! 

Suresh.


On Wed, Aug 10, 2016 at 11:46 AM, Hock, Hans Henrich <hhhock@illinois.edu> wrote:
Thanks, Suresh.

Would that things were that simple. The proper transcription, , with candra, but no bindu, can be produced through what are called rendering machines; but when sending files back and forth between different platforms for the volume The languages and linguistic of South Asia, we found that this character, as well as characters such as ā́, did not always survive intact but instead showed up as blanks. Unlike single-glyph characters, these are not stable. We therefore always made sure to add a pdf version, so that messed-up characters could be restored. 

All the best,

Hans


On 10 Aug 2016, at 09:13, Suresh Kolichala <suresh.kolichala@gmail.com> wrote:

Dear Rolf and others,

Sorry for a late response, but I just saw this thread. Since I have some experience with Unicode as a member of the Unicode Consortium, Here are my quick remarks:
  1. Please don't use ṁ for transcribing the half nasal. ṁ is used for anusvāra in ISO-15919. Sinhala should be using ISO-15919, not IAST. ṃ is not used in ISO-15919.
  2. I don't know what work/application you need this symbol for, but the use of combining diacritical marks are very common in Unicode, and widely used for Indological purposes, as Hans pointed out. 
  3. m̐ should serve your purposes for representing the labial prenasalized consonant ඹ (U+0DB9 SINHALA LETTER AMBA BAYANNA)
  4. In the world of Unicode, nobody should be bothering about individual glyphs. If you work requires it, then, it is perhaps using an outdated technology. Abandon it :).
  5. If you use Unicode, it should not matter whether you are using a Mac, PC or any other publishing software. The text should be as transportable as the ordinary English Text (which follows ASCII standard). Unicode is an international standard, and almost all the applications and operating systems developed in the last decade should fully support it.
Regards,
Suresh.