Are there any plans to support reading of unicode text files? I realize this would require a totally new internal structure of character strings, but what about a "brute force" reading where all unicode characters are converted to ASCII characters? I do not know how much extra it would take; maybe the underlying libraries already support this? Why do one want this dumb translation? It would useful if the data file I am reading is 99.99% numerical and I could live with the fact that some character fields *might* get funny looking values. This is just a query of the status of unicode and not a request of an implementation. There are plenty of software that can read Unicode text files and save them as plain ASCII so I am doing fine myself. Have a great day Henrik Bengtsson Dept. of Mathematical Statistics @ Centre for Mathematical Sciences Lund Institute of Technology/Lund University, Sweden (+2h UTC) +46 46 2229611 (off), +46 708 909208 (cell), +46 46 2224623 (fax) h b @ m a t h s . l t h . s e, http://www.maths.lth.se/bioinformatics/ -.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.- r-help mailing list -- Read http://www.ci.tuwien.ac.at/~hornik/R/R-FAQ.html Send "info", "help", or "[un]subscribe" (in the "body", not the subject !) To: r-help-request at stat.math.ethz.ch _._._._._._._._._._._._._._._._._._._._._._._._._._._._._._._._._._._._._._._._