[Top][All Lists]

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

bug#14403: 24.3.50; [regression] Typing non-ascii characters on a non-GU

From: Eli Zaretskii
Subject: bug#14403: 24.3.50; [regression] Typing non-ascii characters on a non-GUI MS-Windows session
Date: Wed, 15 May 2013 11:13:03 +0300

> From: Stefan Monnier <address@hidden>
> Cc: address@hidden,  address@hidden
> Date: Tue, 14 May 2013 18:28:35 -0400
> > This is an exact duplicate of 14368, discussed just yesterday.
> Hmm, this looks different, since it doesn't seem to use quail

Different, yes.  But caused by the same change in keyboard.c.

> Is it the case that the non-GUI code in Windows receives decoded
> chars in `read_char', contrary to the posix code which receives
> encoded chars there?

Both GUI and non-GUI keyboard input on Windows produce Unicode
codepoints of the characters the user types.  The produced input
events have the 'kind' of either ASCII_KEYSTROKE_EVENT or
MULTIBYTE_CHAR_KEYSTROKE_EVENT.  They are returned via this call

  read_char -> kbd_buffer_get_event -> make_lispy_event

I guess the solution is to tell read_decoded_char more about the event
that produced the character?

Or maybe we should not set the keyboard encoding to the console
codepage on Windows (although I have no idea what kind of breakage
this could cause)?  What setting of keyboard-coding-system tells the
condition below that no decoding is needed?

Of course, we could always augment this condition:

      if (!((FRAME_TERMCAP_P (frame) || FRAME_MSDOS_P (frame))
            && (TERMINAL_KEYBOARD_CODING (terminal)->common_flags
        return nextevt;         /* No decoding needed.  */

to do something special for MS-Windows, but that sounds kludgey.

reply via email to

[Prev in Thread] Current Thread [Next in Thread]