NetBSD/gnu/dist/gettext/doc/gettext.texi
2001-02-17 19:06:18 +00:00

4803 lines
209 KiB
Plaintext

\input texinfo @c -*-texinfo-*-
@c %**start of header
@setfilename gettext.info
@settitle GNU @code{gettext} utilities
@finalout
@c %**end of header
@include version.texi
@dircategory GNU Gettext Utilities
@direntry
* Gettext: (gettext). GNU gettext utilities.
* gettextize: (gettext)gettextize Invocation. Prepare a package for gettext.
* msgfmt: (gettext)msgfmt Invocation. Make MO files out of PO files.
* msgmerge: (gettext)msgmerge Invocation. Update two PO files into one.
* xgettext: (gettext)xgettext Invocation. Extract strings into a PO file.
@end direntry
@ifinfo
This file provides documentation for GNU @code{gettext} utilities.
It also serves as a reference for the free Translation Project.
Copyright (C) 1995, 1996, 1997 Free Software Foundation, Inc.
Permission is granted to make and distribute verbatim copies of
this manual provided the copyright notice and this permission notice
are preserved on all copies.
@ignore
Permission is granted to process this file through TeX and print the
results, provided the printed document carries copying permission
notice identical to this one except for the removal of this paragraph
(this paragraph not being relevant to the printed manual).
@end ignore
Permission is granted to copy and distribute modified versions of this
manual under the conditions for verbatim copying, provided that the entire
resulting derived work is distributed under the terms of a permission
notice identical to this one.
Permission is granted to copy and distribute translations of this manual
into another language, under the above conditions for modified versions,
except that this permission notice may be stated in a translation approved
by the Foundation.
@end ifinfo
@titlepage
@title GNU gettext tools, version @value{VERSION}
@subtitle Native Language Support Library and Tools
@subtitle Edition @value{EDITION}, @value{UPDATED}
@author Ulrich Drepper
@author Jim Meyering
@author Fran@,{c}ois Pinard
@page
@vskip 0pt plus 1filll
Copyright @copyright{} 1995, 1996, 1997 Free Software Foundation, Inc.
Permission is granted to make and distribute verbatim copies of
this manual provided the copyright notice and this permission notice
are preserved on all copies.
Permission is granted to copy and distribute modified versions of this
manual under the conditions for verbatim copying, provided that the entire
resulting derived work is distributed under the terms of a permission
notice identical to this one.
Permission is granted to copy and distribute translations of this manual
into another language, under the above conditions for modified versions,
except that this permission notice may be stated in a translation approved
by the Foundation.
@end titlepage
@ifinfo
@node Top, Introduction, (dir), (dir)
@top GNU @code{gettext} utilities
@menu
* Introduction:: Introduction
* Basics:: PO Files and PO Mode Basics
* Sources:: Preparing Program Sources
* Initial:: Making the Initial PO File
* Updating:: Updating Existing PO Files
* Binaries:: Producing Binary MO Files
* Users:: The User's View
* Programmers:: The Programmer's View
* Translators:: The Translator's View
* Maintainers:: The Maintainer's View
* Conclusion:: Concluding Remarks
* Country Codes:: ISO 639 country codes
--- The Detailed Node Listing ---
Introduction
* Why:: The Purpose of GNU @code{gettext}
* Concepts:: I18n, L10n, and Such
* Aspects:: Aspects in Native Language Support
* Files:: Files Conveying Translations
* Overview:: Overview of GNU @code{gettext}
PO Files and PO Mode Basics
* Installation:: Completing GNU @code{gettext} Installation
* PO Files:: The Format of PO Files
* Main PO Commands:: Main Commands
* Entry Positioning:: Entry Positioning
* Normalizing:: Normalizing Strings in Entries
Preparing Program Sources
* Triggering:: Triggering @code{gettext} Operations
* Mark Keywords:: How Marks Appears in Sources
* Marking:: Marking Translatable Strings
* c-format:: Telling something about the following string
* Special cases:: Special Cases of Translatable Strings
Making the Initial PO File
* xgettext Invocation:: Invoking the @code{xgettext} Program
* C Sources Context:: C Sources Context
* Compendium:: Using Translation Compendiums
Updating Existing PO Files
* msgmerge Invocation:: Invoking the @code{msgmerge} Program
* Translated Entries::
* Fuzzy Entries:: Fuzzy translated Entries
* Untranslated Entries:: Untranslated Entries
* Obsolete Entries:: Obsolete Entries
* Modifying Translations:: Modifying Translations
* Modifying Comments:: Modifying Comments
* Auxiliary:: Consulting Auxiliary PO Files
Producing Binary MO Files
* msgfmt Invocation:: Invoking the @code{msgfmt} Program
* MO Files:: The Format of GNU MO Files
The User's View
* Matrix:: The Current @file{ABOUT-NLS} Matrix
* Installers:: Magic for Installers
* End Users:: Magic for End Users
The Programmer's View
* catgets:: About @code{catgets}
* gettext:: About @code{gettext}
* Comparison:: Comparing the two interfaces
* Using libintl.a:: Using libintl.a in own programs
* gettext grok:: Being a @code{gettext} grok
* Temp Programmers:: Temporary Notes for the Programmers Chapter
About @code{catgets}
* Interface to catgets:: The interface
* Problems with catgets:: Problems with the @code{catgets} interface?!
About @code{gettext}
* Interface to gettext:: The interface
* Ambiguities:: Solving ambiguities
* Locating Catalogs:: Locating message catalog files
* Optimized gettext:: Optimization of the *gettext functions
Temporary Notes for the Programmers Chapter
* Temp Implementations:: Temporary - Two Possible Implementations
* Temp catgets:: Temporary - About @code{catgets}
* Temp WSI:: Temporary - Why a single implementation
* Temp Notes:: Temporary - Notes
The Translator's View
* Trans Intro 0:: Introduction 0
* Trans Intro 1:: Introduction 1
* Discussions:: Discussions
* Organization:: Organization
* Information Flow:: Information Flow
Organization
* Central Coordination:: Central Coordination
* National Teams:: National Teams
* Mailing Lists:: Mailing Lists
National Teams
* Sub-Cultures:: Sub-Cultures
* Organizational Ideas:: Organizational Ideas
The Maintainer's View
* Flat and Non-Flat:: Flat or Non-Flat Directory Structures
* Prerequisites:: Prerequisite Works
* gettextize Invocation:: Invoking the @code{gettextize} Program
* Adjusting Files:: Files You Must Create or Alter
Files You Must Create or Alter
* po/POTFILES.in:: @file{POTFILES.in} in @file{po/}
* configure.in:: @file{configure.in} at top level
* aclocal:: @file{aclocal.m4} at top level
* acconfig:: @file{acconfig.h} at top level
* Makefile:: @file{Makefile.in} at top level
* src/Makefile:: @file{Makefile.in} in @file{src/}
Concluding Remarks
* History:: History of GNU @code{gettext}
* References:: Related Readings
@end menu
@end ifinfo
@node Introduction, Basics, Top, Top
@chapter Introduction
@quotation
This manual is still in @emph{DRAFT} state. Some sections are still
empty, or almost. We keep merging material from other sources
(essentially e-mail folders) while the proper integration of this
material is delayed.
@end quotation
In this manual, we use @emph{he} when speaking of the programmer or
maintainer, @emph{she} when speaking of the translator, and @emph{they}
when speaking of the installers or end users of the translated program.
This is only a convenience for clarifying the documentation. It is
@emph{absolutely} not meant to imply that some roles are more appropriate
to males or females. Besides, as you might guess, GNU @code{gettext}
is meant to be useful for people using computers, whatever their sex,
race, religion or nationality!
This chapter explains the goals sought in the creation
of GNU @code{gettext} and the free Translation Project.
Then, it explains a few broad concepts around
Native Language Support, and positions message translation with regard
to other aspects of national and cultural variance, as they apply to
to programs. It also surveys those files used to convey the
translations. It explains how the various tools interact in the
initial generation of these files, and later, how the maintenance
cycle should usually operate.
Please send suggestions and corrections to:
@example
@group
@r{Internet address:}
bug-gnu-utils@@prep.ai.mit.edu
@end group
@end example
@noindent
Please include the manual's edition number and update date in your messages.
@menu
* Why:: The Purpose of GNU @code{gettext}
* Concepts:: I18n, L10n, and Such
* Aspects:: Aspects in Native Language Support
* Files:: Files Conveying Translations
* Overview:: Overview of GNU @code{gettext}
@end menu
@node Why, Concepts, Introduction, Introduction
@section The Purpose of GNU @code{gettext}
Usually, programs are written and documented in English, and use
English at execution time to interact with users. This is true
not only of GNU software, but also of a great deal of commercial
and free software. Using a common language is quite handy for
communication between developers, maintainers and users from all
countries. On the other hand, most people are less comfortable with
English than with their own native language, and would prefer to
use their mother tongue for day to day's work, as far as possible.
Many would simply @emph{love} to see their computer screen showing
a lot less of English, and far more of their own language.
However, to many people, this dream might appear so far fetched that
they may believe it is not even worth spending time thinking about
it. They have no confidence at all that the dream might ever
become true. Yet some have not lost hope, and have organized themselves.
The Translation Project is a formalization of this hope into a
workable structure, which has a good chance to get all of us nearer
the achievement of a truly multi-lingual set of programs.
GNU @code{gettext} is an important step for the Translation Project,
as it is an asset on which we may build many other steps. This package
offers to programmers, translators and even users, a well integrated
set of tools and documentation. Specifically, the GNU @code{gettext}
utilities are a set of tools that provides a framework within which
other free packages may produce multi-lingual messages. These tools
include a set of conventions about how programs should be written to
support message catalogs, a directory and file naming organization for the
message catalogs themselves, a runtime library supporting the retrieval of
translated messages, and a few stand-alone programs to massage in various
ways the sets of translatable strings, or already translated strings.
A special mode for GNU Emacs also helps ease interested parties into
preparing these sets, or bringing them up to date.
GNU @code{gettext} is designed to minimize the impact of
internationalization on program sources, keeping this impact as small
and hardly noticeable as possible. Internationalization has better
chances of succeeding if it is very light weighted, or at least,
appear to be so, when looking at program sources.
The Translation Project also uses the GNU @code{gettext}
distribution as a vehicle for documenting its structure and methods.
This goes beyond the strict technicalities of documenting the GNU @code{gettext}
proper. By so doing, translators will find in a single place, as
far as possible, all they need to know for properly doing their
translating work. Also, this supplemental documentation might also
help programmers, and even curious users, in understanding how GNU
@code{gettext} is related to the remainder of the Translation
Project, and consequently, have a glimpse at the @emph{big picture}.
@node Concepts, Aspects, Why, Introduction
@section I18n, L10n, and Such
Two long words appear all the time when we discuss support of native
language in programs, and these words have a precise meaning, worth
being explained here, once and for all in this document. The words are
@emph{internationalization} and @emph{localization}. Many people,
tired of writing these long words over and over again, took the
habit of writing @dfn{i18n} and @dfn{l10n} instead, quoting the first
and last letter of each word, and replacing the run of intermediate
letters by a number merely telling how many such letters there are.
But in this manual, in the sake of clarity, we will patiently write
the names in full, each time@dots{}
By @dfn{internationalization}, one refers to the operation by which a
program, or a set of programs turned into a package, is made aware of and
able to support multiple languages. This is a generalization process,
by which the programs are untied from calling only English strings or
other English specific habits, and connected to generic ways of doing
the same, instead. Program developers may use various techniques to
internationalize their programs. Some of these have been standardized.
GNU @code{gettext} offers one of these standards. @xref{Programmers}.
By @dfn{localization}, one means the operation by which, in a set
of programs already internationalized, one gives the program all
needed information so that it can adapt itself to handle its input
and output in a fashion which is correct for some native language and
cultural habits. This is a particularisation process, by which generic
methods already implemented in an internationalized program are used
in specific ways. The programming environment puts several functions
to the programmers disposal which allow this runtime configuration.
The formal description of specific set of cultural habits for some
country, together with all associated translations targeted to the
same native language, is called the @dfn{locale} for this language
or country. Users achieve localization of programs by setting proper
values to special environment variables, prior to executing those
programs, identifying which locale should be used.
In fact, locale message support is only one component of the cultural
data that makes up a particular locale. There are a whole host of
routines and functions provided to aid programmers in developing
internationalized software and which allow them to access the data
stored in a particular locale. When someone presently refers to a
particular locale, they are obviously referring to the data stored
within that particular locale. Similarly, if a programmer is referring
to ``accessing the locale routines'', they are referring to the
complete suite of routines that access all of the locale's information.
One uses the expression @dfn{Native Language Support}, or merely NLS,
for speaking of the overall activity or feature encompassing both
internationalization and localization, allowing for multi-lingual
interactions in a program. In a nutshell, one could say that
internationalization is the operation by which further localizations
are made possible.
Also, very roughly said, when it comes to multi-lingual messages,
internationalization is usually taken care of by programmers, and
localization is usually taken care of by translators.
@node Aspects, Files, Concepts, Introduction
@section Aspects in Native Language Support
For a totally multi-lingual distribution, there are many things to
translate beyond output messages.
@itemize @bullet
@item
As of today, GNU @code{gettext} offers a complete toolset for
translating messages output by C programs. Perl scripts and shell
scripts will also need to be translated. Even if there are today some hooks
by which this can be done, these hooks are not integrated as well as they
should be.
@item
Some programs, like @code{autoconf} or @code{bison}, are able
to produce other programs (or scripts). Even if the generating
programs themselves are internationalized, the generated programs they
produce may need internationalization on their own, and this indirect
internationalization could be automated right from the generating
program. In fact, quite usually, generating and generated programs
could be internationalized independently, as the effort needed is
fairly orthogonal.
@item
A few programs include textual tables which might need translation
themselves, independently of the strings contained in the program
itself. For example, @w{RFC 1345} gives an English description for each
character which GNU @code{recode} is able to reconstruct at execution.
Since these descriptions are extracted from the RFC by mechanical means,
translating them properly would require a prior translation of the RFC
itself.
@item
Almost all programs accept options, which are often worded out so to
be descriptive for the English readers; one might want to consider
offering translated versions for program options as well.
@item
Many programs read, interpret, compile, or are somewhat driven by
input files which are texts containing keywords, identifiers, or
replies which are inherently translatable. For example, one may want
@code{gcc} to allow diacriticized characters in identifiers or use
translated keywords; @samp{rm -i} might accept something else than
@samp{y} or @samp{n} for replies, etc. Even if the program will
eventually make most of its output in the foreign languages, one has
to decide whether the input syntax, option values, etc., are to be
localized or not.
@item
The manual accompanying a package, as well as all documentation files
in the distribution, could surely be translated, too. Translating a
manual, with the intent of later keeping up with updates, is a major
undertaking in itself, generally.
@end itemize
As we already stressed, translation is only one aspect of locales.
Other internationalization aspects are not currently handled by GNU
@code{gettext}, but perhaps may be handled in future versions. There
are many attributes that are needed to define a country's cultural
conventions. These attributes include beside the country's native
language, the formatting of the date and time, the representation of
numbers, the symbols for currency, etc. These local @dfn{rules} are
termed the country's locale. The locale represents the knowledge
needed to support the country's native attributes.
There are a few major areas which may vary between countries and
hence, define what a locale must describe. The following list helps
putting multi-lingual messages into the proper context of other tasks
related to locales, and also presents some other areas which GNU
@code{gettext} might eventually tackle, maybe, one of these days.
@table @emph
@item Characters and Codesets
The codeset most commonly used through out the USA and most English
speaking parts of the world is the ASCII codeset. However, there are
many characters needed by various locales that are not found within
this codeset. The 8-bit @w{ISO 8859-1} code set has most of the special
characters needed to handle the major European languages. However, in
many cases, the @w{ISO 8859-1} font is not adequate. Hence each locale
will need to specify which codeset they need to use and will need
to have the appropriate character handling routines to cope with
the codeset.
@item Currency
The symbols used vary from country to country as does the position
used by the symbol. Software needs to be able to transparently
display currency figures in the native mode for each locale.
@item Dates
The format of date varies between locales. For example, Christmas day
in 1994 is written as 12/25/94 in the USA and as 25/12/94 in Australia.
Other countries might use @w{ISO 8061} dates, etc.
Time of the day may be noted as @var{hh}:@var{mm}, @var{hh}.@var{mm},
or otherwise. Some locales require time to be specified in 24-hour
mode rather than as AM or PM. Further, the nature and yearly extent
of the Daylight Saving correction vary widely between countries.
@item Numbers
Numbers can be represented differently in different locales.
For example, the following numbers are all written correctly for
their respective locales:
@example
12,345.67 English
12.345,67 French
1,2345.67 Asia
@end example
Some programs could go further and use different unit systems, like
English units or Metric units, or even take into account variants
about how numbers are spelled in full.
@item Messages
The most obvious area is the language support within a locale. This is
where GNU @code{gettext} provides the means for developers and users to
easily change the language that the software uses to communicate to
the user.
@end table
In the near future we see no chance that components of locale outside of
message handling will be made available for use in other
packages. The reason for this is that most modern systems provide
a more or less reasonable support for at least some of the missing
components. Another point is that the GNU @code{libc} and Linux will get
a new and complete implementation of the whole locale functionality
which could be adopted by system lacking a reasonable locale support.
@node Files, Overview, Aspects, Introduction
@section Files Conveying Translations
The letters PO in @file{.po} files means Portable Object, to
distinguish it from @file{.mo} files, where MO stands for Machine
Object. This paradigm, as well as the PO file format, is inspired
by the NLS standard developed by Uniforum, and implemented by Sun
in their Solaris system.
PO files are meant to be read and edited by humans, and associate each
original, translatable string of a given package with its translation
in a particular target language. A single PO file is dedicated to
a single target language. If a package supports many languages,
there is one such PO file per language supported, and each package
has its own set of PO files. These PO files are best created by
the @code{xgettext} program, and later updated or refreshed through
the @code{msgmerge} program. Program @code{xgettext} extracts all
marked messages from a set of C files and initializes a PO file with
empty translations. Program @code{msgmerge} takes care of adjusting
PO files between releases of the corresponding sources, commenting
obsolete entries, initializing new ones, and updating all source
line references. Files ending with @file{.pot} are kind of base
translation files found in distributions, in PO file format, and
@file{.pox} files are often temporary PO files.
MO files are meant to be read by programs, and are binary in nature.
A few systems already offer tools for creating and handling MO files
as part of the Native Language Support coming with the system, but the
format of these MO files is often different from system to system,
and non-portable. They do not necessary use @file{.mo} for file
extensions, but since system libraries are also used for accessing
these files, it works as long as the system is self-consistent about
it. If GNU @code{gettext} is able to interface with the tools already
provided with systems, it will consequently let these provided tools
take care of generating the MO files. Or else, if such tools are not
found or do not seem usable, GNU @code{gettext} will use its own ways
and its own format for MO files. Files ending with @file{.gmo} are
really MO files, when it is known that these files use the GNU format.
@node Overview, , Files, Introduction
@section Overview of GNU @code{gettext}
The following diagram summarizes the relation between the files
handled by GNU @code{gettext} and the tools acting on these files.
It is followed by a somewhat detailed explanations, which you should
read while keeping an eye on the diagram. Having a clear understanding
of these interrelations would surely help programmers, translators
and maintainers.
@example
@group
Original C Sources ---> PO mode ---> Marked C Sources ---.
|
.---------<--- GNU gettext Library |
.--- make <---+ |
| `---------<--------------------+-----------'
| |
| .-----<--- PACKAGE.pot <--- xgettext <---' .---<--- PO Compendium
| | | ^
| | `---. |
| `---. +---> PO mode ---.
| +----> msgmerge ------> LANG.pox --->--------' |
| .---' |
| | |
| `-------------<---------------. |
| +--- LANG.po <--- New LANG.pox <----'
| .--- LANG.gmo <--- msgfmt <---'
| |
| `---> install ---> /.../LANG/PACKAGE.mo ---.
| +---> "Hello world!"
`-------> install ---> /.../bin/PROGRAM -------'
@end group
@end example
The indication @samp{PO mode} appears in two places in this picture,
and you may safely read it as merely meaning ``hand editing'', using
any editor of your choice, really. However, for those of you being
the lucky users of GNU Emacs, PO mode has been specifically created
for providing a cozy environment for editing or modifying PO files.
While editing a PO file, PO mode allows for the easy browsing of
auxiliary and compendium PO files, as well as for following references into
the set of C program sources from which PO files have been derived.
It has a few special features, among which are the interactive marking
of program strings as translatable, and the validatation of PO files
with easy repositioning to PO file lines showing errors.
As a programmer, the first step to bringing GNU @code{gettext}
into your package is identifying, right in the C sources, those strings
which are meant to be translatable, and those which are untranslatable.
This tedious job can be done a little more comfortably using emacs PO
mode, but you can use any means familiar to you for modifying your
C sources. Beside this some other simple, standard changes are needed to
properly initialize the translation library. @xref{Sources}, for
more information about all this.
For newly written software the strings of course can and should be
marked while writing the it. The @code{gettext} approach makes this
very easy. Simply put the following lines at the beginning of each file
or in a central header file:
@example
@group
#define _(String) (String)
#define N_(String) (String)
#define textdomain(Domain)
#define bindtextdomain(Package, Directory)
@end group
@end example
@noindent
Doing this allows you to prepare the sources for internationalization.
Later when you feel ready for the step to use the @code{gettext} library
simply remove these definitions, include @file{libintl.h} and link
against @file{libintl.a}. That is all you have to change.
Once the C sources have been modified, the @code{xgettext} program
is used to find and extract all translatable strings, and create an
initial PO file out of all these. This @file{@var{package}.pot} file
contains all original program strings. It has sets of pointers to
exactly where in C sources each string is used. All translations
are set to empty. The letter @kbd{t} in @file{.pot} marks this as
a Template PO file, not yet oriented towards any particular language.
@xref{xgettext Invocation}, for more details about how one calls the
@code{xgettext} program. If you are @emph{really} lazy, you might
be interested at working a lot more right away, and preparing the
whole distribution setup (@pxref{Maintainers}). By doing so, you
spare yourself typing the @code{xgettext} command, as @code{make}
should now generate the proper things automatically for you!
The first time through, there is no @file{@var{lang}.po} yet, so the
@code{msgmerge} step may be skipped and replaced by a mere copy of
@file{@var{package}.pot} to @file{@var{lang}.pox}, where @var{lang}
represents the target language.
Then comes the initial translation of messages. Translation in
itself is a whole matter, still exclusively meant for humans,
and whose complexity far overwhelms the level of this manual.
Nevertheless, a few hints are given in some other chapter of this
manual (@pxref{Translators}). You will also find there indications
about how to contact translating teams, or becoming part of them,
for sharing your translating concerns with others who target the same
native language.
While adding the translated messages into the @file{@var{lang}.pox}
PO file, if you do not have GNU Emacs handy, you are on your own
for ensuring that your efforts fully respect the PO file format, and quoting
conventions (@pxref{PO Files}). This is surely not an impossible task,
as this is the way many people have handled PO files already for Uniforum or
Solaris. On the other hand, by using PO mode in GNU Emacs, most details
of PO file format are taken care of for you, but you have to acquire
some familiarity with PO mode itself. Besides main PO mode commands
(@pxref{Main PO Commands}), you should know how to move between entries
(@pxref{Entry Positioning}), and how to handle untranslated entries
(@pxref{Untranslated Entries}).
If some common translations have already been saved into a compendium
PO file, translators may use PO mode for initializing untranslated
entries from the compendium, and also save selected translations into
the compendium, updating it (@pxref{Compendium}). Compendium files
are meant to be exchanged between members of a given translation team.
Programs, or packages of programs, are dynamic in nature: users write
bug reports and suggestion for improvements, maintainers react by
modifying programs in various ways. The fact that a package has
already been internationalized should not make maintainers shy
of adding new strings, or modifying strings already translated.
They just do their job the best they can. For the Translation
Project to work smoothly, it is important that maintainers do not
carry translation concerns on their already loaded shoulders, and that
translators be kept as free as possible of programmatic concerns.
The only concern maintainers should have is carefully marking new
strings as translatable, when they should be, and do not otherwise
worry about them being translated, as this will come in proper time.
Consequently, when programs and their strings are adjusted in various
ways by maintainers, and for matters usually unrelated to translation,
@code{xgettext} would construct @file{@var{package}.pot} files which are
evolving over time, so the translations carried by @file{@var{lang}.po}
are slowly fading out of date.
It is important for translators (and even maintainers) to understand
that package translation is a continuous process in the lifetime of a
package, and not something which is done once and for all at the start.
After an initial burst of translation activity for a given package,
interventions are needed once in a while, because here and there,
translated entries become obsolete, and new untranslated entries
appear, needing translation.
The @code{msgmerge} program has the purpose of refreshing an already
existing @file{@var{lang}.po} file, by comparing it with a newer
@file{@var{package}.pot} template file, extracted by @code{xgettext}
out of recent C sources. The refreshing operation adjusts all
references to C source locations for strings, since these strings
move as programs are modified. Also, @code{msgmerge} comments out as
obsolete, in @file{@var{lang}.pox}, those already translated entries
which are no longer used in the program sources (@pxref{Obsolete
Entries}). It finally discovers new strings and inserts them in
the resulting PO file as untranslated entries (@pxref{Untranslated
Entries}). @xref{msgmerge Invocation}, for more information about what
@code{msgmerge} really does.
Whatever route or means taken, the goal is to obtain an updated
@file{@var{lang}.pox} file offering translations for all strings.
When this is properly achieved, this file @file{@var{lang}.pox} may
take the place of the previous official @file{@var{lang}.po} file.
The temporal mobility, or fluidity of PO files, is an integral part of
the translation game, and should be well understood, and accepted.
People resisting it will have a hard time participating in the
Translation Project, or will give a hard time to other participants! In
particular, maintainers should relax and include all available official
PO files in their distributions, even if these have not recently been
updated, without banging or otherwise trying to exert pressure on the
translator teams to get the job done. The pressure should rather come
from the community of users speaking a particular language, and
maintainers should consider themselves fairly relieved of any concern
about the adequacy of translation files. On the other hand, translators
should reasonably try updating the PO files they are responsible for,
while the package is undergoing pretest, prior to an official
distribution.
Once the PO file is complete and dependable, the @code{msgfmt} program
is used for turning the PO file into a machine-oriented format, which
may yield efficient retrieval of translations by the programs of the
package, whenever needed at runtime (@pxref{MO Files}). @xref{msgfmt
Invocation}, for more information about all modalities of execution
for the @code{msgfmt} program.
Finally, the modified and marked C sources are compiled and linked
with the GNU @code{gettext} library, usually through the operation of
@code{make}, given a suitable @file{Makefile} exists for the project,
and the resulting executable is installed somewhere users will find it.
The MO files themselves should also be properly installed. Given the
appropriate environment variables are set (@pxref{End Users}), the
program should localize itself automatically, whenever it executes.
The remainder of this manual has the purpose of explaining in depth the various
steps outlined above.
@node Basics, Sources, Introduction, Top
@chapter PO Files and PO Mode Basics
The GNU @code{gettext} toolset helps programmers and translators
at producing, updating and using translation files, mainly those
PO files which are textual, editable files. This chapter stresses
the format of PO files, and contains a PO mode starter. PO mode
description is spread throughout this manual instead of being concentrated
in one place. Here we present only the basics of PO mode.
@menu
* Installation:: Completing GNU @code{gettext} Installation
* PO Files:: The Format of PO Files
* Main PO Commands:: Main Commands
* Entry Positioning:: Entry Positioning
* Normalizing:: Normalizing Strings in Entries
@end menu
@node Installation, PO Files, Basics, Basics
@section Completing GNU @code{gettext} Installation
Once you have received, unpacked, configured and compiled the GNU
@code{gettext} distribution, the @samp{make install} command puts in
place the programs @code{xgettext}, @code{msgfmt}, @code{gettext}, and
@code{msgmerge}, as well as their available message catalogs. To
top off a comfortable installation, you might also want to make the
PO mode available to your GNU Emacs users.
During the installation of the PO mode, you might want modify your
file @file{.emacs}, once and for all, so it contains a few lines looking
like:
@example
(setq auto-mode-alist
(cons '("\\.po[tx]?\\'\\|\\.po\\." . po-mode) auto-mode-alist))
(autoload 'po-mode "po-mode")
@end example
Later, whenever you edit some @file{.po}, @file{.pot} or @file{.pox}
file, or any file having the string @samp{.po.} within its name,
Emacs loads @file{po-mode.elc} (or @file{po-mode.el}) as needed, and
automatically activates PO mode commands for the associated buffer.
The string @emph{PO} appears in the mode line for any buffer for
which PO mode is active. Many PO files may be active at once in a
single Emacs session.
If you are using Emacs version 20 or better, and have already installed
the appropriate international fonts on your system, you may also manage
for the these fonts to be automatically loaded and used for displaying
the translations on your Emacs screen, whenever necessary. For this to
happen, you might want to add the lines:
@example
(autoload 'po-find-file-coding-system "po-mode")
(modify-coding-system-alist 'file "\\.po[tx]?\\'\\|\\.po\\."
'po-find-file-coding-system)
@end example
@noindent
to your @file{.emacs} file.
@node PO Files, Main PO Commands, Installation, Basics
@section The Format of PO Files
A PO file is made up of many entries, each entry holding the relation
between an original untranslated string and its corresponding
translation. All entries in a given PO file usually pertain
to a single project, and all translations are expressed in a single
target language. One PO file @dfn{entry} has the following schematic
structure:
@example
@var{white-space}
# @var{translator-comments}
#. @var{automatic-comments}
#: @var{reference}@dots{}
#, @var{flag}@dots{}
msgid @var{untranslated-string}
msgstr @var{translated-string}
@end example
The general structure of a PO file should be well understood by
the translator. When using PO mode, very little has to be known
about the format details, as PO mode takes care of them for her.
Entries begin with some optional white space. Usually, when generated
through GNU @code{gettext} tools, there is exactly one blank line
between entries. Then comments follow, on lines all starting with the
character @kbd{#}. There are two kinds of comments: those which have
some white space immediately following the @kbd{#}, which comments are
created and maintained exclusively by the translator, and those which
have some non-white character just after the @kbd{#}, which comments
are created and maintained automatically by GNU @code{gettext} tools.
All comments, of either kind, are optional.
After white space and comments, entries show two strings, giving
first the untranslated string as it appears in the original program
sources, and then, the translation of this string. The original
string is introduced by the keyword @code{msgid}, and the translation,
by @code{msgstr}. The two strings, untranslated and translated,
are quoted in various ways in the PO file, using @kbd{"}
delimiters and @kbd{\} escapes, but the translator does not really
have to pay attention to the precise quoting format, as PO mode fully
intend to take care of quoting for her.
The @code{msgid} strings, as well as automatic comments, are produced
and managed by other GNU @code{gettext} tools, and PO mode does not
provide means for the translator to alter these. The most she can
do is merely deleting them, and only by deleting the whole entry.
On the other hand, the @code{msgstr} string, as well as translator
comments, are really meant for the translator, and PO mode gives her
the full control she needs.
The comment lines beginning with @kbd{#,} are special because they are
not completely ignored by the programs as comments generally are. The
comma separated list of @var{flag}s is used by the @code{msgfmt}
program to give the user some better disgnostic messages. Currently
there are two forms of flags defined:
@table @kbd
@item fuzzy
This flag can be generated by the @code{msgmerge} program or it can be
inserted by the translator herself. It shows that the @code{msgstr}
string might not be a correct translation (anymore). Only the translator
can judge if the translation requires further modification, or is
acceptable as is. Once satisfied with the translation, she then removes
this @kbd{fuzzy} attribute. The @code{msgmerge} programs inserts this
when it combined the @code{msgid} and @code{msgstr} entries after fuzzy
search only. @xref{Fuzzy Entries}.
@item c-format
@itemx no-c-format
These flags should not be added by a human. Instead only the
@code{xgettext} program adds them. In an automatized PO file processing
system as proposed here the user changes would be thrown away again as
soon as the @code{xgettext} program generates a new template file.
In case the @kbd{c-format} flag is given for a string the @code{msgfmt}
does some more tests to check to validity of the translation.
@xref{msgfmt Invocation}.
@end table
It happens that some lines, usually whitespace or comments, follow the
very last entry of a PO file. Such lines are not part of any entry,
and PO mode is unable to take action on those lines. By using the
PO mode function @w{@kbd{M-x po-normalize}}, the translator may get
rid of those spurious lines. @xref{Normalizing}.
The remainder of this section may be safely skipped by those using
PO mode, yet it may be interesting for everybody to have a better
idea of the precise format of a PO file. On the other hand, those
not having GNU Emacs handy should carefully continue reading on.
Each of @var{untranslated-string} and @var{translated-string} respects
the C syntax for a character string, including the surrounding quotes
and imbedded backslashed escape sequences. When the time comes
to write multi-line strings, one should not use escaped newlines.
Instead, a closing quote should follow the last character on the
line to be continued, and an opening quote should resume the string
at the beginning of the following PO file line. For example:
@example
msgid ""
"Here is an example of how one might continue a very long string\n"
"for the common case the string represents multi-line output.\n"
@end example
@noindent
In this example, the empty string is used on the first line, to
allow better alignment of the @kbd{H} from the word @samp{Here}
over the @kbd{f} from the word @samp{for}. In this example, the
@code{msgid} keyword is followed by three strings, which are meant
to be concatenated. Concatenating the empty string does not change
the resulting overall string, but it is a way for us to comply with
the necessity of @code{msgid} to be followed by a string on the same
line, while keeping the multi-line presentation left-justified, as
we find this to be a cleaner disposition. The empty string could have
been omitted, but only if the string starting with @samp{Here} was
promoted on the first line, right after @code{msgid}.@footnote{This
limitation is not imposed by GNU @code{gettext}, but comes from the
@code{msgfmt} implementation on Solaris.} It was not really necessary
either to switch between the two last quoted strings immediately after
the newline @samp{\n}, the switch could have occurred after @emph{any}
other character, we just did it this way because it is neater.
One should carefully distinguish between end of lines marked as
@samp{\n} @emph{inside} quotes, which are part of the represented
string, and end of lines in the PO file itself, outside string quotes,
which have no incidence on the represented string.
Outside strings, white lines and comments may be used freely.
Comments start at the beginning of a line with @samp{#} and extend
until the end of the PO file line. Comments written by translators
should have the initial @samp{#} immediately followed by some white
space. If the @samp{#} is not immediately followed by white space,
this comment is most likely generated and managed by specialized GNU
tools, and might disappear or be replaced unexpectedly when the PO
file is given to @code{msgmerge}.
@node Main PO Commands, Entry Positioning, PO Files, Basics
@section Main PO mode Commands
After setting up Emacs with something similar to the lines in
@ref{Installation}, PO mode is activated for a window when Emacs finds a
PO file in that window. This puts the window read-only and establishes a
po-mode-map, which is a genuine Emacs mode, in a way that is not derived
from text mode in any way. Functions found on @code{po-mode-hook},
if any, will be executed.
When PO mode is active in a window, the letters @samp{PO} appear
in the mode line for that window. The mode line also displays how
many entries of each kind are held in the PO file. For example,
the string @samp{132t+3f+10u+2o} would tell the translator that the
PO mode contains 132 translated entries (@pxref{Translated Entries},
3 fuzzy entries (@pxref{Fuzzy Entries}), 10 untranslated entries
(@pxref{Untranslated Entries}) and 2 obsolete entries (@pxref{Obsolete
Entries}). Zero-coefficients items are not shown. So, in this example, if
the fuzzy entries were unfuzzied, the untranslated entries were translated
and the obsolete entries were deleted, the mode line would merely display
@samp{145t} for the counters.
The main PO commands are those which do not fit into the other categories of
subsequent sections. These allow for quitting PO mode or for managing windows
in special ways.
@table @kbd
@item U
Undo last modification to the PO file.
@item Q
Quit processing and save the PO file.
@item q
Quit processing, possibly after confirmation.
@item O
Temporary leave the PO file window.
@item ?
@itemx h
Show help about PO mode.
@item =
Give some PO file statistics.
@item V
Batch validate the format of the whole PO file.
@end table
The command @kbd{U} (@code{po-undo}) interfaces to the GNU Emacs
@emph{undo} facility. @xref{Undo, , Undoing Changes, emacs, The Emacs
Editor}. Each time @kbd{U} is typed, modifications which the translator
did to the PO file are undone a little more. For the purpose of
undoing, each PO mode command is atomic. This is especially true for
the @kbd{@key{RET}} command: the whole edition made by using a single
use of this command is undone at once, even if the edition itself
implied several actions. However, while in the editing window, one
can undo the edition work quite parsimoniously.
The commands @kbd{Q} (@code{po-quit}) and @kbd{q}
(@code{po-confirm-and-quit}) are used when the translator is done with the
PO file. The former is a bit less verbose than the latter. If the file
has been modified, it is saved to disk first. In both cases, and prior to
all this, the commands check if some untranslated message remains in the
PO file and, if yes, the translator is asked if she really wants to leave
off working with this PO file. This is the preferred way of getting rid
of an Emacs PO file buffer. Merely killing it through the usual command
@w{@kbd{C-x k}} (@code{kill-buffer}) is not the tidiest way to proceed.
The command @kbd{O} (@code{po-other-window}) is another, softer way,
to leave PO mode, temporarily. It just moves the cursor to some other
Emacs window, and pops one if necessary. For example, if the translator
just got PO mode to show some source context in some other, she might
discover some apparent bug in the program source that needs correction.
This command allows the translator to change sex, become a programmer,
and have the cursor right into the window containing the program she
(or rather @emph{he}) wants to modify. By later getting the cursor back
in the PO file window, or by asking Emacs to edit this file once again,
PO mode is then recovered.
The command @kbd{h} (@code{po-help}) displays a summary of all available PO
mode commands. The translator should then type any character to resume
normal PO mode operations. The command @kbd{?} has the same effect
as @kbd{h}.
The command @kbd{=} (@code{po-statistics}) computes the total number of
entries in the PO file, the ordinal of the current entry (counted from
1), the number of untranslated entries, the number of obsolete entries,
and displays all these numbers.
The command @kbd{V} (@code{po-validate}) launches @code{msgfmt} in verbose
mode over the current PO file. This command first offers to save the
current PO file on disk. The @code{msgfmt} tool, from GNU @code{gettext},
has the purpose of creating a MO file out of a PO file, and PO mode uses
the features of this program for checking the overall format of a PO file,
as well as all individual entries.
The program @code{msgfmt} runs asynchronously with Emacs, so the
translator regains control immediately while her PO file is being studied.
Error output is collected in the GNU Emacs @samp{*compilation*} buffer,
displayed in another window. The regular GNU Emacs command @kbd{C-x`}
(@code{next-error}), as well as other usual compile commands, allow the
translator to reposition quickly to the offending parts of the PO file.
Once the cursor is on the line in error, the translator may decide on
any PO mode action which would help correcting the error.
@node Entry Positioning, Normalizing, Main PO Commands, Basics
@section Entry Positioning
The cursor in a PO file window is almost always part of
an entry. The only exceptions are the special case when the cursor
is after the last entry in the file, or when the PO file is
empty. The entry where the cursor is found to be is said to be the
current entry. Many PO mode commands operate on the current entry,
so moving the cursor does more than allowing the translator to browse
the PO file, this also selects on which entry commands operate.
Some PO mode commands alter the position of the cursor in a specialized
way. A few of those special purpose positioning are described here,
the others are described in following sections.
@table @kbd
@item .
Redisplay the current entry.
@item n
@itemx n
Select the entry after the current one.
@item p
@itemx p
Select the entry before the current one.
@item <
Select the first entry in the PO file.
@item >
Select the last entry in the PO file.
@item m
Record the location of the current entry for later use.
@item l
Return to a previously saved entry location.
@item x
Exchange the current entry location with the previously saved one.
@end table
Any GNU Emacs command able to reposition the cursor may be used
to select the current entry in PO mode, including commands which
move by characters, lines, paragraphs, screens or pages, and search
commands. However, there is a kind of standard way to display the
current entry in PO mode, which usual GNU Emacs commands moving
the cursor do not especially try to enforce. The command @kbd{.}
(@code{po-current-entry}) has the sole purpose of redisplaying the
current entry properly, after the current entry has been changed by
means external to PO mode, or the Emacs screen otherwise altered.
It is yet to be decided if PO mode helps the translator, or otherwise
irritates her, by forcing a rigid window disposition while she
is doing her work. We originally had quite precise ideas about
how windows should behave, but on the other hand, anyone used to
GNU Emacs is often happy to keep full control. Maybe a fixed window
disposition might be offered as a PO mode option that the translator
might activate or deactivate at will, so it could be offered on an
experimental basis. If nobody feels a real need for using it, or
a compulsion for writing it, we should drop this whole idea.
The incentive for doing it should come from translators rather than
programmers, as opinions from an experienced translator are surely
more worth to me than opinions from programmers @emph{thinking} about
how @emph{others} should do translation.
The commands @kbd{n} (@code{po-next-entry}) and @kbd{p}
(@code{po-previous-entry}) move the cursor the entry following,
or preceding, the current one. If @kbd{n} is given while the
cursor is on the last entry of the PO file, or if @kbd{p}
is given while the cursor is on the first entry, no move is done.
The commands @kbd{<} (@code{po-first-entry}) and @kbd{>}
(@code{po-last-entry}) move the cursor to the first entry, or last
entry, of the PO file. When the cursor is located past the last
entry in a PO file, most PO mode commands will return an error saying
@samp{After last entry}. Moreover, the commands @kbd{<} and @kbd{>}
have the special property of being able to work even when the cursor
is not into some PO file entry, and one may use them for nicely
correcting this situation. But even these commands will fail on a
truly empty PO file. There are development plans for the PO mode for it
to interactively fill an empty PO file from sources. @xref{Marking}.
The translator may decide, before working at the translation of
a particular entry, that she needs to browse the remainder of the
PO file, maybe for finding the terminology or phraseology used
in related entries. She can of course use the standard Emacs idioms
for saving the current cursor location in some register, and use that
register for getting back, or else, use the location ring.
PO mode offers another approach, by which cursor locations may be saved
onto a special stack. The command @kbd{m} (@code{po-push-location})
merely adds the location of current entry to the stack, pushing
the already saved locations under the new one. The command
@kbd{r} (@code{po-pop-location}) consumes the top stack element and
reposition the cursor to the entry associated with that top element.
This position is then lost, for the next @kbd{r} will move the cursor
to the previously saved location, and so on until no locations remain
on the stack.
If the translator wants the position to be kept on the location stack,
maybe for taking a look at the entry associated with the top
element, then go elsewhere with the intent of getting back later, she
ought to use @kbd{m} immediately after @kbd{r}.
The command @kbd{x} (@code{po-exchange-location}) simultaneously
reposition the cursor to the entry associated with the top element of
the stack of saved locations, and replace that top element with the
location of the current entry before the move. Consequently, repeating
the @kbd{x} command toggles alternatively between two entries.
For achieving this, the translator will position the cursor on the
first entry, use @kbd{m}, then position to the second entry, and
merely use @kbd{x} for making the switch.
@node Normalizing, , Entry Positioning, Basics
@section Normalizing Strings in Entries
There are many different ways for encoding a particular string into a
PO file entry, because there are so many different ways to split and
quote multi-line strings, and even, to represent special characters
by backslahsed escaped sequences. Some features of PO mode rely on
the ability for PO mode to scan an already existing PO file for a
particular string encoded into the @code{msgid} field of some entry.
Even if PO mode has internally all the built-in machinery for
implementing this recognition easily, doing it fast is technically
difficult. To facilitate a solution to this efficiency problem,
we decided on a canonical representation for strings.
A conventional representation of strings in a PO file is currently
under discussion, and PO mode experiments with a canonical representation.
Having both @code{xgettext} and PO mode converging towards a uniform
way of representing equivalent strings would be useful, as the internal
normalization needed by PO mode could be automatically satisfied
when using @code{xgettext} from GNU @code{gettext}. An explicit
PO mode normalization should then be only necessary for PO files
imported from elsewhere, or for when the convention itself evolves.
So, for achieving normalization of at least the strings of a given
PO file needing a canonical representation, the following PO mode
command is available:
@table @kbd
@item M-x po-normalize
Tidy the whole PO file by making entries more uniform.
@end table
The special command @kbd{M-x po-normalize}, which has no associate
keys, revises all entries, ensuring that strings of both original
and translated entries use uniform internal quoting in the PO file.
It also removes any crumb after the last entry. This command may be
useful for PO files freshly imported from elsewhere, or if we ever
improve on the canonical quoting format we use. This canonical format
is not only meant for getting cleaner PO files, but also for greatly
speeding up @code{msgid} string lookup for some other PO mode commands.
@kbd{M-x po-normalize} presently makes three passes over the entries.
The first implements heuristics for converting PO files for GNU
@code{gettext} 0.6 and earlier, in which @code{msgid} and @code{msgstr}
fields were using K&R style C string syntax for multi-line strings.
These heuristics may fail for comments not related to obsolete
entries and ending with a backslash; they also depend on subsequent
passes for finalizing the proper commenting of continued lines for
obsolete entries. This first pass might disappear once all oldish PO
files would have been adjusted. The second and third pass normalize
all @code{msgid} and @code{msgstr} strings respectively. They also
clean out those trailing backslashes used by XView's @code{msgfmt}
for continued lines.
Having such an explicit normalizing command allows for importing PO
files from other sources, but also eases the evolution of the current
convention, evolution driven mostly by aesthetic concerns, as of now.
It is easy to make suggested adjustments at a later time, as the
normalizing command and eventually, other GNU @code{gettext} tools
should greatly automate conformance. A description of the canonical
string format is given below, for the particular benefit of those not
having GNU Emacs handy, and who would nevertheless want to handcraft
their PO files in nice ways.
Right now, in PO mode, strings are single line or multi-line. A string
goes multi-line if and only if it has @emph{embedded} newlines, that
is, if it matches @samp{[^\n]\n+[^\n]}. So, we would have:
@example
msgstr "\n\nHello, world!\n\n\n"
@end example
but, replacing the space by a newline, this becomes:
@example
msgstr ""
"\n"
"\n"
"Hello,\n"
"world!\n"
"\n"
"\n"
@end example
We are deliberately using a caricatural example, here, to make the
point clearer. Usually, multi-lines are not that bad looking.
It is probable that we will implement the following suggestion.
We might lump together all initial newlines into the empty string,
and also all newlines introducing empty lines (that is, for @w{@var{n}
> 1}, the @var{n}-1'th last newlines would go together on a separate
string), so making the previous example appear:
@example
msgstr "\n\n"
"Hello,\n"
"world!\n"
"\n\n"
@end example
There are a few yet undecided little points about string normalization,
to be documented in this manual, once these questions settle.
@node Sources, Initial, Basics, Top
@chapter Preparing Program Sources
@c FIXME: Rewrite (the whole chapter).
For the programmer, changes to the C source code fall into three
categories. First, you have to make the localization functions
known to all modules needing message translation. Second, you should
properly trigger the operation of GNU @code{gettext} when the program
initializes, usually from the @code{main} function. Last, you should
identify and especially mark all constant strings in your program
needing translation.
Presuming that your set of programs, or package, has been adjusted
so all needed GNU @code{gettext} files are available, and your
@file{Makefile} files are adjusted (@pxref{Maintainers}), each C module
having translated C strings should contain the line:
@example
#include <libintl.h>
@end example
The remaining changes to your C sources are discussed in the further
sections of this chapter.
@menu
* Triggering:: Triggering @code{gettext} Operations
* Mark Keywords:: How Marks Appears in Sources
* Marking:: Marking Translatable Strings
* c-format:: Telling something about the following string
* Special cases:: Special Cases of Translatable Strings
@end menu
@node Triggering, Mark Keywords, Sources, Sources
@section Triggering @code{gettext} Operations
The initialization of locale data should be done with more or less
the same code in every program, as demonstrated below:
@example
@group
int
main (argc, argv)
int argc;
char argv;
@{
@dots{}
setlocale (LC_ALL, "");
bindtextdomain (PACKAGE, LOCALEDIR);
textdomain (PACKAGE);
@dots{}
@}
@end group
@end example
@var{PACKAGE} and @var{LOCALEDIR} should be provided either by
@file{config.h} or by the Makefile. For now consult the @code{gettext}
sources for more information.
The use of @code{LC_ALL} might not be appropriate for you.
@code{LC_ALL} includes all locale categories and especially
@code{LC_CTYPE}. This later category is responsible for determining
character classes with the @code{isalnum} etc. functions from
@file{ctype.h} which could especially for programs, which process some
kind of input language, be wrong. For example this would mean that a
source code using the @,{c} (c-cedilla character) is runnable in
France but not in the U.S.
Some systems also have problems with parsing number using the
@code{scanf} functions if an other but the @code{LC_ALL} locale is used.
The standards say that additional formats but the one known in the
@code{"C"} locale might be recognized. But some systems seem to reject
numbers in the @code{"C"} locale format. In some situation, it might
also be a problem with the notation itself which makes it impossible to
recognize whether the number is in the @code{"C"} locale or the local
format. This can happen if thousands separator characters are used.
Some locales define this character according to the national
conventions to @code{'.'} which is the same character used in the
@code{"C"} locale to denote the decimal point.
So it is sometimes necessary to replace the @code{LC_ALL} line in the
code above by a sequence of @code{setlocale} lines
@example
@group
@{
@dots{}
setlocale (LC_TIME, "");
setlocale (LC_MESSAGES, "");
@dots{}
@}
@end group
@end example
@noindent
or to switch for and back to the character class in question. On all
POSIX conformant systems the locale categories @code{LC_CTYPE},
@code{LC_COLLATE}, @code{LC_MONETARY}, @code{LC_NUMERIC}, and
@code{LC_TIME} are available. On some modern systems there is also a
locale @code{LC_MESSAGES} which is called on some old, XPG2 compliant
systems @code{LC_RESPONSES}.
@node Mark Keywords, Marking, Triggering, Sources
@section How Marks Appears in Sources
All strings requiring translation should be marked in the C sources. Marking
is done in such a way that each translatable string appears to be
the sole argument of some function or preprocessor macro. There are
only a few such possible functions or macros meant for translation,
and their names are said to be marking keywords. The marking is
attached to strings themselves, rather than to what we do with them.
This approach has more uses. A blatant example is an error message
produced by formatting. The format string needs translation, as
well as some strings inserted through some @samp{%s} specification
in the format, while the result from @code{sprintf} may have so many
different instances that it is impractical to list them all in some
@samp{error_string_out()} routine, say.
This marking operation has two goals. The first goal of marking
is for triggering the retrieval of the translation, at run time.
The keyword are possibly resolved into a routine able to dynamically
return the proper translation, as far as possible or wanted, for the
argument string. Most localizable strings are found in executable
positions, that is, attached to variables or given as parameters to
functions. But this is not universal usage, and some translatable
strings appear in structured initializations. @xref{Special cases}.
The second goal of the marking operation is to help @code{xgettext}
at properly extracting all translatable strings when it scans a set
of program sources and produces PO file templates.
The canonical keyword for marking translatable strings is
@samp{gettext}, it gave its name to the whole GNU @code{gettext}
package. For packages making only light use of the @samp{gettext}
keyword, macro or function, it is easily used @emph{as is}. However,
for packages using the @code{gettext} interface more heavily, it
is usually more convenient to give the main keyword a shorter, less
obtrusive name. Indeed, the keyword might appear on a lot of strings
all over the package, and programmers usually do not want nor need
their program sources to remind them forcefully, all the time, that they
are internationalized. Further, a long keyword has the disadvantage
of using more horizontal space, forcing more indentation work on
sources for those trying to keep them within 79 or 80 columns.
Many packages use @samp{_} (a simple underline) as a keyword,
and write @samp{_("Translatable string")} instead of @samp{gettext
("Translatable string")}. Further, the coding rule, from GNU standards,
wanting that there is a space between the keyword and the opening
parenthesis is relaxed, in practice, for this particular usage.
So, the textual overhead per translatable string is reduced to
only three characters: the underline and the two parentheses.
However, even if GNU @code{gettext} uses this convention internally,
it does not offer it officially. The real, genuine keyword is truly
@samp{gettext} indeed. It is fairly easy for those wanting to use
@samp{_} instead of @samp{gettext} to declare:
@example
#include <libintl.h>
#define _(String) gettext (String)
@end example
@noindent
instead of merely using @samp{#include <libintl.h>}.
Later on, the maintenance is relatively easy. If, as a programmer,
you add or modify a string, you will have to ask yourself if the
new or altered string requires translation, and include it within
@samp{_()} if you think it should be translated. @samp{"%s: %d"} is
an example of string @emph{not} requiring translation!
@node Marking, c-format, Mark Keywords, Sources
@section Marking Translatable Strings
In PO mode, one set of features is meant more for the programmer than
for the translator, and allows him to interactively mark which strings,
in a set of program sources, are translatable, and which are not.
Even if it is a fairly easy job for a programmer to find and mark
such strings by other means, using any editor of his choice, PO mode
makes this work more comfortable. Further, this gives translators
who feel a little like programmers, or programmers who feel a little
like translators, a tool letting them work at marking translatable
strings in the program sources, while simultaneously producing a set of
translation in some language, for the package being internationalized.
The set of program sources, targetted by the PO mode commands describe
here, should have an Emacs tags table constructed for your project,
prior to using these PO file commands. This is easy to do. In any
shell window, change the directory to the root of your project, then
execute a command resembling:
@example
etags src/*.[hc] lib/*.[hc]
@end example
@noindent
presuming here you want to process all @file{.h} and @file{.c} files
from the @file{src/} and @file{lib/} directories. This command will
explore all said files and create a @file{TAGS} file in your root
directory, somewhat summarizing the contents using a special file
format Emacs can understand.
For packages following the GNU coding standards, there is
a make goal @code{tags} or @code{TAGS} which construct the tag files in
all directories and for all files containing source code.
Once your @file{TAGS} file is ready, the following commands assist
the programmer at marking translatable strings in his set of sources.
But these commands are necessarily driven from within a PO file
window, and it is likely that you do not even have such a PO file yet.
This is not a problem at all, as you may safely open a new, empty PO
file, mainly for using these commands. This empty PO file will slowly
fill in while you mark strings as translatable in your program sources.
@table @kbd
@item ,
Search through program sources for a string which looks like a
candidate for translation.
@item M-,
Mark the last string found with @samp{_()}.
@item M-.
Mark the last string found with a keyword taken from a set of possible
keywords. This command with a prefix allows some management of these
keywords.
@end table
The @kbd{,} (@code{po-tags-search}) command search for the next
occurrence of a string which looks like a possible candidate for
translation, and displays the program source in another Emacs window,
positioned in such a way that the string is near the top of this other
window. If the string is too big to fit whole in this window, it is
positioned so only its end is shown. In any case, the cursor
is left in the PO file window. If the shown string would be better
presented differently in different native languages, you may mark it
using @kbd{M-,} or @kbd{M-.}. Otherwise, you might rather ignore it
and skip to the next string by merely repeating the @kbd{,} command.
A string is a good candidate for translation if it contains a sequence
of three or more letters. A string containing at most two letters in
a row will be considered as a candidate if it has more letters than
non-letters. The command disregards strings containing no letters,
or isolated letters only. It also disregards strings within comments,
or strings already marked with some keyword PO mode knows (see below).
If you have never told Emacs about some @file{TAGS} file to use, the
command will request that you specify one from the minibuffer, the
first time you use the command. You may later change your @file{TAGS}
file by using the regular Emacs command @w{@kbd{M-x visit-tags-table}},
which will ask you to name the precise @file{TAGS} file you want
to use. @xref{Tags, , Tag Tables, emacs, The Emacs Editor}.
Each time you use the @kbd{,} command, the search resumes from where it was
left by the previous search, and goes through all program sources,
obeying the @file{TAGS} file, until all sources have been processed.
However, by giving a prefix argument to the command @w{(@kbd{C-u
,})}, you may request that the search be restarted all over again
from the first program source; but in this case, strings that you
recently marked as translatable will be automatically skipped.
Using this @kbd{,} command does not prevent using of other regular
Emacs tags commands. For example, regular @code{tags-search} or
@code{tags-query-replace} commands may be used without disrupting the
independent @kbd{,} search sequence. However, as implemented, the
@emph{initial} @kbd{,} command (or the @kbd{,} command is used with a
prefix) might also reinitialize the regular Emacs tags searching to the
first tags file, this reinitialization might be considered spurious.
The @kbd{M-,} (@code{po-mark-translatable}) command will mark the
recently found string with the @samp{_} keyword. The @kbd{M-.}
(@code{po-select-mark-and-mark}) command will request that you type
one keyword from the minibuffer and use that keyword for marking
the string. Both commands will automatically create a new PO file
untranslated entry for the string being marked, and make it the
current entry (making it easy for you to immediately proceed to its
translation, if you feel like doing it right away). It is possible
that the modifications made to the program source by @kbd{M-,} or
@kbd{M-.} render some source line longer than 80 columns, forcing you
to break and re-indent this line differently. You may use the @kbd{O}
command from PO mode, or any other window changing command from
GNU Emacs, to break out into the program source window, and do any
needed adjustments. You will have to use some regular Emacs command
to return the cursor to the PO file window, if you want command
@kbd{,} for the next string, say.
The @kbd{M-.} command has a few built-in speedups, so you do not
have to explicitly type all keywords all the time. The first such
speedup is that you are presented with a @emph{preferred} keyword,
which you may accept by merely typing @kbd{@key{RET}} at the prompt.
The second speedup is that you may type any non-ambiguous prefix of the
keyword you really mean, and the command will complete it automatically
for you. This also means that PO mode has to @emph{know} all
your possible keywords, and that it will not accept mistyped keywords.
If you reply @kbd{?} to the keyword request, the command gives a
list of all known keywords, from which you may choose. When the
command is prefixed by an argument @w{(@kbd{C-u M-.})}, it inhibits
updating any program source or PO file buffer, and does some simple
keyword management instead. In this case, the command asks for a
keyword, written in full, which becomes a new allowed keyword for
later @kbd{M-.} commands. Moreover, this new keyword automatically
becomes the @emph{preferred} keyword for later commands. By typing
an already known keyword in response to @w{@kbd{C-u M-.}}, one merely
changes the @emph{preferred} keyword and does nothing more.
All keywords known for @kbd{M-.} are recognized by the @kbd{,} command
when scanning for strings, and strings already marked by any of those
known keywords are automatically skipped. If many PO files are opened
simultaneously, each one has its own independent set of known keywords.
There is no provision in PO mode, currently, for deleting a known
keyword, you have to quit the file (maybe using @kbd{q}) and reopen
it afresh. When a PO file is newly brought up in an Emacs window, only
@samp{gettext} and @samp{_} are known as keywords, and @samp{gettext}
is preferred for the @kbd{M-.} command. In fact, this is not useful to
prefer @samp{_}, as this one is already built in the @kbd{M-,} command.
@node c-format, Special cases, Marking, Sources
@section Special Comments preceding Keywords
@c FIXME document c-format and no-c-format.
In C programs strings are often used within calls of functions from the
@code{printf} family. The special thing about these format strings is
that they can contain format specifiers introduced with @kbd{%}. Assume
we have the code
@example
printf (gettext ("String `%s' has %d characters\n"), s, strlen (s));
@end example
@noindent
A possible German translation for the above string might be:
@example
"%d Zeichen lang ist die Zeichenkette `%s'"
@end example
A C programmer, even if he cannot speak German, will recognize that
there is something wrong here. The order of the two format specifiers
is changed but of course the arguments in the @code{printf} don't have.
This will most probably lead to problems because now the length of the
string is regarded as the address.
To prevent errors at runtime caused by translations the @code{msgfmt}
tool can check statically whether the arguments in the original and the
translation string match in type and number. If this is not the case a
warning will be given and the error cannot causes problems at runtime.
@noindent
If the word order in the above German translation would be correct one
would have to write
@example
"%2$d Zeichen lang ist die Zeichenkette `%1$s'"
@end example
@noindent
The routines in @code{msgfmt} know about this special notation.
Because not all strings in a program must be format strings it is not
useful for @code{msgfmt} to test all the strings in the @file{.po} file.
This might cause problems because the string might contain what looks
like a format specifier, but the string is not used in @code{printf}.
Therefore the @code{xgettext} adds a special tag to those messages it
thinks might be a format string. There is no absolute rule for this,
only a heuristic. In the @file{.po} file the entry is marked using the
@code{c-format} flag in the @kbd{#,} comment line (@pxref{PO Files}).
The careful reader now might say that this again can cause problems.
The heuristic might guess it wrong. This is true and therefore
@code{xgettext} knows about special kind of comment which lets
the programmer take over the decision. If in the same line or
the immediately preceding line of the @code{gettext} keyword
the @code{xgettext} program find a comment containing the words
@kbd{xgettext:c-format} it will mark the string in any case with
the @kbd{c-format} flag. This kind of comment should be used when
@code{xgettext} does not recognize the string as a format string but
is really is one and it should be tested. Please note that when the
comment is in the same line of the @code{gettext} keyword, it must be
before the string to be translated.
This situation happens quite often. The @code{printf} function is often
called with strings which do not contain a format specifier. Of course
one would normally use @code{fputs} but it does happen. In this case
@code{xgettext} does not recognize this as a format string but what
happens if the translation introduces a valid format specifier? The
@code{printf} function will try to access one of the parameter but none
exists because the original code does not refer to any parameter.
@code{xgettext} of course could make a wrong decision the other way
round. A string marked as a format string is not really a format
string. In this case the @code{msgfmt} might give too many warnings and
would prevent translating the @file{.po} file. The method to prevent
this wrong decision is similar to the one used above, only the comment
to use must contain the string @kbd{xgettext:no-c-format}.
If a string is marked with @kbd{c-format} and this is not correct the
user can find out who is responsible for the decision. @xref{xgettext
Invocation} to see how the @kbd{--debug} option can be used for solving
this problem.
@node Special cases, , c-format, Sources
@section Special Cases of Translatable Strings
The attentive reader might now point out that it is not always possible
to mark translatable string with @code{gettext} or something like this.
Consider the following case:
@example
@group
@{
static const char *messages[] = @{
"some very meaningful message",
"and another one"
@};
const char *string;
@dots{}
string
= index > 1 ? "a default message" : messages[index];
fputs (string);
@dots{}
@}
@end group
@end example
While it is no problem to mark the string @code{"a default message"} it
is not possible to mark the string initializers for @code{messages}.
What is to be done? We have to fulfill two tasks. First we have to mark the
strings so that the @code{xgettext} program (@pxref{xgettext Invocation})
can find them, and second we have to translate the string at runtime
before printing them.
The first task can be fulfilled by creating a new keyword, which names a
no-op. For the second we have to mark all access points to a string
from the array. So one solution can look like this:
@example
@group
#define gettext_noop(String) (String)
@{
static const char *messages[] = @{
gettext_noop ("some very meaningful message"),
gettext_noop ("and another one")
@};
const char *string;
@dots{}
string
= index > 1 ? gettext ("a default message") : gettext (messages[index]);
fputs (string);
@dots{}
@}
@end group
@end example
Please convince yourself that the string which is written by
@code{fputs} is translated in any case. How to get @code{xgettext} know
the additional keyword @code{gettext_noop} is explained in @ref{xgettext
Invocation}.
The above is of course not the only solution. You could also come along
with the following one:
@example
@group
#define gettext_noop(String) (String)
@{
static const char *messages[] = @{
gettext_noop ("some very meaningful message",
gettext_noop ("and another one")
@};
const char *string;
@dots{}
string
= index > 1 ? gettext_noop ("a default message") : messages[index];
fputs (gettext (string));
@dots{}
@}
@end group
@end example
But this has some drawbacks. First the programmer has to take care that
he uses @code{gettext_noop} for the string @code{"a default message"}.
A use of @code{gettext} could have in rare cases unpredictable results.
The second reason is found in the internals of the GNU @code{gettext}
Library which will make this solution less efficient.
One advantage is that you need not make control flow analysis to make
sure the output is really translated in any case. But this analysis is
generally not very difficult. If it should be in any situation you can
use this second method in this situation.
@node Initial, Updating, Sources, Top
@chapter Making the Initial PO File
@c FIXME: Rewrite.
@menu
* xgettext Invocation:: Invoking the @code{xgettext} Program
* C Sources Context:: C Sources Context
* Compendium:: Using Translation Compendiums
@end menu
@node xgettext Invocation, C Sources Context, Initial, Initial
@section Invoking the @code{xgettext} Program
@c FIXME: Rewrite.
@example
xgettext [@var{option}] @var{inputfile} @dots{}
@end example
@table @samp
@item -a
@itemx --extract-all
Extract all strings.
@item -c [@var{tag}]
@itemx --add-comments[=@var{tag}]
Place comment block with @var{tag} (or those preceding keyword lines)
in output file.
@item -C
@itemx --c++
Recognize C++ style comments.
@itemx --debug
Use the flags @kbd{c-format} and @kbd{possible-c-format} to show who was
responsible for marking a message as a format string. The later form is
used if the @code{xgettext} program decided, the format form is used if
the programmer prescribed it.
By default only the @kbd{c-format} form is used. The translator should
not have to care about these details.
@item -d @var{name}
@itemx --default-domain=@var{name}
Use @file{@var{name}.po} for output (instead of @file{messages.po}).
The special domain name @file{-} or @file{/dev/stdout} means to write
the output to @file{stdout}.
@item -D @var{directory}
@itemx --directory=@var{directory}
Change to @var{directory} before beginning to search and scan source
files. The resulting @file{.po} file will be written relative to the
original directory, though.
@item -f @var{file}
@itemx --files-from=@var{file}
Read the names of the input files from @var{file} instead of getting
them from the command line.
@itemx --force
Always write output file even if no message is defined.
@item -h
@itemx --help
Display this help and exit.
@item -I @var{list}
@itemx --input-path=@var{list}
List of directories searched for input files.
@item -j
@itemx --join-existing
Join messages with existing file.
@item -k @var{word}
@itemx --keyword[=@var{word}]
Additonal keyword to be looked for (without @var{word} means not to
use default keywords).
The default keywords, which are always looked for if not explicitly
disabled, are @code{gettext}, @code{dgettext}, @code{dcgettext} and
@code{gettext_noop}.
@item -m [@var{string}]
@itemx --msgstr-prefix[=@var{string}]
Use @var{string} or "" as prefix for msgstr entries.
@item -M [@var{string}]
@itemx --msgstr-suffix[=@var{string}]
Use @var{string} or "" as suffix for msgstr entries.
@item --no-location
Do not write @samp{#: @var{filename}:@var{line}} lines.
@item -n
@itemx --add-location
Generate @samp{#: @var{filename}:@var{line}} lines (default).
@item --omit-header
Don't write header with @samp{msgid ""} entry.
This is useful for testing purposes because it eliminates a source
of variance for generated @code{.gmo} files. We can ship some of
these files in the GNU @code{gettext} package, and the result of
regenerating them through @code{msgfmt} should yield the same values.
@item -p @var{dir}
@itemx --output-dir=@var{dir}
Output files will be placed in directory @var{dir}.
@item -s
@itemx --sort-output
Generate sorted output and remove duplicates.
@item --strict
Write out strict Uniforum conforming PO file.
@item -v
@itemx --version
Output version information and exit.
@item -x @var{file}
@itemx --exclude-file=@var{file}
Entries from @var{file} are not extracted.
@end table
Search path for supplementary PO files is:
@file{/usr/local/share/nls/src/}.
If @var{inputfile} is @samp{-}, standard input is read.
This implementation of @code{xgettext} is able to process a few awkward
cases, like strings in preprocessor macros, ANSI concatenation of
adjacent strings, and escaped end of lines for continued strings.
@node C Sources Context, Compendium, xgettext Invocation, Initial
@section C Sources Context
PO mode is particularily powerful when used with PO files
created through GNU @code{gettext} utilities, as those utilities
insert special comments in the PO files they generate.
Some of these special comments relate the PO file entry to
exactly where the untranslated string appears in the program sources.
When the translator gets to an untranslated entry, she is fairly
often faced with an original string which is not as informative as
it normally should be, being succinct, cryptic, or otherwise ambiguous.
Before chosing how to translate the string, she needs to understand
better what the string really means and how tight the translation has
to be. Most of times, when problems arise, the only way left to make
her judgment is looking at the true program sources from where this
string originated, searching for surrounding comments the programmer
might have put in there, and looking around for helping clues of
@emph{any} kind.
Surely, when looking at program sources, the translator will receive
more help if she is a fluent programmer. However, even if she is
not versed in programming and feels a little lost in C code, the
translator should not be shy at taking a look, once in a while.
It is most probable that she will still be able to find some of the
hints she needs. She will learn quickly to not feel uncomfortable
in program code, paying more attention to programmer's comments,
variable and function names (if he dared chosing them well), and
overall organization, than to programmation itself.
The following commands are meant to help the translator at getting
program source context for a PO file entry.
@table @kbd
@item s
Resume the display of a program source context, or cycle through them.
@item M-s
Display of a program source context selected by menu.
@item S
Add a directory to the search path for source files.
@item M-S
Delete a directory from the search path for source files.
@end table
The commands @kbd{s} (@code{po-cycle-reference}) and @kbd{M-s}
(@code{po-select-source-reference}) both open another window displaying
some source program file, and already positioned in such a way that
it shows an actual use of the string to be translated. By doing
so, the command gives source program context for the string. But if
the entry has no source context references, or if all references
are unresolved along the search path for program sources, then the
command diagnoses this as an error.
Even if @kbd{s} (or @kbd{M-s}) opens a new window, the cursor stays
in the PO file window. If the translator really wants to
get into the program source window, she ought to do it explicitly,
maybe by using command @kbd{O}.
When @kbd{s} is typed for the first time, or for a PO file entry which
is different of the last one used for getting source context, then the
command reacts by giving the first context available for this entry,
if any. If some context has already been recently displayed for the
current PO file entry, and the translator wandered off to do other
things, typing @kbd{s} again will merely resume, in another window,
the context last displayed. In particular, if the translator moved
the cursor away from the context in the source file, the command will
bring the cursor back to the context. By using @kbd{s} many times
in a row, with no other commands intervening, PO mode will cycle to
the next available contexts for this particular entry, getting back
to the first context once the last has been shown.
The command @kbd{M-s} behaves differently. Instead of cycling through
references, it lets the translator choose of particular reference among
many, and displays that reference. It is best used with completion,
if the translator types @kbd{TAB} immediately after @kbd{M-s}, in
response to the question, she will be offered a menu of all possible
references, as a reminder of which are the acceptable answers.
This command is useful only where there are really many contexts
available for a single string to translate.
Program source files are usually found relative to where the PO
file stands. As a special provision, when this fails, the file is
also looked for, but relative to the directory immediately above it.
Those two cases take proper care of most PO files. However, it might
happen that a PO file has been moved, or is edited in a different
place than its normal location. When this happens, the translator
should tell PO mode in which directory normally sits the genuine PO
file. Many such directories may be specified, and all together, they
constitute what is called the @dfn{search path} for program sources.
The command @kbd{S} (@code{po-consider-source-path}) is used to interactively
enter a new directory at the front of the search path, and the command
@kbd{M-S} (@code{po-ignore-source-path}) is used to select, with completion,
one of the directories she does not want anymore on the search path.
@node Compendium, , C Sources Context, Initial
@section Using Translation Compendiums
@c FIXME: Rewrite.
Compendiums are yet to be implemented.
An incoming PO mode feature will let the translator maintain a
compendium of already achieved translations. A @dfn{compendium}
is a special PO file containing a set of translations recurring in
many different packages. The translator will be given commands for
adding entries to her compendium, and later initializing untranslated
entries, or updating already translated entries, from translations
kept in the compendium. For this to work, however, the compendium
would have to be normalized. @xref{Normalizing}.
@c It is not useful that I modify the @file{lib/} routines if not done in
@c the true sources. How do you/I/they proceed for getting this job done?
@c I presume that @file{lib/} routines will all use @code{gettext} for
@c the time being.
@node Updating, Binaries, Initial, Top
@chapter Updating Existing PO Files
@c FIXME: Rewrite.
@menu
* msgmerge Invocation:: Invoking the @code{msgmerge} Program
* Translated Entries::
* Fuzzy Entries::
* Untranslated Entries:: Untranslated Entries
* Obsolete Entries:: Obsolete Entries
* Modifying Translations:: Modifying Translations
* Modifying Comments:: Modifying Comments
* Auxiliary:: Consulting Auxiliary PO Files
@end menu
@node msgmerge Invocation, Translated Entries, Updating, Updating
@section Invoking the @code{msgmerge} Program
@c FIXME: Rewrite.
@c @example
@c tupdate --help
@c tupdate --version
@c tupdate @var{new} @var{old}
@c @end example
@c File @var{new} is the last created PO file (generally by
@c @code{xgettext}). It need not contain any translations. File
@c @var{old} is the PO file including the old translations which will
@c be taken over to the newly created file as long as they still match.
@c When English messages change in the programs, this is reflected in
@c the PO file as extracted by @code{xgettext}. In large messages, that
@c can be hard to detect, and will obviously result in an incomplete
@c translation. One of the virtues of @code{tupdate} is that it detects
@c such changes, saving the previous translation into a PO file comment,
@c so marking the entry as obsolete, and giving the modified string with
@c an empty translation, that is, marking the entry as untranslated.
@node Translated Entries
@section Translated Entries
Each PO file entry for which the @code{msgstr} field has been filled with
a translation, and which is not marked as fuzzy (@pxref{Fuzzy Entries}),
is a said to be a @dfn{translated} entry. Only translated entries will
later be compiled by GNU @code{msgfmt} and become usable in programs.
Other entry types will be excluded; translation will not occur for them.
Some commands are more specifically related to translated entry processing.
@table @kbd
@item t
Find the next translated entry.
@item M-t
Find the previous translated entry.
@end table
The commands @kbd{t} (@code{po-next-translated-entry}) and @kbd{M-t}
(@code{po-previous-transted-entry}) move forwards or backwards, chasing
for an translated entry. If none is found, the search is extended and
wraps around in the PO file buffer.
Translated entries usually result from the translator having edited in
a translation for them, @ref{Modifying Translations}. However, if the
variable @code{po-auto-fuzzy-on-edit} is not @code{nil}, the entry having
received a new translation first becomes a fuzzy entry, which ought to
be later unfuzzied before becoming an official, genuine translated entry.
@xref{Fuzzy Entries}.
@node Fuzzy Entries, Untranslated Entries, Translated Entries, Updating
@section Fuzzy Entries
Each PO file entry may have a set of @dfn{attributes}, which are
qualities given an name and explicitely associated with the entry
translation, using a special system comment. One of these attributes
has the name @code{fuzzy}, and entries having this attribute are said
to have a fuzzy translation. They are called fuzzy entries, for short.
Fuzzy entries, even if they account for translated entries for
most other purposes, usually call for revision by the translator.
Those may be produced by applying the program @code{msgmerge} to
update an older translated PO files according to a new PO template
file, when this tool hypothesises that some new @code{msgid} has
been modified only slightly out of an older one, and chooses to pair
what it thinks to be the old translation for the new modified entry.
The slight alteration in the original string (the @code{msgid} string)
should often be reflected in the translated string, and this requires
the intervention of the translator. For this reason, @code{msgmerge}
might mark some entries as being fuzzy.
Also, the translator may decide herself to mark an entry as fuzzy
for her own convenience, when she wants to remember that the entry
has to be later revisited. So, some commands are more specifically
related to fuzzy entry processing.
@table @kbd
@item f
Find the next fuzzy entry.
@item M-f
Find the previous fuzzy entry.
@item TAB
Remove the fuzzy attribute of the current entry.
@end table
The commands @kbd{f} (@code{po-next-fuzzy}) and @kbd{M-f}
(@code{po-previous-fuzzy}) move forwards or backwards, chasing for
a fuzzy entry. If none is found, the search is extended and wraps
around in the PO file buffer.
The command @kbd{TAB} (@code{po-unfuzzy}) removes the fuzzy
attribute associated with an entry, usually leaving it translated.
Further, if the variable @code{po-auto-select-on-unfuzzy} has not
the @code{nil} value, the @kbd{TAB} command will automatically chase
for another interesting entry to work on. The initial value of
@code{po-auto-select-on-unfuzzy} is @code{nil}.
The initial value of @code{po-auto-fuzzy-on-edit} is @code{nil}. However,
if the variable @code{po-auto-fuzzy-on-edit} is set to @code{t}, any entry
edited through the @kbd{RET} command is marked fuzzy, as a way to ensure
some kind of double check, later. In this case, the usual paradigm is
that an entry becomes fuzzy (if not already) whenever the translator
modifies it. If she is satisfied with the translation, she then uses
@kbd{TAB} to pick another entry to work on, clearing the fuzzy attribute
on the same blow. If she is not satisfied yet, she merely uses @kbd{SPC}
to chase another entry, leaving the entry fuzzy.
The translator may also use the @kbd{DEL} command
(@code{po-fade-out-entry}) over any translated entry to mark it as being
fuzzy, when she wants to easily leave a trace she wants to later return
working at this entry.
Also, when time comes to quit working on a PO file buffer with the @kbd{q}
command, the translator is asked for confirmation, if fuzzy string
still exists.
@node Untranslated Entries, Obsolete Entries, Fuzzy Entries, Updating
@section Untranslated Entries
When @code{xgettext} originally creates a PO file, unless told
otherwise, it initializes the @code{msgid} field with the untranslated
string, and leaves the @code{msgstr} string to be empty. Such entries,
having an empty translation, are said to be @dfn{untranslated} entries.
Later, when the programmer slightly modifies some string right in
the program, this change is later reflected in the PO file
by the appearance of a new untranslated entry for the modified string.
The usual commands moving from entry to entry consider untranslated
entries on the same level as active entries. Untranslated entries
are easily recognizable by the fact they end with @w{@samp{msgstr ""}}.
The work of the translator might be (quite naively) seen as the process
of seeking after an untranslated entry, editing a translation for
it, and repeating these actions until no untranslated entries remain.
Some commands are more specifically related to untranslated entry
processing.
@table @kbd
@item u
Find the next untranslated entry.
@item M-u
Find the previous untranslated entry.
@item k
Turn the current entry into an untranslated one.
@end table
The commands @kbd{u} (@code{po-next-untranslated-entry}) and @kbd{M-u}
(@code{po-previous-untransted-entry}) move forwards or backwards,
chasing for an untranslated entry. If none is found, the search is
extended and wraps around in the PO file buffer.
An entry can be turned back into an untranslated entry by
merely emptying its translation, using the command @kbd{k}
(@code{po-kill-msgstr}). @xref{Modifying Translations}.
Also, when time comes to quit working on a PO file buffer
with the @kbd{q} command, the translator is asked for confirmation,
if some untranslated string still exists.
@node Obsolete Entries, Modifying Translations, Untranslated Entries, Updating
@section Obsolete Entries
By @dfn{obsolete} PO file entries, we mean those entries which are
commented out, usually by @code{msgmerge} when it found that the
translation is not needed anymore by the package being localized.
The usual commands moving from entry to entry consider obsolete
entries on the same level as active entries. Obsolete entries are
easily recognizable by the fact that all their lines start with
@kbd{#}, even those lines containing @code{msgid} or @code{msgstr}.
Commands exist for emptying the translation or reinitializing it
to the original untranslated string. Commands interfacing with the
kill ring may force some previously saved text into the translation.
The user may interactively edit the translation. All these commands
may apply to obsolete entries, carefully leaving the entry obsolete
after the fact.
Moreover, some commands are more specifically related to obsolete
entry processing.
@table @kbd
@item o
Find the next obsolete entry.
@item M-o
Find the previous obsolete entry.
@item DEL
Make an active entry obsolete, or zap out an obsolete entry.
@end table
The commands @kbd{o} (@code{po-next-obsolete-entry}) and @kbd{M-o}
(@code{po-previous-obsolete-entry}) move forwards or backwards,
chasing for an obsolete entry. If none is found, the search is
extended and wraps around in the PO file buffer.
PO mode does not provide ways for un-commenting an obsolete entry
and making it active, because this would reintroduce an original
untranslated string which does not correspond to any marked string
in the program sources. This goes with the philosophy of never
introducing useless @code{msgid} values.
However, it is possible to comment out an active entry, so making
it obsolete. GNU @code{gettext} utilities will later react to the
disappearance of a translation by using the untranslated string.
The command @kbd{DEL} (@code{po-fade-out-entry}) pushes the current entry
a little further towards annihilation. If the entry is active (it is a
translated entry), then it is first made fuzzy. If it is already fuzzy,
then the entry is merely commented out, with confirmation. If the entry
is already obsolete, then it is completely deleted from the PO file.
It is easy to recycle the translation so deleted into some other PO file
entry, usually one which is untranslated. @xref{Modifying Translations}.
Here is a quite interesting problem to solve for later development of
PO mode, for those nights you are not sleepy. The idea would be that
PO mode might become bright enough, one of these days, to make good
guesses at retrieving the most probable candidate, among all obsolete
entries, for initializing the translation of a newly appeared string.
I think it might be a quite hard problem to do this algorithmically, as
we have to develop good and efficient measures of string similarity.
Right now, PO mode completely lets the decision to the translator,
when the time comes to find the adequate obsolete translation, it
merely tries to provide handy tools for helping her to do so.
@node Modifying Translations, Modifying Comments, Obsolete Entries, Updating
@section Modifying Translations
PO mode prevents direct edition of the PO file, by the usual
means Emacs give for altering a buffer's contents. By doing so,
it pretends helping the translator to avoid little clerical errors
about the overall file format, or the proper quoting of strings,
as those errors would be easily made. Other kinds of errors are
still possible, but some may be caught and diagnosed by the batch
validation process, which the translator may always trigger by the
@kbd{V} command. For all other errors, the translator has to rely on
her own judgment, and also on the linguistic reports submitted to her
by the users of the translated package, having the same mother tongue.
When the time comes to create a translation, correct an error diagnosed
mechanically or reported by a user, the translators have to resort to
using the following commands for modifying the translations.
@table @kbd
@item RET
Interactively edit the translation.
@item LFD
Reinitialize the translation with the original, untranslated string.
@item k
Save the translation on the kill ring, and delete it.
@item w
Save the translation on the kill ring, without deleting it.
@item y
Replace the translation, taking the new from the kill ring.
@end table
The command @kbd{RET} (@code{po-edit-msgstr}) opens a new Emacs window
containing a copy of the translation taken from the current PO file entry,
all ready for edition, fully modifiable and with the complete extent of
GNU Emacs modifying commands. The string is presented to the translator
expunged of all quoting marks, and she will modify the @emph{unquoted}
string in this window to heart's content. Once done, the regular Emacs
command @kbd{M-C-c} (@code{exit-recursive-edit}) may be used to return the
edited translation into the PO file, replacing the original translation.
The keys @w{@kbd{C-c C-c}} are bound so they have the same effect as
@kbd{M-C-c}.
If the translator becomes unsatisfied with her translation to the extent
she prefers keeping the translation which was existent prior to the
@kbd{RET} command, she may use the standard Emacs command @kbd{C-]}
(@code{abort-recursive-edit}) to merely get rid of edition, while
preserving the original translation. The keys @w{@kbd{C-c C-k}} are
bound so they have the same effect as @kbd{C-]}. Another way would
be for her to exit normally with @w{@kbd{C-c C-c}}, then type @code{U}
once for undoing the whole effect of last edition.
Functions found on @code{po-subedit-mode-hook}, if any, are executed after
the string has been inserted in the edit buffer and before recursive edit
is entered.
While editing her translation, the translator should pay attention to
not inserting unwanted @kbd{@key{RET}} (carriage returns) characters at
the end of the translated string if those are not meant to be there,
or to removing such characters when they are required. Since these
characters are not visible in the editing buffer, they are easily
introduced by mistake. To help her, @kbd{@key{RET}} automatically puts
the character @kbd{<} at the end of the string being edited, but this
@kbd{<} is not really part of the string. On exiting the editing
window with @w{@kbd{C-c C-c}}, PO mode automatically removes such
@kbd{<} and all whitespace added after it. If the translator adds
characters after the terminating @kbd{<}, it looses its delimiting
property and integrally becomes part of the string. If she removes
the delimiting @kbd{<}, then the edited string is taken @emph{as
is}, with all trailing newlines, even if invisible. Also, if the
translated string ought to end itself with a genuine @kbd{<}, then the
delimiting @kbd{<} may not be removed; so the string should appear,
in the editing window, as ending with two @kbd{<} in a row.
When a translation (or a comment) is being edited, the translator
may move the cursor back into the PO file buffer and freely
move to other entries, browsing at will. The edited entry will
be recovered as soon as the edit ceases, because it is this entry
only which is being modified. If, with an edition still opened, the
translator wanders in the PO file buffer, she cannot modify
any other entry. If she tries to, PO mode will react by suggesting
that she abort the current edit, or else, by inviting her to finish
the current edit prior to any other modification.
The command @kbd{LFD} (@code{po-msgid-to-msgstr}) initializes, or
reinitializes the translation with the original string. This command
is normally used when the translator wants to redo a fresh translation
of the original string, disregarding any previous work.
It is possible to arrange so, whenever editing an untranslated
entry, the @kbd{LFD} command be automatically executed. If you set
@code{po-auto-edit-with-msgid} to @code{t}, the translation gets
initialised with the original string, in case none exist already.
The default value for @code{po-auto-edit-with-msgid} is @code{nil}.
In fact, whether it is best to start a translation with an empty
string, or rather with a copy of the original string, is a matter of
taste or habit. Sometimes, the source language and the
target language are so different that is simply best to start writing
on an empty page. At other times, the source and target languages
are so close that it would be a waste to retype a number of words
already being written in the original string. A translator may also
like having the original string right under her eyes, as she will
progressively overwrite the original text with the translation, even
if this requires some extra editing work to get rid of the original.
The command @kbd{k} (@code{po-kill-msgstr}) merely empties the
translation string, so turning the entry into an untranslated
one. But while doing so, its previous contents is put apart in
a special place, known as the kill ring. The command @kbd{w}
(@code{po-kill-ring-save-msgstr}) has also the effect of taking a
copy of the translation onto the kill ring, but it otherwise leaves
the entry alone, and does @emph{not} remove the translation from the
entry. Both commands use exactly the Emacs kill ring, which is shared
between buffers, and which is well known already to GNU Emacs lovers.
The translator may use @kbd{k} or @kbd{w} many times in the course
of her work, as the kill ring may hold several saved translations.
From the kill ring, strings may later be reinserted in various
Emacs buffers. In particular, the kill ring may be used for moving
translation strings between different entries of a single PO file
buffer, or if the translator is handling many such buffers at once,
even between PO files.
To facilitate exchanges with buffers which are not in PO mode, the
translation string put on the kill ring by the @kbd{k} command is fully
unquoted before being saved: external quotes are removed, multi-lines
strings are concatenated, and backslashed escaped sequences are turned
into their corresponding characters. In the special case of obsolete
entries, the translation is also uncommented prior to saving.
The command @kbd{y} (@code{po-yank-msgstr}) completely replaces the
translation of the current entry by a string taken from the kill ring.
Following GNU Emacs terminology, we then say that the replacement
string is @dfn{yanked} into the PO file buffer.
@xref{Yanking, , , emacs, The Emacs Editor}.
The first time @kbd{y} is used, the translation receives the value of
the most recent addition to the kill ring. If @kbd{y} is typed once
again, immediately, without intervening keystrokes, the translation
just inserted is taken away and replaced by the second most recent
addition to the kill ring. By repeating @kbd{y} many times in a row,
the translator may travel along the kill ring for saved strings,
until she finds the string she really wanted.
When a string is yanked into a PO file entry, it is fully and
automatically requoted for complying with the format PO files should
have. Further, if the entry is obsolete, PO mode then appropriately
push the inserted string inside comments. Once again, translators
should not burden themselves with quoting considerations besides, of
course, the necessity of the translated string itself respective to
the program using it.
Note that @kbd{k} or @kbd{w} are not the only commands pushing strings
on the kill ring, as almost any PO mode command replacing translation
strings (or the translator comments) automatically save the old string
on the kill ring. The main exceptions to this general rule are the
yanking commands themselves.
To better illustrate the operation of killing and yanking, let's
use an actual example, taken from a common situation. When the
programmer slightly modifies some string right in the program, his
change is later reflected in the PO file by the appearance
of a new untranslated entry for the modified string, and the fact
that the entry translating the original or unmodified string becomes
obsolete. In many cases, the translator might spare herself some work
by retrieving the unmodified translation from the obsolete entry,
then initializing the untranslated entry @code{msgstr} field with
this retrieved translation. Once this done, the obsolete entry is
not wanted anymore, and may be safely deleted.
When the translator finds an untranslated entry and suspects that a
slight variant of the translation exists, she immediately uses @kbd{m}
to mark the current entry location, then starts chasing obsolete
entries with @kbd{o}, hoping to find some translation corresponding
to the unmodified string. Once found, she uses the @kbd{DEL} command
for deleting the obsolete entry, knowing that @kbd{DEL} also @emph{kills}
the translation, that is, pushes the translation on the kill ring.
Then, @kbd{r} returns to the initial untranslated entry, @kbd{y}
then @emph{yanks} the saved translation right into the @code{msgstr}
field. The translator is then free to use @kbd{@key{RET}} for fine
tuning the translation contents, and maybe to later use @kbd{u},
then @kbd{m} again, for going on with the next untranslated string.
When some sequence of keys has to be typed over and over again, the
translator may find it useful to become better acquainted with the GNU
Emacs capability of learning these sequences and playing them back under
request. @xref{Keyboard Macros, , , emacs, The Emacs Editor}.
@node Modifying Comments, Auxiliary, Modifying Translations, Updating
@section Modifying Comments
Any translation work done seriously will raise many linguistic
difficulties, for which decisions have to be made, and the choices
further documented. These documents may be saved within the
PO file in form of translator comments, which the translator
is free to create, delete, or modify at will. These comments may
be useful to herself when she returns to this PO file after a while.
Comments not having whitespace after the initial @samp{#}, for example,
those beginning with @samp{#.} or @samp{#:}, are @emph{not} translator
comments, they are exclusively created by other @code{gettext} tools.
So, the commands below will never alter such system added comments,
they are not meant for the translator to modify. @xref{PO Files}.
The following commands are somewhat similar to those modifying translations,
so the general indications given for those apply here. @xref{Modifying
Translations}.
@table @kbd
@item #
Interactively edit the translator comments.
@item K
Save the translator comments on the kill ring, and delete it.
@item W
Save the translator comments on the kill ring, without deleting it.
@item Y
Replace the translator comments, taking the new from the kill ring.
@end table
These commands parallel PO mode commands for modifying the translation
strings, and behave much the same way as they do, except that they handle
this part of PO file comments meant for translator usage, rather
than the translation strings. So, if the descriptions given below are
slightly succinct, it is because the full details have already been given.
@xref{Modifying Translations}.
The command @kbd{#} (@code{po-edit-comment}) opens a new Emacs
window containing a copy of the translator comments on the current
PO file entry. If there are no such comments, PO mode
understands that the translator wants to add a comment to the entry,
and she is presented with an empty screen. Comment marks (@kbd{#}) and
the space following them are automatically removed before edition,
and reinstated after. For translator comments pertaining to obsolete
entries, the uncommenting and recommenting operations are done twice.
Once in the editing window, the keys @w{@kbd{C-c C-c}} allow the
translator to tell she is finished with editing the comment.
Functions found on @code{po-subedit-mode-hook}, if any, are executed after
the string has been inserted in the edit buffer and before recursive edit
is entered.
The command @kbd{K} (@code{po-kill-comment}) get rid of all
translator comments, while saving those comments on the kill ring.
The command @kbd{W} (@code{po-kill-ring-save-comment}) takes
a copy of the translator comments on the kill ring, but leaves
them undisturbed in the current entry. The command @kbd{Y}
(@code{po-yank-comment}) completely replaces the translator comments
by a string taken at the front of the kill ring. When this command
is immediately repeated, the comments just inserted are withdrawn,
and replaced by other strings taken along the kill ring.
On the kill ring, all strings have the same nature. There is no
distinction between @emph{translation} strings and @emph{translator
comments} strings. So, for example, let's presume the translator
has just finished editing a translation, and wants to create a new
translator comment to document why the previous translation was
not good, just to remember what was the problem. Foreseeing that she
will do that in her documentation, the translator may want to quote
the previous translation in her translator comments. To do so, she
may initialize the translator comments with the previous translation,
still at the head of the kill ring. Because editing already pushed the
previous translation on the kill ring, she merely has to type @kbd{M-w}
prior to @kbd{#}, and the previous translation will be right there,
all ready for being introduced by some explanatory text.
On the other hand, presume there are some translator comments already
and that the translator wants to add to those comments, instead
of wholly replacing them. Then, she should edit the comment right
away with @kbd{#}. Once inside the editing window, she can use the
regular GNU Emacs commands @kbd{C-y} (@code{yank}) and @kbd{M-y}
(@code{yank-pop}) to get the previous translation where she likes.
@node Auxiliary, , Modifying Comments, Updating
@section Consulting Auxiliary PO Files
PO mode is able to help the knowledgeable translator, being fluent in
many languages, at taking advantage of translations already achieved
in other languages she just happens to know. It provides these other
language translations as additional context for her own work. Moreover,
it has features to ease the production of translations for many languages
at once, for translators preferring to work in this way.
An @dfn{auxiliary} PO file is an existing PO file meant for the same
package the translator is working on, but targeted to a different mother
tongue language. Commands exist for declaring and handling auxiliary
PO files, and also for showing contexts for the entry under work.
Here are the auxiliary file commands available in PO mode.
@table @kbd
@item a
Seek auxiliary files for another translation for the same entry.
@item M-a
Switch to a particular auxiliary file.
@item A
Declare this PO file as an auxiliary file.
@item M-A
Remove this PO file from the list of auxiliary files.
@end table
Command @kbd{A} (@code{po-consider-as-auxiliary}) adds the current
PO file to the list of auxiliary files, while command @kbd{M-A}
(@code{po-ignore-as-auxiliary} just removes it.
The command @kbd{a} (@code{po-cycle-auxiliary}) seeks all auxiliary PO
files, round-robin, searching for a translated entry in some other language
having an @code{msgid} field identical as the one for the current entry.
The found PO file, if any, takes the place of the current PO file in
the display (its window gets on top). Before doing so, the current PO
file is also made into an auxiliary file, if not already. So, @kbd{a}
in this newly displayed PO file will seek another PO file, and so on,
so repeating @kbd{a} will eventually yield back the original PO file.
The command @kbd{M-a} (@code{po-select-auxiliary}) asks the translator
for her choice of a particular auxiliary file, with completion, and
then switches to that selected PO file. The command also checks if
the selected file has an @code{msgid} field identical as the one for
the current entry, and if yes, this entry becomes current. Otherwise,
the cursor of the selected file is left undisturbed.
For all this to work fully, auxiliary PO files will have to be normalized,
in that way that @code{msgid} fields should be written @emph{exactly}
the same way. It is possible to write @code{msgid} fields in various
ways for representing the same string, different writing would break the
proper behaviour of the auxiliary file commands of PO mode. This is not
expected to be much a problem in practice, as most existing PO files have
their @code{msgid} entries written by the same GNU @code{gettext} tools.
However, PO files initially created by PO mode itself, while marking
strings in source files, are normalised differently. So are PO
files resulting of the the @samp{M-x normalize} command. Until these
discrepancies between PO mode and other GNU @code{gettext} tools get
fully resolved, the translator should stay aware of normalisation issues.
@node Binaries, Users, Updating, Top
@chapter Producing Binary MO Files
@c FIXME: Rewrite.
@menu
* msgfmt Invocation:: Invoking the @code{msgfmt} Program
* MO Files:: The Format of GNU MO Files
@end menu
@node msgfmt Invocation, MO Files, Binaries, Binaries
@section Invoking the @code{msgfmt} Program
@c FIXME: Rewrite.
@example
Usage: msgfmt [@var{option}] @var{filename}.po @dots{}
@end example
@table @samp
@item -a @var{number}
@itemx --alignment=@var{number}
Align strings to @var{number} bytes (default: 1).
@c Currently the README mentions that this constant could be changed by
@c the installer by changing the value in config.h. Should this go away?
@item -h
@itemx --help
Display this help and exit.
@item --no-hash
Binary file will not include the hash table.
@item -o @var{file}
@itemx --output-file=@var{file}
Specify output file name as @var{file}.
@itemx --strict
Direct the program to work strictly following the Uniforum/Sun
implementation. Currently this only affects the naming of the output
file. If this option is not given the name of the output file is the
same as the domain name. If the strict Uniforum mode is enable the
suffix @file{.mo} is added to the file name if it is not already
present.
We find this behaviour of Sun's implementation rather silly and so by
default this mode is @emph{not} selected.
@item -v
@itemx --verbose
Detect and diagnose input file anomalies which might represent
translation errors. The @code{msgid} and @code{msgstr} strings are
studied and compared. It is considered abnormal that one string
starts or ends with a newline while the other does not.
Also, if the string represents a format sring used in a
@code{printf}-like function both strings should have the same number of
@samp{%} format specifiers, with matching types. If the flag
@code{c-format} or @code{possible-c-format} appears in the special
comment @key{#,} for this entry a check is performed. For example, the
check will diagnose using @samp{%.*s} against @samp{%s}, or @samp{%d}
against @samp{%s}, or @samp{%d} against @samp{%x}. It can even handle
positional parameters.
Normally the @code{xgettext} program automatically decides whether a
string is a format string or not. This algorithm is not perfect,
though. It might regard a string as a format string though it is not
used in a @code{printf}-like function and so @code{msgfmt} might report
errors where there are none. Or the other way round: a string is not
regarded as a format string but it is used in a @code{printf}-like
function.
So solve this problem the programmer can dictate the decision to the
@code{xgettext} program (@pxref{c-format}). The translator should not
consider removing the flag from the @key{#,} line. This "fix" would be
reversed again as soon as @code{msgmerge} is called the next time.
@item -V
@itemx --version
Output version information and exit.
@end table
If input file is @samp{-}, standard input is read. If output file
is @samp{-}, output is written to standard output.
@node MO Files, , msgfmt Invocation, Binaries
@section The Format of GNU MO Files
The format of the generated MO files is best described by a picture,
which appears below.
The first two words serve the identification of the file. The magic
number will always signal GNU MO files. The number is stored in the
byte order of the generating machine, so the magic number really is
two numbers: @code{0x950412de} and @code{0xde120495}. The second
word describes the current revision of the file format. For now the
revision is 0. This might change in future versions, and ensures
that the readers of MO files can distinguish new formats from old
ones, so that both can be handled correctly. The version is kept
separate from the magic number, instead of using different magic
numbers for different formats, mainly because @file{/etc/magic} is
not updated often. It might be better to have magic separated from
internal format version identification.
Follow a number of pointers to later tables in the file, allowing
for the extension of the prefix part of MO files without having to
recompile programs reading them. This might become useful for later
inserting a few flag bits, indication about the charset used, new
tables, or other things.
Then, at offset @var{O} and offset @var{T} in the picture, two tables
of string descriptors can be found. In both tables, each string
descriptor uses two 32 bits integers, one for the string length,
another for the offset of the string in the MO file, counting in bytes
from the start of the file. The first table contains descriptors
for the original strings, and is sorted so the original strings
are in increasing lexicographical order. The second table contains
descriptors for the translated strings, and is parallel to the first
table: to find the corresponding translation one has to access the
array slot in the second array with the same index.
Having the original strings sorted enables the use of simple binary
search, for when the MO file does not contain an hashing table, or
for when it is not practical to use the hashing table provided in
the MO file. This also has another advantage, as the empty string
in a PO file GNU @code{gettext} is usually @emph{translated} into
some system information attached to that particular MO file, and the
empty string necessarily becomes the first in both the original and
translated tables, making the system information very easy to find.
The size @var{S} of the hash table can be zero. In this case, the
hash table itself is not contained in the MO file. Some people might
prefer this because a precomputed hashing table takes disk space, and
does not win @emph{that} much speed. The hash table contains indices
to the sorted array of strings in the MO file. Conflict resolution is
done by double hashing. The precise hashing algorithm used is fairly
dependent of GNU @code{gettext} code, and is not documented here.
As for the strings themselves, they follow the hash file, and each
is terminated with a @key{NUL}, and this @key{NUL} is not counted in
the length which appears in the string descriptor. The @code{msgfmt}
program has an option selecting the alignment for MO file strings.
With this option, each string is separately aligned so it starts at
an offset which is a multiple of the alignment value. On some RISC
machines, a correct alignment will speed things up.
Nothing prevents a MO file from having embedded @key{NUL}s in strings.
However, the program interface currently used already presumes
that strings are @key{NUL} terminated, so embedded @key{NUL}s are
somewhat useless. But MO file format is general enough so other
interfaces would be later possible, if for example, we ever want to
implement wide characters right in MO files, where @key{NUL} bytes may
accidently appear.
This particular issue has been strongly debated in the GNU
@code{gettext} development forum, and it is expectable that MO file
format will evolve or change over time. It is even possible that many
formats may later be supported concurrently. But surely, we have to
start somewhere, and the MO file format described here is a good start.
Nothing is cast in concrete, and the format may later evolve fairly
easily, so we should feel comfortable with the current approach.
@example
@group
byte
+------------------------------------------+
0 | magic number = 0x950412de |
| |
4 | file format revision = 0 |
| |
8 | number of strings | == N
| |
12 | offset of table with original strings | == O
| |
16 | offset of table with translation strings | == T
| |
20 | size of hashing table | == S
| |
24 | offset of hashing table | == H
| |
. .
. (possibly more entries later) .
. .
| |
O | length & offset 0th string ----------------.
O + 8 | length & offset 1st string ------------------.
... ... | |
O + ((N-1)*8)| length & offset (N-1)th string | | |
| | | |
T | length & offset 0th translation ---------------.
T + 8 | length & offset 1st translation -----------------.
... ... | | | |
T + ((N-1)*8)| length & offset (N-1)th translation | | | | |
| | | | | |
H | start hash table | | | | |
... ... | | | |
H + S * 4 | end hash table | | | | |
| | | | | |
| NUL terminated 0th string <----------------' | | |
| | | | |
| NUL terminated 1st string <------------------' | |
| | | |
... ... | |
| | | |
| NUL terminated 0th translation <---------------' |
| | |
| NUL terminated 1st translation <-----------------'
| |
... ...
| |
+------------------------------------------+
@end group
@end example
@node Users, Programmers, Binaries, Top
@chapter The User's View
When GNU @code{gettext} will truly have reached is goal, average users
should feel some kind of astonished pleasure, seeing the effect of
that strange kind of magic that just makes their own native language
appear everywhere on their screens. As for naive users, they would
ideally have no special pleasure about it, merely taking their own
language for @emph{granted}, and becoming rather unhappy otherwise.
So, let's try to describe here how we would like the magic to operate,
as we want the users' view to be the simplest, among all ways one
could look at GNU @code{gettext}. All other software engineers:
programmers, translators, maintainers, should work together in such a
way that the magic becomes possible. This is a long and progressive
undertaking, and information is available about the progress of the
Translation Project.
When a package is distributed, there are two kind of users:
@dfn{installers} who fetch the distribution, unpack it, configure
it, compile it and install it for themselves or others to use; and
@dfn{end users} that call programs of the package, once these have
been installed at their site. GNU @code{gettext} is offering magic
for both installers and end users.
@menu
* Matrix:: The Current @file{ABOUT-NLS} Matrix
* Installers:: Magic for Installers
* End Users:: Magic for End Users
@end menu
@node Matrix, Installers, Users, Users
@section The Current @file{ABOUT-NLS} Matrix
Languages are not equally supported in all packages using GNU
@code{gettext}. To know if some package uses GNU @code{gettext}, one
may check the distribution for the @file{ABOUT-NLS} information file, for
some @file{@var{ll}.po} files, often kept together into some @file{po/}
directory, or for an @file{intl/} directory. Internationalized packages
have usually many @file{@var{ll}.po} files, where @var{ll} represents
the language. @ref{End Users} for a complete description of the format
for @var{ll}.
More generally, a matrix is available for showing the current state
of the Translation Project, listing which packages are prepared for
multi-lingual messages, and which languages is supported by each.
Because this information changes often, this matrix is not kept within
this GNU @code{gettext} manual. This information is often found in
file @file{ABOUT-NLS} from various distributions, but is also as old as
the distribution itself. A recent copy of this @file{ABOUT-NLS} file,
containing up-to-date information, should generally be found on the
Translation Project sites, and also on most GNU archive sites.
@node Installers, End Users, Matrix, Users
@section Magic for Installers
By default, packages fully using GNU @code{gettext}, internally,
are installed in such a way that they to allow translation of
messages. At @emph{configuration} time, those packages should
automatically detect whether the underlying host system provides usable
@code{catgets} or @code{gettext} functions. If neither is present,
the GNU @code{gettext} library should be automatically prepared
and used. Installers may use special options at configuration
time for changing this behavior. The command @samp{./configure
--with-included-gettext} bypasses system @code{catgets} or @code{gettext} to
use GNU @code{gettext} instead, while @samp{./configure --disable-nls}
produces program totally unable to translate messages.
Internationalized packages have usually many @file{@var{ll}.po}
files. Unless
translations are disabled, all those available are installed together
with the package. However, the environment variable @code{LINGUAS}
may be set, prior to configuration, to limit the installed set.
@code{LINGUAS} should then contain a space separated list of two-letter
codes, stating which languages are allowed.
@node End Users, , Installers, Users
@section Magic for End Users
We consider here those packages using GNU @code{gettext} internally,
and for which the installers did not disable translation at
@emph{configure} time. Then, users only have to set the @code{LANG}
environment variable to the appropriate @samp{@var{ll}} prior to
using the programs in the package. @xref{Matrix}. For example,
let's presume a German site. At the shell prompt, users merely have to
execute @w{@samp{setenv LANG de}} (in @code{csh}) or @w{@samp{export
LANG; LANG=de}} (in @code{sh}). They could even do this from their
@file{.login} or @file{.profile} file.
@node Programmers, Translators, Users, Top
@chapter The Programmer's View
@c FIXME: Reorganize whole chapter.
One aim of the current message catalog implementation provided by
GNU @code{gettext} was to use the systems message catalog handling, if the
installer wishes to do so. So we perhaps should first take a look at
the solutions we know about. The people in the POSIX committee does not
manage to agree on one of the semi-official standards which we'll
describe below. In fact they couldn't agree on anything, so nothing
decide only to include an example of an interface. The major Unix vendors
are split in the usage of the two most important specifications: X/Opens
catgets vs. Uniforums gettext interface. We'll describe them both and
later explain our solution of this dilemma.
@menu
* catgets:: About @code{catgets}
* gettext:: About @code{gettext}
* Comparison:: Comparing the two interfaces
* Using libintl.a:: Using libintl.a in own programs
* gettext grok:: Being a @code{gettext} grok
* Temp Programmers:: Temporary Notes for the Programmers Chapter
@end menu
@node catgets, gettext, Programmers, Programmers
@section About @code{catgets}
The @code{catgets} implementation is defined in the X/Open Portability
Guide, Volume 3, XSI Supplementary Definitions, Chapter 5. But the
process of creating this standard seemed to be too slow for some of
the Unix vendors so they created their implementations on preliminary
versions of the standard. Of course this leads again to problems while
writing platform independent programs: even the usage of @code{catgets}
does not guarantee a unique interface.
Another, personal comment on this that only a bunch of committee members
could have made this interface. They never really tried to program
using this interface. It is a fast, memory-saving implementation, an
user can happily live with it. But programmers hate it (at least me and
some others do@dots{})
But we must not forget one point: after all the trouble with transfering
the rights on Unix(tm) they at last came to X/Open, the very same who
published this specifications. This leads me to making the prediction
that this interface will be in future Unix standards (e.g. Spec1170) and
therefore part of all Unix implementation (implementations, which are
@emph{allowed} to wear this name).
@menu
* Interface to catgets:: The interface
* Problems with catgets:: Problems with the @code{catgets} interface?!
@end menu
@node Interface to catgets, Problems with catgets, catgets, catgets
@subsection The Interface
The interface to the @code{catgets} implementation consists of three
functions which correspond to those used in file access: @code{catopen}
to open the catalog for using, @code{catgets} for accessing the message
tables, and @code{catclose} for closing after work is done. Prototypes
for the functions and the needed definitions are in the
@code{<nl_types.h>} header file.
@code{catopen} is used like in this:
@example
nl_catd catd = catopen ("catalog_name", 0);
@end example
The function takes as the argument the name of the catalog. This usual
refers to the name of the program or the package. The second parameter
is not further specified in the standard. I don't even know whether it
is implemented consistently among various systems. So the common advice
is to use @code{0} as the value. The return value is a handle to the
message catalog, equivalent to handles to file returned by @code{open}.
This handle is of course used in the @code{catgets} function which can
be used like this:
@example
char *translation = catgets (catd, set_no, msg_id, "original string");
@end example
The first parameter is this catalog descriptor. The second parameter
specifies the set of messages in this catalog, in which the message
described by @code{msg_id} is obtained. @code{catgets} therefore uses a
three-stage addressing:
@display
catalog name @result{} set number @result{} message ID @result{} translation
@end display
@c Anybody else loving Haskell??? :-) -- Uli
The fourth argument is not used to address the translation. It is given
as a default value in case when one of the addressing stages fail. One
important thing to remember is that although the return type of catgets
is @code{char *} the resulting string @emph{must not} be changed. It
should better @code{const char *}, but the standard is published in
1988, one year before ANSI C.
@noindent
The last of these function functions is used and behaves as expected:
@example
catclose (catd);
@end example
After this no @code{catgets} call using the descriptor is legal anymore.
@node Problems with catgets, , Interface to catgets, catgets
@subsection Problems with the @code{catgets} Interface?!
Now that this descriptions seemed to be really easy where are the
problem we speak of. In fact the interface could be used in a
reasonable way, but constructing the message catalogs is a pain. The
reason for this lies in the third argument of @code{catgets}: the unique
message ID. This has to be a numeric value for all messages in a single
set. Perhaps you could imagine the problems keeping such list while
changing the source code. Add a new message here, remove one there. Of
course there have been developed a lot of tools helping to organize this
chaos but one as the other fails in one aspect or the other. We don't
want to say that the other approach has no problems but they are far
more easily to manage.
@node gettext, Comparison, catgets, Programmers
@section About @code{gettext}
The definition of the @code{gettext} interface comes from a Uniforum
proposal and it is followed by at least one major Unix vendor
(Sun) in its last developments. It is not specified in any official
standard, though.
The main points about this solution is that it does not follow the
method of normal file handling (open-use-close) and that it does not
burden the programmer so many task, especially the unique key handling.
Of course here is also a unique key needed, but this key is the
message itself (how long or short it is). @xref{Comparison} for a
more detailed comparison of the two methods.
The following section contains a rather detailed description of the
interface. We make it that detailed because this is the interface
we chose for the GNU @code{gettext} Library. Programmers interested
in using this library will be interested in this description.
@menu
* Interface to gettext:: The interface
* Ambiguities:: Solving ambiguities
* Locating Catalogs:: Locating message catalog files
* Optimized gettext:: Optimization of the *gettext functions
@end menu
@node Interface to gettext, Ambiguities, gettext, gettext
@subsection The Interface
The minimal functionality an interface must have is a) to select a
domain the strings are coming from (a single domain for all programs is
not reasonable because its construction and maintenance is difficult,
perhaps impossible) and b) to access a string in a selected domain.
This is principally the description of the @code{gettext} interface. It
has an global domain which unqualified usages reference. Of course this
domain is selectable by the user.
@example
char *textdomain (const char *domain_name);
@end example
This provides the possibility to change or query the current status of
the current global domain of the @code{LC_MESSAGE} category. The
argument is a null-terminated string, whose characters must be legal in
the use in filenames. If the @var{domain_name} argument is @code{NULL},
the function return the current value. If no value has been set
before, the name of the default domain is returned: @emph{messages}.
Please note that although the return value of @code{textdomain} is of
type @code{char *} no changing is allowed. It is also important to know
that no checks of the availability are made. If the name is not
available you will see this by the fact that no translations are provided.
@noindent
To use a domain set by @code{textdomain} the function
@example
char *gettext (const char *msgid);
@end example
is to be used. This is the simplest reasonable form one can imagine.
The translation of the string @var{msgid} is returned if it is available
in the current domain. If not available the argument itself is
returned. If the argument is @code{NULL} the result is undefined.
One things which should come into mind is that no explicit dependency to
the used domain is given. The current value of the domain for the
@code{LC_MESSAGES} locale is used. If this changes between two
executions of the same @code{gettext} call in the program, both calls
reference a different message catalog.
For the easiest case, which is normally used in internationalized
packages, once at the beginning of execution a call to @code{textdomain}
is issued, setting the domain to a unique name, normally the package
name. In the following code all strings which have to be translated are
filtered through the gettext function. That's all, the package speaks
your language.
@node Ambiguities, Locating Catalogs, Interface to gettext, gettext
@subsection Solving Ambiguities
While this single name domain work good for most applications there
might be the need to get translations from more than one domain. Of
course one could switch between different domains with calls to
@code{textdomain}, but this is really not convenient nor is it fast. A
possible situation could be one case discussing while this writing: all
error messages of functions in the set of common used functions should
go into a separate domain @code{error}. By this mean we would only need
to translate them once.
@noindent
For this reasons there are two more functions to retrieve strings:
@example
char *dgettext (const char *domain_name, const char *msgid);
char *dcgettext (const char *domain_name, const char *msgid,
int category);
@end example
Both take an additional argument at the first place, which corresponds
to the argument of @code{textdomain}. The third argument of
@code{dcgettext} allows to use another locale but @code{LC_MESSAGES}.
But I really don't know where this can be useful. If the
@var{domain_name} is @code{NULL} or @var{category} has an value beside
the known ones, the result is undefined. It should also be noted that
this function is not part of the second known implementation of this
function family, the one found in Solaris.
A second ambiguity can arise by the fact, that perhaps more than one
domain has the same name. This can be solved by specifying where the
needed message catalog files can be found.
@example
char *bindtextdomain (const char *domain_name,
const char *dir_name);
@end example
Calling this function binds the given domain to a file in the specified
directory (how this file is determined follows below). Especially a
file in the systems default place is not favored against the specified
file anymore (as it would be by solely using @code{textdomain}). A
@code{NULL} pointer for the @var{dir_name} parameter returns the binding
associated with @var{domain_name}. If @var{domain_name} itself is
@code{NULL} nothing happens and a @code{NULL} pointer is returned. Here
again as for all the other functions is true that none of the return
value must be changed!
It is important to remember that relative path names for the
@var{dir_name} parameter can be trouble. Since the path is always
computed relative to the current directory different results will be
achieved when the program executes a @code{chdir} command. Relative
paths should always be avoided to avoid dependencies and
unreliabilities.
@node Locating Catalogs, Optimized gettext, Ambiguities, gettext
@subsection Locating Message Catalog Files
Because many different languages for many different packages have to be
stored we need some way to add these information to file message catalog
files. The way usually used in Unix environments is have this encoding
in the file name. This is also done here. The directory name given in
@code{bindtextdomain}s second argument (or the default directory),
followed by the value and name of the locale and the domain name are
concatenated:
@example
@var{dir_name}/@var{locale}/LC_@var{category}/@var{domain_name}.mo
@end example
The default value for @var{dir_name} is system specific. For the GNU
library, and for packages adhering to its conventions, it's:
@example
/usr/local/share/locale
@end example
@noindent
@var{locale} is the value of the locale whose name is this
@code{LC_@var{category}}. For @code{gettext} and @code{dgettext} this
locale is always @code{LC_MESSAGES}. @code{dcgettext} specifies the
locale by the third argument.@footnote{Some
system, eg Ultrix, don't have @code{LC_MESSAGES}. Here we use a more or
less arbitrary value for it.} @footnote{When the system does not support
@code{setlocale} its behavior in setting the locale values is simulated
by looking at the environment variables.}
@node Optimized gettext, , Locating Catalogs, gettext
@subsection Optimization of the *gettext functions
At this point of the discussion we should talk about an advantage of the
GNU @code{gettext} implementation. Some readers might have pointed out
that an internationalized program might have a poor performance if some
string has to be translated in an inner loop. While this is unavoidable
when the string varies from one run of the loop to the other it is
simply a waste of time when the string is always the same. Take the
following example:
@example
@group
@{
while (@dots{})
@{
puts (gettext ("Hello world"));
@}
@}
@end group
@end example
@noindent
When the locale selection does not change between two runs the resulting
string is always the same. One way to use this is:
@example
@group
@{
str = gettext ("Hello world");
while (@dots{})
@{
puts (str);
@}
@}
@end group
@end example
@noindent
But this solution is not usable in all situation (e.g. when the locale
selection changes) nor is it good readable.
The GNU C compiler, version 2.7 and above, provide another solution for
this. To describe this we show here some lines of the
@file{intl/libgettext.h} file. For an explanation of the expression
command block see @ref{Statement Exprs, , Statements and Declarations in
Expressions, gcc, The GNU CC Manual}.
@example
@group
# if defined __GNUC__ && __GNUC__ == 2 && __GNUC_MINOR__ >= 7
extern int _nl_msg_cat_cntr;
# define dcgettext(domainname, msgid, category) \
(__extension__ \
(@{ \
char *result; \
if (__builtin_constant_p (msgid)) \
@{ \
static char *__translation__; \
static int __catalog_counter__; \
if (! __translation__ \
|| __catalog_counter__ != _nl_msg_cat_cntr) \
@{ \
__translation__ = \
dcgettext__ ((domainname), (msgid), (category)); \
__catalog_counter__ = _nl_msg_cat_cntr; \
@} \
result = __translation__; \
@} \
else \
result = dcgettext__ ((domainname), (msgid), (category)); \
result; \
@}))
# endif
@end group
@end example
The interesting thing here is the @code{__builtin_constant_p} predicate.
This is evaluated at compile time and so optimization can take place
immediately. Here two cases are distinguished: the argument to
@code{gettext} is not a constant value in which case simply the function
@code{dcgettext__} is called, the real implementation of the
@code{dcgettext} function.
If the string argument @emph{is} constant we can reuse the once gained
translation when the locale selection has not changed. This is exactly
what is done here. The @code{_nl_msg_cat_cntr} variable is defined in
the @file{loadmsgcat.c} which is available in @file{libintl.a} and is
changed whenever a new message catalog is loaded.
@node Comparison, Using libintl.a, gettext, Programmers
@section Comparing the Two Interfaces
@c FIXME: arguments to catgets vs. gettext
@c Partly done 950718 -- drepper
The following discussion is perhaps a little bit colored. As said
above we implemented GNU @code{gettext} following the Uniforum
proposal and this surely has its reasons. But it should show how we
came to this decision.
First we take a look at the developing process. When we write an
application using NLS provided by @code{gettext} we proceed as always.
Only when we come to a string which might be seen by the users and thus
has to be translated we use @code{gettext("@dots{}")} instead of
@code{"@dots{}"}. At the beginning of each source file (or in a central
header file) we define
@example
#define gettext(String) (String)
@end example
Even this definition can be avoided when the system supports the
@code{gettext} function in its C library. When we compile this code the
result is the same as if no NLS code is used. When you take a look at
the GNU @code{gettext} code you will see that we use @code{_("@dots{}")}
instead of @code{gettext("@dots{}")}. This reduces the number of
additional characters per translatable string to @emph{3} (in words:
three).
When now a production version of the program is needed we simply replace
the definition
@example
#define _(String) (String)
@end example
@noindent
by
@example
#include <libintl.h>
#define _(String) gettext (String)
@end example
@noindent
Additionally we run the program @file{xgettext} on all source code file
which contain translatable strings and that's it: we have a running
program which does not depend on translations to be available, but which
can use any that becomes available.
The same procedure can be done for the @code{gettext_noop} invocations
(@pxref{Special cases}). First you can define @code{gettext_noop} to a
no-op macro and later use the definition from @file{libintl.h}. Because
this name is not used in Suns implementation of @file{libintl.h},
you should consider the following code for your project:
@example
#ifdef gettext_noop
# define N_(String) gettext_noop (String)
#else
# define N_(String) (String)
#endif
@end example
@code{N_} is a short form similar to @code{_}. The @file{Makefile} in
the @file{po/} directory of GNU gettext knows by default both of the
mentioned short forms so you are invited to follow this proposal for
your own ease.
Now to @code{catgets}. The main problem is the work for the
programmer. Every time he comes to a translatable string he has to
define a number (or a symbolic constant) which has also be defined in
the message catalog file. He also has to take care for duplicate
entries, duplicate message IDs etc. If he wants to have the same
quality in the message catalog as the GNU @code{gettext} program
provides he also has to put the descriptive comments for the strings and
the location in all source code files in the message catalog. This is
nearly a Mission: Impossible.
But there are also some points people might call advantages speaking for
@code{catgets}. If you have a single word in a string and this string
is used in different contexts it is likely that in one or the other
language the word has different translations. Example:
@example
printf ("%s: %d", gettext ("number"), number_of_errors)
printf ("you should see %d %s", number_count,
number_count == 1 ? gettext ("number") : gettext ("numbers"))
@end example
Here we have to translate two times the string @code{"number"}. Even
if you do not speak a language beside English it might be possible to
recognize that the two words have a different meaning. In German the
first appearance has to be translated to @code{"Anzahl"} and the second
to @code{"Zahl"}.
Now you can say that this example is really esoteric. And you are
right! This is exactly how we felt about this problem and decide that
it does not weight that much. The solution for the above problem could
be very easy:
@example
printf ("%s %d", gettext ("number:"), number_of_errors)
printf (number_count == 1 ? gettext ("you should see %d number")
: gettext ("you should see %d numbers"),
number_count)
@end example
We believe that we can solve all conflicts with this method. If it is
difficult one can also consider changing one of the conflicting string a
little bit. But it is not impossible to overcome.
@c Should this be here?
Translator note: It is perhaps appropriate here to tell those English
speaking programmers that the plural form of a noun cannot be formed by
appending a single `s'. Most other languages use different methods.
Even the above form is not general enough to cope with all languages.
Rafal Maszkowski <rzm@@mat.uni.torun.pl> reports:
@quotation
In Polish we use e.g. plik (file) this way:
@example
1 plik
2,3,4 pliki
5-21 pliko'w
22-24 pliki
25-31 pliko'w
@end example
and so on (o' means 8859-2 oacute which should be rather okreska,
similar to aogonek).
@end quotation
A workable approach might be to consider methods like the one used for
@code{LC_TIME} in the POSIX.2 standard. The value of the
@code{alt_digits} field can be up to 100 strings which represent the
numbers 1 to 100. Using this in a situation of an internationalized
program means that an array of translatable strings should be indexed by
the number which should represent. A small example:
@example
@group
void
print_month_info (int month)
@{
const char *month_pos[12] =
@{ N_("first"), N_("second"), N_("third"), N_("fourth"),
N_("fifth"), N_("sixth"), N_("seventh"), N_("eighth"),
N_("ninth"), N_("tenth"), N_("eleventh"), N_("twelfth") @};
printf (_("%s is the %s month\n"), nl_langinfo (MON_1 + month),
_(month_pos[month]));
@}
@end group
@end example
@noindent
It should be obvious that this method is only reasonable for small
ranges of numbers.
@c catgets allows same original entry to have different translations
@node Using libintl.a, gettext grok, Comparison, Programmers
@section Using libintl.a in own programs
Starting with version 0.9.4 the library @code{libintl.h} should be
self-contained. I.e., you can use it in your own programs without
providing additional functions. The @file{Makefile} will put the header
and the library in directories selected using the @code{$(prefix)}.
One exception of the above is found on HP-UX systems. Here the C library
does not contain the @code{alloca} function (and the HP compiler does
not generate it inlined). But it is not intended to rewrite the whole
library just because of this dumb system. Instead include the
@code{alloca} function in all package you use the @code{libintl.a} in.
@node gettext grok, Temp Programmers, Using libintl.a, Programmers
@section Being a @code{gettext} grok
To fully exploit the functionality of the GNU @code{gettext} library it
is surely helpful to read the source code. But for those who don't want
to spend that much time in reading the (sometimes complicated) code here
is a list comments:
@itemize @bullet
@item Changing the language at runtime
For interactive programs it might be useful to offer a selection of the
used language at runtime. To understand how to do this one need to know
how the used language is determined while executing the @code{gettext}
function. The method which is presented here only works correctly
with the GNU implementation of the @code{gettext} functions. It is not
possible with underlying @code{catgets} functions or @code{gettext}
functions from the systems C library. The exception is of course the
GNU C Library which uses the GNU @code{gettext} Library for message handling.
In the function @code{dcgettext} at every call the current setting of
the highest priority environment variable is determined and used.
Highest priority means here the following list with decreasing
priority:
@enumerate
@item @code{LANGUAGE}
@item @code{LC_ALL}
@item @code{LC_xxx}, according to selected locale
@item @code{LANG}
@end enumerate
Afterwards the path is constructed using the found value and the
translation file is loaded if available.
What is now when the value for, say, @code{LANGUAGE} changes. According
to the process explained above the new value of this variable is found
as soon as the @code{dcgettext} function is called. But this also means
the (perhaps) different message catalog file is loaded. In other
words: the used language is changed.
But there is one little hook. The code for gcc-2.7.0 and up provides
some optimization. This optimization normally prevents the calling of
the @code{dcgettext} function as long as no new catalog is loaded. But
if @code{dcgettext} is not called the program also cannot find the
@code{LANGUAGE} variable be changed (@pxref{Optimized gettext}). A
solution for this is very easy. Include the following code in the
language switching function.
@example
/* Change language. */
setenv ("LANGUAGE", "fr", 1);
/* Make change known. */
@{
extern int _nl_msg_cat_cntr;
++_nl_msg_cat_cntr;
@}
@end example
The variable @code{_nl_msg_cat_cntr} is defined in @file{loadmsgcat.c}.
The programmer will find himself in need for a construct like this only
when developing programs which do run longer and provide the user to
select the language at runtime. Non-interactive programs (like all
these little Unix tools) should never need this.
@end itemize
@node Temp Programmers, , gettext grok, Programmers
@section Temporary Notes for the Programmers Chapter
@menu
* Temp Implementations:: Temporary - Two Possible Implementations
* Temp catgets:: Temporary - About @code{catgets}
* Temp WSI:: Temporary - Why a single implementation
* Temp Notes:: Temporary - Notes
@end menu
@node Temp Implementations, Temp catgets, Temp Programmers, Temp Programmers
@subsection Temporary - Two Possible Implementations
There are two competing methods for language independent messages:
the X/Open @code{catgets} method, and the Uniforum @code{gettext}
method. The @code{catgets} method indexes messages by integers; the
@code{gettext} method indexes them by their English translations.
The @code{catgets} method has been around longer and is supported
by more vendors. The @code{gettext} method is supported by Sun,
and it has been heard that the COSE multi-vendor initiative is
supporting it. Neither method is a POSIX standard; the POSIX.1
committee had a lot of disagreement in this area.
Neither one is in the POSIX standard. There was much disagreement
in the POSIX.1 committee about using the @code{gettext} routines
vs. @code{catgets} (XPG). In the end the committee couldn't
agree on anything, so no messaging system was included as part
of the standard. I believe the informative annex of the standard
includes the XPG3 messaging interfaces, ``@dots{}as an example of
a messaging system that has been implemented@dots{}''
They were very careful not to say anywhere that you should use one
set of interfaces over the other. For more on this topic please
see the Programming for Internationalization FAQ.
@node Temp catgets, Temp WSI, Temp Implementations, Temp Programmers
@subsection Temporary - About @code{catgets}
There have been a few discussions of late on the use of
@code{catgets} as a base. I think it important to present both
sides of the argument and hence am opting to play devil's advocate
for a little bit.
I'll not deny the fact that @code{catgets} could have been designed
a lot better. It currently has quite a number of limitations and
these have already been pointed out.
However there is a great deal to be said for consistency and
standardization. A common recurring problem when writing Unix
software is the myriad portability problems across Unix platforms.
It seems as if every Unix vendor had a look at the operating system
and found parts they could improve upon. Undoubtedly, these
modifications are probably innovative and solve real problems.
However, software developers have a hard time keeping up with all
these changes across so many platforms.
And this has prompted the Unix vendors to begin to standardize their
systems. Hence the impetus for Spec1170. Every major Unix vendor
has committed to supporting this standard and every Unix software
developer waits with glee the day they can write software to this
standard and simply recompile (without having to use autoconf)
across different platforms.
As I understand it, Spec1170 is roughly based upon version 4 of the
X/Open Portability Guidelines (XPG4). Because @code{catgets} and
friends are defined in XPG4, I'm led to believe that @code{catgets}
is a part of Spec1170 and hence will become a standardized component
of all Unix systems.
@node Temp WSI, Temp Notes, Temp catgets, Temp Programmers
@subsection Temporary - Why a single implementation
Now it seems kind of wasteful to me to have two different systems
installed for accessing message catalogs. If we do want to remedy
@code{catgets} deficiencies why don't we try to expand @code{catgets}
(in a compatible manner) rather than implement an entirely new system.
Otherwise, we'll end up with two message catalog access systems installed
with an operating system - one set of routines for packages using GNU
@code{gettext} for their internationalization, and another set of routines
(catgets) for all other software. Bloated?
Supposing another catalog access system is implemented. Which do
we recommend? At least for Linux, we need to attract as many
software developers as possible. Hence we need to make it as easy
for them to port their software as possible. Which means supporting
@code{catgets}. We will be implementing the @code{glocale} code
within our @code{libc}, but does this mean we also have to incorporate
another message catalog access scheme within our @code{libc} as well?
And what about people who are going to be using the @code{glocale}
+ non-@code{catgets} routines. When they port their software to
other platforms, they're now going to have to include the front-end
(@code{glocale}) code plus the back-end code (the non-@code{catgets}
access routines) with their software instead of just including the
@code{glocale} code with their software.
Message catalog support is however only the tip of the iceberg.
What about the data for the other locale categories. They also have
a number of deficiencies. Are we going to abandon them as well and
develop another duplicate set of routines (should @code{glocale}
expand beyond message catalog support)?
Like many parts of Unix that can be improved upon, we're stuck with balancing
compatibility with the past with useful improvements and innovations for
the future.
@c @node Temp DLS, Temp Notes, Temp WSI, Temp Programmers
@c @subsection Temporary - Double layer solution
@c GNU locale implements a @code{gettext}-style interface on top of a
@c @code{catgets}-style interface.
@c This is not needless complexity. It is absolutely vital, because
@c it enables @code{gettext} to run on top of @code{catgets}, which
@c enables Linux International to recommend users use it @emph{today}.
@c Rewriting @code{gettext} so that it could use @emph{either}
@c @code{catgets} @emph{or} some simpler mechanism would not break
@c anything, but would not reduce complexity either. It might be
@c worth doing, but it isn't urgent.
@c In general, simplicity is not enough of a reason to rewrite a
@c program that works. Simplicity is just one desirable thing.
@c It is not overridingly important.
@node Temp Notes, , Temp WSI, Temp Programmers
@subsection Temporary - Notes
X/Open agreed very late on the standard form so that many
implementations differ from the final form. Both of my system (old
Linux catgets and Ultrix-4) have a strange variation.
OK. After incorporating the last changes I have to spend some time on
making the GNU/Linux @code{libc} @code{gettext} functions. So in future
Solaris is not the only system having @code{gettext}.
@node Translators, Maintainers, Programmers, Top
@chapter The Translator's View
@c FIXME: Reorganize whole chapter.
@menu
* Trans Intro 0:: Introduction 0
* Trans Intro 1:: Introduction 1
* Discussions:: Discussions
* Organization:: Organization
* Information Flow:: Information Flow
@end menu
@node Trans Intro 0, Trans Intro 1, Translators, Translators
@section Introduction 0
GNU is going international! The Translation Project is a way
to get maintainers, translators and users all together, so GNU will
gradually become able to speak many native languages.
The GNU @code{gettext} tool set contains @emph{everything} maintainers
need for internationalizing their packages for messages. It also
contains quite useful tools for helping translators at localizing
messages to their native language, once a package has already been
internationalized.
To achieve the Translation Project, we need many interested
people who like their own language and write it well, and who are also
able to synergize with other translators speaking the same language.
If you'd like to volunteer to @emph{work} at translating messages,
please send mail to your translating team.
Each team has its own mailing list, courtesy of Linux
International. You may reach your translating team at the address
@file{@var{ll}@@li.org}, replacing @var{ll} by the two-letter @w{ISO 639}
code for your language. Language codes are @emph{not} the same as
country codes given in @w{ISO 3166}. The following translating teams
exist:
@quotation
Chinese @code{zh}, Czech @code{cs}, Danish @code{da}, Dutch @code{nl},
Esperanto @code{eo}, Finnish @code{fi}, French @code{fr}, Irish
@code{ga}, German @code{de}, Greek @code{el}, Italian @code{it},
Japanese @code{ja}, Indonesian @code{in}, Norwegian @code{no}, Polish
@code{pl}, Portuguese @code{pt}, Russian @code{ru}, Spanish @code{es},
Swedish @code{sv} and Turkish @code{tr}.
@end quotation
@noindent
For example, you may reach the Chinese translating team by writing to
@file{zh@@li.org}. When you become a member of the translating team
for your own language, you may subscribe to its list. For example,
Swedish people can send a message to @w{@file{sv-request@@li.org}},
having this message body:
@example
subscribe
@end example
Keep in mind that team members should be interested in @emph{working}
at translations, or at solving translational difficulties, rather than
merely lurking around. If your team does not exist yet and you want to
start one, please write to @w{@file{gnu-translation@@prep.ai.mit.edu}};
you will then reach the GNU coordinator for all translator teams.
A handful of GNU packages have already been adapted and provided
with message translations for several languages. Translation
teams have begun to organize, using these packages as a starting
point. But there are many more packages and many languages for
which we have no volunteer translators. If you would like to
volunteer to work at translating messages, please send mail to
@file{gnu-translation@@prep.ai.mit.edu} indicating what language(s)
you can work on.
@node Trans Intro 1, Discussions, Trans Intro 0, Translators
@section Introduction 1
This is now official, GNU is going international! Here is the
announcement submitted for the January 1995 GNU Bulletin:
@quotation
A handful of GNU packages have already been adapted and provided
with message translations for several languages. Translation
teams have begun to organize, using these packages as a starting
point. But there are many more packages and many languages
for which we have no volunteer translators. If you'd like to
volunteer to work at translating messages, please send mail to
@samp{gnu-translation@@prep.ai.mit.edu} indicating what language(s)
you can work on.
@end quotation
This document should answer many questions for those who are curious
about the process or would like to contribute. Please at least skim
over it, hoping to cut down a little of the high volume of e-mail
generated by this collective effort towards GNU internationalization.
Most free programming which is widely shared is done in English, and
currently, English is used as the main communicating language between
national communities collaborating to the GNU project. This very document
is written in English. This will not change in the foreseeable future.
However, there is a strong appetite from national communities for
having more software able to write using national language and habits,
and there is an on-going effort to modify GNU software in such a way
that it becomes able to do so. The experiments driven so far raised
an enthusiastic response from pretesters, so we believe that GNU
internationalization is dedicated to succeed.
For suggestion clarifications, additions or corrections to this
document, please e-mail to @file{gnu-translation@@prep.ai.mit.edu}.
@node Discussions, Organization, Trans Intro 1, Translators
@section Discussions
Facing this internationalization effort, a few users expressed their
concerns. Some of these doubts are presented and discussed, here.
@itemize @bullet
@item Smaller groups
Some languages are not spoken by a very large number of people, so people
speaking them sometimes consider that there may not be all that much
demand such versions of free software packages. Moreover, many people
being @emph{into computers}, in some countries, generally seem to prefer
English versions of their software.
On the other end, people might enjoy their own language a lot, and be
very motivated at providing to themselves the pleasure of having their
beloved free software speaking their mother tongue. They do themselves
a personal favor, and do not pay that much attention to the number of
people beneficiating of their work.
@item Misinterpretation
Other users are shy to push forward their own language, seeing in this
some kind of misplaced propaganda. Someone thought there must be some
users of the language over the networks pestering other people with it.
But any spoken language is worth localization, because there are
people behind the language for whom the language is important and
dear to their hearts.
@item Odd translations
The biggest problem is to find the right translations so that
everybody can understand the messages. Translations are usually a
little odd. Some people get used to English, to the extent they may
find translations into their own language ``rather pushy, obnoxious
and sometimes even hilarious.'' As a French speaking man, I have
the experience of those instruction manuals for goods, so poorly
translated in French in Korea or Taiwan@dots{}
The fact is that we sometimes have to create a kind of national
computer culture, and this is not easy without the collaboration of
many people liking their mother tongue. This is why translations are
better achieved by people knowing and loving their own language, and
ready to work together at improving the results they obtain.
@item Dependencies over the GPL
Some people wonder if using GNU @code{gettext} necessarily brings their package
under the protective wing of the GNU General Public License, when they
do not want to make their program free, or want other kinds of freedom.
The simplest answer is yes.
The mere marking of localizable strings in a package, or conditional
inclusion of a few lines for initialization, is not really including
GPL'ed code. However, the localization routines themselves are under
the GPL and would bring the remainder of the package under the GPL
if they were distributed with it. So, I presume that, for those
for which this is a problem, it could be circumvented by letting to
the end installers the burden of assembling a package prepared for
localization, but not providing the localization routines themselves.
@end itemize
@node Organization, Information Flow, Discussions, Translators
@section Organization
On a larger scale, the true solution would be to organize some kind of
fairly precise set up in which volunteers could participate. I gave
some thought to this idea lately, and realize there will be some
touchy points. I thought of writing to Richard Stallman to launch
such a project, but feel it might be good to shake out the ideas
between ourselves first. Most probably that Linux International has
some experience in the field already, or would like to orchestrate
the volunteer work, maybe. Food for thought, in any case!
I guess we have to setup something early, somehow, that will help
many possible contributors of the same language to interlock and avoid
work duplication, and further be put in contact for solving together
problems particular to their tongue (in most languages, there are many
difficulties peculiar to translating technical English). My Swedish
contributor acknowledged these difficulties, and I'm well aware of
them for French.
This is surely not a technical issue, but we should manage so the
effort of locale contributors be maximally useful, despite the national
team layer interface between contributors and maintainers.
The Translation Project needs some setup for coordinating language
coordinators. Localizing evolving programs will surely
become a permanent and continuous activity in the free software community,
once well started.
The setup should be minimally completed and tested before GNU
@code{gettext} becomes an official reality. The e-mail address
@file{translation@@iro.umontreal.ca} has been setup for receiving
offers from volunteers and general e-mail on these topics. This address
reaches the Translation Project coordinator.
@menu
* Central Coordination:: Central Coordination
* National Teams:: National Teams
* Mailing Lists:: Mailing Lists
@end menu
@node Central Coordination, National Teams, Organization, Organization
@subsection Central Coordination
I also think GNU will need sooner than it thinks, that someone setup
a way to organize and coordinate these groups. Some kind of group
of groups. My opinion is that it would be good that GNU delegates
this task to a small group of collaborating volunteers, shortly.
Perhaps in @file{gnu.announce} a list of this national committee's
can be published.
My role as coordinator would simply be to refer to Ulrich any German
speaking volunteer interested to localization of free software packages, and
maybe helping national groups to initially organize, while maintaining
national registries for until national groups are ready to take over.
In fact, the coordinator should ease volunteers to get in contact with
one another for creating national teams, which should then select
one coordinator per language, or country (regionalized language).
If well done, the coordination should be useful without being an
overwhelming task, the time to put delegations in place.
@node National Teams, Mailing Lists, Central Coordination, Organization
@subsection National Teams
I suggest we look for volunteer coordinators/editors for individual
languages. These people will scan contributions of translation files
for various programs, for their own languages, and will ensure high
and uniform standards of diction.
From my current experience with other people in these days, those who
provide localizations are very enthusiastic about the process, and are
more interested in the localization process than in the program they
localize, and want to do many programs, not just one. This seems
to confirm that having a coordinator/editor for each language is a
good idea.
We need to choose someone who is good at writing clear and concise
prose in the language in question. That is hard---we can't check
it ourselves. So we need to ask a few people to judge each others'
writing and select the one who is best.
I announce my prerelease to a few dozen people, and you would not
believe all the discussions it generated already. I shudder to think
what will happen when this will be launched, for true, officially,
world wide. Who am I to arbitrate between two Czekolsovak users
contradicting each other, for example?
I assume that your German is not much better than my French so that
I would not be able to judge about these formulations. What I would
suggest is that for each language there is a group for people who
maintain the PO files and judge about changes. I suspect there will
be cultural differences between how such groups of people will behave.
Some will have relaxed ways, reach consensus easily, and have anyone
of the group relate to the maintainers, while others will fight to
death, organize heavy administrations up to national standards, and
use strict channels.
The German team is putting out a good example. Right now, they are
maybe half a dozen people revising translations of each other and
discussing the linguistic issues. I do not even have all the names.
Ulrich Drepper is taking care of coordinating the German team.
He subscribed to all my pretest lists, so I do not even have to warn
him specifically of incoming releases.
I'm sure, that is a good idea to get teams for each language working
on translations. That will make the translations better and more
consistent.
@menu
* Sub-Cultures:: Sub-Cultures
* Organizational Ideas:: Organizational Ideas
@end menu
@node Sub-Cultures, Organizational Ideas, National Teams, National Teams
@subsubsection Sub-Cultures
Taking French for example, there are a few sub-cultures around computers
which developed diverging vocabularies. Picking volunteers here and
there without addressing this problem in an organized way, soon in the
project, might produce a distasteful mix of internationalized programs,
and possibly trigger endless quarrels among those who really care.
Keeping some kind of unity in the way French localization of
internationalized programs is achieved is a difficult (and delicate) job.
Knowing the latin character of French people (:-), if we take this
the wrong way, we could end up nowhere, or spoil a lot of energies.
Maybe we should begin to address this problem seriously @emph{before}
GNU @code{gettext} become officially published. And I suspect that this
means soon!
@node Organizational Ideas, , Sub-Cultures, National Teams
@subsubsection Organizational Ideas
I expect the next big changes after the official release. Please note
that I use the German translation of the short GPL message. We need
to set a few good examples before the localization goes out for true
in the free software community. Here are a few points to discuss:
@itemize @bullet
@item
Each group should have one FTP server (at least one master).
@item
The files on the server should reflect the latest version (of
course!) and it should also contain a RCS directory with the
corresponding archives (I don't have this now).
@item
There should also be a ChangeLog file (this is more useful than the
RCS archive but can be generated automatically from the later by
Emacs).
@item
A @dfn{core group} should judge about questionable changes (for now
this group consists solely by me but I ask some others occasionally;
this also seems to work).
@end itemize
@node Mailing Lists, , National Teams, Organization
@subsection Mailing Lists
If we get any inquiries about GNU @code{gettext}, send them on to:
@example
@file{translation@@iro.umontreal.ca}
@end example
The @file{*-pretest} lists are quite useful to me, maybe the idea could
be generalized to many GNU, and non-GNU packages. But each maintainer
his/her way!
Fran@,{c}ois, we have a mechanism in place here at
@file{gnu.ai.mit.edu} to track teams, support mailing lists for
them and log members. We have a slight preference that you use it.
If this is OK with you, I can get you clued in.
Things are changing! A few years ago, when Daniel Fekete and I
asked for a mailing list for GNU localization, nested at the FSF, we
were politely invited to organize it anywhere else, and so did we.
For communicating with my pretesters, I later made a handful of
mailing lists located at iro.umontreal.ca and administrated by
@code{majordomo}. These lists have been @emph{very} dependable
so far@dots{}
I suspect that the German team will organize itself a mailing list
located in Germany, and so forth for other countries. But before they
organize for true, it could surely be useful to offer mailing lists
located at the FSF to each national team. So yes, please explain me
how I should proceed to create and handle them.
We should create temporary mailing lists, one per country, to help
people organize. Temporary, because once regrouped and structured, it
would be fair the volunteers from country bring back @emph{their} list
in there and manage it as they want. My feeling is that, in the long
run, each team should run its own list, from within their country.
There also should be some central list to which all teams could
subscribe as they see fit, as long as each team is represented in it.
@node Information Flow, , Organization, Translators
@section Information Flow
There will surely be some discussion about this messages after the
packages are finally released. If people now send you some proposals
for better messages, how do you proceed? Jim, please note that
right now, as I put forward nearly a dozen of localizable programs, I
receive both the translations and the coordination concerns about them.
If I put one of my things to pretest, Ulrich receives the announcement
and passes it on to the German team, who make last minute revisions.
Then he submits the translation files to me @emph{as the maintainer}.
For free packages I do not maintain, I would not even hear about it.
This scheme could be made to work for the whole Translation Project,
I think. For security reasons, maybe Ulrich (national coordinators,
in fact) should update central registry kept at the Translation Project
(Jim, me, or Len's recruits) once in a while.
In December/January, I was aggressively ready to internationalize
all of GNU, giving myself the duty of one small GNU package per week
or so, taking many weeks or months for bigger packages. But it does
not work this way. I first did all the things I'm responsible for.
I've nothing against some missionary work on other maintainers, but
I'm also loosing a lot of energy over it---same debates over again.
And when the first localized packages are released we'll get a lot of
responses about ugly translations :-). Surely, and we need to have
beforehand a fairly good idea about how to handle the information
flow between the national teams and the package maintainers.
Please start saving somewhere a quick history of each PO file. I know
for sure that the file format will change, allowing for comments.
It would be nice that each file has a kind of log, and references for
those who want to submit comments or gripes, or otherwise contribute.
I sent a proposal for a fast and flexible format, but it is not
receiving acceptance yet by the GNU deciders. I'll tell you when I
have more information about this.
@node Maintainers, Conclusion, Translators, Top
@chapter The Maintainer's View
The maintainer of a package has many responsibilities. One of them
is ensuring that the package will install easily on many platforms,
and that the magic we described earlier (@pxref{Users}) will work
for installers and end users.
Of course, there are many possible ways by which GNU @code{gettext}
might be integrated in a distribution, and this chapter does not cover
them in all generality. Instead, it details one possible approach which
is especially adequate for many free software distributions following GNU
standards, or even better, Gnits standards, because GNU @code{gettext}
is purposely for helping the internationalization of the whole GNU
project, and as many other good free packages as possible. So, the
maintainer's view presented here presumes that the package already has
a @file{configure.in} file and uses GNU Autoconf.
Nevertheless, GNU @code{gettext} may surely be useful for free packages
not following GNU standards and conventions, but the maintainers of such
packages might have to show imagination and initiative in organizing
their distributions so @code{gettext} work for them in all situations.
There are surely many, out there.
Even if @code{gettext} methods are now stabilizing, slight adjustments
might be needed between successive @code{gettext} versions, so you
should ideally revise this chapter in subsequent releases, looking
for changes.
@menu
* Flat and Non-Flat:: Flat or Non-Flat Directory Structures
* Prerequisites:: Prerequisite Works
* gettextize Invocation:: Invoking the @code{gettextize} Program
* Adjusting Files:: Files You Must Create or Alter
@end menu
@node Flat and Non-Flat, Prerequisites, Maintainers, Maintainers
@section Flat or Non-Flat Directory Structures
Some free software packages are distributed as @code{tar} files which unpack
in a single directory, these are said to be @dfn{flat} distributions.
Other free software packages have a one level hierarchy of subdirectories, using
for example a subdirectory named @file{doc/} for the Texinfo manual and
man pages, another called @file{lib/} for holding functions meant to
replace or complement C libraries, and a subdirectory @file{src/} for
holding the proper sources for the package. These other distributions
are said to be @dfn{non-flat}.
For now, we cannot say much about flat distributions. A flat
directory structure has the disadvantage of increasing the difficulty
of updating to a new version of GNU @code{gettext}. Also, if you have
many PO files, this could somewhat pollute your single directory.
In the GNU @code{gettext} distribution, the @file{misc/} directory
contains a shell script named @file{combine-sh}. That script may
be used for combining all the C files of the @file{intl/} directory
into a pair of C files (one @file{.c} and one @file{.h}). Those two
generated files would fit more easily in a flat directory structure,
and you will then have to add these two files to your project.
Maybe because GNU @code{gettext} itself has a non-flat structure,
we have more experience with this approach, and this is what will be
described in the remaining of this chapter. Some maintainers might
use this as an opportunity to unflatten their package structure.
Only later, once gained more experience adapting GNU @code{gettext}
to flat distributions, we might add some notes about how to proceed
in flat situations.
@node Prerequisites, gettextize Invocation, Flat and Non-Flat, Maintainers
@section Prerequisite Works
There are some works which are required for using GNU @code{gettext}
in one of your package. These works have some kind of generality
that escape the point by point descriptions used in the remainder
of this chapter. So, we describe them here.
@itemize @bullet
@item
Before attempting to use you should install some other packages first.
Ensure that recent versions of GNU @code{m4}, GNU Autoconf and GNU
@code{gettext} are already installed at your site, and if not, proceed
to do this first. If you got to install these things, beware that
GNU @code{m4} must be fully installed before GNU Autoconf is even
@emph{configured}.
To further ease the task of a package maintainer the @code{automake}
package was designed and implemented. GNU @code{gettext} now uses this
tool and the @file{Makefile}s in the @file{intl/} and @file{po/}
therefore know about all the goals necessary for using @code{automake}
and @file{libintl} in one project.
Those four packages are only needed to you, as a maintainer; the
installers of your own package and end users do not really need any of
GNU @code{m4}, GNU Autoconf, GNU @code{gettext}, or GNU @code{automake}
for successfully installing and running your package, with messages
properly translated. But this is not completely true if you provide
internationalized shell scripts within your own package: GNU
@code{gettext} shall then be installed at the user site if the end users
want to see the translation of shell script messages.
@item
Your package should use Autoconf and have a @file{configure.in} file.
If it does not, you have to learn how. The Autoconf documentation
is quite well written, it is a good idea that you print it and get
familiar with it.
@item
Your C sources should have already been modified according to
instructions given earlier in this manual. @xref{Sources}.
@item
Your @file{po/} directory should receive all PO files submitted to you
by the translator teams, each having @file{@var{ll}.po} as a name.
This is not usually easy to get translation
work done before your package gets internationalized and available!
Since the cycle has to start somewhere, the easiest for the maintainer
is to start with absolutely no PO files, and wait until various
translator teams get interested in your package, and submit PO files.
@end itemize
It is worth adding here a few words about how the maintainer should
ideally behave with PO files submissions. As a maintainer, your role is
to authentify the origin of the submission as being the representative
of the appropriate translating teams of the Translation Project (forward
the submission to @file{translation@@iro.umontreal.ca} in case of doubt),
to ensure that the PO file format is not severely broken and does not
prevent successful installation, and for the rest, to merely to put these
PO files in @file{po/} for distribution.
As a maintainer, you do not have to take on your shoulders the
responsibility of checking if the translations are adequate or
complete, and should avoid diving into linguistic matters. Translation
teams drive themselves and are fully responsible of their linguistic
choices for the Translation Project. Keep in mind that translator teams are @emph{not}
driven by maintainers. You can help by carefully redirecting all
communications and reports from users about linguistic matters to the
appropriate translation team, or explain users how to reach or join
their team. The simplest might be to send them the @file{ABOUT-NLS} file.
Maintainers should @emph{never ever} apply PO file bug reports
themselves, short-cutting translation teams. If some translator has
difficulty to get some of her points through her team, it should not be
an issue for her to directly negotiate translations with maintainers.
Teams ought to settle their problems themselves, if any. If you, as
a maintainer, ever think there is a real problem with a team, please
never try to @emph{solve} a team's problem on your own.
@node gettextize Invocation, Adjusting Files, Prerequisites, Maintainers
@section Invoking the @code{gettextize} Program
Some files are consistently and identically needed in every package
internationalized through GNU @code{gettext}. As a matter of
convenience, the @code{gettextize} program puts all these files right
in your package. This program has the following synopsis:
@example
gettextize [ @var{option}@dots{} ] [ @var{directory} ]
@end example
@noindent
and accepts the following options:
@table @samp
@item -c
@itemx --copy
Copy the needed files instead of making symbolic links. Using links
would allow the package to always use the latest @code{gettext} code
available on the system, but it might disturb some mechanism the
maintainer is used to apply to the sources. Because running
@code{gettextize} is easy there shouldn't be problems with using copies.
@item -f
@itemx --force
Force replacement of files which already exist.
@item -h
@itemx --help
Display this help and exit.
@item --version
Output version information and exit.
@end table
If @var{directory} is given, this is the top level directory of a
package to prepare for using GNU @code{gettext}. If not given, it
is assumed that the current directory is the top level directory of
such a package.
The program @code{gettextize} provides the following files. However,
no existing file will be replaced unless the option @code{--force}
(@code{-f}) is specified.
@enumerate
@item
The @file{ABOUT-NLS} file is copied in the main directory of your package,
the one being at the top level. This file gives the main indications
about how to install and use the Native Language Support features
of your program. You might elect to use a more recent copy of this
@file{ABOUT-NLS} file than the one provided through @code{gettextize},
if you have one handy. You may also fetch a more recent copy of file
@file{ABOUT-NLS} from Translation Project sites, and from most GNU
archive sites.
@item
A @file{po/} directory is created for eventually holding
all translation files, but initially only containing the file
@file{po/Makefile.in.in} from the GNU @code{gettext} distribution.
(beware the double @samp{.in} in the file name). If the @file{po/}
directory already exists, it will be preserved along with the files
it contains, and only @file{Makefile.in.in} will be overwritten.
@item
A @file{intl/} directory is created and filled with most of the files
originally in the @file{intl/} directory of the GNU @code{gettext}
distribution. Also, if option @code{--force} (@code{-f}) is given,
the @file{intl/} directory is emptied first.
@end enumerate
If your site support symbolic links, @code{gettextize} will not
actually copy the files into your package, but establish symbolic
links instead. This avoids duplicating the disk space needed in
all packages. Merely using the @samp{-h} option while creating the
@code{tar} archive of your distribution will resolve each link by an
actual copy in the distribution archive. So, to insist, you really
should use @samp{-h} option with @code{tar} within your @code{dist}
goal of your main @file{Makefile.in}.
It is interesting to understand that most new files for supporting
GNU @code{gettext} facilities in one package go in @file{intl/}
and @file{po/} subdirectories. One distinction between these two
directories is that @file{intl/} is meant to be completely identical
in all packages using GNU @code{gettext}, while all newly created
files, which have to be different, go into @file{po/}. There is a
common @file{Makefile.in.in} in @file{po/}, because the @file{po/}
directory needs its own @file{Makefile}, and it has been designed so
it can be identical in all packages.
@node Adjusting Files, , gettextize Invocation, Maintainers
@section Files You Must Create or Alter
Besides files which are automatically added through @code{gettextize},
there are many files needing revision for properly interacting with
GNU @code{gettext}. If you are closely following GNU standards for
Makefile engineering and auto-configuration, the adaptations should
be easier to achieve. Here is a point by point description of the
changes needed in each.
So, here comes a list of files, each one followed by a description of
all alterations it needs. Many examples are taken out from the GNU
@code{gettext} @value{VERSION} distribution itself. You may indeed
refer to the source code of the GNU @code{gettext} package, as it
is intended to be a good example and master implementation for using
its own functionality.
@menu
* po/POTFILES.in:: @file{POTFILES.in} in @file{po/}
* configure.in:: @file{configure.in} at top level
* aclocal:: @file{aclocal.m4} at top level
* acconfig:: @file{acconfig.h} at top level
* Makefile:: @file{Makefile.in} at top level
* src/Makefile:: @file{Makefile.in} in @file{src/}
@end menu
@node po/POTFILES.in, configure.in, Adjusting Files, Adjusting Files
@subsection @file{POTFILES.in} in @file{po/}
The @file{po/} directory should receive a file named
@file{POTFILES.in}. This file tells which files, among all program
sources, have marked strings needing translation. Here is an example
of such a file:
@example
@group
# List of source files containing translatable strings.
# Copyright (C) 1995 Free Software Foundation, Inc.
# Common library files
lib/error.c
lib/getopt.c
lib/xmalloc.c
# Package source files
src/gettextp.c
src/msgfmt.c
src/xgettext.c
@end group
@end example
@noindent
Dashed comments and white lines are ignored. All other lines
list those source files containing strings marked for translation
(@pxref{Mark Keywords}), in a notation relative to the top level
of your whole distribution, rather than the location of the
@file{POTFILES.in} file itself.
@node configure.in, aclocal, po/POTFILES.in, Adjusting Files
@subsection @file{configure.in} at top level
@enumerate
@item Declare the package and version.
This is done by a set of lines like these:
@example
PACKAGE=gettext
VERSION=@value{VERSION}
AC_DEFINE_UNQUOTED(PACKAGE, "$PACKAGE")
AC_DEFINE_UNQUOTED(VERSION, "$VERSION")
AC_SUBST(PACKAGE)
AC_SUBST(VERSION)
@end example
@noindent
Of course, you replace @samp{gettext} with the name of your package,
and @samp{@value{VERSION}} by its version numbers, exactly as they
should appear in the packaged @code{tar} file name of your distribution
(@file{gettext-@value{VERSION}.tar.gz}, here).
@item Declare the available translations.
This is done by defining @code{ALL_LINGUAS} to the white separated,
quoted list of available languages, in a single line, like this:
@example
ALL_LINGUAS="de fr"
@end example
@noindent
This example means that German and French PO files are available, so
that these languages are currently supported by your package. If you
want to further restrict, at installation time, the set of installed
languages, this should not be done by modifying @code{ALL_LINGUAS} in
@file{configure.in}, but rather by using the @code{LINGUAS} environment
variable (@pxref{Installers}).
@item Check for internationalization support.
Here is the main @code{m4} macro for triggering internationalization
support. Just add this line to @file{configure.in}:
@example
AM_GNU_GETTEXT
@end example
@noindent
This call is purposely simple, even if it generates a lot of configure
time checking and actions.
@item Have output files created.
The @code{AC_OUTPUT} directive, at the end of your @file{configure.in}
file, needs to be modified in two ways:
@example
AC_OUTPUT([@var{existing configuration files} intl/Makefile po/Makefile.in],
@var{existing additional actions}])
@end example
The modification to the first argument to @code{AC_OUTPUT} asks
for substitution in the @file{intl/} and @file{po/} directories.
Note the @samp{.in} suffix used for @file{po/} only. This is because
the distributed file is really @file{po/Makefile.in.in}.
@end enumerate
@node aclocal, acconfig, configure.in, Adjusting Files
@subsection @file{aclocal.m4} at top level
If you do not have an @file{aclocal.m4} file in your distribution,
the simplest is taking a copy of @file{aclocal.m4} from
GNU @code{gettext}. But to be precise, you only need macros
@code{AM_LC_MESSAGES}, @code{AM_WITH_NLS} and @code{AM_GNU_GETTEXT},
and @code{AM_PATH_PROG_WITH_TEST}, which is called by @code{AM_WITH_NLS},
so you may use an editor and remove macros you do not need.
If you already have an @file{aclocal.m4} file, then you will have
to merge the said macros into your @file{aclocal.m4}. Note that if
you are upgrading from a previous release of GNU @code{gettext}, you
should most probably @emph{replace} the said macros, as they usually
change a little from one release of GNU @code{gettext} to the next.
Their contents may vary as we get more experience with strange systems
out there.
These macros check for the internationalization support functions
and related informations. Hopefully, once stabilized, these macros
might be integrated in the standard Autoconf set, because this
piece of @code{m4} code will be the same for all projects using GNU
@code{gettext}.
@node acconfig, Makefile, aclocal, Adjusting Files
@subsection @file{acconfig.h} at top level
If you do not have an @file{acconfig.h} file in your distribution, the
simplest is use take a copy of @file{acconfig.h} from GNU
@code{gettext}. But to be precise, you only need the lines and comments
for @code{ENABLE_NLS}, @code{HAVE_CATGETS}, @code{HAVE_GETTEXT} and
@code{HAVE_LC_MESSAGES}, @code{HAVE_STPCPY}, @code{PACKAGE} and
@code{VERSION}, so you may use an editor and remove everything else. If
you already have an @file{acconfig.h} file, then you should merge the
said definitions into your @file{acconfig.h}.
@node Makefile, src/Makefile, acconfig, Adjusting Files
@subsection @file{Makefile.in} at top level
Here are a few modifications you need to make to your main, top-level
@file{Makefile.in} file.
@enumerate
@item
Add the following lines near the beginning of your @file{Makefile.in},
so the @samp{dist:} goal will work properly (as explained further down):
@example
PACKAGE = @@PACKAGE@@
VERSION = @@VERSION@@
@end example
@item
Add file @file{ABOUT-NLS} to the @code{DISTFILES} definition, so the file gets
distributed.
@item
Wherever you process subdirectories in your @file{Makefile.in}, be sure
you also process dir subdirectories @samp{intl} and @samp{po}. Special
rules in the @file{Makefiles} take care for the case where no
internationalization is wanted.
If you are using Makefiles, either generated by automake, or hand-written
so they carefully follow the GNU coding standards, the effected goals for
which the new subdirectories must be handled include @samp{installdirs},
@samp{install}, @samp{uninstall}, @samp{clean}, @samp{distclean}.
Here is an example of a canonical order of processing. In this
example, we also define @code{SUBDIRS} in @code{Makefile.in} for it
to be further used in the @samp{dist:} goal.
@example
SUBDIRS = doc lib @@INTLSUB@@ src @@POSUB@@
@end example
@noindent
that you will have to adapt to your own package.
@item
A delicate point is the @samp{dist:} goal, as both
@file{intl/Makefile} and @file{po/Makefile} will later assume that the
proper directory has been set up from the main @file{Makefile}. Here is
an example at what the @samp{dist:} goal might look like:
@example
distdir = $(PACKAGE)-$(VERSION)
dist: Makefile
rm -fr $(distdir)
mkdir $(distdir)
chmod 777 $(distdir)
for file in $(DISTFILES); do \
ln $$file $(distdir) 2>/dev/null || cp -p $$file $(distdir); \
done
for subdir in $(SUBDIRS); do \
mkdir $(distdir)/$$subdir || exit 1; \
chmod 777 $(distdir)/$$subdir; \
(cd $$subdir && $(MAKE) $@@) || exit 1; \
done
tar chozf $(distdir).tar.gz $(distdir)
rm -fr $(distdir)
@end example
@end enumerate
@node src/Makefile, , Makefile, Adjusting Files
@subsection @file{Makefile.in} in @file{src/}
Some of the modifications made in the main @file{Makefile.in} will
also be needed in the @file{Makefile.in} from your package sources,
which we assume here to be in the @file{src/} subdirectory. Here are
all the modifications needed in @file{src/Makefile.in}:
@enumerate
@item
In view of the @samp{dist:} goal, you should have these lines near the
beginning of @file{src/Makefile.in}:
@example
PACKAGE = @@PACKAGE@@
VERSION = @@VERSION@@
@end example
@item
If not done already, you should guarantee that @code{top_srcdir}
gets defined. This will serve for @code{cpp} include files. Just add
the line:
@example
top_srcdir = @@top_srcdir@@
@end example
@item
You might also want to define @code{subdir} as @samp{src}, later
allowing for almost uniform @samp{dist:} goals in all your
@file{Makefile.in}. At list, the @samp{dist:} goal below assume that
you used:
@example
subdir = src
@end example
@item
You should ensure that the final linking will use @code{@@INTLLIBS@@} as
a library. An easy way to achieve this is to manage that it gets into
@code{LIBS}, like this:
@example
LIBS = @@INTLLIBS@@ @@LIBS@@
@end example
In most packages internationalized with GNU @code{gettext}, one will
find a directory @file{lib/} in which a library containing some helper
functions will be build. (You need at least the few functions which the
GNU @code{gettext} Library itself needs.) However some of the functions
in the @file{lib/} also give messages to the user which of course should be
translated, too. Taking care of this it is not enough to place the support
library (say @file{libsupport.a}) just between the @code{@@INTLLIBS@@}
and @code{@@LIBS@@} in the above example. Instead one has to write this:
@example
LIBS = ../lib/libsupport.a @@INTLLIBS@@ ../lib/libsupport.a @@LIBS@@
@end example
@item
You should also ensure that directory @file{intl/} will be searched for
C preprocessor include files in all circumstances. So, you have to
manage so both @samp{-I../intl} and @samp{-I$(top_srcdir)/intl} will
be given to the C compiler.
@item
Your @samp{dist:} goal has to conform with others. Here is a
reasonable definition for it:
@example
distdir = ../$(PACKAGE)-$(VERSION)/$(subdir)
dist: Makefile $(DISTFILES)
for file in $(DISTFILES); do \
ln $$file $(distdir) 2>/dev/null || cp -p $$file $(distdir); \
done
@end example
@end enumerate
@node Conclusion, Country Codes, Maintainers, Top
@chapter Concluding Remarks
We would like to conclude this GNU @code{gettext} manual by presenting
an history of the Translation Project so far. We finally give
a few pointers for those who want to do further research or readings
about Native Language Support matters.
@menu
* History:: History of GNU @code{gettext}
* References:: Related Readings
@end menu
@node History, References, Conclusion, Conclusion
@section History of GNU @code{gettext}
Internationalization concerns and algorithms have been informally
and casually discussed for years in GNU, sometimes around GNU
@code{libc}, maybe around the incoming @code{Hurd}, or otherwise
(nobody clearly remembers). And even then, when the work started for
real, this was somewhat independently of these previous discussions.
This all began in July 1994, when Patrick D'Cruze had the idea and
initiative of internationalizing version 3.9.2 of GNU @code{fileutils}.
He then asked Jim Meyering, the maintainer, how to get those changes
folded into an official release. That first draft was full of
@code{#ifdef}s and somewhat disconcerting, and Jim wanted to find
nicer ways. Patrick and Jim shared some tries and experimentations
in this area. Then, feeling that this might eventually have a deeper
impact on GNU, Jim wanted to know what standards were, and contacted
Richard Stallman, who very quickly and verbally described an overall
design for what was meant to become @code{glocale}, at that time.
Jim implemented @code{glocale} and got a lot of exhausting feedback
from Patrick and Richard, of course, but also from Mitchum DSouza
(who wrote a @code{catgets}-like package), Roland McGrath, maybe David
MacKenzie, Fran@,{c}ois Pinard, and Paul Eggert, all pushing and
pulling in various directions, not always compatible, to the extent
that after a couple of test releases, @code{glocale} was torn apart.
While Jim took some distance and time and became dad for a second
time, Roland wanted to get GNU @code{libc} internationalized, and
got Ulrich Drepper involved in that project. Instead of starting
from @code{glocale}, Ulrich rewrote something from scratch, but
more conformant to the set of guidelines who emerged out of the
@code{glocale} effort. Then, Ulrich got people from the previous
forum to involve themselves into this new project, and the switch
from @code{glocale} to what was first named @code{msgutils}, renamed
@code{nlsutils}, and later @code{gettext}, became officially accepted
by Richard in May 1995 or so.
Let's summarize by saying that Ulrich Drepper wrote GNU @code{gettext}
in April 1995. The first official release of the package, including
PO mode, occurred in July 1995, and was numbered 0.7. Other people
contributed to the effort by providing a discussion forum around
Ulrich, writing little pieces of code, or testing. These are quoted
in the @code{THANKS} file which comes with the GNU @code{gettext}
distribution.
While this was being done, Fran@,{c}ois adapted half a dozen of
GNU packages to @code{glocale} first, then later to @code{gettext},
putting them in pretest, so providing along the way an effective
user environment for fine tuning the evolving tools. He also took
the responsibility of organizing and coordinating the Translation
Project. After nearly a year of informal exchanges between people from
many countries, translator teams started to exist in May 1995, through
the creation and support by Patrick D'Cruze of twenty unmoderated
mailing lists for that many native languages, and two moderated
lists: one for reaching all teams at once, the other for reaching
all willing maintainers of internationalized free software packages.
Fran@,{c}ois also wrote PO mode in June 1995 with the collaboration
of Greg McGary, as a kind of contribution to Ulrich's package.
He also gave a hand with the GNU @code{gettext} Texinfo manual.
@node References, , History, Conclusion
@section Related Readings
Eugene H. Dorr (@file{dorre@@well.com}) maintains an interesting
bibliography on internationalization matters, called
@cite{Internationalization Reference List}, which is available as:
@example
ftp://ftp.ora.com/pub/examples/nutshell/ujip/doc/i18n-books.txt
@end example
Michael Gschwind (@file{mike@@vlsivie.tuwien.ac.at}) maintains a
Frequently Asked Questions (FAQ) list, entitled @cite{Programming for
Internationalisation}. This FAQ discusses writing programs which
can handle different language conventions, character sets, etc.;
and is applicable to all character set encodings, with particular
emphasis on @w{ISO 8859-1}. It is regularly published in Usenet
groups @file{comp.unix.questions}, @file{comp.std.internat},
@file{comp.software.international}, @file{comp.lang.c},
@file{comp.windows.x}, @file{comp.std.c}, @file{comp.answers}
and @file{news.answers}. The home location of this document is:
@example
ftp://ftp.vlsivie.tuwien.ac.at/pub/8bit/ISO-programming
@end example
Patrick D'Cruze (@file{pdcruze@@li.org}) wrote a tutorial about NLS
matters, and Jochen Hein (@file{Hein@@student.tu-clausthal.de}) took
over the responsibility of maintaining it. It may be found as:
@example
ftp://sunsite.unc.edu/pub/Linux/utils/nls/catalogs/Incoming/...
...locale-tutorial-0.8.txt.gz
@end example
@noindent
This site is mirrored in:
@example
ftp://ftp.ibp.fr/pub/linux/sunsite/
@end example
A French version of the same tutorial should be findable at:
@example
ftp://ftp.ibp.fr/pub/linux/french/docs/
@end example
@noindent
together with French translations of many Linux-related documents.
@node Country Codes, , Conclusion, Top
@appendix Country Codes
The @w{ISO 639} standard defines two character codes for many countries.
All abreviations for countries or languages used in the Translation
Project should come from this standard.
@table @samp
@include iso-apdx.texi
@end table
@contents
@bye
@c Local variables:
@c texinfo-column-for-description: 32
@c End: