7646404196
ClutterScript should be able to automatically call gettext() and friends on strings loaded from a UI definition, prior to passing the string to the object it is constructing. The basic implementation is trivial: - set a translation domain on the ClutterScript instance - mark the translatable strings inside the JSON data, like: "property" : { "translatable" : true, "string" : "a translatable string" } The hard part is now getting the tools we use to extract the translatable strings to understand the JSON format we use inside ClutterScript. |
||
---|---|---|
.. | ||
accessibility | ||
conform | ||
data | ||
interactive | ||
micro-bench | ||
performance | ||
Makefile.am | ||
README |
Outline of test categories: The conform/ tests should be non-interactive unit-tests that verify a single feature is behaving as documented. See conform/ADDING_NEW_TESTS for more details. The performance/ tests are performance tests, both focused tests testing single metrics and larger tests. These tests are used to report one or more performance markers for the build of Clutter. Each performance marker is picked up from the standard output of running the tests from strings having the form "\n@ marker-name: 42.23" where 'marker-name' and '42.23' are the key/value pairs of a single metric. Each test can provide multiple key/value pairs. Note that if framerate is the feedback metric the test should forcibly enable FPS debugging itself. The file test-common.h contains utility function helping to do fps reporting. The interactive/ tests are any tests whose status can not be determined without a user looking at some visual output, or providing some manual input etc. This covers most of the original Clutter tests. Ideally some of these tests will be migrated into the conformance/ directory so they can be used in automated nightly tests. The accessibility/ tests are tests created to test the accessibility support of clutter, testing some of the atk interfaces. The data/ directory contains optional data (like images and ClutterScript definitions) that can be referenced by a test. Other notes: • All tests should ideally include a detailed description in the source explaining exactly what the test is for, how the test was designed to work, and possibly a rationale for the approach taken for testing. • When running tests under Valgrind, you should follow the instructions available here: http://live.gnome.org/Valgrind and also use the suppression file available inside the data/ directory.