mirror of
https://github.com/brl/mutter.git
synced 2024-11-27 02:20:43 -05:00
7d674db9d2
The ClutterColor API has some inconsistencies: - the string deserialization function does not match the rest of the conversion function naming policy; the naming should be: clutter_color_parse() -> clutter_color_from_string() and the first parameter should be the ClutterColor that will be set from the string, not the string itself (a GDK-ism). - the fixed point API should not be exposed, especially in the form of ClutterFixed values - the non-fixed point HLS conversion functions do not make any sense. The values returned should be: hue := range [ 0, 360 ] luminance := range [ 0, 1 ] saturation := range [ 0, 1 ] like the current fixed point API does. Returning a value in the [ 0, 255 ] range is completely useless - the clutter_color_equal() should be converted for its use inside a GHashTable; a clutter_color_hash() should be added as well - the second parameter of the clutter_color_shade() function should be the shading factor, not the result (another GDK-ism). this way the function call can be translated from this: color.shade(out result, factor) to the more natural: color.shade(factor, out result) This somewhat large commit fixes all these issues and updates the internal users of the API. |
||
---|---|---|
.. | ||
conform | ||
data | ||
interactive | ||
micro-bench | ||
tools | ||
.gitignore | ||
Makefile.am | ||
README |
Outline of test categories: The conform/ tests should be non-interactive unit-tests that verify a single feature is behaving as documented. See conform/ADDING_NEW_TESTS for more details. The micro-bench/ tests should be focused perfomance test, ideally testing a single metric. Please never forget that these tests are synthetec and if you are using them then you understand what metric is being tested. They probably don't reflect any real world application loads and the intention is that you use these tests once you have already determined the crux of your problem and need focused feedback that your changes are indeed improving matters. There is no exit status requirements for these tests, but they should give clear feedback as to their performance. If the framerate is the feedback metric, then the test should forcibly enable FPS debugging. The interactive/ tests are any tests whos status can not be determined without a user looking at some visual output, or providing some manual input etc. This covers most of the original Clutter tests. Ideally some of these tests will be migrated into the conformance/ directory so they can be used in automated nightly tests. Other notes: All tests should ideally include a detailed description in the source explaining exactly what the test is for, how the test was designed to work, and possibly a rationale for the aproach taken for testing.