config root man

Current Path : /home/usr.opt/mysql57/mysql-test/suite/parts/inc/

FreeBSD hs32.drive.ne.jp 9.1-RELEASE FreeBSD 9.1-RELEASE #1: Wed Jan 14 12:18:08 JST 2015 root@hs32.drive.ne.jp:/sys/amd64/compile/hs32 amd64
Upload File :
Current File : //home/usr.opt/mysql57/mysql-test/suite/parts/inc/partition.pre

################################################################################
# inc/partition.pre                                                            #
#                                                                              #
# Purpose:                                                                     #
#   Auxiliary script creating prerequisites needed by the partitioning tests   #
#   The name of the toplevel scripts sourcing this one is                      #
#         t/partition_<feature>_<storage engine>.test                          #
#                                                                              #
# Several parameters have to be set before this file is sourced.               #
# Please refer to the README.                                                  #
#                                                                              #
# The README for the partitioning testcases is at the end of this file.        #
#                                                                              #
#------------------------------------------------------------------------------#
# Original Author: mleich                                                      #
# Original Date:   2006-03-05                                                  #
# Change Author:   mleich                                                      #
# Change Date:     2007-10-08                                                  #
# Change:          Minor cleanup and fix for                                   #
#                  Bug#31243 Test "partition_basic_myisam" truncates path names#
#                  - Blow column file_list up to VARBINARY(10000)              #
#                  - remove reference to fixed bugs #17455, #19305             #
################################################################################

# Set the session storage engine
eval SET @@session.default_storage_engine = $engine;

##### Disabled/affected testcases, because of open bugs #####
# --echo
# --echo #------------------------------------------------------------------------
# --echo # There are several testcases disabled because of the open bugs
# if (`SELECT @@session.default_storage_engine IN('ndbcluster')`)
# {
# --echo # #18730
# }
# --echo #------------------------------------------------------------------------
# # Attention: Only bugs appearing in all storage engines should be mentioned above.
# #            The top level test wrapper (example: t/partition_basic_ndb.test)
# #            may set the $fixed_bug<nnnnn> variable to 0 after sourcing
# #            this file.
# # Bug#18730: Partitions: NDB, crash on SELECT MIN(<unique column>)
# # Attention: NDB testcases set this variable later to 0
# let $fixed_bug18730= 1;

--echo
--echo #------------------------------------------------------------------------
--echo #  0. Setting of auxiliary variables + Creation of an auxiliary tables
--echo #     needed in many testcases
--echo #------------------------------------------------------------------------
# Set the variable $no_debug depending on the current value of $debug;
--disable_query_log
eval SET @aux = $debug;
let $no_debug= `SELECT @aux = 0`;
--enable_query_log
if ($debug)
{
--echo # Attention: Script debugging is swiched on.
--echo #       - all statements will be protocolled
--echo #        - some additional will be executed
--echo #       It is to be expected, that we get huge differences.
}

let $ER_DUP_KEY=                          1022;
let $ER_GET_ERRNO=                        1030;
let $ER_BAD_NULL_ERROR=                   1048;
let $ER_DUP_ENTRY=                        1062;
let $ER_PARSE_ERROR=                      1064;
let $ER_TOO_MANY_PARTITIONS_ERROR=        1499;
let $ER_UNIQUE_KEY_NEED_ALL_FIELDS_IN_PF= 1503;
let $ER_NO_PARTS_ERROR=                   1504;
let $ER_DROP_PARTITION_NON_EXISTENT=      1507;
let $ER_SAME_NAME_PARTITION=              1517;
let $ER_NO_PARTITION_FOR_GIVEN_VALUE=     1526;

# Set the variable $engine_other to a storage engine <> $engine
--disable_query_log
eval SELECT UPPER($engine) = 'MEMORY' INTO @aux;
let $aux= `SELECT @aux`;
if ($aux)
{
   let $engine_other= 'MyISAM';
}
if (!$aux)
{
   let $engine_other= 'MEMORY';
}
--enable_query_log

# Numbers used for
# - partitioning           Example: ... PARTITION part1 VALUES LESS THAN ($max_row_div2)
# - INSERT/SELECT/UPDATE/DELETE    Example: ... WHERE f_int1 > @max_row_div3
let $max_row= `SELECT @max_row`;
SELECT @max_row DIV 2 INTO @max_row_div2;
let $max_row_div2= `SELECT @max_row_div2`;
SELECT @max_row DIV 3 INTO @max_row_div3;
let $max_row_div3= `SELECT @max_row_div3`;
SELECT @max_row DIV 4 INTO @max_row_div4;
let $max_row_div4= `SELECT @max_row_div4`;
SET @max_int_4 = 2147483647;
let $max_int_4= `SELECT @max_int_4`;

# Three insert statements used in many testcases.
let $insert_first_half= INSERT INTO t1(f_int1,f_int2,f_char1,f_char2,f_charbig)
SELECT f_int1,f_int2,f_char1,f_char2,f_charbig FROM t0_template
WHERE f_int1 BETWEEN 1 AND @max_row_div2 - 1;
let $insert_second_half= INSERT INTO t1(f_int1,f_int2,f_char1,f_char2,f_charbig)
SELECT f_int1,f_int2,f_char1,f_char2,f_charbig FROM t0_template
WHERE f_int1 BETWEEN @max_row_div2 AND @max_row;
#
let $insert_first_third= INSERT INTO t1(f_int1,f_int2,f_char1,f_char2,f_charbig)
SELECT f_int1,f_int2,f_char1,f_char2,f_charbig FROM t0_template
WHERE f_int1 BETWEEN 1 AND @max_row_div3 - 1;
let $insert_second_third= INSERT INTO t1(f_int1,f_int2,f_char1,f_char2,f_charbig)
SELECT f_int1,f_int2,f_char1,f_char2,f_charbig FROM t0_template
WHERE f_int1 BETWEEN @max_row_div3 AND 2 * @max_row_div3 - 1;
let $insert_third_third= INSERT INTO t1(f_int1,f_int2,f_char1,f_char2,f_charbig)
SELECT f_int1,f_int2,f_char1,f_char2,f_charbig FROM t0_template
WHERE f_int1 BETWEEN 2 * @max_row_div3 AND @max_row;
#
let $insert_all= INSERT INTO t1(f_int1,f_int2,f_char1,f_char2,f_charbig)
SELECT f_int1,f_int2,f_char1,f_char2,f_charbig FROM t0_template;

# Column list with definition for all tables to be checked
let $column_list= f_int1 INTEGER,
f_int2 INTEGER,
f_char1 CHAR(20),
f_char2 CHAR(20),
f_charbig VARCHAR(1000);

# Currently (April 2006) the default compiled NDB cannot manage
#    no_of_partitions (no subpartitioning) > 8
#    no_of_partitions * no_of_subpartitions > 8
# This NDB specific limitation will cause
#   1005: Can't create table 'test.t1' (errno: 1224)
# in partition_methods[1|2].inc and partition_alter_1[1|3].inc
# when $sub_part_no is set to >= 3.
let $sub_part_no= 3;
if (`SELECT @@session.default_storage_engine = 'ndbcluster'`)
{
  let $sub_part_no= 2;
}

# Auxiliary table used for many experiments (INSERT INTO t1 ... SELECT ...)
# on the tables to be checked
--disable_warnings
DROP TABLE IF EXISTS t0_template;
--enable_warnings
eval CREATE TABLE t0_template (
$column_list ,
PRIMARY KEY(f_int1))
ENGINE = MEMORY;
--echo #     Logging of <max_row> INSERTs into t0_template suppressed
--disable_query_log
let $num= `SELECT @max_row`;
while ($num)
{
  eval INSERT INTO t0_template
SET f_int1 = $num, f_int2 = $num, f_char1 = '$num', f_char2 = '$num',
f_charbig = '===$num===';

  dec $num;
}
--enable_query_log

# Auxiliary table used for comparisons of table definitions and file lists
--disable_warnings
DROP TABLE IF EXISTS t0_definition;
--enable_warnings
CREATE TABLE t0_definition (
state CHAR(3),
create_command VARBINARY(5000),
file_list      VARBINARY(10000),
PRIMARY KEY (state)
) ENGINE = MEMORY;

# Auxiliary table used for trigger experiments
--disable_warnings
DROP TABLE IF EXISTS t0_aux;
--enable_warnings
eval CREATE TABLE t0_aux ( $column_list )
ENGINE = MEMORY;

# Prevent that a change of defaults breaks the tests.
SET AUTOCOMMIT= 1;
SET @@session.sql_mode= '';

--echo # End of basic preparations needed for all tests
--echo #-----------------------------------------------

if (0)
{
# README for the partioning tests (t/partition_<feature>_<engine>.test)
# ========================================================================
#
# 1. Explanation of the variables to be assigned in the top-level storage engine
#    specific scripts
#------------------------------------------------------------------------------#
#
# Options, for mostly test(script+logic+result) debugging support:
#     $debug= 0 (default)
#          --> The protocolling of auxiliary stuff is suppressed.
#              The file with expected results fits to this setting.
#     $debug= 1
#          --> All executed statements will be printed into the protocol.
#              That means statements which
#              - are most time of low interest and do auxiliary stuff
#                like generating the next SQL statement to be executed
#              - additional statements giving informations about table
#                contents or the value of some variables
#                You will get huge differences, because the file with the
#                expected results was created with $debug = 0 .
#
#     $with_partitioning= 1 (default)
#          --> Do the test with really partitioned tables.
#     $with_partitioning= 0
#          --> Do not use partitioned tables. This means omit the
#              "PARTITION BY ... SUBPARTITION BY ..." part of the CREATE TABLE
#              statement. This setting has only an effect on tests where
#              partition_methods1.inc and/or partition_methods2.inc are sourced.
#
#              You will get differences when the CREATE TABLE statements
#              and table related files are printed or testcases check
#              partition borders, but most server responses and result
#              sets should be usable as reference for the test with the
#              partioned tables.
#              Please make a run with $with_partitioning= 0, whenever
#              - you do not trust the scripts (routines checking server codes/
#                result sets)
#              - fear that there is a new bug affecting partitioned and non
#                partitioned tables
#
#
# Execute the test of "table" files
#     $do_file_tests= 1 (default for
#                        - all storage engines within the extended QA test
#                        - only MyISAM within the main regression tests)
#          --> Collect the file list and compare the file list before and after
#              OPTIMIZE/REPAIR/TRUNCATE
#     $do_file_tests= 0 (default for non MyISAM storage engines within the
#                        main regression tests)
#          --> Do not collect the file list.
#     Only MyISAM has files per PARTITION/SUBPARTITION, PRIMARY KEY, INDEX, ..
#     There is a low probability that this tests detects bugs when used in
#     connection with other storage engines.
#
# Option, for displaying files:
#     $ls= 1 (default)
#          --> Display the table related directory content via
#                "ls $MYSQLTEST_VARDIR/mysqld.1/data/test/t1*"
#              if these informations were collected.
#              This is probably not portable to some OS.
#     $ls= 0
#          --> Omit displaying the directory
#
#
# Number of rows for the INSERT/UPDATE/DELETE/SELECT experiments
# on partitioned tables:
#     @max_row is the number of rows which will be filled into the auxiliary
#     MEMORY table t0_template. This table is used for INSERT ... SELECT
#     experiments. The typical test table t1 contains most of the time
#     about @max_row DIV 2 rows.
#     Too small values of @max_row should be avoided, because some statements
#     should affect several rows and partitions.
#     Too big values of @max_row should be avoided, because of runtime issues.
#     @max_row= 20 (default for the main regression tests)
#              The file with expected results fits to this amount of rows.
#     @max_row= 300 (default for extended QA test)
#          --> Use <number rows>.
#              There should be only a few systematic differences to the file
#              with expected results, because most SQL statements use @max_row and
#              variables like max_row_div2 instead of a constant with the actual
#              number of rows.
#              I assume a value of 300 rows should be
#
#
# Perform the variant with extended tests:
#     $more_trigger_tests, $more_pk_ui_tests(PK=PRIMARY KEY,UI=UNIQUE INDEX),
#       =0 (default for the main regression tests)
#           - There is a very low probability, that the omitted tests reveal a
#             bug which cannot be detected with the other tests.
#           - Limiting the partitioning tests solves issues with runtime and
#             protocol size.
#       =1 (default for extended QA test)
#
#
# Perform PRIMARY KEY specific tests:
#     $do_pk_tests= 0;
#          --> Do not execute the PRIMARY KEY related tests.
#     $do_pk_tests= 1 (default for extended QA test)
#          --> Execute the PRIMARY KEY related tests.
#     The default setting for the main regression tests depends on the
#     storage engine. The PRIMARY KEY tests must be executed for every storage
#     engine, where the existence of a PRIMARY KEY affects the kind how the
#     table rows are stored.
#     Examples for the main rgression tests:
#       InnoDB - The PRIMARY KEY is a clustered index where the data for the
#                rows are stored.       $do_pk_tests= 1
#       NDB    - The PRIMARY KEY is used for implicit partitioning (NDB).
#                                       $do_pk_tests= 1
#       MyISAM - AFAIK there is no effect on the tree containing the rows.
#                                       $do_pk_tests= 0
#
# Assign a big number smaller than the maximum value for partitions
# and smaller than the maximum value of SIGNED INTEGER
# The NDB handler only supports 32 bit integers in VALUES
#     2147483647 seems to be too big.
#     $MAX_VALUE= (2147483646);
#
#
# 2. Typical architecture of a test:
#------------------------------------------------------------------------------#
# 2.1. storage engine specific script on top level
#      (t/partition_<feature>_<engine>.test)
#      a) General not engine specific settings and requirements
#         $debug, $ls, @max_row, $more_trigger_tests, .....
#         --source inc/have_partition.inc
#      b) Engine specific settings and requirements
#         $do_pk_tests, $MAX_VALUE, $engine
#         SET SESSION default_storage_engine
#         $engine_other
#      c) Generate the prerequisites ($variables, @variables, tables) needed
#         via
#         --source inc/partition.pre
#      d) Set "fixed_bug<number>" variables to 1 if there are open engine
#         specific bugs which need worarounds.
#      e) Execute the feature specific testscript via
#         --source inc/partition_<feature>.inc
#      f) Perform a cleanup by removing all objects created within the tests
#         --source inc/partition_cleanup.inc
#
# 2.2. script generating the prerequisites needed in all tests
#      (inc/partition.pre)
#      a) Message about open bugs causing that
#         - some testcases are disabled
#         - it cannot be avoided that the file with expected results suffers
#           from open bugs
#           This should not occur often !
#           Example: There is extreme often an auxiliary testscript sourced,
#                    but the the conditions vary. We get under a certain combination
#                  of conditions a wrong result set or server response.
#      b) Set "fixed_bug<number>" variables to 0 if there are open engine
#         specific bugs. They are later set to 1 within the toplevel script.
#         Set "fixed_bug<number>" variables to 1 if there are open NOT engine
#         specific bugs.
#      c) Setting of auxiliary variables
#      d) Creation of auxiliary tables ....
#
# 2.3. script checking a feature
#      (inc/partition_<feature.inc>.inc)
#      Example:
#      a) "set/compute" a CREATE TABLE t1 .. and an ALTER TABLE ... statement
#      b) CREATE TABLE t1 ...
#      c) INSERT INTO t1 (.....) SELECT .... FROM t0_template WHERE ...
#         The first 50 % of all t0_template rows will be inserted into t1.
#      d) ALTER TABLE t1 (Example: ADD/DROP UNIQUE INDEX)
#      e) INSERT INTO t1 (.....) SELECT .... FROM t0_template WHERE ...
#         The second 50 % of all t0_template rows will be inserted into t1.
#      Now t1 and t0_template should have the same content.
#      f) Check the "usability" of the current table t1
#         via
#         --source inc/partition_check.pre
#      g) DROP TABLE t1
#      Switch to other CREATE and ALTER statements and run sequence a)-g) again
#      ...
#
# 2.4. script checking if a certain table shows the expected behaviour
#      ("usability" check):   inc/partition_check.inc
#      - SELECT/INSERT/UPDATE/DELETE affecting single and multiple records
#      - check of values of special interest like NULL etc.
#      - INSERT/UPDATE with BEFORE/AFTER triggers
#      - violations of UNIQUE constraints, if there are any defined
#      - transactions ...
#      - TRUNCATE/OPTIMIZE/..
#      - ...
#
#
# 2.5. There are some auxiliary scripts with sub tests where we cannot predict
#      if we get an error and if we get one, which one.
#      Example: INSERT a record where the value for a certain column equals
#         some existing record.
#         Depending on existing/not existing PRIMARY KEYs, UNIQUE INDEXes
#         the response might be "no error", ER_DUP_KEY, ER_DUP_ENTRY.
#      Our requirements:
#      1. We cannot abort whenever get an error message from the server.
#      2. We want the exact server message into the protocol.
#      3. We want abort testing if we know that a certain error must not happen.
#      Common but unusable Solutions:
#      a) --error 0, ER_DUP_KEY, ER_DUP_ENTRY
#         <statment>
#         We get no error message even if the statement fails.
#      b) --error ER_DUP_KEY, ER_DUP_ENTRY
#         <statment>
#         We might get "got one of the expected errors".
#         There are situations where the statement must be successful.
#      c) --disable_abort_on_error
#         <statment>
#         --enable_abort_on_error
#         And nothing extra
#         We do not abort in case of unexpected server errors.
#
#      Final solution:
#         --disable_abort_on_error
#         <statment>
#         --enable_abort_on_error
#         Check via error number if the error is not totally unexpected.
#         The sub tests use $ER_DUP_KEY, $ER_DUP_ENTRY, etc.
#         Assignment of values happen in this file.
#
#
# 3. How to analyze a partitioning bug revealed with these tests/ How to build
#    a small replay script from the monstrous protocols ?
#------------------------------------------------------------------------------#
# a) crash    -- use the file var/mysqld.1/data/mysql/general_log.CSV
# b) no crash, but unexpected server response (there is no "reject file)
#             -- use the file r/<testcase>.log
#                Please be aware that the option $debug= 0 suppresses the
#                protocolling of some queries.
# c) no crash, but unexpected result set
#             -- use the file r/<testcase>.reject
#                Please be aware that the option $debug= 0 suppresses the
#                protocolling of some queries.
# In most cases you will find that the r/<testcase>.<log/reject> contains at
# least a line "#       # check <something>:       0".
# That means that a check within inc/partition_check did not got the
# expected result.
# A good start for a replay script would be
#   1. Copy t/partition_<feature>_<engine>.test to t/my_test.test
#   2. Edit t/my_test.test
#      - set $debug to 1
#      - replace the line
#        "--source inc/partition_<feature>.inc"
#        with all statements between the last
#        CREATE TABLE t1 statement (included this)
#        and the line
#        "# Start usability test (inc/partition_check.inc)"
#      - add the content of inc/partition_check.inc at the end.
#
# Please excuse that the partitioning tests generate such huge protocols which
# and are not very handy when it comes to bug analysis. I tried to squeez out
# as much test coverage as possible by writing some hopefully smart routines
# and reusing them in various combinations.
#
# Matthias
#
}

Man Man