Rechercher dans le manuel MySQL
22.5.13.1 NDB Cluster Disk Data Objects
NDB Cluster Disk Data storage is implemented using the following objects:
Tablespace: Acts as containers for other Disk Data objects. A tablespace contains one or more data files and one or more undo log file groups.
Data file: Stores column data. A data file is assigned directly to a tablespace.
Undo log file: Contains undo information required for rolling back transactions. Assigned to an undo log file group.
log file group: Contains one or more undo log files. Assigned to a tablespace.
Undo log files and data files are actual files in the file
system of each data node; by default they are placed in
ndb_
in
the node_id
_fsDataDir
specified in the NDB
Cluster config.ini
file, and where
node_id
is the data node's node
ID. It is possible to place these elsewhere by specifying either
an absolute or relative path as part of the filename when
creating the undo log or data file. Statements that create these
files are shown later in this section.
Undo log files are used only by Disk Data tables, and are not
needed or used by NDB
tables that are stored
in memory only.
NDB Cluster tablespaces and log file groups are not implemented as files.
Although not all Disk Data objects are implemented as files,
they all share the same namespace. This means that
each Disk Data object must be uniquely
named (and not merely each Disk Data object of a given type).
For example, you cannot have a tablespace and a log file group
both named dd1
.
Assuming that you have already set up an NDB Cluster with all nodes (including management and SQL nodes), the basic steps for creating an NDB Cluster table on disk are as follows:
Create a log file group, and assign one or more undo log files to it (an undo log file is also sometimes referred to as an undofile).
Create a tablespace; assign the log file group, as well as one or more data files, to the tablespace.
Create a Disk Data table that uses this tablespace for data storage.
Each of these tasks can be accomplished using SQL statements in the mysql client or other MySQL client application, as shown in the example that follows.
We create a log file group named
lg_1
usingCREATE LOGFILE GROUP
. This log file group is to be made up of two undo log files, which we nameundo_1.log
andundo_2.log
, whose initial sizes are 16 MB and 12 MB, respectively. (The default initial size for an undo log file is 128 MB.) Optionally, you can also specify a size for the log file group's undo buffer, or permit it to assume the default value of 8 MB. In this example, we set the UNDO buffer's size at 2 MB. A log file group must be created with an undo log file; so we addundo_1.log
tolg_1
in thisCREATE LOGFILE GROUP
statement:To add
undo_2.log
to the log file group, use the followingALTER LOGFILE GROUP
statement:Some items of note:
The
.log
file extension used here is not required. We employ it merely to make the log files easily recognizable.Every
CREATE LOGFILE GROUP
andALTER LOGFILE GROUP
statement must include anENGINE
option. The only permitted values for this option areNDBCLUSTER
andNDB
.ImportantThere can exist at most one log file group in the same NDB Cluster at any given time.
When you add an undo log file to a log file group using
ADD UNDOFILE '
, a file with the namefilename
'filename
is created in thendb_
directory within thenode_id
_fsDataDir
of each data node in the cluster, wherenode_id
is the node ID of the data node. Each undo log file is of the size specified in the SQL statement. For example, if an NDB Cluster has 4 data nodes, then theALTER LOGFILE GROUP
statement just shown creates 4 undo log files, 1 each on in the data directory of each of the 4 data nodes; each of these files is namedundo_2.log
and each file is 12 MB in size.UNDO_BUFFER_SIZE
is limited by the amount of system memory available.See Section 13.1.16, “CREATE LOGFILE GROUP Syntax”, and Section 13.1.6, “ALTER LOGFILE GROUP Syntax”, for more information about these statements.
Now we can create a tablespace—an abstract container for files used by Disk Data tables to store data. A tablespace is associated with a particular log file group; when creating a new tablespace, you must specify the log file group it uses for undo logging. You must also specify at least one data file; you can add more data files to the tablespace after the tablespace is created. It is also possible to drop data files from a tablespace (see example later in this section).
Assume that we wish to create a tablespace named
ts_1
which useslg_1
as its log file group. We want the tablespace to contain two data files, nameddata_1.dat
anddata_2.dat
, whose initial sizes are 32 MB and 48 MB, respectively. (The default value forINITIAL_SIZE
is 128 MB.) We can do this using two SQL statements, as shown here:- CREATE TABLESPACE ts_1
- ADD DATAFILE 'data_1.dat'
- USE LOGFILE GROUP lg_1
- INITIAL_SIZE 32M
- ENGINE NDBCLUSTER;
- ALTER TABLESPACE ts_1
- ADD DATAFILE 'data_2.dat'
- INITIAL_SIZE 48M;
The
CREATE TABLESPACE
statement creates a tablespacets_1
with the data filedata_1.dat
, and associatests_1
with log file grouplg_1
. TheALTER TABLESPACE
adds the second data file (data_2.dat
).Some items of note:
As is the case with the
.log
file extension used in this example for undo log files, there is no special significance for the.dat
file extension; it is used merely for easy recognition.When you add a data file to a tablespace using
ADD DATAFILE '
, a file with the namefilename
'filename
is created in thendb_
directory within thenode_id
_fsDataDir
of each data node in the cluster, wherenode_id
is the node ID of the data node. Each data file is of the size specified in the SQL statement. For example, if an NDB Cluster has 4 data nodes, then theALTER TABLESPACE
statement just shown creates 4 data files, 1 each in the data directory of each of the 4 data nodes; each of these files is nameddata_2.dat
and each file is 48 MB in size.CREATE TABLESPACE
statements must contain anENGINE
clause; only tables using the same storage engine as the tablespace can be created in the tablespace. ForALTER TABLESPACE
, anENGINE
clause is accepted but is deprecated and subject to removal in a future release. ForNDB
tablespaces, the only permitted values for this option areNDBCLUSTER
andNDB
.For more information about the
CREATE TABLESPACE
andALTER TABLESPACE
statements, see Section 13.1.21, “CREATE TABLESPACE Syntax”, and Section 13.1.10, “ALTER TABLESPACE Syntax”.
Now it is possible to create a table whose unindexed columns are stored on disk using files in tablespace
ts_1
:- )
- TABLESPACE ts_1 STORAGE DISK
- ENGINE NDBCLUSTER;
TABLESPACE ts_1 STORAGE DISK
tells theNDB
storage engine to use tablespacets_1
for data storage on disk.Once table
ts_1
has been created as shown, you can performINSERT
,SELECT
,UPDATE
, andDELETE
statements on it just as you would with any other MySQL table.It is also possible to specify whether an individual column is stored on disk or in memory by using a
STORAGE
clause as part of the column's definition in aCREATE TABLE
orALTER TABLE
statement.STORAGE DISK
causes the column to be stored on disk, andSTORAGE MEMORY
causes in-memory storage to be used. See Section 13.1.20, “CREATE TABLE Syntax”, for more information.
You can obtain information about the NDB
disk
data files and undo log files just created by querying the
FILES
table in the
INFORMATION_SCHEMA
database, as shown here:
For more information and examples, see Section 25.11, “The INFORMATION_SCHEMA FILES Table”.
Indexing of columns implicitly stored on disk.
For table dt_1
as defined in the example
just shown, only the dob
and
joined
columns are stored on disk. This is
because there are indexes on the id
,
last_name
, and
first_name
columns, and so data belonging
to these columns is stored in RAM. Only nonindexed columns can
be held on disk; indexes and indexed column data continue to
be stored in memory. This tradeoff between the use of indexes
and conservation of RAM is something you must keep in mind as
you design Disk Data tables.
You cannot add an index to a column that has been explicitly
declared STORAGE DISK
, without first changing
its storage type to MEMORY
; any attempt to do
so fails with an error. A column which
implicitly uses disk storage can be
indexed; when this is done, the column's storage type is
changed to MEMORY
automatically. By
“implicitly”, we mean a column whose storage type
is not declared, but which is which inherited from the parent
table. In the following CREATE TABLE statement (using the
tablespace ts_1
defined previously), columns
c2
and c3
use disk storage
implicitly:
- -> c4 INT
- -> )
- -> STORAGE DISK
- -> TABLESPACE ts_1
- Query OK, 0 rows affected (1.31 sec)
Because c2
, c3
, and
c4
are themselves not declared with
STORAGE DISK
, it is possible to index them.
Here, we add indexes to c2
and
c3
, using, respectively, CREATE
INDEX
and ALTER TABLE
:
- Query OK, 0 rows affected (2.72 sec)
- Query OK, 0 rows affected (0.92 sec)
SHOW CREATE TABLE
confirms that
the indexes were added.
- *************************** 1. row ***************************
- Table: ti
You can see using ndb_desc that the indexed columns (emphasized text) now use in-memory rather than on-disk storage:
shell> ./ndb_desc -d test t1
-- t1 --
Version: 33554433
Fragment type: HashMapPartition
K Value: 6
Min load factor: 78
Max load factor: 80
Temporary table: no
Number of attributes: 4
Number of primary keys: 1
Length of frm data: 317
Max Rows: 0
Row Checksum: 1
Row GCI: 1
SingleUserMode: 0
ForceVarPart: 1
PartitionCount: 4
FragmentCount: 4
PartitionBalance: FOR_RP_BY_LDM
ExtraRowGciBits: 0
ExtraRowAuthorBits: 0
TableStatus: Retrieved
Table options:
HashMap: DEFAULT-HASHMAP-3840-4
-- Attributes --
c1 Int PRIMARY KEY DISTRIBUTION KEY AT=FIXED ST=MEMORY
c2 Int NULL AT=FIXED ST=MEMORY
c3 Int NULL AT=FIXED ST=MEMORY
c4 Int NULL AT=FIXED ST=DISK
-- Indexes --
PRIMARY KEY(c1) - UniqueHashIndex
i2(c3) - OrderedIndex
PRIMARY(c1) - OrderedIndex
i1(c2) - OrderedIndex
NDBT_ProgramExit: 0 - OK
Performance note. The performance of a cluster using Disk Data storage is greatly improved if Disk Data files are kept on a separate physical disk from the data node file system. This must be done for each data node in the cluster to derive any noticeable benefit.
You can use absolute and relative file system paths with
ADD UNDOFILE
and ADD
DATAFILE
; relative paths are calculated with respect
to the data node's data directory. You can also use
symbolic links to point to a directory on a different disk; see
Section 22.5.13.2, “Using Symbolic Links with Disk Data Objects”, for more
information and examples.
A log file group, a tablespace, and any Disk Data tables using these must be created in a particular order. This is also true for dropping these objects, sibject to the following constraints:
A log file group cannot be dropped as long as any tablespaces use it.
A tablespace cannot be dropped as long as it contains any data files.
You cannot drop any data files from a tablespace as long as there remain any tables which are using the tablespace.
It is not possible to drop files created in association with a different tablespace other than the one with which the files were created.
For example, to drop all the objects created so far in this section, you can use the following statements:
- mysql> DROP LOGFILE GROUP lg_1
These statements must be performed in the order shown, except
that the two ALTER TABLESPACE ... DROP
DATAFILE
statements may be executed in either order.
Document created the 26/06/2006, last modified the 26/10/2018
Source of the printed document:https://www.gaudry.be/en/mysql-rf-mysql-cluster-disk-data-objects.html
The infobrol is a personal site whose content is my sole responsibility. The text is available under CreativeCommons license (BY-NC-SA). More info on the terms of use and the author.
References
These references and links indicate documents consulted during the writing of this page, or which may provide additional information, but the authors of these sources can not be held responsible for the content of this page.
The author This site is solely responsible for the way in which the various concepts, and the freedoms that are taken with the reference works, are presented here. Remember that you must cross multiple source information to reduce the risk of errors.