Skip to main content

MySQL NDB Cluster Installation for Data, SQL and Management Node

Management node: It manages the other nodes within the NDB Cluster, provides configuration data, start and stop nodes, running backups. As it manages the configuration of the other nodes, it should be started first, before any other node. Use ndb_mgmd to start it.
It also maintains a log of cluster activities. Management clients can connect to the management server and check the cluster's status.
Data node: It stores cluster data. There are as many data nodes as there are replicas, times the number of fragments. For example, with two replicas, each having two fragments, you need four data nodes. One replica is good for data storage, but no redundancy. For redundancy and high availability 2 more nodes are recommended.To start data node use ndbd or ndbmtd. Tables are normally stored completely in memory, that's why it can be refer as in-memory database. Some data can be stored on disk.
SQL node: It accesses the cluster data. It is a traditional MySQL server that uses the NDBCLUSTER storage engine. An SQL node is a mysqld process started with the --ndbcluster and --ndb-connect string options. It is API node, which designates any application which accesses NDB Cluster data. Another example of an API node is the ndb_restore utility that is used to restore a cluster backup. 
Note: Use multiple data and SQL nodes. The use of multiple management nodes is also highly recommended for production environment, redundancy. 

SQL Node:- Installation of MySQL cluster
Following command will install mysql cluster in directory /usr/local
$ tar -C /usr/local -xzvf mysql-cluster-gpl-7.5.7-linux2.6.tar.gz
Create link
$ ln -s /usr/local/mysql-cluster-gpl-7.5.7-linux2.6-i686 /usr/local/mysql

Set up the system databases using mysqld
$ cd mysql
$ mysqld --initialize

Set the necessary permissions for the MySQL server and data directories:
$ chown -R root .
$ chown -R mysql data
$ chgrp -R mysql .

Copy the MySQL startup script to the appropriate directory, make it executable, and set it to start when the operating system is booted up:
$ cp support-files/mysql.server /etc/rc.d/init.d/
$ chmod +x /etc/rc.d/init.d/mysql.server
$ chkconfig --add mysql.server

Data nodes:-
Data nodes does not require the mysqld binary. Only the NDB Cluster data node executable ndbd (single-threaded) or ndbmtd (multi-threaded) is required. These binaries can also be found in the .tar.gz archive.

Install the Data node binaries:-

Change location to the /var/tmp directory, and extract the ndbd and ndbmtd binaries from the archive into a suitable directory such as /usr/local/bin:
$ cd /var/tmp
$ tar -zxvf mysql-5.7.18-ndb-7.5.7-linux-i686-glibc23.tar.gz
$ cd mysql-5.7.18-ndb-7.5.7-linux-i686-glibc23
$ cp bin/ndbd /usr/local/bin/ndbd
$ cp bin/ndbmtd /usr/local/bin/ndbmtd
Change location to the directory into which you copied the files, and then make both of them executable:
$ cd /usr/local/bin
$ chmod +x ndb*
The preceding steps should be repeated on each data node host.
Note:-The data directory on each machine hosting a data node is /usr/local/mysql/data. This piece of information is essential when configuring the management node. 

Management nodes:-

It does not require the mysqld binary. Only the NDB Cluster management server (ndb_mgmd) is required; you most likely want to install the management client (ndb_mgm) as well. Both of these binaries also be found in the .tar.gz archive.
Extract the ndb_mgm and ndb_mgmd from the archive into a suitable directory such as /usr/local/bin:
$ cd /var/tmp
$ tar -zxvf mysql-5.7.18-ndb-7.5.7-linux2.6-i686.tar.gz
$ cd mysql-5.7.18-ndb-7.5.7-linux2.6-i686
$ cp bin/ndb_mgm* /usr/local/bin
Make both of them executable:
$ cd /usr/local/bin
$ chmod +x ndb_mgm*

Comments

Popular posts from this blog

Oralce Pluggable Database

Pluggable database administration:-

Open all pluggable databases:
Connect to container database using oraenv
$. oraenv=DB_Name
$sqlplus "/as sysdba"
SQL> alter pluggable database all open;
Get pluggable database status:
SQL> select con_id, dbid, name, open_mode from v$pdbs;
Switch to Puggable database from container :
SQL> alter session set container = <Puggable_DB>;
Get status of Pluggable database
SELECT v.name, v.open_mode, NVL(v.restricted, 'n/a') "RESTRICTED", d.status
FROM v$pdbs v, dba_pdbs d WHERE v.guid = d.guid ORDER BY v.create_scn;
Get file location for pluggable database:
SQL>select file_name from dba_data_files;
Create table space for pluggable datatbase:
SQL> alter session set container=<Pluggable_DB>;
Session altered.
SQL> CREATE TABLESPACE TB_Name LOGGING DATAFILE '/<Path>/<File_Name>.dbf' SIZE 1024M REUSE AUTOEXTEND ON NEXT 8192K MAXSIZE UNLIMITED EXTENT MANAGEMENT LOCAL UNIFORM SIZE 1M;
Set new t…

Oracle Daily SQL

-- Enable System-wide tracing:
alter system set events '10046 trace name context forever,level 12';

-- Disabled in all sessions
alter system set events '10046 trace name context off';

-- Session Tracing
alter session set tracefile_identifier='10046';
alter session set timed_statistics = true;
alter session set statistics_level=all;
alter session set max_dump_file_size = unlimited;
alter session set events '10046 trace name context forever,level 12';
-- crsctl for Oracle RAC

./crsctl start cluster
./crsctl check crs

- Make column to CLOB

ALTER TABLE TABLE_NAME MODIFY (COLUMN_NAME LONG);
ALTER TABLE TABLE_NAME MODIFY (COLUMN_NAME CLOB);
RENAME <TABLE_NAME> TO <NEW_TABLE_NAME>;

- Rename column, change data type, drop column

ALTER TABLE TABLE_NAME RENAME COLUMN COLUMN_NAME TO COLUMN_NAME1;
ALTER TABLE TABLE_NAME MODIFY (COLUMN_NAME VARCHAR2(10));
ALTER TABLE TABLE_NAME DROP COLUMN COLUMN_NAME;

-- Create Primary Key

ALTER TABLE TABLE_NAME ADD C…

Oracle Golden Gate Basic Classic

Create environment
01. Source & target - Create GG tablespace.
02. Source & target - Create the GoldenGate Schema Owner
03. Grant privilege including DBA to GG schema owner
04. Add schema owner to global parameter file ./GLOBALS
05. Execute role set up script to create role GGS_GGSUSER_ROLE
06. Grant role GGS_GGSUSER_ROLE to GG user at both Source and target
Configure GG  Extract
01. Source and Target - Configure Manager Parameters
$ ggsci
GGSCI 1> EDIT PARAMS MGR
PORT 7809
DYNAMICPORTLIST 7810-7820
02. Source, create parameter file for Extract - ex1,
EXTRACT ex1
USERID <id>, PASSWORD <pwd>
EXTTRAIL /home/oracle/goldengate/dirdat/ex
TABLE <Schema>.*;
03. Source, configure Data Pump Parameters
EXTRACT dp1
USERID <id>, PASSWORD <pwd>
RMTHOST <hostname>, MGRPORT 7809
RMTTRAIL /home/oracle/goldengate/dirdat/rt
TABLE <schema>.*;
Target - Create check point table
$ ggsci
GGSCI 1> DBLOGIN USERID <id>, PASSWORD <pwd>
GGSCI 2>…