Dear Student,
Greetings from www.etraining.guru.
We provide THE BEST IBM Netezza DBA/Development training in the Industry by THE real time expert. Our Trainer comes with 10+ years of IT working experience on open systems of project team sizes ranging from 2 to 70+, in which 5+ years is on IBM Netezza/Bigdata.
Course Highlights:
****************
(1) 24 hours of course content:
DBA: http://etraining.guru/online-training-ibm-netezza-puredata-dba.aspx
Development: http://etraining.guru/online-training-ibm-puredata-netezza-development.aspx
(2) 2 free classes on Unix & Shell scripting of topics of student's interest
(3) 6 hours of Free job support. Beyond these hours, job support is provided for a nominal charge of $10/hour
(4) Course fees: $350 (or) 20000 INR
(5) Class Timing: Weekends. Next batch is starting on 21/Feb/2015 8:00 A.M IST. We also provide adhoc training on weekdays based upon trianers availability.
(6) Email: info@etraining.guru
Please feel free to let us know if any further details needed.
Thanks,
Bhavana
www.etraining.guru
Email:info@etraining.guru
3. Single chassis, single host system
Appliance for data warehouse test and development environments
It’s a non High-Availability (HA) system
Shares the same architecture and software as IBM Netezza 1000
(Twinfin) model
5. Purpose built, standards based data warehouse appliance that
architecturally integrates database, server, storage and advanced
analytics capabilities into a single, easy-to-manage system
It supports High Availability (HA)
Fast load speeds: > 4TB/Hour
Fast backup rates: > 4TB/Hour
1:1:1 ratio between disks, CPU’s, FPGA’s
10. Update to the IBM Netezza 1000 model, with same architecture
On 09/Oct/2012, IBM announced the name changes from TwinFin
(1000) to PureData System for Analytics, as below:
1:1:1 ratio between Disks, CPU’s, FPGA’s
13. C1000 models are similar to N1001 systems, but with more storage/rack
Scales to more than 10 petabytes of user data capacity
Each Netezza C1000 rack has one S-Blade chassis that contains 4 S-blades. Each S-blade
has 8 CPU/FPGA processors. Same as N1001
1 Rack = 4 Storage Arrays (or Storage Groups)
1 Storage Array = 1 disk raid controller + 2 disk enclosures
1 disk raid controller = 12 disks; 1 disk enclosure = 12 disks
So, 1 Storage array/group = 12 + (2*12) = 36 disks
So, 1 Rack = 4 Storage array/groups = 4 * 36 disks = 144 TB
Note: 2 spare disks per each storage array
C1000-4:
1 Rack,
1 S-Blade Chassis 8 CPU, 8 FPGA
4 Storage groups 144 TB (8 Spares)
C1000-8:
2 Racks
2 S-Blade Chassis 16 CPU, 16 FPGA
8 Storage groups 288 TB (16 Spares)
Note: There is no 1:1:1 ratio between Disks, CPU’s, FPGA’s
16. 3x faster analytics performance & 50% more usable capacity per rack
128 GB/Sec scan rate
1:1:1 ration between disks, FPGA engines, and CPU cores doesn’t apply
1 Rack = 7 S-blades + 288 disks
1 S-blade = 16 CPU cores + 16 FPGA engines
288 disks = 240 active disks + 34 spare + 14 used for swap/log space
Note: Each disk size in striper: 600 GB (user space:200GB, Mirror:200GB, Temp: 200GB)
Why no 1:1:1 ratio?
There is a 1:1 ratio between FPGA:CPU.
There is no 1:1 ratio between FPGA and disks because FPGA can
process
the data much faster than the disk can produce.
2 S-blades have 40 disks; 5 S-blades have 32 disks
Striper configurations: ½ rack, 1 rack, 2 rack, or 4 rack.
We no longer have a ¼ rack configuration as we did with the Twinfin
22. When the system starts up, 32 or 40 dataslices are assigned to each s-blade
At no point during operation can one s-blade access the data on a dataslice which has been assigned to another s-
blade
If an s-blade fails, the dataslices which were assigned to the failed s-blade will be rebalanced and assigned to the
remaining still operational s-blades. That is exactly the same as the system worked in later releases of NPS on the
TwinFin architecture as well.
There in no attachment of CPUs to disks. The only attachment is that dataslice is assigned to 1 S-blade when the
system starts. CPU and FPGA resources are assigned as they are available.
When a query starts, NPS will start 240 processes, one for each dataslice.
The processes start reading data off disk; As each page of data (128KB) comes off disk, that page gets assigned to
the first available FPGA on that s-blade.
FPGA decompresses and filters data and passes the result back to the process.
The Linux CPU scheduler assigns the process to one of the CPUs which processes the remaining data that came out
of the FPGA.
Once complete, the next 128KB is read off disk and that continues until all of the data has been processed for the
table being scanned.
24. option1: nz_get_model
[/export/home/nz]$nz_get_model
IBM PureData System for Analytics N2001-010
option2: select * from _t_environ where name like ‘%NPS%’
/export/home/nz->nzsql -c “select * from _t_environ where name like ‘NPS%’;”
NAME | VAL
————–+————
NPS_PLATFORM | xs
NPS_MODEL | P1000X_A_E
NPS_FAMILY | Pseries
Note: Pseries is for Twinfin; Qseries is for Striper
The 14 disks are dedicated for partitions to store swap/log space required by the linux operating system that runs on the S-Blades, this has been separated out from the disks containing the partitions with user data primary/mirror and the database temp space, in the former TwinFin architecture all of these partitions were spread across all the disks, i.e. they weren't separated out.