Step 1

arghtalentData Management

Jan 31, 2013 (4 years and 8 months ago)

264 views





DS/AS REPLACEMENT

STEP 1

CRITICAL DESIGN REVIEW

8/10/04

Tim Hopkins

NWS SYSTEMS ENGINEERING CENTER

2

Deirdre Jones


Opening Remarks


3

Objectives


Critical Design Review for Step 1


Preliminary Design Review for entire plan


Will follow up with PDR and CDR for Step 2, etc.

-
Each Design Review will include an overview of entire plan with
progress and milestones.


30
-
50 field sites deployed per Step 2, by Feb. 15, 2005


Schedule accelerated at direction of Corporate Board, to
enhance system performance b/4 next severe weather
season.


4

Team Organization and
Responsibilities


Tim Hopkins (OST31)


Project Lead


Chuck Piercy (OST11)


Program Management and
Budget


Walter Scott (OST11)


COTR


Dave Holloran (NGIT)


Project Manager


Leigh Dominy (NGIT)


Project Engineer


Franz Zichy/Karthik Srinivasan (OST31/OPS12)


Maintenance and Logistics


Mary Buckingham (OPS24)


Operational Assessment
Test Director


Carl McCalla/Ed Mandell (OST31)


Software
Development and Integration Coordinator




5

Chuck Piercy


Program Management and Budget




6

AWIPS System Architecture

NEXRAD

GOES

ASOS

NCEP Weather Models

Buoys,

River Gauges,

Lightning, etc.

Terrestrial and
Satellite
Communications
Systems

Network Control Facility

AWIPS

Firewall

Retrieval,
Processing &
Storage
Servers

Workstations

Workstations

Workstations

Workstations

LAN

7

Legacy Architecture

PLAINTREE

SWITCH

HP

DS

HP

AS

HP

WS

100 Mbps FDDI

HP
-
RT

CP

SIMPACT

NEXRAD

RPG

FIREWALL

LINUX

WS

LDAD

Server

10 Mbps

Ethernet

10 Mbps

Ethernet

WAN

ROUTER

SIMPACT

Total Processing Capacity: Approx. 900 MFLOPS

8

Linux Phase I Architecture

HP

DS

HP

AS

HP

WS

100 Mbps FDDI

LINUX

CP

NEXRAD

RPG

FIREWALL

LINUX

WS

LDAD

Server

10 Mbps

Ethernet

10 Mbps

Ethernet

WAN

LINUX

PX

ROUTER

100

100

100

PLAINTREE

SWITCH

10/100/1000 Mbps

Ethernet Switch

SIMPACT

Phase I Enhancements

Total Processing Capacity: Approx. 3750 MFLOPS

9

Linux Phase II Target Architecture

ROUTER

LINUX

WS

NEXRAD

RPG

WAN

LINUX CLUSTER(s)

Ethernet 1000/100/10 Mbps SWITCH

LINUX

PX

AS

REPLACE
-

MENT

DB

SERVER

LDAD

FW

LDAD

SERVER

LINUX

CP

Total Processing Capacity: Approx. 16,000+ MFLOPS

NEXRAD

RPG

AHPS

Equip


SBN IP

Multicast

WFO/RFC

Archive

“Local Apps” Server CLUSTER

10

Expected WFO/RFC Hardware
Architecture ~July ’05

DS1

HP
-
UX10.2

Plaintree

Switch

Ethernet 1000/100/10 Mbps SWITCH

LDAD

Firewall

LINUX

PX1

LINUX

XT

LINUX

SX1

Serial MUX

PV

LINUX

DX1

LINUX

DX2

Network Attached

Storage (NAS)

Linux

CP

Router

WAN

AWIPS

WAN

Linux

AX

LDAD

SERVER

RFC

REP

Nexrad

RPG

< 1 yr old

1
-
2 yr old

2
-
3 yr old

3+ yr old

Legend

100Mbps FDDI

DS2

HP
-
UX10.2

LINUX

LX

11

Operational Concept



Same operational concept, with enhanced computing
resources additional storage space and centralized,
shared mass storage using Network Attached Storage
device.

12

DS/AS Replacement

Roadmap




Leigh Dominy (NGIT)


13


Overview


An analysis of technology and development of a “to
-
be” architecture
was initially briefed to the AWIPS Systems Engineering Team (SET)
on 2/3/04.


To
-
be architecture provides the framework to complete individual
product improvement tasks while maintaining a common goal.


X
-
terminals


Redundant LDAD firewalls


DS/AS replacement

Redundant LDAD servers


Serial mux upgrade

Full DVB deployment


To
-
be architecture facilitates development of a roadmap for schedule,
budget and deployment planning.


Roadmap assists in issue and dependency identification and resource
planning for risk reduction.

14

AWIPS To
-
Be Architecture


Hardware

-
Utilize Network Attached Storage (NAS) technology

-
Deploy commodity servers on GbE LAN

-
Incrementally deployed and activated

-
Promote reuse of select hardware

-
Remove limitations of direct attached storage


Software

-
For availability, move from COTS solution to use of public
domain utilities


Some experience with NCF and REP


Can be decoupled from operating system upgrades


Supports NAS environments


Can be augmented for load balancing if required

-
Deploy low cost Linux database engine


15

AWIPS To
-
Be Architecture


Methodology

-
Stage hardware, move software when ready

-
Reuse hardware to ease transition and allow planned
decommissioning

-
Deploy flexible availability framework


Goals

-
Decommission AS, AdvancedServer, DS and FDDI

-
Support universal deployment of single Linux distribution
(currently RHE3.0 with OB6)

-
Provide dedicated resource for local applications

-
Expandable within framework (easy to add servers)

-
Deploy to subset of sites prior to February 2005


16

Step 1


Initial Hardware Deployment


Release OB4 is a prerequisite


Initial staging of hardware

-
New rack

-
NAS w/LTO
-
2 tape (~400GB storage and backup)

-
2 commodity servers


DX1/DX2

-
2 GbE switches and associated cables, etc

-
2 8
-
port serial mux replacements (installed in PX1/2)


NAS key to incremental deployment and activation


Serial mux replacements installed but not

activated


LTO
-
2 drive for site backup


New hardware and PX1/PX2 on GbE LAN


LDAD firewall upgrade deployed independently

17

Step 1 and Release OB4

Hardware Architecture

DS1


Ethernet 1000/100/10 Mbps SWITCH

LDAD

Firewall

PX1

LINUX

XT

PX2

Serial MUX

PV

DX1

DX2

Network Attached

Storage (NAS)

Linux

CP

Router

WAN

AWIPS

WAN

Linux

AX

LS1

Nexrad

RPG

100Mbps FDDI

DS2

LINUX

LX

AS1

AS2

MassStorage

Benefits


Solves Active/Active PX problem


DS Mass Storage and AutoLoader can be decommissioned

Issues


Will plaintree swtch handle nfs traffic from DS/AS to NAS?


Does site have footprint for another rack before any old ones are decommissioned?


What is the file structure of the NAS?


Do most decoders use internal storage for temp files?


What kind of NCF monitoring will be available for the NAS?

plaintree

Light Blue or Clear boxes indicates new hardware component

7/30/04

18

Step 1 and Release OB4

Software and Data Architecture

DS1


PX1

PX2

PV

LINUX

DX1

LINUX

DX2

NAS

WAN

DS2

AS1

AS2


Decoders


Radar Processing


Dial Radar


wfoAPI


RMRserver


Radar Server


Radar Text
Decoder


WWA Processes


Database Engine


TexDB read/write


OH cron/apps


Shef decoder


MTA/MHS
processes


LDAD server


LDAD routers


Listener


Local ldad
applications


NWWS Product


LAMP


Netmetrix


Trap
Interceptor
(ITO)


asynchScheduler


Decoders


Metar


Synioptic


RAMOS


Redbook


RadarStorage


DNS


NIS


NTP


JetAdmin


SNMP Agent


NWWS Scheduler


NotificationServer


DamCat(OH)


LAPS


Data Storage


Site Backup


IFP


GFE


Decoders


Satellite


Grib


BinLightinig


Maritime


Profiler


Web Server


SafeSeas


CommsRouter


FFMP/SCAN


MSAS


Decoders


BufMosDecoder


WarnDB


StdDB


Collective


Raob


Aircraft


ACARS


Bufr Drivers


textNotificationserver


DNS Phase I


DNS Phase I

Bold Underlined text indicates newly ported or moved processes or data

Light Blue or Clear boxes indicate new hardware component

Processes and/or data are listed underneath corresponding hardware component

7/30/04

19

Step 1
-

Incremental Activation


Activate NAS

-
Point DS to NAS


Allows DS1/DS2 to be used as active/active pair using existing
MC/ServiceGuard infrastructure


Decommission DS mass storage and autoloader

-
Point PX1/PX2 to NAS


Maintain active/active pair


Deliver new availability mechanism


De
-
activate cluster management portion of AS2.1


PowerVault (PV) deactivated for near
-
term


Connect PX1/PX2 to GbE LAN and install PCI
-
X cards as future AS serial mux
replacement once APS port is complete


Most data to the NAS, temporary files continue to be written to local disk


Planning early October OAT to verify DS1/DS2 and PX1/PX2 failover and
NAS data availability


Require a full deployment decision in mid
-
October to complete deployment of
initial 64 sites by 1 Feb 05




20

Step 1
-

Considerations


Current plan is to move data
from DS mass storage and
PX PowerVault to NAS

-
Limited to 2GB tar files

-
Some sites will need to
clean up px1data and
px2data prior to NAS
installation


Rack placement at WFOs

-
Early sites will have
“extra” rack longer
than sites deployed
later in 2005

NAS1
10 x 72 GB Raw Disks (RAID 4)
Disks (10): Root/Data (7) + Parity (2) + Hot Spare (1)
Qtree Name
Mount point
Quota (GB)
WFO
Quota (GB)
RFC
/vol0/.snapshot
-- 10 % --
41.6
41.6
/vol0/home
/home
2.0
8.0
/vol0/data_fxa
/data/fxa
60.0
60.0
/vol0/data_x400
/data/x400
0.3
0.6
/vol0/awips_GFESuite
/awips/GFESuite
35.0
0
/vol0/awips_adapt
/awips/adapt
1.0
0
/vol0/data_adapt
/data/adapt
2.5
0
/vol0/awips_hydroapps
/awips/hydroapps
1.5
55.0
/vol0/awips_gis
/awips/gis
0
3.0
/vol0/awips_dev
/awips/dev
0.3
0.3
/vol0/data_local
/data/local
9.0
1.5
/vol0/DS_shared
/DS_shared
1.4
1.4
Total Space Used
154.6
171.4
WFO
RFC
Low level Format Capacity (GB per disk)
66.0
66.0
WAFL requirement of 10% (GB per disk)
6.6
6.6
Raw Root/Data Available Capacity (GB)
415.8
415.8
Allocated Capacity (GB)
154.6
171.4
Unallocated Capacity (GB)
261.2
244.4
Network Attached Storage
21

Step 2
-

Decommission AS1/AS2


Release OB4 maintenance release (OB4.X) of stable OB5 software
required for AS decommissioning


Activate DX1/DX2

-
Infrastructure/decoders move to DX1

-
IFP/GFE to DX2

-
Newly ported functionality to DX1


Reuse PX1 and PX2 as PX1 and SX1

-
PX1 for applications/processes using processed data

-
SX1 for Web Server, local applications and eventually LDAD

-
Activate serial mux replacement and ported APS


Non
-
ported software from AS1 to DS2


DX1/DX2 deployed with RH 7.2 as risk reduction


22

Step 2 and OB4.AS/OB5 (w/RH7.2)

Hardware Architecture

DS1


Ethernet 1000/100/10 Mbps SWITCH

LDAD

Firewall

PX1

LINUX

XT

SX1

Serial MUX

PV

DX1

DX2

Network Attached

Storage (NAS)

Linux

CP

Router

WAN

AWIPS

WAN

Linux

AX

LS1

Nexrad

RPG

100Mbps FDDI

DS2

LINUX

LX

Benefits


Could stop maintenance on AS’s


Could eliminate 1 rack of equipment


Could decommission FDDI at this point if 100MB boards were cheaper than maintenance.


Start migration of MHS

Issues


Start running X.400 and SMTP in parallel.


Should PX and SX be retrofitted with extra disks?

plaintree

Light Blue or Clear boxes indicates new hardware component

7/30/04

23

Step 2 and OB4.AS/OB5 (w/RH7.2)

Software and Data Architecture

DS1


PX1

SX1

PV

DX1

DX2

NAS

WAN

DS2


Radar Processing


Dial Radar


wfoAPI


RMRserver


Radar Server


WWA Processes


Database Engine


TexDB read/write


OH cron/apps


Shef decoder


MTA/MHS
processes


LDAD server


LDAD routers


Listener


Local ldad
applications


NWWS Product


LAMP


Decoders


Metar


Synoptic


RAMOS


Netmetrix


Trap
Interceptor
(ITO)


Data Storage


Site Backup



FFMP/SCAN


SafeSeas


LAPS


MSAS


Text
notificationserver


asynchScheduler


NWWS
Scheduler


Notification
server


DamCrest(OH)**


Web Server


routerStoreNetcdf**


Decoders


BinLightning


Satellite


Grib


Maritime


WarnDB


StdDB


Collective


BufMostDecoder


Raob


Aircraft


ACARS


Profiler


Redbook


RadarStorage


RadarMsgHandler


RadarTextDecoder


HandleGeneric


Bufr Drivers


CommsRouter


DNS


NIS


NTP


JetAdmin
replacement


SNMP Agent


SMTP MTA/MHS


Xyplex hosting




IFPServer


GFE

Bold Italicized text indicates moved processes Bold and underline indicates newly ported processes

Light Blue or Clear boxes indicate new hardware component

Processes and/or data are listed underneath corresponding hardware component

7/30/04

*Serial MUX is an AS Decommissioning dependency

**Requires prototype testing

24

Step 2
-

Incremental Activation


Failover scheme

-
DX1 to DX2

-
DX2 to DX1

-
PX1 applications and servers to DX2

-
PX1 processes APS and NWWSProduct to SX1 (require PCI
-
X
card access)

-
SX1 baseline software to PX1


Decommission AS1/AS2 and excess rack


Linux SMTP MTA deployed as start of migration from x.400 (required
for DS decommission)


Provide sites with some level of

performance improvement as early as
software readiness allows

25

Step 3


Deploy Data Base and OS


Deploy Linux PostgreSQL data base engine to DX1

-
Move existing PV to new rack and connect to DX1/2

-
Reconfigure PV (possibly into 2 separate direct attached disk
farms, one for each DX)

-
Database availability via mirrored or replicated databases is TBD
at this time

-
May be accelerated to OB5 for fxatext database


Migrate ported databases


Upgrade operating system (currently RHE3.0) on all applicable hosts



LX/XT


CP (if full DVB deployment complete)


DX


AX


PX/SX


RP (RFCs only)

26

Step 3 and OB6 (w/RHE3.0)

Hardware Architecture

DS1


Ethernet 1000/100/10 Mbps SWITCH

LDAD

Firewall

PX1

LINUX

XT

SX1

Serial MUX

DX1

DX2

Network Attached

Storage (NAS)

Linux

CP

Router

WAN

AWIPS

WAN

Linux

AX

LS1

Nexrad

RPG

100Mbps FDDI

DS2

LINUX

LX

Benefits


Start Migration of Databases

Issues


When will apps be ready to use PostgreSQL?


Data reliability with PostgreSQL


Will there be issues with running Informix and PostgreSQL in parallel?


What is the performance of PostgreSQL?

plaintree

Light Blue or Clear boxes indicates new hardware component

7/30/04

PV

PV

27

Step 3 and OB6 (w/RHE3.0)

Software and Data Architecture

DS1


PX1

SX1

DX1

DX2

NAS

WAN

DS2


Radar Processing


Dial Radar


wfoAPI


RMRserver


Radar Server


WWA Processes


Informix Database
Engine


LDAD server


LDAD routers


Listener


Local ldad
applications


NWWS Product


Decoders


Metar


Synoptic


RAMOS


Netmetrix


Trap
Interceptor
(ITO)


Data Storage


Site Backup



FFMP/SCAN


SafeSeas


LAPS


MSAS


Text
notificationserver


asynchScheduler


NWWS Scheduler


Notification server


DamCrest(OH)**


LAMP


Web Server


routerStoreNetcdf**


ldadMon


PostgreSQL
Database Engine


OH cron/apps


Shef decoder


TextDB read/write


Decoders


BinLightning


Satellite


Grib


Maritime


WarnDB


StdDB


Collective


BufMostDecoder


Raob


Aircraft


ACARS


Profiler


Redbook


RadarStorage


RadarMsgHandler


RadarTextDecoder


HandleGeneric


Bufr Drivers


CommsRouter


DNS


NIS


NTP


JetAdmin
replacement


SNMP Agent


SMTP MTA/MHS


Xyplex hosting


MHS Processes




IFPServer


GFE

Bold and underline indicates newly ported processes

Light Blue or Clear boxes indicate new hardware component

Processes and/or data are listed underneath corresponding hardware component

7/30/04

PV

PV


PostgreSQL Database

28

Step 3


Incremental Activation


Decommission AS 2.1 and HP Informix


HP Informix engine can remain on DS for site use


Transition to SMTP and decommission x.400


Consideration
-

Can/should PostgreSQL be delivered early to RH7.2
DXs?

-
Most sites run software against GFS databases, not their own
databases.

-
Database and software will be tested with RHE3 only as part of
OB6.

-
If databases delivered early how/when does

parallel ingest get
developed and tested?

-
Should RFCs be handled differently?

29

Step 4


Deploy LDAD Upgrade


Deploy LS1 and LS2

-
Deploy redundant server pair (requirements still tbd)

-
Could reuse PX1/SX1 as LS1/LS2 and use new generation
hardware for PX1/SX1


Activate LS1/LS2

-
Migrate internal LDAD processing to SX1

-
Some internal and external LDAD processing must transition at
same time


Reuse existing HP LS on internal LAN

-
Existing 10/100 MB LAN card

30

Step 4 and OBx

Hardware Architecture

DS1


Ethernet 1000/100/10 Mbps SWITCH

LDAD

Firewall

PX1

LINUX

XT

SX1

Serial MUX

DX1

DX2

Network Attached

Storage (NAS)

Linux

CP

Router

WAN

AWIPS

WAN

Linux

AX

Nexrad

RPG

100Mbps FDDI

DS2

LINUX

LX

Benefits


Move old LS1 to internal LAN

Issues


Transition from LS to LSx needs to consider how/when local ldad apps will be done.


Do we run old LS and new LSx in parallel while sites convert their local software?


Int/Ext Ldad software must be ready to transition at the same time.


Could reuse PX1 and SX1 for LSx1/2 and buy higher end machines for PX1 and SX1


plaintree

Light Blue or Clear boxes indicates new hardware component

7/30/04

PV

PV

LSx1

LS1

31

Step 4 and OBx

Software and Data Architecture

DS1


PX1

SX1

DX1

DX2

NAS

WAN

DS2


Radar Processing


Dial Radar


wfoAPI


RMRserver


Radar Server


WWA Processes


NWWS Product


Decoders


Metar


Synoptic


RAMOS


Netmetrix


Trap
Interceptor
(ITO)


Data Storage


Site Backup



FFMP/SCAN


SafeSeas


LAPS


MSAS


Text
notificationserver


asynchScheduler


NWWS Scheduler


Notification server


LAMP


DamCrest(OH)**


Web Server


routerStoreNetcdf**


ldadMon


LDAD Server


LDAD routers


Listener


Local ldad applications


PostgreSQL
Database Engine


OH cron/apps


Shef decoder


TextDB read/write


Decoders


BinLightning


Satellite


Grib


Maritime


WarnDB


StdDB


Collective


BufMostDecoder


Raob


Aircraft


ACARS


Profiler


Redbook


RadarStorage


RadarMsgHandler


RadarTextDecoder


HandleGeneric


Bufr Drivers


CommsRouter


DNS


NIS


NTP


SNMP Agent


SMTP MTA/MHS


MHS Processes


JetAdmin
replacement


Xyplex hosting





IFPServer


GFE

Bold and underline indicates newly ported processes

Light Blue or Clear boxes indicate new hardware component

Processes and/or data are listed underneath corresponding hardware component

7/30/04

PV

PV


PostgreSQL Database

LSx1


LDAD
Server
Applications

32

Step 5


Decommission DS1/DS2


Continue to move ported software/databases to Linux servers

-
Combine with step 6 if additional DXs are required

-
DX1 for database server and infrastructure

-
DX2 for IFP/GFE

-
DXn for decoders


All Linux devices on GbE LAN


Remaining functionality on LS on 100MB LAN

-
DialRadar/wfoAPI (tied to FAA/DoD requirements) may be OBE
at this point. If so,

Simpacts and LS can be decommissioned at all
non
-
hub sites.

-
Netmetrix (required at hub sites only)


33

Step 5 and OBx

Hardware Architecture

Ethernet 1000/100/10 Mbps SWITCH

LDAD

Firewall

PX1

LINUX

XT

SX1

Serial MUX

DX1

DX2

Network Attached

Storage (NAS)

Linux

CP

Router

WAN

AWIPS

WAN

Linux

AX

Nexrad

RPG

LINUX

LX

Benefits


FDDI and DS decommissioned.


Reuse of LS for SW that isn’t proted yet. (non
-
redundant)


Rack consolidation

Issues


Dialup and wfoAPI may still be needed on an HP
-
UX machine


Netmetrix at HUB sites needs to be on HP
-
UX


Light Blue or Clear boxes indicates new hardware component

7/30/04

PV

PV

LSx1

LS1

Not Upgraded


Printers


Xyplex


Modems


VIR’s


Simpacts

DX3

(optional)

34

Step 5 and OBx

Software and Data Architecture

PX1

SX1

DX1

DX2

NAS

WAN

LS1


Radar Processing


wfoAPI


RMRserver


Radar Server


WWA Processes


NWWS Product


Decoders


Metar


Synoptic


RAMOS


Data Storage


Site Backup


Trap Interceptor (ITO)



FFMP/SCAN


SafeSeas


LAPS


MSAS


Text
notificationserver


asynchScheduler


NWWS Scheduler


Notification server


LAMP


DamCrest(OH)**


Web Server


routerStoreNetcdf**


ldadMon


LDAD Server


LDAD routers


Listener


Local ldad applications


PostgreSQL
Database Engine


OH cron/apps


Shef decoder


TextDB read/write


Decoders


BinLightning


Satellite


Grib


Maritime


WarnDB


StdDB


Collective


BufMostDecoder


Raob


Aircraft


ACARS


Profiler


Redbook


RadarStorage


RadarMsgHandler


RadarTextDecoder


HandleGeneric


Bufr Drivers


CommsRouter


DNS


NIS


NTP


SNMP Agent


SMTP MTA/MHS


MHS Processes


JetAdmin
replacement


Xyplex hosting





IFPServer


GFE

Bold and underline indicates newly ported processes

Light Blue or Clear boxes indicate new hardware component

Processes and/or data are listed underneath corresponding hardware component

7/30/04

PV

PV


PostgreSQL Database

LSx1


LDAD
Server
Applications


DialRadar


Netmetrix

DX3

(optional)

35

Step 6


Process Loading


Incrementally add DX hosts for load
balancing for new functionality and data
sets.

-
DX1 for database server and infrastructure

-
DX2 for IFP/GFE

-
DXn for decoders

36

Step 6 and OBx

Hardware Architecture

Ethernet 1000/100/10 Mbps SWITCH

LDAD

Firewall

PX1

LINUX

XT

SX1

Serial MUX

DX1

DX2

Network Attached

Storage (NAS)

Linux

CP

Router

WAN

AWIPS

WAN

Linux

AX

Nexrad

RPG

LINUX

LX

Light Blue or Clear boxes indicates new hardware component

7/30/04

PV

PV

LSx1

LS1

Not Upgraded


Printers


Xyplex


Modems


VIR’s


Simpacts

DX3

(optional)

37

Step 6 and OBx

Software and Data Architecture

PX1

SX1

DX1

DX2

NAS

WAN

LS1


Radar Processing


wfoAPI


RMRserver


Radar Server


WWA Processes


NWWS Product


Data Storage


Site Backup


Trap Interceptor (ITO)



FFMP/SCAN


SafeSeas


LAPS


MSAS


Text
notificationserver


asynchScheduler


NWWS Scheduler


Notification server


LAMP


DamCrest(OH)**


Web Server


routerStoreNetcdf**


ldadMon


LDAD Server


LDAD routers


Listener


Local ldad applications


PostgreSQL
Database Engine


OH cron/apps


Shef decoder


TextDB read/write


RadarStorage


RadarMsgHandler


RadarTextDecoder


HandleGeneric


DNS


NIS


NTP


SNMP Agent


SMTP MTA/MHS


MHS Processes


JetAdmin
replacement


Xyplex hosting





IFPServer


GFE

Bold and italicized indicates newly ported processes

Light Blue or Clear boxes indicate new hardware component

Processes and/or data are listed underneath corresponding hardware component

7/30/04

PV

PV


PostgreSQL Database

LSx1


LDAD
Server
Applications


DialRadar


Netmetrix

DX3

(optional)


Decoders


BinLightning


Satellite


Grib


Maritime


WarnDB


StdDB


Collective


BufMostDecoder


Raob


Aircraft


ACARS


Profiler


Metar


Synoptic


Redbook


RAMOS


Bufr Drivers


CommsRouter




38

High Level Schedule


Key dates


-
Step 1 Deployment Decision by 21 October
-
necessary to complete initial deployment by 1 Feb

-
Step 2 OB4.X


will provide performance improvements to Step 1 sites

-
OB6 check
-
in




Task
OB5 Check-in
notificationServer
Step 1-OAT
Step 1-Initial Deploy
*
Deploy Decision
OB5 Beta
Step 2-OB4.x Test

**Accelerate if possible
Step 2-OB4.x-Deploy
Step 1- Full Deploy
OB5 Full Deployment
Step 3-DB Availability
Step 3- RHE3 procure
OB6 code check-in
OB6 Beta
Jul-05
Aug-05
Sep-05
Mar-05
Apr-05
May-05
Jun-05
Nov-04
Dec-04
Jan-05
Feb-05
Jul-04
Aug-04
Sep-04
Oct-04
39

DX/NAS OAT

Mary Buckingham

OPS24

8/10/04

40

DX/NAS OAT


OAT begins about 9/30


Need for early hardware deployment decision (by mid Oct)

-
Risk hardware issues might not surface in so short a time

-
Can’t stress hardware much within these constraints

-
Possible a severe problem surfaces in operations
-

would affect
entire schedule


Will take 60 days to finish OAT to ensure installation or
software issues are found and fixed


Combine with Serial MUX Replacement OAT

41


Preliminary Mod Note proof on NMTW


12 Test sites


1 RH

(1st site, safest)


1 RFC


1 NCEP


9 WFO


OB4 Prerequisite (Begin 9/23)
-

Will affect pace of installations


Conflicts with VTEC and Watch by County ORDs (8/30
-
10/13)


Sites selected based on:

-
History of poor performance and risk in severe wx

-
Cover the spectrum of site types

-
Space for new rack

DX/NAS OAT

42


Risk Factors

-
Expect similar installation issues as encountered in PX install

-
Repointing all databases to new location (NAS)

-
Unknown local mounts

-
Rush on decisions and limited sites mean risk of not finding install or operational
issues until deployment


Unknown how much relief step 1 will give sites

-
Faster I/O should help but more problems might lurk


OCONUS sites: unknown whether standard system will help their severe
problems

-
Additional engineering might be required

-
OAT will give info on whether more is needed

DX/NAS OAT

43

DS/AS Replacement Project

Step 1

Software Development and Integration


No software development needed for Step 1


44

DS/AS Replacement Project

Step 1

Maintenance and Logistics


Franz Zichy/Karthik Srinivasan
(OST31/OPS12)


Maintenance and
Logistics


45

Maintenance and Logistics Concept


Integrated logistic support


Maintenance and logistics concept


Maintenance policy and procedure


Disposal Procedure


M&C of NAS, DX1 and DX2


46

Integrated Logistics Support


Supply Support


Determine LRU


DX1, DX2 cpu’s


NAS


Ge Switch


Miscellaneous (cables, manuals, etc)

-
Assign Agency Stock Numbers (ASN)

-
Catalog [(p/n, price, manufacturer)]

47

Integrated Logistics Support


OPS12 provides TIP (Technical Information Package)

-
Inform field


Schedule


Technical specifications


Maintenance concept


Logistic concept


NGIT will provide NRC with DX1/2 Checkout procedures


NRC receives spare DX1/2servers

48

Integrated Logistics Support


failed component tested.

-
Vendor notified of failed component


Components divided into LRUs and placed into
stock at NLSC.

-
servers

-
switches


49

Maintenance Concept

Linux device faults
Hardware
Failure?
Software
Failure?
Ask site to order
new Linux device
from CLS
Network
Failure?
Call NCF
open trouble
ticket
Can it be repaired with
NGIT/SST resources?
Yes
Linux Hardware Resolution Flowchart
Is it external to the
device?
Yes
Use NGIT
resources to repair
Yes
End
Close Trouble
Ticket
Yes
Does new
device work
Yes
Continue
Troubleshooting
No
A
A
Does device
work
No
Yes
Open DR/provide
work around until
next MR
No
No
No
Use NGIT/SST
resources to repair
Yes
No
50

Maintenance Concept


Linux device fails

-
Open trouble ticket with NCF


Hardware


Network


Software

-
Hardware failure


Order new from CLS

-
Network failure


If external, use NGIT resources
to repair


If internal, order new from CLS

51

Maintenance Concept


Software failure

-
Open DR and/or provide work around until next MR

-
OS only in step 1


Linux hardware failure

-
Order new LRU from CLS

-
Once received, ship defective unit back in same container

-
Call NCF

-
Install new device

-
Site works with NCF to recover/configure device


52

Disposal of Existing Equipment


No disposal in Step 1





53

Monitor and Control


M&C of DX1 and DX2 will be through
Xyplex


M&C of NAS will be through Xyplex

54

Security



Complying or will comply with all existing
pertinent security policies, directives and
rules


No known issues with Step 1

55

Installation Coordination Issues


Direct Ship from staging site in Chantilly to sites


New Rack Requires Space and Power

-
Similar to REP


Maintenance & Logistics issues

-
Spares for infant failures

-
Deployment support

56

AWIPS DS/AS Replacment Step 1

Major Schedule Milestones

Initial SET Brief

PECP delivered

Task Authorized

Critical Design Review(CDR)

FMK Development

Mod Note Complete

Operational Assessment Test

Full Deployment


Feb04 Apr04 Jun04 Aug04 Sep04 Oct04
-
Feb05

8/10

9/28

8/30

9/29

1/6/05

2/16/05

10/6

10/29

9/28

6/23

2/3

4/26

57

Risks


NAS Single point of failure


Old HP’s and FDDI to new GE and NAS


Re
-
engineer of failover


58

Action Item Review





59

Deirdre Jones


Closing Remarks