The present invention relates generally to launch systems, and more particularly to an architecture for such systems.
The Mk 41 Vertical Launching System (VLS) is a canister launching system that provides a rapid-fire launch capability against hostiles. The US Navy currently deploys MK 41 VLS on AEGIS-equipped Ticonderoga-class cruisers and Spruance- and Arleigh Burke-class destroyers, and plans to use it on next generation of surface ships. Additionally, MK 41 VLS is the choice of eight other international navies, including Canada, Japan, Germany, Turkey, Spain, Netherlands, Australia and New Zealand.
The basic element of the MK 41 VLS is an eight-cell launcher module. Each module is a complete, standalone dual-redundant launcher. Each module includes a launch control system, gas management system, missile canisters, ballistic deck & hatches with deluge and sprinklers, and walkways. Electronic equipment mounted on the 8-Cell Module monitors the stored missile canisters and the module components and assists in launching the missiles. Modules can be combined to form launchers tailored in size to meet individual combatant mission requirements. For example, The MK 41 VLS is currently deployed at sea in 13 different configurations, ranging from a single module with eight cells to a system having 16 modules with 128 cells.
Three components are required for firing a missile using the MK 41 VLS: a Weapon Control System (WCS), a Launch Control Unit (LCU), and a Launch Sequencer (LS). This architecture is depicted in
Weapon Control System 100 is the man-machine interface for the MK 41 VLS weapons system.
Launch Sequencer 104, which is a part of the eight-cell launcher module, is the communication link between the upstream fire control systems and the missile itself.
Launch control unit 102, which is part of the eight-cell launcher module, maintains simultaneous interfaces with the various weapon control systems to provide simultaneous multi-mode launch coordination and reports inventory and launcher status. During normal operations, each Launch Control Unit 102 controls half of Launch Sequencers 104 in the launcher module. But if one of Launch Control Units 102 is offline, the other assumes control of all Launch Sequencers 104 in the launcher.
Launch Control Unit 102 contains a software component called the “Launch Control Computer Programs” (“LCCP”). This software component supports communication with Weapons Control System 100 over two NTDS serial communication lines, one for each direction. The LCCP support two-way communications with Launch Sequencer 104 over a redundant Ethernet LAN.
When a launch order is given, Weapons Control System 100 sends a signal to one of two parallel Launch Control Units 102 (only one of which is depicted) in each eight-cell launcher module. The Launch Control Unit then issues pre-launch and launch commands for the selected missile. Launch Sequencer 104 responds to the commands (issued by Launch Control Unit 102) by preparing the eight-cell missile module and missiles for launch and then launching the selected missile.
The MK-41 VLS has been in production since 1982 and is continually upgraded to incorporate new technology. Yet, in the MK-41 VLS, as in most launch systems, there is a dependency or fixed operational relationship between Weapons Control System 100 and Launcher Sequencer 104. This dependency arises from the use of proprietary and non-open protocols and services, as provided by Launch Control Unit 102.
As a consequence of the fixed operational relationship between the Weapons Control System and the Launch Sequencer, the architecture of the launch system is not flexible. That is, it is not scaleable for single cell launchers or other variations. It will support only one type of launch system (e.g., the MK41 VLS, etc.) and is platform dependent (i.e., operating system and processor). This limits the ability to incorporate new technologies into the MK-41 VLS and, to the extent that such integration is even possible, substantially complicates the integration process.
What is needed, therefore, is launching-system architecture that has a flexible framework that enables it to adapt to different launch systems, weapon control systems, and the like.
The present invention provides a scalable and distributable architecture for use in conjunction with various different weapons control systems and launch systems. The architecture eliminates the need, for example, for the Launch Control Unit in the MK-41 VLS.
The inventive architecture, which is software-based, discards the proprietary and non-open protocols and services that characterize the typical Launch Control Unit and replaces them with open source adaptive and middleware components. The inventive architecture is structured to expose potential points of variation. That is, the architecture is structured to avoid paths (i.e., hardware or software solutions) that, once implemented, dictate downstream or upstream systems and components.
In the illustrative embodiment of the invention, the inventive architecture is implemented as a Launch Control Module that separates different layers of responsibility within a prior-art launch control unit (e.g., launch control unit 102, see
In the illustrative embodiment of the invention, the layers of responsibility are separated by defining three “layers” within the Launch Control Module, including:
The Launch Controller manages a logical grouping of weapon systems or launch sequencers (see,
The Module Controller manages multiple groupings of Cell Controllers as well as their interdependent hardware components. The Module Controller is also responsible for rules regarding safety and other issues related to the Cell Controllers.
The Cell Controller oversees missile-specific sequence control and interface with the launch hardware. Cell safety is managed at this layer as well. Types of Cell Controllers are configured as a function of physical missile types and their number in the system.
In some embodiments, the Launch Control Module is not distributed; it is hosted in a single hardware device (e.g., processor, etc.). For example, in some embodiments, the Launch Control Module is hosted by a single processor within the launch sequencer hardware. In some other embodiments, in particular those in which fault tolerance is a concern, it is desirable to provide a replica of the Launch Control Module on a second processor. The replica can be hosted by another processor within the launch sequencer hardware or on any hardware platform that is accessible to the weapons-system network. This can be done because the Launch Control Module disclosed herein is platform independent.
Platform independence is provided through the use of commercial off-the-shelf open-source adaptive and distribution middleware that provides services for platform independence and relocateability. A proxy pattern is employed to support independent interfaces. New and legacy interfaces, protocols, and communications schemes are handled at this level. A data distribution service model is used to communicate weapon availability to networked clients and redundant Launch Controllers and Module Controllers.
In some other embodiments, the Launch Control Module is distributed. In these embodiments, one or more of the “components” (e.g., the layers, or portions of the layers, etc.) of the Launch Control Module are deployed on more than one hardware device. Distribution of functionality is desirable for fault tolerance and in applications in which a particular component of the Launch Control Module is more closely associated with another subsystem (e.g., the WCS, etc.).
In addition to the three layers of the Launch Control Module, a Data Store component is defined. The Data Store supports the data distribution service model for remote monitoring and to support fault tolerance. Furthermore, the Data Store maintains configuration, mode, and availability/status information concerning the launching system.
The Launch Control Module includes one or more of the following capabilities, including the ability to:
As a consequence of its flexible and distributable nature, the architecture described herein solves a number of legacy problems that are associated with launch systems, such as operating system and network dependencies, the tight coupling of software components, and the high costs of adding new launcher capabilities. The flexibility provided by the inventive architecture enables a launch system to accept new capabilities without impacting existing behavior and performance requirements.
The following terms are defined for use in this Specification, including the appended claims:
Launch Controller 312 manages a logical grouping of weapon systems or launch sequencers. The Launch Controller further acts as the focal point for redundancy in a fault tolerant architecture/application, as required (see, e.g.
Module Controller 314 manages multiple groupings of Cell Controllers 316 as well as their interdependent hardware components. Module Controller 314 is also responsible for rules regarding safety and other issues related to the Cell Controllers 316.
Cell Controller 316 oversees missile-specific sequence control and interface with the launch hardware. Cell safety is managed at this layer as well. Various types of Cell Controllers 316 are configured as a function of physical missile types and number in the system.
Data Containers 320 are objects of information that are exchanged between two other components on the diagram (e.g., between Weapons Control System Proxy 310 and Launch Controller 312, etc.). In fact, in some embodiments, all communications within the inventive architecture use data containers. In some other embodiments, Data Containers are replaced with method calls using an RPC or client-server mechanism. The flow of data in the Data Containers is shown in the directed lines.
Data Store 322 supports the Distribution Middleware feature of a Data Distribution Service (DDS). In some alternative embodiments, data store 322 is replaced with a commercial off-the-shelf or Object Management Group (OMG) compliant service. No lines of communication are depicted between Data Store 322 and other components for the sake of clarity. In fact, Data Store 322 receives registration requests (subscriptions) and publications from many of the components of Launch Control Module 202 (e.g., Launch Controller 312, Module Controller 314, Cell Controller 316, etc.), as needed. Data Store 322 then sends instances of the published data to all subscribers. This is the Data Distribution Model.
Weapon Control System Proxy 310 supports two-way communications between Launch Control Module 202 and Weapon Control System 200. Launch Sequencer Proxy 318 performs the same role for the communications with Launch Sequencer 204.
The first is between Weapon Control System IDSIM 330 and Weapon Control System Proxy 310, which in some embodiments communicate over an Ethernet interface. The second is between Launch Sequence Proxy 318 and Launch Sequencer IDSIM 332. It is notable that “IDSIM” is a simulation of those components indicated and can be substituted for test purposes. The third interface shows the relationship between Launch Control Module Monitor 334 and Data Store 322. The Launch Control Module Monitor uses the Data Distribution Service to subscribe and receive data published by other components within Launch Control Module 202 (e.g., Launch Controller 312, Module Controller 314, Cell Controllers 316-1, 316-2, etc.).
It is notable that the dependency relationship between some of the legacy components in
Regarding items that have not previously been described, simulation controller 530 provides a simulation of the cell control functionality at the sub-launch level. This capability is used for upper layer validation and training. This component, like the other elements, can be allocated to any network processor. Simulation controller 530 is invoked by the Launch Control 312 when commanded by Weapon Control System 200. Simulation controller 530 is an optional component; in some embodiments it is included and in some other embodiments it is not. This could be performed statically or with dynamic composition.
In the “Framework/Infrastructure” layer, software package entitled main 534 provides a common service that is required on most operating systems. Variation from one operating system to another for initiation of the application is performed via this package.
Generic Control 532 is a package of software components in the “Framework/Infrastructure” layer. This package provides a common set of services required by all controllers in the architecture. It provides the pattern for implementing a controller and is the point of variation required when underlying operating system services require a change. This package isolates those changes from the application component in the next higher layer.
The layer called “LCCP Legacy Components” shows one embodiment of the architecture wherein some components are reused from the existing Launch Control Computer Program (LCCP) in the prior art. In some embodiments, this is a transitory path wherein message validation occurs.
Key features of the embodiment of Launch Control Module 202 that is depicted in
The Launcher Electronics that are depicted in
Event Services 852 is a software package that provides for an exchange of information between two systems, typically upon a change in state or an “event.” This is often used as a generic term, but actually originates from the OMG CORBA specification. In a more recent version of the OMG specification, which is based on the “publish-subscribed” paradigm, event services are replaced with the Data Distribution Service (DDS).
It is to be understood that the above-described embodiments are merely illustrative of the present invention and that many variations of the above-described embodiments can be devised by those skilled in the art without departing from the scope of the invention. For example, in this Specification, numerous specific details are provided in order to provide a thorough description and understanding of the illustrative embodiments of the present invention. Those skilled in the art will recognize, however, that the invention can be practiced without one or more of those details, or with other methods, materials, components, etc.
Furthermore, in some instances, well-known structures, materials, or operations are not shown or described in detail to avoid obscuring aspects of the illustrative embodiments. It is understood that the various embodiments shown in the Figures are illustrative, and are not necessarily drawn to scale. Reference throughout the specification to “one embodiment” or “an embodiment” or “some embodiments” means that a particular feature, structure, material, or characteristic described in connection with the embodiment(s) is included in at least one embodiment of the present invention, but not necessarily all embodiments. Consequently, the appearances of the phrase “in one embodiment,” “in an embodiment,” or “in some embodiments” in various places throughout the Specification are not necessarily all referring to the same embodiment. Furthermore, the particular features, structures, materials, or characteristics can be combined in any suitable manner in one or more embodiments. It is therefore intended that such variations be included within the scope of the following claims and their equivalents.
This case claims priority of U.S. Provisional Patent Application Ser. No. 60/778,764, which was filed on Mar. 3, 2006 and is incorporated by reference herein in its entirety.
Number | Date | Country | |
---|---|---|---|
60778764 | Mar 2006 | US |