Survey
* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
IDM UID 333K4C VERSION CREATED ON / VERSION / STATUS 08 Feb 2013 / 2.0/ Approved EXTERNAL REFERENCE How To Guideline for Fast Controllers, I/O Bus Systems and Communication Methods between Interconnected Systems This is the ITER I&C guideline to Fast Controllers. It is part of the Plant System Design Handbook (27LH2V) documentation set. The document summarizes the various input and output (I/O) bus systems and communication methods between interconnected I/O systems. Potential alternatives are listed. A recommendation is made for the selection of specifications of both electrical and communications protocols to be used in ITER Instrumentation and Control Systems. A short list of suitable systems and ... Author CoAuthor Reviewers Approver Read Access Name Makijarvi P. Approval Process Action 08-Feb-2013:signed Affiliation IO/DG/DIP/CHD/CSD/CDC Wallander A. Thomas P. 11-Feb-2013:recommended IO/DG/DIP/CHD/CSD 24-Mar-2013:approved IO/DG/DIP/CHD Document Security: level 1 (IO unclassified) RO: Makijarvi Petri AD: ITER, AD: External Collaborators, AD: Division - Control System Division - EXT, AD: Section CODAC - EXT, AD: Section - CODAC, project administrator, RO, LG: CODAC Fast Controllers, LG: PCDH Fast Controllers, LG: CODAC team PDF generated on 24-Mar-2013 DISCLAIMER : UNCONTROLLED WHEN PRINTED – PLEASE CHECK THE STATUS OF THE DOCUMENT IN IDM Change Log Title (Uid) Versio n Latest Status Issue Date Description of Change Guideline for Fast Controllers, I/O Bus Systems and Communication Methods between Interconnected Systems (333K4C_v2_0) v2.0 Approved 08 Feb 2013 For PCDH v7 Guideline for Fast Controllers, I/O Bus Systems and Communication Methods between Interconnected Systems (333K4C_v1_3) v1.3 Approved 05 Jan 2011 for PCDH 6.0 - after internalreview and with corrections from the proof reader (J . Poole) Guideline for Fast Controllers, I/O Bus Systems and Communication Methods between Interconnected Systems (333K4C_v1_2) v1.2 Approved 01 Feb 2010 After PCDH 5.1 review, approved modifications done, links updated. Guideline for Fast Controllers, I/O Bus Systems and Communication Methods between Interconnected Systems (333K4C_v1_1) v1.1 Signed 14 Dec 2009 Layout settings, IDM numbers and references corrected. Guideline for Fast Controllers, I/O Bus Systems and Communication Methods between Interconnected Systems (333K4C_v1_0) v1.0 In Work 07 Dec 2009 PDF generated on 24-Mar-2013 DISCLAIMER : UNCONTROLLED WHEN PRINTED – PLEASE CHECK THE STATUS OF THE DOCUMENT IN IDM 1 EXECUTIVE SUMMARY The ITER Organization has selected PCI Express as the base I/O bus technology for use in ITER Instrumentation and Control System fast controllers. o The use of Conventional PCI (32-bit parallel bus I/O devices) and all of the spin-offs (CompactPCI, PXI, ..) are facilitated by the backward compatibility of PCI Express. The main recommended items of the ITER Instrumentation and Control System fast controller catalog are the PXI, Compact PCI and PXI Express I/O boards. A PXI Express chassis is used to carry the I/O boards. The PXI Express I/O chassis is separate from the CPU chassis, which will be a standard industrial computer. o The two systems are interconnected using a PCI Express link. o The industrial computer contains CPU, disk, memory and network connectivity. o The configuration with separate I/O chassis and CPU/Network minimizes the ties to single vendors (vendor lock-in) and simplifies maintenance. The usage of PXI Express chassis and PXI I/O boards creates a strong relationship with National Instruments and other members of the PXI Systems Alliance. By using an external CPU/Network unit, the vendor lock-in is relaxed for the rapidly changing part of the hardware. By running Linux and EPICS on external CPU/Network units the ITER Organization emphasizes its position of being against the use of proprietary software solutions on embedded hardware. Performance upgrades will be easier because the I/O chassis will never be touched; only the industrial computer with CPU/Networking will be replaced, thus avoiding modifications and manipulation of the I/O cabling. xTCA (µTCA and AdvancedTCA) are proposed as the platform for diagnostics which require high-end electronics design. This type of diagnostic application is not well covered by the CODAC Catalog of Commodity off the Shelf (COTS) items. System integration services will be essential to configure and build I&C systems with fast controllers and they should be offered to all industrial partners by the ITER Organization. o o It would be unrealistic to suppose that all plant system manufacturers would be able to purchase, build and configure fast controller systems to comply with ITER CODAC requirements entirely by themselves. The ITER Organization should not only make catalogs but have industrial partners who would be able to propose system integration services to plant system manufacturers. 1 Table of contents 1 EXECUTIVE SUMMARY..................................................................................................................1 2 INTRODUCTION ............................................................................................................................4 3 2.1 PCDH context ...........................................................................................................................4 2.2 Objectives ................................................................................................................................5 2.3 Scope of the Document ............................................................................................................5 2.4 Acronyms .................................................................................................................................5 2.5 References ...............................................................................................................................6 REVIEW OF BUS TECHNOLOGIES ...................................................................................................7 3.1 VMEbus systems ......................................................................................................................7 3.2 PCI – Peripheral Component Interconnect ................................................................................7 3.3 PCI in Industry and in Science ...................................................................................................9 3.3.1 3.3.2 3.3.3 3.3.4 3.4 PCI Express in Industry and in Science.....................................................................................13 3.4.1 3.4.2 3.4.3 3.4.4 3.4.6 4 PICMG 1.0 and 1.1.....................................................................................................................9 CompactPCI ...............................................................................................................................9 PXI 10 PCI Express ..............................................................................................................................11 PICMG 1.3................................................................................................................................13 CompactPCI Express, CompactPCI Plus and CompactPCI PlusIO.............................................14 PXI Express ..............................................................................................................................15 ATCA ........................................................................................................................................16 µTCA ........................................................................................................................................18 I/O INTERCOMMUNICATIONS.....................................................................................................20 4.1 PCI Express I/O intercommunications .....................................................................................20 4.2 Ethernet I/O intercommunications .........................................................................................21 5 SIGNAL CONNECTIVITY ...............................................................................................................22 6 EXAMPLE CONFIGURATIONS.......................................................................................................23 6.1 General Purpose Fast I/O........................................................................................................23 6.1.1 6.1.2 6.2 How-to Build a General Purpose Fast Controller ....................................................................23 I&C Grade Fast Controller requiring large number of I/O modules ........................................25 DO’s and DON’Ts with I/O cards based on Conventional PCI ...................................................26 6.2.1 6.2.2 DO: Integrate a DAQ System requiring 6U CompactPCI digitizers ..........................................26 DON’T: General Purpose I/O requiring Conventional PCI or direct PCI Express I/O................27 6.3 High-End fast controllers for specific projects .........................................................................28 6.4 Alternative solutions for Plasma Control front-end computers ...............................................29 6.4.1 6.4.2 6.4.3 PCS - Activating the 10 Gb/s Ethernet Option of Basic Fast Controller ...................................29 PCS – GPU Processing Expansion ............................................................................................29 PCS – Total segregation of pre-processing from front-end computer ....................................30 2 6.4.4 7 8 PCS – PCI Express Expansion in ATCA based systems..............................................................31 SYSTEM INTEGRATION................................................................................................................32 7.1 Opportunities for Industrial System Integrators......................................................................32 7.2 CODAC I&C Integration Kit......................................................................................................33 7.3 How to make new I/O cards compliant with the ITER I&C Standards ......................................34 CONCLUSIONS ............................................................................................................................36 3 2 INTRODUCTION The ITER project’s long time span and the nature of the instrumentation and control (I&C) procurement procedures for the plant systems require that the ITER Organization defines and follows well recognized standards which are used both by industry and in physics experiments. For I&C manufacturers, one of the most important standards is the computer bus specification which is needed for the selection of input and output (I/O) devices, CPU boards and other components used in plant systems. However, today’s I/O specifications cover not only the electrical and mechanical aspects of an I/O bus system but also the interconnection and communication protocol issues between sub-systems. The ITER Organization is following a top-down approach in the specification process. In the first stage, the communication protocol specifications and electrical specifications for I/O connectivity will be chosen. The second stage will be the selection of the mechanical form factors and commercial products for the chassis, CPU, I/O and communication modules. 2.1 PCDH context The Plant Control Design Handbook (PCDH) [1] defines methodology, standards, specifications and interfaces applicable to the whole life cycle of ITER plant instrumentation and control (I&C) systems. I&C standards are essential for ITER to: Integrate all plant systems in one control system. Maintain all plant systems after delivery acceptance. Contain cost by economy of scale. PCDH comprises a core document which presents the plant system I&C life cycle and recaps the main rules to be applied to the plant system I&Cs for conventional controls, interlocks and safety controls. Some I&C topics will be explained in greater detail in dedicated documents associated with PCDH as presented in the picture below. This document is one of them. 4 2.2 Objectives At the current first stage of the standards selection process, the primary objective of this document is to present the ITER Organization’s decision concerning communication protocols and the bus to use with input and output systems. This gives a solid base for selection of the fundamental technologies. The secondary objective is to give guidelines for the second stage of the selection process to allow rapid evaluation of commercial products which could – but not necessarily will – make their way into the ITER CODAC certified hardware catalog. This catalog is to be published in continuous editions, together with the Plant Control Design Handbook (PCDH) [1]. 2.3 Scope of the Document The standards and specifications presented are applicable to all CODAC input and output systems. However, control systems within a plant system containing exclusively slow I/O, will be built using Siemens Series 7 PLC devices and these are therefore out of the scope of this document. The document mainly addresses the definitions and standards to be used with plant system fast controllers. In exceptional cases, this document is applicable for slow I/O, when the system contains both fast and slow I/O (see PCDH [1] to learn more about slow and fast I/O definitions). The network interconnect is out of scope of this document and the starting point of the Plant Operation Network interconnect is discussed in the ‘Integration Kit for PS I&C’ satellite document [9]. 2.4 Acronyms AMC ATCA CPU CODAC COTS I&C I/O ISA µTCA PCDH PCI PCI Express PCI-X PICMG PLC PMC PROFINET PSH PXI PXI Express SAS Advanced Mezzanine Card – often used in ATCA and in µTCA Advanced Telecommunications Computing Architecture (PICMG 3.0) Central Processing Unit COntrol, Data Access and Communication Commodity off the Shelf Instrumentation and Control Input and Output Computer bus standard found in early PCs microTCA – simplified specification from ATCA Plant Control Design Handbook Peripheral Component Interconnect – computer bus standard Serialized communications protocol for computer I/O Extended PCI – computer bus standard PCI Industrial Computer Manufacturers Group Programmable Logic Controller PCI Mezzanine Card – a piggy-bag form factor Industrial Fieldbus communications protocol on Ethernet Plant System Host – a dedicated CODAC computer for Slow I/O management PCI Extensions for Instrumentation An evolution of PXI using PCI Express technologies Serial Attached Small Computer System Interface – computer disk connection 5 SATA TCP/IP VLAN VME Serial Advanced Technology Attachment – computer disk connection Also called Internet Protocol Stack – A communication protocol Virtual Local Area Network Versa Module Eurocard – computer bus standard 2.5 References 1. 2. 3. 4. 5. 6. 7. 8. 9. Plant Control Design Handbook (27LH2V v7.0) CODAC Catalog of I&C products – Fast Controllers (345X28 v2.0) CODAC Conceptual Design (27LBW6 v1.1) Fast Controller Survey (2M3YGM v2.2) Technology Survey Report for Plant System Simulator project (ITER/CT/09/4100000771) (2NC4LT v2.0) Get CODAC Core System http://www.iter.org/org/team/chd/cid/codac/coresystem Cubicle Products (35LXVZ v2.1) I&C cubicle internal configuration (4H5DW6 v2.0) Integration Kit for PS I&C (C8X9AE v1.0) 6 3 REVIEW OF BUS TECHNOLOGIES This chapter makes a short review of the different input and output technologies and communications methods considered during the design phase of the CODAC system. Those familiar with the industrial I/O bus systems can safely skip this chapter. The intention is to give background information on the proposed specifications and standards in a simplified manner. The ITER Organization has conducted technology surveys on plant system fast controllers (see [4] and [5]) which should be consulted for an in-depth analysis of the technologies presented. Those who are familiar with PCI Express I/O bus system can safely skip this chapter. 3.1 VMEbus systems VMEbus (VME) was created in early 80’s mainly as a result of Motorola’s effort to push for an open embedded bus standard based on their popular 680x0 family of processors. It is a parallel, asynchronous bus (in its original form) with one bus arbiter but allowing several bus masters for multiprocessor applications. Data and address lines are separated, with three different address spaces. Various form factors and variations have been produced since the introduction of VMEbus. VMEbus has been widely successful and a large number of manufacturers have produced a huge selection of processor and I/O boards with excellent interoperability. Many high-energy physics experimental facilities have adopted VMEbus technology for their embedded front-end control and data acquisition systems. The golden era of VMEbus based systems was in the 1990’s in its 6U height format (as shown) withP1/P2 connectors and a 32-bit asynchronous bus. Single board computers for VMEbus were mainly 680x0 or PowerPC based. The ITER Organization does not consider VMEbus based systems suitable for the ITER project for two reasons. Firstly, all attempts to move the VMEbus design from 32-bit asynchronous bus to 64-bit synchronous bus have failed to gain ground commercially. Secondly, there was no serial communications protocol specification for VMEbus available at the time when the PCDH [1] document was created. Guideline 1-1. VMEbus based systems will not be used in the ITER Instrumentation and Control System or in any of its subsystems. 3.2 PCI – Peripheral Component Interconnect There is no doubt that PCI is the all-time winner in computer peripheral bus systems from the sheer number of installations and I/O cards available. It has a quite unexciting specification of synchronous, parallel data bus with 33 MHz and 32 bit data width with multiplexed data and address cycles. The theoretical bandwidth for a single card has a maximum of 132 MB/s. 7 PCI bus was originally developed by Intel while the PCI-SIG industrial interest group (www.pcisig.com) maintains the specification. The most important specification, PCI 2.0, was published in 1993. The name has recently been changed to Conventional PCI to make a distinction with the PCI Express specification. What makes the PCI-bus a true winner is its capability of device enumeration and identification. This has allowed the creation of today’s plug-and-play enabled operating systems. In real-life systems the performance is hampered by bus arbitration because of the shared connection topology. The number of PCI cards attached to a PCI bridge is limited to five. In practice, systems will contain several PCI bridges connected in series to increase the total number of PCI cards. This will further increase the latency time of the most distant PCI devices. On the right is a diagram of the device topology that was used on all PC based systems for well over a decade. Note the presence of what is called North (core system) Bridge and South (I/O) Bridge. They are part of the CPU chipset (Intel) and to a large extent their quality and characteristics determine the performance and the scalability of the entire system. There have been several design iterations to try to resolve the most evident performance bottleneck, the shared connection topology, by simply putting in more horsepower. As a result there is a 64-bit PCI, 3.3 Volt 64-bit PCI and last but not least the PCI-X specification which can do a whopping 2.1 GB/s at 266 MHz. However, the physical limits of the bus make every step forward extremely painful. In spite of its evident drawbacks, the Conventional PCI specification has been proposed by the ITER CODAC team because it has such a dominant market position. Further PCI specification variants are studied in the following sections. Guideline 1-2. Conventional PCI will not be used in the ITER Instrumentation and Control System as a plain parallel bus, nor as a unique mechanical form factor. Guideline 1-3. The ITER Instrumentation and Control System will contain fast controllers that would allow usage of Conventional PCI cards. Their usage is, however discouraged. 8 Guideline 1-4. The ITER Instrumentation and Control System will contain fast controllers that would allow usage of 32-bit and 64-bit PCI-X cards. Their usage is, however discouraged. 9 3.3 PCI in Industry and in Science The PCI bus found its way into industrial products very quickly. The governing specification body for PCI bus industrial applications is the PICMG Interest Group (www.picmg.org). 3.3.1 PICMG 1.0 and 1.1 Two of the most successful specifications are PICMG 1.0 (1994) and PICMG 1.1 (1995). These specifications are still widely used in industrial grade computers. They define a passive PCI/ISA bus backplane on which an industrial grade single board computer can be plugged. Then everything can be wrapped in an industrial grade, rackable enclosure. For example, to have a 4U height industrial computer with 4 ISA slots and 7 PCI slots, the following components would be selected: Select motherboard Select CPU board Select 2U or 4U casing The ITER Organization does not consider PICMG 1.0 or 1.1 suitable for CODAC system’s needs for two reasons. Firstly, it is based on Conventional PCI bus which is 32 bit technology. Secondly the Team estimates that although the industry’s needs would keep the market alive, there will be no further innovation. Guideline 1-5. PICMG 1.0 or 1.1 will not be used in the ITER Instrumentation and Control System. 3.3.2 CompactPCI Other, more rugged industrial standards followed PICMG 1.0. The CompactPCI (PICMG 2.0) specification defines a Eurocard PCI card form factor, assigning a PCI pinout on 2 millimeter hard metric connectors. J1 = 110 pins for PCI, J2 = 110 pins defined in different specifications, for Telecom, Ethernet, ANSI I/O, and so on. 3U and 6U form factors are defined, where the 6U format has additional J3/J4/J5 connectors that are used in many telecommunication and signal processing applications. PICMG 2.1 defines the hot swappable CompactPCI. 10 CompactPCI is a clever specification which was in direct concurrence with the VMEbus based systems. The limitations of the PCI bus have always hampered the popularity of CompactPCI while the PC compatibility of the CPU board, operating system and drivers, the J2-J5 connectors' massive and reliable extra pin population and excellent mechanical conception have been the attractive factors. ITER Organization will not consider using CompactPCI as an I/O bus system because the only commercially viable specification is 32-bit and because the market position in the long term is questionable due to lack of innovation. The ITER Organization recognizes the availability and reliability of I/O cards based on CompactPCI specification. Therefore the Team wishes to promote more advanced I/O bus solutions which allow backward compatibility with the CompactPCI I/O cards. Guideline 1-6. CompactPCI will not be used as an I/O bus or as a dominant form factor in the ITER Instrumentation and Control System. Guideline 1-7. The ITER Instrumentation and Control System allows using 3U height CompactPCI I/O cards within other, backward compatible form factors recommended in the CODAC fast controller catalog. 3.3.3 PXI PICMG 2.8 defines PXI, namely the use of the J2 connector of the CompactPCI for instrumentation purposes. National Instruments split from the specification in 1997, creating PXI System Alliance (www.pxisa.org). The rest is a commercial success story, where the driving force is the National Instruments software flag-ship, LabView. Although PXI is backward compatible with CompactPCI, the majority of the more than 1000 PXI I/O cards available come from either National Instrument's own catalog or from their commercial partners who have an interest to see their instrumentation solution integrated into the National Instruments’ products. For testing, laboratory instrumentation or similar applications there is a high probability that PXI/LabView can provide a solution to meet the requirements. Despite PXI’s commercial success, the ITER Organization is not considering defining the use of pure PXI based chassis and CPUs because they are based on 32-bit PCI bus technology. Guideline 1-8. The ITER Instrumentation and Control System discourage the use of 32-bit, parallel bus PXI based chassis and CPUs Guideline 1-9. The ITER Instrumentation and Control System allows using 3U height parallel bus PXI I/O modules within other, backward compatible form factors recommended in the CODAC fast controller catalog. 11 3.3.4 PCI Express The ITER Organization has studied the use of serial I/O bus protocols because of their impressive performance and rapid market penetration. PCI Express combines many assets of the old and new technology because it wraps the Conventional PCI specification in a communication protocol specification. Banging their heads on the wall with PCI-X, Intel came up with completely new thinking. Their idea was to create a completely new approach for peripheral communications: Telecom techniques Serial communications Communication protocol Bus data is the payload Switches and routers Point-to-point connections Backward compatible with Conventional PCI No special Graphics bus Very high bandwidth Scalable The requirements for serial communication and the scalability for different performance levels come together in the PCI Express specification. A connection between two devices is done with lanes. A lane contains two pairs of connection media (copper, fibre): one pair for transmitting and one pair for receiving. Lanes can be added between devices giving a gain in performance. Data transmission multiplexing is done automatically by switching. Therefore for example, a 4x lane device can be operated with just 1x lane – it still works, but it is slower. This incredible scalability can be used, for example to move a device hundreds of meters away from the host device, using a single pair of optical fibres, carrying 1x lane of PCI Express. 12 PCI Express following: layers are the Software layer Transaction layer Data link layer Physical layer Logical Electrical Mechanical layer The data payload is the actual bus data which is, of course, backward compatible with the PCI bus data. Therefore it is trivial to create a bridge between a PCI Express switch and a Conventional PCI bus. In fact, all the current CPU chipsets provide one such a bridge by default. Therefore Conventional PCI bus I/O cards can be connected to any PCI Express enabled computer The different layers create about a 20% overhead on data transmission efficiency. But despite that, the capacity of a single PCI Express 2.0 lane for 5 GT/s (giga transactions per second) means some 500 MB/s of data transfer speed per lane. Another important feature of PCI Express is the virtual channeling which allows prioritized routing of data packets. This ability to have virtual point to point connections is used today to create high speed networks between systems. In fact, the performance of 4x or 8x PCI Express based networks carrying TCP/IP as payload are much better than those of a 10 Gb/s Ethernet. Guideline 1-10. PCI Express will be used in the ITER Instrumentation and Control System and in its sub-systems as the principal I/O Bus System and as an alternative communication method between interconnected systems. As discussed in section 3.2, Conventional PCI based I/O cards are available in different form factors in abundant quantities and the ITER Organization plans their use in the ITER Instrumentation and Control System and its different sub-systems. Only the pure Conventional PCI bus based form factors will be forbidden. 13 Guideline 1-11. Conventional PCI form factor cards cannot be used in the ITER Instrumentation and Control System and in its sub-systems while CompactPCI and PXI based cards can be used. 3.4 PCI Express in Industry and in Science In industry, PCI Express has mainly been pushed by the PICMG interest group. There is a single but big obstacle to the fast propagation of the PCI Express in industrial world: the huge amount of Conventional PCI I/O used in industry. Luckily the excellent backward compatibility of PCI Express is helping the transition. Instead of immediately building pure PCI Express based systems, all manufacturers propose mixed systems so that the mechanical appearance and the number of Conventional PCI / PCI Express slots can be selected according to end user's needs. 3.4.1 PICMG 1.3 The most obvious transition path from Conventional PCI to PCI Express in industrial applications is provided by the PICMG 1.3 specification. Its extensions even allow running an old PICMG 1.0 system using a PICMG 1.3 leveraged PCI Express enabled System Host Board (single host board (SHB) or single board computer). As with PICMG 1.0, an industrial computer integrator and PICMG 1.3 components would be selected according the particular requirements. Select motherboard (here a butterfly model for 2U) Select CPU board Select 2U or 4U casing (or even Shoebox!) PICMG 1.3 based systems inherit the mechanical design that is largely inferior to form factors such as CompactPCI. The ITER Organization understands the obvious problems this presents for cabling and reliability of some type of I/O cards. It is not suggested that the ITER Instrumentation and Control System could be built solely with PICMG 1.3 computers, but PICMG 1.3 based industrial grade computers have the following advantages that would make them excellent elements in the CODAC fast controller catalog: Large casings with good cooling capacity allow low-cost solutions with high-end processors, such as Intel® Xeon® with excellent power / price ratio Extreme flexibility to resolve upcoming configuration problems, ranging from systems with PCIExpress-only-connectivity to systems with PCI-Express, Conventional PCI and even 64-bit PCI-X Longevity and availability proven by PICMG 1.0 Guideline 1-12. PICMG 1.3 based systems will be used in the ITER Instrumentation and Control System. 14 Guideline 1-13. PICMG 1.3 based systems will be used for their CPU power and PCI Express connectivity in mixed, interconnected systems. Guideline 1-14. PICMG 1.3 based systems will not be used to host high-end I/O cards which require either front-panel connectors or large, heavy or otherwise complicated connectors. 3.4.2 CompactPCI Express, CompactPCI Plus and CompactPCI PlusIO CompactPCI Express is an extension of the PICMG 2.0 definition to define PCI Express-ready connectors and backplane on CompactPCI standard. CompactPCI Plus (PICMG 2.30) is a supplement to the PICMG 2.0 CompactPCI Standard. The workgroup started its work at the end of 2008 and the aim is to add a standard pin-out for BP (User I/O) signals for serial high-speed interconnects. This would allow four 1x lane PCI Express, four SATA, four USB and two Ethernet serialized connections to “satellite” slots, as the following picture illustrates. CompactPCI PlusIO standardizes the above interfaces at the rear J2 connector of CompactPCI, maintaining full compatibility with the CompactPCI specification and thus allowing a migration path to CompactPCI Plus, which will the major new standard for CompactPCI. The ITER Organization welcomes the initiative and finds it an attractive way to modernize the CompactPCI standard while maintaining the compatibility with the existing, huge catalog of CompactPCI I/O cards. In particular, the presence of USB 2.0 in the standard is considered very interesting. The fact that the PCI Express is limited to 1x lane only will limit the use to general purpose I/O. Guideline 1-15. The ITER Organization will consider adding CompactPCI PlusIO compliant items into the CODAC fast controller catalog when compatible COTS products will become available. 15 3.4.3 PXI Express The PXI System Alliance (www.pxisa.org) has homologated a definition named PXI-5, which is the current definition of PXI Express. As the alliance states, the specification allows increasing PXI system’s bandwidth from 132 MB/s to 6 GB/s by taking the advantage of PCI Express in the backplane. In fact, the specification is derived from the CompactPCI Express specification, maintaining backward compatibility with it (although not many CompactPCI Express I/O cards actually exist). The specification maintains backward compatibility with 32-bit CompactPCI and PXI cards allowing manufacturers (for any practical purpose, National Instruments) to create hybrid chassis allowing the combination of different standards within the same chassis. The PXI Express chassis manufacturers have been scarce, practically only National Instruments. Other manufacturers, such as Agilent Technologies have come up with their own PXI Express chassis. The ITER Organization considers PXI Express as a good platform to be used in the ITER Instrumentation and Control System for following reasons: Excellent compliance with 32-bit PXI and CompactPCI standards, allowing the usage of an extremely wide range of COTS I/O cards Availability of some high-end PXI Express compliant I/O cards from National Instruments and Agilent Technologies Possibility to operate a PXI crate as a PCI Express extension (without National Instrument’s CPU) from another computer, for example from an industrial grade PICMG 1.3 computer Guideline 1-16. PXI Express chassis hosted I/O sub-systems will be used in the ITER Instrumentation and Control System. In order to minimize the vendor lock-in and to maximize flexibility in the upgrade path regarding the operating system and the CPU power, the ITER Organization considers it necessary to avoid the use of National Instrument’s embedded CPU controllers in the PXI Express chassis, 16 Guideline 1-17. PXI Express chassis is recommended to be used as PCI Express extension of another, CPU oriented industrial computer, such as PICMG 1.3 based system. 3.4.4 ATCA Advanced Telecommunications Computing Architecture (AdvancedTCA or ATCA) is another of the PICMG consortium specifications (PICMG 3.0 being the base specification). It contains innovative chassis specification with no bus but a fabric connection between ATCA boards (aka. blades). ATCA equipment originally targeted telecom applications where high speed interconnect communications are essential. Other criteria were reliability, availability and serviceability. An ATCA chassis is highly manageable, which means that almost every aspect of its operation can be monitored remotely and actions can be taken, for example if an environmental sensor indicates an alert, or if a watchdog reports a nonfunctional blade. This, together with the hot-swappable electronic units allows creation of systems with high availability (minimum 99.999 % - "five nines" – or 5 minutes nonavailability per year). The most common ATCA blade fabric interconnect method is, by far, Ethernet (GbE), specified by PICMG 3.1 standard. PICMG 3.4 specifies the PCI Express use in the ATCA chassis for fabric interconnection. There are other communication interconnection methods available e.g. the RapidIO (PICMG 3.5) which is used for packet switched interconnection at FPGA chip level. The selection of the interconnection method depends of the application and the data exchange requirements. For example, TCP/IP based communications without high performance criteria are easiest to set up with Ethernet fabric which is readily available in commercial products. PCI Express, on the other hand allows the performance to be increased simply by adding lanes, as seen above. For example, it is possible to create a TCP/IP network connection between two blades simply by passing the TCP/IP as payload data in PCI Express. By using, say eight lanes the data transmission performance can easily be multiplied compared to a single link GbE connection between two blades. In general, PCI Express performs better with larger data packets. RapidIO performs well with small data packets. However, only Ethernet based backplane fabrics are supported in most COTS systems on the market. COTS I/O card availability in ATCA is poor for classic control system needs. There are many in-house ATCA I/O card designs in control systems for physics experiments. They take advantage of the large printed circuit area of the ATCA boards but on the other hand they also have to resolve the connectivity issues, which often lead to other specific design problems when the signals have to be routed through a dedicated ATCA Rear Transition Module (RTM). Some ATCA blades have dedicated Advanced Mezzanine Card (AMC) or even PCI Mezzanine Card (PMC) carriers. With these cards, the COTS availability is certainly considerably expanded but only the AMC 17 boards fall into the category of PCI Express enabled devices. It is precisely on the AMC side where the spectrum of I/O cards available is significantly smaller. Even when it is possible to find a suitable a COTS AMC I/O module, the problem of finding a suitable carrier and routing the I/O through an RTM module of specific design remains, as illustrated in the below on the left (courtesy of W. Koprek & T. Jezynski, DESY). xTCA Extensions for Physics are currently created under a PICMG Technical Subcommittee. The PDG.0 R1.0 standard addresses the physics clocks, gates, triggers and timing on PICMG 3.0, amongst others. The standard will be available for ATCA in 2013 but the ITER Organization estimates that it will take a long time before its recommendations find their way into COTS products. AXIe is another standard based on AdvancedTCA with extensions for instrumentation and tests (http://www.axiestandard.org/). Today’s situation is that ATCA based systems cannot be considered for use as general purpose fast controllers in ITER instrumentation and control systems. However, it looks as if many ongoing efforts to bring out standards making the ATCA based systems suit the high energy physics experiments better will make a brake through, especially for the “high-end” applications such as plasma diagnostics. This is true in particular where the performance is not the only issue but also redundancy and high availability are important. The ITER Organization is actively prototyping the ATCA platform presented in the fast controller catalogue in different pilot projects related to the data acquisition use cases. When applicable, xTCA Extensions for Physics and AXIe standards are also studied in related projects. Timing and PCI Express interconnect solutions have been made available by external companies. The above observations and the somewhat uncertain COTS market situation lead the ITER Organization to publish the following, quite generic guideline regarding the use of ATCA based systems in the ITER Instrumentation and Control System. Guideline 1-18. ATCA with xTCA extensions for Physics (PDG.0 R1.0 or greater) is a recommended platform for high end diagnostics applications, such as DAQ applications participating in the Plasma Control System. 18 19 3.4.6 µTCA Industry has quickly taken advantage of the ATCA specification and the outcome is a simplified µTCA (microTCA) specification from PICMG, the MTCA.0. A microTCA system is basically a simplified carrier for Advanced Mezzanine Cards (AMC). The chassis management is maintained, although much simplified by concentrating multiple tasks into a µTCA carrier hub. As the name implies, the hub carries from one to a maximum of twelve AMC mezzanine cards, depending on the mechanical arrangement. The PCI Express (and Ethernet) connectivity remains in the microTCA zone connectors, but as the picture illustrates, the main job of the µTCA carrier hub is to make SATA, Ethernet and PCI Express packet switching between the AMC slots. One of the arguments to use PCI Express is the possibility to use it for system extension. One could extend a µTCA chassis, for example, by connecting it to a PXI Express chassis which would considerably enlarge the available COTS standard I/O selection. Some attempts have been made as the picture on the right illustrates but again, but the general availability of the PCI Express extension products for µTCA is non-existent. The specification for xTCA Extension for Physics, mentioned above is more advanced in the µTCA platform. It is now possible to purchase a full width / full height µTCA shelf, compliant with the xTCA Extension for Physics, MTCA.4 from Schroff or ELMA, for example. Compared to the 6U VME and CompactPCI this platform is particularly interesting with its µRTM (Rear Transition Module) illustrated below. Together with the AMC board, the total real estate area is considerable and can be used for 20 modular design, for example, doing the signal conditioning on the µRTM module. The ITER Organization is aware of the µTCA’s potential to become the “next VME” when it comes to interoperability and availability of I/O cards. The problem is that it has not yet reached that stage. Concerning the industry, one can only wait until the market situation clarifies and winning technologies come out. The ITER Organization is closely following the xTCA Extensions for Physics Standard and is conducting pilot projects to evaluate the µTCA/µRTM platform suitability especially for diagnostics applications. Timing and PCI Express interconnect solutions have been made available by external companies. Guideline 1-19. MTCA.4 is a recommended platform for high end diagnostics applications, especially as design solutions where various signal conditioning requirements are fulfilled on separate µRTM modules while the DAQ FPGA base board on AMC side remains the same. 21 4 I/O INTERCOMMUNICATIONS The preceding chapter explained the transition from parallel bus systems to serial data connections where the actual bus information is wrapped in a communication protocol. This means that geographical limitations of input and output devices are largely relaxed and remote I/O is possible over hundreds of metres using fibre links. 4.1 PCI Express I/O intercommunications A typical use of PCI Express I/O intercommunications is with I/O expansion. An industrial system which is PCI Express enabled can easily interconnect with an I/O chassis having a different form factor as long as it has a PCI Express slot and/or perhaps a bridge towards its Conventional PCI bus slots. On the right is an example how PXI Express chassis can be controlled from any PCI Express enabled computer, such as PICMG 1.3 or even from an office workstation. The picture on the left presents a similar expansion card in AMC format which allows bridging of an ATCA AMC carried blade with an industrial computer which has Conventional PCI and PCI Express I/O cards. From the CPU and operating system’s point of view, all I/O boards are seen as any PCI device, there is no geographical notion of cards being remote instead of local. Guideline 1-20. PCI Express will be used in the ITER Instrumentation and Control System for I/O expansion. Another way to use PCI Express to interconnect systems is through an external PCI Express switch, as illustrated in the below picture. The result is a meshed network of computers with shared memory or reflective memory structure. This structure can be used in many ways, one of which is to create a high performance, real-time network between computers by running a network protocol stack on it. 22 Guideline 1-21. PCI Express is a recommend communication method between interconnected I/O systems in ITER Instrumentation and Control System but only if latency and jitter requirements exceed those of the ITER CODAC Synchronous Databus Network (SDN). 4.2 Ethernet I/O intercommunications There are several cases where Ethernet is an adequate communication method between interconnected I/O systems. One example is the Advanced TCA meshed backplane, PICMG 3.1 (see section 3.4.4). If one takes the commercial reality into account, there is a much higher probability of finding a suitable ATCA shelf product with Ethernet backplane fabric than anything else. Direct interconnections between I/O cards could be therefore be made using Ethernet; some designs have already been exploiting this possibility successfully. Another important use is various fieldbus protocols on Ethernet, taking the example of PROFINET. The next diagram illustrates how a fieldbus communication can be organized using VLANs on commercial Ethernet L2 level switches. ID2 Physical Location #1 MPI Eth. PLC PSH Networkc Services: DHCP (MAC) DNS (name/IP) NTP (time) Plant System A Building and Power – Cubicle A ID1002 ID1002 Ethernet Fieldbus ID2 PLC Remote I/O Module CAS/CAC Controller IOC PLC PLC Remote I/O Module Plant System A Building and Power - Cubicle B Guideline 1-22. Switched and meshed Ethernet is a recommend communication method between interconnected I/O systems in the ITER Instrumentation and Control System. 23 5 SIGNAL CONNECTIVITY The connectivity between the I/O signals and fast controller I/O cards in many cases resembles the slow controller connectivity problems. The ITER Organization is coordinating the choice of signal connectivity products so that whenever possible, similar signal patch panels and other products can be used with both slow and fast controllers. Guideline 1-23. The slow and fast controller catalogs published by the ITER Organization will make reference to a common way of building ITER I&C Cubicles, including the cabling and signal interfacing. For more information please consult the I&C cubicle internal configuration (4H5DW6 v2.0)guide [8]. 24 6 EXAMPLE CONFIGURATIONS This chapter gives some block diagrams for possible configurations based on the conclusions of the preceding chapters. 6.1 General Purpose Fast I/O Guideline 1-24. Separated I/O chassis and CPU/Network chassis is the preferred solution in the ITER Instrumentation and Control System. 1 Gb/s Ethernet From Catalog: - 4U PICMG 1.3 chassis Optional 10 Gb/s Ethernet Bus Extension PCI-Express x1 From catalog: - PXIe Chassis I/O from catalog: - PXI - CompactPCI - PXI Express The lifespan of the PXIe chassis and the I/O units will be the lifetime of the ITER project. The CPU/Network chassis’ lifetime is estimated to be 7 to 10 years after which it will be replaced to provide the latest CPU and networking power for the system and control software. The “PICMG 1.3 chassis” is an industrial level computer but not particularly designed for redundancy. However, the most obvious unreliability issue of spinning hard disk is resolved by using two, RAID-1 configured SSD (solid state) disks. The dual power supply is usually available except for systems with most powerful GPU cards. The ITER Organization estimates that some 60% of fast I/O problems can be resolved with the simple type of configuration above. 6.1.1 How-to Build a General Purpose Fast Controller This section simplifies the detailed instructions which are available in the I&C cubicle internal configuration (4H5DW6 v2.0) guide [8]. We will point in parenthesis to the datasheets in IDM folders of proposed items. The following items from the PCDH Fast Controller Catalog [2] have to be obtained: 1. Standard ITER cubicle from the Cubicle Products Catalog [7] 25 2. 3. 4. 5. 6. 7. a. Accessories: four L-shaped glider shelves (to install two computer chassis) and caged counter nuts to attach them into the mounting rails. I&C Grade PICMG 1.3 PC 4U industrial computer (BKLT5Z) PXI-Express hybrid chassis, for example PXIe-1065 (BW8XTV) PCI-Express connectivity between the industrial computer and the PXI-Express chassis, for example one PCI-Express lane version (BW8XTV) – 192 MB/s General purpose I&C Multi-I/O board with full CODAC Core System support, for example PXI-6259 (C8S269) 2.8 MS/s total maximum sampling rate shared by all channels. Synchronization and timing board for system synchronization using PTP protocol, time based triggering and for automated time stamping in CODAC Core System (provided by the ITER Organization, together with the I&C Integration Kit [9] for procurement arrangements) (34RRE2 v1.2). Two termination blocks for the PXI-6259 Multi-I/O board, NI SMB-2090A (incorporated in ITER part number for PXI-6259 board kit, see the reference above for the board). It may be desirable to purchase the following additional item(s) for convenience: 8. An CD-ROM (DVD) disk drive with USB connectivity for the CODAC Core System installation The pictures below illustrate the example configuration installed into an ITER standard cubicle: General purpose fast I/O PXI 6259 SMB 2090A patch panel in the back. The following are required in order to get the CODAC Core System installed: 1. If the institute or company has not been registered as a CODAC Core System user the instructions for registration published on the ITER Organization’s web site [6] should be followed. 2. Within a few days instructions on how to download an ISO image from the internet together with the installation instructions will be delivered. 3. The ISO image should be used to burn a CD. Connect the CD player to the USB port of the industrial computer. 4. Ensure that the system’s upper network plug is connected to the organization’s network 26 a. The network plug is allowed to connect to the Internet and reach the CODAC Core System distribution server (http://codac-dist.iter.org) – this can be easily verified by connecting a laptop or workstation to the plug and using the browser to connect to the above server b. In case of doubt, contact your IT or network personnel and explain what you want to do – they may need to disable the network proxy usage for the port you are using, depending of your corporate network policy. c. If the system does not boot on the USB CD player, press the Del (Delete) key when you see the boot splash screen. Go to the BIOS settings, Boot section and verify that the USB device has been detected, recognized and enabled as boot device. 5. Follow the instructions on the CD’s boot screen to select the appropriate system role (e.g. usually, fast controller with development role). When the connection to the CODAC Core System distribution server has been established, the installation will take anything from two to several hours, depending of the network connection speed. 6. Once the system has been installed, log in using the credentials provided by the CODAC Core System support team. Browse through the example applications to an example application which makes use of the PXI-6259 multi-I/O board in this configuration. a. Study the application and make some modifications to learn how to interface hardware and signals with the CODAC Core System. 6.1.2 I&C Grade Fast Controller requiring large number of I/O modules As the example below illustrates, a very large number of PXI, PXI Express and CompactPCI 3U I/O cards can be interfaced by adding a second 18 slot crate to the system and using the double channel PCI Express connectivity product from the catalog. What is the total number of PXI Express chassis that can be connected into a single PICMG 1.3 industrial computer? In practice only two or four, depending of the fast controller catalog PICMG1.3 items, because each high-speed link requires four lanes (x4) and two chassis eight lanes (x8) of PCI Express. The available number of x8 PCI Express slots in the PICMG 1.3 passive backplane is the limitation. Backplanes exist which can take multiple SHB computers (not in the catalogue), each of which drives its own set of PCI Express extension slots. Using this type of motherboard, a total of twelve PXI Express chassis can be connected to a single PC 4U PICMG 1.3 computer, running a total six SHB computers, each connecting with two PXI Express chassis. Contact the ITER Organization’s CODAC Section ([email protected]) to discuss your problems and obtain more information about this type of configuration. 27 1 Gb/s Ethernet From catalog: - 4U PICMG 1.3 chassis Bus Extension 2 x PCI-Express x4 Optional 10 Gb/s Ethernet From catalog: - PXIe Chassis (hybrid) - PXIe Chassis I/O from catalog: - PXI - CompactPCI - PXI Express 6.2 DO’s and DON’Ts with I/O cards based on Conventional PCI As explained in the chapters above, the ITER Organization does not promote or encourage the use of I/O cards based on Conventional PCI but neither does it forbid them completely. For example, PXI cards and CompactPCI cards can be used in the 3U hybrid PXIe chassis found in the fast controller catalog. There are other ways to connect some interesting, existing, I/O cards such as 6U CompactPCI cards. These use PCI Express to PCI bridges and expansion chassis, which are readily available. This section gives some examples of the good - and the bad - practices. 6.2.1 DO: Integrate a DAQ System requiring 6U CompactPCI digitizers In the example below, there is an external chassis, carrying 6U CompactPCI digitizers in a 9U chassis from a well-known instrumentation manufacturer, with 80 channels, each sampling with 1GS/s. Without discussing if the same performance could be achieved some other way, assume that the chassis is here and it needs to be integrated in a fast controller solution from the ITER catalog: select a solution with no embedded CPU from the manufacturers catalog but with a chassis bus extension module to make a cPCI to Conventional PCI connection and put the latter into the PICMG 1.3 industrial computer. It now sees both the PXI Express hybrid chassis and the cPCI chassis. The software integration of the non-standard cPCI chassis into the CODAC Core System software, EPICS control system frame work and the Red Hat Linux operating system needs to be done. For non-catalog items, the ITER Organization has no ready solutions available so the integration works remain under the user’s responsibility. 28 I/O outside of the catalog: - example: 6U CompactPCI Digitizers Conventional PCI / cPCI Bus Extension 1 Gb/s Ethernet From catalog: - 4U PICMG 1.3 chassis Optional 10 Gb/s Ethernet Bus Extension PCI-Express x4 From catalog: - PXIe Chassis (hybrid) I/O from catalog: - PXI / PXI Express - CompactPCI 6.2.2 DON’T: General Purpose I/O requiring Conventional PCI or direct PCI Express I/O If there are some old, maybe in-house built conventional PCI card designs, or even a modern PCI Express board the temptation to put this card into the PICMG 1.3 industrial computer should be avoided. This would be acting against two guidelines: Guideline 1-11 (page 13) explains that Conventional PCI form factor cards should not be used in the ITER I&C. Guideline 1-14 (page 14) explains that PICMG1.3 based systems are used to host the CPU, networking, disk and chassis interconnection components, not I/O. I/O outside of the catalog: - example: specific PCI design DON’T ! 1 Gb/s Ethernet From catalog: - 4U PICMG 1.3 chassis 29 6.3 High-End fast controllers for specific projects At later date, the fast controller catalog will certainly contain ATCA CPU blades and a managed ATCA shelf solution as base components for more ambitious electronics projects. The ITER Organization has prototyped some example configurations for data acquisition use cases and pre-processing in Diagnostics Plant System I&C using the ATCA products in the fast controller catalog. Configuration guidelines based on the experience of these pilot projects are available on demand; the CODAC team can point you to relevant reports in in IDM. 30 6.4 Alternative solutions for Plasma Control front-end computers It might not be necessary or even possible to use a high-end ATCA platform for all diagnostics and even in the Plasma Control System (PCS) it might be necessary to have alternative, less demanding solutions available. The proposed, basic fast I/O solution is expandable. The pictures below illustrate some alternatives as to how the pre-processing and the actual data acquisition can be physically separated in different entities. 6.4.1 PCS - Activating the 10 Gb/s Ethernet Option of Basic Fast Controller Network communication performance can be boosted simply by installing the 10 Gb/s Ethernet option on a standard catalog PC 2U PICMG 1.3 computer. 1 Gb/s Ethernet From catalog: - 4U PICMG 1.3 chassis 10 Gb/s Ethernet (requires x4 PCIe) From catalog: - PXIe Chassis 6.4.2 PCS – GPU Processing Expansion Parallel processing power can be added to a front-end computer by adding GPU processor units (ex. 2 x 120 GPU processors). 31 From catalog: - 2 x 120 GPU Unit - Requires 2x x8 or 2x x16 PCIe Bus Extension PCI-Express x8 1 Gb/s Ethernet From catalog: - 4U PICMG 1.3 chassis 10 Gb/s Ethernet (requires x4 PCIe) From catalog: - PXIe Chassis 6.4.3 PCS – Total segregation of pre-processing from front-end computer Switched PCI Express techniques can be used to create shared memory zones between the front-end computer and the pre-processing computer (Dolphin, One Stop Systems). 32 10 Gb/s Ethernet 2 x 120 GPU Unit SDN 10 Gb/s Ethernet 8 x8 lane PCIe switch - Optional, point-topoint connectivity solutions do exist. PCS Pre-processing Bus Extension PCI-Express x8 PCS front-end 1 Gb/s Ethernet From catalog: - 4U PICMG 1.3 chassis 10 Gb/s Ethernet (requires x4 PCIe) From catalog: - PXIe Chassis 6.4.4 PCS – PCI Express Expansion in ATCA based systems As stated in [4.1], should the Plasma Control System fast controllers be constructed using ATCA, it is necessary to arrange PCI Express connectivity towards the future HPN solutions. Two solutions can be envisaged: 1. AMC PCI-Express expansion card (picture) 2. Purpose-designed Rear-Transition Module RTM (design work has been ordered, and it will be added to the catalogue). This solution is proposed in the hardware catalog [2]. 33 7 SYSTEM INTEGRATION 7.1 Opportunities for Industrial System Integrators Building fast controllers for plant systems is a demanding task. It is not simply a question of selecting hardware from a catalog but different integration skills will be needed to make sure that the system fulfills all the requirements. I/O Bus Extension I/O Chassis Signal Interfacing CPU/Networking Operating System(s) CODAC Core System I/O Modules Fast Controller Device Drivers EPICS PCDH Satellite Catalogues and Documents Plant System Design Handbook Plant System Requirements ITER Instrumentation and Control System Requirements Obviously a plant system manufacturer has a good understanding in his specific domain and could easily come to a conclusion regarding signal interfacing and other issues close to the local control problem. But many other aspects of the integration process require specific knowledge about operating systems, networking, EPICS and the ITER Instrumentation and Control System in general. It is unlikely that all plant system manufacturers could obtain a similar level of knowledge by simply reading the documentation provided by ITER Organization. Therefore system integrators who can propose complete systems for the specific needs of each plant system are needed. System integrators would work closely with the ITER Organization and would be certified both by the ITER Organization and by the Domestic Agencies for selling complete, CODAC compliant solutions for plant system manufacturers who could then concentrate on their specific domain of expertise. 34 7.2 CODAC I&C Integration Kit The CODAC I&C integration kit is an organizational concept to distribute some of the main components of the previously described control system to the plant system manufacturers. The concept has been created by the ITER Organization to help the Domestic Agencies and their contractors in their work to build CODAC compatible I&C systems. The use of identical CODAC I&C integration kits will enforce the correct level of standardization in the plant systems. A CODAC I&C integration kit will be loaned to all organizations and entities who have signed a procurement arrangement to build an ITER plant system. It contains all the software tools necessary to develop and manage the construction phase of the plant system I&C. It contains connectivity tools, allowing the plant system manufacturer to use CODAC collaborative tools. These tools are available over the internet, creating a worldwide system to support the I&C software development, configuration, testing, acceptance and integration functions. Physically, the I&C integration kit is a small set of hardware and it contains: 1. A single industrial computer 2. An industrial grade network switch 3. Network connected cubicle health monitoring unit of ITER design 4. Set of accessories and cables 5. [optionally] ITER High Performance Network interface boards, delivered for some plant systems The kit allows the rapid commissioning of the 1st I&C Cubicle of the plant system, constructed by the plant system manufacturer. Read more from the Integration Kit for PS I&C (C8X9AE v1.0) [9]. Cubicle Monitoring From The I & C Integration Kit HPN cards The 1st I&C cubicle of the Plant System Not in the Kit 35 7.3 How to make new I/O cards compliant with the ITER I&C Standards The ITER Organization is fully aware that there could never be such a thing as a fast controller catalog which would cover all of the I&C and especially diagnostics needs. New I/O cards will appear and it is in the general interest to get them integrated in both the ITER I&C system and the CODAC Core System and finally make them appear in the fast controller catalog in order to promote standardization. The picture below illustrates how, ideally, all solutions would be found from PCDH catalogues so that the best, integrated solutions would be selected by the plant system designer together with the responsible officers and experts from the ITER Organization. If no solution is available, the COTS market must be searched with help from an external expert for a suitable candidate product to resolve the given problem. Use Case PS Designer Fast I/O or DAQ Requirements IO ROs PCDH Catalogues IO Expert No New solution required ? EXT Expert Yes Solution Proposal COTS Market When a suitable candidate product has been found, it must be prototyped. The ITER Organization is willing to help in the prototyping effort. Since all products will ultimately lead to their use in the ITER I&C, software items such as Linux device drivers and EPICS Device Support must be developed at the earliest possible moment. The CODAC Core System is the framework for the development and therefore also for the prototyping. Since ITER is not only a scientific project but also a huge industrial and engineering challenge, the successful prototype must be developed further and industrialized. The picture on the next page illustrates the prototyping phase which should contain the Linux and EPICS software support from the very first step of the project. 36 Solution Proposal EPICS Linux Drv HW/SW Prototyping Industrialization The industrialization of a prototype I/O board is a major project which reveals the actual total cost of ownership (TCO) of the proposed I/O board when used through the lifespan of the ITER Project. For installation purposes, the rack mounting, cabling, signal termination and EMC issues must be specified, planned and documented. The software support must reach production level quality and it must be maintained at that level. The procurement of the I/O board must be arranged worldwide and it must be made consistent in time using, for example ITER part numbers. The calibration, replacement and maintenance issues must be planned beyond the commissioning, over the time span of several years, even over decades. The picture below illustrates this continuous process, which is under the control of the ITER Organization, for all items which appear in the PCDH related catalogues. PCDH Fast Controller Catalog Industrialization Linux Drv EPICS Software Support Hardware Support SDD Tests Document s Software maintenance Ex. new CODAC versions Cubicle Integration Hardware and Software integration Procurement Lifecycle Management ITER part numbers Obsolescence management 37 8 CONCLUSIONS The ITER Organization has selected PCI Express as the base I/O bus technology to be used in ITER instrumentation and control system fast controllers. The selection is based on the excellent backward compatibility with different Conventional PCI bus technologies and on the dominant market position of PCI Express in COTS systems when compared to other serial I/O bus technologies. PXI, CompactPCI and PXI Express I/O boards will be used as the main, recommended items for the ITER Instrumentation and Control System fast controller catalog. PXI Express chassis is used to carry the I/O boards. Segregation between the CPU/Network power and the I/O sub-system is promoted by providing PCI Express I/O expansion systems. In this model, an industrial computer with good price/performance ratio is used to provide the CPU power and network connectivity. The I/O will be done through a PXI Express chassis which is connected to the industrial computer, typically over an 8x lane PCI Express link. Since the CPU/Network power requirements are evolving more rapidly than the basic I/O requirements, the industrial computer unit can be replaced when necessary without touching the I/O and most importantly, the cabling. The ITER Organization is promoting the use of Ethernet based interconnection techniques between I/O subsystems together with the PCI Express based I/O. For high-end I/O requirements in diagnostics systems, AdvancedTCA and MTCA.4 (µTCA) systems are considered as the strongest candidates, together with high end PXI Express data acquisition solutions. ITER Organization has created a PCDH satellite document, “ITER Instrumentation and Control System Catalog – fast controllers” to standardize the selection of hardware as far as possible. The various example configurations given in this document emphasize the flexibility of the technical selections made. The complexity of the system selection, purchasing, construction and configuration is recognized. A role of a system integrator to provide this work as a service is suggested. To maximize the speed of penetration of CODAC hardware and software standards into the plant system development and manufacturing cycle, the document introduces the concept of ITER CODAC I&C Integration Kit, which is a complete, turn-key CODAC Core System development solution with all the necessary hardware and software components. 38