Parsytec


ISRA VISION PARSYTEC AG is a company of ISRA VISION AG and was founded in 1985 as Parsytec in Aachen, Germany.
Parsytec has become known in the late 1980s and early 1990s as a manufacturer of transputer-based parallel systems. Products ranged from a single transputer plug-in board for the IBM PC up to large massively-parallel systems with thousands of transputers such as the Parsytec GC. Some sources call the latter ultracomputer sized, scalable multicomputers .
As part of the ISRA VISION AG, today the company focusses on solutions in the machine vision and industrial image procession sector. The ISRA Parsytec products are used for quality and surface inspection especially in the metal and paper industries.

History

In 1985, Parsytec was founded by Falk-Dietrich Kübler, Gerhard H. Peise, and Bernd Wolff in Aachen, Germany, with an 800000 DM grant from Federal Ministry for Research and Technology.
In contrast to SUPRENUM, Parsytec directly aimed their systems at industrial applications such as surface inspection. Therefore, they not only had a substantial market share in the European academia but they could also win many industrial customers. This included many customers outside Germany. In 1988, export accounted for roughly a third of Parsytec's turnover.
Turnover figures were: nil in 1985, 1.5M DM in 1986, 5.2M DM in 1988, 9M DM in 1989, and 15M DM in 1990, 17M USD in 1991.
In order to focus Parsytec on research and development, ParaCom was founded. ParaCom thence took care of the sales and marketing side of the business.
Parsytec/ParaCom's headquarters were maintained in Aachen, however they had subsidiary sales offices in Chemnitz, Southampton, Chicago, St Petersburg and Moscow. In Japan, the machines were sold by Matsushita.
Between 1988 and 1994, Parsytec built quite an impressive range of transputer based computers having its peak in the "Parsytec GC" which was available in versions using 64 up to 16384 transputers.
Parsytec had its IPO in mid-1999 at the German Stock Exchange in Frankfurt.
On Apr, 30, 2006 founder Falk-D. Kübler left the company.
In July 2007, 52.6% of the Parsytec AG were acquired by ISRA VISION AG. The delisting of Parsytec shares from the stock market started December the same year. And since 18 April 2008, the Parsytec share is no longer listed on the stock exchange.
Whilst workforce at Parsytec was roundabout 130 staff in the early 1990s, the ISRA VISION Group had more than 500 employees in 2012/2013.
Today, the core business of ISRA Parsytec within the ISRA VISION Group is the development and distribution of surface inspection systems for strip products in the metal and paper industries.

Products/Computers

Parsytec's product range included:
In total, some 700 stand-alone systems had been shipped.
In the beginning, Parsytec had participated in the GPMIMD project under the umbrella of the ESPRIT project, both being funded by the European Commission's Directorate for Science.
However, after substantial divisions with the other participants, Meiko, Parsys, Inmos and Telmat, as regards the choice of a common physical architecture, Parsytec left the project and announced a T9000-based machine of their own, i.e. the GC. But due to Inmos' problems with the T9000, they were forced to change to the ensemble Motorola MPC 601 CPUs and Inmos. This led to Parsytec's "hybrid" systems degrading transputers to communication processors whilst the compute work was offloaded to the PowerPCs.
Parsytec's cluster systems were operated by an external workstation, typically a SUN workstation.
There is a substantial confusion as regards the names of the Parsytec products.
On the one hand this has to do with the architecture, but on the other hand it had to do with the aforementioned non-availability of Inmos T9000 that forced Parsytec to use the and the PowerPC instead. Systems that were equipped with PowerPC processors had the prefix "Power".
For what concerns the architecture of GC systems, an entire GigaCluster is made up of self-contained GigaCubes.
The basic architectural element of a Parsytec system was a cluster which consisted inter alia of four transputers/processors.
A GigaCube consisted of four clusters with 16 Inmos transputers, RAM, plus a further redundant , the local link connections and four Inmos C004 routing chips. Hardware fault tolerance was provided by linking each of the to a different C004.
The unusual spelling of x'plorer led to xPlorer and the Gigacluster is sometimes referred to as the Gigacube or Grand Challenge.

Megaframe

Megaframe was the product name of a family of transputer based parallel processing modules.
Some of which could be used to upgrade an IBM PC.
As a standalone system, a Megaframe system could hold up to ten processor modules. Different versions of the modules were available, for example, one with a 32-bit transputer, floating-point hardware Motorola 68881, 1 MB of RAM and a throughput of 10 MIPS, or one with four 16-bit transputers with 64 kB of RAM.
Also cards for special features were on offer, such as a graphics processor with a resolution of 1280 x 1024 pixels or I/O-"cluster" with terminal and SCSI interface.

Multicluster

MultiCluster-1 series were statically configurable systems and could be tailored to specific user requirements such as number of processors, amount of memory, and I/0 configuration, as weil as system topology. The required processor topology could be configured by using UniLink connection; fed through the special back plane. In addition, four external sockets were provided.
Multicluster-2 used network configuration units that provided flexible, dynamically configurable interconnection networks. The multiuser envirorunent could support up to eight users by using Parsytec's multiple virtual architecture software.
The NCU design was based on the Inmos crossbar switch, the C004, which gives full crossbar connectivity for up to 16 transputers. Each NCU, made of C004s, connected up to 96 UniLinks that link internal as weil as external transputers and other I/0 subsystems.
MultiCluster-2 provided the ability to configure a variety of fixed interconnection topologies such as tree or mesh structures.

SuperCluster

SuperCluster had a hierarchical, cluster-based design. A basic unit was a 16-transputer, fully connected cluster; larger systems had additional levels of NCUs to form necessary connections. The Network Configuration Manager software controlled the NCUs and dynamically established the required Connections. Each transputer could be equipped with 1 to 32 MB of dynamicic RAM
with single-error correction and double-error detection.

GigaCluster

The GigaCluster was a parallel computer which was produced in the early 1990s. A Giga Cluster was made up of Giga Cubes.
Being designed for the Inmos T9000-transputers, it could never be launched as such, since the Inmos T9000 transputers itself never made it to the market in good time.
This led to the development of the GC/PP in which two Motorola MPC 601 were used as the dedicated CPUs supported by four transputers
Whilst the GC/PP was a hybrid system, the GCel was based on only. The GCel was supposed to be upgradeable to the T9000 transputers, thus becoming a full GC. As the T9000 was Inmos' evolutionary successor of the, upgrading was planned to be simple and straightforward because, firstly, both transputers shared the same instruction set and, secondly, they also had quite a similar performance ratio of compute power versus communication throughput. Therefore, a theoretical a speed-up factor of 10 was expected but in the end it was never reached.
The network structure of the GC was a two-dimensional lattice with an inter-communication speed between the nodes of 20 Mbit/s.
For the time, the concept of the GC was exceptionally modular and thus scalable.
A so-called GigaCube was a module that was already a one gigaflop system; furthermore, it was the building block for greater systems.
A module contained
of which each was equipped with
By combining modules one could theoretically connect up to 16384 processors to a very powerful system together.
Typical installations were:
SystemNumber of CPUsNumber of GigaCubes
GC-1641
GC-22564
GC-3102416
GC-4409648
GC-516384256

The two largest installations of the GC, which were actually shipped, had 1024 processors and were operated at the data centers of the Universities of Cologne and Paderborn.
In October 2004, the latter had been given to the Heinz Nixdorf Museums Forum where it is inoperable now.
The power consumption of a system with 1024 processors was approximately 27 kW, the weight was almost a ton. In 1992, the system priced about 1.5M DM. While the smaller versions up to GC-3 were air-cooled, water cooling was mandatory for the larger systems.
In 1992, a GC with 1024 processors reached a placement in the TOP500 list of the world's fastest supercomputer installations. In Germany alone, it was number 22 of the fastest computers.
In 1995, there were nine Parsytec computers in the Top500 list of which two GC/PP 192 installations ranked 117 and 188 in the TOP500 list.
And in 1996, they still ranked 230 and 231 in the TOP500 list.

x'plorer

The x'plorer model came in two versions: The initial version was featuring 16 transputers, each having access to 4MB RAM and called just x'plorer. Later when Parsytec generally switched to the PPC architecture, it was called POWERx'plorer and featured 8 MPC 601 CPUs. Both models came in the same gorgeous desktop case.
In any model, the x'plorer was more or less a single "slice" — Parsytec called them cluster — of a GigaCube, which used 4 of those clusters in its smallest version. Thus, some call it a "GC-0.25".
The POWERx'plorer was based on 8 processing units arranged in a 2D mesh. Each processing unit had
  1. one 80 MHz MPC 601 processor,
  2. 8 MB of local memory and
  3. a transputer for establishing and maintaining communication links.

    Cognitive Computer

The Parsytec CC system was an autonomous unit at the card rack level.
The CC card rack subsystem provided the system with its infrastructure including power supply and cooling. The system could be configured as a standard 19'' rack mountable unit which accepted the various 6U plug-in modules.
The CC system was a distributed memory, message passing parallel computer and is globally classified into the MIMD category of parallel
computers.
There were two different versions available
In all CC-systems, the nodes were directly connected to the same router which implemented an active hardware 8 by 8 crossbar switch for up to 8 connections using the 40 MBytes/s high-speed link.
For what concerns the CCe, the software was based on IBM's AIX 4.1 UNIX operating system together with Parsytec's parallel programming environment Embedded PARIX.
Thus, it combined a standard UNIX environment with an advanced software programming development environment. The system was integrated to the local area network using standard Ethernet. Therefore, a CC node had a peak performance of 266 MFlops. The peak performance of the 8-node CC system installed at Geneva University Hospital was therefore 2.1 GFlops.

Powermouse

Powermouse was another scalable system that consisted of modules and individual components. It was a straightforward extension of the x'plorer-system. Each module contained four MPC 604 processors and 64 MB RAM attaining a peak performance of 2.4 Gflop/s.
A separate communication processor ) equipped with 4 MB RAM, controlled the data flow in four directions to other modules in the system. The bandwidth of a single node was 9 MB/s
For about 35000 DM a basic system consisting of 16 CPUs could provide a total computing power of 9.6 Gflop/s. As was with all Parsytec products, Powermouse required a Sun Sparcstation as the front-end.
All software was included.

Operating system

The operating system used was PARIX . Based on UNIX, PARIX supported remote procedure calls, it was compliant with the POSIX standard.
PARIX provided UNIX functionality at the front-end with library extensions for the needs of the parallel system at the backend which was precisely the Parsytec product itself. The Parix software package comprised components for the program development environment and runtime environment. PARIX offered different types of synchronous and asynchronous communication.
In addition, Parsytec provided a parallel programming environment called Embedded PARIX.
To develop parallel applications using EPX, data streams and function tasks were allocated to a network of nodes. The data handling between
processors required just a few system calls.
Standard routines for synchronous communication such as send and receive were available as well as asynchronous system calls. The full set of EPX calls established the EPX application programming interface. The destination for any message transfer was defined through a virtual channel that ended at any user defined process. Virtual channels were user defined and managed by EPX.
The actual message delivery system software utilised the router.
Moreover, one could also run COSY and Helios on the machines. Helios supported the special reset-mechanism of Parsytec out of the box.