A study of intermittent faults in digital computers

Similar documents
Design of Fault Coverage Test Pattern Generator Using LFSR

VLSI System Testing. BIST Motivation

Scan. This is a sample of the first 15 pages of the Scan chapter.

VLSI Test Technology and Reliability (ET4076)

Impact of Intermittent Faults on Nanocomputing Devices

An MFA Binary Counter for Low Power Application

Modular redundancy without voters decreases complexity of restoring organ

Objectives. Combinational logics Sequential logics Finite state machine Arithmetic circuits Datapath

Jin-Fu Li Advanced Reliable Systems (ARES) Laboratory. National Central University

At-speed testing made easy

FLIP-FLOPS AND RELATED DEVICES

Design for Testability

Avoiding False Pass or False Fail

Sharif University of Technology. SoC: Introduction

data and is used in digital networks and storage devices. CRC s are easy to implement in binary

DESIGN AND SIMULATION OF A CIRCUIT TO PREDICT AND COMPENSATE PERFORMANCE VARIABILITY IN SUBMICRON CIRCUIT

Chapter 5 Flip-Flops and Related Devices

Bit Swapping LFSR and its Application to Fault Detection and Diagnosis Using FPGA

Random Access Scan. Veeraraghavan Ramamurthy Dept. of Electrical and Computer Engineering Auburn University, Auburn, AL

Training Note TR-06RD. Schedules. Schedule types

Testing Digital Systems II

Low Power VLSI Circuits and Systems Prof. Ajit Pal Department of Computer Science and Engineering Indian Institute of Technology, Kharagpur

The outputs are formed by a combinational logic function of the inputs to the circuit or the values stored in the flip-flops (or both).

Design for Test. Design for test (DFT) refers to those design techniques that make test generation and test application cost-effective.

BUSES IN COMPUTER ARCHITECTURE

UNIT IV CMOS TESTING. EC2354_Unit IV 1

BTC and SMT Rework Challenges

Full Disclosure Monitoring

A Design Language Based Approach

VLSI Design: 3) Explain the various MOSFET Capacitances & their significance. 4) Draw a CMOS Inverter. Explain its transfer characteristics

Testing Digital Systems II

Lecture 18 Design For Test (DFT)


A Briefing on IEEE Standard Test Access Port And Boundary-Scan Architecture ( AKA JTAG )

BOARD TEST The powerful combination of flying probe test and JTAG test speeds up testing

Synchronous Sequential Logic

Highly Accelerated Stress Screening of the Atlas Liquid Argon Calorimeter Front End Boards

Using on-chip Test Pattern Compression for Full Scan SoC Designs

VLSI Technology used in Auto-Scan Delay Testing Design For Bench Mark Circuits

At-speed Testing of SOC ICs

Course 10 The PDH multiplexing hierarchy.

Testability: Lecture 23 Design for Testability (DFT) Slide 1 of 43

Synchronous Sequential Logic

TODAY, the use of embedded systems in safety-critical

Children cannot always recognize potential hazards properly. This 5.1 system is not designed for operation in a heavy industry environment.

EXA PH200/400 and EXA PH202/402 Troubleshooting and Error Code Guide

Optimization of Multi-Channel BCH Error Decoding for Common Cases. Russell Dill Master's Thesis Defense April 20, 2015

Chapter 4. Logic Design

NH 67, Karur Trichy Highways, Puliyur C.F, Karur District UNIT-III SEQUENTIAL CIRCUITS

EE241 - Spring 2001 Advanced Digital Integrated Circuits. References

Implementation of a High-Speed Distribution Network Reconfiguration Scheme by Greg Hataway, Ted Warren, and Chris Stephens.

(Refer Slide Time 1:58)

Self Restoring Logic (SRL) Cell Targets Space Application Designs

Counter/timer 2 of the 83C552 microcontroller

ALGORHYTHM. User Manual. Version 1.0

* This configuration has been updated to a 64K memory with a 32K-32K logical core split.

MC9211 Computer Organization

Previous Lecture Sequential Circuits. Slide Summary of contents covered in this lecture. (Refer Slide Time: 01:55)

Failure Analysis Technology for Advanced Devices

[Krishna*, 4.(12): December, 2015] ISSN: (I2OR), Publication Impact Factor: 3.785

Vignana Bharathi Institute of Technology UNIT 4 DLD

International Journal of Scientific & Engineering Research, Volume 5, Issue 9, September ISSN

Fieldbus Testing with Online Physical Layer Diagnostics

Failure Modes, Effects and Diagnostic Analysis

1C.5.1 Voltage Fluctuation and Flicker

User Manual CC DC 24 V 5A. Universal Control Unit UC-1-E. General Information SET. Universal Control Unit UC-1 Of Central Lubrication PAUSE CONTACT

This Chapter describes the concepts of scan based testing, issues in testing, need

LFSR Counter Implementation in CMOS VLSI

Fault Detection And Correction Using MLD For Memory Applications

Processor time 9 Used memory 9. Lost video frames 11 Storage buffer 11 Received rate 11

Bringing an all-in-one solution to IoT prototype developers

Y. Tsiatouhas. VLSI Systems and Computer Architecture Lab. Built-In Self Test 2

Long and Fast Up/Down Counters Pushpinder Kaur CHOUHAN 6 th Jan, 2003

ORM0022 EHPC210 Universal Controller Operation Manual Revision 1. EHPC210 Universal Controller. Operation Manual

Enhancing Performance in Multiple Execution Unit Architecture using Tomasulo Algorithm

EMI/EMC diagnostic and debugging

(Refer Slide Time: 2:00)

Timing Error Detection: An Adaptive Scheme To Combat Variability EE241 Final Report Nathan Narevsky and Richard Ott {nnarevsky,

Combinational vs Sequential

DISTRIBUTION STATEMENT A 7001Ö

2 Work Package and Work Unit descriptions. 2.8 WP8: RF Systems (R. Ruber, Uppsala)

Based on slides/material by. Topic 14. Testing. Testing. Logic Verification. Recommended Reading:

CMOS Testing-2. Design for testability (DFT) Design and Test Flow: Old View Test was merely an afterthought. Specification. Design errors.

DESIGNING AN ECU CPU FOR RADIATION ENVIRONMENT. Matthew G. M. Yee College of Engineering University of Hawai`i at Mānoa Honolulu, HI ABSTRACT

Hardware Implementation of Viterbi Decoder for Wireless Applications

A video signal processor for motioncompensated field-rate upconversion in consumer television

Logic Design ( Part 3) Sequential Logic- Finite State Machines (Chapter 3)

MICROMASTER Encoder Module

High Performance Dynamic Hybrid Flip-Flop For Pipeline Stages with Methodical Implanted Logic

Sequential Logic Basics

Gamma instabus. Technical product information

BER MEASUREMENT IN THE NOISY CHANNEL

An Efficient High Speed Wallace Tree Multiplier

A New "Duration-Adapted TR" Waveform Capture Method Eliminates Severe Limitations

Choosing an Oscilloscope

Flip-Flops and Related Devices. Wen-Hung Liao, Ph.D. 4/11/2001

Precision testing methods of Event Timer A032-ET

Dimming actuators GDA-4K KNX GDA-8K KNX

Product Update. JTAG Issues and the Use of RT54SX Devices

How to Predict the Output of a Hardware Random Number Generator

Transcription:

A study of intermittent faults in digital computers by OMUR T ASAR and VEHBI T ASAR University of Detroit Detroit, Michigan ABSTRACT Definition of intermittent faults in digital computer systems and their possible causes are given. The effects of intermittent faults on the performance of systems and alternatives to overcome these effects have been examined. Present attempts to cope with the intermittent failures and future research areas have been explored. INTRODUCTION The advanced system architecture of the current computers has had an impact on reliability and maintainability concepts. Although there have been evolutionary inventions in hardware and software built into computers, the complexity of the whole system makes the problem of fault isolation difficult. It is of vital importance to keep a system running as well as producing the system with the utilization of the best engineering knowledge. Hence one should be concerned with providing high availability and easy maintainability at the design stage. Due to insufficient attention devoted to this aspect, many of the machines in the field ~ ha.\i,e>heeil.~~,frowe.poor <Ieliability.awi maintainability. In the so-called space age, it is hard to believe that systems can go down unexpectedly or cause interrupts for no apparent reason and groups of people wander around for many hours, even days to find the problem. All the maintenance procedures embodied in the machine do not give any hint. The field engineer does his best and the machine does not respond. At some point in time, the machine chooses to run due to an unknown action and there is almost no information on why, how, where and when this happened. This dramatized description is quite realistic for a number of current installations. The problem causing the above situation is called the intermittent fault. In most systems 80 to 90 percent of faults are estimated to be intermittent. These faults account for more than 90 percent of total maintenance expense because they are difficult to detect and isolate. INTERMITTENT PROBLEM DEFINITION Intermittents were first defined as faults over which the user had little or no control. This definition still holds true. Today some people define intermittents as failures that are not reproducible since the conditions causing or surrounding an intermittent are often not known. The inherent inconsistency in the occurrence of such faults is a major hindrance to a detailed analysis of these conditions. Intermittents can be defined as random failures that prevent the proper operation of a unit for a short period implying that the duration of failures is not long enough for the application of a test procedure designed for permanent faults. Some intermittents are known to occur due to external effects such as temperature, humidity, vibration, power fluctuation, pollution, pressure, and electromagnetic fields. Assuming that the system runs under specified external conditions, this class of faults constitutes a small portion of intermittents. The nonenvironmental intermittents, namely the faults that are within the system, seem to be the basic source of irritation to the user. There is a difference of opinion as to the behavior of this class of intermittents. Some believe there actually exists a permanent failure which is stimulated onlywj.der. a ~er1ain sequenc.eof ey~nts.~~~(l~eqqeiltly.j it appears as intermittent. The other opinion is that portions or components of the system malfunction intermittently. For example, loose connections, resistance variations and partially defective components may cause such faults. Examples of both opinions have been observed in practice. There are also intermittents that eventually become permanent failures. Such intermittents are caused by deteriorating or aging components. Once they become permanent, they can be detected by existing procedures. However this transition can take from a few minutes to several months during which the frequency of intermittents increases intolerably. To speed up this transition, environmental conditions can be drastically changed to unfavorable levels. Sometimes this technique, known as the stress technique, works but it may inflict new damage in other parts of the system. The source of nonenvironmental intermittents can be 807

808 National Computer Conference, 1977 software as well as hardware. It has not yet been established which causes the major part of such intermittents. However, the general trend is to treat them as hardware oriented. REVIEW OF LITERATURE ON INTERMITTENT FAULTS The first experiments on the effects of intermittent faults were conducted by Ball, Hardie and Suhocki in 1966. 1,2 A highly sophisticated logic simulator was developed for the purpose of normal and/or fault simulation of the Saturn V Launch Vehicle aerospace computer which was being designed by IBMo The computer is a binary; fixed point, serial machine employing triple modular redundancy to provide very high reliability. The simulator was capable of analyzing single and multiple, permanent and intermittent faults. Nearly 800,000 intermittents were simu!ated to obtain a reliable statistical sample. The duration of these intermittents varied between 500 nanoseconds (one clock time) to five milliseconds. They were injected at randomly selected points of combinational and sequential logic circuits in the arithmetic-instruction and multiply and divide units, at the execution time of the simulator. The analysis of the intermittents included a record of the time of occurrence, time and number of detections and the number of intermittents that caused a difference from the correct output. The probability of detection was calculated based upon this information. The authors revealed interesting as well as important conclusions from the stimulation output. These can be summarized as follows. Only 8.3 percent of the total intermittents injected caused the computer to perform incorrectly. The combinational logic was less sensitive to intermittents. In other words, the probability of detection in combinational logic was smaller compared to sequential logic. A single intermittent fault of one clock period was almost undetectable. Single intermittents of longer duration had very low probabilities of preventing the correct operation. The probability of detection was directly proportional to the duration of intermittents and highly dependent on computer modules. The authors experienced that field failures in aerospace computers tend to be intermittent. In commercial computers, intermittents have been estimated to constitute 90 percent of all field failures. Hence, maintenance cost is mainly due to intermittents. The authors believed that most intermittents were due to insufficient testing at the factory. They proposed to develop better detection procedures as a partial solution to the intermittent problem. The other approach was to design equipment insensitive to intermittents introducing retry and/or redundancy. IBM has also tried to deal with intermittents in systems 360 and 370. 3,4 Here the objective was to reduce the number and length of unscheduled maintenances and to minimize the impact of intermittents on system availability. Automatic error recording at the instant of discovery followed by error recovery was proposed as a feasible solution. The so called Recovery Management of IBM follows several steps to achieve the above objectives. Instruction retry can be affected in the 110 area, central processor and main storage areas. Selective termination enables the system to examine the failing environment while all other jobs are running. These functions are incorporated in Recovery Management in four steps. First, a functional recovery is attempted. If the retry of the interrupted operation is successful, the fault becomes transparent to the user. If not, the next function is System Recovery where a selective termination is effected to analyze the failure. Then the system-supported restart is tried without stopping for repair. If a stop is required, system repair utilizes all the detailed error analysis records. To perform a!! these steps, Recover}' I'.,1anagement has 1'0 De\"icc/Unit Recovery, Channel Recovery, I/O Recovery Management, CPU/Processor Storage Recovery, System Associated Recovery and Error Record Retrieval facilities. Honeywell's proposal to overcome the intermittent problem was also retry. The Honeywell 6000 was implemented as a retryable processor. Maestri discusses the problems encountered and the design proposal to avoid such problems. 5 Some instructions cause a destructive read of a memory location. Therefore, it is necessary to restore that memory location before retry can be attempted. For this purpose, a buffer register has to be added. Data can be held in this register until error detection and correction codes report correct data recovery. Instruction retry for a MOVE may cause problems if a data block overlays another block, partly destroying the latter. To avoid this, snapshot registers are added where at the occurrence of a fault the state of the cycle control flags and address register could be saved. Snapshot registers can also be used as a diagnostic aid. To retry instructions with indirect addressing, several methods are proposed. One way is to obtain a pointer to the first indirect word. Restoring should take precautions against parity errors in the updated word and double updating. The second approach requires a scratchpad memory where the state of sequence control flags and memory addresses are saved for every cycle of an instruction. Software can then handle restoration of indirect words assuring no error. Multicycle instructions change the contents of registers with the speed of the adder cycle. In order to retry such instructions, intermediate registers are needed to protect the contents of primary registers and to keep up processor speed. These can be placed at the inputs of the adder on lines from memory and primary registers. Data can be held here until error checking is completed. To solve problems due to instruction overlap, four instruction counters are added for simultaneously executed instructions. If an error occurs, the instruction counter of the failing cycle can be identified either selectively or by a program utilizing a failure flag included in the scratchpad memory. The addition of the above mentioned registers increases the cost of implementation. However, the new processor is almost 100 percent effective in doing instruction retry. The cost advantage can be justified bearing in mind that instruction retry is 80 to 90 percent successful. Honeywell now

A Study of Intermittent Faults in Digital Computers 809 even states, in sales literature, that the 6000 system is capable of handling 95 to 97 percent of all intermittent failures. The other approach to coping with faults in digital computers has been the introduction of redundancy. Research has shown that Triple Modular Redundancy (TMR) can mask solid as well as intermittent faults. If TMR is applied at the module level, then the effect of a single intermittent can be permanent in a sequential circuit. 6 In case an intermittent induces a faulty state, it must be corrected with a resynchronization sequence before a second failure occurs. TMR has been used in some space projects; however, there has been no commercial application. Totally self-checking check circuits have been designed. Such circuits are capable of detecting failures in themselves during normal operation. 7 If intermittents last long enough, they affect the output in such a way that they can be detected. Theoretical modeling of intermittent faults has been attempted in recent years. Breuer presented a two-state first order Markov model to represent the intermittents. 8 Figure 1 shows the two states: the faulty state (N 1) and the normal state (N2). Since the Markov model is probabilistic, each state and every transition is associated with probabilities. With respect to intermittents, a circuit either operates normally or it possesses an intermittent. Let p be the probability of having an intermittent, then the normal state will have a probability of (l-p). To apply this model to any circuit, first of all the parameters, namely p, r, s, have to be estimated. A time interval (T,T') is considered for fault analysis. Fault patterns are defined over this interval as follows: d=( d Q, d Q - l,, d 2, d l ). The subscripts refer to the number of clocks in this interval. d=(o, 0,..., 0, 0) implies a normal circuit. d=(1, 1,..., 1, 1) represents a solid fault. d=(o, 0,..., 1, 0, 0) is the fault pattern of an intermittent occurring in the third clock period. The probability of each fault pattern can be calculated using the Markov model. A complete test set will be obtained if test sets are generated for all possible fault patterns. A modified f'#.f;l:~~bi,"!l~ -for- f@si ~aeratiqq that makes. us.e of two types of gates called d-or and d-and. d-or has the property that its output is ad, if and only if at least one of its inputs is a d or d; otherwise its output is O. In other words, if any input of d-or receives a fault, it can propagate it. The output of d-and is ad, if and only if all of its inputs have a d; otherwise, it is a O. Test generation 5 reg : ~s = (1-5) r = (l-r) Figure I-Markov model requires that the combinational logic be evaluated the same number of times as the number of fault patterns present. In sequential logic analysis, -the circuit evaluation should be repeated for every fault pattern in every clock period within the time interval (T,T'). Hence, even for a small circuit the test generatio~ is a very lengthy procedure. Every test generated has a certain probability of detection. If a confidence level can be specified, then a lower bound for the number of tests that are required to detect an intermittent can be calculated. Kamal's model is also probabilistic. 9 It is based on pattern recognition techniques. The analysis covers well behaved signal independent single intermittents in nonredundant circuits. A state space includes the state of being faultfree-the states possessing an intermittent that do not affect the output and the states possessing an intermittent that affect the output. The test set is the set of tests that are generated to detect permanent faults. The proposed solution is the repeated application of these tests. It is proven in the paper that an intermittent can be detected by infinite number of repetitions of a test. The probabilistic model is utilized to find the finite number of repetitions satisfying a confidence level. The model requires estimation of the probability that an intermittent occurs and the conditional probability of the output being affected given an intermittent has already occurred. The probability of detection given an intermittent is calculated according to Bayes' rule. After each application of a test, this figure is updated. This probability approaches 1 with an infinite number of applications. However, the procedure can be stopped once the confidence level is reached or the number of repetitions can be calculated in terms of the confidence level and the other probabilistic parameters. In most cases this number is very large. Kamal provides an optimization procedure by the use of integer programming. Diagnosis procedure follows the same approach.lo Here it is assumed that an intermittent has been detected. Using the test sets and the set of permanent faults, a coverage table is generated. From this table a test set is chosen that covers the permanent faults. These tests are repeatedly applied until one fails. This reduces the fault table and the po~ iqle f<;lult.. The aboye cycle is repeated until the fault set consists of a single fault. -If th~re- ~~ists no test set'to cover all possible faults, the diagnosis experiment is again terminated. The length of the experiment is calculated making use of the model and is shown to be finite. The fault table is also analyzed to obtain minimum diagnostic resolution. Optimization of the diagnosis procedure is not offered. There are two major difficulties in the application of Breuer's and Kamal's proposals. First, it is very hard to obtain realistic estimates of the model parameters using the current available data. Even if an extensive experiment is conducted on a certain system, the parameters would apply to this particular system. The unpredictable behavior of intermittents may also impose an update on the parameters. The second problem is the huge number of tests required to detect and diagnose intermittents. In this respect, the proposed procedures may not be economically feasible. Even the detection of solid faults are far more expensive then desired.

810 National Computer Conference, 1977 Intermittent faults have been analyzed at the circuit level. Yen discusses intermittents due to noise in four phase MOS gates ll which are widely used in LSI circuits. The problem is a charge redistribution problem. It is possible that the charge at some output node will decrease to critical levels causing intermittents. The situation does not fit the stuck-at model and testing will not be successful in detecting this type of failure. Yen proposes the addition of prechargers to avoid the problem. The number of the prechargers can be optimized by the use of computer-aided circuit simulation in order to decrease the additional hardware cost. ALTERNATIVES FOR SOLUTION No ideal method has been established to diagnose intermittent faults so far. Academic work offers solutions on restricted models that do not realistically represent the intermittent space. Industrial efforts concentrate on increasing the availability, hence they are mainly concerned with the minimization of the effects of intermittents. Up to now, the former has found no application. On the other hand, commercial manufacturers introduced retryable systems that recover from intermittents in a very short period. Although retryability is not a method to diagnose or to avoid the intermittents, it seems to be the only feasible way of recovering from such faults with minimum impact. Indeed, it is difficult to detect and diagnose something that goes on and off unpredictably, and retry ability seems to be the only solution at the systems level currently and in the near future. Intermittents at the circuit or card level should be viewed as a completely different problem. At factory testing, intermittents account for only 30 percent of all faults, whereas in the field they are estimated to constitute 90 percent of all faults. Furthermore, the consequences of intermittents occurring at the system level are far more serious and costly compared to their effects at the testing level. Therefore, our efforts should concentrate on methods dealing with system intermittents. Here we discuss retry and other possible alternatives and where and how they can be effective in fighting intermittents. Instruction retry is not a new concept. It has been a regular procedure that is carried out whenever an error occurs in the 110 area in reading or writing a tape. The new attempt is to extend this feature to central processor and memory units. IBM has been working on this since 1967; Honeywell since the early 70's. Burroughs has recently expended efforts on making its large scale processors retryable. The outcome indicates that these systems can survive 95 percent of intermittents. 100 percent effectiveness is also possible. Since the chance of success in retrying is.8 to.9, the system will have 75 to 85 percent less faults that are apparent to the outside world. In other words, instead of giving an interrupt, the system continues to operate successfully with a few microseconds or milliseconds of delay, caused by the execution of retry. Recovery alone does not help to isolate the intermittent, but an immediate remedy is achieved with respect to the operation of the system. Adequate error recording serves as a diagnostic aid during scheduled maintenance. Utilization of these records toward a test program will be discussed later. Retry requires both hardware and software support. Intermediate and snapshot registers and scratchpad memories have to be added in order to be able to save the status of the system prior to an intermittent. Only then can the system go back and retry the failing state. Even with the addition of the sophisticated retry software, the implementation is not costly compared to the maintenance cost due to intermittents. Today it is common practice to use stress testing to change an intermittent to a solid failure. Marginal voltage and timing conditions can be set or thermal stress can be applied such that slow rise times, low switching thresholds and race c.onditions are amplified. If test and diagnostic programs are run during the application of mechanical stress, faults due to loose connections, defective connectors or printed circuits can be caught. The disadvantage of this method is the dedication of the processor to maintenance procedures as well as the possible infliction of new damage due to thermal and mechanical stress. By all means retry is a better approach than this technique. Retry is aimed at continuing the operation of the system. Whether it is successful or not, a detailed error recording is a major part of this scheme. If retry is not successful, the analysis of the error recording prior and at the instant of error occurrence would assist the personnel to determine the cause of the error in an easier and faster way. If retry is successful, then the error recording dump would be analyzed during scheduled maintenance to determine the cause of the intermittent. It may be advantageous to replace or repair the part causing the intermittent to avoid future possible intermittents. Although retry requires a short amount of time, it is more desirable not to have any interruptions. If this policy is continued persistently, the rate of interruptions due to intermittents may gradually decrease. Another important outcome of detailed error recording is the possible generation of a test program that diagnoses intermittent faults. The accumulated error recording of a certain period may well include information about the most frequent intermittents such as the rate of occurrence, the effect of the intermittent, the symptoms and the fastest correction procedure. Similar to the detection procedure, given an intermittent and its symptom(s), we can run the above mentioned program and come out with an intermittent or a set of possible intermittents that can be examined. Isolation of the intermittent will be an easy matter if it actually belongs to th~ group of the intermittents under consideration. If not, the intermittent can be located by existing means and the test program can be expanded to include the new situation. As time progresses, one may come up with a complete intermittent diagnostic routine. Realization of a diagnostic routine for intermittents requires extensive error recording on a retry able processor. Diagnosis can be achieved by another practical method called dynamic monitoring. As the name implies, dynamic monitoring involves a continuous scan of the machine to be able to detect a fault at the instant of occurrence. Other-

A Study of Intermittent Faults in Digital Computers 811 wise, there is practically no way of reproducing the same event for purposes of detection and diagnosis. This scheme can be realized by placing test points and interrogating the values of these test points at every clock period. If the machine has retry capability, dynamic monitoring can be very successful in isolating intermittents. The values of test points can be continuously stored in a monitor memory, the size of which depends on the number of test points and the number of clock periods that have to be considered. In case of an interrupt, one of the test points will have a faulty value. These values can be stored in a memory for further reference. If the retry is successful, the particular test point will have the correct value. Comparing the two maps of test points before and after retry, the faulty test point can be isolated. To locate the fault causing component, it would suffice to examine the components feeding that test point. Although dynamic monitoring requires test points and a memory of considerabie size, the advantages are apparent. The fault is actually isolated during the operation of the system. All intermittents are detected regardless of their cause. If statistics are maintained, deteriorating components can be isolated and future faults predicted. The feasibility, development and application of this technique has to be considered in conjunction with retry capability. If retry is not available, the technique can be implemented by straight dumping of the recorded memory. The basic problem in realizing this scheme is the optimal location of the test points. The above discussion refers to the solution of intermittents at the systems level. Dynamic monitoring can also be employed in testers that are used to screen circuits and cards. At the circuit level, the models presented by Kamal and Breuer may also be applicable. Current automatic testers operate at a very high speed. Hence, a very large number of tests can be applied in a short period. If careful factory screening is desired, these techniques can be incorporated with the testers. Although the cost of testing will increase, less faults will escape the factory screening. Economic feasibility of these two models can be studied by the use of computer-aided simulation. ACKNOWLEDGMENT The authors wish to thank Mr. R. E. Stackhouse for his guidance and comments during the preparation of this paper. This work was supported by the Burroughs Co. through an internship program. REFERENCES 1. Ball, M. and F. Hardie, "Effects and Detection oflntennittent Failures in Digital Systems," 1969 Fall Joint Computer Conference, AFIPS Conference Proceedings, Vol. 35, Montvale, N. J., AFIPS Press, 1969, pp. 329-335. 2. Hardie, F. H. and R, J. Suchocki, "Design and Use of Fault Simulation for Saturn Computer Design," IEEE Trans. Computers, Vol. EC-16, August 1967, pp. 412-429. 3. Carter, W. C., H. C. Montgomery, R. J. Preiss and H. J. Reinheimer, "Design of Serviceability Features for the IBM SystemJ360," IBM Journal, April 1964, pp. 115-126. 4. Droulette, D. L., "Recovery through Programming SystemJ360-SystemJ 370," 1971 Spring Joint Computer Conference, AFIPS Conference Proceedings, Vol. 38, Montvale, N.J., AFIPS Press, 1971, pp. 467-476. 5. Maestri, G. H., "The Retryable Processor," 1972 Fall Joint Computer Conference, AFIPS Conference Proceedings, Vol. 41, Montvale, N.J., AFIPS Press, 1972, pp. 273-277. 6. Wakerly, J. F., "Transient Failures in Triple Modular Redundancy Systems with Sequential Modules," IEEE Trans. Computers, Vol. C-24. May 1975, pp. 570-573. 7. Anderson, D. A., and G. Metze, "Design of Totally Self-Checking Check Circuits for m-out-of-n Codes," IEEE Trans. Computers, Vol. C- 22, March 1973, pp. 263-269. 8. Breuer, M. A., "Testing for Intennittent Faults in Digital Circuits," IEEE Trans. Computers, Vol. C-22, March 1973, pp. 241-246. 9. Kamal, S., and C. V. Page, "Intermittent Faults: A Model and Detection Procedure," IEEE Trans. Computers, Vol. C-23, July 1974, pp. 713-719. 10. Kamal, S., "An Approach to the Diagnosis of Intermittent Faults." IEEE Trans. Computers, Vol. C-24, May 1975, pp. 461-467. II. Yen, Y. T., "Intermittent Failure Problems of Four-phase MOS Circuits," IEEE Journal of Solid-State Circuits, Vol. SC-4, June 1969, pp. 107-110.