Illumination-based Real-Time Contactless Synchronization of High-Speed Vision Sensors

Similar documents
Illumination-Based Synchronization of High-Speed Vision Sensors

1ms Column Parallel Vision System and It's Application of High Speed Target Tracking

Efficient 500 MHz Digital Phase Locked Loop Implementation sin 180nm CMOS Technology

PEP-II longitudinal feedback and the low groupdelay. Dmitry Teytelman

Application Note AN-708 Vibration Measurements with the Vibration Synchronization Module

Data flow architecture for high-speed optical processors

INTERNATIONAL JOURNAL OF ELECTRONICS AND COMMUNICATION ENGINEERING & TECHNOLOGY (IJECET)

A New Overlap-Scan Circuit for High Speed and Low Data Voltage in Plasma-TV

A Real Time Infrared Imaging System Based on DSP & FPGA

Communication Lab. Assignment On. Bi-Phase Code and Integrate-and-Dump (DC 7) MSc Telecommunications and Computer Networks Engineering

A 5-Gb/s Half-rate Clock Recovery Circuit in 0.25-μm CMOS Technology

Chapter 5 Flip-Flops and Related Devices

Logic Analyzer Triggering Techniques to Capture Elusive Problems

Fig. 1. The Front Panel (Graphical User Interface)

IT T35 Digital system desigm y - ii /s - iii

ECE 5765 Modern Communication Fall 2005, UMD Experiment 10: PRBS Messages, Eye Patterns & Noise Simulation using PRBS

Chapter 6: Real-Time Image Formation

TV Synchronism Generation with PIC Microcontroller

Rec. ITU-R BT RECOMMENDATION ITU-R BT * WIDE-SCREEN SIGNALLING FOR BROADCASTING

2 MHz Lock-In Amplifier

Major Differences Between the DT9847 Series Modules

CS3350B Computer Architecture Winter 2015

Investigation of Digital Signal Processing of High-speed DACs Signals for Settling Time Testing

Asynchronous inputs. 9 - Metastability and Clock Recovery. A simple synchronizer. Only one synchronizer per input

CS61C : Machine Structures

UNIT V 8051 Microcontroller based Systems Design

Clocking Spring /18/05

LOCAL DECODING OF WALSH CODES TO REDUCE CDMA DESPREADING COMPUTATION. Matt Doherty Introductory Digital Systems Laboratory.

Large Area, High Speed Photo-detectors Readout

ME EN 363 ELEMENTARY INSTRUMENTATION Lab: Basic Lab Instruments and Data Acquisition

Techniques for Extending Real-Time Oscilloscope Bandwidth

Integration of Virtual Instrumentation into a Compressed Electricity and Electronic Curriculum

Reading an Image using CMOS Linear Image Sensor. S.R.Shinthu 1, P.Maheswari 2, C.S.Manikandababu 3. 1 Introduction. A.

Journal of Theoretical and Applied Information Technology 20 th July Vol. 65 No JATIT & LLS. All rights reserved.

Realizing Waveform Characteristics up to a Digitizer s Full Bandwidth Increasing the effective sampling rate when measuring repetitive signals

Camera Interface Guide

Experiment 13 Sampling and reconstruction

A MISSILE INSTRUMENTATION ENCODER

Figure 1 shows a simple implementation of a clock switch, using an AND-OR type multiplexer logic.

PICOSECOND TIMING USING FAST ANALOG SAMPLING

AD9884A Evaluation Kit Documentation

Laboratory 1 - Introduction to Digital Electronics and Lab Equipment (Logic Analyzers, Digital Oscilloscope, and FPGA-based Labkit)

Modified Sigma-Delta Converter and Flip-Flop Circuits Used for Capacitance Measuring

Data Conversion and Lab (17.368) Fall Lecture Outline

CS 110 Computer Architecture. Finite State Machines, Functional Units. Instructor: Sören Schwertfeger.

A low-power portable H.264/AVC decoder using elastic pipeline

A New Hardware Implementation of Manchester Line Decoder

Objectives. Combinational logics Sequential logics Finite state machine Arithmetic circuits Datapath

CS61C : Machine Structures

Design of High Speed Phase Frequency Detector in 0.18 μm CMOS Process for PLL Application

AN ARTISTIC TECHNIQUE FOR AUDIO-TO-VIDEO TRANSLATION ON A MUSIC PERCEPTION STUDY

Hardware Implementation of Viterbi Decoder for Wireless Applications

The XYZs of Logic Analyzers

Samsung VTU11A0 Timing Controller

More Digital Circuits

Efficient Architecture for Flexible Prescaler Using Multimodulo Prescaler

Digital Signal Processing Laboratory 7: IIR Notch Filters Using the TMS320C6711

Chapter 4: One-Shots, Counters, and Clocks

8 DIGITAL SIGNAL PROCESSOR IN OPTICAL TOMOGRAPHY SYSTEM

Digital Systems Based on Principles and Applications of Electrical Engineering/Rizzoni (McGraw Hill

Reconfigurable FPGA Implementation of FIR Filter using Modified DA Method

CMS Conference Report

Area-Efficient Decimation Filter with 50/60 Hz Power-Line Noise Suppression for ΔΣ A/D Converters

SEQUENTIAL LOGIC. Satish Chandra Assistant Professor Department of Physics P P N College, Kanpur

Lab experience 1: Introduction to LabView

Loop Bandwidth Optimization and Jitter Measurement Techniques for Serial HDTV Systems

Introduction. NAND Gate Latch. Digital Logic Design 1 FLIP-FLOP. Digital Logic Design 1

4. ANALOG TV SIGNALS MEASUREMENT

A dedicated data acquisition system for ion velocity measurements of laser produced plasmas

16 Stage Bi-Directional LED Sequencer

Tiptop audio z-dsp.

CHARACTERIZATION OF END-TO-END DELAYS IN HEAD-MOUNTED DISPLAY SYSTEMS

THE USE OF forward error correction (FEC) in optical networks

EE241 - Spring 2005 Advanced Digital Integrated Circuits

V6118 EM MICROELECTRONIC - MARIN SA. 2, 4 and 8 Mutiplex LCD Driver

Name Of The Experiment: Sequential circuit design Latch, Flip-flop and Registers

NAPIER. University School of Engineering. Advanced Communication Systems Module: SE Television Broadcast Signal.

VARIABLE FREQUENCY CLOCKING HARDWARE

Digitization: Sampling & Quantization

T ips in measuring and reducing monitor jitter

XC-77 (EIA), XC-77CE (CCIR)

BER MEASUREMENT IN THE NOISY CHANNEL

CCD 143A 2048-Element High Speed Linear Image Sensor

Flip-Flops and Related Devices. Wen-Hung Liao, Ph.D. 4/11/2001

Combinational vs Sequential

CPS311 Lecture: Sequential Circuits

Long and Fast Up/Down Counters Pushpinder Kaur CHOUHAN 6 th Jan, 2003

DPD80 Infrared Datasheet

Doubletalk Detection

Chapter 6. sequential logic design. This is the beginning of the second part of this course, sequential logic.

Area Efficient Pulsed Clock Generator Using Pulsed Latch Shift Register

An Improved Recursive and Non-recursive Comb Filter for DSP Applications

A low jitter clock and data recovery with a single edge sensing Bang-Bang PD

Laser Beam Analyser Laser Diagnos c System. If you can measure it, you can control it!

Removal of Decaying DC Component in Current Signal Using a ovel Estimation Algorithm

technical note flicker measurement display & lighting measurement

Synchronous Sequential Logic

2.6 Reset Design Strategy

Logic Design II (17.342) Spring Lecture Outline

TERRESTRIAL broadcasting of digital television (DTV)

Transcription:

Proceedings of the 2008 IEEE International Conference on Robotics and Biomimetics Bangkok, Thailand, February 21-26, 2009 Illumination-based Real-Time Contactless Synchronization of High-Speed Vision Sensors Lei Hou, Shingo Kagami, and Koichi Hashimoto Graduation School of Information Sciences Tohoku University 6-6-01 Aramaki Aza Aoba, Aoba-ku, Sendai 980-8579, Japan {lei, swk, koichi}@ic.is.tohoku.ac.jp Abstract To acquire images of a scene from multiple points of view simultaneously, the acquisition time of vision sensors should be synchronized. In this paper, an illumination-based synchronization technique derived from the phase-locked loop algorithm is proposed. Both Simulation results and experimental results show that the operation of vision sensor can be successfully locked to the corresponding edges of an intensitymodulated LED illumination signal in real time, as long as the feedback gain is empirically chosen with the help of MATLAB simulation. Index Terms Camera Synchronization, Phase-Locked Loop, Visible Light Communication I. INTRODUCTION The emphasis on the vision sensor technology becomes more and more evident in various fields such as automotive, human machine interface, surveillance and security, and industry control. When vision sensors are used to take multiple points of view simultaneously, the acquisition time of the sensors should be synchronized. Usually this synchronization is done through a dedicated bus or a wired network [1], [2]. In this paper, the authors explore into an illumination-based synchronization, rather than the conventional synchronization based on wired buses. This breakthrough is believed to contribute to convenient distribution of vision sensors, meanwhile reducing the propagation delay of synchronization signals as much as possible to ensure small synchronization errors. To realize the illumination-based synchronization, an algorithm based on the phase-locked loop (PLL) theory [3] is proposed, which can generate steadily synchronized output signals in real time. Specifically, in the contactless illuminationbased synchronization system, as shown in Fig. 1 the reference signal is an optical one, in a specific case, an intensitymodulated LED illumination. The output signal of PLL corresponds to a series of the electronic shutter timing of the vision sensor. If a phase error between the reference and the reference signal (illumination) Vision Sensor (Software PLPLL L inside) synchronized frame timing (shutter timing) Fig. 1. Conceptual diagram of the proposed illumination-based synchronization. output builds up, it will be negatively fed back to the frame time length of vision sensor in such a way that the phase error is again reduced to a minimum in real time. In this way, the vision sensor can fire automatically synchronized according to the properties of the modulated reference signal. To implement the proposed technique, we require an illumination source which is intensity-modulated at a frequency on the order of the target vision frame rate (namely, half of the frame rate). If the target frame rate is, for example, 30 Hz, the illumination has to be modulated at 15 Hz and this kind of low-frequency blinking may be annoying to human eyes, and thus impacts on the popularization of the illumination-based synchronization technique. One way to address this issue is to use an invisible light such as an infrared light, and another way is to focus on high frame rate vision techniques [4] so that the modulation effect of illumination is not perceptible to human eyes at all. In this paper, the authors focus on the latter because the former is trivial, and also because synchronizing high frame rate vision sensors is particularly challenging comparing to the state of arts synchronization techniques. II. SYNCHRONIZATION ALGORITHM A. Digital PLL theory PLL was introduced in 1932 by de Bellescize. Precisely, a PLL is a circuit synchronizing an output signal with a 978-4244-2679-9/08/$25.00 2008 IEEE 1750

referenc e s ignal Phase Detector Fig. 2. Low pass filter L(s) Block diagram of DPLL. Voltage- Controlle d Oscillato r output signal reference or input signal in frequency as well as in phase. In particular, a digital PLL (DPLL) is a PLL in which the Fig. 3. shift. Fig. 4. shift. Fig. 5. shift. phase error π/4 Example of the computation of the phase errors, with pi/4 phase phase error π/2 (locked) Example of the computation of the phase errors, with pi/2 phase phase error 3π/4 Example of the computation of the phase errors, with 3pi/4 phase input and the output are binary signals consisting of HIGH or LOW values. The block diagram of a standard DPLL is shown in Fig. 2. The function of the PD is to compute the product of the reference signal and the output signal. The average output of the phase detector, which is generated by the LPF, depends on the phase error θ e between and as shown in Fig. 3, Fig. 4 and Fig. 5. The voltage-controlled oscillator (VCO) generates a square-wave signal of which the frequency is determined by the input in real time. When the difference between the phases of the reference and the output signals is π/2, as shown in Fig. 4, is zero and thus the output frequency is stable at the central frequency of VCO. This state is called the locked state. Otherwise, as shown in Fig. 3 and Fig. 5, the average output of the phase detector exhibits positive or negative values, which makes VCO generate a higher or lower frequency than its central one so that the phase deference can converge to π/2 consequently after some running time. At the present time, due to the large impact of microprocessors, the logic and/or arithmetic operations within the building blocks can be executed by software if the desired operation frequency is not so high. The algorithm proposed in the following parts of this paper is a software PLL for the most components, while the functions of the phase detector and part of the low pass filter are achieved by way of optical analog processing. Software-based function blocks of DPLL, such as phase detector, low pass filter, and digitally controlled oscillator can be designed according to its DPLL prototype. B. Proposed algorithm A digital PLL makes use of the average phase error between the reference signal and the output signal, to adjust the output signal frequency. This average phase error can be expressed as a time-domain correlation integral 1 t f(τ)g(τ) dτ (1) T of them. Here, T is the period of the correlation time window, which should be sufficiently longer than a vision frame period. In a typical DPLL, both and are considered to be square waves of which the highest and the lowest values are 1 and 1, respectively. However, when the reference signal is expressed in an intensity-modulated light, negative signal values are unachievable. Instead, the authors employ a square wave of which the low value is 0 and the high value is 1 as the reference signal. Let f (t) be the original square wave whose 1751

amplitude set is { 1, 1}, then the new is = 1 2 f (t)+ 1 (2) 2 where the illumination is on (bright) while =1, and off (dark) while =0. The DPLL will work on this new reference signal just in the same way as on f (t) as long as is a 50 % dutyratio square wave whose amplitude set is { 1, 1} because 1 T = 1 T = 1 2T = 1 2T where the integration result of 1 2T f(τ)g(τ)dτ (3) (1/2f (τ)/2)g(τ)dτ (4) f (τ)g(τ)dτ + 1 g(τ)dτ (5) 2T f (τ)g(τ)dτ (6) g(τ)dτ = 0. This signal modification is commonly used in many optical implementations of the time-domain correlation [5], [6]. We assign the odd-number frames of vision sensor to the periods where = 1 and the even-number frames to = 1, referring to as the frame state signal. This is shown in Fig. 6. Therefore, the frame rate of vision sensor will be twice the frequency of the illumination signal. Next, we have to consider how the product is computed. Most vision sensors operate in frame-based manners, that is, time-domain integral of incident light brightness over one frame time is obtained as a pixel value at the end of the frame time, and therefore the input value or the product at any arbitrary time instant is not available. However, by considering that is a constant during one frame period, we can obtain the time correlation as 1 f(τ)g(τ)dτ (7) T = 1 ( 1) i 1 F [i] (8) T i where i is the frame number index and F (i) is the pixel value (more precisely, sum of the pixel values is used in the implementation) obtained within frame i. Whereafter, summation over a correlation time window normalized with the window size T is replaced by an IIR low pass filter: LPF[i] =k LPF[i 2] + (1 k) (F [i 1] F [i]) (9) which is recursively computed every two successive frames, where k is the coefficient that determines the characteristics Illumination brightness 0 off frame state on on on off off off odd even odd even odd even summation and low-pass filtering Fig. 6. t t t The proposed algorithm. Integrated Photocurrent Pixel output average output of phase detector of the filter, and LPF[i] is the low pass filter output at the frame i. The term (F [i 1] F [i]) expresses the phase error obtained over the recent two frames. The length of frame time is then adjusted according to this output of low pass filter. The whole procedure is depicted in Fig. 6. To implement this algorithm, we do not need any dedicated pixel structures, unlike some prior proposals [5], [6]. The only issue we have to assume is that we can precisely adjust the frame time length of the sensor in real time, for example, with the help of built-in camera functions or by controlling external trigger signals. To put it the other way around, it is difficult to use the proposed method when there is no means to control the frame time length. It should be noted that the blinking illumination does not disturb the visibility of the scene. This is because the vision sensor gets locked with the π/2 phase shift, the sensor operates so that a half of every frame time is covered by the period when the illumination is on, and the accumulated incident light within one frame time is always constant. Thus the synchronization can be done simultaneously with image acquisition. In the following simulation and experiments, the adjustment of frequency was realized by either prolonging or shortening every frame period in real time during the nonintegration time of a frame. More specifically, a frame time is composed of an integration period in which photons are accumulated and an immediate non-integration period. III. SIMULATION RESULTS The authors carried out numerical simulation for the proposed algorithm to explore feasible parameters and analyze their behaviors. The proposed algorithm was implemented in MATLAB (R2007a) along with the generation of input signal. Con- 1752

Fig. 7. Simulated result of the case 2. The upper figure shows the low pass filter output and the lower shows the relative phase shift of the output signal to the reference. sidering the specification of the vision sensor used in the real experiment described in the next section, we adopted the following setup parameters: basic reference frequency: 500 Hz (that is, the basic vision frame rate is 1,000 Hz), number of pixels: 64 64, pixel value when the illumination is on: 63, pixel value when the illumination is off: 0, low pass filter coefficient (k): 3/4. The sum of all the pixel values in the images is computed and it is used as F [i]. At the primary stage, all sorts of background illumination is not taken into account. However, a comprehensive analysis and discussion will be carried out in the following section of this paper. The setups for the simulation and the results are summarized in Table I. Here, the gain is the ratio of the frame time growth (in 10 6 s) 1 to the output of low pass filter. One of the simulated results, with the case 2 where the gain =0.04, is shown in Fig. 7. The upper figure shows the output of low pass filter and the lower shows the output phase relative to the reference signal. In this case, adjustment of frequency finally converged to the desired value after some slight vibrations. As long as the gain value was within the correct range, the output signal of PLL got synchronized for the most cases. The optimal value will be determined later within this range when being applied to real hardware use. The ratio of the upper and lower limits of the gain with which the system 1 The reason the authors measure the time in 10 6 s is because it is the clock cycle time of the vision sensor used in the real experiment. Arbitrary Wave Generator LED driver reference signal LED Fig. 8. Vision Chip Experiment set up. Oscilloscope frame time signal (reset photodiode) was operational was 20, which means the system works well for a reasonably wide range of the gain setting. As an exception, the case 4 exhibited an unstable behavior where the relative phase shift diverged, although the gain was within the correct range. The cause is not clear yet, and further investigation is needed. IV. EXPERIMENT AND DISCUSSION A. Structure of the experiment system Figure 8 shows the block diagram of the experiment system. A high-speed vision system VCS-IV [4] was used as a vision sensor, which includes a digital vision chip, a CMOS image sensor each pixel of which has a programmable processing element. The reason the authors chose it is because it can operate at high frame rate (e.g. 1,000 fps) and also its frame time, either the integration period or the non-integration period, can be easily adjusted by software. Although it has a capability to execute pixel-parallel image processing programs on the focal-plane processing element array, it is not utilized in this experiment but instead only the summation of the digital pixel values over the array is computed on the focal plane. This pixel sum is used as F [i] just in the same way as done in the simulation. The illumination system consists of a Nissin Electronics LDR-90 LED arrray and an LPR-30W-D power supply system, which are externally drived by a Tektronics AFG3102 arbitrary wave generator. The operation of the vision system was measured by observing the pixel reset signal of the sensor, whose positive edge corresponds to the beginning of a frame, with a Tektronics TDS3034 oscilloscope. If the operation of the vision system is locked to the illumination, synchronized waveforms of the pixel reset and the reference signal are observed in the oscilloscope. B. Experimental results According to the proposed algorithm, the flowchart of the experimental algorithm is shown in Fig. 10. When the LED light directly shed on the vision sensor, as shown in Fig. 9, as long as appropriate parameters such as 1753

Table I SIMULATION SETUPS AND TRACKING RESULTS case reference frequency gain initial phase difference output frequency state 1 500 Hz 0.002 1/5π 1,000 Hz locked 2 500 Hz 0.04 1/5π 1,000 Hz locked 3 500 Hz 0.05 1/5π fluctuant unlocked 4 500 Hz 0.031 1/5π fluctuant unlocked 5 500 Hz 0.04 (n +3/2)π 1,000 Hz locked 6 510 Hz 0.04 1/5π fluctuant unlocked LPF out Start : = 0Signal : = 0; Reset timer A Integration of pixelcurrent N Even frame Y Signal += pixel sum Signal -= pixel sum Fig. 9. Experimental setup of a direct LED illumination case. LPF out (t) : = k LPF + (1 k) Signal out (t 1) the gain were selected, the electronic shuttering time of the vision sensor got synchronized with the π/2 phase shift and twice the frequency of the input signal as shown in Fig. 11. The peak-to-peak jitter of the output signal was also measured with the help of oscilloscope. Although the jitter varied depending on experimental conditions, its typical valve was 32.0 μs, as shown in Fig. 12, Since the expected frequency of vision frame is 1,000 Hz, the jitter is 3.2 %of the frame time, which approves the possibility of illumination-based synchronization. The ranges of appropriate parameters depended on the experimental conditions such as the illuminance on the focal plane, as is instinctively anticipated. Because the proposed algorithm takes difference of pixel values in successive frames as its input, the results do not depend on background illumination, and thus the synchronization was also successful even when the sensor was indirectly illuminated, that is, when the vision sensor observed a scene illuminated by the LED, as long as all the parameters were chosen appropriately. In virtue of the proposed algorithm, the synchronization is robust to small internal frequency errors. Actually, the operation of the vision sensor was stable when the frequency range of the modulated LED light was between 343 Hz and 606 Hz. However, the larger the frequency difference is, the Compute the elapsed time from A Compute the processing time left Frequency adjustment by running empty for loop s Fig. 10. The flowchart of the algorithm. larger the steady-state phase error becomes. The most fatal problem in the proposed system is that it is not robust to changes in illumination signal brightness because the gain parameter must be set in accordance with it, as shown in Fig. 13. This situation can easily happen, for example, when the scene includes moving objects or the vision sensor is moving. This problem must be overcome by introducing a signal normalization technique. V. CONCLUSION An illumination-based synchronization technique based on PLL for high-speed vision sensors has been described. Experimental results show that the sensor operation can be successfully locked to an LED illumination signal as long as 1754

Fig. 11. Successfully synchronized electronic shutter, compared to the reference signal. future work. REFERENCES [1] Point Grey Research Inc., Dragonfly Camera Synchronization, http:// www.ptgrey.com/newsletters/feb2002.html. (as of 2008/01/02) [2] P. K. Rai, K. Tiwari, P. Guha, and A. Mukerjee, A Cost-effective Multiple Camera Vision System using FireWire Cameras and Software Synchronization, 10th Intl. Conf. on High Performance Computing (HiPC 2003), 2003. [3] R. E. Best, Phase-Locked Loops Theory, Design, and Applications, McGraw-Hill, 1976. [4] S. Kagami, T. Komuro, and M. Ishikawa, A High-Speed Vision System with In-Pixel Programmable ADCs and PEs for Real-Time Visual Sensing, 8th IEEE Intl. Workshop on Advanced Motion Control, pp. 439 443, 2004. [5] S. Ando, and A. Kimachi, Correlation Image Sensor: Two-Dimensional Matched Detection of Amplitude-Modulated Light, IEEE Trans. on Electron Devices, Vol. 50, No. 10, 2003. [6] J. Ohta, K. Yamamoto, T. Hirai, and K. Watanabe, An Image Sensor With an In-Pixel Demodulation Function for Detecting the Intensity of a Modulated Light Signal, IEEE Trans. on Electron Devices, Vol. 50, No. 1, 2003. Fig. 12. Peak-to-peak jitter ( readout) of the vision frame. the gain parameter was carefully chosen to fit the illumination signal brightness. This dependency should be removed in Fig. 13. Fluctuant electronic shutter of vision chip because of dynamic disturbance that affected the integration value of average brightness. 1755