Is image manipulation necessary to interpret digital mammographic images efficiently?

Similar documents
Breast screening: visual search as an aid for digital mammographic interpretation training

Usage of any items from the University of Cumbria s institutional repository Insight must conform to the following fair usage guidelines.

The human factors surrounding system change in breast cancer screening: a case study

What to consider when choosing a mammography display

Guidelines for Assuring Softcopy Image Quality

Role of Color in Telemedicine Applications. Elizabeth A. Krupinski, PhD

+ Human method is pattern recognition based upon multiple exposure to known samples.

Intuitive Workflow by Barco. Designed for the way you work, naturally.

Display Quality Assurance: Considerations When Establishing a Display QA Program. Mike Silosky, M.S. 8/3/2017

Coronis 5MP Mammo. The standard of care for digital mammography

Optimizing the Workflow of Radiologists

The Diagnosis of Small Solitary Pulmonary Nodule:

Initial Experience With Automatic Image Transmission to an Intensive Care Unit Using Picture Archiving and Communications System Technology

Ofcom's proposed guidance on regional production and regional programming

White Paper. Uniform Luminance Technology. What s inside? What is non-uniformity and noise in LCDs? Why is it a problem? How is it solved?

Forensic Video Analysis Technical Procedure Manual Page 1

World First Slim Cassette Type Digital Mammo. Upgrade Solution

Raspberry Pi driven digital signage

Guidance for Quality Assurance of PACS Diagnostic Display Devices

BBC Television Services Review

Coronis Uniti (MDMC-12133) 12MP diagnostic display system for PACS and breast imaging

Full-Field Digital Mammography on LCD Versus CRT Monitors

3/2/2016. Medical Display Performance and Evaluation. Objectives. Outline

Clarity and confidence. Raising the bar in diagnostic imaging with Barco radiology and mammography displays

Breast Imaging Essentials

Nio. Industry-standard diagnostic display systems

Mammo Tomosynthesis 5MP

FDA MAMMOSCAN FULL FIELD DIGITAL MAMMOGRAPHY SYSTEM MAMMOGRAPHY

Detectability of breast cancer and reading time on screening mammography: Comparison between 5-MP and 8-MP LCD monitors.

RX850. Your advantages. 8MP Medical-Display

RoMEO Studies 8: Self-archiving when Yellow and Blue make Green: the logic behind the colour-coding used in the Copyright Knowledge Bank

MX215. Your advantages. 2MP Medical-Display

FFDM Quality Control in Canada - a Vendor Neutral Approach

Randell R, Ambepitiya T, Mello-Thoms C, Ruddle R, Brettle D, Thomas R, Treanor D. (2014)

QUALITY CONTROL AND PATIENT DOSES FROM X-RAY EXAMINATIONS IN SOME HOSPITALS IN THAILAND

Coronis Fusion multi-modality displays. The ultimate in diagnostic flexibility

Display Quality Assurance: Recommendations from AAPM TG270 for Tests, Tools, Patterns, and Performance Criteria

Display Quality Assurance: Recommendations from AAPM TG270 for Tests, Tools, Patterns, and Performance Criteria

MODELLING IMPLICATIONS OF SPLITTING EUC BAND 1

MQSA Quality Control Manual

D PSB Audience Impact. PSB Report 2011 Information pack June 2012

1.1 What is CiteScore? Why don t you include articles-in-press in CiteScore? Why don t you include abstracts in CiteScore?

Pearson BTEC Nationals in Music Technology New for applications

Radiography and Mammography Viewing Systems

Assessment of the Impact of Dark Signal on Image Quality in Computerized Mammography. Introduction

Bibliometrics and the Research Excellence Framework (REF)

B - PSB Audience Impact. PSB Report 2013 Information pack August 2013

Hospital Wide. Healthcare Display Solutions DICOM Displays, Large Screen Displays and Projectors

MX215. Your advantages. 2MP Medical-Display

Welsh print online THE INSPIRATION THE THEATRE OF MEMORY:

The Versatile and Powerful ACLxy. ACLxy

S M A R T U LT R A S O U N D

BBC Red Button: Service Review

In accordance with the Trust s Syndication Policy for BBC on-demand content. 2

MX194. Your advantages. 1MP Medical-Display

BBC Trust Review of the BBC s Speech Radio Services

White Paper. Diagnostic Color Displays. What s inside?

Libraries as Repositories of Popular Culture: Is Popular Culture Still Forgotten?

Student Laboratory Experiments Exploring Optical Fibre Communication Systems, Eye Diagrams and Bit Error Rates

The Music-Related Quality of Life (MuRQoL) questionnaire INSTRUCTIONS FOR USE

Selection and Quality Assurance of Monitors for Mammogram Diagnosis

BROADCAST. The following concepts help ensure the way we distribute revenue to members is equitable.

Medical Monitor Solutions

Design and Implementation of a Digital Teleultrasound System for Real-Time Remote Diagnosis

Solution for Nonuniformities and Spatial Noise in Medical LCD Displays by Using Pixel-Based Correction

BBC Trust Changes to HD channels Assessment of significance

FIM INTERNATIONAL SURVEY ON ORCHESTRAS

Don t Skip the Commercial: Televisions in California s Business Sector

explain manufacturers specifications for a range of audio recording equipment P1 In order to get a merit in this task you will need to

The BBC s management of the costs of producing continuing drama

The Effects of Web Site Aesthetics and Shopping Task on Consumer Online Purchasing Behavior

Working With Music Notation Packages

Suggested Publication Categories for a Research Publications Database. Introduction

BERMUDA STATUTORY INSTRUMENT BR 25/1987 TELEVISION BROADCASTING SERVICE REGULATIONS 1987

Interactive Virtual Laboratory for Distance Education in Nuclear Engineering. Abstract

MANAGING HDR CONTENT PRODUCTION AND DISPLAY DEVICE CAPABILITIES

Entry Level Assessment Blueprint Audio-Visual Communications Technology

Comed Medical Systems Co., Ltd. Office 707, Woolim Lion s Valley I, 311-3, Sangdaewon-dong, Seongnam-si, Gyeonggi-do, Korea Tel:

Nio. Industry-standard diagnostic display systems

Global Trade Medical Supplies

White Paper : Achieving synthetic slow-motion in UHDTV. InSync Technology Ltd, UK

Request for Proposals

CARESTREAM DIRECTVIEW Elite CR System

PROCEEDINGS OF SPIE. Event: SPIE Defense, Security, and Sensing, 2013, Baltimore, Maryland, United States

CARESTREAM DIRECTVIEW Elite CR System

Making Progress With Sounds - The Design & Evaluation Of An Audio Progress Bar

Public Service Broadcasting Annual Report 2011

Semi-automating the manual literature search for systematic reviews increases efficiency

Instructions to Authors

Part III: How to Present in the Health Sciences

Quantify. The Subjective. PQM: A New Quantitative Tool for Evaluating Display Design Options

digital television digital radio analogue spectrum planning

An Efficient Multi-Target SAR ATR Algorithm

Acquisition Control System Design Requirement Document

A HIGHLY INTERACTIVE SYSTEM FOR PROCESSING LARGE VOLUMES OF ULTRASONIC TESTING DATA. H. L. Grothues, R. H. Peterson, D. R. Hamlin, K. s.

Questions to Ask Before Beginning a Digital Audio Project

MUSICAL MOODS: A MASS PARTICIPATION EXPERIMENT FOR AFFECTIVE CLASSIFICATION OF MUSIC

A Comparison of Case Retrieval Times: Film Versus Picture Archiving and Communications Systems

High Frequency X-Ray Technology. Wide Screen Control with Touch-Technology

Equipment Quality Control for Digital Radiography February 22, Imaging Physics CancerCare Manitoba

Transcription:

Loughborough University Institutional Repository Is image manipulation necessary to interpret digital mammographic images efficiently? This item was submitted to Loughborough University's Institutional Repository by the/an author. Citation: CHEN, Y.... et al, 2011. Is image manipulation necessary to interpret digital mammographic images efficiently? IN: Manning, D.J. and Abbey, C.K. (eds.). Proceedings of SPIE, Vol. 7996, Medical Imaging 2011: Image Perception, Observer Performance, and Technology Assessment, Buena Vista, United States, 12th-17th February, pp. 79660S-1-79660S-10. Additional Information: This is a conference paper. Further details of the conference can be found at: http://spie.org/medical-imaging.xml. Copyright c 2011 Society of Photo-Optical Instrumentation Engineers. One print or electronic copy may be made for personal use only. Systematic reproduction and distribution, duplication of any material in this paper for a fee or for commercial purposes, or modification of the content of the paper are prohibited. Metadata Record: https://dspace.lboro.ac.uk/2134/8251 Version: Published Publisher: c The Society of Photo-Optical Instrumentation Engineers (SPIE) Please cite the published version.

This item was submitted to Loughborough s Institutional Repository (https://dspace.lboro.ac.uk/) by the author and is made available under the following Creative Commons Licence conditions. For the full text of this licence, please go to: http://creativecommons.org/licenses/by-nc-nd/2.5/

Is image manipulation necessary to interpret digital mammographic images efficiently? Yan Chen a, Alastair Gale a, Anne Turnbull b, Jonathan James c a Applied Vision Research Centre, Loughborough University, Loughborough, UK; b Breast Unit, Derby Royal Hospital, Derby, UK; c Nottingham Breast Institute, Nottingham City Hospital, Nottingham, UK ABSTRACT With the introduction of digital breast screening across the UK, screeners need to learn how best to inspect these images. A key advantage over mammographic film is the facility to use workstation image manipulation tools. Forty two-view FFDM screening cases, representing malignant, normal and benign appearances were examined by fourteen radiologists and advanced practitioners from two UK screening centres. For half the cases, the mammography workstation image manipulation tools could be employed and for the other half these were not used. Participants classified each case and indicated whether an abnormality was present. Throughout the study the participants visual search behaviour as well as their image manipulations was recorded. Whether or not image manipulation tools were used made very little difference to overall performance (t-test, p>.05) as confirmed by JAFROC analysis Figure-Of-Merit values of 0.816 and 0.838 (with and without tools respectively); performance not using tools was better. However, using tools significantly increased inspection time (p<0.5) as well as participants confidence. Detailed examination of participants image inspection behaviour elicited that the average time on each case in the different viewing conditions differed significantly between the high experienced readers and low experienced readers. The visual data analysis revealed that the participants made similar overall pattern of errors on both modalities. The visual search behaviour on both modalities are surprisingly similar. Keywords: Mammographic interpretation training, low resolution devices, eye movements, HCI 1. INTRODUCTION Breast screening has been undertaken across the UK for over 20 years using mammographic film as the imaging medium [1]. Recent developments will see the age range being increased to encompass woman aged 47-73 years [2]. To cope with the additional workload Full Field Digital Mammographic imaging is being rolled out so that soon all 110 screening centres will have some digital imaging ability with nationwide full digital imaging ensuing. The change to digital allows current screening personnel to manipulate images to optimise appearance. However, there has been little research to show that the image manipulation tools support better performance and also whether it is an efficient way of examining mammographic images. At the same time, clinical workstations are in great demand and if other less expensive monitors could be employed for training, or other non-clinical purposes then this would be very useful. As reported previously [3] a long and detailed study was performed examining screening performance on a workstation, standard computer monitor and an iphone where participants either were able to manipulate the displayed images using typical DICOM viewing software or else were not permitted to manipulate the images in such a manner. Here, more detailed data regarding the comparison of performance and the related underlying factors on the workstation and monitor are presented. In particular the observers visual search and interaction behaviour whilst examining the images on these displays is concentrated upon, particularly where false negative responses were made and where the participants manipulated the images. The fundamental purpose of the overall study was to examine how screeners performance on different display devices varied and whether non-clinical workstation performance was at all at a level to imply that these could be used instead of workstations to offer training (potentially outside the radiology department). Here the purpose of the investigation is to report in detail how performance on a workstation and a monitor varied. A group of screeners was presented with sets of recent difficult FFDM screening cases on a mammographic workstation and asked to examine these images either with or without using any image manipulation functions. Their performance Medical Imaging 2011: Image Perception, Observer Performance, and Technology Assessment, edited by David J. Manning, Craig K. Abbey, Proc. of SPIE Vol. 7966, 79660S 2011 SPIE CCC code: 0277-786X/11/$18 doi: 10.1117/12.878753 Proc. of SPIE Vol. 7966 79660S-1

and visual search behaviour were measured to determine how using these functions affected their case reading behaviour and abilities. Three UK screening centres were selected which each had had GE digital mammography equipment for at least four years and all screeners were familiar with the appearance of images generated by these systems. Therefore, nineteen screeners from these three major breast screening centres were approached with detailed study-related information. For logistical reasons (the large amount of experimental equipment required to be transported from centre to centre), it eventually transpired that to conduct the study at one of the centres was not practical. Consequently, fourteen participants from two of the major UK breast screening centres took part. The participants were divided into two groups according to the different screening centres where they primarily worked. 2. METHOD 2.1. Experimental Design Participants: fourteen experienced screeners (including nine consultant radiologists and five advanced practitioners [i.e. technologists]) from two major UK breast screening centres volunteered to undertake the experiments. Visual Stimuli: two sets of 20 challenging digital screening cases (MLO and CC views) which demonstrated difficult examples of normal, benign and malignant appearances were selected and matched by feature type (i.e. masses, calcifications and architectural distortions) and difficulty. Hardware: (1) GE digital mammography workstations (with 5 megapixel dual monitors; resolution 2,048 x 2,560 pixels each); (2) a standard LCD monitor (images were shown using a DICOM viewer running on a laptop, screen size: 21.5, resolution: 1,050 x1,680). Eye Tracking Device: A head mounted eye tracker (ASL 504) was used to monitor the visual search behaviour of the experienced screeners in this study. Procedure: each participant examined the two sets of images with/without using the image manipulation tool (HCI) on both digital workstation and standard monitor, however with a minimum two month gap in between. The participants behaviour whilst examining each image was recorded for subsequent detailed analyses. The procedure has been reported elsewhere in details [3]. 2.2 Behaviour Data Analysis Detailed examination of the sequence of the image manipulation (HCI) data was then carried out. For each participant, their behavioural data for each study round was recorded on two separate videotapes; one for the case set viewed with the support of the image manipulation tools and the other for the case set viewed without the support of such tools. Each tape was transferred to computer disk and saved as an.avi file for data analysis. Each of the.avi files was then reviewed and coded using Captiv L-2100 software. Figure 1 illustrates the video analysis process. This software allows the experimenter to replay the video record of each participant s behaviour and code various predetermined actions as and when they occur during the participant s visual examination of each mammographic case. For instance the start and end of examining each mammographic view, use of image zoom and pan etc. are all recorded. These data are then exported to an Excel spreadsheet for further detailed analyses. Also, to check on any potential subjective interpretation of the video records, one video recording was first analysed using the Captiv software and then the same video was reviewed again independently by another experimenter using the software. The results showed no individual differences in identifying the time sequencing of the actions/events and consequently it was decided that all of the tapes would be analysed by one individual. Proc. of SPIE Vol. 7966 79660S-2

Figure 1. Screen shot of analysis of the imaging interpretation behaviour data using Captiv L-2100 software. The symbol codes on the right were used to identify specific observer actions (e.g. image zooming). 2.3. Visual Data Analysis Participants eye movements were recorded throughout the study. A key part of the interaction with the images was how individuals visually examined and interacted with the cases when examining images on the diagnostic workstation and the standard monitor when the image manipulation tools were used. The raw eye movement data were clustered into fixations using criteria (e.g. number of data samples to define a fixation, the spatial separation of each data sample, and a minimum fixation duration of 200ms) adapted from the ASL eye tracking Eyenal software. Of particular interest was the eye fixation data on all the false negative cases made when participants were examining cases on the workstation using image manipulation and also examining cases on the standard monitor again while using image manipulation. The eye movement data were used to classify the false negative responses into the following three categories: Search error: neither the location of the abnormality, nor the area of interest, attracted any visual attention; Detection error: the location of the abnormality was not reported and the area of interest only attracted visual attention for a short (<1000ms) period of time; Interpretation error: the location of the abnormality was not reported even though the area of interest attracted visual attention for longer than 1000ms. Other information was also collected and analysed, such as: Time to first hit, which is how long and also how many fixations it took from when the image appeared to first hit the area-of-interest (AOI); Dwell time corresponding to the total amount of time spent in the area-of-interest. Proc. of SPIE Vol. 7966 79660S-3

3. RESULTS Overall performance between the digital workstation, whether or not image manipulation tools were used, and the standard monitor, when image manipulation tools were used, was not significantly different (p>.05) as confirmed by JAFROC analysis Figure-Of-Merit values of 0.816, 0.838 and 0.827. These performance data have been reported elsewhere [3]. The key interest here is the participants behaviour data and visual search data analyses. Participants data were also considered as grouped into high (where mammography experience was 10 years and above) and low (i.e. mammography experience was less than 10 years) experience. 3.1. Behaviour Data Data were examined across all participants with regard to the average examination time spent on each case. t-tests showed non-significant differences (p=n.s.) between the time spent examining cases on the standard monitor, using image manipulation tools, and on the workstation using, or not using, the image manipulation tools. See figure 2 for details. 01:43.7 01:26.4 Aveage time spent on each case (mm:ss.0) 01:09.1 00:51.8 00:34.6 00:17.3 00:00.0 SM-HCI W/S-HCI W/S-nonHCI Viewing conditions Figure 2. Average time spent per case comparison between the three viewing conditions. A 2 x 3 mixed design ANOVA revealed that there was a non-significant main effect of mammography experience on the average time spent on each case [F(1, 10)=4.17, p=n.s., r =.54] and a significant main effect of viewing conditions on the average time spent on each case, [F(2, 20)=3.9, p <.05, r =.40]. Contrast revealed that the average time spent on each case while examining on the standard monitor with image manipulation tools was significant longer than if examining on the workstation without using the image manipulation tools [F(2, 20)=6.39, p <.05, r =.49]. There was a significant interaction effect between viewing conditions and participants mammography experience level, [F(2, 20)=3.9, p <.05, r =.40]. This indicates that the average time on each case in the different viewing conditions differed significantly between the high experienced readers and low experienced readers. To break down the interaction, contrast was performed which revealed significant interactions when comparing high and low experienced observers average time on each case: while examining images on the standard monitor with image manipulation tools as compared to examining images on the workstation without image manipulation tools, [F(1, 10)=7.84, p <.05, r =.66] Proc. of SPIE Vol. 7966 79660S-4

and whilst examining images on the workstation with image manipulation tools compared to examining images on workstation without image manipulationn tools [F(1, 10)=14.937, p <.05, r =.77]. Details are shown in figure 3. Figure 3. Average time spent per case comparison between the three conditions x mammography experience 3.2. Eye Movement Data A two-way repeated-measures ANOVAA revealed that there was a non-significant effect of viewing modality on false negative error percentage, F(1, 14) = 3.5, p= n.s.,, r =.13; there was also a non-significant effect of visual errorr classification on false negative error percentage, F(2, 28) = 1.05, p= n.s, r =.19. Visual search error was the main false negative error (48% of false negative errors were search errors on the standardd monitor; 31% of false negative errors were search errors on the digital workstation). The percentages of detection and interpretationn errors were similar (32.2% of detection error and 19.6% of interpretation error on the standard monitor; 15% of detection error and 33.3% of interpretation error on the workstation). There was a non-significant interactionn effect between the types of viewing modality and the types of visual errors, F(2, 28) = 1.11, p= n.s, r =.19. The percentages of each false negative error type on the two different viewing modalities are shown in figures 4 & 5. Proc. of SPIE Vol. 7966 79660S-5

70% 60% FN'Search' Error FN'Detection' Error FN'Interpretation' Error % of False Negative Errors 50% 40% 30% 20% 10% 0% Standard Monitor Viewing conditions Workstation Figure 4. False negative errors made on each modality Figure 5. Proportion of FN errors split by error type Proc. of SPIE Vol. 7966 79660S-6

On average, the number of times (approximately 4-5 times) that participants fixated within the AOI while examining images on a standard monitor (M = 4.6, SE = 1.5) was not significantly different from examining on the workstation (M = 5.2, SE = 3.1, t(12)= -.162, p=n.s., r =.05). Similarly, on average, participants eye dwell time within the AOI while examining images on the standard monitor (M = 4.82, SE=2.88) was not significantly different from examining on the workstation (M = 3.27, SE=1.86, t(12)= =.448, p=n.s., r =.13) although in terms of mean values they dwelt within the AOI for about 2s longer on the monitor than the workstation. Details are shown in figure 6 & figure 7. 9 Number of times eye fixed within the AOI 8 7 6 5 4 3 2 1 0 Standard Monitor Workstation Viewing conditions Figure 6. Number of times participants fixated within the AOI while examining images on the standard monitor and the workstation On average, the number of fixations before the eye first fixated within the AOI while examining images on the standardd monitor (M = 87.09, SE = 26.01) was not significantly different from examining on the workstation (M = 175.57, SE= 79.08, t(6)=-1.023, p=n.s., r=.39). Similarly, on average, the participants average duration between image onset to first hit the AOI while examining images on the standardd monitor (M = 38.46, SE = 13.20) was not significantly different from examining on the workstation (M = 95.33, SE = 59.4, t(6)=-.897, p=n.s., r= =.34). Details are shown in figures 8 & 9. Proc. of SPIE Vol. 7966 79660S-7

9 8 7 Dwell time (Second) 6 5 4 3 2 1 0 Standard Monitor Viewing conditions Workstation Figure 7. Dwell time comparison between examining images on the standard monitor and the workstation 250 Number of fixation before hit AOI 2000 150 1000 50 0 Standard Monitor Workstation Viewing conditions Figure 8. Comparison of number of fixations before the eye first fixated within the AOI while examining images on the standard monitor and workstation Proc. of SPIE Vol. 7966 79660S-8

160 140 120 Time to hit (Second) 100 80 60 40 20 0 Standard Monitor Workstation Viewing conditions Figure 9. Comparison of time to hit between examining images on standard monitor and workstation 4. DISCUSSION This study examined how radiologists and advanced practitioner radiographers (technologists) performed when examining sets of difficult recent screening cases on different modalities. The research interest here is the differencee between participants manipulation behaviour under different viewing conditions; also the comparison of participants visual search behaviour under the different conditions. Participants image manipulation behaviour on the workstation and the standard monitor was compared. The important finding was that the average time on each case in the different viewing conditions differed significantly between the highh experienced readers and low experienced readers. Compared with the low experienced participants, highh experienced ones spent less time on the standard monitor and workstation while the image manipulation tools were used. However, high experienced participants spent more time than the inexperienced participants while the image manipulation tools were not used. This could be because the experienced readers became more careful and therefore spent a longer time examining images while the normal image manipulation tools weree not available. The visual data analysis revealed that the participants made similar overall pattern of errors on both modalities. Theree were no significant differences between the three different visual errors on either modality. Furthermore, there was no significant difference between modalities in terms of these error types. On average, the number of times participants fixated within the AOI and participants eye dwell time within the AOI while examining images on a standard monitor was not significantly different from examining images on the workstation. Similarly, the number of fixations made before the eye first fixated within the AOI and the averagee duration to first hit the AOI while examining images on the standard monitor was not significantly different from examining images on the workstation. These findings are really somewhat surprising given the differences in monitor and workstation display sizes and resolutions. This points to similarity in how they inspected images on the two displays. Participants were overall faster on the monitor to hit the AOI and made fewer eye fixations before they hit the AOI. However, once within the AOI then they spent more time dwelling there on the monitor. This may well reflect that the abnormalities were able to be seen on either display type but thatt once identified then it took longer to examine in detail on the monitor. Proc. of SPIE Vol. 7966 79660S-9

5. CONCLUSION A standard monitor with the support of image manipulation achieved similar performance as on the digital workstation. Visual data analysis on the false negative errors suggested that participants visual examination behaviour on the standard monitor was very similar to their behaviour on the digital workstation. Also, their screen interaction behaviour on these two modalities was shown to be comparable. This implies that the standard monitor can potentially be considered to be a good alternative to a clinical workstation for the purpose of delivering mammographic interpretation training if image manipulation tools are made available and the training is carefully planned. ACKNOWLEDGEMENTS This work is partly supported by the UK National Health Service Breast Screening Programme. REFERENCES [1] Department of Health, "Breast Cancer Screening Report to the Health Ministers of England, Wales, Scotland and Northern Ireland by a Working Group Chaired by Sir Patrick Forrest," London, HMSO, 1986. [2] Department of Health, "Cancer Reform Strategy," London, 2007. [3] Chen, Y., Gale, A.G., James, J., Evans, A. & Turnbull, A., "Breast Screening: visual search as an aid for digital mammographic interpretation training," In B. Sahiner & D.J. Manning (Eds.) Proceeding of SPIE Medical Imaging 2010: Image Perception, Observer Performance, and Technology Assessment, Vol 7627, 2010. Proc. of SPIE Vol. 7966 79660S-10