International Journal of Biomedical Imaging

International Journal of Biomedical Imaging / 2013 / Article

Research Article | Open Access

Volume 2013 |Article ID 472971 | 10 pages | https://doi.org/10.1155/2013/472971

Endoscopy-MR Image Fusion for Image Guided Procedures

Academic Editor: Tiange Zhuang
Received11 Apr 2013
Revised27 Aug 2013
Accepted13 Sep 2013
Published02 Nov 2013

Abstract

Minimally invasive endoscope based abdominal procedures provide potential advantages over conventional open surgery such as reduced trauma, shorter hospital stay, and quick recovery. One major limitation of using this technique is the narrow view of the endoscope and the lack of proper 3D context of the surgical site. In this paper, we propose a rapid and accurate method to align intraoperative stereo endoscopic images of the surgical site with preoperative Magnetic Resonance (MR) images. Gridline light pattern is projected on the surgical site to facilitate the registration. The purpose of this surface-based registration is to provide 3D context of the surgical site to the endoscopic view. We have validated the proposed method on a liver phantom and achieved the surface registration error of  mm.

1. Introduction

In this paper, we develop a new method for endoscopy-MR image fusion of the liver organ for minimally invasive endoscope based surgery. Image guidance is an essential tool in minimally invasive endoscope based abdominal procedures [1]. Effective image guidance can compensate the restricted perception during the operation, which is considered a major limitation in endoscopic procedures. Without image guidance, the surgeon cannot see through the surface of the operation site and may accidentally cause damages to the critical structures of the patient. A typical procedure in image guidance is to map pre-operative high quality MR images to intra-operative endoscopic video images, or the patient thereby provides a good quality context to the real-time endoscopic images. Thus, the surgeon will be able to visually access the operation site during the procedure. As a result, the damage to the critical organs or tissues will be substantially minimized.

Fusion of endoscopic video images with high quality MR images requires good match of these two modalities. In this paper, we adopt a surface based image fusion because the two modalities are different in acquisition and nature [2, 3]. In order to find the corresponding 3D surface model from endoscopic images, we utilize stereovision to snapshot the surgical site from two different angles and compute the 3D location by using triangulation [4]. Cameras are calibrated before triangulation is used [5, 6].

Although a liver phantom is used to validate the proposed technique, our method is not restricted to the liver surgery. The integrated image guidance can also be applied to other endoscopic procedures. This paper is organized as follows. Section 2 introduces the experimental setup and the camera calibration of the stereo endoscope. Section 3 discusses automatic surface reconstruction, and Section 4 presents surface based registration and experimental results of image fusion. Section 5 discusses the issues in this study. Section 6 presents the conclusion and future work.

2. Experimental Setup and Camera Calibration

Experimental setup is shown in Figure 1. In the experiments of this study, we use the following major components: a Visionsense VSII stereo endoscope, an Optoma PK301 Pocket Projector, a liver phantom, and a chessboard calibration pattern. Optoma PK301 Pocket Projector is a small size projector and can be easily mounted. The resolution of this projector is 848 by 480. The liver phantom was printed using a 3D printer based on the liver model that was segmented from MR images of a human subject.

In this study, robust 3D surface reconstruction requires accurate camera calibration of the stereo endoscope. The calibration process aims to find intrinsic parameters and correct the optical distortion inherent in the endoscope and to compute extrinsic parameters to capture the spatial relationship between left and right cameras of the stereo endoscope. We have modified the Camera Calibration Toolbox for MATLAB [7] and performed calibration of the stereo endoscope using a chessboard calibration pattern.

3. Surface Reconstruction from Stereo Endoscope

In this section, we propose a novel approach to reconstruct the surface of the surgical site from two stereo endoscopic images. The reconstruction procedure is shown in Figure 2. First, a gridline pattern is projected on the surgical site, and both left and right images are acquired at the same time. Second, the intersection points of the gridlines are automatically detected and matched in both images. Then we reconstruct the surface with the matched intersection points. We will describe major steps in detail in the following sections.

3.1. Conversion of Input Images to Grayscale Images

In order to detect the intersections of the grid lines pattern of an image, we use the image of binary format as algorithm input. We first convert the color images acquired from the stereo endoscope to grayscale images (as shown in Figures 3 and 4). According to the thinning algorithm used in the proposed system, the gridlines of the light pattern should be bright to detect their intersections. The grayscale image is thus inverted to meet this constraint. In this process, the dark areas become bright and vice versa. Next, multiple steps are employed to obtain good binary images.

3.2. Intensity Correction

The image intensity of the endoscopic images is not uniform given that variation in illumination and ambient lights exist. As a consequence, conventional threshold methods cannot be directly used to achieve good binary images which can successfully separates gridlines from the background. In this paper, we present an intensity correction technique to improve the image. The improvement aims to equalize the contrast between gridlines and background over the whole image. The new corrected pixel value is calculated by where is the intensity value of the current pixel, is the avarage intensity of its neighbourhood pixels, and is the new intensity value after correction.

Figure 5 shows the image after intensity correction, in which the contrast between the gridlines and background is more uniform compared with the image before correction in Figure 4(b). Figure 6 shows binary images by thresholding, which will be used for intersection detection. With intensity correction, all gridlines are clearly shown in the binary image, while without intensity correction, only a part of the gridlines is shown in the cluttered binary image. The intensity correction also significantly improves the detection and matching accuracy with the successful rate of 98% versus 57% without intensity correction (see Table 1).

(a) With intensity correction

ImageNumber of pointsDetected pointsCorrect pointsSensitivityFPPFNP

Im177 73730.948104
Im277 74740.96103
Im377 74740.96103
Im477 73730.948104
Im577 76760.98701
Im677 78771.010
Im777 77760.98711
Im877 76760.98701
Im977 76760.98701
Im1077 76760.98701
Im1177 76760.98701
Im1277 76760.98701
Im1377 77771.000
Im1477 76760.98701
Im1577 77771.000
Im1677 76760.98701
Im1777 76760.98701
Im1877 76760.98701
Im1977 76760.98701

Average0.9822

(b) Without intensity correction

ImageNumber of pointsDetected pointsCorrect pointsSensitivityFPPFNP

Im177 51510.66026
Im277 48480.62029
Im377 25240.31153
Im477 53520.67125
Im577 53530.68024
Im677 39380.49139
Im777 42420.54035
Im877 39390.50038
Im977 60600.77017
Im1077 56550.71122
Im1177 53530.68024
Im1277 52520.67025
Im1377 44440.57033
Im1477 42410.53136
Im1577 42400.51237
Im1677 41400.51137
Im1777 36350.45142
Im1877 36350.45142
Im1977 41330.42244

Average0.5653

3.3. Detection of Region of Interest

In this paper, region of interest (ROI) is defined as the region which only covers the projected gridline light pattern in the endoscopic image. Automatic detection of ROI is critical for accurate detection and matching of intersection points in the gridline pattern. During the image preprocessing step, the area out of ROI should be cut out. ROI detection leads to automatic removal of unwanted areas. This step significantly improves the correctness of gridline intersection detection as well as the processing speed.

The ROI detection process aims to generate a mask of the grid lines pattern. Following intensity correction, we threshold the images in order to convert the corrected grayscale image into a binary image for further processing. Next, the dilation and the erosion operations are performed. Eventually, as the consequence of dilation and erosion processes, we obtain a binary mask image only covering the region of the projected gridline pattern. Then, we apply the mask image to the intensity corrected image to produce a cropped image within the desired ROI. The cropped image is then converted to a binary image by applying a threshold to it, which is used for feature detection. Figure 7(a) shows the detected ROI of the input image (ROI mask image), and Figure 7(b) shows the cropped image within ROI.

3.4. Image Dilation and Erosion

Because of the conversion to a binary image, some white pixels in the binary image are far away from the gridlines. Hence, these types of pixels could cause false positive pixels in the thinning process. By using dilation, we can expand the gridlines to fill the gaps between them and the protrusions pixels. Dilation process followed by erosion process is used to return the structure to its original state by removing the added structure of the gridlines. As a result, we have smoother gridlines without holes and protrusion pixels. Figure 8 shows the binary images before and after dilation and erosion process, respectively.

3.5. Thinning and Intersection Detection

In order to detect the intersections of the gridline pattern, we used a thinning process applied to the above processed binary image. The thinning process generates an image with one pixel width; that is, it generates a skeleton image of the input binary image. Then we proceeded to detect the intersections of the image gridlines. This process was accomplished by applying a hybrid approach for cross-point detection called the combined cross-point number (CCN) method [6]. The CCN method uses two techniques to detect intersections of gridlines: simple cross-point number (SCN) and the modified cross-point number (MCN). The CCN algorithm is used to detect the intersection points of the gridlines.

In simple cross-point number, the image is iterated with a small window of size 3 by 3 pixels [6], as a result we have eight pixels surrounding the tested pixel. To test if the center pixel of the 3 by 3 window is a cross-point pixel, we iterate this window on the image and get the cross-point number (CPN) for the center pixel. The CPN is calculated by where is the pixel value of th pixel of the 3 by 3 window and . A point is considered a cross point if its CPN is four.

In modified cross-point number method, the image is iterated by a window of size 5 by 5 pixels surrounding the center pixel. The CPN is calculated by where . The pixel is considered a cross-point pixel if . In the combined cross-point number both simple cross-point number and modified cross-point number methods are used. The simple cross-point number is used in the inner neighbors of the center pixel, while the modified cross-point number is used in the outer neighbors of the center pixel. Each pixel in the image has been tested against CPN using the modified cross-point number method, in which it is considered a cross point if and only if it satisfies and . Because of the low quality of the images, we adjust the CPN of the combined CPN to be in the range of 3.0 and 4.0. Figure 9(a) shows the left skeleton image by thinning operation. Figure 9(b) shows the detected intersection points plotted on the left image. Figure 9(c) shows the detected intersection points and plotted on the right image. Notice that there are false positive points in both images, and as shown in Figure 10, these false points are eliminated by our method using the epipolar geometry matching constraint.

3.6. Matching Grid Points

In order to reconstruct the surface within the ROI using the triangulation technique, we need to find the corresponding intersection points in the left and right images. Since these grid points have similar features, these correspondence relationships cannot be effectively obtained using conventional feature matching methods such as Scale Invariant Feature Transform (SIFT) and Speeded Up Robust Features (SURF) based techniques. In this paper, we adopt the method of epipolar constraints [2]. The intersection points are matched column-by-column to achieve good matching in our study.

We have validated the proposed approach using 19 pairs of images acquired at different position and orientation of the stereo endoscope. Figure 10 shows the matched grid points superimposed in one image. Table 1 shows the actual number of points in each image, the number of the points detected, the number of correct points, the number of false positive points (FPP), and the number of false negative points (FNP). The average of sensitivity detection of the proposed method is 0.9822.

3.7. Points/Surface Reconstruction

In general, a video image generated from the endoscope is a 2D projection of the 3D scene. This process can be represented using the pinhole camera model [8]. After we obtain the camera calibration parameters, we reconstruct a 3D point from left and right image projections by using stereo triangulation. A smooth surface can be reconstructed by fitting these reconstructed 3D grid points as shown in Figure 11(a). This figure has demonstrated that the proposed method can achieve accurate surface reconstruction from the stereo endoscopic images.

In order to investigate the impact of the previous image processing procedures on the reconstructed surface, we have performed the following experiments. We use image intensity correction as an example to examine the effects in detail. We repeat the entire process of the surface reconstruction as shown in Figure 2, except that no intensity correction is performed. Figure 11(b) shows the reconstructed surface without intensity correction, and Figure 14 shows the corresponding average registration error without intensity correction. Comparing Figures 11, 13, and 14, we can clearly see that intensity correction has significantly improved the reconstruction accuracy. With intensity correction, the average surface reconstruction error reduces from 1.86 mm to 0.76 mm. Similarly, the proposed automatic detection method for ROI improve the reconstruction accuracy as well. In Table 2, we show how the average surface error is affected by previous image processing procedures.


CaseMean (mm)Standard deviation (mm)

Proposed method0.760.11
Without intensity correction1.860.68
Without detection of ROI1.240.73

4. Surface Based Registration

4.1. ICP Registration

The Iterative Closest Point (ICP) algorithm is widely employed to align two three-dimensional surfaces. The ICP algorithm was first proposed by Besl and McKay [9], which is an iterative two-step method. The first step is to establish point correspondences by finding the corresponding point closest to the second surface for each point in the first surface. The second step is to calculate a transformation based on these matched points, which produces incremental transformations whose composition is the registration results.

In this study, the ICP is employed to register the reconstructed surface from endoscopic images with the surface extracted from MR images. Figure 12 shows the overlay of 3D surfaces after surface registration.

4.2. Registration Accuracy

The projected gridline pattern used to test the proposed approach consists of seven rows and eleven columns, and we have 77 intersection points to detect in each image. We used 19 pairs of left and right images acquired by the stereo endoscope at different poses for validation. After ICP surface registration, we calculated the average surface distance (ASD) between two corresponding surfaces as registration accuracy. The resulting ASD is  mm (see Figure 13). Figure 14 shows the corresponding registration accuracy without intensity correction.

After surface based registration, we are able to fuse the reconstructed surface from stereo endoscopic images with pre-operative high quality MR images and the corresponding patient-specific models such as vessel centerlines as shown in Figure 15. This will enable surgeons to see through critical structures beyond the operational site surface.

5. Discussion

Developing a rapid and accurate approach to reconstruct the surface from stereo endoscopic images is a very challenging task especially for soft tissues with few features. Many techniques have been developed to acquire or reconstruct surgical surface such as using laser scanners and Time-of-Flight (ToF) cameras [1012]. Hayashibe et al. used a laser-scan endoscope technique to reconstruct the shape and texture of the area of interest [13]. A laser scanner was proposed to acquire the liver surface for image-guided liver surgery, but it took about 5–20 seconds [10]. Therefore it is not suitable for free-breathing patients since average respiratory rates of children are 16–30 breaths per minute. ToF cameras produce a depth map that can be immediately used to generate a 3D surface model in real time, but current devices are too large for endoscopic procedures [11, 12]. For this study, we employed a small stereo endoscope with the diameter of 4.9 mm, which can be used with a typical 5 mm trocar in clinical practice. For proof of concept, we used a general purpose projector to project the light pattern of gridlines. In the future, this easy-to-implement light pattern can be generated by using a very small lithographic pattern generator with 10 μm thick lines at a distance of 50 μm [14], which can generate a pattern of lines within the size of 3 mm × 3 mm.

Many works [1517], proposed to reconstruct the soft tissue structures of the abdomen using stereovision. However, it is difficult to find the correspondences between the two images, even when taking into account epipolar constraints. Moreover, 3D surface reconstruction for abdominal procedures is more challenging due to few or no features on the surface of some organs such as the liver. To tackle this, structured light based methods were presented in [18]. Traditionally, the structured light technique projects the coded pattern onto the object and substitutes one camera in the stereovision with a projector [19]. Consequently, the correspondence problem becomes a decoding problem, and we can determine the correspondences between the acquired image and the original known coded pattern. Many light patterns are proposed for 3D surface reconstruction in the last decades [20, 21]. The design and realization of a new endoscopic device by means of a robust structured light coding are presented in [4]. However, the coded pattern employed in [4] is not commercially available and is not easy to be implemented for clinical use.

Since we mainly consider minimally invasive image guided procedures, one major criterion of selecting light patterns is easy implementation. In this study, we select the gridline light pattern, which can be easily generated by a commercial projector or a special device. However, this choice of light pattern poses a great challenge for matching feature points (i.e., intersections) of gridlines due to symmetry and similar features of gridline points. Conventional methods [22, 23] such as SIFT cannot be employed effectively for surface reconstruction in our study. Based on the specific characteristics of the gridline light pattern in this paper, we adopt a new method for robust feature detection and matching.

In this study, we have proposed to use dedicated gridline light patterns to create noninvasive artificial features on the tissue surface, which is then used to robustly reconstruct 3D surface from stereo endoscopic images. This is a robust 3D surface reconstruction technique for procedures involving soft tissue organs especially with few surface features. Another feature of the proposed system is the use of a stereo endoscope for 3D surface reconstruction. The major advantage of using stereo endoscope is that it can acquire two synchronized images simultaneously, which provides necessary information for 3D surface reconstruction and eliminates the challenging temporal synchronization problem inherently with mono endoscope for moving deformable targets. The stereo endoscope not only provides two synchronized images at the same time but also reduces intra-operative image acquisition time and eliminates unnecessary motion of mono endoscope to acquire two images at different poses which is required for robust 3D surface reconstruction.

Effective and good display of virtual reality (VR) is an important factor for clinicians to accept and support the multimodality image guidance system in the clinical environment. It is still an active research topic as to how to present and fuse real images with patient-specific pre-operative images and models in an optimal way so that they register correctly in the physician’s brain. Stereo display is one effective way to present the information to physicians. Some 3D video stereo monitors do not require a separate apparatus such as synch box, ZScreen, or active glasses. In order to effectively use our stereo endoscope, we still need 3D glasses in the current configuration; however, it would be more convenient for physicians to watch 3D video stereo monitor without glasses. Fused anatomy and models can provide physicians with more information where all supporting information becomes available. Multiple monitors can be used to display different information separately. For example, the first monitor can display real-time video stereo endoscopic images, while a second monitor is used to display the fused images/models. Thus in challenging scenarios, doctors can selectively watch different monitors to acquire needed information including surgical plans and other critical structures such as tumors and blood vessels without interference from unnecessary information.

6. Conclusion

In this study, we proposed a novel approach to match stereo endoscopic images and MR images. The proposed surface-based registration has proved to be an effective method for registering these images of different imaging mechanisms. Moreover, the light patterns of the gridlines facilitated the surface reconstruction of surgical sites with few surface features. In this paper, we validated the proposed method with static objects; however, our method has the potential to be extended to procedures involving moving organs.

We have demonstrated the effectiveness of our technique in registration of the reconstructed surface with the surface extracted from MR images of a liver phantom. We have shown that various image processing techniques we used before the image registration have a significant impact on the resulting registration accuracy. We have achieved a surface registration accuracy of  mm. The proposed technique has the potential to be used in clinical practice to improve image guidance in endoscope based minimally invasive procedures. The fused image guidance may also be applied to the endoscopic procedures of other organs in the abdomen, chest cavity, and pelvis such as the kidneys and the lungs.

Future work includes integration of our method into the clinical image guidance system and further validation by animal study and clinical study.

Acknowledgments

The authors would like to thank Dr. James Drake, Thomas Looi, and other members at CIGITI in the Hospital for Sick Children for the valuable assistance and support.

References

  1. M. Baumhauer, M. Feuerstein, H.-P. Meinzer, and J. Rassweiler, “Navigation in endoscopic soft tissue surgery: perspectives and limitations,” Journal of Endourology, vol. 22, no. 4, pp. 751–766, 2008. View at: Publisher Site | Google Scholar
  2. M. Sonka, V. Hlavac, and B. Roger, Image Processing Analysis and Machine Vision, 3rd edition, 2008.
  3. J. L. Prince and J. M. Links, Medical Imaging Signals and Systems, Printice Hall, Upper Saddle River, New Jersey, NJ, USA, 2006.
  4. X. Maurice, C. Albitar, C. Doignon, and M. De Mathelin, “A structured light-based laparoscope with real-time organs' surface reconstruction for minimally invasive surgery,” in Proceedings of the Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC '12), pp. 5769–55772, 2012. View at: Google Scholar
  5. M. Feuerstein, T. Mussack, S. M. Heining, and N. Navab, “Intraoperative laparoscope augmentation for port placement and resection planning in minimally invasive liver resection,” IEEE Transactions on Medical Imaging, vol. 27, no. 3, pp. 355–369, 2008. View at: Publisher Site | Google Scholar
  6. V. Bevilacqua, S. Cambo, L. Cariello, and G. Mastronardi, “A combined method to detect retinal fundus features,” in European Conference on Emergent Aspects In Clinical Data Analysis, pp. 1–6, Pisa, Italy, 2005. View at: Google Scholar
  7. J. Y. Bouguet, Camera Calibration Toolbox for Matlab, http://www.vision.caltech.edu/bouguetj, 2010.
  8. D. Dey, D. G. Gobbi, P. J. Slomka, K. J. M. Surry, and T. M. Peters, “Automatic fusion of freehand endoscopic brain images to three-dimensional surfaces: creating stereoscopic panoramas,” IEEE Transactions on Medical Imaging, vol. 21, no. 1, pp. 23–30, 2002. View at: Publisher Site | Google Scholar
  9. P. J. Besl and N. D. McKay, “A method for registration of 3-D shapes,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 14, no. 2, pp. 239–256, 1992. View at: Publisher Site | Google Scholar
  10. L. W. Clements, W. C. Chapman, B. M. Dawant, R. L. Galloway Jr., and M. I. Miga, “Robust surface registration using salient anatomical features for image-guided liver surgery: algorithm and validation,” Medical Physics, vol. 35, no. 6, pp. 2528–2540, 2008. View at: Publisher Site | Google Scholar
  11. L. Maier-Hein, M. Schmidt, A. M. Franz et al., “Accounting for anisotropic noise in fine registration of time-of-flight range data with high-resolution surface data,” Medical Image Computing and Computer-Assisted Intervention, vol. 6361, no. 1, pp. 251–258, 2010. View at: Publisher Site | Google Scholar
  12. A. Groch, A. Seitel, S. Hempel et al., “3D surface reconstruction for laparo-scopic computer-assisted interventions: comparison of state-of-the-art methods,” in Proceedings of the Medical Imaging 2011: Visualization, Image-Guided Procedures, and Modeling, vol. 7964, 2011. View at: Google Scholar
  13. M. Hayashibe, N. Suzuki, and Y. Nakamura, “Laser-scan endoscope system for intraoperative geometry acquisition and surgical robot safety management,” Medical Image Analysis, vol. 10, no. 4, pp. 509–519, 2006. View at: Publisher Site | Google Scholar
  14. M. Proesmans, L. J. Van Gool, and A. J. Oosterlinck, “Active acquisition of 3D shape for moving objects,” in Proceedings of the IEEE International Conference on Image Processing (ICIP '96), pp. 647–650, September 1996. View at: Google Scholar
  15. M. Hu, G. Penney, P. Edwards, and D. Hawkes, “3D reconstruction of internal organ surfaces for minimal invasive surgery,” in International Conference on Medical Image Computing and Computer-Assisted Intervention, pp. 68–77, Brisbane, Australia, 2007. View at: Google Scholar
  16. W. W. Lau, N. A. Ramey, J. J. Corso, N. V. Thakor, and G. D. Hager, “Stereo-based endoscopie tracking of cardiac surface deformation,” in Proceedings of the 7th International Conference on Medical Image Computing and Computer-Assisted Intervention (MICCAI 2004), pp. 494–501, St Malo, France, September 2004. View at: Google Scholar
  17. D. Stoyanov, A. Darzi, and G. Z. Yang, “Dense 3D depth recovery for soft tissue deformation during robotically assisted laparoscopic surgery,” in Proceedings of the 7th International Conference Medical Image Computing and Computer-Assisted Intervention (MICCAI '04), pp. 41–48, St Malo, France, September 2004. View at: Google Scholar
  18. M. Hayashibe, N. Suzuki, and Y. Nakamura, “Laser-scan endoscope system for intraoperative geometry acquisition and surgical robot safety management,” Medical Image Analysis, vol. 10, no. 4, pp. 509–519, 2006. View at: Publisher Site | Google Scholar
  19. J. Salvi, J. Pagès, and J. Batlle, “Pattern codification strategies in structured light systems,” Pattern Recognition, vol. 37, no. 4, pp. 827–849, 2004. View at: Publisher Site | Google Scholar
  20. A. Bhatti, Current Advancements in Stereo Vision, InTech, 2012.
  21. J. Salvi, S. Fernandez, T. Pribanic, and X. Llado, “A state of the art in structured light patterns for surface profilometry,” Pattern Recognition, vol. 43, no. 8, pp. 2666–2680, 2010. View at: Publisher Site | Google Scholar
  22. P. Mountney, D. Stoyanov, and G.-Z. Yang, “Three-dimensional tissue deformation recovery and tracking,” IEEE Signal Processing Magazine, vol. 27, no. 4, pp. 14–24, 2010. View at: Publisher Site | Google Scholar
  23. D. G. Lowe, “Distinctive image features from scale-invariant keypoints,” International Journal of Computer Vision, vol. 60, no. 2, pp. 91–110, 2004. View at: Publisher Site | Google Scholar

Copyright © 2013 Anwar Abdalbari et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

1305 Views | 811 Downloads | 2 Citations
 PDF  Download Citation  Citation
 Download other formatsMore
 Order printed copiesOrder