Open Access
13 May 2021 Reduced motion artifacts and speed improvements in enhanced line-scanning fiber bundle endomicroscopy
Author Affiliations +
Abstract

Significance: Confocal laser scanning enables optical sectioning in fiber bundle endomicroscopy but limits the frame rate. To be able to better explore tissue morphology, it is useful to stitch sequentially acquired frames into a mosaic. However, low frame rates limit the maximum probe translation speed. Line-scanning (LS) confocal endomicroscopy provides higher frame rates, but residual out-of-focus light degrades images. Subtraction-based approaches can suppress this residue at the expense of introducing motion artifacts.

Aim: To generate high-frame-rate endomicroscopy images with improved optical sectioning, we develop a high-speed subtraction method that only requires the acquisition of a single camera frame.

Approach: The rolling shutter of a CMOS camera acts as both the aligned and offset detector slits required for subtraction-based sectioning enhancement. Two images of the bundle are formed on different regions of the camera, allowing both images to be acquired simultaneously.

Results: We confirm improved optical sectioning compared to conventional LS, particularly far from focus, and show that motion artifacts are not introduced. We demonstrate high-speed mosaicing at frame rates of up to 240 Hz.

Conclusion: High-speed acquisition of optically sectioned images using the new subtraction based-approach leads to improved mosaicing at high frame rates.

1.

Introduction

A conventional biopsy involves a clinician extracting suspect tissue from the body. The tissue is subsequently processed and examined in a pathology lab for morphological or cellular changes to determine whether a disease is present or stage its progression. Fluorescence endomicroscopy, a probe-based endoscopic technique with cellular-scale resolution, provides a means to perform a comparable analysis in vivo and in real-time. Fluorescence imaging is sometimes feasible using the autofluorescence of endogenous tissue structures such as collagen, but more commonly requires application of topical or intravenous contrast agents.1

Endomicroscopy is often performed with a flexible probe that is guided through the working channel of an endoscope. Working channels are typically <4  mm in diameter, with some variation among manufacturers and clinical applications. Therefore, the endomicroscopy probes must have diameters of around 3.5 mm or smaller and a sufficiently short, rigid tip to traverse the endoscope as it bends.

As with all in vivo microscopy techniques, optical sectioning is beneficial in fluorescence endomicroscopy to reduce corruption of images by out-of-focus fluorescence. Clinical studies with intravenous fluorescein are generally conducted using optical sectioning endomicroscopes. While careful use of topical stains can allow non-sectioning endomicroscopes to resolve surface features successfully,2,3 there is still a clear benefit to optical sectioning in some cases.4

As a means of providing sectioning, confocal laser scanning microscopy has been adapted to endomicroscopy using both distal and proximal scanning architectures.5,6 Distally scanned systems use a single-core fiber to act as both the point source and confocal detection pinhole. This approach requires precise miniaturized distal scanning elements to be integrated at the tip of the probe, and the only commercially available system developed and subsequently withdrawn (ISC-1000, Optiscan/Pentax) provided frame rates of only up to 2 Hz.7 Proximal scanning architectures instead make use of a multicore (30,000 core) fiber-optic imaging bundle. The imaging bundle relays excitation light from a bulk proximal scanning system (such as galvanometer scanning mirrors) and collects the returning fluorescence emission before de-scanning.5 The currently commercially available endomicroscope using this approach (Cellvizio, Mauna Kea Technologies) achieves frame rates of up to 12 Hz.1 However, the resolution is limited by the spacing between the fiber cores (down to 3  μm), although this can be further improved at the expense of the field-of-view using a distal objective with non-unity magnification.

The field-of-view of endomicroscopy probes is typically <1  mm, far smaller than a typical histology slice. It is possible to increase the effective image size by collecting a time-series of images as the probe is moved and then computationally stitching these images into a mosaic. Mosaicing requires the probe shift between image frames to be determined, which can be done in real-time using pairwise rigid registration.8 The need for reliable mosaics that resemble the ground truth is inhibited by tissue deformation, such as shearing, and cumulative error introduced by pairwise registration. More computationally intensive approaches, which are typically done offline, can, for example, register frames globally,9 correct for non-rigid distortions,8,10 or embed low-resolution images into a higher precision mesh and hence improve resolution.11

In both real-time and offline mosaicing, the imaging frame rate determines the maximum speed at which the probe can be moved due to the need to ensure sufficient overlap between images for registration. For example, for a probe with a 500-μm diameter field-of-view, and assuming that there can be a maximum shift of 1/3 of the image diameter between consecutive images, the maximum speed the probe can be translated at for acquisition frame rates of 2 and 12 Hz are, respectively, 0.33 and 2  mm/s. These low translation speeds are difficult to achieve reliably and consistently in realistic clinical scenarios. If distal optics are used to improve resolution and thus decrease the field-of-view, the allowable speeds are reduced even further by a factor equal to the magnification. Therefore, high frame rates are an important factor determining the feasibility of mosaicing.12

A higher frame-rate alternative to conventional confocal endomicroscopy—and the topic of this paper—is to use line-scanning (LS) illumination. Images are formed by scanning a laser line over the fiber bundle’s (FB) proximal face, and the fluorescence emission returning from the sample is then imaged onto a detection slit and subsequently onto a camera.13 This technique, known as LS or slit-scanning confocal microscopy, was first adapted to endomicroscopy by Sabharwal et al.14 The method of Sabharwal et al.14 used a configuration where returning fluorescence was de-scanned and imaged through a physical detection slit before being re-scanned onto a 2D camera. Similar systems were used for several subsequent studies, including multispectral imaging.15 A faster configuration using a 1D linear CCD, and hence not requiring re-scanning, was later reported by Hughes and Yang12 achieving a frame rate of over 100 Hz.

An alternative to a physical detection slit or linear camera is to use the rolling shutter of a CMOS camera as an electronically variable moving slit with a width that can be changed on-the-fly. The use of the electronic slit was first demonstrated for benchtop systems by Mei et al.,16 and later adapted to endomicroscopy by Hughes and Yang17; providing that the camera readout is synchronized with the scanning laser line, this avoids the need for returning fluorescence to be de-scanned by the scanner; the bundle can be imaged directly onto the CMOS camera (via appropriate fluorescence filters). The frame rate is then limited only by the camera readout speed; Hughes and Yang17 achieved 120 Hz.

While LS provides a degree of optical sectioning, it is not as effective as conventional point-scanning confocal microscopy.18 While close to focus the performance is similar, far from focus the amount of out-of-focus light rejected only increases approximately linearly with the distance from focus rather than quadratically as in point-scanning confocal. Fluorescence from far-from-focus depths can, therefore, still significantly degrade the image.

To mitigate the far-from-focus light degrading images, a two-frame subtraction-based approach, previously shown in principle for benchtop systems, was adapted to endomicroscopy.17,18 The first frame (the confocal image) of the sequence is acquired with the rolling shutter virtual detector slit aligned with the laser line as it is scanned, and the second (the residual image) is acquired with a fixed offset between the laser line and detector slit. The residual or offset image provides a first-order estimate of the residual out-of-focus signal not rejected by the slit. The residual image is then subtracted from the confocal image to generate an enhanced line-scan (ELS) image with improved optical sectioning. The full theory demonstrating how the subtraction of the residual image leads to an image with optical sectioning properties comparable to point-scanning confocal, albeit with inferior signal-to-noise ratio, is presented in Poher et al.18

Since two frames are required to synthesize each ELS image, this subtraction approach leads to a halving of the frame rate and, as with any multi-frame approach, introduces motion artifacts when the probe shifts between acquisitions. Motion artifacts present particular problems for mosaicing, where such shifts are required and deliberately introduced. A two-camera approach was recently proposed, which avoids these inter-frame motion artefacts.19 Rather than acquiring the two images sequentially, the beam is split between two rolling shutter cameras, one with triggering timing adjusted to provide an aligned-slit (the confocal image) and one providing an offset-slit (the residual image). The authors used an LED source to generate the excitation line, with the scanning line created by displaying a series of patterns on a digital micromirror device (DMD), a programmable array of mirrors. A DMD has the advantage of removing the need for a laser or scanning system but has several limitations, including very low throughput of power from the LED (since at any one time most of the light is blocked) and that the DMD can only store a limited number of patterns onboard, insufficient to generate a full set of scanned line positions. Multiple parallel scanning lines were used to overcome this, rather than a single swept line, allowing the same DMD pattern to be used for multiple scan positions. Parallel scanning has several drawbacks: the width of the illumination lines is constrained by the number of lines that can fit onto the DMD array, the size of the detection slit is limited by the spacing between the illumination lines, and it requires the excitation of regions where light is not being collected which may contribute to scattering and photo-bleaching.

A two-camera approach with a galvanometer scanner used to scan a laser line is feasible but would require the inclusion and fine alignment of a second camera. Cameras operated in free-run mode with camera-initiated triggering are typically able to run at twice the frame rate of cameras that are triggered. External triggering is a requirement of the two-camera approach. Instead, we now report a single camera configuration of the ELS approach for a galvanometer-based system that restores the full frame rate and eliminates motion artifacts. We denote this approach as ELS+. The system projects two images of the bundle onto different parts of a camera to introduce a controllable offset. Hence, using the camera’s rolling shutter as both the aligned and offset detection slits, the confocal and residual images can be provided simultaneously. The system can achieve enhanced sectioning at frame rates of up to 240 Hz with a lateral resolution of 8.8  μm. Since the two images are acquired simultaneously, the approach does not suffer from inter-frame motion artifacts. The optical sectioning performance is essentially identical to ELS using sequential/alternating acquisition of the two frames, which we denote sequential acquisition enhanced line-scanning (ELSA). ELS+ provides a frame rate improvement over the previously reported single-camera approach17 (60 Hz) by a factor of four and over the two-camera DMD approach19 (15 Hz) by a factor of 16. We also demonstrate real-time image registration and mosaicing at the full frame rate, which to our knowledge, is the fastest endomicroscopy mosaicing system reported to-date.

2.

Methods

A schematic diagram of the ELS+ endomicroscopy system is shown in Fig. 1(a). The output from a fiber-coupled 488-nm laser diode (Thorlabs L488P60) is collimated by a 4× microscope objective and passes onto a single galvanometer (galvo) scanning mirror (Thorlabs GVS001), then onto the back aperture of a 10× infinity-corrected objective (Nikon PLAN N RMS10X) via a 50-mm focal length cylindrical lens (Thorlabs LJ169SRM-A), a 50-mm focal length achromatic doublet lens (Thorlabs AC254-050-A-ML) and a dichroic mirror (Thorlabs DMLP490). A laser line is produced on the face of a 30,000 core fiber optic bundle (Fujikura FIGH-30-650S), which can be scanned by the galvo mirror.

Fig. 1

(a) ELS+ system schematic: L1, 4× finite conjugate objective, L2, f=50-mm cylindrical lens (power axis only is shown), L3, f=50-mm lens, L4, 10× infinity-corrected objective, L5, f=40-mm tube lens, GS, galvanometer scanning mirror (galvo), GSC, galvanometer scanner controller, DAQ, data acquisition card, SF, single mode fiber, FB, fiber bundle (scanning direction across bundle shown), DBS, dichroic beam-splitter, F1, emission filter, BS, beam-splitter, M1, stationary mirror, M2, adjustable tilt mirror with image offset axis and galvanometer axis (SEP) shown. (b) Illustration of camera plane showing the offset residual image and an example of the positions of the scanning illumination line and detection slit due to the rolling shutter.

JBO_26_5_056501_f001.png

The fiber probe does not use any distal optics, and so the probe is operated in direct contact with the tissue for imaging. The bundle’s image diameter is 620  μm, and the average spacing between the bundle cores is 3.2  μm. Fluorescence returning through the bundle is collected by the objective and transmitted by the dichroic before passing through an emission filter (Thorlabs DMLP490). It is then split into two paths by a beam-splitter onto a fixed mirror and a tilting mirror. The mirror configuration is designed to guide identical images of the bundle onto the left and right regions of a rolling shutter CMOS camera (Flea 3, FLIR FL3-U3-13S2M-CS) via a 75-mm tube lens (Thorlabs AC254-075-A-ML).

The rolling shutter camera uses sequential line readout, which progresses at a rate of 7.63  μs per line. The exposure determines the slit width by controlling how many lines are exposed simultaneously. The magnification factor between the sample and the camera was 3.8. Therefore, the minimum slit width, determined by the dimensions of a camera pixel, was 3.63  μm at the camera plane and 0.96  μm at the bundle plane. The relationship between the number of simultaneous active lines, N, and exposure, E in μs is

Eq. (1)

N=E7.63  μs,
where N is the number of active lines. The slit width at the bundle plane, W, is therefore given as

Eq. (2)

W=MΔpE7.63  μs,
where M is the magnification factor, and Δp is the pixel size.

The camera is operated in free-run mode, and the strobe signal from the camera is used to trigger the generation of each galvo scanning waveform signal via a USB DAQ Card (National Instruments). Synchronization between the ramp waveform driving the galvo and the camera readout is achieved via a calibration procedure described below.

Two images of a bundle are shown illustratively on the camera plane in Fig. 1(b). The offset of the bundle’s right-hand image (the residual image) on the camera is controllable with the tilting mirror (M2), allowing the rolling shutter to act as both the aligned and offset detection slit with a manually adjustable offset.

The camera can be run at the standard frame rate (120 Hz) and at a faster rate (240 Hz), in which case only a smaller region of interest can be used. When running at 120 Hz, the setup described above gives 6.8 camera pixels per average fiber core spacing (high sampling), leading to a lateral resolution of 5.52  μm, as measured using a back-illuminated United States Air Force (USAF) resolution target. To operate at 240 Hz, the 10× imaging objective was replaced with a 4× objective (finite conjugate). The smaller bundle image led to slight under-sampling of the bundle by the camera; the bundle was sampled by 2.4 camera pixels per fiber core spacing (low sampling), leading to a slightly reduced resolution (due to under-sampling) of 8.8  μm. Alternative approaches include cropping the bundle or using a high pixel density camera capable of high speeds. To crop the bundle a mask could be inserted at a conjugate plane, allowing the full resolution to be maintained at the cost of slight reduction in the field-of-view. Camera selection can be guided by sampling requirements. Due to the variation in the distance between cores, and the quasi-periodic nature of the bundle, optimal sampling is obtained around 3 pixels per core spacing, for the bundle used in this experiment that implies a square window of around 550×550 pixels for each bundle image.

The power delivered to the sample was measured to be 1 mW. The losses introduced by the beam-splitter configuration alone were 50%, and each image has a further 50% drop when the intensity is divided between the two images. It would be possible to minimize the 50% loss at the beam-splitter using a polarizing beam-splitter and quarter-wave plate configuration. Excitation arm losses before the bundle include reflections, filter absorption losses, and those associated with coupling to the bundle; these can easily be compensated by increasing the laser power. While maximum permissible exposure standards are not defined by the International Organization for Standards for esophageal imaging, the power used here is less than the commercially available clinical Cellvizio system (7 mW at 488 nm).20

2.1.

Image Processing

The raw image consists of two images of the FB, the LS confocal image, on the left side of the camera, and the residual image, on the right. A background estimation is acquired by placing the probe in a dark tube and averaging 50 images. The laser was on during background acquisition to account for auto-fluorescence generated in the fiber and room light and stray reflections in the optical system. The two images are then extracted, using a prior position calibration discussed in Sec. 2.2, and the residual is subtracted from the confocal to generate the ELS+ image. The image is then smoothed with a 3×3 kernel median filter, similar to the method described by Pierce et al.,21 to eliminate the honeycomb fiber core structure, especially noticeable in the high sampling mode/low frame rate images. Images acquired with low sampling had inherent smoothing due to the under-sampling of the bundle pixels by the camera and so were not filtered further.

Processed ELS+ images can be stitched together into a mosaic in real-time. The images are first down-sampled to 250×250  pixels and then circularly cropped to 90% of the bundle diameter. The shift between each pair of successive images is then registered using normalized cross-correlation between a template extracted from one image and the entire second image.22 The current image is then inserted dead-leaf at the correct position in the mosaic, where the pixel values of the new frame completely overwrite any previous pixel values in the mosaic. This algorithm was implemented mostly in Labview (National Instruments, Austin, Texas) while the normalized cross-correlation was performed using a DLL written in C++ calling OpenCV libraries. The processing frame rate was sufficient to allow mosaicing at the maximum frame rate (240 Hz) on a PC with an Intel i7-7700 Quad-Core processor and 8 GB RAM. However, with the exception of the demonstration of real-time mosaicing at 240 Hz, mosaics shown in this article were constructed off-line from saved stacks of images, stored as a lossless audio video interleave (AVI) file, to allow for detailed analysis and archiving of raw data.

2.2.

System Calibration

The system is calibrated using a custom-built LabVIEW VI. The process consists of three steps: (i) setting the slit-offset for the residual image to the desired value, (ii) synchronizing the laser LS with the camera readout by mapping the galvanometer drive signal voltage to the location of the line on the camera (in pixels), and (iii) precisely locating the bundle images on the CMOS so that subtraction can be performed. These calibration steps only need to be performed once for a given setup.

The first step is to set the offset of the residual image from the confocal image so as to achieve the desired optical sectioning performance (which, as discussed later, must be balanced with signal-to-noise requirements). The offset calibration procedure is illustrated in Figs. 2(a)2(c). Since the offset between the images should be perpendicular to the laser line (“vertical” direction as shown in Fig. 2), the image of a non-scanning laser line is used as a common reference point. (The laser line is visible in the image due to the excitation of autofluorescence from the bundle itself). Therefore, the vertical difference in positions between the laser lines in the two images is the offset in pixels.

Fig. 2

System calibration. (a)–(c) Calibration of line offset. (a) Non-scanning laser line visible on the bundle for the aligned image (left) and offset image (right). (b) Intensity profiles were taken perpendicular to the laser lines for the two images. (c) A difference of the central values determined by Gaussian fits of the two intensity profiles (the offset) reported in real-time. (d) and (e) Mapping of galvo voltage to pixels on CMOS camera. The line is slowly swept across the bundle face in the direction of the arrow, and (d) at each position step, an intensity profile is generated across the line. (e) The pixel position with the highest intensity is reported for each voltage, and a line is fit with the slope and intercept, determining the mapping between galvo voltage and line position in pixels. A typical mapping is shown, slope=0.00126  V/pixel, offset=0.459  V, and R2=0.99. (f) Bundle image locator routine. The bundle is illuminated from the distal end with an LED, and a binary threshold filter is applied. The centers of the two bundle images are returned as the rows and columns of the bundle with the greatest number of pixels above the threshold.

JBO_26_5_056501_f002.png

The first step in determining the offset was to set the galvo voltage at a value which produces a laser line approximately in the center of the bundle. For both images, intensity profiles are plotted along a line of pixels perpendicular to the laser line. A threshold filter is applied to remove the background, and each profile is then fitted with a Gaussian, and the difference in the center-of-mass positions between the two Gaussian fits is reported as the offset between the images. The line position is updated in real-time to allow instant feedback. The required offset can then be set by manually adjusting the tilting mirror until the desired value is obtained.

The second calibration step, shown in Figs. 2(d)2(e), is a fully automated mapping of the galvo drive signal (in volts) to the position of the laser line on the CMOS camera (in pixels). Mapping is done by sweeping through at least 400 voltage values (and hence line positions) and capturing an image at each voltage. If the laser line is visible, an intensity profile is taken across the line to locate the pixel number at the center of the line. A plot is produced of pixel number versus galvo voltage. The slope of a linear fit to this plot gives the scaling between voltage and pixel value, and the intercept gives the offset. These values, combined with the rolling shutter’s known progression speed, allow a galvo drive voltage ramp to be generated, which, when triggered by the camera’s strobe output trigger, results in the laser line being synchronized with the position of the rolling shutter.

The third step [Fig. 2(f)] is to locate the two bundle images on the CMOS camera. First, the bundle is pointed towards a light source and the pair of images recorded by the camera. A binary threshold mask is then used to identify each bundle. Each bundle’s center is then taken to be the row and column with the largest number of non-zero pixel values.

As a quality check, we imaged a USAF 1951 resolution target in transmission and compared the confocal and residual images extracted using the calibrated bundle positions. Visually there was some misalignment between fiber cores near one edge, but this did not exceed a distance of 1 core. Additionally, variation in transmission and reflection at the beam-splitter interface would lead to intensity differences in the images, although this was not observed in the images. The combined effect can be quantified using the mean of the absolute differences between the images, which was found to be 5.5%. After applying the filter, used to remove the core patterns during imaging, this was further improved to 0.74%.

3.

Results and Discussion

3.1.

Optical Sectioning and Signal Drop

The optical sectioning performance was characterized by generating profiles of the collected signal as a function of defocus. These profiles were generated by translating the FB tip using a mechanical translation stage toward a metal plate stained with a fluorescent highlighter. The mean intensity value of the central 250×250  pixel region (240×240  μm) of the bundle was extracted for each position of the translation stage.

A representative axial profile for a 2.9-μm slit width and a 29-μm slit offset is shown in Fig. 3. Plots for both conventional LS confocal and ELS are shown. Both plots were independently normalized to give a signal of 1 at focus. The 3 and 10 dB drop-off distances occur at the half-width half-maximum and half-width tenth-maximum of the LS and ELS curves. Also shown is the residual image profile (acquired with the offset slit), normalized to the same scale as the LS curve. The ELS curve is, therefore, the (re-normalized) result of subtraction of the residual curve from the LS curve. A clear improvement both in the 3 dB and particularly the 10-dB drop-off distances can be seen.

Fig. 3

Representative axial response profile for a 29-μm slit width and a 29-μm offset. LS and ELS curves were independently normalized, the residual curve was normalized using the same factor as the LS curve for comparison. The 3 and 10 dB drop-off distances for LS and ELS can be read-off at the intersection with the respective horizontal lines.

JBO_26_5_056501_f003.png

The change in the optical sectioning as a function of residual image offset (with a constant slit width) is shown in Figs. 4(a) and 4(b), parameterized by both the 3 and 10 dB drop-off distances. For comparison, the 3- and 10-dB drop-off distances are also shown for LS operation (i.e., without subtraction of the residual image, in which case the image offset has no effect). The ELS plots show that a smaller offset leads to greater optical sectioning strength improvement, shown by a decrease in both the 3- and 10-dB drop-off distances. As the offset increases, the 3-dB drop-off distance for ELS converges with the LS much faster than the 10-dB drop-off distance, so that even with very large offsets, there remains an optical sectioning improvement far from focus.

Fig. 4

A 3- and 10-dB drop-off distances as a function of offset and slit width. (a) and (b) Constant FL 29-μm slit width. (c) and (d) Constant FL 77-μm slit offset. The LS reference drop-off distances in (a) and (b) were determined by averaging drop-off distances over 7 runs, with mean 3-dB drop-off distance of 58±3  μm and mean 10-dB drop-off distance of 550±20  μm.

JBO_26_5_056501_f004.png

The change in the optical sectioning as a function of slit width (with a constant residual image offset) is shown in Figs. 4(b) and 4(c). Reducing the slit width improves both the 3- and 10-dB drop-off distances for both LS and ELS images, as would be expected. It is also apparent that the 10-dB drop-off distance is much more drastically affected by the slit width for the LS case than the ELS case; this is because ELS is very successful at removing far-from-focus light even when the slit width is large. The 10 dB drop-off improvement shows that, at least in principle, ELS can allow larger slit widths to be used than LS.

The pixelation of the fiber imposes an additional constraint on sectioning strength: once the offset or the slit width approaches the inter-core spacing of the bundle, there are virtually no gains in optical sectioning strength. The laser line at the distal end of the fiber cannot be made significantly smaller than the core spacing, and so reducing the detection slit below this has little effect. Similarly, if the slit separation falls below the inter-core spacing, 3  μm, the offset slit begins primarily sampling in-focus light and so subtraction lowers the signal levels. This effect is compounded by the fact the intensity across the line has a Gaussian profile, so that the intensity from the Gaussian tail bleeds into other cores, and the effects from any cross-core coupling.

To demonstrate this, axial response profiles with a constant offset of 29  μm and slit widths of 1.9, 2.9, and 5.8  μm were generated; the respective LS 3 dB drop-off distances were 70, 70, and 67  μm, and ELS distances were 120, 120, and 120  μm. Offsets of 2.9, 5.8, and 8.6  μm have respective 3-dB drop-off distances of 13, 9.6, and 12.5  μm and respective 10-dB drop-off distances of 59, 42, and 45  μm, again showing that there is no continued improvement from smaller offsets once the core spacing is approached.

The ELS subtraction approach is not without penalty. Since the residual image does not have exactly zero intensity from in-focus depths, there is a drop in the in-focus signal relative to LS when subtraction is performed. The signal drop occurs primarily because the illumination line is not arbitrarily thin; it has a Gaussian profile such that the region of the bundle sampled by the offset slit will still receive some illumination. Scattering or core leakage within the bundle will also tend to broaden the line further.

The signal level at focus was compared between LS and ELS using the un-normalized values of the axial response profiles when the fiber was in contact with the metal plate (i.e., at zero defocus). The resulting drop of the in-focus signal as a function of slit width and offset are shown in Fig. 5. With constant slit width, it is clear that the smallest in-focus signal drop occurs at the largest offset, as would be expected. At a 77-μm offset, there appears to be virtually no price paid in terms of relative in-focus signal drop when using ELS instead of LS. It may therefore be tempting to choose a large offset to minimize this loss of signal while still obtaining improved sectioning. However, as discussed in Sec. 3.3, a large offset leads to significant edge artifacts.

Fig. 5

Percentage of in-focus signal maintained (relative to LS) when varying slit width and offset. (a) Shown as a function of slit width with constant 77-μm offset. (b) Shown as a function of slit offset with a constant 2.9-μm slit width. Fit determined through non-linear least squares curve fit α=0.05. Signal reported as relative intensity of ELS to LS when bundle is in contact with mirror.

JBO_26_5_056501_f005.png

3.2.

Reduction of Motion Artifacts

The primary advantage of the simultaneous acquisition of the two images in the ELS+ approach reported here, compared to the previously reported ELSA approaches using sequential acquisition, is the reduction of motion artifacts. Motion artifacts are a function of frame rate, and so to show this most clearly, images of fluorescently stained lens tissue paper placed over a fluorescent background were acquired at a frame rate of 10 Hz with a moving probe. Figure 6(a) shows an image acquired using LS only (no subtraction); the presence of background and an out-of-focus signal is evident. In Fig. 6(b), the ELSA subtraction (using sequential acquisition) has led to noticeable motion artifacts. In Fig. 6(c), the ELS+ image (with the two images acquired simultaneously) shows the removal of the out-of-focus background without the motion artifacts. For comparison, the LS and ELS+ images were individually normalized, allowing 0.01% saturation, while the same normalization factor scaled the ELSA as ELS+.

Fig. 6

Images of lens tissue paper stained with fluorescent highlighter overlaid on a fluorescent background. (a) LS; (b) ELSA; (c) ELS+; and (d) intensity profiles. The probe was moved by hand. The imaging frame rate was set to 10 Hz. In (a) lower contrast is caused by light returning far-from-focus, while in (b) subtraction fails to achieve improved sectioning. In (d), square brackets indicate improved sectioning with ELS+, and arrows indicate some regions where motion artifacts appear in ELSA images. LS and ELS+ images were independently normalized to 0.01% saturation, which led to the intensity of the ELS+ image being greater than the LS image in the marked region (*). The same factor normalized ELSA images as ELS+ images. Acquired at 10-Hz reconstructed from recording.

JBO_26_5_056501_f006.png

3.3.

Edge Effect Artifacts

While the ELS+ approach does not introduce motion artifacts, there are some observed edge artifacts that are shared with the previously reported ELSA approach, and some which are common to most FB endomicroscopes. Edge artifacts limit the visual quality of mosaics because the same region of tissue is samples with different parts of the bundle; this is particularly apparent when using real-time mosaicing with dead-leaf image insertion rather than blending.

First, all FB systems that are not fully confocal have inherently better sectioning near the edge of the bundle than at the center for geometrical reasons (simply, less excitation light reaches the out-of-focus depths in the edge areas). A system with better sectioning performance inherently reduces the relative effect of this artifact, so ELS+ is an improvement over conventional LS in this regard.

Specifically for ELS (both ELSA and ELS+), there are further edge artifacts, due to the shape of the bundle, which are shown in Fig. 7, and which have not been commented on previously. Because the detection slit is always leading the illumination line, when the detection slit is near the top of the bundle, the illumination line is above the bundle, and no light reaches the tissue. Hence, near the top of the bundle, the residual image has close to zero intensity, and subtraction does not improve optical sectioning in this area. This can clearly be seen in the residual images shown in Fig. 7(a), although it is not particularly noticeable in individual ELS+ images.

Fig. 7

Demonstration of edge effect artifacts in residual images. (a) and (b) Average of 300 residual images taken with a moving probe over paper stained with a yellow highlighter, acquired using a constant slit width (8.8  μm) and varying offset: (a) 74 and (b) 12  μm. The intensity of (a) was scaled to match and (b) in the central 100×100  pixel region. Experimental causes of artifacts: (c) illumination line clipping, seen in image (a*); (d) greater detection area in residual images (a-arrow); (e) greater illumination area in residual images—less pronounced in (a).

JBO_26_5_056501_f007.png

A similar artifact appears even once the illumination line has reached the bundle. Because the bundle is circular, the width of the line changes as it scans through the tissue. Therefore, the left and right parts of the residual images will tend to under-estimate the residual signal in the top half of the bundle and over-estimate it in the bottom part. This is shown in Figs. 7(d) and 7(e).

Therefore, while a large slit offset for the residual image appears to deliver good performance in terms of the 10-dB drop-off distance, with minimal loss of in-focus signal, this comes at the expense of worsening the edge artifacts. Reducing the slit offset results in better sectioning and minimizes artifacts, but also results in a lower signal-to-noise ratio. Figure 7(b) shows this very clearly; when a smaller offset of 11.53  μm is used, there are virtually no artifacts from clipping or unequal illumination areas compared to when a larger offset of 73.7  μm is used as in Fig. 7(a).

A flat-field correction can compensate for illumination differences but would not correct the clipping artifact since there is no way to sample the residual image in the edge regions. Therefore, clipping artifacts can be considered a fundamental limitation of the ELS technique, although not a serious one in practice providing a careful choice of offset is made. In cases where a larger offset is desired, and cropping cannot be tolerated, two-sided illumination, with lines above and below the detection slit, would reduce the clipping artifacts.

3.4.

Ultra-High-Speed Mosaicing

Using the system set up in 240-Hz mode, we collected mosaics from lens tissue paper, stained as earlier. Figures 8(a) and 8(b) show mosaics generated from recordings for LS and ELS+, allowing a direct comparison, while Fig. 8(c) shows an ELS+ mosaic generated and displayed in real-time. To the best of our knowledge, this is the highest speed mosaicing reported from FB endomicroscopy and is an order of magnitude faster than typically reported rates. The slight worsening in lateral resolution at 240 Hz is purely a consequence of the need to use a smaller region of the camera, this could be mitigated by selecting a rolling shutter camera that can operate at high-speed with a greater number of pixels. Therefore, this higher mosaicing speed provides a significant potential improvement in the usability of mosaicing during high-speed motion of the probe.

Fig. 8

Lens paper stained with yellow highlighter overlaid on a fluorescent background. Acquired at 240 Hz. (a) LS; (b) ELS+ reconstructed from recording; (c) ELS+ real-time mosaicing. 8.8-μm resolution.

JBO_26_5_056501_f008.png

3.5.

Tissue Imaging

We validate the ELS+ approach for tissue imaging using ex vivo porcine esophagus stained with acriflavine hydrochloride. The confocal image and the residual image were simply extracted from alternating frames to simulate ELSA (sequential acquisition) imaging for comparison. Example individual frames (during motion) are shown in Figs. 9(a)9(c) for ELS, ELSA, and ELS+, respectively. The presence of motion artifacts in the ELSA image, which are not present in the ELS+ image, demonstrates the new approach’s value. The improved sectioning of ELS+ over LS can be seen most clearly in the mosaics of Figs. 9(d) and 9(e). The camera can acquire the full field of view at 120 Hz. However, to ensure no dropped frames during recording, these images were acquired at 60 Hz. We have previously demonstrated high-sampling 120 Hz mosaicing is also achievable.12

Fig. 9

Porcine esophagus: labeled with acriflavine hydrochloride. (a)–(c) Selected regions of interest from individual LS, ELSA, and ELS+ frames. Circled region indicates light that has been incorrectly subtracted due to a shift between acquisitions. LS and ELS+ independently normalized to 0.01% saturation, ELSA normalized to same factor as ELS+. Mosaics (d) LS and (e) ELS+. Scale bars 100  μm, and 6.8 pixels per core. Acquired at 60 Hz, reconstructed from recording.

JBO_26_5_056501_f009.png

4.

Conclusions

The newly reported approach of simultaneous acquisition ELS endomicroscopy (ELS+) provides improved optical sectioning over LS endomicroscopy while maintaining the same high frame rate and (most importantly) does not introduce artifacts due to inter-frame motion. Some edge artifacts due to the ELS approach remain, but these can be mitigated through a careful choice of operating parameters, mainly the slit offset.

The system was validated with topically stained ex vivo porcine esophagus and showed improved sectioning compared with LS endomicroscopy and a reduction in motion artifacts compared to ELSA. ELS+ has not yet been tested in vivo with intravenous fluorescein, so it cannot be confirmed whether the optical sectioning is sufficient in this scenario.

The ELS+ approach also allowed the frame rate to be further improved to 240 Hz using a smaller camera region. This enables mosaicing of optically sectioned images at 240 Hz, faster than previously reported approaches and more than an order of magnitude faster than commercially available systems.

The approach was tested using a bare FB imaging probe without any distal optics. In several respects, this underplays the potential advantage of ELS+. First, a bare probe with sharp edges tends to glide over tissue less smoothly than a lensed probe, meaning that the maximum probe speed was as much constrained by this as it was by the need to avoid artifacts. One way to correct this is using medical grade epoxy to form a ball on the distal tip of a bundle, the bundle can then be polished to the bundle’s face leaving the rounded edges. This is shown in Fig. 2(c) of Ref. 21. For a lensed system, it would be possible to scan the probe much faster, and with higher magnifications the removal of motion artifacts becomes more critical.

Lensed probes also usually demagnify the FB onto the tissue, improving the resolution but decreasing the field of view. For example, ELS endomicroscopy’s initial demonstration used a probe with a field-of-view of just 240  μm.17 In this case, the higher frame rate becomes more critical for achieving mosaicing to allow sufficient overlap between images, so the benefit of ELS+ will be more pronounced. A high frame rate is also important if multi-frame techniques such as resolution enhancement11 or axial scanning are to be implemented, with the latter also standing to benefit from the improved optical sectioning of ELS+.

Disclosures

The authors declare that there are no conflicts of interest related to this article.

Acknowledgments

This work was supported by the Engineering and Physical Sciences Research Council (EPSRC); “Ultrathin fluorescence microscope in a needle” Award No. EP/R019274/1. Andrew Thrapp is supported by a University of Kent Vice Chancellor’s PhD Scholarship.

Code, Data, and Materials Availability

Example data and code showing reconstruction of ELS+ images from confocal and residual images of tissue paper and ex vivo animal tissue is available at https://doi.org/10.6084/m9.figshare.13303211

References

1. 

W. A. Wells et al., “In vivo and ex vivo microscopy: moving toward the integration of optical imaging technologies into pathology practice,” Arch. Pathol. Lab. Med., 143 (3), 288 –298 (2019). https://doi.org/10.5858/arpa.2018-0298-RA APLMAS 0003-9985 Google Scholar

2. 

M. Pierce, D. Yu and R. Richards-Kortum, “High-resolution fiber-optic microendoscopy for in situ cellular imaging,” J. Vis. Exp., (47), 2306 (2011). https://doi.org/10.3791/2306 Google Scholar

3. 

M. H. Koucky and M. C. Pierce, “Axial response of high-resolution microendoscopy in scattering media,” Biomed. Opt. Express, 4 (10), 2247 –2256 (2013). https://doi.org/10.1364/BOE.4.002247 BOEICL 2156-7085 Google Scholar

4. 

P. A. Keahey et al., “Optimizing modulation frequency for structured illumination in a fiber-optic microendoscope to image nuclear morphometry in columnar epithelium,” Biomed. Opt. Express, 6 (3), 870 –80 (2015). https://doi.org/10.1364/BOE.6.000870 BOEICL 2156-7085 Google Scholar

5. 

A. F. Gmitro and D. Aziz, “Confocal microscopy through a fiber-optic imaging bundle,” Opt. Lett., 18 (8), 565 (1993). https://doi.org/10.1364/OL.18.000565 OPLEDP 0146-9592 Google Scholar

6. 

J. M. Jabbour et al., “Confocal endomicroscopy: instrumentation and medical applications,” Ann. Biomed. Eng., 40 (2), 378 –397 (2012). https://doi.org/10.1007/s10439-011-0426-y ABMECF 0090-6964 Google Scholar

7. 

D. B. Ellebrecht et al., “Towards an optical biopsy during visceral surgical interventions,” Visc. Med., 36 (2), 70 –79 (2020). https://doi.org/10.1159/000505938 Google Scholar

8. 

T. Vercauteren et al., “Real time autonomous video image registration for endomicroscopy: fighting the compromises,” Proc. SPIE, 6861 68610C (2008). https://doi.org/10.1117/12.763089 PSISDG 0277-786X Google Scholar

9. 

T. Vercauteren et al., “Mosaicing of confocal microscopic in vivo soft tissue video sequences,” Med. Image Comput. Comput. Assist. Interv., 8 (Pt. 1), 753 –760 (2005). https://doi.org/10.1007/11566465_93 Google Scholar

10. 

T. Vercauteren et al., “Robust mosaicing with correction of motion distortions and tissue deformations for in vivo fibered microscopy,” Med. Image Anal., 10 (5), 673 –692 (2006). https://doi.org/10.1016/j.media.2006.06.006 Google Scholar

11. 

K. Vyas et al., “Fiber bundle shifting endomicroscopy for high-resolution imaging,” Biomed. Opt. Express, 9 (10), 4649 –4664 (2018). https://doi.org/10.1364/BOE.9.004649 BOEICL 2156-7085 Google Scholar

12. 

M. Hughes and G. Z. Yang, “High speed, line-scanning, fiber bundle fluorescence confocal endomicroscopy for improved mosaicking,” Biomed. Opt. Express, 6 (4), 1241 –1252 (2015). https://doi.org/10.1364/BOE.6.001241 BOEICL 2156-7085 Google Scholar

13. 

C. J. R. Sheppard and X. Q. Mao, “Confocal microscopes with slit apertures,” J. Mod. Opt., 35 (7), 1169 –1185 (1988). https://doi.org/10.1080/09500348814551251 JMOPEW 0950-0340 Google Scholar

14. 

Y. S. Sabharwal et al., “Slit-scanning confocal microendoscope for high-resolution in vivo imaging,” Appl. Opt., 38 (34), 7133 –7144 (1999). https://doi.org/10.1364/AO.38.007133 APOPAI 0003-6935 Google Scholar

15. 

H. Makhlouf et al., “Multispectral confocal microendoscope for in vivo and in situ imaging,” J. Biomed. Opt., 13 (4), 044016 (2008). https://doi.org/10.1117/1.2950313 JBOPFO 1083-3668 Google Scholar

16. 

E. Mei et al., “A line scanning confocal fluorescent microscope using a cmos rolling shutter as an adjustable aperture,” J. Microsc., 247 (3), 269 –276 (2012). https://doi.org/10.1111/j.1365-2818.2012.03642.x JMICAR 0022-2720 Google Scholar

17. 

M. Hughes and G. Z. Yang, “Line-scanning fiber bundle endomicroscopy with a virtual detector slit,” Biomed. Opt. Express, 7 (6), 2257 –2268 (2016). https://doi.org/10.1364/BOE.7.002257 BOEICL 2156-7085 Google Scholar

18. 

V. Poher et al., “Improved sectioning in a slit scanning confocal microscope,” Opt. Lett., 33 (16), 1813 –1815 (2008). https://doi.org/10.1364/OL.33.001813 OPLEDP 0146-9592 Google Scholar

19. 

Y. Tang et al., “Simple differential digital confocal aperture to improve axial response of line-scanning confocal microendoscopes,” Opt. Lett., 44 (18), 4519 –4522 (2019). https://doi.org/10.1364/OL.44.004519 OPLEDP 0146-9592 Google Scholar

20. 

B. Obstoy et al., “Safety and performance analysis of acriflavine and methylene blue for in vivo imaging of precancerous lesions using fibered confocal fluorescence microscopy (FCFM): an experimental study,” BMC Pulm. Med., 15 30 (2015). https://doi.org/10.1186/s12890-015-0020-4 Google Scholar

21. 

M. C. Pierce et al., “A pilot study of low-cost, high-resolution microendoscopy as a tool for identifying women with cervical precancer,” Cancer Prev. Res., 5 (11), 1273 –1279 (2012). https://doi.org/10.1158/1940-6207.CAPR-12-0221 Google Scholar

22. 

B. Kai and D. H. Uwe, “Template matching using fast normalized cross correlation,” Proc. SPIE, 4387 95 –102 (2001). https://doi.org/10.1117/12.421129 PSISDG 0277-786X Google Scholar

Biography

Andrew D. Thrapp is a PhD student at the University of Kent. He received his bachelor’s degree in physics with a minor in mathematics from the University of Colorado in 2017. His current research interests include optical-sectioning in microscopy and endomicroscopy.

Michael R. Hughes is a lecturer in applied optics in the School of Physical Sciences at the University of Kent. His research interests are in endoscopic and needle-based microscopes and computational approaches to low-cost and point-of-care microscopy.

CC BY: © The Authors. Published by SPIE under a Creative Commons Attribution 4.0 Unported License. Distribution or reproduction of this work in whole or in part requires full attribution of the original publication, including its DOI.
Andrew D. Thrapp and Michael R. Hughes "Reduced motion artifacts and speed improvements in enhanced line-scanning fiber bundle endomicroscopy," Journal of Biomedical Optics 26(5), 056501 (13 May 2021). https://doi.org/10.1117/1.JBO.26.5.056501
Received: 1 December 2020; Accepted: 22 March 2021; Published: 13 May 2021
Lens.org Logo
CITATIONS
Cited by 3 scholarly publications.
Advertisement
Advertisement
KEYWORDS
Cameras

Electrophoretic light scattering

Endomicroscopy

Confocal microscopy

Camera shutters

Tissues

Imaging systems

Back to Top