
In recent years, minimally invasive surgery has developed rapidly in the clinical practice of surgery and has gradually become one of the critical surgical techniques. Compared with traditional surgery, the advantages of minimally invasive surgery include small incisions and less pain during the operation, and the patients recover faster after surgery. With the expansion of minimally invasive surgery in several medical fields, traditional minimally invasive techniques have bottlenecks in clinical practice, such as the inability of the endoscope to determine the depth information of the lesion area from the two-dimensional images obtained, the difficulty in locating the endoscopic position information and the inability to get a complete view of the overall situation in the cavity. This paper uses a visual simultaneous localization and mapping (SLAM) approach to achieve endoscope localization and reconstruction of the surgical region in a minimally invasive surgical environment. Firstly, the K-Means algorithm combined with the Super point algorithm is used to extract the feature information of the image in the lumen environment. Compared with Super points, the logarithm of successful matching points increased by 32.69%, the proportion of effective points increased by 25.28%, the error matching rate decreased by 0.64%, and the extraction time decreased by 1.98%. Then the iterative closest point method is used to estimate the position and attitude information of the endoscope. Finally, the disparity map is obtained by the stereo matching method, and the point cloud image of the surgical area is finally recovered.
Citation: Liwei Deng, Zhen Liu, Tao Zhang, Zhe Yan. Study of visual SLAM methods in minimally invasive surgery[J]. Mathematical Biosciences and Engineering, 2023, 20(3): 4388-4402. doi: 10.3934/mbe.2023203
[1] | Yongli Yan, Tiansheng Sun, Teng Ren, Li Ding . Enhanced grip force estimation in robotic surgery: A sparrow search algorithm-optimized backpropagation neural network approach. Mathematical Biosciences and Engineering, 2024, 21(3): 3519-3539. doi: 10.3934/mbe.2024155 |
[2] | Xiaoming Liu, Zhonghan Liu, Yaqin Pan, Yufeng Huang, Desheng Wu, Zhaoyu Ba . Full-endoscopic transforaminal procedure to treat the single-level adjacent segment disease after posterior lumbar spine fusion: 1–2 years follow-up. Mathematical Biosciences and Engineering, 2019, 16(6): 7829-7838. doi: 10.3934/mbe.2019393 |
[3] | Lufeng Yao, Haiqing Wang, Feng Zhang, Liping Wang, Jianghui Dong . Minimally invasive treatment of calcaneal fractures via the sinus tarsi approach based on a 3D printing technique. Mathematical Biosciences and Engineering, 2019, 16(3): 1597-1610. doi: 10.3934/mbe.2019076 |
[4] | Xuefei Deng, Yu Liu, Hao Chen . Three-dimensional image reconstruction based on improved U-net network for anatomy of pulmonary segmentectomy. Mathematical Biosciences and Engineering, 2021, 18(4): 3313-3322. doi: 10.3934/mbe.2021165 |
[5] | Akim Kapsalyamov, Shahid Hussain, Prashant K. Jamwal . A novel compliant surgical robot: Preliminary design analysis. Mathematical Biosciences and Engineering, 2020, 17(3): 1944-1958. doi: 10.3934/mbe.2020103 |
[6] | Hafiz Muhammad Muzzammil, Yong-De Zhang, Hassan Ejaz, Qihang Yuan, Muhammad Muddassir . A review on tissue-needle interaction and path planning models for bevel tip type flexible needle minimal intervention. Mathematical Biosciences and Engineering, 2024, 21(1): 523-561. doi: 10.3934/mbe.2024023 |
[7] | Wenlu Zhang, Ziyue Ma, Hong Wang, Juan Deng, Pengfei Li, Yu Jia, Yabin Dong, Hong Sha, Feng Yan, Wenjun Tu . Study on automatic ultrasound scanning of lumbar spine and visualization system for path planning in lumbar puncture surgery. Mathematical Biosciences and Engineering, 2023, 20(1): 613-623. doi: 10.3934/mbe.2023028 |
[8] | Maher Alwuthaynani, Raluca Eftimie, Dumitru Trucu . Inverse problem approaches for mutation laws in heterogeneous tumours with local and nonlocal dynamics. Mathematical Biosciences and Engineering, 2022, 19(4): 3720-3747. doi: 10.3934/mbe.2022171 |
[9] | Tianran Yuan, Hongsheng Zhang, Hao Liu, Juan Du, Huiming Yu, Yimin Wang, Yabin Xu . Watertight 2-manifold 3D bone surface model reconstruction from CT images based on visual hyper-spherical mapping. Mathematical Biosciences and Engineering, 2021, 18(2): 1280-1313. doi: 10.3934/mbe.2021068 |
[10] | Tian Ma, Boyang Meng, Jiayi Yang, Nana Gou, Weilu Shi . A half jaw panoramic stitching method of intraoral endoscopy images based on dental arch arrangement. Mathematical Biosciences and Engineering, 2024, 21(1): 494-522. doi: 10.3934/mbe.2024022 |
In recent years, minimally invasive surgery has developed rapidly in the clinical practice of surgery and has gradually become one of the critical surgical techniques. Compared with traditional surgery, the advantages of minimally invasive surgery include small incisions and less pain during the operation, and the patients recover faster after surgery. With the expansion of minimally invasive surgery in several medical fields, traditional minimally invasive techniques have bottlenecks in clinical practice, such as the inability of the endoscope to determine the depth information of the lesion area from the two-dimensional images obtained, the difficulty in locating the endoscopic position information and the inability to get a complete view of the overall situation in the cavity. This paper uses a visual simultaneous localization and mapping (SLAM) approach to achieve endoscope localization and reconstruction of the surgical region in a minimally invasive surgical environment. Firstly, the K-Means algorithm combined with the Super point algorithm is used to extract the feature information of the image in the lumen environment. Compared with Super points, the logarithm of successful matching points increased by 32.69%, the proportion of effective points increased by 25.28%, the error matching rate decreased by 0.64%, and the extraction time decreased by 1.98%. Then the iterative closest point method is used to estimate the position and attitude information of the endoscope. Finally, the disparity map is obtained by the stereo matching method, and the point cloud image of the surgical area is finally recovered.
Compared to traditional surgery, the advantages of minimally invasive surgery are minor trauma, minor infection, and minor pain [1], a trend of surgical treatment in modern medicine. In modern minimally invasive surgery technology, the endoscope and specialist instruments are inserted into the internal cavity through a small incision in the surgical site. The images captured by the endoscope are displayed on the screen in real-time [2]. The surgeon can monitor the patient's entire operation through the endoscopic transmission of images and can determine the patient's surgical condition [3]. For traditional minimally invasive surgery, there are still limitations in obtaining image information from the endoscope, such as the narrow field of view, uneven illumination and high reflection intensity, and the difficulty in obtaining depth information of the target area and locating the endoscope accurately. Under such complex environmental conditions, higher requirements are put forward for computer vision processing systems. Therefore, SLAM system is innovatively applied to the field of minimally invasive surgery to solve the bottleneck faced by modern minimally invasive surgery [4]. The endoscope acts as a robot in the field of minimally invasive surgery. It moves through the cavity of the human body, measures spatial information of the visible area of the cavity, completes the reconstruction of soft tissue, and displays it to the doctor. Minimally invasive surgery can help doctors make the right decisions by combining SLAM with better positioning of the endoscope and depth of the target area. It has crucial research significance and application value [5].
The concept of minimally invasive surgery and the issue of visual SLAM were first proposed in 1983 and 1986, respectively. The use of vision technology based on endoscope images in minimally invasive surgery has obvious advantages without introducing additional equipment to the already very complex surgical device [6]. SLAM is most frequently used in minimally invasive surgery in the abdomen, where deformation and tissue movement are minimal. At the initial stage of application, SLAM was added to the algorithm as an auxiliary means for positioning [7]. In 2009, Grasa et al. used monocular laparoscopy to generate a sparse map of the environment from SLAM and then estimated the three-dimensional information of the scene through sequential images. In 2009 and 2010, the Mountney P team respectively proposed the application of an SLAM system based on the monocular endoscope and binocular endoscope in minimally invasive surgery for endoscope positioning and 3D modeling, and the monocular system based on the framework of SLAM to establish a motion model to estimate the motion of the camera and soft tissue effectively [8]. The binocular system constructs a 3D texture model of a minimally invasive surgery environment, a method model of dynamic view expansion [9]. In recent years, there have been more and more studies on the application of visual SLAM in minimally invasive surgery [10,11,12], most of which are devoted to solving the bottleneck of minimally invasive surgery. Vision-based techniques are good at restoring 3D structures and estimating endoscope motion. However, this research has a long way to go to solve the problems of accuracy [13].
This paper mainly studies visual SLAM in minimally invasive surgery. For the deficiencies of traditional minimally invasive surgery, visual SLAM technology is used to complete the positioning and 3D mapping of the endoscope [14,15]. In combination with the K-means algorithm and the Super point algorithm, the feature information of the image in the inner cavity is extracted. It solves the problem that it is difficult to obtain the depth information of the target area and accurately locate the endoscope position in the environment with narrow vision, uneven illumination and high reflection intensity of minimally invasive surgical endoscope. Then, the iterative nearest point method was used to estimate the position and attitude of the endoscope. Finally, the stereo matching method was used to reconstruct the lumen image, and the point cloud image of the surgical area was recovered. Improve surgical accuracy and shorten the surgical time, thus optimizing the operation of the entire system [16].
The structure of this paper is as follows. Section 2 discusses the visual SLAM and the method of extracting image features through Superpoint algorithm, as well as the method of extracting image feature points based on Superpoint algorithm and k-means algorithm. In Section 3, experiments are carried out to verify the effectiveness of our proposed method. Section 4 summarizes the whole article.
The final output of the Super point (self supervised interest point detection and description) algorithm is the feature point algorithm with descriptors [17]. The network diagram is shown in Figure 1.
In this network, the initial feature detector is obtained by training the basic graph with undisputed feature points in a self-supervised way, and the image feature information is further extracted by the neural network. The stability of feature extraction is improved while the sensitivity of the feature extraction algorithm to illumination is reduced. The network is divided into two parts [18], the network for detecting basic graph corners and the network for extracting image features and outputting descriptors. The specific steps are as follows:
(1) Feature point detector pre-training. For an image, it is difficult to define which points are the real feature points, but if it is a simple graph with defined corner positions, it is easy to determine the truth. Therefore, the pre-training is carried out by extracting undisputed feature points from an unlabelled virtual simple image, obtaining pseudo-truths, and using the resulting pseudo-truths to determine the true feature points. Then, the pseudo-true values are combined with the true values to retrain the feature detector to obtain the extracted feature. The resulting pseudo-truths are combined with the true values to retrain the feature detector to obtain the extracted feature point model.
(2) The feature point detection network is divided into two parts: encoding and decoding. Image coding is to input the image to be processed into the shared coding network and reduce the dimension of the image to reduce the computation of the network. After dimensionality reduction, the image size is 1/8 of the original image. The main function of the encoder in the network is to map the input image to the spatial tensor, which has a smaller spatial dimension and larger channel depth. A probability is calculated for each pixel, which represents the probability that this pixel is a feature point. When decoding feature points, subpixel convolution is adopted to avoid the possibility of network overload caused by excessive computation in the process of extracting feature points. The input dimension of the decoder is
(3) Description sub detection network. Extracting the descriptor is a decoding operation. The descriptors are obtained from feature points. First, the image size and feature point position are normalized.
K-Means clustering algorithm will use all the data as a whole, the initial clustering center is
After the Super point algorithm extracts the feature points and descriptors of the image, the heatmap of the feature map can be obtained. Each feature point in the heatmap has an index value, and the K-means method is used to cluster these index values. When feature points are matched, a confidence probability of a correct match is set for each index value. According to this probability, whether this point is the correct match of the points to be matched is determined. Set the maximum number of matching point pairs and match according to the index of feature points. When all feature points are matched, or the maximum number of matching points is reached, the matching graph is output. The K-means clustering algorithm is applied to avoid the successful matching between the reflective spot and the strong light point in the inner cavity and reduce the probability of mismatching. Because the k-means algorithm is very simple, very fast, time complexity is nearly linear, and it is suitable for mining large-scale datasets, combining it with the super point algorithm not only improves the matching degree, but also shortens the time it takes to compute the k-means algorithm.
The pose estimation process is actually to find the corresponding point between the 2D image and the 3D spatial position. When representing an object's 3D spatial position and orientation, a coordinate system rotating with the object is usually established with the object as the coordinate origin. The transformation relationship between the rotating coordinate system and the reference coordinate system can determine the spatial position of the object. That is, the solution of pose estimation discusses the transformation relationship between coordinate systems.
In visual SLAM, there are usually two methods to estimate the pose of the sensor. One is to use a linear method to estimate the pose information of the visual sensor. After obtaining the initial pose, the pose information is further optimized by constructing and solving the least square problem. The other is to directly integrate the position of the space point and the position of the sensor to solve the pose information. In the estimation of the pose information of the endoscope, since the position of the endoscope is not continuous and uniform, the first method is selected to solve the pose information of the endoscope. The initial pose is estimated first, and then the results are optimized to increase the accuracy and robustness of the SLAM system.
The iterative closest point (ICP) is a standard algorithm that estimates pose between completed 3D Point pairs. Suppose there are a pair of 3D point pairs matched and their centroids:
P={p1,…,pn},P′={p′1,…,p′n}p=1nn∑i=1(pi),p′=1nn∑i=1(p′i) | (1) |
where:
To perform pose estimation is actually to find a Euclidean transform that satisfies:
∀i,pi=Rp′i+t | (2) |
The result of the above formula is solved through the algebraic method of singular value decomposition, and the error term of the point
ei=pi−(Rpi′+t) | (3) |
Constructing the least-squares problem:
minR,tJ=12n∑i=1‖pi−(Rp′i+t)‖22=12n∑i=1‖pi−p−R(p′i−p′)‖22+‖p−Rp′−t‖22 | (4) |
Solve the above equation and get
qi=pi−p, q′i=p′i−p′ | (5) |
The rotation matrix and translation vector are calculated according to the following formula:
{R∗=argminR12n∑i=1‖qi−Rq′i‖22t∗=p−Rp′ | (6) |
Define matrix:
W=n∑i=1qiqi′T | (7) |
R=UVT | (8) |
Substitute the solved rotation matrix into Eq (8) to solve the translation vector.
Firstly, the disparity value of the target region is obtained and optimized by the SGBM stereo matching method; Secondly, the depth value is calculated by using the parallax value to form the depth image; Finally, the 3D point cloud data are obtained by further calculation.
Parallax refers to that when binocular stereo vision obtains two frames of images at the same time through the left and right cameras, the corresponding relationship between features is established at the imaging points on the left and right imaging planes of the camera through the corresponding three-dimensional space points so that the human eye can feel the depth. Observe the imaging difference of the image on the camera plane. This difference is called parallax. In short, parallax describes the horizontal pixel difference between the corresponding imaging points of the left and right eyes [21].
There is a point
The parallax values in the parallax map can be calculated to obtain the corresponding depth values. The depth map shows the form of the image by integrating these depth values. The depth map contains the depth information of each pixel value, and the pixel value in the depth map is the depth value. According to the geometric relation of parallel binocular vision, the relation between parallax value and object depth information is as follows:
bdep=(b+xr)−xldep−f | (9) |
where:
dep=b×fxl−xr=f×bd | (10) |
where,
According to the coordinate transformation relation, the transformation relation between the world coordinate system and the pixel coordinate system can be obtained as follows:
zp[uv1]=[fx0cx0fycy001][R,t][XYZ1] | (11) |
where: The first matrix on the right of the equation is the endoscope internal parameter; matrix
{z=zpx=(u−cx)×zp/fxy=(v−cy)×zp/fy | (12) |
The calculated data can be saved in the created point cloud data, and the point cloud image can be displayed by using some libraries. The depth image of point cloud data can be calculated through the reverse calculation of the above operations.
In order to verify whether the improved algorithm has better feature extraction and matching effect in the strong reflection environment in minimally invasive surgery, the feature extraction results of the improved algorithm and the improved algorithm in the lumen environment are compared, and the simulation results are compared and analyzed. The original image pairs for feature extraction are shown in a) and b) Figure 5. By comparing the feature extraction results of various methods on the image pairs, the effectiveness of the improved extraction network is shown [22].
The results of image feature extraction using the original Super point algorithm are shown in c) in Figure 5, and the results of feature extraction using the improved Super point algorithm are shown in d) in Figure 5. As can be seen from the figure, the number of feature points extracted by the Super point algorithm in the inner cavity environment under a strong reflection environment is more and more uniform than that extracted by the traditional algorithm, but there are too many failed matching points and many invalid points. These matching failed points and invalid points are caused by the narrow field of vision, uneven lighting and high reflection intensity of minimally invasive surgical endoscope, The numerical value of the depth map physically means the distance from the camera. A lot of black and a lot of white are similar distances from you. The feature points are difficult to extract, and the resolution of the depth map is low, which reduces the number of feature points that can be extracted. False matching situations are reasonable. However, the distribution of feature points extracted by the improved Super point algorithm is more uniform, and effective features can be extracted from the edge part of the insufficient light. e) in Figure 5 is the matching result of feature points extracted by the original Super point algorithm, and f) in Figure 5 is the matching result of feature points extracted by the improved Super point algorithm. It can be easily seen from the figure that the improved algorithm has a better matching effect on feature points, and the number of invalid feature points is 0, which is better than the original algorithm and the traditional algorithm. As can be seen from Figure 5 and Table 1, after the Super point algorithm and clustering algorithm are combined and improved, the number of feature extraction effects of the inner cavity image increases and is evenly distributed, without any redundant points that fail to be matched successfully, and the percentage of successfully matched points in the total number of extractions is the highest and the false matching rate is the lowest. Although the proportion of effective points extracted by SIFT, SURF and ORB algorithms can reach 100%, their speed is very slow when there are too many matching points, and there are many wrong matching point pairs. The original Superpoint algorithm can not only ensure the number of feature points, but also ensure the speed, but it has a very low proportion of effective points. Compared with the traditional feature extraction algorithm and the original Super point algorithm, it shows the best effect.
SIFT | SURF | ORB | Superpoint | Improving Superpoint | |
Extracting time /s | 1.282 | 0.182 | 0.125 | 0.103 | 0.101 |
Match time /s | 0.187 | 0.078 | 0.031 | 0.018 | 0.021 |
Extract point pairs/pairs | 334 | 121 | 268 | 352 | 349 |
Match point pairs/pairs | 334 | 121 | 268 | 263 | 349 |
False match point pair/pair | 79 | 25 | 31 | 13 | 15 |
Effective point proportion /% | 100.00 | 100.00 | 100.00 | 74.72 | 100.00 |
False match rate /% | 23.65 | 20.66 | 11.57 | 4.94 | 4.30 |
Table 1 shows the numerical results of each algorithm. From the two indicators of Extracting time and Match time in the table, the time of the improved algorithm is significantly shorter than that of other algorithms, which also proves that the improved algorithm has the minimum time consumption, the maximum number of matching success points and the minimum mismatch rate, and the effective point ratio is significantly improved compared with that before the improvement.
Different from large-scale scenarios, when the endoscope is in the special environment of minimally invasive surgery, the estimation of the position information of the endoscope only needs numerical values. Since the movement process of the endoscope is irregular, it is not meant to form the map of its position information [23]. In this topic, will the SLAM system of camera pose estimation algorithm used in endoscopic minimally invasive surgery, using the image information of the real environment of minimally invasive surgery for endoscopic frame only when the estimate judgment in endoscopic movement between rotation and translation in the process of calculation is accurate, lumen mapping results of image feature points as shown in Figure 6.
The blue point in Figure 6 is the feature point of the current frame, and the green point is the feature point of the next frame. The current frame is directly processed without any operation. The position difference between blue and green is why the camera moves when shooting different frames, and the red point is the position of the green point after the rotation matrix and the motion vector move left. The distance between blue dots and red dots is smaller than that between green dots, with an average error of 5.43 pixels.
The disparity image and depth image before and after optimization are obtained by using the algorithm in this chapter as follows:
Figure 7 shows the point cloud before and after optimization of the lumen image in the actual scene, which is displayed in the gray form, and the cavity area is more obvious. The image information with too dark illumination in the original image has the condition that the visual difference has not been calculated, so the depth value cannot be calculated, which is reflected as the whole area in the point cloud image. The point cloud image before optimization has obvious segmentation lines. After optimization, this phenomenon is weakened, and most of the information is effectively restored. The details of the main areas are more delicate, and some smaller blood vessels that are not easy to extract can also be restored [24].
Figure 8 is a pure point cloud image without adding mesh and texture. The color depth in the figure indicates the distance between the point and the endoscope. Although the edge results of the disparity map and depth map are not satisfactory, it can be seen from b) of Figures 7 and 8 that the reconstruction method adopted can have a good effect on the restoration of images in the lumen environment.
As can be seen from a) and b) in Figure 9, except for the empty part with extremely weak light, the image has a filling effect, and the edge is partially optimized. Although the "boundary" between the core area and the fat part of the image does not have an excellent restoration result, the edge is smoothed. c) is a pure point cloud image without mesh and texture, and the depth of color in the figure represents the distance between the point and the endoscope. d) is the point cloud diagram of the inner cavity image in the real scene. In the case of the image information with a too-dark light in the original image, the parallax value is not calculated, so the depth value cannot be calculated, which is reflected in the point cloud image as the empty area in the figure. Most of the information in the image can be effectively restored. The details of the main areas are delicate, and some small vessels that are not easy to be extracted can also be recovered. The reconstruction method used in this paper can restore the image in the inner cavity with good effect.
In view of the problem that it is difficult to obtain depth information on the target area and accurately locate the position of the endoscope in the environment of the narrow field of vision, uneven illumination, and high reflection intensity of minimally invasive surgery endoscope, the method based on the combination of K-means and Super point algorithm is first used to extract image feature points and match. Compared with the traditional algorithm, the simulation results are more accurate and stable. Compared with the Super point, the logarithm of successful matching points is increased by 32.69%, the proportion of effective points is increased by 25.28%, the false matching rate is reduced by 0.64%, and the extraction time is reduced by 1.98%. In the real minimally invasive surgery scene, the feature points in the endovascular image were mapped to verify the accuracy of the calculated rotation and translation, with an average error of 5 pixels. Finally, the stereo matching algorithm is used to calculate the disparity map of the original image, and then the disparity map is used to calculate the depth information so as to calculate the point cloud image with depth information, which can well reconstruct the small features such as blood vessels in the image, and also have a good reconstruction effect on the edge part of the insufficient light [25].
Funding: This research was supported by the Natural Science Foundation of Heilongjiang Province (LH2019F024), China, 2019–2022, and Key R & D Program Guidance Projects of Heilongjiang Province (Grant No. GZ20210065), 2021–2024
The authors declare there is no conflict of interest.
[1] |
T. N. Robinson, G. V. Stiegmann, Minimally invasive surgery, Endoscopy, 36 (2004), 48–51. https://doi.org/10.1055/s-2004-814113 doi: 10.1055/s-2004-814113
![]() |
[2] | J. Perissat, Laparoscopic cholecystectomy, a treatment for gallstones: From idea to reality, World J. Surg., 23 (1999), 328–331. |
[3] | J. Chen, Y. Guo, Observation and study on the therapeutic effect of endoscopy combined with minimally invasive abdominal surgery robot in the treatment of gallstones, in Proceedings of 2019 International Conference on Biology, Chemistry and Medical Engineering Francis, (2019), 79–84. |
[4] | J. Xiao, Q. Wu, D. Sun, C. He, Y. Chen, Classifications and functions of vitreoretinal surgery assisted robots-a review of the state of the Art, in 2019 International Conference on Intelligent Transportation, Big Data & Smart City (ICITBS) IEEE, (2019), 474–484. https://doi.org/10.1109/ICITBS.2019.00122 |
[5] |
C. Siristatidis, C. Chrelias, Feasibility of office hysteroscopy through the "see and treat technique" in private practice: A prospective observational study, Arch. Gynecol. Obstet., 283 (2011), 819–823. https://doi.org/10.1007/s00404-010-1431-3 doi: 10.1007/s00404-010-1431-3
![]() |
[6] | P. Cheeseman, R. Smith, M. Self, A stochastic map for uncertain spatial relationships, in 4th International Symposium on Robotic Research, (1987), 467–474. |
[7] | P. Mountney, D. Stoyanov, A. Davison, G. Yang, Simultaneous stereoscope localization and soft-tissue mapping for minimal invasive surgery, in International Conference on Medical Image Computing and Computer-Assisted Intervention Springer, Berlin, Heidelberg, (2006), 347–354. https://doi.org/10.1007/11866565_43 |
[8] |
G. Mattioli, V. Rossi, F. Palo, M. C. Y. Wong, P. Gandullia, S. Arrigo, et al., Minimal invasive approach to paediatric colorectal surgery, J. Ped. Endosc. Surg., 3 (2021), 129–139. https://doi.org/10.1007/s42804-020-00090-6 doi: 10.1007/s42804-020-00090-6
![]() |
[9] | P. Mountney, G. Z. Yang, Dynamic view expansion for minimally invasive surgery using simultaneous localization and mapping, in 2009 Annual International Conference of the IEEE Engineering in Medicine and Biology Society IEEE, (2009), 1184–1187. https://doi.org/10.1109/IEMBS.2009.5333939 |
[10] | B. Lin, Visual SLAM and Surface Reconstruction for Abdominal Minimally Invasive Surgery, University of South Florida, 2015. |
[11] | L. Chen, W.Tang, N. W. John, T. R. Wan, J. J. Zhang, Augmented reality for depth cues in monocular minimally invasive surgery, preprint, arXiv: 1703.01243. |
[12] | A. Marmol, P. Corke, T. Pinot, ArthroSLAM: Multi-sensor robust visual localization for minimally invasive orthopedic surgery, in 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) IEEE, (2018), 3882–3889. https://doi.org/10.1109/IROS.2018.8593501 |
[13] | C. Girerd, A. V. Kudryavtsev, P. Rougeot, P. Renaud, K. Rabenorosoa, B. Tamadazte, Automatic tip-steering of concentric tube robots in the trachea based on visual slam, in IEEE Transactions on Medical Robotics and Bionics, 2 (2020), 582–585. https://doi.org/10.1109/TMRB.2020.3034720 |
[14] | I. Font, S. Weiland, M. Franken, M. Steinbuch, L. Rovers, Haptic feedback designs in teleoperation systems for minimal invasive surgery, in 2004 IEEE International Conference on Systems, Man and Cybernetics (IEEE Cat. No.04CH37583), (2004), 2513–2518. https://doi.org/10.1109/ICSMC.2004.1400707 |
[15] | S. Seung, B. Kang, H. Je, J. Park, K. Kim, S. Park, Tele-operation master-slave system for minimal invasive brain surgery, in 2009 IEEE International Conference on Robotics and Biomimetics (ROBIO), (2009), 177–182. https://doi.org/10.1109/ROBIO.2009.5420619 |
[16] | D. DeTone, T. Malisiewicz, A. Rabinovich, Superpoint: Self-supervised interest point detection and description, in Proceedings of The IEEE Conference on Computer Vision and Pattern Recognition Workshops, (2018), 224–236. https://doi.org/10.1109/CVPRW.2018.00060 |
[17] | E. Mühe, Laparoscopic cholecystectomy—late results, in Die Chirurgie und ihre Spezialgebiete Eine Symbiose, (1991), 416–423. https://doi.org/10.1007/978-3-642-95662-1_189 |
[18] |
J. Gimenez, A. Amicarelli, J. M. Toibero, F. di Sciascio, R. Carelli, Iterated conditional modes to solve simultaneous localization and mapping in markov random fields context, Int. J. Autom. Comput., 15 (2018), 310–324. https://doi.org/10.1007/s11633-017-1109-4 doi: 10.1007/s11633-017-1109-4
![]() |
[19] | N. Ketkar, J. Moolayil, Convolutional neural networks, in Deep Learning with Python, Springer International Publishing, (2017), 197–242. https://doi.org/10.1007/978-1-4842-5364-9_6 |
[20] | Z. Zhang, Flexible camera calibration by viewing a plane from unknown orientations, in Proceedings of the Seventh Ieee International Conference on Computer Vision, (1999), 666–673. |
[21] |
D. G. Lowe, Distinctive image features from scale-invariant keypoints, Int. J. Comput. Vision, 60 (2004), 91–110. https://doi.org/10.1023/B:VISI.0000029664.99615.94 doi: 10.1023/B:VISI.0000029664.99615.94
![]() |
[22] | T. Qin, P. Li, S. Shen, Vins-mono: a robust and versatile monocular visual-inertial state estimator, in IEEE Transactions on Robotics, 34 (2018), 1004–1020. https://doi.org/10.1109/TRO.2018.2853729 |
[23] | O. Garcıa, J. Civera, A. Gueme, V. Munoz, J. M. M. Montiel, Real-time 3d modeling from endoscope image sequences, Adv. Sens. Sensor Integr. Med. Robot., (2009), 1–3. |
[24] |
M. A. Fischler, R. C. Bolles, Random sample consensus: A paradigm for model fitting with applications to image analysis and automated cartography, Commun. ACM, 24 (1981), 381–395. https://doi.org/10.1145/358669.358692 doi: 10.1145/358669.358692
![]() |
[25] | F. Wolfgang, B. P. Wrobel, Bundle adjustment, in Photogrammetric Computer Vision, (2016), 643–725. https://doi.org/10.1007/978-3-319-11550-4_15 |
1. | Le Ren, Peng Yang, Rongchuan Sun, Shumei Yu, Gang Wang, Lining Sun, 2024, Semi-dense Map Reconstruction of Bronchus Based on Prior Feature Correlation, 979-8-3503-6479-8, 263, 10.1109/ISoIRS63136.2024.00058 | |
2. | Bowen Deng, 2024, 3144, 0094-243X, 030008, 10.1063/5.0215489 | |
3. | Hao Qu, 2024, 3194, 0094-243X, 020017, 10.1063/5.0222938 | |
4. | Hao Qu, 2024, 3194, 0094-243X, 020006, 10.1063/5.0222939 |
SIFT | SURF | ORB | Superpoint | Improving Superpoint | |
Extracting time /s | 1.282 | 0.182 | 0.125 | 0.103 | 0.101 |
Match time /s | 0.187 | 0.078 | 0.031 | 0.018 | 0.021 |
Extract point pairs/pairs | 334 | 121 | 268 | 352 | 349 |
Match point pairs/pairs | 334 | 121 | 268 | 263 | 349 |
False match point pair/pair | 79 | 25 | 31 | 13 | 15 |
Effective point proportion /% | 100.00 | 100.00 | 100.00 | 74.72 | 100.00 |
False match rate /% | 23.65 | 20.66 | 11.57 | 4.94 | 4.30 |
SIFT | SURF | ORB | Superpoint | Improving Superpoint | |
Extracting time /s | 1.282 | 0.182 | 0.125 | 0.103 | 0.101 |
Match time /s | 0.187 | 0.078 | 0.031 | 0.018 | 0.021 |
Extract point pairs/pairs | 334 | 121 | 268 | 352 | 349 |
Match point pairs/pairs | 334 | 121 | 268 | 263 | 349 |
False match point pair/pair | 79 | 25 | 31 | 13 | 15 |
Effective point proportion /% | 100.00 | 100.00 | 100.00 | 74.72 | 100.00 |
False match rate /% | 23.65 | 20.66 | 11.57 | 4.94 | 4.30 |