-
Notifications
You must be signed in to change notification settings - Fork 0
/
Copy pathconfig.json
54 lines (54 loc) · 3.49 KB
/
config.json
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
{
"id": "ce585219-d617-4864-b225-06b52532ea95",
"meta": {
"name": "sfm-learner-pose",
"application_area": "Capsule Endoscopy",
"task": "Pose & Depth Estimation",
"task_extended": "Unsupervised Pose & Depth Estimation",
"data_type": "Image/Photo",
"data_source": "http://www.cvlibs.net/datasets/kitti/"
},
"publication": {
"title": "Unsupervised Odometry and Depth Learning for Endoscopic Capsule Robots",
"source": "arxiv",
"year": 2018,
"authors": "Turan Mehmet, Ornek E. Pinar, Ibrahimli Nail, Giracoglu Can, Almalioglu Yasin, Yanik M. Fatih, Sitti Metin",
"email": "[email protected]",
"abstract": "In the last decade, many medical companies and research groups have tried to convert passive capsule endoscopes as an emerging and minimally invasive diagnostic technology into actively steerable endoscopic capsule robots which will provide more intuitive disease detection, targeted drug delivery and biopsy-like operations in the gastrointestinal(GI) tract. In this study, we introduce a fully unsupervised, real-time odometry and depth learner for monocular endoscopic capsule robots. We establish the supervision by warping view sequences and assigning the re-projection minimization to the loss function, which we adopt in multi-view pose estimation and single-view depth estimation network. Detailed quantitative and qualitative analyses of the proposed framework performed on non-rigidly deformable ex-vivo porcine stomach datasets proves the effectiveness of the method in terms of motion estimation and depth recovery.",
"url": "https://arxiv.org/abs/1803.01047",
"google_scholar": "https://scholar.google.com/scholar?oi=bibs&hl=en&cites=16480963845558763827&as_sdt=5",
"bibtex": "@ARTICLE{2018arXiv180301047T, author = {{Turan}, Mehmet and {Pinar Ornek}, Evin and {Ibrahimli}, Nail and {Giracoglu}, Can and {Almalioglu}, Yasin and {Yanik}, Mehmet Fatih and {Sitti}, Metin}, title = {Unsupervised Odometry and Depth Learning for Endoscopic Capsule Robots}, journal = {arXiv e-prints}, keywords = {Computer Science - Robotics}, year = 2018, month = Mar, eid = {arXiv:1803.01047}, pages = {arXiv:1803.01047}, archivePrefix = {arXiv}, eprint = {1803.01047}, primaryClass = {cs.RO}}"
},
"model": {
"description": "The model consists of two networks trained together, first one being single-view depth network and the second one pose-reliability network. Both of them have decoder-encoder design, a stack of convolutional networks.",
"provenance": "contributed by author",
"architecture": "Convolutional Neural Network(CNN), Decoder/Encoder",
"learning_type": "Unsupervised learning",
"format": ".pb",
"io": {
"input": {
"description": "The SFM-Learner input is an image that consists 3 consecutive frames in which the pose estimation will be applied on. Hence, a video can be fed to model by extracting the frames and patching each 3 of them into an image through sliding the frames for each image.",
"format": ["image/png", "image/jpg", "image/jpeg"],
"dim_limits": [
{
"min": 3
},
{
"min": 128
},
{
"min": 416
}
]
},
"output": [
{
"name": "Pose outputs",
"type": "vector",
"description": "The output consists of the estimated pose parameters: translation_x, translation_y, translation_z and the rotation values in a quaternion format."
}
]
}
},
"modelhub": {}
}