Skip to content

Commit ac56d5c

Browse files
committed
update pubtabnet dataset
1 parent d3081fc commit ac56d5c

27 files changed

+400
-77
lines changed
Lines changed: 54 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,54 @@
1+
# dataset settings
2+
dataset_type = 'publaynetDataset'
3+
data_root = '/data4T/ntanh'
4+
img_norm_cfg = dict(
5+
mean=[123.675, 116.28, 103.53], std=[58.395, 57.12, 57.375], to_rgb=True)
6+
crop_size = (512, 512)
7+
train_pipeline = [
8+
dict(type='LoadImageFromFile'),
9+
dict(type='LoadAnnotations'),
10+
dict(type='Resize', img_scale=(2048, 512), ratio_range=(0.5, 2.0)),
11+
dict(type='RandomCrop', crop_size=crop_size, cat_max_ratio=0.75),
12+
dict(type='RandomFlip', flip_ratio=0.5),
13+
dict(type='PhotoMetricDistortion'),
14+
dict(type='Normalize', **img_norm_cfg),
15+
dict(type='Pad', size=crop_size, pad_val=0, seg_pad_val=255),
16+
dict(type='DefaultFormatBundle'),
17+
dict(type='Collect', keys=['img', 'gt_semantic_seg']),
18+
]
19+
test_pipeline = [
20+
dict(type='LoadImageFromFile'),
21+
dict(
22+
type='MultiScaleFlipAug',
23+
img_scale=(2048, 512),
24+
# img_ratios=[0.5, 0.75, 1.0, 1.25, 1.5, 1.75],
25+
flip=False,
26+
transforms=[
27+
dict(type='Resize', keep_ratio=True),
28+
dict(type='RandomFlip'),
29+
dict(type='Normalize', **img_norm_cfg),
30+
dict(type='ImageToTensor', keys=['img']),
31+
dict(type='Collect', keys=['img']),
32+
])
33+
]
34+
data = dict(
35+
samples_per_gpu=4,
36+
workers_per_gpu=4,
37+
train=dict(
38+
type=dataset_type,
39+
data_root=data_root,
40+
img_dir='publaynet/train',
41+
ann_dir='publaynet_gen_gt_oct2.1/train/label',
42+
pipeline=train_pipeline),
43+
val=dict(
44+
type=dataset_type,
45+
data_root=data_root,
46+
img_dir='publaynet/val',
47+
ann_dir='publaynet_gen_gt_oct2.1/val/label',
48+
pipeline=test_pipeline),
49+
test=dict(
50+
type=dataset_type,
51+
data_root=data_root,
52+
img_dir='publaynet/val',
53+
ann_dir='publaynet_gen_gt_oct2.1/val/label',
54+
pipeline=test_pipeline))
Lines changed: 54 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,54 @@
1+
# dataset settings
2+
dataset_type = 'publaynet_split1Dataset'
3+
data_root = '/data20.04/data/doc_structure/publaynet'
4+
img_norm_cfg = dict(
5+
mean=[123.675, 116.28, 103.53], std=[58.395, 57.12, 57.375], to_rgb=True)
6+
crop_size = (512, 512)
7+
train_pipeline = [
8+
dict(type='LoadImageFromFile'),
9+
dict(type='LoadAnnotations'),
10+
dict(type='Resize', img_scale=(2048, 512), ratio_range=(0.5, 2.0)),
11+
dict(type='RandomCrop', crop_size=crop_size, cat_max_ratio=0.75),
12+
dict(type='RandomFlip', flip_ratio=0.5),
13+
dict(type='PhotoMetricDistortion'),
14+
dict(type='Normalize', **img_norm_cfg),
15+
dict(type='Pad', size=crop_size, pad_val=0, seg_pad_val=255),
16+
dict(type='DefaultFormatBundle'),
17+
dict(type='Collect', keys=['img', 'gt_semantic_seg']),
18+
]
19+
test_pipeline = [
20+
dict(type='LoadImageFromFile'),
21+
dict(
22+
type='MultiScaleFlipAug',
23+
img_scale=(2048, 512),
24+
# img_ratios=[0.5, 0.75, 1.0, 1.25, 1.5, 1.75],
25+
flip=False,
26+
transforms=[
27+
dict(type='Resize', keep_ratio=True),
28+
dict(type='RandomFlip'),
29+
dict(type='Normalize', **img_norm_cfg),
30+
dict(type='ImageToTensor', keys=['img']),
31+
dict(type='Collect', keys=['img']),
32+
])
33+
]
34+
data = dict(
35+
samples_per_gpu=4,
36+
workers_per_gpu=8,
37+
train=dict(
38+
type=dataset_type,
39+
data_root=data_root,
40+
img_dir='img_dir/train',
41+
ann_dir='ann_dir/train',
42+
pipeline=train_pipeline),
43+
val=dict(
44+
type=dataset_type,
45+
data_root=data_root,
46+
img_dir='img_dir/val',
47+
ann_dir='ann_dir/val',
48+
pipeline=test_pipeline),
49+
test=dict(
50+
type=dataset_type,
51+
data_root=data_root,
52+
img_dir='img_dir/val',
53+
ann_dir='ann_dir/val',
54+
pipeline=test_pipeline))

configs/_base_/datasets/table_structure1.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,6 @@
11
# dataset settings
22
dataset_type = 'table_structure1Dataset'
3-
data_root = 'data/table_structure1'
3+
data_root = '/home/cuongnd/PycharmProjects/open-mmlab/mmsegmentation/data/table_structure1'
44
img_norm_cfg = dict(
55
mean=[123.675, 116.28, 103.53], std=[58.395, 57.12, 57.375], to_rgb=True)
66
crop_size = (512, 512)

configs/_base_/models/fast_scnn.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,5 @@
11
# model settings
2-
norm_cfg = dict(type='SyncBN', requires_grad=True, momentum=0.01)
2+
norm_cfg = dict(type='BN', requires_grad=True, momentum=0.01)
33
model = dict(
44
type='EncoderDecoder',
55
backbone=dict(

configs/_base_/schedules/schedule_20k.py

Lines changed: 3 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -5,5 +5,6 @@
55
lr_config = dict(policy='poly', power=0.9, min_lr=1e-4, by_epoch=False)
66
# runtime settings
77
total_iters = 20000
8-
checkpoint_config = dict(by_epoch=False, interval=2000)
9-
evaluation = dict(interval=2000, metric='mIoU')
8+
interval=1000
9+
checkpoint_config = dict(by_epoch=False, interval=interval)
10+
evaluation = dict(interval=interval, metric='mIoU')
Lines changed: 9 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,9 @@
1+
# optimizer
2+
optimizer = dict(type='SGD', lr=0.001, momentum=0.9, weight_decay=0.0005)
3+
optimizer_config = dict()
4+
# learning policy
5+
lr_config = dict(policy='poly', power=0.9, min_lr=1e-5, by_epoch=False)
6+
# runtime settings
7+
total_iters = 320000
8+
checkpoint_config = dict(by_epoch=False, interval=32000)
9+
evaluation = dict(interval=320000, metric='mIoU')

configs/_base_/schedules/schedule_40k.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -5,5 +5,5 @@
55
lr_config = dict(policy='poly', power=0.9, min_lr=1e-4, by_epoch=False)
66
# runtime settings
77
total_iters = 40000
8-
checkpoint_config = dict(by_epoch=False, interval=4000)
9-
evaluation = dict(interval=4000, metric='mIoU')
8+
checkpoint_config = dict(by_epoch=False, interval=interval)
9+
evaluation = dict(interval=interval, metric='mIoU')

configs/_base_/schedules/schedule_80k.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -6,4 +6,4 @@
66
# runtime settings
77
total_iters = 80000
88
checkpoint_config = dict(by_epoch=False, interval=8000)
9-
evaluation = dict(interval=8000, metric='mIoU')
9+
evaluation = dict(interval=80000, metric='mIoU')
Lines changed: 10 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,10 @@
1+
_base_ = [
2+
'../_base_/models/fast_scnn.py', '../_base_/datasets/publaynet.py',
3+
'../_base_/default_runtime.py', '../_base_/schedules/schedule_320k.py'
4+
]
5+
6+
# Re-config the data sampler.
7+
data = dict(samples_per_gpu=16, workers_per_gpu=4)
8+
9+
# Re-config the optimizer.
10+
optimizer = dict(type='SGD', lr=0.12, momentum=0.9, weight_decay=4e-5)
Lines changed: 10 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,10 @@
1+
_base_ = [
2+
'../_base_/models/fast_scnn.py', '../_base_/datasets/publaynet_split1.py',
3+
'../_base_/default_runtime.py', '../_base_/schedules/schedule_80k.py'
4+
]
5+
6+
# Re-config the data sampler.
7+
data = dict(samples_per_gpu=16, workers_per_gpu=4)
8+
9+
# Re-config the optimizer.
10+
optimizer = dict(type='SGD', lr=0.12, momentum=0.9, weight_decay=4e-5)

0 commit comments

Comments
 (0)