-
Notifications
You must be signed in to change notification settings - Fork 0
Expand file tree
/
Copy pathcrawl_data.py
More file actions
29 lines (25 loc) · 1.23 KB
/
crawl_data.py
File metadata and controls
29 lines (25 loc) · 1.23 KB
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
import argparse
from src.data_helper import DataProcessor
def parse_args():
parser = argparse.ArgumentParser()
parser.add_argument('-b', '--batch-size', type=int, default=100, help='batch size')
parser.add_argument('-s', '--skiprows', type=int, default=26560, help='skip size')
parser.add_argument('-d', '--data-dir', type=str, default='data', help='data directory')
parser.add_argument('-o', '--output-dir', type=str, default='input/data/64x64/', help='output directory')
parser.add_argument('-t', '--type', type=str, default='train', help='data type')
parser.add_argument('-f', '--first-batch', type=int, default=0, help='first batch')
args = parser.parse_args()
return args
def main():
args = parse_args()
file_dir = "input/2017_11/" + args.type + "/"
# file_dir = "input/data/stl10_binary/"
file_name = "images.csv"
DataProcessor.read_url_csv(file_dir, file_name,
chunksize=args.batch_size,
skiprows=args.skiprows,
first_batch=args.first_batch,
output_dir=args.output_dir,
only_save_img=True)
if __name__ == "__main__":
main()