forked from kk7nc/Text_Classification
-
Notifications
You must be signed in to change notification settings - Fork 0
Expand file tree
/
Copy pathDownload_Glove.py
More file actions
78 lines (57 loc) · 2.41 KB
/
Download_Glove.py
File metadata and controls
78 lines (57 loc) · 2.41 KB
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
'''''''''''''''''''''''''''''''''''''''''''''''''''''''''''''''''
RMDL: Random Multimodel Deep Learning for Classification
* Copyright (C) 2018 Kamran Kowsari <kk7nc@virginia.edu>
* Last Update: 04/25/2018
* This file is part of RMDL project, University of Virginia.
* Free to use, change, share and distribute source code of RMDL
* Refrenced paper : RMDL: Random Multimodel Deep Learning for Classification
* Refrenced paper : An Improvement of Data Classification using Random Multimodel Deep Learning (RMDL)
* Comments and Error: email: kk7nc@virginia.edu
'''''''''''''''''''''''''''''''''''''''''''''''''''''''''''''''''
from __future__ import print_function
import os, sys, tarfile
import numpy as np
import zipfile
if sys.version_info >= (3, 0, 0):
import urllib.request as urllib # ugly but works
else:
import urllib
print(sys.version_info)
# image shape
# path to the directory with the data
DATA_DIR = '.\Glove'
# url of the binary data
# path to the binary train file with image data
def download_and_extract(data='Wikipedia'):
"""
Download and extract the GloVe
:return: None
"""
if data=='Wikipedia':
DATA_URL = 'http://nlp.stanford.edu/data/glove.6B.zip'
elif data=='Common_Crawl_840B':
DATA_URL = 'http://nlp.stanford.edu/data/wordvecs/glove.840B.300d.zip'
elif data=='Common_Crawl_42B':
DATA_URL = 'http://nlp.stanford.edu/data/wordvecs/glove.42B.300d.zip'
elif data=='Twitter':
DATA_URL = 'http://nlp.stanford.edu/data/wordvecs/glove.twitter.27B.zip'
else:
print("prameter should be Twitter, Common_Crawl_42B, Common_Crawl_840B, or Wikipedia")
exit(0)
dest_directory = DATA_DIR
if not os.path.exists(dest_directory):
os.makedirs(dest_directory)
filename = DATA_URL.split('/')[-1]
filepath = os.path.join(dest_directory, filename)
print(filepath)
path = os.path.abspath(dest_directory)
if not os.path.exists(filepath):
def _progress(count, block_size, total_size):
sys.stdout.write('\rDownloading %s %.2f%%' % (filename,
float(count * block_size) / float(total_size) * 100.0))
sys.stdout.flush()
filepath, _ = urllib.urlretrieve(DATA_URL, filepath)#, reporthook=_progress)
zip_ref = zipfile.ZipFile(filepath, 'r')
zip_ref.extractall(DATA_DIR)
zip_ref.close()
return path