1
1
import cudf
2
2
import time
3
3
import pandas as pd
4
+ import cudf
4
5
5
6
6
7
# Measure and print the time taken to read and parse each JSON file
@@ -10,6 +11,7 @@ def measure_parsing_time(json_file_path, description, query_version):
10
11
# Measure loading and conversion time
11
12
start_time = time .time ()
12
13
df = cudf .read_json (json_file_path , lines = True )
14
+ # df = cudf.read_json(json_file_path, lines=True)
13
15
end_time = time .time ()
14
16
total_time_ms = (end_time - start_time ) * 1000
15
17
@@ -145,33 +147,9 @@ def measure_parsing_time(json_file_path, description, query_version):
145
147
146
148
147
149
# Paths to JSON files
148
- # json_file_path_nspl = '/rhome/aveda002/bigdata/Test-Files/wiki_small_records_remove.json'
149
- # json_file_path_wiki = '/rhome/aveda002/bigdata/Test-Files/wiki_small_records_remove.json'
150
- # json_file_path_walmart = '/rhome/aveda002/bigdata/Test-Files/walmart_small_records_remove.json'
151
- # json_file_path_twitter = '/rhome/aveda002/bigdata/Test-Files/twitter_small_records_remove.json'
152
- # json_file_path_google = '/rhome/aveda002/bigdata/Test-Files/google_map_small_records_remove.json'
153
- # json_file_path_bestbuy = '/rhome/aveda002/bigdata/Test-Files/bestbuy_small_records_remove.json'
154
- # json_file_path_nspl = '../../../dataset/nspl_small_records_remove.json'
155
- # json_file_path_wiki = '../../../dataset/wiki_small_records_remove.json'
156
- # json_file_path_walmart = '../../../dataset/walmart_small_records_remove.json'
150
+ # json_file_path_twitter = '../../../dataset/twitter_small_records.json'
157
151
json_file_path_twitter = '../../../dataset/twitter_small_records_remove.json'
158
- # json_file_path_twitter = '../../../dataset/twitter_sample_small_records.json' # /rhome/aveda002/bigdata/cuJSON/dataset/twitter_sample_large_record.json
159
- # json_file_path_google = '../../../dataset/google_map_small_records_remove.json'
160
- # json_file_path_bestbuy = '../../../dataset/bestbuy_small_records_remove.json'
161
-
162
-
163
-
164
- # Measure and print the parsing times
165
- # measure_parsing_time(json_file_path_wiki, "nspl", 0)
166
152
167
153
measure_parsing_time (json_file_path_twitter , "twitter" , 1 )
168
- # measure_parsing_time(json_file_path_twitter, "twitter", 2)
169
- # measure_parsing_time(json_file_path_twitter, "twitter", 3)
170
- # measure_parsing_time(json_file_path_twitter, "twitter", 4)
171
- # measure_parsing_time(json_file_path_walmart, "walmart", 5)
172
- # measure_parsing_time(json_file_path_wiki, "wiki", 6)
173
- # measure_parsing_time(json_file_path_wiki, "wiki", 7)
174
- # measure_parsing_time(json_file_path_google, "google", 8)
175
- # measure_parsing_time(json_file_path_google, "google", 9)
176
- # measure_parsing_time(json_file_path_bestbuy, "bestbuy", 10)
177
- # measure_parsing_time(json_file_path_bestbuy, "bestbuy", 11)
154
+
155
+ print (cudf .__version__ )
0 commit comments