@@ -52,7 +52,8 @@ def test_hdfs_arguments(self):
52
52
"logPath" : "/tmp",
53
53
"bandwidthPerMap" : "20",
54
54
"preserveXAttrs" : false,
55
- "exclusionFilters" : ["ac"]
55
+ "exclusionFilters" : ["ac"],
56
+ "raiseSnapshotDiffFailures" : false
56
57
}'''
57
58
args = utils .deserialize (RAW , ApiHdfsReplicationArguments )
58
59
self .assertEquals ('vst2' , args .sourceService .peerName )
@@ -73,6 +74,7 @@ def test_hdfs_arguments(self):
73
74
self .assertFalse (args .skipTrash )
74
75
self .assertEquals ('DYNAMIC' , args .replicationStrategy )
75
76
self .assertFalse (args .preserveXAttrs )
77
+ self .assertFalse (args .raiseSnapshotDiffFailures )
76
78
77
79
def test_hdfs_cloud_arguments (self ):
78
80
RAW = '''{
@@ -298,6 +300,8 @@ def test_hive_results(self):
298
300
def test_schedule (self ):
299
301
RAW = '''{
300
302
"id" : 39,
303
+ "displayName" : "testrun",
304
+ "description" : "a sample description",
301
305
"startTime" : "2012-12-10T23:11:31.041Z",
302
306
"interval" : 1,
303
307
"intervalUnit" : "DAY",
@@ -342,6 +346,9 @@ def test_schedule(self):
342
346
"numFilesCopyFailed" : 0,
343
347
"numBytesCopyFailed" : 0,
344
348
"dryRun" : false,
349
+ "remainingTime" : 10000,
350
+ "throughput" : 5.10,
351
+ "estimatedCompletionTime" : "2018-06-11T18:22:25.123Z",
345
352
"failedFiles": [ ]
346
353
},
347
354
"dryRun" : false,
@@ -371,23 +378,34 @@ def test_schedule(self):
371
378
"skipTrash" : false,
372
379
"preserveXAttrs" : false
373
380
},
381
+ "runInvalidateMetadata" : true,
382
+ "numThreads" : 4,
374
383
"dryRun" : false
375
384
}
376
385
}'''
377
386
sched = utils .deserialize (RAW , ApiReplicationSchedule )
378
387
self .assertEqual (39 , sched .id )
388
+ self .assertEqual ("testrun" , sched .displayName )
389
+ self .assertEqual ("a sample description" , sched .description )
379
390
self .assertEqual (self ._parse_time ("2012-12-10T23:11:31.041Z" ), sched .startTime )
380
391
self .assertEqual ('DAY' , sched .intervalUnit )
381
392
self .assertEqual (1 , sched .interval )
382
393
self .assertFalse (sched .paused )
383
394
self .assertEqual (self ._parse_time ("2013-01-15T23:11:31.041Z" ), sched .nextRun )
384
395
self .assertFalse (sched .alertOnStart )
385
396
self .assertIsNotNone (sched .hiveArguments )
397
+ self .assertTrue (sched .hiveArguments .runInvalidateMetadata )
398
+ self .assertEqual (4 , sched .hiveArguments .numThreads )
386
399
387
400
self .assertEqual (1 , len (sched .history ))
388
401
self .assertIsInstance (sched .history [0 ], ApiReplicationCommand )
389
402
self .assertEqual ('default' , sched .history [0 ].hiveResult .tables [0 ].database )
390
403
self .assertEqual (92158 , sched .history [0 ].hiveResult .dataReplicationResult .numBytesSkipped )
404
+ self .assertEqual (10000 , sched .history [0 ].hiveResult .dataReplicationResult .remainingTime )
405
+ self .assertEqual (5.10 , sched .history [0 ].hiveResult .dataReplicationResult .throughput )
406
+ self .assertEqual ("2018-06-11T18:22:25.123Z" ,
407
+ sched .history [0 ].hiveResult .dataReplicationResult .estimatedCompletionTime
408
+ )
391
409
self .assertEqual (3 , sched .history [0 ].hiveResult .tableCount )
392
410
self .assertEqual (0 , sched .history [0 ].hiveResult .errorCount )
393
411
0 commit comments