Implemented recursive resource segmentation for large transfers

This commit is contained in:
Mark Qvist 2020-08-12 21:49:59 +02:00
parent 73a3516db8
commit e4dfd052e6
2 changed files with 26 additions and 19 deletions

View File

@ -251,8 +251,10 @@ def client(destination_hexhash, configpath):
# Requests the specified file from the server # Requests the specified file from the server
def download(filename): def download(filename):
global server_link, menu_mode, current_filename global server_link, menu_mode, current_filename, transfer_size, download_started
current_filename = filename current_filename = filename
download_started = 0
transfer_size = 0
# We just create a packet containing the # We just create a packet containing the
# requested filename, and send it down the # requested filename, and send it down the
@ -457,9 +459,11 @@ def download_began(resource):
global menu_mode, current_download, download_started, transfer_size, file_size global menu_mode, current_download, download_started, transfer_size, file_size
current_download = resource current_download = resource
if download_started == 0:
download_started = time.time() download_started = time.time()
transfer_size = resource.size
file_size = resource.uncompressed_size transfer_size += resource.size
file_size = resource.total_size
menu_mode = "downloading" menu_mode = "downloading"

View File

@ -62,6 +62,7 @@ class Resource:
resource.flags = adv.f resource.flags = adv.f
resource.size = adv.t resource.size = adv.t
resource.total_size = adv.d
resource.uncompressed_size = adv.d resource.uncompressed_size = adv.d
resource.hash = adv.h resource.hash = adv.h
resource.original_hash = adv.o resource.original_hash = adv.o
@ -122,6 +123,7 @@ class Resource:
resource_data = None resource_data = None
if hasattr(data, "read"): if hasattr(data, "read"):
data_size = os.stat(data.name).st_size data_size = os.stat(data.name).st_size
self.total_size = data_size
self.grand_total_parts = math.ceil(data_size/Resource.SDU) self.grand_total_parts = math.ceil(data_size/Resource.SDU)
if data_size <= Resource.MAX_EFFICIENT_SIZE: if data_size <= Resource.MAX_EFFICIENT_SIZE:
@ -144,6 +146,7 @@ class Resource:
elif isinstance(data, bytes): elif isinstance(data, bytes):
data_size = len(data) data_size = len(data)
self.grand_total_parts = math.ceil(data_size/Resource.SDU) self.grand_total_parts = math.ceil(data_size/Resource.SDU)
self.total_size = data_size
resource_data = data resource_data = data
self.total_segments = 1 self.total_segments = 1
@ -467,7 +470,7 @@ class Resource:
RNS.log("Error while cleaning up resource files, the contained exception was:", RNS.LOG_ERROR) RNS.log("Error while cleaning up resource files, the contained exception was:", RNS.LOG_ERROR)
RNS.log(str(e)) RNS.log(str(e))
else: else:
RNS.log("Resource segment "+str(self.segment_index)+" of "+str(self.total_segments)+" received, waiting for next segment to be announced", RNS.LOG_VERBOSE) RNS.log("Resource segment "+str(self.segment_index)+" of "+str(self.total_segments)+" received, waiting for next segment to be announced", RNS.LOG_DEBUG)
def prove(self): def prove(self):
@ -728,7 +731,7 @@ class Resource:
self.processed_parts = (self.segment_index-1)*math.ceil(Resource.MAX_EFFICIENT_SIZE/Resource.SDU) self.processed_parts = (self.segment_index-1)*math.ceil(Resource.MAX_EFFICIENT_SIZE/Resource.SDU)
self.processed_parts += self.received_count self.processed_parts += self.received_count
if self.split: if self.split:
self.progress_total_parts = float((self.total_segments-1)*math.ceil(Resource.MAX_EFFICIENT_SIZE/Resource.SDU)+self.total_parts) self.progress_total_parts = float(math.ceil(self.total_size/Resource.SDU))
else: else:
self.progress_total_parts = float(self.total_parts) self.progress_total_parts = float(self.total_parts)
@ -747,7 +750,7 @@ class ResourceAdvertisement:
def __init__(self, resource=None): def __init__(self, resource=None):
if resource != None: if resource != None:
self.t = resource.size # Transfer size self.t = resource.size # Transfer size
self.d = resource.uncompressed_size # Data size self.d = resource.total_size # Total uncompressed data size
self.n = len(resource.parts) # Number of parts self.n = len(resource.parts) # Number of parts
self.h = resource.hash # Resource hash self.h = resource.hash # Resource hash
self.r = resource.random_hash # Resource random hash self.r = resource.random_hash # Resource random hash