transfer/client.py

276 lines
11 KiB
Python
Raw Normal View History

2016-12-16 14:59:07 +08:00
from tornado.httpclient import HTTPRequest, AsyncHTTPClient
2016-12-19 14:11:17 +08:00
from tornado.simple_httpclient import SimpleAsyncHTTPClient
2016-12-16 14:59:07 +08:00
import tornado.ioloop
import tornado.web
2017-03-16 19:57:43 +08:00
import os,sys,re,threading,time
2016-12-16 14:59:07 +08:00
from tornado import gen
from functools import partial
2017-03-10 10:25:52 +08:00
import urllib.parse
import mimetypes
import math
import uuid
from concurrent.futures import ThreadPoolExecutor
import tornado.iostream
from tornado.escape import utf8
from tornado.log import gen_log
2017-03-28 11:02:57 +08:00
from ctypes import *
import os
2017-03-27 15:57:44 +08:00
#import time,datat
2016-12-19 14:11:17 +08:00
2017-03-27 15:57:44 +08:00
readchunky = False
2016-12-16 14:59:07 +08:00
total_downloaded = 0
2017-03-10 10:25:52 +08:00
threadpool = ThreadPoolExecutor(1) # A thread for reading chunks from the file
DEBUG = False
class NoQueueTimeoutHTTPClient(SimpleAsyncHTTPClient):
def fetch_impl(self, request, callback):
key = object()
self.queue.append((key, request, callback))
self.waiting[key] = (request, callback, None)
self._process_queue()
if self.queue:
gen_log.debug("max_clients limit reached, request queued. %d active, %d queued requests." % (len(self.active), len(self.queue)))
class Client:
def _gen_boundary(self, file_size):
if file_size < 1024:
blen = 10
else:
blen = math.ceil(math.log(file_size, 2))
bcnt = max(blen / 32, 1)
return "".join([str(uuid.uuid1()).replace("-", "") for _ in range(bcnt)])
2017-03-16 19:57:43 +08:00
print("_gen_boundary file_size",file_size)
def put_stream(self, url, pos, size, filename, on_response, chunk_size=8192):
2017-03-10 10:25:52 +08:00
"""Uploads file to provided url.
:param url: URL to PUT the file data to
:param filename: Name of the file (Content-Disposition header)
:param file_size: Size of the file in bytes. Used to produce a Content-Size header from file_size.
:param on_response: See AsyncHTTPClient.fetch
:return: Future content value under the provided url as binary string.
"""
2017-03-16 19:57:43 +08:00
uploadpos = pos
#file_size = os.path.getsize(filename) # Be aware: this could also block! In production, you need to do this in a separated thread too!
file_size = size # Be aware: this could also block! In production, you need to do this in a separated thread too!
2017-03-10 10:25:52 +08:00
2017-03-16 19:57:43 +08:00
print("filesize in put_stream is :",file_size)
2017-03-10 10:25:52 +08:00
ext = os.path.splitext(filename)[1].lower()
if ext:
content_type = mimetypes.types_map.get(ext, "application/octet-stream")
else:
content_type = "application/octet-stream"
enc_filename = urllib.parse.quote(filename)
# http://stackoverflow.com/questions/4526273/what-does-enctype-multipart-form-data-mean/28380690#28380690
boundary = self._gen_boundary(file_size)
CRLF = '\r\n'
post_head = b''.join(map(utf8, [
'--', boundary, CRLF,
# See https://tools.ietf.org/html/rfc5987 section 3.2.2 examples
'Content-Disposition: form-data; name="file"; filename*=utf-8\'\'%s' % enc_filename, CRLF,
'Content-Type: ', content_type, CRLF,
'Content-Transfer-Encoding: binary', CRLF,
CRLF,
]))
post_tail = b''.join(map(utf8, [
CRLF, '--', boundary, '--', CRLF
]))
content_length = len(post_head) + int(file_size) + len(post_tail)
2017-03-27 15:57:44 +08:00
#print("content_length is:",content_length)
2017-03-10 10:25:52 +08:00
headers = {
'Content-Type': 'multipart/form-data; boundary=' + boundary,
'Content-Transfer-Encoding': 'binary',
'Content-Length': str(content_length),
}
@gen.coroutine
def body_producer(write):
if DEBUG:
sys.stdout.write(post_head.decode('ascii'))
write(post_head)
remaining = file_size
2017-03-27 15:57:44 +08:00
#print("remaining in body_produceris :",remaining)
2017-03-10 10:25:52 +08:00
with open(filename, "rb") as fileobj:
2017-03-16 19:57:43 +08:00
fileobj.seek(int(uploadpos))
2017-03-10 10:25:52 +08:00
while remaining > 0:
2017-03-16 19:57:43 +08:00
#print("uploadpos in while is :",int(uploadpos))
# data = yield threadpool.submit(fileobj.read(int(remaining)), chunk_size)
data = fileobj.read(int(remaining))
# print(str(data,encoding = "utf-8"))
2017-03-10 10:25:52 +08:00
if data:
remaining -= len(data)
2017-03-16 19:57:43 +08:00
print("len(data) in if is :",len(data),"uploadpos in while is :",int(uploadpos),"remaining in if is :",len(data))
#print("data in if is :",data)
#print("remaining in if is :",len(data))
2017-03-10 10:25:52 +08:00
if DEBUG:
sys.stdout.write(data.decode('utf-8'))
yield write(data)
else:
break
if DEBUG:
sys.stdout.write(post_tail.decode('ascii'))
write(post_tail)
2017-03-16 19:57:43 +08:00
request = tornado.httpclient.HTTPRequest(url=url, request_timeout=1200, method='POST', headers=headers,body_producer=body_producer)
2017-03-10 10:25:52 +08:00
return tornado.httpclient.AsyncHTTPClient().fetch(request, callback=on_response)
2017-03-13 15:49:41 +08:00
def geturlread(action,host,filepath,uid,gid,pos,size):
2016-12-16 14:59:07 +08:00
if action=="read":
2017-03-27 15:57:44 +08:00
url = "http://"+host+"/read?filepath="+filepath+"&uid="+uid+"&gid="+gid+"&pos="+str(pos)+"&size="+str(size)
2016-12-16 14:59:07 +08:00
print(url)
return url
2017-03-16 19:57:43 +08:00
def geturlupload(action,host,targetpath,pos,size,totalsize):
2017-03-10 10:25:52 +08:00
if action=="upload":
2017-03-16 19:57:43 +08:00
url = "http://"+host+"/upload?targetpath="+targetpath+"&pos="+str(pos)+"&size="+str(size)+"&totalsize="+str(totalsize)
2017-03-10 10:25:52 +08:00
print(url)
return url
2016-12-16 14:59:07 +08:00
2017-03-27 15:57:44 +08:00
def chunky(path,pos,totalsize,chunk):
2017-03-13 15:49:41 +08:00
# print("self._max_body_size",self._max_body_size)
2017-03-10 10:25:52 +08:00
global total_downloaded
2017-03-27 15:57:44 +08:00
global readchunky
if readchunky == False and os.path.exists(path):
os.remove(path)
readchunky = True
if not os.path.exists(path):
f = open(path,'w')
f.close()
f = open(path,'ab+')
f.seek(int(pos))
f.write(chunk)
pos = pos+len(chunk)
f.flush()
if pos==totalsize:
f.close()
2017-03-10 10:25:52 +08:00
total_downloaded += len(chunk)
2017-03-13 15:49:41 +08:00
# print("chunk size",len(chunk))
2016-12-16 14:59:07 +08:00
# the OS blocks on file reads + writes -- beware how big the chunks is as it could effect things
2017-03-27 15:57:44 +08:00
def sizebwchunky(chunk):
global FILESIZE
FILESIZE = int(chunk)
@gen.coroutine
def sizebw(host,filepath):
url = "http://"+host+"/sizebw?filepath="+filepath
print(url)
request = HTTPRequest(url, streaming_callback=partial(sizebwchunky), request_timeout=300)
AsyncHTTPClient.configure('tornado.simple_httpclient.SimpleAsyncHTTPClient', max_body_size=1024*1024*1024)
http_client = AsyncHTTPClient(force_instance=True)
#AsyncHTTPClient.configure("tornado.curl_httpclient.CurlAsyncHTTPClient")
#http_client = AsyncHTTPClient()
response = yield http_client.fetch(request)
tornado.ioloop.IOLoop.instance().stop()
2016-12-16 14:59:07 +08:00
@gen.coroutine
2017-03-10 10:25:52 +08:00
def writer(host,filepath,targetdir,uid,gid,pos,size):
2017-03-27 15:57:44 +08:00
#print("writer function")
#tornado.ioloop.IOLoop.instance().start()
2016-12-19 14:11:17 +08:00
file_name = targetdir+os.path.basename(filepath)
2017-03-27 15:57:44 +08:00
'''if os.path.exists(targetdir):
2016-12-19 14:11:17 +08:00
pass
else:
2016-12-27 17:24:32 +08:00
os.makedirs(targetdir)
2016-12-19 14:11:17 +08:00
f = open(file_name,'w')
2017-03-27 15:57:44 +08:00
f.close()'''
request = HTTPRequest(geturlread("read",host,filepath,uid,gid,pos,size), streaming_callback=partial(chunky, file_name, pos, size), decompress_response=True, request_timeout=300)
AsyncHTTPClient.configure('tornado.simple_httpclient.SimpleAsyncHTTPClient', max_body_size=1024*1024*1024)
http_client = AsyncHTTPClient(force_instance=True)
#AsyncHTTPClient.configure("tornado.curl_httpclient.CurlAsyncHTTPClient")
#http_client = AsyncHTTPClient()
2016-12-16 14:59:07 +08:00
response = yield http_client.fetch(request)
2017-03-27 15:57:44 +08:00
# tornado.ioloop.IOLoop.instance().stop()
2016-12-16 14:59:07 +08:00
print("total bytes downloaded was", total_downloaded)
2017-03-27 15:57:44 +08:00
if total_downloaded==FILESIZE:
tornado.ioloop.IOLoop.instance().stop()
2017-03-28 11:02:57 +08:00
lib=cdll.LoadLibrary('./libpycall.so')
print("start")
func=lib.update_bitmap
func.argtypes=(c_int,c_int,c_char_p)
func=lib.update_bitmap(int(pos),int(size),filepath.encode("utf-8"))
print("finish")
2016-12-16 14:59:07 +08:00
2017-03-10 10:25:52 +08:00
@gen.coroutine
2017-03-16 19:57:43 +08:00
def upload(host,filepath,targetpath,pos,size):
2017-03-10 10:25:52 +08:00
def on_response(request):
print("=============== GOT RESPONSE ======")
print(request.body.decode('ascii'))
#print(request.body)
print("===================================")
2017-03-16 19:57:43 +08:00
tornado.ioloop.IOLoop.current().stop() # Stop the loop when the upload is complete.
2017-03-10 10:25:52 +08:00
2017-03-16 19:57:43 +08:00
#tornado.ioloop.IOLoop.instance().start()
2017-03-10 10:25:52 +08:00
client = Client()
2017-03-16 19:57:43 +08:00
#tornado.ioloop.IOLoop.instance().start()
yield client.put_stream(geturlupload("upload",host,targetpath,pos,size,os.path.getsize(filepath)), pos, size, filepath, on_response)
# print(geturlupload("upload",host,targetpath,pos,size))
#tornado.ioloop.IOLoop.instance().start()
# print("ioloop has already started")
2017-03-10 10:25:52 +08:00
def readentrance(host,filepath,targetdir,uid,gid,pos,size):
2017-03-27 15:57:44 +08:00
sizebw(host,filepath)
tornado.ioloop.IOLoop.instance().start()
filesize = FILESIZE
streamno = 2
if(int(size)>=filesize):
streamsize = (filesize-int(pos)) // (streamno-1)
else:
streamsize = (int(size)) // (streamno-1)
i = 0
threads = []
while i < (streamno-1):
threads.append(threading.Thread(target=writer,args=(host,filepath,targetdir,uid,gid,streamsize*i,streamsize)))
# print(streamsize*i,streamsize)
i=i+1
if (streamsize*i) < filesize:
threads.append(threading.Thread(target=writer,args=(host,filepath,targetdir,uid,gid,streamsize*i,filesize-streamsize*i)))
# print(streamsize*i,filesize-streamsize*i)
for t in threads:
t.setDaemon(True)
t.start()
# tornado.ioloop.IOLoop.instance().start()
# t.join()
2017-03-10 10:25:52 +08:00
tornado.ioloop.IOLoop.instance().start()
2017-03-27 15:57:44 +08:00
i=0
for t in threads:
t.join()
#tornado.ioloop.IOLoop.instance().stop()
2017-03-10 10:25:52 +08:00
def uploadentrance(host,filepath,targetpath):
2017-03-27 15:57:44 +08:00
streamno = 2
2017-03-16 19:57:43 +08:00
filesize = os.path.getsize(filepath)
streamsize = filesize // (streamno-1)
i = 0
threads = []
while i < (streamno-1):
threads.append(threading.Thread(target=upload,args=(host,filepath,targetpath,streamsize*i,streamsize)))
i=i+1
if (streamsize*i) < filesize:
threads.append(threading.Thread(target=upload,args=(host,filepath,targetpath,streamsize*i,filesize-streamsize*i)))
for t in threads:
t.setDaemon(True)
t.start()
2017-03-27 15:57:44 +08:00
# tornado.ioloop.IOLoop.instance().start()
#t.join()
#print("upload all!")
tornado.ioloop.IOLoop.instance().start()
2017-03-16 19:57:43 +08:00
t.join()
2017-03-27 15:57:44 +08:00
'''if __name__=="__main__":
2017-03-16 19:57:43 +08:00
uploadentrance("202.122.37.90:28003","/root/leaf/transfer/night.mkv","/home/wangcong/leaf/upload/night.mkv")
#uploadentrance("202.122.37.90:28006","/root/leaf/test.cpp","/root/leaf/pytoc/upload/test.cpp")
2017-03-27 15:57:44 +08:00
tornado.ioloop.IOLoop.instance().start()'''