13.1 事件循环
asyncio
包含各种特定系统实现的模块化事件循环 传输和协议抽象 对TCP、UDP、SSL、子进程、延时调用以及其他的具体支持 模仿futures模块但适用于事件循环使用的Future类 基于 yield from 的协议和任务,可以让你用顺序的方式编写并发代码 必须使用一个将产生阻塞IO的调用时,有接口可以把这个事件转移到线程池 模仿threading模块中的同步原语、可以用在单线程内的协程之间
事件循环+回调(驱动生成器)+epoll(IO多路复用)
asyncio是python用于解决异步io编程的一整套解决方案 tornado、gevent、twisted(scrapy, django channels) torando(实现web服务器), django+flask(uwsgi, gunicorn+nginx) tornado可以直接部署, nginx+tornadoimport asyncioimport time # 不再这使用同步阻塞的timeasync def get_html(url): print("start get url") await asyncio.sleep(2) # time.sleep(2) 不要这样写 print("end get url")if __name__ == "__main__": start_time = time.time() loop = asyncio.get_event_loop() tasks = [get_html("http://www.imooc.com") for i in range(10)] loop.run_until_complete(asyncio.wait(tasks)) print(time.time() - start_time)"""start get urlstart get urlstart get urlstart get urlstart get urlstart get urlstart get urlstart get urlstart get urlstart get urlend get urlend get urlend get urlend get urlend get urlend get urlend get urlend get urlend get urlend get url2.001918077468872"""
import asyncioimport timefrom functools import partial # 偏函数async def get_html(url): print("start get url") await asyncio.sleep(2) return "lewen"def callback(url, future): print(url) print("send callback email to lewen")if __name__ == "__main__": start_time = time.time() loop = asyncio.get_event_loop() # 事件循环 # task = asyncio.ensure_future(get_html("http://www.imooc.com")) # 任务的两种不同写法 task = loop.create_task(get_html("http://www.imooc.com")) task.add_done_callback(partial(callback, "http://www.imooc.com")) loop.run_until_complete(task) print(task.result())"""start get urlhttp://www.imooc.comsend callback email to lewenlewen"""
import asyncioimport timeasync def get_html(url): print("start get url") await asyncio.sleep(2) print("end get url")if __name__ == "__main__": start_time = time.time() loop = asyncio.get_event_loop() tasks = [get_html("http://www.imooc.com") for i in range(10)] # loop.run_until_complete(asyncio.gather(*tasks)) loop.run_until_complete(asyncio.wait(tasks)) # print(time.time()-start_time) # gather和wait的区别 # gather更加高层 high-level 分组 group1 = [get_html("http://projectsedu.com") for i in range(2)] group2 = [get_html("http://www.imooc.com") for i in range(2)] group1 = asyncio.gather(*group1) group2 = asyncio.gather(*group2) # group2.cancel() #取消 loop.run_until_complete(asyncio.gather(group1, group2)) print(time.time() - start_time)
13.2 协程嵌套
13.3 call_soon、call_later、call_at、call_soon_threadsafe
import asynciodef callback(sleep_times, loop): print("success time {}".format(loop.time()))def stoploop(loop): loop.stop()# call_later, call_atif __name__ == "__main__": loop = asyncio.get_event_loop() # 马上执行队列里面的task # loop.call_soon(callback, 4, loop) # loop.call_soon(stoploop, loop) # call_later() 等待多少秒后执行 # loop.call_later(2, callback, 2, loop) # loop.call_later(1, callback, 1, loop) # loop.call_later(3, callback, 3, loop) # call_at() 在某一时刻执行 now = loop.time() loop.call_at(now+2, callback, 2, loop) loop.call_at(now+1, callback, 1, loop) loop.call_at(now+3, callback, 3, loop) loop.run_forever() # loop.call_soon_threadsafe()
13.4 ThreadPoolExecutor+asyncio
# 使用多线程:在协程中集成阻塞io# 数据库等阻塞式IOimport asynciofrom concurrent.futures import ThreadPoolExecutorimport socketfrom urllib.parse import urlparsedef get_url(url): # 通过socket请求html url = urlparse(url) host = url.netloc path = url.path if path == "": path = "/" # 建立socket连接 client = socket.socket(socket.AF_INET, socket.SOCK_STREAM) # client.setblocking(False) client.connect((host, 80)) # 阻塞不会消耗cpu # 不停的询问连接是否建立好, 需要while循环不停的去检查状态 # 做计算任务或者再次发起其他的连接请求 client.send("GET {} HTTP/1.1\r\nHost:{}\r\nConnection:close\r\n\r\n".format(path, host).encode("utf8")) data = b"" while True: d = client.recv(1024) if d: data += d else: break data = data.decode("utf8") html_data = data.split("\r\n\r\n")[1] print(html_data) client.close()if __name__ == "__main__": import time start_time = time.time() loop = asyncio.get_event_loop() executor = ThreadPoolExecutor(3) # 线程池 tasks = [] for url in range(20): url = "http://www.baidu.com/s?wd={}/".format(url) task = loop.run_in_executor(executor, get_url, url) # 将阻塞的放到执行器里面 tasks.append(task) loop.run_until_complete(asyncio.wait(tasks)) print("last time:{}".format(time.time() - start_time))# 将线程池直接应用到协程里面
13.5 asyncio模拟http请求
# coding=utf-8# asyncio 没有提供http协议的接口 aiohttpimport asynciofrom urllib.parse import urlparseasync def get_url(url): # 通过socket请求html url = urlparse(url) host = url.netloc path = url.path if path == "": path = "/" # 建立socket连接 reader, writer = await asyncio.open_connection(host, 80) writer.write("GET {} HTTP/1.1\r\nHost:{}\r\nConnection:close\r\n\r\n".format(path, host).encode("utf8")) all_lines = [] async for raw_line in reader: data = raw_line.decode("utf8") all_lines.append(data) html = "\n".join(all_lines) return htmlasync def main(): tasks = [] for url in range(20): url = "http://www.baidu.com/s?wd={}/".format(url) tasks.append(asyncio.ensure_future(get_url(url))) for task in asyncio.as_completed(tasks): result = await task print(result)if __name__ == "__main__": import time start_time = time.time() loop = asyncio.get_event_loop() loop.run_until_complete(main()) print('last time:{}'.format(time.time() - start_time))
13.6 future和task
future 结果容器
task 是 future 的子类,协程和future之间的桥梁,启动协程
13.7 asyncio同步和通信
total = 0async def add(): # 1,dosomething1 # 2.io操作 # 1.dosomething3 global total for i in range(100000): total += 1async def desc(): global total for i in range(100000): total -= 1if __name__ == "__main__": import asyncio tasks = [add(), desc()] loop = asyncio.get_event_loop() loop.run_until_complete(asyncio.wait(tasks)) print(total)
13.8 aiohttp实现高并发爬虫