Added aiohttp support for asynchronous peers requests
This commit is contained in:
pare
5288dde459
commit
2283983227
S'han modificat 1 arxius amb 77 adicions i 72 eliminacions
117
getworld.py
117
getworld.py
|
@ -20,15 +20,19 @@ import psycopg2
|
|||
from itertools import product
|
||||
|
||||
from multiprocessing import Pool, Lock, Process, Queue, current_process
|
||||
import queue # imported for using queue.Empty exception
|
||||
import queue
|
||||
import multiprocessing
|
||||
|
||||
from decimal import *
|
||||
getcontext().prec = 2
|
||||
import aiohttp
|
||||
import aiodns
|
||||
import asyncio
|
||||
from aiohttp import ClientError, ClientSession, ClientConnectionError, ClientConnectorError, ClientSSLError, ClientConnectorSSLError, ServerTimeoutError
|
||||
from asyncio import TimeoutError
|
||||
import socket
|
||||
from socket import gaierror, gethostbyname
|
||||
|
||||
###############################################################################
|
||||
# INITIALISATION
|
||||
###############################################################################
|
||||
updated_at = datetime.now()
|
||||
peers_api = '/api/v1/instance/peers?'
|
||||
|
||||
def is_json(myjson):
|
||||
try:
|
||||
|
@ -37,24 +41,49 @@ def is_json(myjson):
|
|||
return False
|
||||
return True
|
||||
|
||||
def getpeers(server):
|
||||
def getserver(server):
|
||||
|
||||
global updated_at
|
||||
if server.find(".") == -1:
|
||||
return
|
||||
if server.find("@") != -1:
|
||||
return
|
||||
if server.find("/") != -1:
|
||||
return
|
||||
if server.find(":") != -1:
|
||||
return
|
||||
|
||||
try:
|
||||
|
||||
res = requests.get('https://' + server + '/api/v1/instance/peers?',timeout=3)
|
||||
loop = asyncio.get_event_loop()
|
||||
coroutines = [getpeers(server)]
|
||||
loop.run_until_complete(asyncio.gather(*coroutines, return_exceptions=True))
|
||||
|
||||
if (res.ok):
|
||||
except:
|
||||
|
||||
if server.find(".") != -1 and server.find("@") == -1:
|
||||
server_peers = res.json()
|
||||
print("Server: " + server + ", " + "federated with " + str(len(server_peers)) + " servers")
|
||||
else:
|
||||
print("Server " + str(server) + " is not a domain")
|
||||
pass
|
||||
|
||||
async def getpeers(server):
|
||||
|
||||
try:
|
||||
|
||||
socket.gethostbyname(server)
|
||||
|
||||
except socket.gaierror:
|
||||
|
||||
return
|
||||
|
||||
url = 'https://' + server
|
||||
|
||||
timeout = aiohttp.ClientTimeout(total=3)
|
||||
async with aiohttp.ClientSession(timeout=timeout) as session:
|
||||
try:
|
||||
async with session.get(url+peers_api) as response:
|
||||
if response.status == 200:
|
||||
try:
|
||||
response_json = await response.json()
|
||||
print("Server: " + server + ", " + "federated with " + str(len(response_json)) + " servers")
|
||||
i = 0
|
||||
while i < len(server_peers) and server.find(".") != -1 and server.find("@") == -1:
|
||||
while i < len(response_json):
|
||||
|
||||
saved_at = datetime.now()
|
||||
insert_sql = "INSERT INTO world(server, federated_with, updated_at, saved_at) VALUES(%s,%s,%s,%s) ON CONFLICT DO NOTHING"
|
||||
|
@ -65,7 +94,7 @@ def getpeers(server):
|
|||
|
||||
cur = conn.cursor()
|
||||
|
||||
cur.execute(insert_sql, (server_peers[i], server, updated_at, saved_at,))
|
||||
cur.execute(insert_sql, (response_json[i], server, updated_at, saved_at,))
|
||||
|
||||
conn.commit()
|
||||
|
||||
|
@ -83,40 +112,17 @@ def getpeers(server):
|
|||
|
||||
i += 1
|
||||
|
||||
except KeyError as e:
|
||||
except:
|
||||
|
||||
pass
|
||||
return
|
||||
|
||||
except ValueError as verr:
|
||||
except aiohttp.ClientConnectorError as err:
|
||||
|
||||
pass
|
||||
return
|
||||
|
||||
except requests.exceptions.SSLError as errssl:
|
||||
|
||||
pass
|
||||
return
|
||||
|
||||
except requests.exceptions.HTTPError as errh:
|
||||
|
||||
pass
|
||||
return
|
||||
|
||||
except requests.exceptions.ConnectionError as errc:
|
||||
|
||||
pass
|
||||
return
|
||||
|
||||
except requests.exceptions.Timeout as errt:
|
||||
|
||||
pass
|
||||
return
|
||||
|
||||
except requests.exceptions.RequestException as err:
|
||||
|
||||
pass
|
||||
return
|
||||
###############################################################################
|
||||
# INITIALISATION
|
||||
###############################################################################
|
||||
|
||||
# Returns the parameter from the specified file
|
||||
def get_parameter( parameter, file_path ):
|
||||
|
@ -162,19 +168,18 @@ mastodon = Mastodon(
|
|||
headers={ 'Authorization': 'Bearer %s'%uc_access_token }
|
||||
|
||||
###############################################################################
|
||||
# get current datetime and Mastodon hostname peers
|
||||
###############################################################################
|
||||
# main
|
||||
|
||||
updated_at = datetime.now()
|
||||
if __name__ == '__main__':
|
||||
|
||||
getpeers(mastodon_hostname)
|
||||
self_peers = mastodon.instance_peers()
|
||||
getserver(mastodon_hostname)
|
||||
self_peers = mastodon.instance_peers()
|
||||
|
||||
###########################################################################
|
||||
###########################################################################
|
||||
|
||||
nprocs = multiprocessing.cpu_count()
|
||||
with multiprocessing.Pool(processes=nprocs) as pool:
|
||||
results = pool.starmap(getpeers, product(self_peers))
|
||||
nprocs = multiprocessing.cpu_count()
|
||||
with multiprocessing.Pool(processes=nprocs) as pool:
|
||||
results = pool.starmap(getserver, product(self_peers))
|
||||
|
||||
exec_time = str(round((time.time() - start_time), 2))
|
||||
print(exec_time)
|
||||
exec_time = str(round((time.time() - start_time), 2))
|
||||
print(exec_time)
|
||||
|
|
Loading…
Referencia en una nova incidència