Files
twitter-api-client/scripts/update.py
2023-12-10 18:36:24 -08:00

108 lines
3.6 KiB
Python

import asyncio
import logging.config
import re
import subprocess
from pathlib import Path
import aiofiles
import orjson
from httpx import AsyncClient, Client
from twitter.constants import *
try:
import nest_asyncio
nest_asyncio.apply()
except:
...
try:
import uvloop
uvloop.install()
except:
...
STRINGS = Path('strings.txt')
PATHS = Path('paths.txt')
JS_FILES_MAP = Path('js.json')
JS_FILES = Path('js')
OPERATIONS = Path('operations')
JS_FILES.mkdir(exist_ok=True, parents=True)
logging.config.dictConfig(LOG_CONFIG)
logger = logging.getLogger('twitter')
def get_operations(session: Client) -> None:
"""
Get operations and their respective queryId and feature definitions
"""
r1 = session.get('https://twitter.com')
m = re.findall('href="(https\:\/\/abs\.twimg\.com\/responsive-web\/client-web\/main\.\w+\.js)"', r1.text)
r2 = session.get(m[0])
tmp = '[{' + re.search('\d+:\w=>\{\w\.exports=\{.*?(?=,\d+:\w=>\{"use strict";)', r2.text).group() + '}]'
tmp = re.sub('\w\.exports=', 'return', tmp)
tmp = re.sub(',\d+:\([\w,]+\).*', '}]', tmp)
js = 'const obj={},out=Object.entries(O[0]).forEach(([e,t])=>{let a=t(),o={};for(let r of a.metadata.featureSwitches)o[r]=!0;obj[a.operationName]={queryId:a.queryId,variables:{},features:o}});require("fs").writeFile("' + OPERATIONS.with_suffix(
'.json').name + '",JSON.stringify(Object.fromEntries(Object.entries(obj).sort())),e=>e);'
js_out = OPERATIONS.with_suffix('.js')
js_out.expanduser().write_text(f"O={tmp};" + js)
subprocess.run(f'node {js_out}', shell=True)
async def process(session: Client, fn: callable, urls: any, **kwargs) -> tuple:
async with AsyncClient(follow_redirects=True, headers=session.headers) as s:
return await asyncio.gather(*(fn(s, u, **kwargs) for u in urls))
async def get(session: AsyncClient, url: str) -> tuple[str, str]:
try:
logger.debug(f"GET {url}")
r = await session.get(url)
async with aiofiles.open(JS_FILES / url.split('/')[-1], 'wb') as f:
await f.write(r.content)
return url, r.text
except Exception as e:
logger.error(f"[{RED}failed{RESET}] Failed to get {url}\n{e}")
def get_strings():
# find strings < 120 chars long
# queryId's are usually 22 chars long
s = set()
for p in JS_FILES.iterdir():
s |= set(x.strip() for x in re.split('["\'`]', p.read_text()) if
# ((len(x) == 22) and (not re.search('[\[\]\{\}\(\)]', x))))
((len(x) < 120) and (not re.search('[\[\]\{\}\(\)]', x))))
STRINGS.write_text('\n'.join(sorted(s, reverse=True)))
PATHS.write_text('\n'.join(sorted(s for s in s if '/' in s)))
def get_features():
operations = orjson.loads(OPERATIONS.with_suffix('.json').read_bytes())
features = {}
for k, v in operations.items():
features |= v.get('features', {})
Path('features.json').write_bytes(orjson.dumps(dict(sorted(features.items())), option=orjson.OPT_INDENT_2))
def main():
session = Client(headers={
'user-agent': 'Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/111.0.0.0 Safari/537.36',
}, follow_redirects=True)
get_operations(session)
# urls = (
# f'{_base}/{k}.{v}{_a}'
# for k, v in orjson.loads(JS_FILES_MAP.read_text()).items()
# if not re.search('participantreaction|\.countries-|emojipicker|i18n|icons\/', k, flags=re.I)
# # if 'endpoint' in k
# )
# asyncio.run(process(session, get, urls))
# get_strings()
# get_features()
if __name__ == '__main__':
main()