2021-05-20 17:29:27 +02:00
|
|
|
import sys
|
|
|
|
import psycopg2
|
|
|
|
import psycopg2.extras
|
|
|
|
import logging
|
|
|
|
import requests
|
|
|
|
import os
|
|
|
|
import dateutil
|
|
|
|
import chardet
|
|
|
|
from benedict import benedict
|
|
|
|
import traceback
|
|
|
|
|
|
|
|
psycopg2.extensions.register_adapter(dict, psycopg2.extras.Json)
|
|
|
|
|
|
|
|
token = f'Bearer {os.getenv("BUILDKITE_API_TOKEN")}'
|
|
|
|
|
2021-05-20 18:11:55 +02:00
|
|
|
|
2021-05-20 17:29:27 +02:00
|
|
|
def connect():
|
|
|
|
return psycopg2.connect(
|
|
|
|
f"host=127.0.0.1 sslmode=disable dbname=stats user=stats password={os.getenv('DB_PASSWORD')}")
|
|
|
|
|
|
|
|
|
|
|
|
def download_text(url):
|
|
|
|
r = requests.get(url, allow_redirects=True,
|
|
|
|
headers={'Authorization': token})
|
|
|
|
if r.status_code != 200:
|
|
|
|
raise Exception(f'response status {r.status_code}')
|
|
|
|
try:
|
|
|
|
return r.content.decode('utf-8').replace("\x00", "\uFFFD"), 'utf-8'
|
|
|
|
except:
|
|
|
|
pass
|
|
|
|
try:
|
|
|
|
return r.content.decode('ascii').replace("\x00", "\uFFFD"), 'ascii'
|
|
|
|
except:
|
|
|
|
pass
|
|
|
|
d = chardet.detect(r.content)
|
|
|
|
return r.content.decode(d['encoding']).replace("\x00", "\uFFFD"), d['encoding']
|
|
|
|
|
|
|
|
|
|
|
|
def download_job_logs(conn):
|
|
|
|
logging.info('downloading job logs')
|
|
|
|
with conn.cursor() as c:
|
|
|
|
c.execute(f"""
|
|
|
|
select j.id, j.raw->>'raw_log_url' url
|
|
|
|
from jobs j
|
|
|
|
left join artifacts a on a.job_id = j.id and a.id=j.id
|
|
|
|
where a.id IS NULL and j.raw->>'raw_log_url' IS NOT NULL
|
|
|
|
""")
|
|
|
|
total = c.rowcount
|
|
|
|
logging.info(f'will download {total} logs')
|
|
|
|
cnt = 0
|
|
|
|
for row in c:
|
|
|
|
cnt += 1
|
|
|
|
job_id = row[0]
|
|
|
|
url = row[1]
|
|
|
|
meta = {'filename': 'stdout'}
|
|
|
|
try:
|
|
|
|
content, en = download_text(url)
|
|
|
|
meta['encoding'] = en
|
|
|
|
with conn.cursor() as i:
|
2021-05-20 18:11:55 +02:00
|
|
|
i.execute('INSERT INTO artifacts (id, job_id, content, meta) VALUES (%s, %s, %s, %s)',
|
|
|
|
[job_id, job_id, content, meta])
|
2021-05-20 17:29:27 +02:00
|
|
|
except:
|
|
|
|
meta['failure'] = traceback.format_exc()
|
|
|
|
logging.error(f'download artifact failed {meta["failure"]} {url}')
|
|
|
|
with conn.cursor() as i:
|
|
|
|
i.execute('INSERT INTO artifacts (id, job_id, meta) VALUES (%s, %s, %s)', [job_id, job_id, meta])
|
|
|
|
if cnt % 100 == 0:
|
|
|
|
logging.info(f'downloaded {cnt}/{total} logs')
|
|
|
|
conn.commit()
|
|
|
|
logging.info(f'downloaded {cnt} logs')
|
|
|
|
return True
|
|
|
|
|
|
|
|
|
|
|
|
def download_job_artifacts(conn):
|
|
|
|
logging.info('downloading job artifacts')
|
|
|
|
with conn.cursor() as c:
|
|
|
|
c.execute(f"""
|
|
|
|
select ja.meta from
|
|
|
|
(select j.key,j.id job_id, a->>'id' aid, a as meta from jobs j, json_array_elements(j.meta->'artifacts') as a) as ja
|
|
|
|
left join artifacts a on a.job_id = ja.job_id and a.id=ja.aid
|
|
|
|
where a.id IS NULL""")
|
|
|
|
total = c.rowcount
|
|
|
|
logging.info(f'will download {total} artifacts')
|
|
|
|
cnt = 0
|
|
|
|
for row in c:
|
|
|
|
meta = benedict(row[0])
|
|
|
|
cnt += 1
|
|
|
|
try:
|
|
|
|
content, en = download_text(meta.get('download_url'))
|
|
|
|
meta['encoding'] = en
|
|
|
|
with conn.cursor() as i:
|
|
|
|
i.execute('INSERT INTO artifacts (id, job_id, content, meta) VALUES (%s, %s, %s, %s)',
|
|
|
|
[meta.get('id'), meta.get('job_id'), content, meta])
|
|
|
|
except:
|
|
|
|
meta['failure'] = traceback.format_exc()
|
|
|
|
logging.error(f'download artifact failed {meta["failure"]} {meta.get("download_url")}')
|
|
|
|
with conn.cursor() as i:
|
|
|
|
i.execute('INSERT INTO artifacts (id, job_id, meta) VALUES (%s, %s, %s)',
|
|
|
|
[meta.get('id'), meta.get('job_id'), meta])
|
|
|
|
if cnt % 100 == 0:
|
|
|
|
logging.info(f'downloaded {cnt}/{total} artifacts')
|
|
|
|
conn.commit()
|
|
|
|
logging.info(f'downloaded {cnt} artifacts')
|
|
|
|
return True
|
|
|
|
|
|
|
|
|
|
|
|
def insert_new_builds(conn):
|
|
|
|
logging.info('inserting new builds')
|
|
|
|
max_pages = 2
|
|
|
|
while max_pages < 1000:
|
|
|
|
logging.info(f'checking page #{max_pages}')
|
|
|
|
re = requests.get('https://api.buildkite.com/v2/organizations/llvm-project/builds',
|
|
|
|
params={'page': max_pages},
|
|
|
|
headers={'Authorization': token})
|
|
|
|
if re.status_code != 200:
|
|
|
|
logging.error(f'list builds response status: {re.status_code}')
|
|
|
|
sys.exit(1)
|
|
|
|
x = re.json()
|
|
|
|
existing = 0
|
|
|
|
new = 0
|
|
|
|
for b in x:
|
|
|
|
if (b['state'] == 'running') or (b['state'] == 'scheduled'):
|
|
|
|
new += 1
|
|
|
|
continue
|
|
|
|
with conn.cursor() as c:
|
|
|
|
c.execute('SELECT count(1) FROM builds WHERE id = %s', (b.get('id'),))
|
|
|
|
if c.fetchone()[0] == 0:
|
|
|
|
new += 1
|
|
|
|
else:
|
|
|
|
existing += 1
|
|
|
|
logging.info(f'new {new} existing {existing}')
|
|
|
|
if new == 0:
|
|
|
|
break
|
|
|
|
max_pages += 10
|
|
|
|
max_pages += 5
|
|
|
|
logging.info(f'will load {max_pages} pages')
|
|
|
|
page = 1
|
|
|
|
all_builds = []
|
|
|
|
# Read #max_pages first in order to not miss any builds that are moved due to new inserts.
|
|
|
|
while page <= max_pages:
|
|
|
|
logging.info(f'loading page {page}')
|
|
|
|
re = requests.get('https://api.buildkite.com/v2/organizations/llvm-project/builds',
|
|
|
|
params={'page': page},
|
|
|
|
headers={'Authorization': token})
|
|
|
|
if re.status_code != 200:
|
|
|
|
print('response status', re.status_code, re)
|
|
|
|
break
|
|
|
|
x = re.json()
|
|
|
|
if x == []:
|
|
|
|
logging.warning('empty response')
|
|
|
|
break
|
|
|
|
all_builds.extend(x)
|
|
|
|
page += 1
|
|
|
|
# Now insert new builds in reverse order so that we can resume correctly if operation has failed.
|
|
|
|
all_builds.reverse()
|
|
|
|
logging.info(f'{len(all_builds)} builds loaded')
|
|
|
|
cnt = 0
|
|
|
|
for b in all_builds:
|
|
|
|
if (b['state'] == 'running') or (b['state'] == 'scheduled'):
|
|
|
|
continue
|
|
|
|
with conn.cursor() as c:
|
|
|
|
c.execute('SELECT count(1) FROM builds WHERE id = %s', (b.get('id'),))
|
|
|
|
if c.fetchone()[0] == 0:
|
|
|
|
c.execute('INSERT INTO builds (id, raw) VALUES (%s, %s)', [b.get('id'), psycopg2.extras.Json(b)])
|
|
|
|
cnt += 1
|
|
|
|
if cnt % 100 == 0:
|
|
|
|
logging.info(f'{cnt} builds inserted')
|
|
|
|
conn.commit()
|
|
|
|
conn.commit()
|
|
|
|
logging.info(f'{cnt} builds inserted')
|
|
|
|
return cnt
|
|
|
|
|
|
|
|
|
|
|
|
def download_job_artifacts_list(conn):
|
|
|
|
logging.info('download jobs artifact lsits')
|
|
|
|
with conn.cursor() as c:
|
|
|
|
c.execute("""
|
|
|
|
SELECT key, raw->>'artifacts_url', meta
|
|
|
|
FROM jobs
|
|
|
|
WHERE (meta->>'artifacts' IS NULL) AND (raw->>'artifacts_url' IS NOT NULL)""")
|
|
|
|
cnt = 0
|
|
|
|
total = c.rowcount
|
|
|
|
logging.info(f'will download {total} artifact lists')
|
|
|
|
for row in c:
|
|
|
|
key = row[0]
|
|
|
|
url = row[1]
|
|
|
|
meta = row[2]
|
|
|
|
if meta is None:
|
|
|
|
meta = {}
|
|
|
|
r = requests.get(url, allow_redirects=True, headers={'Authorization': token})
|
|
|
|
if r.status_code != 200:
|
|
|
|
logging.error(f'cannot load artifacts_url {r.status_code} {url}')
|
|
|
|
continue
|
|
|
|
meta['artifacts'] = r.json()
|
|
|
|
with conn.cursor() as i:
|
|
|
|
i.execute('UPDATE jobs SET meta = %s WHERE key = %s', (meta, key))
|
|
|
|
cnt += 1
|
|
|
|
if cnt % 100 == 0:
|
|
|
|
logging.info(f'downloaded {cnt}/{total} artifact lists')
|
|
|
|
conn.commit()
|
|
|
|
logging.info(f'downloaded {cnt} artifact lists')
|
|
|
|
conn.commit()
|
|
|
|
|
|
|
|
|
|
|
|
def insert_new_jobs(conn):
|
|
|
|
logging.info('inserting new jobs')
|
|
|
|
with conn.cursor() as c:
|
|
|
|
c.execute("""select bj.id, bj.jid, bj.job from
|
|
|
|
(select b.id, j->>'id' jid, j as job from builds b, json_array_elements(b.raw->'jobs') as j) as bj
|
|
|
|
left join jobs j on j.id = bj.jid
|
|
|
|
where j.id IS NULL""")
|
|
|
|
total = c.rowcount
|
|
|
|
cnt = 0
|
|
|
|
logging.info(f'will insert {total} jobs')
|
|
|
|
for row in c:
|
|
|
|
build_id = row[0]
|
|
|
|
job_id = row[1]
|
|
|
|
job = benedict(row[2])
|
|
|
|
meta = {}
|
|
|
|
# durations
|
|
|
|
runnable_at = job.get('runnable_at')
|
|
|
|
started_at = job.get('started_at')
|
|
|
|
finished_at = job.get('finished_at')
|
|
|
|
if (runnable_at is not None) and (started_at is not None) and (finished_at is not None):
|
|
|
|
runnable_at = dateutil.parser.parse(runnable_at)
|
|
|
|
started_at = dateutil.parser.parse(started_at)
|
|
|
|
finished_at = dateutil.parser.parse(finished_at)
|
|
|
|
meta['queue_time'] = (started_at - runnable_at).total_seconds()
|
|
|
|
meta['run_time'] = (finished_at - started_at).total_seconds()
|
|
|
|
meta['total_time'] = (finished_at - runnable_at).total_seconds()
|
|
|
|
# agent data
|
|
|
|
for e in job.get('agent.meta_data', []):
|
|
|
|
p = e.split('=')
|
|
|
|
if p[0] == 'queue':
|
|
|
|
meta['agent_queue'] = p[1]
|
|
|
|
if p[0] == 'name':
|
|
|
|
meta['agent_name'] = p[1]
|
|
|
|
with conn.cursor() as i:
|
|
|
|
i.execute('INSERT INTO jobs (id, build_id, raw, meta) VALUES (%s, %s, %s, %s)',
|
|
|
|
[job_id, build_id, job, meta])
|
|
|
|
cnt += 1
|
|
|
|
if cnt % 100 == 0:
|
|
|
|
logging.info(f'inserted {cnt}/{total} jobs')
|
|
|
|
conn.commit()
|
|
|
|
logging.info(f'inserted {cnt} jobs')
|
|
|
|
conn.commit()
|
|
|
|
|
|
|
|
|
|
|
|
if __name__ == '__main__':
|
|
|
|
logging.basicConfig(level='INFO', format='%(levelname)-7s %(message)s')
|
|
|
|
cn = connect()
|
|
|
|
logging.info('downloading buildkite data')
|
|
|
|
insert_new_builds(cn)
|
|
|
|
insert_new_jobs(cn)
|
|
|
|
download_job_artifacts_list(cn)
|
|
|
|
download_job_artifacts(cn)
|
|
|
|
download_job_logs(cn)
|