import base64 import json import re import socket import requests from lib import Client from patch_makers import (AnalyticsPatchMaker, CxPatchMaker, DnsPatchMaker, WikimediaMessagesPatchMaker) final_text = '' gerrit_path = 'https://gerrit.wikimedia.org/g/' client = Client.newFromCreds() def get_checklist_text(url, text, checked): if checked: return ' [x] [[{}|{}]]'.format(url, text) else: return ' [] [[{}|{}]]'.format(url, text) def get_file_from_gerrit(path): gerrit_url = 'https://gerrit.wikimedia.org/g/' url = gerrit_url + '{0}?format=TEXT'.format(path) r = requests.get(url) if r.status_code == 200: return base64.b64decode(r.text).decode('utf-8') else: return '' def get_gerrit_path(repo, filename): return repo + '/+/master/' + filename def get_github_url(repo, filename): return 'https://raw.githubusercontent.com/wikimedia/{}/master/{}'.format( repo, filename ) class PostCreationHandler(object): def __init__(self, phid, db_name, url, language_code, parts): self.main_pid = phid self.db_name = db_name self.url = url self.parts = parts self.language_code = language_code self.post_ticket_bug_id = '' self.post_ticket_text = '' self.checkers = [ self._check_restbase, self._check_cx, self._check_analytics, self._check_pywikibot, self._check_wikidata, ] self.handlers = [ self._handle_restbase, self._handle_cx, self._handle_analytics, self._handle_pywikibot, self._handle_wikidata, self._handle_wikistats, ] self.handlers_needed = {} def handle(self): for checker in self.checkers: checker() self.add_text(' [] Import from Incubator') self.add_text(' [] Clean up old interwiki links') self.add_text(' [] Propose the implementation of the standard bot policy') self.add_text(' [] Inform the [[ https://meta.wikimedia.org/wiki/Talk:Countervandalism_Network | CVN ]] project for IRC monitoring') self._create_ticket() for handler in self.handlers: handler() def add_text(self, a): self.post_ticket_text += a + '\n' def add_checklist(self, url, text, checked): self.add_text(get_checklist_text(url, text, checked)) def _create_ticket(self): result = client.createParentTask( self.post_ticket_text, [ 'PHID-PROJ-2fuv7mxzjnpjfuojdnfd', # wiki-setup 'PHID-PROJ-2b7oz62ylk3jk4aus262', # platform-engineering ], self.main_pid, 'Post-creation work for {}'.format(self.db_name))['object'] self.post_ticket_phid = result['phid'] self.post_ticket_bug_id = 'T' + str(result['id']) def _check_restbase(self): path = get_gerrit_path( 'mediawiki/services/restbase/deploy', 'scap/vars.yaml' ) restbase = get_file_from_gerrit(path) self.add_checklist(gerrit_path + path, 'RESTbase', self.url in restbase) self.handlers_needed['restbase'] = self.url not in restbase def _handle_restbase(self): if not self.handlers_needed['restbase']: return client.createSubtask( 'Per https://wikitech.wikimedia.org/wiki/Add_a_wiki once the wiki has been created', ['PHID-PROJ-mszihytuo3ij3fcxcxgm'], self.post_ticket_phid, 'Add {} to RESTBase'.format(self.db_name)) def _check_cx(self): path = get_gerrit_path( 'mediawiki/services/cxserver', 'config/languages.yaml' ) cxconfig = get_file_from_gerrit(path) cx = '\n- ' + self.language_code in cxconfig self.add_checklist(gerrit_path + path, 'CX Config', cx) self.handlers_needed['cx'] = not cx def _handle_cx(self): if not self.handlers_needed['cx']: return r = requests.get( 'https://gerrit.wikimedia.org/r/changes/' '?q=bug:{}+project:mediawiki/services/cxserver'.format(self.post_ticket_bug_id)) b = json.loads('\n'.join(r.text.split('\n')[1:])) if b: return maker = CxPatchMaker(self.language_code, self.post_ticket_bug_id) maker.run() def _check_analytics(self): path = get_gerrit_path( 'analytics/refinery', 'static_data/pageview/whitelist/whitelist.tsv' ) url = '.'.join(self.parts[:2]) refinery_whitelist = get_file_from_gerrit(path) self.add_checklist(gerrit_path + path, 'Analytics refinery', url in refinery_whitelist) self.handlers_needed['analytics'] = url not in refinery_whitelist def _handle_analytics(self): if not self.handlers_needed['analytics']: return url = '.'.join(self.parts[:2]) r = requests.get( 'https://gerrit.wikimedia.org/r/changes/' '?q=bug:{}+project:analytics/refinery'.format(self.post_ticket_bug_id)) b = json.loads('\n'.join(r.text.split('\n')[1:])) if b: return maker = AnalyticsPatchMaker(url, self.post_ticket_bug_id) maker.run() def _check_pywikibot(self): path = get_gerrit_path( 'pywikibot/core', 'pywikibot/families/{}_family.py'.format(self.parts[1]) ) pywikibot = get_file_from_gerrit(path) self.add_checklist(gerrit_path + path, 'Pywikibot', "'{}'".format(self.language_code) in pywikibot) self.handlers_needed['pywikibot'] = "'{}'".format(self.language_code) not in pywikibot def _handle_pywikibot(self): if not self.handlers_needed['pywikibot']: return client.createSubtask( 'Per https://wikitech.wikimedia.org/wiki/Add_a_wiki once the wiki has been created', ['PHID-PROJ-orw42whe2lepxc7gghdq'], self.post_ticket_phid, 'Add support for {} to Pywikibot'.format(self.db_name)) def _check_wikidata(self): url = 'https://www.wikidata.org/w/api.php' wikidata_help_page = requests.get(url, params={ 'action': 'help', 'modules': 'wbgetentities' }).text self.add_checklist(url, 'Wikidata', self.db_name in wikidata_help_page) def _handle_wikidata(self): client.createSubtask( 'Per https://wikitech.wikimedia.org/wiki/Add_a_wiki once the wiki has been created', ['PHID-PROJ-egbmgxclscgwu2rbnotm', 'PHID-PROJ-7ocjej2gottz7cikkdc6'], self.post_ticket_phid, 'Add Wikidata support for {}'.format(self.db_name)) def _handle_wikistats(self): client.createSubtask("Please add new wiki `%s` to Wikistats, once it is created. Thanks!" % self.db_name, [ 'PHID-PROJ-6sht6g4xpdii4c4bga2i' # VPS-project-Wikistats ], self.post_ticket_phid, 'Add %s to wikistats' % self.db_name) def add_text(a): global final_text final_text += a + '\n' def add_checklist(url, text, checked): add_text(get_checklist_text(url, text, checked)) def hostname_resolves(hostname): try: socket.gethostbyname(hostname) except socket.error: return False return True def handle_special_wiki_apache(parts): file_path = 'hieradata/common/mediawiki.yaml' apache_file = get_file_from_gerrit( 'operations/puppet/+/production/' + file_path) url = '.'.join(parts) return url in apache_file def post_a_comment(comment): comment = 'Hello, I am helping on creating this wiki. ' + comment + \ ' ^_^ Sincerely, your Fully Automated Resource Tackler' pass def handle_subticket_for_cloud(task_details, db_name, wiki_status): hasSubtasks = client.getTaskSubtasks(task_details['phid']) if hasSubtasks: return client.createSubtask("The new wiki's visibility will be: **%s**." % wiki_status, [ 'PHID-PROJ-hwibeuyzizzy4xzunfsk', # DBA 'PHID-PROJ-bj6y6ks7ampcwcignhce' # Data services ], task_details['phid'], 'Prepare and check storage layer for ' + db_name) def handle_ticket_for_wikistats(task_details, db_name): client.createParentTask("Please add new wiki `%s` to Wikistats, once it is created. Thanks!" % db_name, [ 'PHID-PROJ-6sht6g4xpdii4c4bga2i' # VPS-project-Wikistats ], task_details['phid'], 'Add %s to wikistats' % db_name) def get_dummy_wiki(shard, family): if family == "wiktionary": return { "s3": "aawiki", "s5": "mhwiktionary", }.get(shard, "?????") else: return { "s3": "aawiki", "s5": "muswiki" }.get(shard, "?????") def create_patch_for_wikimedia_messages( db_name, english_name, url, lang, bug_id): if not english_name: return r = requests.get( 'https://gerrit.wikimedia.org/r/changes/?q=' 'bug:{}+project:mediawiki/extensions/WikimediaMessages'.format(bug_id)) b = json.loads('\n'.join(r.text.split('\n')[1:])) if b: return maker = WikimediaMessagesPatchMaker( db_name, english_name, url, lang, bug_id) maker.run() def handle_dns(special, url, language_code, task_tid): dns_path = get_gerrit_path( 'operations/dns', 'templates/wikimedia.org' if special else 'templates/helpers/langlist.tmpl') dns_url = gerrit_path + dns_path dns = hostname_resolves(url) print(url) if not dns: print('dns not found') if not special: print('not special') create_patch_for_dns(language_code, task_tid) add_checklist(dns_url, 'DNS', dns) return dns def handle_apache(special, parts): if not special: add_text(' [x] Apache config (Not needed)') return True file_path = 'hieradata/common/mediawiki.yaml' apache_url = gerrit_path + \ 'operations/puppet/+/production/' + file_path if not handle_special_wiki_apache(parts): apache = False else: apache = True add_checklist(apache_url, 'Apache config', apache) return apache def handle_langdb(language_code): langdb_url = get_github_url('language-data', 'data/langdb.yaml') r = requests.get(langdb_url) config = 'Language configuration in language data repo' if re.search(r'\n *?' + language_code + ':', r.text): langdb = True else: langdb = False add_checklist(langdb_url, config, langdb) return langdb def handle_wikimedia_messages_one( db_name, wiki_spec, url, language_code, task_tid): path = get_gerrit_path( 'mediawiki/extensions/WikimediaMessages', 'i18n/wikimediaprojectnames/en.json' ) wikimedia_messages_data = get_file_from_gerrit(path) wikimedia_messages_data = json.loads(wikimedia_messages_data) if 'project-localized-name-' + db_name in wikimedia_messages_data: wikimedia_messages_one = True else: wikimedia_messages_one = False english_name = wiki_spec.get('Project name (English)') create_patch_for_wikimedia_messages( db_name, english_name, url, language_code, task_tid) add_checklist(gerrit_path + path, 'Wikimedia messages configuration', wikimedia_messages_one) url = 'https://en.wikipedia.org/wiki/' + \ 'MediaWiki:Project-localized-name-' + db_name r = requests.get(url) if 'Wikipedia does not have a' not in r.text: wikimedia_messages_one_deployed = True add_text(' [x] [[{}|deployed]]'.format(url)) else: wikimedia_messages_one_deployed = False add_text(' [] [[{}|deployed]]'.format(url)) return wikimedia_messages_one and wikimedia_messages_one_deployed def handle_wikimedia_messages_two(db_name, parts): config = 'Wikimedia messages (interwiki search result) configuration' if parts[1] != 'wikipedia': add_text(' [x] {} (not needed)'.format(config)) return True path = get_gerrit_path( 'mediawiki/extensions/WikimediaMessages', 'i18n/wikimediainterwikisearchresults/en.json' ) search_messages_data = json.loads(get_file_from_gerrit(path)) if 'search-interwiki-results-' + db_name in search_messages_data: wikimedia_messages_two = True else: wikimedia_messages_two = False add_checklist( gerrit_path + path, config, wikimedia_messages_two) url = 'https://en.wikipedia.org/wiki/' + \ 'MediaWiki:Search-interwiki-results-' + db_name r = requests.get(url) if 'Wikipedia does not have a' not in r.text: wikimedia_messages_two_deployed = True add_text(' [x] [[{}|deployed]]'.format(url)) else: wikimedia_messages_two_deployed = False add_text(' [] [[{}|deployed]]'.format(url)) return wikimedia_messages_two and wikimedia_messages_two_deployed def create_patch_for_dns(lang, bug_id): r = requests.get( 'https://gerrit.wikimedia.org/r/changes/' '?q=bug:{}+project:operations/dns'.format(bug_id)) b = json.loads('\n'.join(r.text.split('\n')[1:])) if b: return maker = DnsPatchMaker(lang, bug_id) maker.run() def handle_core_lang(language_code): core_messages_url = get_github_url( 'mediawiki', 'languages/messages/Messages{}.php'.format( language_code[0].upper() + language_code[1:])) r = requests.get(core_messages_url) if r.status_code == 200: core_lang = True else: core_lang = False add_checklist(core_messages_url, 'Language configuration in mediawiki core', core_lang) return core_lang def get_db_name(wiki_spec, parts): db_name = wiki_spec.get('Database name') if not db_name: if parts[1] == 'wikipedia': db_name = parts[0].replace('-', '_') + 'wiki' else: db_name = parts[0].replace('-', '_') + parts[1] return db_name def sync_file(path, summary): return '`scap sync-file {} "{}"`'.format(path, summary) def add_create_instructions(parts, shard, language_code, db_name, task_tid): add_text('\n-------') add_text('**Step by step commands**:') dummy_wiki = get_dummy_wiki(shard, parts[1]) add_text('On deployment host:') add_text('`cd /srv/mediawiki-staging/`') add_text('`git fetch`') add_text('`git log -p HEAD..@{u}`') add_text('`git rebase`') add_text('On mwmaint1002:') add_text('`scap pull`') addwiki_path = 'mwscript extensions/WikimediaMaintenance/addWiki.php' add_text( '`{addwiki_path} --wiki={dummy} {lang} {family} {db} {url}`'.format( addwiki_path=addwiki_path, dummy=dummy_wiki, lang=language_code, family=parts[1], db=db_name, url='.'.join(parts))) summary = 'Creating {db_name} ({phab})'.format( db_name=db_name, phab=task_tid) add_text('On deployment host:') if shard != "s3": add_text(sync_file('wmf-config/db-production.php', summary)) add_text(sync_file('dblists', summary)) add_text('`scap sync-wikiversions "{}"`'.format(summary)) if parts[1] == 'wikimedia': add_text(sync_file('multiversion/MWMultiVersion.php', summary)) add_text(sync_file('static/images/project-logos/', summary)) add_text(sync_file('wmf-config/logos.php', summary)) add_text(sync_file('wmf-config/InitialiseSettings.php', summary)) if parts[1] != 'wikimedia': add_text(sync_file('langlist', summary)) add_text('On mwmaint1002:') add_text('`{search_path} --wiki={dbname} --cluster=all`'.format( search_path='mwscript extensions/CirrusSearch/maintenance/UpdateSearchIndexConfig.php', dbname=db_name, )) add_text('On deployment host:') add_text('`scap update-interwiki-cache`') def update_task_report(task_details): global final_text if not final_text: return old_report = re.findall( r'(\n\n------\n\*\*Pre-install automatic checklist:' r'\*\*.+?\n\*\*End of automatic output\*\*\n)', task_details['description'], re.DOTALL) if not old_report: print('old report not found, appending') client.setTaskDescription( task_details['phid'], task_details['description'] + final_text) else: if old_report[0] != final_text: print('Updating old report') client.setTaskDescription( task_details['phid'], task_details['description'].replace( old_report[0], final_text)) def hande_task(task_details): global final_text final_text = '' print('Checking T%s' % task_details['id']) task_tid = 'T' + task_details['id'] # Extract wiki config wiki_spec = {} for case in re.findall( r'\n- *?\*\*(.+?):\*\* *?(.+)', task_details['description']): wiki_spec[case[0].strip()] = case[1].strip() language_code = wiki_spec.get('Language code') if not language_code: print('lang code not found, skipping') return url = wiki_spec.get('Site URL') if not url: print('url not found, skipping') return parts = url.split('.') if len(parts) != 3 or parts[2] != 'org': print('the url looks weird, skipping') return db_name = get_db_name(wiki_spec, parts) shard = wiki_spec.get('Shard', 'TBD') visibility = wiki_spec.get('Visibility', 'unknown') shardDecided = shard != "TBD" special = parts[1] == 'wikimedia' add_text('\n\n------\n**Pre-install automatic checklist:**') if shardDecided: add_text(' [X] #DBA decided about the shard') else: add_text(' [] #DBA decided about the shard') dns = handle_dns(special, url, language_code, task_tid) if not special and wiki_spec.get('Special', '').lower() != 'yes': handle_subticket_for_cloud(task_details, db_name, visibility) apache = handle_apache(special, parts) langdb = handle_langdb(language_code) core_lang = handle_core_lang(language_code) wm_message_one = handle_wikimedia_messages_one( db_name, wiki_spec, url, language_code, task_tid ) wm_message_two = handle_wikimedia_messages_two(db_name, parts) if dns and apache and langdb and core_lang and wm_message_one and \ wm_message_two and shardDecided: add_text('**The Wiki is ready to be created.**') else: add_text('**The creation is blocked until these part are all done.**') if visibility.lower() != 'private' and not client.getTaskParents(task_details['phid']): handler = PostCreationHandler(task_details['phid'], db_name, url, language_code, parts) handler.handle() add_create_instructions(parts, shard, language_code, db_name, task_tid) add_text('\n**End of automatic output**') def main(): open_create_wikis_phid = 'PHID-PROJ-kmpu7gznmc2edea3qn2x' for phid in client.getTasksWithProject( open_create_wikis_phid, statuses=['open']): task_details = client.taskDetails(phid) hande_task(task_details) update_task_report(task_details) if __name__ == "__main__": main()