I hereby claim:
- I am harej on github.
- I am harej (https://keybase.io/harej) on keybase.
- I have a public key ASDTkxXZLirRVO3P-J1MWRftp-qZeodb4Aeq9GZSazULIwo
To claim this, I am signing this object:
| import irc.client | |
| import requests | |
| import re | |
| from credentials import ( | |
| IRC_SERVER, IRC_CHANNEL, IRC_NICKNAME, | |
| PHABRICATOR_API_KEY, PHID | |
| ) | |
| authorized_hostnames = [ | |
| "user/hare", |
I hereby claim:
To claim this, I am signing this object:
| import requests | |
| import json | |
| import sys | |
| from multiprocessing.dummy import Pool as ThreadPool | |
| from wikidataintegrator import wdi_core, wdi_login | |
| from wikidataintegrator.wdi_core import WDItemEngine | |
| mediawiki_api_url = 'https://iagraph.wiki.opencura.com/w/api.php' | |
| sparql_endpoint_url = 'https://iagraph.wiki.opencura.com/query/sparql' |
| import requests | |
| import json | |
| from time import sleep | |
| from pprint import pprint | |
| # This is a quick script I came up with for ingesting "munged" Wikidata TTL dumps | |
| # into Amazon Neptune, one at a time, going as fast as possible while respecting | |
| # queue limits. | |
| for i in range(0, 4243): |
| jh@Scatter-1 ~ % curl -i http://localhost:63342/Cyberbot_II/IABot/www/setup.php\?_ijt=ek248v577c3ch1l8u1c3mq48gb | |
| HTTP/1.1 200 OK | |
| X-Powered-By: PHP/7.2.28 | |
| Set-Cookie: IABotManagementConsole=dbgrl4h62b779fr9luufh5qmm4; expires=Sun, 12-Apr-2020 20:54:31 GMT; Max-Age=2592000; path=/Cyberbot_II/IABot/www | |
| Cache-Control: no-store, must-revalidate | |
| server: PhpStorm 2019.3.3 | |
| content-length: 9486 | |
| set-cookie: Phpstorm-e21bdce2=b191c1a9-572c-4e8a-b862-7d21cf880eae; Max-Age=315360000; Expires=Mon, 11 Mar 2030 20:54:31 GMT; Path=/; HTTPOnly; SameSite=strict | |
| Set-Cookie: IABotManagementConsole=dbgrl4h62b779fr9luufh5qmm4; expires=Sun, 12-Apr-2020 20:54:31 GMT; Max-Age=2592000; path=/Cyberbot_II/IABot/www |
| # Timeout | |
| Processing wikidump-000001399.ttl.gz | |
| SPARQL-UPDATE: updateStr=LOAD <file:///srv/mungeOut//wikidump-000001399.ttl.gz> | |
| java.util.concurrent.TimeoutException | |
| at java.util.concurrent.FutureTask.get(FutureTask.java:205) | |
| at com.bigdata.rdf.sail.webapp.BigdataServlet.submitApiTask(BigdataServlet.java:292) | |
| at com.bigdata.rdf.sail.webapp.QueryServlet.doSparqlUpdate(QueryServlet.java:460) | |
| at com.bigdata.rdf.sail.webapp.QueryServlet.doPost(QueryServlet.java:241) | |
| at com.bigdata.rdf.sail.webapp.RESTServlet.doPost(RESTServlet.java:269) | |
| at com.bigdata.rdf.sail.webapp.MultiTenancyServlet.doPost(MultiTenancyServlet.java:195) |
| import random | |
| import requests | |
| from time import sleep | |
| while True: | |
| random_id = ''.join(random.choices('0123456789ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz-_', k=11)) | |
| url = 'https://www.youtube.com/watch?v=' + random_id | |
| r = requests.get(url) | |
| if r.text.find('This video is unavailable') == -1: | |
| print(url) |
| import html | |
| import requests | |
| import threading | |
| class AskPubMed(threading.Thread): | |
| def __init__ (self, threadID, name, packages): | |
| threading.Thread.__init__(self) | |
| self.threadID = threadID | |
| self.name = name | |
| self.packages = packages |
| import requests | |
| from bs4 import BeautifulSoup | |
| niosh_mode = False | |
| if niosh_mode == True: | |
| seed = "https://query.wikidata.org/sparql?format=json&query=SELECT%20%3Fitem%20WHERE%20%7B%0A%20%20%3Fitem%20wdt%3AP932%20%3Fdummy0%20.%0A%20%20%3Fitem%20wdt%3AP859%20wd%3AQ60346%20.%0A%7D" | |
| else: | |
| seed = "https://query.wikidata.org/sparql?format=json&query=SELECT%20%3Fitem%20WHERE%20%7B%0A%20%20%3Fitem%20wdt%3AP932%20%3Fdummy0%20.%0A%20%20MINUS%20%7B%20%3Fitem%20wdt%3AP859%20wd%3AQ60346%20%7D%0A%7D" |
| import requests | |
| import time | |
| import csv | |
| from bs4 import BeautifulSoup | |
| def main(sourcefile): | |
| url_template = "https://tools.wmflabs.org/sourcemd/?id={0}&doit=Check+source" | |
| with open(sourcefile) as f: | |
| csvdump = csv.reader(f) |