|  | import argparse | 
					
						
						|  | import csv | 
					
						
						|  | import os | 
					
						
						|  | import json | 
					
						
						|  | import sys | 
					
						
						|  | import time | 
					
						
						|  |  | 
					
						
						|  | import requests | 
					
						
						|  | import requests_ratelimiter | 
					
						
						|  | import tqdm | 
					
						
						|  |  | 
					
						
						|  |  | 
					
						
						|  | def slurp(outfile): | 
					
						
						|  | query = ( | 
					
						
						|  | 'count:all file:\.json$ content:\'{\n  "$schema": "https://json-schema.org/\'' | 
					
						
						|  | ) | 
					
						
						|  |  | 
					
						
						|  | session = requests.Session() | 
					
						
						|  | adapter = requests_ratelimiter.LimiterAdapter(per_second=2) | 
					
						
						|  | session.mount("http://", adapter) | 
					
						
						|  | session.mount("https://", adapter) | 
					
						
						|  | matches = 0 | 
					
						
						|  |  | 
					
						
						|  | with session.get( | 
					
						
						|  | "https://sourcegraph.com/.api/search/stream", | 
					
						
						|  | params={"q": query}, | 
					
						
						|  | headers={ | 
					
						
						|  | "Accept": "text/event-stream", | 
					
						
						|  | "Authorization": "token " + os.environ["SRC_ACCESS_TOKEN"], | 
					
						
						|  | }, | 
					
						
						|  | stream=True, | 
					
						
						|  | ) as resp, open(outfile, "w") as f: | 
					
						
						|  | pbar = tqdm.tqdm() | 
					
						
						|  | writer = csv.writer(f) | 
					
						
						|  | writer.writerow( | 
					
						
						|  | ["repository", "repoStars", "repoLastFetched", "commit", "path"] | 
					
						
						|  | ) | 
					
						
						|  | event = None | 
					
						
						|  | for line in resp.iter_lines(): | 
					
						
						|  | if not line: | 
					
						
						|  | continue | 
					
						
						|  | time.sleep(0.1) | 
					
						
						|  | line = line.decode("utf-8").strip() | 
					
						
						|  |  | 
					
						
						|  | if line.startswith("event:"): | 
					
						
						|  | event = line.split(":", maxsplit=1)[1].strip() | 
					
						
						|  | if event != "matches": | 
					
						
						|  | sys.stderr.write(event + "\n") | 
					
						
						|  | elif line.startswith("data:"): | 
					
						
						|  | data = line.split(":", maxsplit=1)[1].strip() | 
					
						
						|  |  | 
					
						
						|  | if event == "filters": | 
					
						
						|  |  | 
					
						
						|  | continue | 
					
						
						|  | if event == "matches": | 
					
						
						|  | record = [ | 
					
						
						|  | ( | 
					
						
						|  | m["repository"], | 
					
						
						|  | m.get("repoStars", ""), | 
					
						
						|  | m.get("repoLastFetched", ""), | 
					
						
						|  | m["commit"], | 
					
						
						|  | m["path"], | 
					
						
						|  | ) | 
					
						
						|  | for m in json.loads(data) | 
					
						
						|  | ] | 
					
						
						|  | writer.writerows(record) | 
					
						
						|  | matches += len(record) | 
					
						
						|  | pbar.update(len(record)) | 
					
						
						|  | elif event == "progress": | 
					
						
						|  | sys.stderr.write(data + "\n") | 
					
						
						|  |  | 
					
						
						|  |  | 
					
						
						|  | if __name__ == "__main__": | 
					
						
						|  | parser = argparse.ArgumentParser() | 
					
						
						|  | parser.add_argument("--outfile", required=True) | 
					
						
						|  | args = parser.parse_args() | 
					
						
						|  |  | 
					
						
						|  | slurp(args.outfile) | 
					
						
						|  |  |