diff --git a/scripts/basejump b/scripts/basejump deleted file mode 100755 index 43fd062..0000000 --- a/scripts/basejump +++ /dev/null @@ -1,37 +0,0 @@ -#! /usr/bin/env python -import argparse -import sys -import logging -import subprocess - - -logger = logging.getLogger(__name__) - - -def main() -> int: - # Handle program arguments - ap = argparse.ArgumentParser(prog="basejump") - ap.add_argument( - "subcommand", help="The subcommand to run", choices=["init", "fetch", "discover"] - ) - ap.add_argument("arguments", nargs=argparse.REMAINDER) - args = ap.parse_args() - - # Configure logging - logging.basicConfig( - level=logging.INFO, - format="%(levelname)s: %(message)s", - ) - - # Execute the appropriate subcommand - real_command_name = f"basejump-{args.subcommand}" - try: - return subprocess.run([real_command_name] + args.arguments).returncode - except FileNotFoundError: - logger.error(f"Unknown subcommand: {args.subcommand}") - logger.error(f"Could not find `{real_command_name}` in $PATH") - return 1 - - -if __name__ == "__main__": - sys.exit(main()) diff --git a/scripts/basejump-discover b/scripts/basejump-discover deleted file mode 100755 index e7cfb19..0000000 --- a/scripts/basejump-discover +++ /dev/null @@ -1,59 +0,0 @@ -#! /usr/bin/env python -import argparse -import sys -import logging -import subprocess -import json -from pprint import pprint -from pathlib import Path - -logger = logging.getLogger(__name__) - - -def main() -> int: - # Handle program arguments - ap = argparse.ArgumentParser( - prog="basejump discover", description="Discover repos in a codebase" - ) - ap.add_argument("root_path", help="The root path of the codebase", type=Path) - ap.add_argument( - "-v", "--verbose", help="Enable verbose logging", action="store_true" - ) - args = ap.parse_args() - - # Configure logging - logging.basicConfig( - level=logging.DEBUG if args.verbose else logging.INFO, - format="%(levelname)s: %(message)s", - ) - - # Find all git repos in the codebase - logger.info(f"Searching for git repos in: {args.root_path}") - repos = [] - for path in args.root_path.rglob(".git"): - repos.append({"path":str(path.parent.absolute())}) - - # For each repo, find the upstream - logger.info("Finding upstream URLs...") - for repo in repos: - # Get the upstream URL - upstream_url = subprocess.run( - ["git", "remote", "get-url", "origin"], - cwd=repo["path"], - stdout=subprocess.PIPE, - stderr=subprocess.PIPE, - encoding="utf-8", - ).stdout.strip() - - # Add the upstream URL to the repo config - repo["upstream"] = upstream_url - - # Print the results - logger.info("Found the following repos:") - print(json.dumps(repos, indent=4)) - - return 0 - - -if __name__ == "__main__": - sys.exit(main()) diff --git a/scripts/basejump-fetch b/scripts/basejump-fetch deleted file mode 100755 index ce050b1..0000000 --- a/scripts/basejump-fetch +++ /dev/null @@ -1,58 +0,0 @@ -#! /usr/bin/env python -import argparse -import sys -import logging -import json -import subprocess -import os -from pathlib import Path - -logger = logging.getLogger(__name__) - - -def main() -> int: - # Handle program arguments - ap = argparse.ArgumentParser( - prog="basejump fetch", description="Fetches all changes for a whole codebase" - ) - ap.add_argument("name", help="The name of the codebase") - ap.add_argument("--pull", help="Perform a full pull", action="store_true") - ap.add_argument( - "-v", "--verbose", help="Enable verbose logging", action="store_true" - ) - args = ap.parse_args() - - # Configure logging - logging.basicConfig( - level=logging.DEBUG if args.verbose else logging.INFO, - format="%(levelname)s: %(message)s", - ) - - # Ensure that the basejump config dir exists - bj_config_dir = Path.home() / ".config" / "basejump" - bj_config_dir.mkdir(parents=True, exist_ok=True) - - # Read the codebase config - codebase_config_path = bj_config_dir / f"{args.name}.codebase.json" - if not codebase_config_path.exists(): - logger.error(f"Codebase `{args.name}` does not exist") - return 1 - config = json.loads(codebase_config_path.read_text()) - - # Handle each repository - for repo in config["repos"]: - logger.info(f"Fetching {repo['path']}") - - # If we are in pull mode, do a git pull - if args.pull: - subprocess.run(["git", "pull"], cwd=repo["path"]) - - # Otherwise fetch all - else: - subprocess.run(["git", "fetch", "--all"], cwd=repo["path"]) - - return 0 - - -if __name__ == "__main__": - sys.exit(main()) diff --git a/scripts/basejump-init b/scripts/basejump-init deleted file mode 100755 index 0013ea7..0000000 --- a/scripts/basejump-init +++ /dev/null @@ -1,75 +0,0 @@ -#! /usr/bin/env python -import argparse -import sys -import logging -import json -import subprocess -import os -from pathlib import Path - -logger = logging.getLogger(__name__) - - -def main() -> int: - # Handle program arguments - ap = argparse.ArgumentParser( - prog="basejump init", description="Creates a new basejump codebase" - ) - ap.add_argument("name", help="The name of the codebase") - ap.add_argument( - "-v", "--verbose", help="Enable verbose logging", action="store_true" - ) - args = ap.parse_args() - - # Configure logging - logging.basicConfig( - level=logging.DEBUG if args.verbose else logging.INFO, - format="%(levelname)s: %(message)s", - ) - - # Ensure that the basejump config dir exists - bj_config_dir = Path.home() / ".config" / "basejump" - bj_config_dir.mkdir(parents=True, exist_ok=True) - - # Create a new codebase definition - codebase_config_path = bj_config_dir / f"{args.name}.codebase.json" - - # If the path already exists, abort - if codebase_config_path.exists(): - logger.error(f"Codebase `{args.name}` already exists") - logger.info(f"Config file at: {codebase_config_path}") - return 1 - - # Create a template codebase config - template_config = { - "name": args.name, - "repos": [ - { - "path": "/tmp/example", - "upstream": "https://github.com/octocat/Hello-World", - } - ], - } - - # Write the template config to disk - codebase_config_path.write_text(json.dumps(template_config, indent=4)) - - # Open $EDITOR (or vim) to edit the config - subprocess.run([os.environ.get("EDITOR", "vim"), str(codebase_config_path)]) - - # Iterate through every repo and clone it - config = json.loads(codebase_config_path.read_text()) - for repo in config["repos"]: - if Path(repo["path"]).exists(): - logger.info(f"Skipping {repo['path']}, already exists") - continue - - # Do a clone - logger.info(f"Cloning {repo['upstream']} into {repo['path']}") - subprocess.run(["git", "clone", repo["upstream"], repo["path"]]) - - return 0 - - -if __name__ == "__main__": - sys.exit(main()) diff --git a/scripts/gp-upload b/scripts/gp-upload deleted file mode 100755 index 7cd8e8d..0000000 --- a/scripts/gp-upload +++ /dev/null @@ -1,217 +0,0 @@ -#! /usr/bin/env python3 -import argparse -import sys -import logging -import requests -import socket -import urllib.parse -from pathlib import Path - -logger = logging.getLogger(__name__) - -G_CLIENT_ID = "107923498573-ruh1uhkfe1t5f18vam6sckq7pqer1vmg.apps.googleusercontent.com" -G_SCOPES = ["https://www.googleapis.com/auth/photoslibrary.appendonly"] -G_REDIRECT_URI = "http://localhost:7842" - - -def get_google_oauth_token() -> str: - """Either log the user in, or used a stored refresh token to get an OAuth token""" - refresh_token_path = Path("~/.config/gp-upload/refresh-token").expanduser() - client_secret_path = Path("~/.config/gp-upload/client-secret").expanduser() - - # Read the client secret - with client_secret_path.open("r") as f: - client_secret = f.read().strip() - - # Check if we have a refresh token - if refresh_token_path.exists(): - logger.info("Using stored refresh token") - - # Read the refresh token - with refresh_token_path.open("r") as f: - refresh_token = f.read().strip() - - # Make the request - response = requests.post( - "https://oauth2.googleapis.com/token", - data={ - "client_id": G_CLIENT_ID, - "grant_type": "refresh_token", - "refresh_token": refresh_token, - "client_secret": client_secret, - }, - ) - - # Check for errors - if response.status_code != 200: - logger.error("Failed to get OAuth token") - logger.error(response.text) - return None - - # Return the OAuth token - return response.json()["access_token"] - - # Otherwise, log the user in - else: - logger.info("Logging user in") - - # Direct the user to Google's login page - logger.info("Please visit the following URL to log in:") - logger.info( - f"https://accounts.google.com/o/oauth2/v2/auth?client_id={G_CLIENT_ID}&response_type=code&scope={'+'.join(G_SCOPES)}&redirect_uri={G_REDIRECT_URI}&access_type=offline&prompt=consent" - ) - - # Open a TCP server to listen for the redirect - with socket.socket(socket.AF_INET, socket.SOCK_STREAM) as s: - s.bind(("localhost", 7842)) - s.listen() - - # Wait for the redirect - conn, addr = s.accept() - with conn: - # Read the request - request = conn.recv(1024).decode("utf-8") - - # Parse the request - request = request.splitlines() - request = [line for line in request if line.startswith("GET")] - request = request[0].split(" ")[1] - request = request.split("?")[1] - request = request.split("&") - request = {key: urllib.parse.unquote(value) for key, value in [pair.split("=") for pair in request]} - - # Check for errors - if "error" in request: - logger.error(f"Failed to log in: {request['error']}") - conn.sendall(b"HTTP/1.1 500 Internal Server Error\n\n