Skip to content

Instantly share code, notes, and snippets.

import scrapy
from scrapy.http import Request
from lxmlhtml import CustomLxmlLinkExtractor as LinkExtractor
from scrapy.conf import settings
from crawling.items import RawResponseItem
from redis_spider import RedisSpider
import boto3
from git import Repo
import json
import os
from docopt import docopt
import shutil
import tarfile
import requests
GIT_API_KEY = os.getenv("GITHUB_API")
Usage:
generate_creds.py init
generate_creds.py show-roles <account>
generate_creds.py generate [--duration=<sec>] [--log-level=<set>] [--account=<name>]
[--access=<level>] [--username=<user>] [--non-interactive]
[--dry-run]
generate_creds.py refresh [--dry-run]
generate_creds.py switch [--account=<name>] [--non-interactive] [--dry-run]
generate_creds.py (-h | --help)
generate_creds.py --version
import datetime
import json
import logging
import os
import socket
import ssl
from urllib2 import Request, urlopen, URLError, HTTPError
import boto3
import boto3
from git import Repo
import os
import sys
from docopt import docopt
import shutil
import tarfile
GIT_API_KEY = os.getenv("GITHUB_API")
BUCKET = 'glomex-github-repo-backup'
# -*- encoding: utf-8 -*-
import os
import boto3
import botocore
import datetime
import difflib
from urllib2 import Request, urlopen, URLError, HTTPError
import json
from __future__ import print_function
from datetime import date
import boto3
import json
import os
bucket_name = os.getenv('BUCKET')
def main():