- Download the docset index XML.
- Find the docset you want (there are some with URL https://apple.com/none.dmg; ignore them - you will find them again further down the file with a working URL).
- Download the dmg. It's probably around a gigabyte or so.
- "Install" the .pkg file somewhere on your disk. If you don't trust the installer, do it manually:
- Find the largest file, named Payload, and extract it using The Unarchiver.
- This creates a new, even larger file, probably named Payload-1.
- Extract Payload-1 using The Unarchiver.
- After many minutes of extracting, we have our .docset file.
// version 2.2 | |
// Updated jQuery to stable head | |
// removed padding changes on hover, too messy | |
// avoid Tweet This, ADN and Facebook Share links | |
// version 2.1 | |
// | |
// Fixed jQuery injection | |
// version 2.0 |
javascript:(function() { | |
function copyToClipboard(text) { | |
if (window.clipboardData && window.clipboardData.setData) { | |
/*IE specific code path to prevent textarea being shown while dialog is visible.*/ | |
return clipboardData.setData("Text", text); | |
} else if (document.queryCommandSupported && document.queryCommandSupported("copy")) { | |
var textarea = document.createElement("textarea"); | |
textarea.textContent = text; |
if ((/^localhost/).test(document.location.host)) { | |
var linkChecker = (function() { | |
"use strict"; | |
let links, errors = [], completed = [], errorCounter = 0; | |
function addStyle(styleString) { | |
const style = document.createElement("style"); | |
style.textContent = styleString; | |
document.head.append(style); | |
} |
If multiple input files are given, pandoc will concatenate them all (with blank lines between them) before parsing. -- from Pandoc website
Pandoc command:
pandoc -s input1.md input2.md input3.md -o output.html
// Blog post: https://rolandleth.com/extracting-and-parsing-tweets-from-your-twitter-archive | |
import Foundation | |
struct Tweet: Codable { | |
let text: String | |
let timestamp: String | |
} |
I'll start off with letting you know this is a fork from someone else. However, for some bizarre reason, this is the one everyone finds, so I better get round to updating this. Credit to Cristiano#2233 for the original idea.
Also, I've had a lot of people saying the rules are to strict. If you pick all the rules here, you're right, it would be very strict. However the rules below are guidelines! They are there for you to pick the ones you desire, you can ignore ones you don't want. Hopefully they might help with rules you wouldn't have thought of otherwise.
package main | |
import ( | |
"encoding/json" | |
"fmt" | |
"log" | |
"net/http" | |
"net/url" | |
"os/exec" | |
"strconv" |
\documentclass[sigconf,anonymous=$anonymous$]{acmart} | |
\usepackage{booktabs} | |
\usepackage{caption} % http://mirror.easyname.at/ctan/macros/latex/contrib/caption/caption-eng.pdf | |
\usepackage{balance} % balancing bibstyles as per request in accepted submission | |
\usepackage{graphicx} | |
% We will generate all images so they have a width \maxwidth. This means | |
% that they will get their normal width if they fit onto the page, but | |
% are scaled down if they would overflow the margins. |
\documentclass[sigconf,anonymous=$anonymous$]{acmart} | |
\usepackage{booktabs} | |
\usepackage{caption} % http://mirror.easyname.at/ctan/macros/latex/contrib/caption/caption-eng.pdf | |
\usepackage{balance} % balancing bibstyles as per request in accepted submission | |
\usepackage{graphicx} | |
% We will generate all images so they have a width \maxwidth. This means | |
% that they will get their normal width if they fit onto the page, but | |
% are scaled down if they would overflow the margins. |