Taken from here
Add remonte branch:
git remote add --track master mleung git://github.com/mleung/feather.git
Verify:
git remote
Taken from here
Add remonte branch:
git remote add --track master mleung git://github.com/mleung/feather.git
Verify:
git remote
Use case: You have repository A with remote location rA, and repository B (which may or may not have remote location rB). You want to do one of two things:
NB: Check out git subtree
/git submodule
and this Stack Overflow question before going through the steps below. This gist is just a record of how I solved this problem on my own one day.
Before starting, make sure your local and remote repositories are up-to-date with all changes you need. The following steps use the general idea of changing the remote origin and renaming the local master branch of one of the repos in order to combine the two master branches.
for family in UIFont.familyNames() { | |
let sName: String = family as String | |
println("family: \(sName)") | |
for name in UIFont.fontNamesForFamilyName(sName) { | |
println("name: \(name as String)") | |
} | |
} |
/////////////////////////////////////////////////////// THEORY /////////////////////////////////////////////////////// | |
// | |
// Source: https://en.wikipedia.org/wiki/Bellman–Ford_algorithm | |
// | |
// | |
// function BellmanFord(list vertices, list edges, vertex source)::distance[],predecessor[] | |
// // This implementation takes in a graph, represented as | |
// // lists of vertices and edges, and fills two arrays | |
// // (distance and predecessor) with shortest-path | |
// // (less cost/distance/metric) information |
#!/usr/bin/env ruby | |
toc = "# Table of Contents\n" | |
newmd = "" | |
ARGF.each_line do |line| | |
newmd << line | |
next if !line.start_with?("#") | |
heading = line.gsub("#", "").strip |
##VGG16 model for Keras
This is the Keras model of the 16-layer network used by the VGG team in the ILSVRC-2014 competition.
It has been obtained by directly converting the Caffe model provived by the authors.
Details about the network architecture can be found in the following arXiv paper:
Very Deep Convolutional Networks for Large-Scale Image Recognition
K. Simonyan, A. Zisserman
let styles: [UIFont.TextStyle] = [ | |
// iOS 17 | |
.extraLargeTitle, .extraLargeTitle2, | |
// iOS 11 | |
.largeTitle, | |
// iOS 9 | |
.title1, .title2, .title3, .callout, | |
// iOS 7 | |
.headline, .subheadline, .body, .footnote, .caption1, .caption2, | |
] |
The iTunes API doesn't provide a way to grab artist images, but the iTunes website uses Open Graph meta tags, which embeds a meta
tag with a property
attribute value set to og:image
. As it turns out, this seems to be the same image used in the iTunes artwork.
The URL structure is similar to the artworkUrl
values returned by the API, but what concerns us here is the part I've indicated at the end of the URL.
http://is3.mzstatic.com/image/thumb/Music7/v4/68/68/41/68684190-833b-bfb4-5018-e5a2e6f69eb0/source/1200x630bf.jpg
└─ widthxheight
"Test_a" = "AAAAA"; | |
"Test_b" = "BBBBB"; | |
"Test_c" = "CCCCC"; | |
"Test_d" = "DDDDD"; |