Skip to content

Instantly share code, notes, and snippets.

@lisongx
lisongx / gist:4050902
Created November 10, 2012 12:16
fp week6
def sentenceAnagrams(sentence: Sentence): List[Sentence] = {
def sen_iter(occur: Occurrences): List[Sentence] = {
if (occur == List())
List(Nil)
else for {
word_occur <- combinations(occur).filter(_ != List())
word <- dictionaryByOccurrences(word_occur)
rest <- sen_iter(subtract(occur, word_occur))
} yield word :: rest
}
import urlparse
from scrapy.contrib.spiders import CrawlSpider, Rule
from scrapy.contrib.linkextractors.sgml import SgmlLinkExtractor
from scrapy.selector import HtmlXPathSelector
from isbullshit.items import IsBullshitItem
class IsBullshitSpider(CrawlSpider):
""" General configuration of the Crawl Spider """
from scrapy.contrib.spiders import CrawlSpider, Rule
from scrapy.contrib.linkextractors.sgml import SgmlLinkExtractor
from scrapy.selector import HtmlXPathSelector
from dmovie.items import MovieItem
import re
class MovieSpider(CrawlSpider):
name = 'douban_movie'
allowed_domains = ["http://movie.example.com/"]
start_urls = ['http://movie.douban.com/']
$('body').append('<div id="tylerdurden"></div>');
$('#tylerdurden').click(function(e){
DBR.act($(this).attr('action'));
});
@lisongx
lisongx / get_xidain_news.py
Created March 24, 2012 11:44
获得最see.xidian.edu.cn近几日的新闻
#!/usr/bin/env python
import requests
from lxml import html
from datetime import datetime, timedelta
# get news in last X days
DAY = 3
# url of the news page
@lisongx
lisongx / gist:1517515
Created December 24, 2011 15:04
download the hacker monthly issue
from fabric.api import local, run
from pyquery import PyQuery as pq
BASE_URL = "http://hackermonthly.com/"
HACKER_URL = "http://hackermonthly.com/issues.html"
d = pq(url=HACKER_URL)
issue_list = d("#issues li a")
def get_issue(inex, node):
d = pq(node)