Hacker Newsnew | past | comments | ask | show | jobs | submit | ycmike's commentslogin

Love the "have a beer while the students use Coursera" approach. As a student this would be amazing.


Yep same for @sas.upenn.edu


Same, sent them an email though.


I'd like to hear the responses to this since I am in a similar position. I was accepted to a startup accelerator at UPenn and have been working full-time the past 8 weeks both learning to code and building the application. Before that, I was working part-time for about 2 months while holding a job.

Lately, I've been in a massive rut and have serious doubts seeing if I have what it takes to make my startup succeed. It is mostly due to the technical side since I'm just graduating a bootcamp but I feel my energy and drive waining.


I also had to learn to code to get my startup off the ground; it took a while and was painful, no doubt about that. Eventually I realized that coding is a process that I could enjoy if I let myself. Same with all the other things I do day to day.

But 8 weeks? That's it? You have a long way to go. It took me about that long to get to where I was good enough to create a basic app on my own. Building a business on top of your coding abilities takes much longer, so try to focus on what you have accomplished. It will always (ALWAYS) take longer than you'd think to get where you'd want to be, so just repeat what everyone else in this thread has been saying: it's a marathon, not a sprint.


I love reading stories like this.


A fellow Starter League alumni did this and I am super impressed. Here is the main site too. http://codemarks.com/codemarks


I can agree somewhat that episode 2 did not have that special something.

An alternative idea I have is that many of use forget what it is like to watch one episode every week instead of getting the whole season at one time. It is painful actually and you really dissect each episode more.


Yeah this is a strange fascination and yeah I can see why they want to add diversity to a show that would naturally not have much.

I must admit Richard is rather frustrating at times. I know I should root for him but man as a leader I hope he really grows. Although I suppose you could see that a little as the episode ended.

The idea of having Big Head on the inside also seems an interesting plot line to me.


HN,

So who do you guys use more? Import.io or Kimono? I have heard good things about both.


I write my own custom scrapers, I prefer the flexibility and feel safer that the service isn't going to disappear any minute.

If anybody is interested, I wrote a detailed article on scraping not so long back that was well received here: http://jakeaustwick.me/python-web-scraping-resource/


I tried Komono, but it cannot auth into the sites I want to pull the data from....

Just grabbed import.io - will see if it can loginto sites and grab the data from services I am already paying thousands per month for.

EDIT:

To add some context: I pay about $3,000 per month for some monitoring services which do not have any real reportin mechanisms. So for my daily and weekly reports, I have to manually compile them and screen shot a ton of things, compose an email and send.

I want to configure a scraper to automatically grab screens of things I want regularly and email them.

I want to have a script that will grab many diff pieces of data (visual graphs, typically) and put them all into one email.

I am working with my monitoring vendors to get them to add reporting... but until that can happen - I am tired of spending a couple hours per week screen capping graphs...


I'm evaluating these to augment a system I'm building on top of casper. This is the first I've seen of this one, but right out of the gate I think I prefer Kimono.


I prefer to rely on code that doesn't rely on an API that could just vanish the next day or cost a bucket to run.


> rely on an API that could just vanish the next day

Kind of ironic that you are saying this about web scraping ...


But then his data source is gone and what he was doing is pointless. Losing your processor of said data source while said data source is still available is frustrating.


What do you use for scraping? I may have a scraping project later this year and would love recommendations.


I've written a couple "polite" crawlers in Go (i.e. obeys robots.txt, delays between requests to the same host).

- Fetchbot: https://github.com/PuerkitoBio/fetchbot

Flexible, similar API to net/http (uses a Handler interface with a simple mux provided, supports middleware, etc.)

- gocrawl: https://github.com/PuerkitoBio/gocrawl

Higher-level, more framework than library.

Coupled with goquery (https://github.com/PuerkitoBio/goquery ) to scrape the dom (well, the net/html nodes), this makes custom scrapers trivial to write.

(sorry for the self-promoting comment, but this is quite on topic)

edit: polite crawlers, not scrapers.


Scrapy gets a solid recommendation from me. http://scrapy.org/


We've got quite an old mailing list full of geeks hand-coding web scrapers, if you want somewhere to ask questions:

https://groups.google.com/forum/#!forum/scraperwiki


I use custom node.js scripts with these libraries:

* request - https://github.com/mikeal/request

* async - https://github.com/caolan/async

* cheerio - https://github.com/cheeriojs/cheerio

* nedb - https://github.com/louischatriot/nedb


Friend of mine was working on this. Don't know if they are still active or if/when they launch. http://pennapps2014s.challengepost.com/submissions/20858-div...


Yeah this is exactly it! Thanks!


This is so useful. I am an extremely deep sleeper - when I do that is - who never hears his multiple alarms. God bless you.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: