top | item 6039665

Reddit Insight

51 points| gdi2290 | 12 years ago |redditinsight.com | reply

54 comments

order
[+] ChuckMcM|12 years ago|reply
So two big design issues, if you scroll down from the wheel of terms you go wobbly, clicking the down arrow below that last wheel, also wobbly. If you use on of the tabs to go to a graph you create a 'back loop' where the back button takes you to the start, then back again takes you back to the graph, then back to the start then back to the graph ....

The data is interesting but I expect to see a bit of analysis in projects like this. Without it the project becomes a sort of data Rorschach test where the viewer projects their perceptions into it.

[+] gdi2290|12 years ago|reply
We are aware of a lot of the issues on the home page and plan to completely redesign it soon. However which graph are you talking about? If it's the post tracking we will definitely look into it.
[+] malcolmmcc|12 years ago|reply
> "Dogs" occurs more times in titles in r/aww than "cats" or "kittens"

That can be a fact.

> Despite their internet popularity cats are not submitted nearly as many times on this cuddly SubReddit.

Or dogs are rare enough that they're worth naming, because cats are default. Seriously, run whatever calculations you want, but be careful about what conclusions you draw from the numbers.

[+] AjithAntony|12 years ago|reply
Yeah, there are probably more cat pictures just in the "If I fits, I sits" category, than dogs altogether
[+] gdi2290|12 years ago|reply
Good point. We like to stir the pot a bit though.
[+] ctide|12 years ago|reply
Nothing like a bunch of comments tearing apart a website built by people who, presumably, learned to code a few months ago and threw it together over the course of a few weeks. Yeah, it has a few issues, but it's significantly better than the first things I ever built and sent off to the world.

Keep it classy, HN.

[+] ineedtosleep|12 years ago|reply
Examples of "tearing apart"? At its worst, the feedback is terse. If all one wants are warm and fuzzies, one wouldn't submit here.
[+] freehunter|12 years ago|reply
What's the point in submitting this to HN if not to get feedback?
[+] scholia|12 years ago|reply
It's fundamentally horrible. The whole design approach is horrible. It's the contemporary equivalent of GeoCities, just using more recent but equally mindless design cliches.

In fact, it's so horrible it makes PowerPoint look good.

[+] Tomdarkness|12 years ago|reply
Site was borderline unusable with Chrome on android. Text was overlapping, animations to display new text were being triggered after I scrolled past them and in some cases the text just flew in one side and right off the other side of the screen before I even got a glimpse of it.
[+] gdi2290|12 years ago|reply
We updated the landing page to better reflect your suggestions.
[+] oliverhunt|12 years ago|reply
> r/Technology leads all SubReddits, with an average of 2,027 Karma per Post

Is this saying that all posts in r/technology get an average of 2027 karma?

[+] malcolmmcc|12 years ago|reply
My prior on that is extremely low, so my guess is that's like, of the posts that make front-page? Or of the top posts. What this really reflects, therefore, is something more like how up-votey people are.
[+] joshu|12 years ago|reply
average is probably the wrong characterization of the distribution of karma per post.
[+] visarga|12 years ago|reply
This is great and all, but if there is one tool I'd like to see is a personal data exporter.

I want to get all the posts and messages I have ever posted, going back many years, and beyond the 1000 cutoff in their comment history feeds. I'd like to run some keyword analysis on my own data, search it, access it however I see fit. As it stands now, there is no way to retrieve it.

[+] gdi2290|12 years ago|reply
Thanks for your feedback, but what you're asking for Reddit doesn't provide in their API
[+] alexleavitt|12 years ago|reply
How much of "the Reddit" did you actually use to calculate these stats? Is this based off a one-time snapshot from the API (which limits to 1000 posts, per type of query) or is it a longitudinal crawl (and how data was produced in that sample)?
[+] gdi2290|12 years ago|reply
One-time snap shot. Given longer time on this project we would love to do a larger analysis.
[+] zoba|12 years ago|reply
You should know this site isn't usable/viewable on iPhone Safari or Chrome.
[+] gdi2290|12 years ago|reply
sorry about that it's our initial prototype
[+] frakkingcylons|12 years ago|reply
For how long did you collect data? During previous crawls, I've found that one of my spider bots can scrape through about 11.5k submissions and 51k comments per day (while observing Reddit's API access rules).
[+] joeblau|12 years ago|reply
This looks awesome. Next step is to make it "real-time."

I worked for an analytics company and got to build some pretty awesome visualizations using D3 and one of the problems I always ran into was that while the visualizations are cool, you rarely get any actionable information from the charts. I feel like this would be a lot better if at the end, there was some call to action.

[+] minimaxir|12 years ago|reply
Real time analytics is tricky due to API limits. (unless you can accept a "real time" of minutes/hours per update)

Example: Twitter's search API is limited to 15 queries of 100 Tweets every 15 minutes. Do you query 100 Tweets every minute, or 1500 Tweets every 15 minutes?

[+] babs474|12 years ago|reply
I've got an experimental, realtime comment search engine over at commentfindder.com. Perhaps some useful visualizations could be created off of it. Do you have any ideas for what kind of actionable chart would be interesting?
[+] gdi2290|12 years ago|reply
Side note: everyone in our team is available for hire
[+] robg|12 years ago|reply
What is your contact info?
[+] shaohua|12 years ago|reply
we should do one for hacker news too.
[+] minimaxir|12 years ago|reply
It's much harder to get data from Hacker News, as HN doesn't have a public API (and HNSearch's API limits make it infeasable)