Hacker Read top | best | new | newcomments | leaders | about | bookmarklet login

but we don't scrape content. our search results do have abstracts but they are smaller than Google's, and in some cases they actually are google's.

so, i think you have the issue confused.

Also, on #1 I think most SEOs would say your wrong. More pages does NOT equal more page rank. It's the EXACT opposite.

MORE QUALITY PAGES and more QUALITY links = bette SEO from what the top SEOs have told me. I hope this helps with your site.

all the best,

Jason



sort by: page size:

Jason, I'm quite sure I'm not confused.

1) "MORE QUALITY PAGES and more QUALITY links = bette SEO from what the top SEOs have told me. "

For this purpose, indexed by Google counts as "quality." Please read the PageRank paper, and you'll quickly see that more pages in the index == more PageRank.

That doesn't always mean putting crap pages into Google is a good idea, but strictly speaking, it will raise the total amount of PageRank your site has. Seriously, read the paper yourself.

2) "but we don't scrape content. our search results do have abstracts but they are smaller than Google's, and in some cases they actually are google's."

Perhaps scrape is the wrong term. But you absolutely do include other people's content and nofollow it, for example all of the images on this page:

http://www.mahalo.com/chupacabra

NoFollow is meant for links to content which you don't want to "vouch" for. If you've included their content in your page, I think you can vouch for it.

Maybe an honest oversight, maybe not. Either way, you know about it now, so fix it.


#13 (write less pages and less text to conserve ranking potential) is directly contrary to both the truth and my experience. The more content I add to my website, the better it ranks in aggregate and the better the front page ranks in particular. If you want I can even show you a graph of amount of content versus traffic as a result of Google. Let me spoil the surprise: it fits n * (# pages) ^ (1 + y).

Your pages do compete against each other in some narrow senses: for any given query, you'll typically be capped at two results in the SERP (results page), so if you had a third best option on your site it gets left out in the cold. However, pages on your website are (and should be treated as) comrades in arms.

The more links my page on, e.g., Jane Austen bingo cards collects, the more trust my domain will have and the more juice that individual page will pass to the literature category its in and the other page it links to. This helps me rank both the other literature pages (which have scant anything to do with Jane Austen) and also the almost wholly unrelated, e.g., Easter bingo.


SEO - Google ranks those pages higher.

If it's getting scraped heavily and published elsewhere, Google will most likely not pass PageRank for any links other than the original once it's recognized as duplicate content on multiple pages.

In fact, in the MarketWire article, having a bunch of keyword-rich text all pointing at the same page will likely have little SEO value, as diminishing returns hits really quickly in a case like this.


This whole thing is insane.... we have "stub" pages just like Wikipedia.

These are topic pages that people are working on and THEY DON'T RANK in search engines until they we get the word count to around 300-500 words.

We are the process of NOINDEXING the pages that are below 300 words just to make Aaron happy... we actually had these noindexed before our last version and that got lost in the shuffle of the new launch (really, it did... when you do new code you might leave something out of the old code).

i'm also getting a list of every page under 300 words and having the page managers build them out in 30 days or deleting them.

Anyway, i thank Aaron for busting out chops and making us better!

The claims that we are "scraping" are absurd... we're using google, bing, twitter, etc. apis to do a comprehensive search page.

i dont know everything about SEO, but i don't understand this claim by Aaron. i think he is trying to start trouble for us... and maybe it will work. Thanks pal!


They do not come naturally, speaking from experience. Our site is placed on the 2nd to 3rd pages for the vital search queries that our product needs for long-term survival despite relatively low competition.

But most of the first page results have many times more backlinks. We've got less than a dozen organically during the last year.


Hey cellis, thanks. What I see in cases where we come up higher in SERPs is that the crawler just hasn't found the best content in google groups yet, if ever. I scratch my head over that one too, but ours is not to reason why. I don't see the index pointing to the same content in the landing pages typically either with my queries.

UI: LOL, yeah, it's a toss up here whether to make it prettier or keep it a bit old school.


Not exactly.... take a look at the how to articles, walkthroughs, buzz news on the homepage and Q&A community. We're doing some amazing content and Aaron is focusing on short pages that are not ranked in any search engine and, that frankly, we don't let folks build any more (we started with an open system like wikipedia where anyone could build and build they did... now we have a system like MTurk where you have to complete the page before we publish it).

the page we get traffic for ten to have a LOT of content... also, if we see a page gets more than 25-50 page views we put a writer on it to build it out.

so, basically nothing that ranks is a short content page (stub in wikipedia terms).

however, the truth is google doesn't rank short pages except for VERY long tail ones.... and even that is rare.

google ranks you for long, deep content that is updated.

Aaron is making a big deal out a problem that doesn't exist because he is mad at me for saying seo is bullshit back in 2006. :-)


Our pages are built by our community, so the quality will vary. That page doesn't have a "vertical manager" yet, but it will. Then we would build out the content a little more.

It's not a perfect system, people can't put multiple search boxes in there.

However, that page will never rank well in a search engine (unless by a fluke). In order to rank well you really need to have 500+ original words.

We're in the process of moving all pages to that standard. It's really a self-regulating thing: if our contributors make short pages they never rank and never make money. They get frustrated and we teach them how to make longer pages and some day they may rank.

... it's really not a problem, and the truth is we rank for three things well:

1. video game walkthroughs (typically 2-10,000 words!) 2. how to articles (typically 800 to 5,000 words) 3. question & answer pages (typically 300 to 10,000 words),

Isn't this basic SEO (and i'm not expert): build original content and you might rank. Build short pages, you don't rank.

All pages start short (just like wikipedia stubs do), and over time we make them longer. that's the normal process.


It's actually true... we can make mistakes.

... the truth is having short content pages indexed works against you--that's why i no indexed them to being with.

Again, I'm not as big of an expert on SEO as Aaron, but I think there is something called "page rank sculpting" in which you push your sites page rank to high-quality pages and no-index the ones that are shorter in terms of original content. We did that because one of our people read about it on a blog.

We're probably holding ourselves back having removed this and we are putting it back on because we didn't realize it was off.

That is why I thanked Aaron.

Is the page rank sculpting thing not a good idea? I thought this was a fairly certain thing: only index the best pages.


Let's be clear here: the content doesn't rank because it's on a subdomain or folder of the strong domain, it ranks because it's there AND gets lots of links from most/all pages of that strong domain. The links are the relevant part, you'd have similar (albeit not as strong) results if you had site-wide links from top domains to your separate domain.

Google already has a rule to deal with this: buying and selling links (and, by extension, renting links) is against their guidelines. They just don't enforce it for big publishing companies.


Site freshness only matters for certain categories/topics. For other topics you could say the older the static content the better. Some of my best ranking/money making pages have not changed in 8 years and rank better every year, without doing anything.

Regarding ranking algos, you did not mention the Trust Rank (there is a patent G filed some years ago). A low Trust Rank is equally bad as a low Page Rank. I don't really know much about it, but afaik it has to do with the trust that G has collected on your site, while Page Rank is based on individual pages. If you link to "bad neigborhoods" (another important term, think casino sites, v*agra, etc.) your TR will go down for example, and its hard to get it back up. If National Geographic links to one of your pages, your site's TR goes way up.

And then there is the number of pages on your site. Google will only crwal a limited amount of pages from your site, depending on your Page Rank and distribution of incoming links. So try to keep the number of crawlable pages focused to the relevant pages (ie the ones you'd like to see in G's index).

After some time and if you have tons of links, you will see a "deep crawl" with a bunch of Gbots hammerings you server and downloading many more pages a day than usual.


I agree. Google nowadays show real SEO results below a huge amount of links. My entire blog is listing on 2nd pages, having the 100/100 pagespeed, best practices and 20 year domain record. Still I see the threshold around the 2nd/3rd page. Who goes there?

> Yes, from a basic standpoint, descriptive URLs are easier to crawl, index, and rank accordingly.

Why are they easier to crawl and index? Most people do shoot themselves in the foot with some duplicate content if they try to implement those speaking URLs.

And "rank accordingly"? Like keyword tags help to rank better?


If I remember correctly, your page ranks better in Google if the search terms are in the URL (even more so if they are in the hostname).

Do any SEOs put thought into the internal linking structure of their sites?

It would seem that, because PageRank is divided amongst the number of links on a page, that a page with fewer links is more effective at passing link juice then a page with many links.


Seeing recent Top news related to SEO on HN it seems like backlinks is the key to rank in Google.

Quality of content and Quantity doesn't matter all what matters is backlinks!


Absolutely, but it's just one piece of a bigger SEO puzzle. The more pieces you figure out, the higher you'll rank.

I really just wanted to point out that the article wasn't 'Proof of how important website titles are for SEO', as there are other elements on that page that could have been just as responsible for the ranking.


If a site ranks high in a SR, isn't it because it must have good backlinks?

Shouldn't that be more important than whether it has original content or not?

Maybe an aggregator displays the content in a more useful way than the originator so it gets linked to more than the originator.

If the content creator doesn't like his content copied he can take it up with the copier. It's not Googles job to get involved in that.

Google's job is to give the searcher a list of the sites that matches his keywords in an unbiased way. They should do that mostly on what the internet thinks is the best site, not what Google thinks is the best site.

next

Legal | privacy