Tuesday, July 22, 2014

Link Building: Why Should You Build Links Today?

Webster's Dictionary defines link building as... it doesn't actually.




Why is there no standard definition? It isn't like link building is a secret. According to a recent survey, 80 percent of companies engaged in SEO are spending more than $1,000 a month on link building.

The lack of a standard definition leads to a large amount of confusion within the industry itself. I've met other SEOs and marketers who had a tenuous grasp on link building at best. It's not like there's an SEO college. Forget that: SEO is barely taught in universities at all.

So what is link building exactly? If you're asking me for a bare bones definition, I would say it's the process of going out of your way to find great links.

But I don't think that definition truly does link building justice: it ignores the importance of link building. It ignores the crucial role it can and should play in any online marketing campaign. In my opinion, the true definition of link building doesn't come from just asking what, it also comes from asking why.

Link building isn't just the aforementioned process of acquiring backlinks that point to your site: it's also a proven marketing tactic that increases brand awareness and conversions.
The Why

So why should you build links today? You should do it for the same reason you should have built links last year. And the year before. And every year since Google dominated the search market.

You should build links because links are still one of the most important ranking signals in Google's algorithm, and are likely to remain so for the foreseeable future.

You don't have to take just my word for it. Listen to Matt Cutts, who emphasized the importance of links at the recent SMX Advanced conference.

Danny Sullivan: Is link building just dead? You keep saying a new tactic is dead, or must be nofollowed, is it really you just don't want people to try to build links at all?

Cutts: No, link building is not dead. And a very small percentage of links on the web are nofollowed. There's a lot of mileage left in links.

Even though it seemed Sullivan, the founder of Search Engine Land, was trying to bait him into saying link building possessed little value, Cutts rebuffed that idea.

It's not just about increasing search visibility and brand exposure, however. Link building can help drive other marketing strategies. There's a lot of talk about content marketing being "the new link building." I believe this to be a falsehood.

The two are completely different. Content marketing has ambitions beyond building links; it's primarily focused on disseminating a brand message.

If you launch a content marketing initiative solely for links, your content will likely suffer. Content marketing and link building may not be synonymous, but they complement each other nicely. Link building can help you to build relationships that can serve to benefit you in other endeavors as well.

Link building is a crucial component to a comprehensive online marketing campaign.
The Ghost of Link Building Past

I can't accurately provide you with the year when link building started, but I can tell you that link building has existed for a long time. Eric Ward, who is known as the godfather of link building, helped Amazon.com create a link building strategy all the way back in 1994. That's four years before Google even existed.

Source: http://searchenginewatch.com/article/2356174/What-is-Link-Building

Tuesday, January 7, 2014

Future-proof Link Building Techniques for 2014

Since Google’s recent updates, link building by spamming does not work anymore and that’s a good thing. To succeed with your website in 2014, you have to do the right things.

1. In 2014, link building is more alive than ever

You’ve probably seen some “link building is dead” theories on the Internet. People who say this usually engaged heavily in link spam activities.

Automatically creating backlinks in bulk does not work anymore. Faking social network profiles to get backlinks does not work anymore. Spamming forums and article websites with fake or low quality content to get backlinks does not work anymore. Automated link networks do not work anymore.
In short: spamming does not work anymore.

Link building, however, is still the most important thing that you have to do if you want to get high rankings on Google. Google’s Matt Cutts made that clear in an interview:

“Links are still the best way that we’ve found to discover [how relevant or important somebody is], and maybe over time social or authorship or other types of markup will give us a lot more information about that.”

2. Relevance has become even more important

“Things, not strings” is one of the most important concepts that Google introduced last year. The context of a link has become even more important.

The links that point to your website should come from pages that are related to the topic of your own website. Google has been saying this for years but now they are serious about it.

If all of the links that point to your website contain exactly the same keyword, you can be sure that this will trigger an ‘unnatural links’ filter. If the links to your website contain keywords that are related to the topic of your website, Google will find your website relevant for that topic.

3. Once again: spamming is risky

Some people still think that they can trick Google’s algorithm with the ‘brand new secret method that will get your site on Google’s first result page with just a few mouse clicks.’ These spam tools do not work anymore.

Google does not like link schemes and you should avoid these link types:
  • Buying or selling links that pass PageRank.
  • Excessive link exchanges.
  • Large-scale article marketing or guest posting campaigns with keyword-rich anchor text links.
  • Using automated programs or services to create links to your site.
  • Text advertisements that pass PageRank.
  • Advertorials or native advertising where payment is received for articles that include links that pass PageRank.
  • Links with optimized anchor text in articles or press releases distributed on other sites.
  • Low-quality directory or bookmark site links.
  • Links embedded in widgets that are distributed across various sites.
  • Widely distributed links in the footers of various sites.
  • Forum comments with optimized links in the post or signature

If you use these methods to build backlinks, it’s likely that your website will be penalized.


Read More: http://blog.seoprofiler.com/2014/01/future-proof-link-building-techniques-2014/

Thursday, January 2, 2014

SEO Updates: Know All Google Algorithm Changes 2013

Unnamed Update — November 14, 2013

Multiple Google trackers picked up unusual activity, which co-occurred with a report of widespread DNS errors in Google Webmaster Tools. Google did not confirm an update, and the cause and nature of this flux was unclear.

Penguin 2.1 (#5) — October 4, 2013

After a 4-1/2 month gap, Google launched another Penguin update. Given the 2.1 designation, this was probably a data update (primarily) and not a major change to the Penguin algorithm. The overall impact seemed to be moderate, although some webmasters reported being hit hard.

Hummingbird — August 20, 2013

Announced on September 26th, Google suggested that the "Hummingbird" update rolled out about a month earlier. Our best guess ties it to a MozCast spike on August 20th and many reports of flux from August 20-22. Hummingbird has been compared to Caffeine, and seems to be a core algorithm update that may power changes to semantic search and the Knowledge Graph for months to come.

In-depth Articles — August 6, 2013

Google added a new type of news result called "in-depth articles", dedicated to more evergreen, long-form content. At launch, it included links to three articles, and appeared across about 3% of the searches that MozCast tracks.

Unnamed Update — July 26, 2013

MozCast tracked a large Friday spike (105° F), with other sources showing significant activity over the weekend. Google has not confirmed this update.

Knowledge Graph Expansion — July 19, 2013

Seemingly overnight, queries with Knowledge Graph (KG) entries expanded by more than half (+50.4%) across the MozCast data set, with more than a quarter of all searches showing some kind of KG entry.

Panda Recovery — July 18, 2013

Google confirmed a Panda update, but it was unclear whether this was one of the 10-day rolling updates or something new. The implication was that this was algorithmic and may have "softened" some previous Panda penalties.

Multi-Week Update — June 27, 2013

Google's Matt Cutts tweeted a reply suggesting a "multi-week" algorithm update between roughly June 12th and "the week after July 4th". The nature of the update was unclear, but there was massive rankings volatility during that time period, peaking on June 27th (according to MozCast data). It appears that Google may have been testing some changes that were later rolled back.

Panda Dance — June 11, 2013

While not an actual Panda update, Matt Cutts made an important clarification at SMX Advanced, suggesting that Panda was still updating monthly, but each update rolled out over about 10 days. This was not the "everflux" many people had expected after Panda #25.

"Payday Loan" Update — June 11, 2013

Google announced a targeted algorithm update to take on niches with notoriously spammy results, specifically mentioning payday loans and porn. The update was announced on June 11th, but Matt Cutts suggested it would roll out over a 1-2 month period.

Penguin 2.0 (#4) — May 22, 2013

After months of speculation bordering on hype, the 4th Penguin update (dubbed "2.0" by Google) arrived with only moderate impact. The exact nature of the changes were unclear, but some evidence suggested that Penguin 2.0 was more finely targeted to the page level.

Domain Crowding — May 21, 2013

Google released an update to control domain crowding/diversity deep in the SERPs (pages 2+). The timing was unclear, but it seemed to roll out just prior to Penguin 2.0 in the US and possibly the same day internationally.

"Phantom" — May 9, 2013

In the period around May 9th, there were many reports of an algorithm update (also verified by high MozCast activity). The exact nature of this update was unknown, but many sites reported significant traffic loss.

Panda #25 — March 14, 2013

Matt Cutts pre-announced a Panda update at SMX West, and suggested it would be the last update before Panda was integrated into the core algorithm. The exact date was unconfirmed, but MozCast data suggests 3/13-3/14.

Panda #24 — January 22, 2013

Google announced its first official update of 2013, claiming 1.2% of queries affected. This did not seem related to talk of an update around 1/17-18 (which Google did not confirm).

Read More at http://moz.com/google-algorithm-change

Wednesday, December 18, 2013

SEO News: Matt Cutts Say "A Little Duplicate Content Won't Hurt Your Rankings"



Duplicate content is always a concern for webmasters. Whether it's a website stealing content from another site, or perhaps a website that hasn't taken an active role in ensuring they get great unique quality content on their site, being duplicated out of the Google index is a problem.

In the latest webmaster help video from Google's Matt Cutts, he addresses how Google handles duplicate content, and when it can negatively impact your search rankings.

Matt Cutts started by explaining what duplicate content is and why duplicate content isn't always a problem, especially when it comes to quoting parts of other web pages.

It's important to realize that if you look at content on the web, something like 25 or 30 percent of all of the web's content is duplicate content. … People will quote a paragraph of a blog and then link to the blog, that sort of thing. So it's not the case that every single time there's duplicate content it's spam, and if we made that assumption the changes that happened as a result would end up probably hurting our search quality rather than helping our search quality.

For several years, Google's stance has been that they try to find the originating source and give that result the top billing, so to speak. After all, Google doesn't want to serve up masses of identical pages to a searcher because it doesn't provide a very good user experience if they click on one page, didn't find what they're looking for, and then go back and click the next result only to discover the identical page, just merely on a different site.

Google looks for duplicate content and where we can find it, we often try to group it all together and treat it as of it's just one piece of content. So most of the time, suppose we're starting to return a set of search results and we've got two pages that are actually kind of identical. Typically we would say, "OK, rather than show both of those pages since they're duplicates, let's just show one of those pages and we'll crowd the other result out," and then if you get to the bottom of the search results and you really want to do an exhaustive search, you can change the filtering so that you can say, "OK, I want to see every single page" and then you'd see that other page. But for the most part, duplicate content isn't really treated as spam. It's just treated as something we need to cluster appropriately and we need to make sure that it ranks correctly, but duplicate content does happen.

Next, Cutts tackles the issue of where duplicate content is spam, such as websites that have scraped content off the original websites or website owner suggests republish a lot of “free articles” that are republished on masses of other websites. These types of sites have the biggest problem with duplicate content because they merely copy content created on other websites.

It's certainly the case that if you do nothing but duplicate content, and you are doing in an abusive, deceptive, malicious, or a manipulative way, we do reserve the right to take action on spam. So someone on Twitter was asking a question about "how can I do an RSS auto blog to a blog site and not have that be viewed as spam," and the problem is that if you are automatically generating stuff that is coming from nothing but an RSS feed, you're not adding a lot of value, so that duplicate content might be a little bit more likely to be viewed as spam.
There are also cases where businesses might legitimately end up with duplicate content that won't necessarily viewed as spam. In some cases, websites end up with duplicate content for usability reasons, rather than SEO. For the most part those websites shouldn't worry either

But if you're just making a regular website and you're worried about whether you'd have something on the .com and the .co.uk, or you might have two versions of your terms and conditions, an older version and a newer version, that sort of duplicate content happens all the time on the web and I really wouldn't get stressed out about the notion that you might have a little bit of duplicate content.

Cutts does caution against local directory types of websites that list masses of cities but serve up empty listings with no true content about what the user might be looking for, as well as sites that create individual pages for every neighborhood they service, even though the content is the same as what's on main city web page.

As long as you're not trying to massively copy for every city in every state in the entire United States, show the same boilerplate text which is, "no dentists found in this city either," for the most part you should be in very good shape not have anything to worry about.

Bottom line: as long as your duplicate content is there for legitimate reasons (e.g., you're quoting another website or you have things like two versions of terms and conditions), you really shouldn't be concerned about duplicate content. However, Google certainly can and will take action against sites utilizing duplicate content in a spammy fashion, because they aren't adding value to the search results.

Source: http://searchenginewatch.com/article/2319706/Googles-Matt-Cutts-A-Little-Duplicate-Content-Wont-Hurt-Your-Rankings

Friday, October 18, 2013

SEO Update: What To Do If Your Website Was Hit By Penguin 2.1?

Will the Penguin Damage or Hurt?


Google lately launched information about the newest in a sequence of junk battling criteria up-dates that the look for massive has applied since 2012 under the ‘Penguin’ name. This time they have marked the upgrade as Penguin 2.1. As with all Penguin criteria up-dates, many incoming promoters are frantically verifying their web page positions right now to discover out if they have been hit with any charges from this new upgrade.

According to Google’s head of web junk, Matt Cutts, about 1% of all queries will be affected by the Penguin 2.1 update, which is quite a bit more than with past minimal Penguin update. So will it effect search results of your site?

What’s Penguin all About Again?

Penguin was presented as a part of Google’s look for criteria that tries to discover sites that are position well in search engines look for because of spam link-building methods. If Penguin recognizes your web page as using spam methods for enhancing positions, you will get punished and your look for engine outcomes positioning positions will fall, possibly by a lot.

Being hit by Penguin can be terrible for your web page, especially if you depend intensely on look for traffic for lead generation.

What to do if Your Website Was Hit?

If your statistics show that your web page positions have considerably decreased since Oct 5th, you might have been punished by the Penguin 2.1 upgrade. If this is the case, you want to act fast. Get rid of any bad back-links using Google’s Disavow device, or other weblink trimming methods. For the long run, move away from concentrating too much on producing back-links. Instead you should create top quality material and provide an excellent consumer encounter. This will help with normally gaining hyperlinks to your website.

Even after eliminating spam hyperlinks you still might not see your website restore from a Penguin hit though. In the past we’ve often seen that sites that are hit by Penguin don’t restore after cleaning up their act until the next Penguin upgrade.

So what Penguin 2.1 means to incoming promoters is that Bing is not done battling web junk just yet. Concentrate on making a top quality website, with excellent material and an similarly excellent consumer encounter and your possibilities to succeed in the long run incoming marketing game will considerably increase.