Internal Links are hyperlinks that point at (target) the same domain as the domain that the link exists on (source). In layman's terms, an internal link is one that points to another page on the same website.
Code Sample <a href="http://www.same-domain.com/" title="Keyword Text">Keyword Text</a> Optimal FormatUse descriptive keywords in anchor text that give a sense of the topic or keywords the source page is trying to target. What is an Internal Link?Internal links are links that go from one page on a domain to a different page on the same domain. They are commonly used in main navigation. These type of links are useful for three reasons:
On an individual page, search engines need to see content in order to list pages in their massive keyword–based indices. They also need to have access to a crawlable link structure—a structure that lets spiders browse the pathways of a website—in order to find all of the pages on a website. (To get a peek into what your site's link structure looks like, try running your site through Open Site Explorer.) Hundreds of thousands of sites make the critical mistake of hiding or burying their main link navigation in ways that search engines cannot access. This hinders their ability to get pages listed in the search engines' indices. Below is an illustration of how this problem can happen: In the example above, Google's colorful spider has reached page "A" and sees internal links to pages "B" and "E." However important pages C and D might be to the site, the spider has no way to reach them—or even know they exist—because no direct, crawlable links point to those pages. As far as Google is concerned, these pages basically don’t exist–great content, good keyword targeting, and smart marketing don't make any difference at all if the spiders can't reach those pages in the first place. The optimal structure for a website would look similar to a pyramid (where the big dot on the top is homepage): This structure has the minimum amount of links possible between the homepage and any given page. This is helpful because it allows link juice (ranking power) to flow throughout the entire site, thus increasing the ranking potential for each page. This structure is common on many high-performing websites (like Amazon.com) in the form of category and subcategory systems. But how is this accomplished? The best way to do this is with internal links and supplementary URL structures. For example, they internally link to a page located at http://www.example.com/mammals/cats.html with the anchor text "cats." Below is the format for a correctly formatted internal link. Imagine this link is on the domain jonwye.com. In the above illustration, the "a" tag indicates the start of a link. Link tags can contain images, text, or other objects, all of which provide a "clickable" area on the page that users can engage to move to another page. This is the original concept of the Internet: "hyperlinks." The link referral location tells the browser—and the search engines—where the link points. In this example, the URL http://www.jonwye.com is referenced. Next, the visible portion of the link for visitors, called "anchor text" in the SEO world, describes the page the link is pointing at. In this example, the page pointed to is about custom belts made by a man named Jon Wye, so the link uses the anchor text "Jon Wye's Custom Designed Belts." The </a> tag closes the link, so that elements later on in the page will not have the link attribute applied to them. This is the most basic format of a link—and it is eminently understandable to the search engines. The search engine spiders know that they should add this link to the engine's link graph of the web, use it to calculate query-independent variables (like MozRank), and follow it to index the contents of the referenced page. Below are some common reasons why pages might not be reachable, and thus, may not be indexed. Links in Submission-Required FormsForms can include elements as basic as a drop–down menu or elements as complex as a full–blown survey. In either case, search spiders will not attempt to "submit" forms and thus, any content or links that would be accessible via a form are invisible to the engines. Links Only Accessible Through Internal Search BoxesSpiders will not attempt to perform searches to find content, and thus, it's estimated that millions of pages are hidden behind completely inaccessible internal search box walls. Links in Un-Parseable JavascriptLinks built using Javascript may either be uncrawlable or devalued in weight depending on their implementation. For this reason, it is recommended that standard HTML links should be used instead of Javascript based links on any page where search engine referred traffic is important. Links in Flash, Java, or Other Plug-InsAny links embedded inside Flash, Java applets, and other plug-ins are usually inaccessible to search engines. Links Pointing to Pages Blocked by the Meta Robots Tag or Robots.txtThe Meta Robots tag and the robots.txt file both allow a site owner to restrict spider access to a page. Links on pages with Hundreds or Thousands of LinksThe search engines all have a rough crawl limit of 150 links per page before they may stop spidering additional pages linked to from the original page. This limit is somewhat flexible, and particularly important pages may have upwards of 200 or even 250 links followed, but in general practice, it's wise to limit the number of links on any given page to 150 or risk losing the ability to have additional pages crawled. Links in Frames or I-FramesTechnically, links in both frames and I-Frames are crawlable, but both present structural issues for the engines in terms of organization and following. Only advanced users with a good technical understanding of how search engines index and follow links in frames should use these elements in combination with internal linking. By avoiding these pitfalls, a webmaster can have clean, spiderable HTML links that will allow the spiders easy access to their content pages. Links can have additional attributes applied to them, but the engines ignore nearly all of these, with the important exception of the rel="nofollow" tag. Want to get a quick glimpse into your site's indexation? Use a tool like Moz Pro, Open Site Explorer, or Screaming Frog to run a site crawl. Then, compare the number of pages the crawl turned up to the number of pages listed when you run a site:search on Google. Rel="nofollow" can be used with the following syntax: <a href="/" rel="nofollow">nofollow this link</a>In this example, by adding the rel="nofollow" attribute to the link tag, the webmaster is telling the search engines that they do not want this link to be interpreted as a normal, juice passing, "editorial vote." Nofollow came about as a method to help stop automated blog comment, guestbook, and link injection spam, but has morphed over time into a way of telling the engines to discount any link value that would ordinarily be passed. Links tagged with nofollow are interpreted slightly differently by each of the engines.
2 Comments
|
Page Traffic By SEOPage Traffic By SEO launched in 2010 as a mainstream of innovative online internet marketing, creativity has always been the lifeline of our business process and key deliverable. We have touched all industries ranging from retail and fashion to travel and finance for offering SEO, Social Media Optimisation, digital marketing and Web Designing services. Categories
All
|