Category Archives: SEO

To Frame or Not To Frame for SEO

  • The best way of optimizing a site that uses frames is to stop using frames. Most crawlers only follow HREF links, not SRC links, which means that the crawler won’t find anything past the page with the frameset.
  • Typically development techniques like tables, server side includes, master pages, etc. will allow developers and designers to build content as if it were using a frameset, and be pre-compiled into one page before it is served to the user or crawler.
  • Other development techniques, like a CSS based table-less layout, increase the content to code ratio and can also mirror the layout of frames
  • If Frames absolutely must be used, be sure to set up the NoFrames section. The best thing to put in the NoFrames section is a full version of the page, with keyword rich headings and a full HTML based navigation menu. This will ensure the crawlers see the whole page, maintain keyword SEO, and can index the rest of your site. This will feel as if you are maintaining 2 pages in 1, and defeats the purpose of using frames, but will keep your site optimized for crawling. Also be sure to include the same content in your NoFrames tag – you do not want to lose ranking for keyword stuffing.
  • A common problem that occurs when you use frames is that the search engines will often display one of the internal pages in your site in response to a query. You may want to include some code to verify that the page is show in the frameset it belongs to, otherwise reload the page with its frameset.

The Secrets of WHOIS on SEO

  • WHOIS is a protocol that is used to identify the owner of a domain name or IP address on the Internet.
  • The WHOIS information is now used for validating domains and their content dynamically by search engines and their crawlers:
    • inception date of the domain (age of the site and its content)
    • expiration of the domain (legitimacy of the site and its content – shorter registrations rank lower)
    • Frequency of changes to the WHOIS data (stability of the site and its content)
    • Information about name servers (purpose of the site and its content)
  • Issues with using WHOIS information
    • Privacy concerns
    • False registrations
    • Inaccurate information
    • Obsolete sites
    • Lack of history
    • Internationalization
    • No central WHOIS server list
    • Differing result formats from different WHOIS servers
  • Some tips that will leverage WHOIS information
    • Register your domains for 5 years or more at a time
    • Avoid changing your WHOIS or registration information unless absolutely necessary

Web 2.0 Goes Corporate – Enterprise 2.0

While I was sitting on the beach of Ocean Grove NJ this week, my wife and I read the Technology section of the Wall Street Journal from June 18. This was a very intriguing article. It describes how IBM has embraced the idea of Web 2.0 . A good definition of Web 2.0 is the new interactive social networking of the Internet. Applications like Blogger, Wikipedia, MySpace, LinkedIn, Google Reader, and Del.icio.us are all examples of Web 2.0 applications. Now imagine how a corporation could leverage each of these.

Before we examine how these could be used within a corporate environment, let’s examine the function that each of these Web 2.0 sites serve (these have been “borrowed” from each of the sites above, and slightly modified to be more generic):

  • Blog – a web site where entries are written in chronological order and displayed in reverse chronological order. Blogs provide commentary or news on a particular subject; some function as more personal online diaries. A typical blog combines text, images, and links to other blogs, web pages, and other media related to its topic. The ability for readers to leave comments in an interactive format is an important part of many blogs.
  • Wiki – a server program that allows users to collaborate in forming the content of a Web site. With a wiki, any user can edit the site content, including other users’ contributions, using a regular Web browser. Basically, a wiki Web site operates on a principle of collaborative trust.
  • Community Space – a private community where you can share photos, journals and interests with your growing network of mutual friends
  • Social Networking – an online network people from around the world. When you join, you create a profile that summarizes aspects about you. Your profile helps you find and be found by friends, family, former colleagues, clients, and partners. You can add more connections by inviting trusted contacts to join your network and connect to you.
  • RSS Aggregator – combines multiple syndicated web content sources such as news headlines, blogs, podcasts, and vlogs in a single location for easy viewing.
  • Favorites – a collection of favorites – yours and everyone else’s. You can use this to:
    • Keep links to your favorite articles, blogs, music, reviews, recipes, and more, and access them from any computer on the web.
    • Share favorites with friends, family, coworkers, and the del.icio.us community.
    • Discover new things. Everything on del.icio.us is someone’s favorite — they’ve already done the work of finding it. So del.icio.us is full of bookmarks about technology, entertainment, useful information, and more. Explore and enjoy.

So let’s now try to find corporate applications for each of these services.

Blogs are easy. Each person can create their own blog. Each person should try to focus their blog on pertinent topics to their daily work. This is a great way to capture tacit knowledge about processes, projects, and subjects of expertise. Blogs get a bit more complicated when blogs are to be used as a method of communication outside the company. Then the messages in the blog will probably be reviewed by a corporate communications team.

Favorites are also easy. You could add them and “tag” them by department, division, feature, function or other category. Keeping your favorites stored online instead of on the individual PC allows the favorites themselves to be accessible from any computer anywhere, and can be searched. If you are looking for the corporate provider of translation services, as an example, searching the corporate favorites would yield that information just as much as any other source.

An online RSS Aggregator would be another simple service to provide. It would need to be web based, so that it would be accessible anywhere in the company. This would also allow for metrics to be collected, such as most frequented feeds, posts, blogs, etc. This would also encourage sources of information to be syndicated, such as blogs, corporate news, internal communications, promotions, industry news, etc.

Social Networking pages are a great way to store your profile – name, address, email, phone, etc. It is also a place where you can list your accomplishments, educational and professional history, what projects you have worked on, who you worked with, and any specific topics that you consider yourself a subject matter expert. When people are looking for qualified people to fill their project, or find internal candidates for open positions, view contacts by organization, or just to find an email or phone number, this would be a great tool for that.

The corporate use of wikis could be a bit more complex. The easiest use of a corporate wiki would be to cerate wiki pages for each ongoing project, and allow all project members to add, modify, and update the project pages. This would unify the source of information for all project work. The next logical step would be to use wiki pages for corporate policies, standard operating procedures, departments, and organizational announcements. The amount of corporate knowledge that could be captured by a wiki is endless… and all of it would be shareable, update-able, and searchable in a very easy format that technical and non-technical people can all use just as easy.

Community spaces would be the glue to all of the Enterprise 2.0 services. Each person in the organization would have their own page. The page would link to your social network pages, profiles, and link to your friends or colleagues. The links would be directly tied to their Instant Messenger ID or their email address for easy access. It would display your most recent blog entries. Your favorites would be shown by tag, by most recent added, or most recent used. It would list your most recently viewed or edited wiki pages. You could view your aggregated RSS content. Your email and calendar would be integrated. Creating additional “widgets” like stock ticker, weather, etc. would be easy.

Some of this is already possible with corporate portals like Microsoft SharePoint and BEA AquaLogic. But most is not. Some of these services are also available as individual disparate systems, but need to be justified, funded, tightly regulated, and monitored. IBM is doing some of this, and is self-monitored with corporate responsibility and common sense instead of corporate policies and Legal Review.

So… what do you think? How long do you think it will take to have an integrated Enterprise 2.0? Is all of this together even possible, or is it just a pipe dream?

5 Serious Server Side SEO Specs

  • Make sure your web server supports the If-Modified-Since HTTP header. This feature allows your web server to tell Google whether your content has changed since your site was crawled last. Supporting this feature saves you bandwidth and overhead.
  • DNS Resolution – If there are any consistent DNS resolution issues while the crawler is trying to index the site, it may run into enough trouble to stop trying to index your site, and stop returning later to try again.
  • Your web server should be configured to accept all requests for a domain whether they have the www or not. For example, non-www requests should be 301 redirected to the www domain.
  • If you have a static IP for your site, ensure that the IP address resolves to the site. If it does not, this will cause problems as well
  • To save time, many crawlers will cache the IP of a domain and simply request the domain by IP rather than a full DNS request each time it wishes to index the site. If there are multiple sites on the same IP address, every other site hosted on that IP has the potential to negatively impact your search engine rankings, because the crawler will think that they are all the same.

I am sure these are just a very few of the server related issues that affect SEO. Are there any other server based issues that you have encountered that affect SEO positively or negatively? Leave me a comment and let me know.

The Relationship Between SEO, SEM and Web Analytics

I think it would be best to start with some definitions so that we have a base level to begin with. So, I have gone to Wikipedia and grabbed their definitions:

  • Search Engine Optimization (SEO) is the process of improving the volume and quality of traffic to a web site from search engines via “natural” (“organic” or “algorithmic”) search results
  • Search Engine Marketing (SEM) is a form of Internet Marketing that seeks to promote web sites by increasing their visibility in the Search Engine results pages (SERPs) and has a proven Return on Investment (ROI).
  • Pay Per Click (PPC), or Paid Placement, is an advertising model used on web sites, advertising networks, and search engines where advertisers only pay when a user actually clicks on an ad to visit the advertiser’s web site.
  • Paid Inclusion is a search engine marketing product where the search engine company charges fees related to inclusion of web sites in their search index.
  • Key Performance Indicators (KPI) are financial and non-financial metrics used to quantify objectives to reflect strategic performance of an organization.
  • Web Analytics is the study of the behavior of web site visitors.

My personal perspective expands the definition of Web Analytics to include the measurement of SEM, including paid inclusion, pay-per-click, organic search results, and SEO.

Here is how I see the relationship between all these:

What do you think? Are there better ways to diagram or describe the relationship? Do you agree or disagree with these definitions? Leave your thoughts here in your comments.

6 Sources for SEO and Successful Server Redirects

  • If you are going to move your site, 301 redirects are the recommended practice. These are best understood by search engine crawlers, and happen before the content of the page is rendered for the crawler or the user
  • Meta redirects are generally considered bad practice for search engine optimization purposes, as they are followed on the client side, and not by the crawlers
  • 302 redirects are considered acceptable practice, but only if the content is moving temporarily, as both pages will remain in the crawler index
  • For ASP.Net users, server.redirect or response.redirect seems to be a legitimate practice, as long as it is not used for “cloaking”, or serving up special content for crawlers. A response.redirect is considered the same as a 302. Some testing will need to be conducted to determine if server.redirect is a 301, but that is my suspicion.

Technorati , Bloglines, & del.icio.us

Scott Hanselman has a great post about how to keep you blog from sucking… http://www.hanselman.com/blog/BlogInteresting32WaysToKeepYourBlogFromSucking.aspx

So… I have followed a number of his suggestions. I have signed up with Technorati.com . Technorati is a great place to list your blog. It is not a search engine, but it is a way to claim your blog and add it to its search listings. Here is my Technorati Profile. And, if you want to view my blog from within Technorati instead, you can do that too.

And what timing, too! When I signed up for Technorati, I found a blog post about my blog in just wonky, Rob Fuller’s blog. Thanks, Rob for the kinds word and the inbound link should help a bunch!

Bloglines is similar to Google Reader and Technorati. It allows you to sign up for blogs and RSS feeds, and aggregates them for you to read. It also allows you to claim your blog, and includes it in its index for searching and advertising. I now have a Bloglines account as well, and have claimed my blog.

I have also started to use del.icio.us to store all of my bookmarks. This makes them available in IE, Netscape, and from any PC anywhere. You can see my bookmarks at http://del.icio.us/brian.whaley . There is a great del.icio.us bookmarks plugin for Firefox that you can use that will make you bookmarks directly available from the menu bar. Very handy.

10 Tips to Easily Blend Your Content and Your SEO Keywords

  • Site content should actually contain the keywords you choose… actually a better way to think about it is that your keywords should be chosen based on your site content
  • It is important to drive home a very important point – forcing keywords into places that do not belong will be clearly evident to indexers, crawlers, and your readers. Be sure your keywords fit naturally within their locations.
  • Think “specific keyword phrases” not “keywords”. Due to the extreme amount of competition for general terms in the search engines, if your keyword phrases are too general it is very unlikely you will rank well in the search engines. You stand a far better chance to rank well for specific phrases where there is less competition. The resulting traffic, since it is more highly targeted, should also be much higher quality too!
  • Keywords should appear within the first paragraph on your page
  • Keywords should also appear in H1, H2, H3, etc. Tags.
  • Only one H1 tag is recommended on a page. Note that changing the style of your header tags (H1, H2, etc) will not affect rankings
  • Bold and strong tags emphasize the importance of words to crawlers too
  • Keyword density should be between 3-10% of the content on your page
  • Keyword Proximity is also important. Be sure to place keywords close to each other
  • Keep in mind that the order in which the terms are typed will affect the search results

10 Essential Site Organization and Structure SEO Tips

  • Organize according to themes and logic based on your keywords phrases (key phrases can also be a single word).
  • Use book-like structure, with chapters and blocks delegating importance.
  • Each page should have one concept, i.e. – one or maximum two main key phrases that are more important than the rest.
  • A single, unified concept (without additional data) will greatly increase weighting (importance) of that page in the world of pages that exist for that key phrase.
  • Content Layering
    • http://www.seomoz.org/blog/layering-content-to-maximize-visibility
    • It is generally recommended to keep as flat a structure as possible when planning your web site presence. That does not mean however you need to throw everything in the root directory for best results. You should not go more than 2-3 levels deep in your directory structure.
    • Each major theme or category should become a subdirectory within the root. This is a simple, easy, efficient way to organize your site.
    • To increase visibility, each layer must act like an independent Web site
    • Search engines tend to treat a sub-domain as its own site. In other words, a search engine sees http://google.searchengines.com and http://searchengines.com as essentially 2 different sites.
    • If you’ve only got 10 or 15 or even 50 pages in your sub-domain, chances are it won’t rank as competitively as it would have as a sub-folder of a larger site. But larger sub-domains are a great idea.
  • Site map pages are extremely important
    • http://www.clickz.com/showPage.html?page=3588136
    • Site maps are a great way to be sure all of your pages are indexed.
    • A link to a site map page on all of your pages is a great idea.
    • They make it easy for the crawler to crawl the whole site
    • Make sure the site map is available to your users, not just the crawlers. This will increase usability, and give your users a list of links that point to the important parts of your site.
    • If the site map is larger than 100 or so links, you may want to break the site map into separate pages.
  • Short link paths are better – users should be able to access sub-pages by traversing a minimal number of pages from the start page. The deeper the page, the more specific the content should be, and the less importance search engines will give it.
  • Don’t use pop-ups. These are penalized heavily by Google.
  • Use header tags (h1, h2, h3) for content hierarchy: h1 tags for the page topic (ie. use only one per page), h2 tags for a topical headings, and h3 tags for sub-topical headings.
  • It is recommended that you have an identical navigation menu on each page, with navigation links based on keywords.

5 More Query String SEO Tips You Need To Read

I have been asked to post some information regarding my resources for my query string post. The original post is located here – http://btw73.blogspot.com/2007/06/seo-tip-querystrings.html

So the resources that I used when compiling my information were the following sites:

Here are also a number of other issues to consider when constructing URLs (note that some of this is subjective and not necessarily related to SEO):

  • http://www.seoconsultants.com/articles/1000/urls.asp
  • Dirty URLs – Complex, hard-to-read URLs are often dubbed dirty URLs because they tend to be littered with punctuation and identifiers that are at best irrelevant to the ordinary user
  • Dirty URLS are bad because:
    • Dirty URLs are difficult to type
    • Dirty URLs do not promote usability
    • Dirty URLs are a security risk
    • Dirty URLs impede abstraction and maintainability
  • Dirty URLs are good because:
    • They are portable
    • They can discourage unwanted reuse
  • How to Clean a URL:
    • Keep them short and sweet
    • Avoid punctuation in file names
    • Use lower case and try to address case sensitivity issues
    • Do not expose technology via directory names
    • Plan for host name typos
    • Plan for domain name typos
    • Support multiple domain forms
    • Add guess-able entry point URLs
    • Where possible, remove query strings by pre-generating dynamic pages