Before you ask, I’ll come out and admit that I did not come up with this information on my own. Instead, I’ve collected information from several reputable sources and have tried to add my own thoughts and references where possible.
My objective is to compile the best information (and most valuable to my site’s users) from the most informative places around the web on commonly used terms in Search Engine Optimization, Marketing, and more. My end goal is to be a highly-valuable resource for all-things SEO-related.
Above the Fold – A traditional marketing term used to describe the top portion of a newspaper. On the web (or in email), it means the area of content in the browser viewable before scrolling.
Absolute Link – A hyperlink which shows the full URL of the page being linked to. For example, some links only show relative link [link] paths instead of having the entire reference URL.
Example of Absolute link:
Example of Relative link [link]:
Affiliate Marketing – Affiliate marketing programs allows merchants to expand their market reach and mindshare by paying independent agents on a cost per action (CPA) basis. Affiliates only get paid if visitors complete an action.
Most affiliates make next to nothing because they are not aggressive marketers, have no real focus, fall for wasting money on instant wealth programs that lead them to buying a bunch of unneeded garbage via other’s affiliate links, and do not attempt to create any real value.
Some power affiliates make hundreds of thousands or millions of dollars per year because they are heavily focused on automation and/or tap large traffic streams. Typically niche affiliate sites make more per unit effort than overtly broad ones because they are easier to focus (and thus have a higher conversion rate).
Some quality affiliate marketing networks:
Age – Domain age [link;], page age, user account age, and related historical data is sometimes taken into account when determining how much to trust that a particular person, website, page, or document. Generally, older domains are more trusted and therefore deemed more authoritative. However, sometimes search engines, like blog search engines, may boost the relevancy of newer documents.
Sometimes, fresh content may temporarily rank better than you might expect, because of citations from other channels which place the content or article on their home page or a well trusted (high PageRank) page on their site. After those sites publish more content and the referenced page falls deeper into their archives, those links typically become less authoritative causing a drop in rank.
Also see: Supplemental Index [link: ]
Alexa – California, U.S.-based subsidiary company of Amazon.com [link:] that is known for its toolbar and website. Once installed, the toolbar collects data on browsing behavior which is transmitted to the website where it is stored and analyzed and is the basis for the company’s web traffic reporting.
Algorithm – An algorithm is a set of finite, ordered steps for solving a mathematical problem. Each Search Engine uses a proprietary algorithm set to calculate the relevance of its indexed web pages to your particular query [link]. The result of this process is a list of sites ranked in the order of relevance as deemed by each particular search engine.
It is important to note:
- Each search engine has a different algorithm that may weigh various factors differently.
- Search engine algorithms are closely guarded in order to prevent exploitation of algorithmic results.
- Search algorithms are also changed frequently to incorporate new data and improve relevancy.
AlltheWeb – Search engine which was created by Fast, then bought by Overture [link], which was bought by Yahoo! [link]. Yahoo! may use AllTheWeb as a test bed for new search technologies and features.
Alt Attribute/Alt Tag – An Alt Attribute (or Alt Tag) is a way for you to give descriptions to your site’s images. Search engines aren’t able to look at an image and tell what it is about. Therefore, it is very important that you provide descriptions within the Alt Attribute so that they can scan, index, and rank your image accordingly.
To a search engine webcrawler and image represents a blank space on the canvas of a website. So by giving each image a description, you provide search engines with a frame of reference to understand what each image on your site is about.
In addition, Alt Tags are necessary from a usability perspective. Blind people (and people with other disabilities) are not able to easily distinguish what is in an image. Using an image alt attribute allows you to help screen readers and search engines understand the function of an image by providing a text equivalent for the object.
Example of an Alt Attribute:
AltaVista – Search engine bought out by Overture (although Overture was later bought by Yahoo). AltaVista was an early powerhouse in searching, but on October 25, 1999 they did a major algorithmic upgrade which caused them to dump many websites. Ultimately the upgrade and brand mismanagement drove themselves toward irrelevant class and they loss market and mind share.
Amazon.com – Amazon.com is the largest internet retailing website. Amazon.com is rich in consumer generated media. Amazon also owns a number of other popular websites, including IMDB [link:www.imdb.com] and Alexa [link:].
Anchor Text – The text that a user clicks on when following a hyperlink. Quality anchor text within inbound links can carry tons of SEO value back to your site. Search engines assume that your page is authoritative for the words that people include in links pointing at your site. When links occur naturally they typically have a wide array of anchor text combinations.
Example of anchor text:
Beware: Too much similar anchor text may be a considered a sign of manipulation, and thus discounted or filtered. Make sure when you are building links that you control that you try to mix up your anchor text. Outside of your core brand terms if you are targeting Google, you probably do not want any more than 10% to 20% of your anchor text to be the same.
Note: In the case the link is an image, the image alt attribute may act in the place of anchor text.
AOL – Stands for America Online. AOL is a popular web portal which merged with Time Warner.
API – Application Program Interface. A series of conventions or routines used to access software functions. Most major search products have an API program.
Arbitrage – Exploiting market inefficiencies by buying and reselling a commodity for a profit. As it relates to the search market, many thin content sites laced with an Overture [link: ] feed or AdSense [link: ] ads buy traffic from the major search engines and hope to send some percent of that traffic clicking out on a higher-priced ad. Shopping search engines generally draw most of their traffic through arbitrage.
- Wolf-Howl: AdSense Arbitrage: Tips, Tricks & Secrets.
- Jeremy Shoemaker interviews Kris Jones and part 2.
- Wikipedia: arbitrage
Ask – Ask is a search engine owned by InterActive Corp. They were originally named Ask Jeeves, but they dumped Jeeves in early 2006. Their search engine is powered by the Teoma.
ASP – Active Server Pages. a dynamic Microsoft programming language.
Authority – The ability of a page or domain to rank well in search engines. Five large factors associated with site and page authority are link equity [link:], site age [link:], traffic trends [link:], site history [link:], and publishing unique original quality content [link:].
In a general sense, search engines try to create algorithms that rank websites that are deemed the most relevant authority on a given subject or query. The goal of every website owner should be to posture their website presence properly, so that they can be perceived as an “Authority” on whatever subject or field they are targeting.
Search engines constantly tweak their algorithms to try to balance relevancy algorithms based on topical authority and overall authority across the entire web. Sites may be considered topical authorities or general authorities. For example, Wikipedia and DMOZ are considered broad general authority sites. Sites such as SEOBook or SEOMoz have topical authority on SEO, but not a broad general authority.
Automated Bid Management Software – Used by Pay per click search engines. Some search engines and third party software developers have created software which makes it easier to control your ad spend. Some of the more advanced tools can integrate with your analytics program and help you focus on conversion, ROI, and earnings elasticity instead of just looking at cost per click.
If you want to program internal bid management software you can get a developer token to use the Google Adwords. A few popular bid management tools are:
Backlink – see Inbound Link [link: ]
Bad Request – The request could not be understood by the server due to malformed syntax. The client should now repeat the request without modifications.
Bait and Switch – Marketing technique where you make something look overtly pure or as though it has another purpose to get people to believe in it or vote for it (by linking at it or sharing it with friends), then switch the intent or purpose of the website after you gain authority [link: ].
It is generally easier to get links to informational websites than commercial sites. Some new sites might gain authority much quicker if they tried looking noncommercial and gaining influence before trying to monetize their market position.
Banned – Also known as delisted or blacklisted, a banned site is a URL that has been removed from a search engine’s Index, typically for engaging in Black Hat SEO. Banned sites are ignored by search engines.
Banner Ad – A banner ad is a rectangular graphic advertisement. Banner ads are one of the commonest forms of online advertising. Their sizes vary, but most measure 468 pixels wide by 60 pixels high. Clicking on a banner ad will direct you to the advertiser’s website or a designated Landing Page.
Example of banner ad: [image:]
Banner Blindness – Banner blindness is a usability phenomenon in which a website visitor completely overlooks a banner. Such a banner may either be an advertising banner from an external site, or a banner that the serving site intends to use to promote content or a navigation link.
During the first web boom many businesses were based on eyeballs more than actually building real value. Many ads were typically quite irrelevant and web users learned to ignore the most common ad types.
In many ways text ads are successful because they are more relevant and look more like content, but with the recent surge in the popularity of text ads some have speculated that in time people may eventually become text ad blind as well.
Behavioral Targeting – Ad targeting based on past recent experience and/or implied intent. For example, if you recently searched online for restaurants then are later reading through another website, the page may still show you restaurant ads.
Best of the Web – Best of the Web Directory is a commercial web directory founded in 1994 providing websites categorized topically and regionally. BOTW allows site owners to submit their websites for an expedited review, and commercial sites are required to pay for the review.
Bid Management Software – see Automated Bid Management Software [link: ]
Bing – Formerly Live Search, Windows Live Search, MSN Search, Bing is the current web search engine (advertised as a “decision engine”) from Microsoft [link:]. Unveiled by Microsoft CEO Steve Ballmer on May 28, 2009 at the All Things Digital conference in San Diego, Bing is a replacement for Live Search. It went fully online on June 3, 2009, with a preview version released on June 1, 2009. In its first few weeks Bing was successful in gaining some market share.
Black-Hat SEO – Refers to SEO tactics that are deemed by search engines as deceptive in nature. Search engines set up guidelines that help them extract billions of dollars of ad revenue from the work of publishers and the attention of searchers.
Within that highly profitable framework, search engines consider certain marketing and optimization techniques deceptive in nature, and label them as Black-Hat SEO. Those tactics which are considered within their guidelines are called White-Hat SEO [link:] techniques.
Note: The search guidelines are not a static set of rules, and things that may be considered legitimate one day may be considered deceptive the next. Search engines are not without flaws in their business models, but there is nothing immoral or illegal about testing search algorithms to understand how search engines work.
As long as you’re not trying to be tricky or deceptive, you should be okay. A good rule of thumb to avoid accidentally using black-hat practices is to design the interface and content of your website with a user-centered focus first and a search-engine focus second.
Blacklisted – Also known as banned [link:] or delisted, a blacklisted site is a URL that has been removed from a search engine’s Index, typically for engaging in Black Hat SEO. Blacklisted sites are ignored by search engines.
Blog – Short for web-log. A periodically updated online journal typically formatted in reverse chronological order. Many blogs not only archive and categorize information, but also provide a feed and allow simple user interaction like leaving comments on the posts.
Most blogs tend to be personal in nature; however more and more businesses are beginning to add and maintain blogs as core online competencies. Blogs can generally quite authoritative with heavy link equity because they give people a reason to frequently come back to their site, read their content, and link to whatever they think is interesting.
Some popular blogging platforms:
Blog Comment Spam – Either manually or automatically (via a software program) adding low value or no value comments to other sites.
Automated blog spam:
Discreet Adult XXX Hardcore Porn Free Shipping
Manual blog spam:
douchebag manual spammer (usually keywords used in name)
Blogger – (1) Blogger is a free blog platform owned by Google [link:]. It allows you to publish sites on a subdomain off of Blogspot.com, or to FTP content to your own domain. Blogger is probably the easiest blogging software tool to use, but it lacks many some features present in other blog platforms.
Note: If you are serious about building a brand or making money online you should publish your content to your own domain, because it can be hard to reclaim a website’s link equity and age related trust if you have built years of link equity into a subdomain on someone else’s website. By default, Blogger blogs go to http://somename.blogspot.com as a URL format. As stated above, this is not optimal for SEO and you should try to get the content to post to your own site.
(2) A blogger is someone (anyone) who posts to their own weblog. This could be you!
Blogroll – Link list on a blog, usually linking to other blogs owned by the same company, friends of that blogger, or blogs that a blogger likes. A blogroll can be used to do link exchanges [link:].
Bold – A way to make words appear in a bolder font. Words that appear in a bolder font are more likely to be read by humans that are scanning a page.
A search engine may also place slightly greater weighting on these words than regular text; but if you write natural page copy and a word or phrase appears on a page many times it probably does not make sense or look natural if you bold ever occurrence. Word to the wise: don’t overuse this as an SEO technique. It will look bad from a user perspective and will saturate the effect from an SEO perspective.
Bookmarks – Bookmark is a feature in most Internet browsers that allows you to save the address of a web site you like. In Microsoft Internet Explorer [link:] bookmarks are known as favorites and a menu option is available for you to view your favorites in the Microsoft Internet Explorer browser. Most browsers [link:] come with the ability to bookmark your favorite pages. Many web based services have also been created to allow you to bookmark.
Boolean Search – A search allowing the inclusion or exclusion of documents containing certain words through the use of operators such as AND, NOT and OR. Search engines like Google and Yahoo make standard use of the AND variable.
- A Google search for Agent SEO will return results for Agent AND SEO.
- A Google search for “Agent SEO” will return results for the phrase Agent SEO.
- A Google search for Agent SEO -Jake will return results containing Agent AND SEO but NOT Jake.
- A Google search for ~SEO -SEO will find results with words related to SEO that do not contain SEO.
- “Boolean Searching on the Internet” via InternetTutorials.net
Brand – The emotional response associated with your company and/or products. A brand is built through controlling customer expectations and the social interactions between customers. Building a brand is what allows you to move away from commodity based pricing and move toward higher margin value based pricing.
- “Best Global Brands 2009” via Interbrand.com
- “40 of the Best Twitter Brands and the People Behind Them” via Mashable.com
Branded Keywords – Keywords or keyword phrases associated with a brand [link:]. Typically branded keywords occur late in the buying cycle [link:], and are some of the highest value and highest converting keywords.
Some affiliate marketing programs prevent affiliates from bidding on the core brand related keywords, while others actively encourage it. Either way can work depending on your business model and marketing savvy, but it is important to ensure there is synergy between internal marketing and affiliate marketing programs.
Breadcrumb Navigation – Navigational technique used to help search engines and website users understand the relationship between pages and website hierarchy.
Example of breadcrumb navigation structure:
Bridge – An entry into a site other than the homepage. It may be a legitimate landing page, used to measure the results of a specific promotion or campaign. But such pages are often created exclusively to spam search engine results, or misrepresent a site’s content. Search engines frown heavily on the use of doorway pages, and may penalize sites if the practice is reported or discovered.
- Doorway pages [link:]
Broken Link – A hyperlink that is not functioning, or a link which does not lead to the desired location. Links may broken for a number of reasons, but four of the most common reasons are:
- a website going offline
- linking to content which is temporary in nature (due to licensing structures or other reasons)
- linking to a page or site that has moved to another location
- linking to a site whose root domain name has changed
Most large websites have some broken links, but if too many of a site’s links are broken it may be an indication of outdated content, and it may provide website users with a poor user experience. Both of which may cause search engines to rank a page as being less relevant.
If you’re running a WordPress blog or website, I recommend using the Broken Link Checker plugin to monitor the status of your links.
Browser – Application used to view the World Wide Web. Some popular browsers include:
- Internet Explorer [link:]
- Firefox [link:]
- Safari [link:]
- Opera [link:]
Buying Cycle – Before making any purchase of goods/services or ideas, customers typically go through a buying cycle, firstly showing interest, conducting research and then making a purchase decision. The time it takes to complete the buying cycle could vary from time to time depending on the complexity of the item being purchased or the price involved.
Cache – Copy of a web page stored by a search engine. When you search the web you are not actively searching the whole web, but are searching files in the search engine index [link:]. Some search engines provide links to cached versions of pages in their search results, and allow you to strip some of the formatting from cached copies of pages.
Canonical URL – Many content management systems are configured with errors which cause duplicate or exceptionally similar content to get indexed under multiple URLs. Many webmasters use inconsistent link structures throughout their site that cause the exact same content to get indexed under multiple URLs.
The canonical version of any URL is the single most authoritative version indexed by major search engines. Search engines typically use PageRank or a similar measure to determine which version of a URL is the canonical URL.
Webmasters should use consistent linking structures throughout their sites to ensure that they funnel the maximum amount of PageRank at the URLs they want indexed. When linking to the root level of a site or a folder index it is best to end the link location at a / instead of placing the index.html or default.asp filename in the URL.
Examples of URLs which may contain the same information in spite of being at different web addresses:
Click-Fraud – The illegal practice of manipulating Cost-Per-Click (CPC)[link:] or Pay-Per-Click (PPC)[link:] revenue sharing agreements. There are numerous types of click fraud, but in a typical scenario the webmaster of a site that earns money from each click of the advertising links it publishes pays individuals a small fee to click those links.
Companies thus pay for advertising to clients who had no intention of buying from them. Some companies have filed class action lawsuits alleging that ad publishers such as Google[link:] and Yahoo![link:] have failed to aggressively confront click fraud because they benefit from increased CPC [link:] revenue.
Click-Through – Refers to a single instance of a user clicking on an advertising link or site listing and moving to a Landing Page. A higher Click-Through Rate (CTR)[link:] is one of the primary goals of Search Engine Optimization[link:].
Click-Through-Rate (CTR) – The percentage of users who click on an advertising link or search engine site listing out of the total number of people who see it, i.e. four click-throughs out of ten views is a 40% CTR.
Cloaking – Displaying different content to search engines and searchers. Depending on the intent of the display discrepancy and the strength of the brand of the person / company cloaking, it may be considered reasonable or it may get a site banned from a search engine. Cloaking is typically done to achieve a higher search engine position or to trick users into visiting a site. In such cases cloaking is considered to be Black Hat SEO and the offending URL could be Blacklisted [link:].
Cloaking has many legitimate uses which are within search guidelines. For example, changing user experience based on location is common on many popular websites.
- The Definitive Guide to Cloaking – Dan Kramer’s guide to cloaking. I also interviewed Dan here.
Clustering – In SERPs[link:], clustering is limiting each represented website to one or two listings. This is done to make the search results appear neat and organized, as well as to ensure diversity amongst the top ranked search results.
Code Bloat – Code bloat is the production of code that is perceived as unnecessarily long, slow, or otherwise wasteful of resources. Code bloat can also be caused by inadequacies in the language in which the code is written, or inadequacies in the compiler used to compile the language.
Content Management System (CMS) – Tool used to help make it easy to update and add information to a website. Blog software programs such as WordPress[link:] are some of the most popular content management systems currently used on the web. In short, a good CMS will make it easier for non-HTML savvy users to create and maintain a website or blog.
Beware: Many content management systems have errors associated with them which make it hard for search engines to index content due to issues such as duplicate content.
Content-Rich Doorway – A doorway page dressed up with graphics, navigation, and linked to from a site map so that it looks like a normal part of a Web site. The copy is written to rank for a single keyword expression.
Contextual Link-Inventory (CLI) – Search engines/advertising networks use their contextual link inventory to match keyword-relevant text-link advertising with site content. CLI is generated based on listings of website pages with content that the ad-server deems a relevant keyword match. Ad networks further refine CLI relevancy by monitoring the Click-Through Rate[link:] of the displayed ads.
Conversion – Conversion is the term used for any significant action a user takes while visiting a site, i.e. making a purchase, requesting information, or registering for an account.
Conversion Rate – The key metric to evaluate the effectiveness of a conversion effort (accepting a free gift, setting an appointment), reflecting the percentage of people converted into buyers (or subscribers, or whatever action is desired) out of the total population exposed to the conversion effort.
For Web sites, the conversion rate is the number of visitors who took the desired action divided by the total number of visitors in a given time period (typically, per month). For example, the conversion rate of visitors that subscribe to a newsletter = number of visitors divided by number of subscribers. If a website has 10,000 visitors and 500 subscribe, the conversion rate equals 1 out of 20 (or 5%).
Cookie – Small data file written to a user’s local machine to track them. Cookies are used to help websites customize your user experience, can help your browser remember passwords, track search history, help affiliate program managers track conversions, and much more.
Cost-Per-Action (CPA) – In a cost-per-action advertising revenue system, advertisers are charged a conversion-based fee (i.e. each time a user buys a product, opens an account, or requests a free trial). Many affiliate marketing programs and contextual ads are structured on a cost per action basis.
Cost-Per-Click (CPC) – See Pay-Per-Click Advertising [link: ]
Cost-Per-Impression – See Cost Per Thousand [link:]
Cost-Per-Thousand (CPM) – Also known as Cost-Per-Impression [link:] or CPM for cost-per-mille (mille is the Latin word for thousand), cost-per-thousand is an advertising revenue system used by search engines and ad networks in which advertising companies pay an agreed amount for every 1,000 users who see their ads, regardless of whether a click-through or conversion is achieved. CPM is typically used for Banner Ad [link:] sales, while Cost-Per-Click [link:] is typically used for text link advertising.
Crawlability – Refers to the ability of a search engine to crawl through the entire text content of your website, easily navigating to every one of your web pages, without encountering an unexpected dead-end or chained redirects.
Crawler – See Web Crawler [link:]
Crawl Depth – How deeply a website is crawled, and how many pages are indexed (or what percentage of the site’s pages).
Since searches which are longer in nature tend to be more targeted, it is important to try to get most or all of a site indexed such that the deeper pages have the ability to rank for relevant long-tail keywords [link:].
A large site needs adequate link equity [link:] to get deeply indexed. Another thing which may prevent a site from being fully indexed is duplicate content [link:] issues.
Crawl Frequency – How frequently a website is crawled. Sites which are well trusted or frequently updated may be crawled more frequently than sites with low Trust Scores [link:] and limited Link Authority [link:].
On the flip side, websites with highly artificial link authority scores (ie: mostly low quality spammy links) or sites which are heavy in duplicate content [link:] or near duplicate content (such as affiliate feed sites) may be crawled less frequently than sites with unique content which are well integrated into the web.
Crawl Page – A document consisting of links to other pages, provided for the sole purpose of giving crawlers (robots) links to follow. Spammers used to submit these pages to the search engines en masse.
CSS – Stands for Cascading Style Sheets, which is a method for adding styles to web documents. Using external CSS files makes it easy to change the design of many pages by editing a single file. You can link to an external CSS file using code similar to the following in the head of your HTML documents:
<link rel=”stylesheet” href=”http://www.seobook.com/style.css” type=”text/css” />
Tip: Trying to stay away from using inline CSS styling, as it causes Code Bloat [link:], which is bad for SEO.
- W3C: CSS – official guidelines for CSS.
- CSS Zen Garden – examples of various CSS layouts.
- Glish.com – examples of various CSS layouts, links to other CSS resources.
- CSS Elite – showcases the best in CSS design & development.
- CSS Remix – gallery of cool CSS designs.
- Best Web Gallery – excellent gallery of CSS designs.
- The Best Designs – displays some of the best CSS designs.
Cybersquatting – Registering domains related to other trademarks or brands in an attempt to cash in on the value created by said trademark or brand.
Dead Link – A link which is no longer functional. Most large, high-quality websites have at least a few dead links in them, but the ratio of good links to dead links can be seen as a sign of solid information quality.
Dedicated Server – Server which is limited to serving one website or a small collection of websites owned by a single person. Dedicated servers tend to be more reliable than shared (or virtual) servers [link:]. Dedicated servers usually run from $100 to $500 a month. Virtual servers typically run from $5 to $50 per month.
Deep Link – A link which points to an interior page within a website. When links grow naturally typically most high quality websites have many links pointing at many different interior pages.
When you request links from other websites it makes sense to request a link from their most targeted relevant page to your most targeted relevant page.
Del.icio.us – Highly popular social bookmarking website.
Delisting – A URL that has been removed from a search engine’s index [link:]. Delisted sites are ignored by search engines. A site may become delisted for many reasons:
- Engaging in Black-Hat SEO[link:] tactics
- Pages on new websites (or sites with limited link authority relative to their size) may be temporarily de-indexed until the search engine does a deep spidering and re-cache of the web.
- Pages which have changed location and are not properly redirected, or pages which are down when a search engine tries to crawl them may be temporarily de-indexed.
- If a website tripped an automatic spam filter it may return to the search index anywhere from a few days to a few months after the problem has been fixed.
- If a website is editorially removed by a human you may need to contact the search engine directly to request re-inclusion.
- During some updates search engines readjust crawl priorities.
Digg – Social news site where users vote (or “Digg”) on which stories get the most exposure and become the most popular.
Directory – A search site whose index is compiled by human editors (as opposed to web spiders). Although editors may pro actively include sites they consider to be of value, most inclusions are the result of submitted requests. The decision to include a site, and its subsequent ranking and categorization, is one of editorial judgment rather than being computed by an algorithm.
Some directories cater to specific niche topics, while others are more comprehensive in nature. Major search engines likely place significant weight on links from DMOZ[link:] and the Yahoo! Directory[link:]. Smaller and less established general directories likely pull less weight. If a directory does not exercise editorial control over listings, search engines will be far less likely to trust their links.
DMOZ – The Open Directory Project is the largest, most trusted human edited directory of websites. DMOZ is owned by AOL, and is primarily ran by volunteer editors. It is likely that search engines place significant weight on links from this directory.
DNS – Domain Name Server or Domain Name System. A naming scheme mechanism used to help resolve a domain name / host name to a specific TCP/IP Address.
DNS Propagation – When a new domain name is registered (or an existing one is transferred to a new DNS), the information must make its way around the entire internet. This process usually takes around 24 hours, during which time the domain will be inaccessible to many or all users.
Domain – Scheme used for logical or location organization of the web. Many people also use the word domain to refer to a specific website (common domain extensions include .com, .net, .org, .edu, and .gov).
Domain Age – See Age[link:]
Domain Authority – See Authority[link:]
Domain Mirror – A domain mirror for your website enables you to host two domain names using the same web content. Typically, a domain mirror will be used to cheaply register variants of a domain name (i.e. – the .com.au and .com) and host them both using the same website. They can also be used to assist in the case of changing your domain since the old domain can be mirrored to the new domain once everything is set up.
Doorway – A document with a small amount of text (usually coherent but sometimes gibberish) intended to rank well specifically for one targeted expression. In the old days, people created as many doorways as they had targeted keywords and search engines to work with.
Doorway Page – Pages designed to rank for highly targeted search queries, typically designed to redirect searchers to a page with other advertisements. Some webmasters cloak [link:] thousands of doorway pages on trusted domains, and rake in a boatload of cash until they are caught and delisted [link:].
If the page would have a unique purpose outside of search then search engines are generally fine with it, but if the page only exists because search engines exist then search engines are more likely to frown on the behavior.
Dreamweaver – Popular web development and editing software offering a WYSIWYG [link:] interface.
Visit Dreamweaver: Official Site
Duplicate Content – Content which is duplicate or near duplicate in nature. Search engines do not want to index multiple versions of similar content.
For example, printer friendly pages may be search engine unfriendly duplicates. Also, many automated content generation techniques rely on recycling content, so some search engines are somewhat strict in filtering out content they deem to be similar or nearly duplicate in nature.
- Duplicate Content Detection – video where Matt Cutts talks about the process of duplicate content detection
- Identifying and filtering near-duplicate documents
- Search Engine Patents on Duplicated Content and Re-Ranking Methods
- Stuntdubl: How to Remedy Duplicate Content
Dynamic Content – Content which changes over time or uses a dynamic language such as PHP to help render the page. In the past search engines were less aggressive at indexing dynamic content than they currently are. While they have greatly improved their ability to index dynamic content it is still preferable to use URL rewriting to help make dynamic content look static in nature.
Dynamic URL – A dynamic URL is the address of a Web page with content that depends on variable parameters that are provided to the server that delivers it.
The parameters may be already present in the URL itself or they may be the result of user input. A dynamic URL can often be recognized by the presence of certain characters or character strings that appear in the URL (visible in the address bar of your browser).
You might use URL parameters in your site to perform various functions apart from modifying the content of the page, such as,
- Session ids – for tracking user sessions
- Source trackers – for tracking the sources which are sending referrals to your pages and site
- Format modifiers – for print formats etc
Example of dynamic URL (with paramters bolded):
It is generally my belief that dynamic URLs, with their long strings of random numbers and variables are not as SEO-friendly as Static URLs [link:].
Editorial Links – Search engines count links as “votes” for a website’s quality. They primarily want to count editorial links – i.e. those links that were earned, rather than links that were bought or bartered.
Many paid links, such as those from quality directories [link: ], still count as signs of votes as long as they are also associated with editorial quality standards. If they are from sites without editorial control, like link farms [link: ], they are not likely to help you rank well. Using an algorithm similar to TrustRank [link: ], some search engines may place more trust on well known sites with strong editorial guidelines.
EMR – A search result that precisely matches a user’s search term.
Entry Page – The page on which a user enters your website. If you’re running pay-per-click[link: ] ads, it is very important to send visitors to the most appropriate and targeted page associated with the keyword they searched for (this strategy is called “deep linking” [link: ]). If you are doing link building it is important to point links at your most appropriate page when possible so that:
- If anyone clicks the link they are sent to the most appropriate/relevant page.
- You help search engines understand what the pages and content on your site are associated with.
Ethical SEO – Search engines like to paint SEO services which manipulate (or take advantage of) their relevancy algorithms as being unethical – i.e. any particular technique is generally not typically associated with ethics, but is either effective or ineffective.
Some search marketers lacking in creativity tend to describe services sold by others as being unethical while their own services are ethical. Any particular technique is generally not typically associated with ethics, but is either effective or ineffective.
The only ethics issues associated with SEO are generally business ethics related issues. Two of the bigger frauds are:
- Not disclosing risks: Some SEOs may use high risk techniques when they are not needed. Some may make that situation even worse by not disclosing potential risks to clients. These techniques may be referred to ask grey-hat or black-hat [link:] SEO.
- Taking money & doing nothing: Many people selling SEO services may not actually know how to competently provide them. Some shady people claim to be SEO’s and bilk money out of unsuspecting small businesses.
As long as the client is aware of potential risks there is nothing unethical about being aggressive.
External Link – Link which references another domain. Some people believe in link hoarding [link:], but linking out to other related resources is a good way to help search engines understand what your site is about.
However, if you link out to lots of low quality sites or primarily rely on low quality reciprocal links some search engines may not rank your site very well. Search engines are more likely to trust high quality editorial links [link: ] (both to and from your site).
Facebook – Facebook is a highly-popular social networking [link: ] website that is operated and privately owned by Facebook, Inc. Users can add friends and send them messages, and update their personal profiles to notify friends about themselves. Additionally, users can join networks organized by city, workplace, and school or college.
A January 2009 Compete.com [link: ] study ranked Facebook as the most used social network by worldwide monthly active users, followed by MySpace [link: ].
Favicon – Favorites Icon is a small icon which appears next to URLs in a web browser. Upload an image named favicon.ico in the root of your site to have your site associated with a favicon.
Feed – Many content management systems such as blogs, allow readers to subscribe to content update notifications via RSS [link: ] or XML feeds.
Feeds can also refer to pay per click syndicated feeds, or merchant product feeds. Merchant product feeds have become less effective as a means of content generation due to improving duplicate content filters.
Feed Aggregator – See Feed Reader [link: ]
Feed Reader – Software or website used to subscribe to feed update notifications.
- Google Reader – Google’s popular web-based feed reader.
- Bloglines – Popular web-based feed reader.
- My Yahoo! – Allows you to subscribe to feed updates.
- Feed Demon – Desktop-based feed reader.
FFA – Free for all pages are pages which allow anyone to add a link to them. Generally these links do not pull much weight in search relevancy algorithms because many automated programs fill these pages with links pointing at low quality websites.
Firefox – Popular extensible open source web browser from Mozilla.
Download Mozilla’s Firefox
Flash – Vector graphics-based animation software which makes it easier to make websites look rich and interactive in nature.
Search engines tend to struggle indexing and ranking flash websites because flash typically contains so little relevant content. If your website uses flash, please ensure that:
- You embed flash files within HTML pages
- You use a noembed element to describe what is in the flash
- You publish your flash content in multiple separate files such that you can embed appropriate flash files in relevant pages
- You supplement the flash with noscript [link: ] content
Frames – A technique created by Netscape [link: ] used to display multiple smaller pages on a single display. This web design technique allows for consistent site navigation, but makes it hard to deep link at relevant content.
Given the popularity of Server Side Includes, Content Management Systems [link: ], and Dynamic Languages [link: ] there really is no legitimate reason to use frames to build a content site today.
- iFrames [link: ]
Fresh Content – Content which is dynamic in nature and gives people a reason to keep paying attention to your website. Blogs [link: ] are a good example of fresh content.
Many SEOs talk up fresh content, but fresh content does not generally mean re-editing old content. It more often refers to creating new content.
The primary advantages to fresh content are:
- Maintain and grow mindshare: If you keep giving people a reason to pay attention to you more and more people will pay attention to you, and link to your site.
- Faster idea spreading: If many people pay attention to your site, when you come out with good ideas they will spread quickly through such things as social networking [link: ].
- Growing archives: If you are a content producer then owning more content means you have more chances to rank. If you keep building additional fresh content eventually that gives you a large catalog of relevant content.
- Frequent crawling: Frequently updated websites are more likely to be crawled frequently [link: ].
FTP – File Transfer Protocol is a protocol for transferring data between computers. Many content management systems (such as blogging platforms) include FTP capabilities.
Web development software such as Dreamweaver also comes with FTP capabilities. There are also a number of free or cheap FTP programs such as Cute FTP, Core FTP, and Leech FTP.
Fuzzy Search – Search which will find matching terms when terms are misspelled (or fuzzy).
Fuzzy search technology is similar to stemming technology, with the exception that fuzzy search corrects the misspellings at the users end and stemming searches for other versions of the same core word within the index.
Gateway Page – Also known as a doorway page or jump page, a gateway page is a URL with minimal content designed to rank highly for a specific keyword and redirect visitors to a homepage or designated Landing Page.
Some search engines frown on gateway pages as a softer form of Cloaking or Spam. However, gateway pages may be legitimate landing pages designed to measure the success of a promotional campaign, and they are commonly allowed in Paid Listings.
- Doorway Pages [link: ]
Geographical Targeting – Geographical targeting is the focusing of Search Engine Marketing efforts on specific states, counties, cities and neighborhoods that are important to a company’s business.
One basic aspect of geographical targeting is adding the names of relevant cities or streets to a site’s keywords, i.e. Hyde Street Chicago apartments. Another important element of geo-targeting is increasing your site’s presence on local search engines such as Google Maps [link: ] and/or Yahoo Local [link: ].
Geographical Segmentation – Geographic segmentation is the use of analytics to categorize a site’s web traffic by the physical locations from which it originated.
Google – The world’s leading search engine in terms of reach. Google was created in 1996 by Stanford students Larry Page and Sergey Brin, and actually started as a research project.
Since its incorporation in 1998, Google has been the on the cutting edge of the search industry. They’ve been famously innovative, pioneering such things as PageRank [link: ] and Google Earth [link: ], as well as acquiring media giants YouTube [link: ] and DoubleClick [link: ].
Search with Google
Googlebot – Google’s search engine spider [link: ] (or webcrawler [link: ]). Google has a shared crawl cache between their various spiders, including vertical search spiders and spiders associated with ad targeting.
Googleplex – Nickname for Google’s headquarters in Mountain View, California.
Google AdWords – Google’s advertisement and link auction network. Most of Google’s ads are keyword targeted and sold on a cost per click basis in an auction which factors in ad clickthrough rate as well as max bid.
Adwords is also a program that determines the advertising rates and keywords used in the Google AdSense [link] program. Advertisers bid on the keywords that are relevant to their businesses. Ranked ads then appear as sponsored links on Google Search Engine Results Pages (SERPS) and Google AdSense host sites.
- Google Adwords [link: ]
- Google Adwords Keyword Tool [link: ]
Google AdSense – Google’s contextual advertising [link: ] network. Publishers large and small may automatically publish relevant advertisements near their content and share the profits from those ad clicks with Google.
AdSense offers a highly scalable automated ad revenue stream which will help some publishers establish a baseline for the value of their ad inventory. In many cases AdSense will be underpriced, but that is the trade off for automating ad sales.
AdSense ad auction formats include:
- Cost Per Click (CPC) [link: ] – advertisers are only charged when ads are clicked on
- Cost Per Impression (CPM) [link:] – advertisers are charged a certain amount per ad impression. Advertisers can target sites based on keyword, category, or demographic information.
AdSense ad formats include:
- Animated Graphics
In some cases, I have seen ads which got a 2 or 3% click through rate (CTR); while sites that are optimized for maximum CTR (through aggressive ad integration) can obtain as high as a 50 or 60% CTR depending on:
- How niche their website is
- How commercially oriented their website is
- The relevancy and depth of advertisers in their vertical
It is also worth pointing out that if you are too aggressive in monetizing your site before it has built up adequate authority your site may never gain enough authority to become highly profitable. Depending on your vertical your most efficient monetization model may be any of the following:
- Affiliate Marketing [link:]
- Direct Ad Sales
- Selling unique products and services
- Mixture of all of the above
- Google AdSense Program – sign up as an ad publisher.
- Google AdSense Heat Map – shows ad clickthrough rate estimates based on ad positioning.
- Google AdWords – buy ads on Google search and / or contextually relevant web pages.
Google Analytics – A free service offered by Google that generates detailed statistics about the visitors to a website. Its main highlight is that the product is aimed at marketers as opposed to webmasters and technologists from which the industry of web analytics originally grew.
It is the most widely used website statistics service, currently in use at around 40% of the 10,000 most popular websites. Google Analytics can track visitors from all referrers, including search engines, display advertising, pay-per-click networks, email marketing and digital collateral such as links within PDF documents.
Visit Google Analytics
Google Base – Free database of semantically structured information created by Google. Google Base may also help Google better understand what types of information are commercial in nature, and how they should structure different vertical search products.
In addition, Google base is an excellent way to get your product inventory to show up within Google’s search engine results pages.
Visit Google Base
Google Bombing – Making a pank rank well for a specific search query by pointing hundreds or thousands of links at it with the keywords in the anchor text.
Google Bowling – Knocking a competitor out of the search results by pointing hundreds or thousands of low-trust, low-quality links at their website.
Typically it is easier to bowl new sites out of the results. Older established sites are much harder to knock out of the search results.
Google Checkout – Payment service provided by Google which helps Google better understand merchant conversion rates and the value of different keywords and markets.
Google Dance – In the past Google updated their index roughly once a month. Those updates were named Google Dances, but since Google shifted to a constantly updating index, Google no longer does what was traditionally called a Google Dance.
Major search indexes are constantly updating. Google refers to this continuous refresh as Everflux [link:].
The second meaning of Google Dance is a yearly party at Google’s corporate headquarters which Google holds for search engine marketers. This party coincides with the San Jose Search Engine Strategies conference.
- Matt Cutts Google Terminology Video – Matt talks about the history of Google Updates and the shift from Google Dances to everflux.
Google Earth – Google Earth is a virtual globe, map and geographic information program that was originally called EarthViewer 3D, and was created by Keyhole, Inc, a company acquired by Google in 2004.
It maps the Earth by the superimposition of images obtained from satellite imagery, aerial photography and GIS 3D globe.
Visit Google Earth
Google Keyword Tool – A keyword research tool provided by Google which estimates the competition and search volume for a keyword, recommends related keywords, and will tell you what keywords Google thinks are relevant to your site or a page on your site.
Try the Google Keyword Tool
Google Maps – Google Maps (for a time named Google Local) is a basic web mapping service application and technology provided by Google, free (for non-commercial use), that powers many map-based services, including the Google Maps website, Google Ride Finder, Google Transit, and maps embedded on third-party websites via the Google Maps API.
It offers street maps, a route planner for traveling by foot, car, or public transport, and an urban business locator for numerous countries around the world.
Visit Google Maps
Google OneBox – Portion of the search results page above the organic search results which Google sometimes uses to display local and vertical search results from Google Maps [link:], Google News, Google Base [link:], and other Google owned vertical search services.
Google Supplemental Index – Index where pages with lower trust scores are stored. Pages may be placed in Google’s Supplemental Index if they consist largely of duplicate content, if the URLs are excessively complex in nature, if the site which hosts them lacks significant trust, or if they are new and still untrusted by Google.
Google Traffic Estimator – Tool which estimates bid prices and how many Google searchers will click on an ad for a particular keyword.
If you do not submit a bid price the tool will return an estimated bid price necessary to rank #1 for 85% of Google’s queries for a particular keyword.
Try Google Traffic Estimator
Google Trends – Tool which allows you to see how Google search volumes for a particular keyword change over time.
Try Google Trends
Google Webmaster Tools – Google Webmaster Tools is a no-charge web service by Google for webmasters. It allows webmasters to check indexing status and optimize visibility of their websites.
It has tools that let the webmaster:
- Check and set the crawl rate.
- List internal and external pages that link to the site.
- See what keyword searches on Google led to the site being listed in the SERPs, and the click through rates of such listings.
- View statistics about how Google indexes the site.
- Submit and check a sitemap.
- Generate and check a robots.txt file.
- Set a preferred domain (i.e prefer agent-seo.com over www.agent-seo.com or vice versa).
Visit Google Webmaster Tools
Graphical Search Inventory – Non-text-based advertising that is displayed based on the relevance of surrounding content. Includes banners, pop-ups, toolbars and rich media.
Google Website Optimizer – Free multi variable testing platform used to help AdWords [link:] advertisers improve their conversion rates.
Try Google Website Optimizer
Gray-Hat SEO – Search Engine Optimization tactics that fall in between Black Hat SEO [link:] and White Hat SEO [link:]. Gray hat SEO techniques can be legitimate in some cases and illegitimate in others. Such techniques include Doorway Pages [link:], Gateway Pages [link:], Cloaking [link:] and duplicate content [link:].
Guestbook Spam – A type of low quality automated link which search engines do not want to place much trust on.
Heading Tags – The heading element briefly describes the subject of the section it introduces, and are generally used to denote places of importance on any given web page. If your website were a book, heading tags would closely resemble chapter headings.
Heading elements go from H1 to H6 with the lower numbered headings being most important (H1 = most important, H6 = least important).
An H1 element source would look like:
Heading elements may be styled using CSS [link:]. Many content management systems [link:] place the same content in the main page heading and the page title, although in many cases it may be preferential to mix them up if possible.
Tip: You should only use a single H1 element on each page, and may want to use multiple other heading elements to structure a document.
Hidden Text – SEO technique used to show search engine spiders text that human visitors do not see.
While some sites may get away with it for a while, generally the risk to reward ratio is inadequate for most legitimate sites to consider using hidden text.
Hits – This generally means all requests from a webserver including requests by a web browser for html pages, jpeg’s, gif’s and other images. Hits is an outdated phrase often thrown around when referring to website traffic, but is generally not very meaningful in quantitfying actual search engine traffic.
Homepage – The main page on your website, which is largely responsible for helping develop your brand and setting up the navigational schemes that will be used to help users and search engines navigate your website.
As far as SEO goes, a home page is typically going to be one of the easier pages to rank for some of your more competitive terms, largely because it is easy to build links at a home page.
You should ensure your homepage stays focused and reinforces your brand though, and do not assume that most of your visitors will come to your site via the home page. If your site is well structured many pages on your site will likely be far more popular and rank better than your home page for relevant queries.
Home Directory – The directory in which your site’s main index page is located. usually named /public_html/, or /www/ or /web/.
.htaccess – Apache directory-level configuration file which can be used to password protect or redirect files.
As a note of caution, make sure you copy your current .htaccess file before editing it, and do not edit it on a site that you can’t afford to have go down unless you know what you are doing.
HTML – HyperText Markup Language is the language in which pages on the World Wide Web are created.
Some newer web pages are also formatted in XHTML.
HTTP – HyperText Transfer Protocol is the foremost used protocol to communicate between servers and web browsers. Hypertext transfer protocol is the means by which data is transferred from its residing location on a server to an active browser.
Hubs – A document that links out to many other documents devoted to a single topic. Think of any category page in a major directory like Yahoo! or DMOZ. All the documents linked to are assumed to be authorities (sort of a circular logic).
Topical hubs are sites which link to well trusted within their topical community. A topical authority is a page which is referenced from many topical hub sites. A topical hub is a page which references many authorities.
- SEOBook’s Hub Finder
- Mike Grehan on Topic Distillation [PDF]
- John Klienberg’s Authoratative Sources in a Hyperlinked Environment [PDF]
- John Klienberg’s Homepage
Hyperlink – Also known as link or HTML link, a hyperlink is an image or portion of text that when clicked on by a user opens another web page or jumps the browser to a different portion of the current page.
Inbound links [link:] with keyword-relevant anchor text [link:] are an important part of SEO strategy.
Code for a hyperlink looks like this:
iFrames – An HTML structure that allows another HTML document to be inserted into an HTML page. The iFrame is set up as a window frame of a specified size that scrolls along with the rest of the page, but the iFrame’s content can itself be scrolled if it is larger than the iFrame window.
Unlike the regular HTML frames function, which is used to divide the screen into multiple windows, the iFrame is typically used to insert an ad or small amount of text in the middle of a page.
iFrames have been widely used within the automotive industry as a way to “frame in” their inventory which is typically generated and maintained by a 3rd-party source. Using iFrames is not a solid SEO practice, as it allows for content not actually residing on a particular website (thus giving the original website no credit for the content).
Inbound Link – Link pointing to one website from another website.
Most search engines allow you to see a sample of links pointing to a document by searching using the link: function. For example, using link:www.agent-seo.com would show pages linking to the homepage of this site (both internal links [link:] and inbound links). Due to canonical URL [link:] issues www.agent-seo.com and agent-seo.com may show different linkage data.
Google typically shows a much smaller sample of linkage data than competing engines do, but Google still knows of and counts many of the links that do not show up when you use their link: function.
Index – Collection of data used as bank to search through to find a match to a user fed query. The larger search engines have billions of documents in their catalogs.
When search engines search they search via reverse indexes by words and return search results based on matching relevancy vectors. Stemming and semantic analysis allow search engines to return near matches.
(2) The term ‘Index’ may also refer to the root of a folder on a web server.
Information Architecture – the art of expressing a model or concept of information used in activities that require explicit details of complex systems. Among these activities are library systems, content management systems [link:], web development, user interactions, database development, programming, technical writing, enterprise architecture, and critical system software design.
Good information architecture considers both how humans and search spiders access a website. Information architecture suggestions:
- Focus each page on a specific topic.
- Use descriptive page titles [link:] and meta descriptions [link:] which describe the content of the page
- Use clean (few or no variables) descriptive file names and folder names
- Use headings [link:] to help break up text and semantically structure a document
- Use breadcrumb navigation [link:] to show page relationships
- Use descriptive link anchor text [link:]
- Link to related information from within the content area of your web pages
- Improve conversion [link:] rates by making it easy for people to take desired actions
- Avoid feeding search engines duplicate or near-duplicate content [link:]
Internal Link – Link from one page on a site to another page on the same site.
It is preferential to use descriptive internal linking to make it easy for search engines to understand what your website is about. Use consistent navigational anchor text for each section of your site, emphasizing other pages within that section. Place links to relevant related pages within the content area of your site to help further show the relationship between pages and improve the usability of your website.
Internet – The Internet is a global system of interconnected computer networks that use the standard Internet Protocol Suite (TCP/IP) to serve billions of users worldwide. It is a network of networks that consists of millions of private and public, academic, business, and government networks of local to global scope that are linked by a broad array of electronic and optical networking technologies.
The Internet carries a vast array of information resources and services, most notably the inter-linked hypertext documents of the World Wide Web (WWW) and the infrastructure to support electronic mail.
Internet Explorer – Microsoft’s web browser. After they beat out Netscape’s browser on the marketshare front, they failed to innovate on any level for about 5 years – until Firefox [link:] forced them to.
Invisible Web – Portions of the web which are not easily accessible to crawlers due to search technology limitations, copyright issues, or information architecture [link:] issues.
IP Address – Internet Protocol Address. Every computer connected to the internet has an IP address. Some websites and servers have unique IP addresses, but most web hosts host multiple websites on a single host.
Many SEO’s refer to unique C class IP addresses. Every site is hosted on a numerical address like aa.bb.cc.dd. In some cases many sites are hosted on the same IP address. It is believed by many SEO’s that if links come from different IP ranges with a different number somewhere in the aa.bb.cc part, then the link may count more than links from the same local range and host.
ISP – Internet Service Providers sell end users access to the web. Some of these companies also sell usage data to web analytics companies.
Jump Page – An entry into a site other than the homepage. It may be a legitimate landing page, used to measure the results of a specific promotion or campaign. However, these pages are often created exclusively to spam search engine results, or misrepresent a site’s content. Search engines frown heavily on the use of these doorway pages [link:], and may penalize sites if the practice is reported and/or discovered.
Keyword – The words or phrases a person types into a search box. Also refers to the word or phrase a site owner wants to be found under.
Search Engine Results Pages (SERP’s) [link:] rank indexed sites according to how relevant they are in relation to the search query (most relevant to least relevant).
Long tail [link:] and brand [link:] related keywords are typically worth more than shorter and vague keywords because they typically occur later in the buying cycle and are associated with a greater level of implied intent.
Keyword Density – An old measure of search engine relevancy based on how prominent keywords appeared within the content of a page. Keyword density is no longer a valid measure of relevancy over a broad open search index though.
When people use keyword stuffed copy it tends to read mechanically (and thus does not convert well and is not link worthy); plus some pages that are crafted with just the core keyword in mind often lack semantically related words and modifiers from the related vocabulary (and that causes the pages to rank poorly as well).
The keyword density formula is the total number of keyword mentions divided by the total number of words on a page. Traditionally it has been said that keywords should fall between 2 and 8% density (although this is not a heavily-used factor anymore).
- The Keyword Density of Non Sense
- Keyword Density Analysis Tool from SEOBook
- “Search Engine Friendly Copywriting: What Does ‘Write Naturally’ Mean for SEO?” from SEOBook
Keyword Effectiveness Index (KEI) – Keyword effectiveness index: A mathematical representation of the popularity of a keyword compared to its popularity mesured as the number of pages in a search engines index.
KEI is a statistical formulation that reveals the most effective keyword phrases and terms to use in optimize your web pages for. Efficiency can be many things. According to KEI, it is efficient to optimize for keywords that have many searchers, but only a few competing pages.
The lower the KEI, the more popular your keywords are, and the less competition they have. That means that you might have a better chance of getting to the top in the search engines and receive a good number of searchers for your effort.
Example: Suppose the number of searches for a keyword is 821 per day and Google displays 224,234 results (pages) for that keyword. Then the ratio between the popularity and competitiveness for that keyword is:
Example of the opposite: Suppose the number of searches for a keyword is 2 per day and Google displays 11,224,234 results (pages) for that keyword. Then the ratio between the popularity and competitiveness for that keyword is:
Explanation: According to the KEI definition, the best keywords are those that have many searches and that don’t have much competition in the search results. A low KEI is therefore preferable.
Keyword Proximity – How close keywords are to each other on web pages.
Keyword Research – The process of discovering relevant keywords and keyword phrases to focus your SEO and PPC marketing campaigns on.
Examples of keyword research & discovery methods:
- Using keyword research tools [link:].
- Reviewing analytics data [link:] or server logs [link:].
- Analyzing ad copy and tags on competing websites.
- Reviewing consumer feedback.
- Interacting with customers to understand how/why the found and/or chose your business.
Keyword Research Tools – Tools which help you discover potential keywords based on past search volumes, search trends, bid prices, and page content from related websites.
Some popular keyword research tools:
Please note that most keyword research tools used alone are going to be highly inaccurate at giving exact quantitative search volumes. The tools are better for qualitative measurements. To test the exact volume for a keyword it may make sense to set up a test Google AdWords [link:] campaign.
Keyword Stuffing – Writing copy that uses excessive amounts of the core keyword(s) and is usually written in a very unnatural, robotic fashion.
When people use keyword stuffed copy it tends to read mechanically (and thus does not convert well and is not link worthy); plus some pages that are crafted with just the core keyword in mind often lack semantically related words and modifiers from the related vocabulary (and that causes the pages to rank poorly as well).
Landing Page – The landing page is the page on which a visitor “lands” after clicking a:
- search engine listing
- email link
- banner ad
- ppc ad
- other ad/link
The landing page can be a site’s homepage, but is usually a page designed to appeal to users who click-through from a specific ad or link. Similar to a doorway page, but a legitimate marketing function — it is used for counting and tracking arrivals and determining the effectiveness of a marketing campaign.
Well designed landing pages that are relevant to a user’s keyword query, will improve conversion rates, and play a critical role in Search Engine Marketing.
Landing Page Quality Scores – A measure used by Google to help filter noisy ads out of their AdWords [link:] program.
When Google AdWords [link:] launched, affiliates and arbitrage players made up a large portion of their ad market — as more mainstream companies have spent on search marketing, Google [link:] has done many measures to try to keep their ads relevant.
- Landing Page & Site Quality Guidelines
- Google Adwords Landing Page Quality Scores
- 9 Myths of Landing Page Quality Score
Latent Semantic Indexing (LSI) – An algebraic model of document retrieval based on a singular value decomposition of the vectorial space of index terms.
Link – See hyperlink [link:]
Link Baiting – The art of targeting, creating, and formatting information that provokes the target audience to point high quality links at your site. Many link baiting techniques are targeted at social media [link:] and bloggers [link:].
The types of link bait vary tremendously, but they include highly informative articles or news stories, useful resources and sometimes controversial or sensationalistic content.
Link baiting is a technique used to help a site improve its Link Popularity [link:] and Page Rank [link:]. Some sites use link baiting as the centerpiece of a website marketing campaign.
Link Building – The process of building high quality inbound links to and from your website. Over time, search engines will evaluate links pointing to your website in order to influence their decisions whether or not to trust that your website is authoritative, relevant, and trustworthy.
- SEOMoz: The Professionals Guide to Link Building — this guide from SEOMoz is only available to Pro users, but I can promise that it is totally worth the money!
- SEOMoz: Link Building from A to Z
- Build conceptually unique linkworthy high-quality content
- Create viral marketing ideas that want to spread and make people talk about you
- Mix your anchor text
- Get deep links
- Try to build at least a few quality links before actively obtaining any low quality links
- Register your site in relevant high quality directories such as DMOZ, the Yahoo! Directory, and Business.com
- When possible try to focus your efforts mainly on getting high quality editorial links
- Create link bait
- Try to get bloggers to mention you on their blogs
- It takes a while to catch up with the competition, but if you work at it long enough and hard enough eventually you can enjoy a self-reinforcing market position
- Filthy Linking Rich [PDF] – Mike Grehan article about how top rankings are self reinforcing
Quick link building tips:
Link Bursts – A rapid increase in the quantity of links pointing at a website.
When links occur naturally they generally develop over time. In some cases it may make sense that popular viral articles receive many links quickly, but in those cases there are typically other signs of quality as well, such as:
- Increased usage data
- Increase in brand related search queries
- Traffic from the link sources to the site being linked at
- Many of the new links coming from new pages on trusted domains
Link Churn – The rate at which a site loses links.
Link Equity – A measure of how strong a site is based on its inbound link popularity and the authority [link:] of the sites providing those links (i.e. the total number and overall quality of the links pointing to your website).
Link Exchange – A link exchange is a quid pro quo arrangement or reciprocal link [link:] exchange between two sites. Reciprocal links usually lead to the home page of the associate site.
Link Farm – Website or group of websites which exercises little to no editorial control when linking to other sites. FFA [link:] pages, for example, are link farms.
Generally, link hoarding is a bad idea for the following reasons:
- Many authority sites were at one point hub sites that freely linked out to other relevant resources
- If you are unwilling to link out to other sites, people are going to be less likely to link to your site
- Outbound links to relevant resources may improve your credibility and/or boost your overall relevancy scores
~ Matt Cutts, Google
- Why Paris Hilton Is Famous (Or Understanding Value In A Post-Madonna World) – article about how being a platform (ie: someone who freely links out) makes it easier to become an authority.
Link Popularity – The number and quality of links pointing at a website. Page Rank is achieved when backlinks [link:] are located on reputable, relevant sites rather than so-called Link Farms. Most search engines use link popularity as a factor in their algorithmic results.
For competitive search queries link quality counts much more than link quantity. Google typically shows a smaller sample of known linkage data than the other engines do, even though Google still counts many of the links they do not show when you do a link: search.
Link Reputation – The combination of your link equity [link:] and anchor text [link:].
Link Rot – A measure of how many and what percent of a website’s links are broken.
Links may broken for a number of reason, but four of the most common reasons are:
- A website going offline
- Linking to content which is temporary in nature (due to licensing structures or other reasons)
- Moving a page’s location
- Changing a domain’s content management system
Most large websites have some broken links, but if too many of a site’s links are broken it may be an indication of outdated content, and it may provide website users with a poor user experience. Both of which may cause search engines to rank a page as being less relevant.
- Xenu Link Sleuth – free software program which crawls websites to find broken links.
- Worpress Plugin: Broken Links Checker
Listings – see SERP [link: ].
Local Search – Local search refers to both the addition of geographical keywords (cities, streets, etc.) to search terms. Also refers to YellowPages-type search engines such as Google Maps [link: ], Yahoo! Local [link: ], etc. to find business services in a particular zip code.
Search engine placement services use local SEO to help traditional “brick and mortar businesses” connect with customers in their community.
Local Search Optimization – The process of increasing the amount of visitors to a website by ranking high for specific locality-based keyword phrases in search engines, plus the addition of geographical keywords in the search phrase (cities, streets, zip codes). Local SEO is a powerful tool attracting local customers, especially for local small businesses.
- Local SEO Guide: Local SEO blog
- Local Search Ranking Factors in 2009: Comprehensive guide from David Mihm
Log Files – Server files which show you what your leading sources of traffic are and what people are search for to find your website.
Log files DO NOT typically show as much data as analytics programs would — if they do, it is generally not in a format that is as useful beyond seeing the top few stats.
In my experience, when trying to compare log files to actual website analytics the waters are generally very muddy. Discrepancies between actual traffic and all server traffic should be expected, making this source of traffic a little unreliable.
Long-Tail Keywords – The phrase “The Long Tail (as a proper noun with capitalized letters)” was first coined by Chris Anderson in a 2004 Wired Magazine article to describe certain business and economic models such as Amazon.com or Netflix. The term long tail is also generally used in statistics, often applied in relation to wealth distributions or vocabulary use.
More recently the word long-tail is being used to describe a (niche) longer keyword focus while writing content for search engine optimization (i.e. a keyword phrase where several words are used to specify the search query).
- Why It Makes Sense To Target Long-Tailed Keywords First
- Long-Tail vs. Short-Tail Keywords
- Targeting Long-Tailed Keywords
- SEOMoz: Illustrating the Long-Tail
Meta Description Tag – The meta description tag is typically a sentence or two of content which describes the content of the page. Search engines may consider or display this tag at their discretion. Relevant meta description tags may appear in search results [link:] as part of the page description below the page title [link:].
Meta Description code is found between the <head></head> tags and looks like this:
A good Meta Description tag should:
- Be highly relevant to the content on the page.
- Serve to reinforce the page title.
- Focus on including offers and secondary keywords & phrases to help add context to the page title.
- Respect character limit best practices – 145-155 for homepage, and not exceeding 300 character spaces for all other pages.
- Not be used as a place to spam keywords.
- Free Meta Tag Generator – Free formatting tool that offers advice on creating meta tags.
Meta Keywords Tag – The meta keywords tag is a tag which can be used to highlight keywords and keyword phrases which the page is targeting.
Meta Keywords code is found between the <head></head> tags and looks like this:
Many people spammed meta keyword tags and searchers typically never see the tag, so most search engines do not place any weight on it. Many SEO professionals no longer bother to consider meta keywords tags as a serious SEO tactic.
- Free Meta Tag Generator – Free formatting tool that offers advice on creating meta tags.
Meta Refresh – A meta tag used to make a browser refresh to another URL location.
Meta Refresh code is found between the <head></head> tags and looks like this:
Generally in most cases it is preferred to use a 301 [link:] or 302 [link:] redirect over a meta refresh.
Meta Robots Tag – A meta robots tag (named for a search engine Crawler or Robot) lets page authors prevent their webpages from being added to a search engine’s Index. Alternatives to a meta robots tag are Robots.txt files and password protection.
Meta Robots code is found between the <head></head> tags and looks like this:
<meta name=”robots” content=”noindex, nofollow” / > or
<meta name=”robots” content=”noindex, follow” / >
Meta Search Engine – A meta search engine derives its listings by running user queries through multiple other search engines and then summarizing the results. A meta search engine does not maintain its own Index.
Listings are displayed by meta search engines either in aggregate or categorized by search engine source.
- Visit Dogpile.com
Meta Tags – People generally refer to meta descriptions and meta keywords as meta tags. Some people also group the page title in with these.
- The page title [link:] is highly important.
- The meta description [link:] tag is somewhat important.
- The meta keywords tag [link:] is not important.
Meta Verification Tag – Meta tags generated by the likes of Google Webmaster Tools [link:], Bing Webmaster, and Yahoo Site Explorer [link:] to enabled webmasters to track important information about their website as it pertains to search engines.
Meta Verification code is found between the <head></head> tags
Google’s tag will look something like this:
Bing’s will look something like this:
Yahoo’s will look something like this:
Microformats – A microformat is a web-based approach to semantic markup that seeks to re-use existing XHTML and HTML tags to convey metadata and other attributes. This approach allows information intended for end-users (such as contact information, geographic coordinates, calendar events, and the like) to also be automatically processed by software.
- Visit Microformats.org
Microsoft (MSN) – Microsoft Corporation is a multinational computer technology corporation that develops, manufactures, licenses, and supports a wide range of software products for computing devices. Headquartered in Redmond, Washington, USA, some of its most well-known products include the Microsoft Windows operating system, the internet browser Internet Explorer, and the Microsoft Office suite of productivity software among others.
Microsoft AdCenter – Microsoft’s cost per click ad network. It includes a few interesting features (dayparting, demographic-based bidding, etc.), but it is still a little behind in quality compared to Google AdWords [link].
Mirror Site – Site which mirrors (or duplicates) the contents of another website. Generally search engines prefer not to index duplicate content [link:]. The one exception to this is that if you are a hosting company it might make sense to offer free hosting or a free mirror site to a popular open source [link:] software site to build significant link equity.
Mod-rewrite – A module or plugin for Apache web servers that can be used to rewrite requested URLs on the fly.
It supports an unlimited number of rules and an unlimited number of attached rule conditions for each rule to provide a flexible and powerful URL manipulation mechanism. Mod-rewrites can be used to alter non search-friendly URLs to make them more search-friendly, thus increasing indexing chances (typically used for a dynamic database driven websites).
Movable Type – For sale blogging software which allows you to host a blog on your website. Movable Type is typically much harder to install that WordPress [link:] is.
Moved Permanently – Called a 301-Redirect or “moved permanently”, means that the file has been moved permanently to a new location. This is the preferred method of redirecting for most pages or websites. Depending on your site authority and crawl frequency it may take anywhere from a few days to a month or so for the 301-Redirect [link:] to be picked up.
Myspace – One of the most popular social networking sites, largely revolving around connecting musicians to fans and having an easy to use blogging platform.
Natural Listings – Web page listings that appear in a search engine’s results, based on the engine’s own proprietary algorithm. The website/web page owner has not paid for these positions in the results, and the listings are ranked in order of relevance to each search query with the most relevant listed first.
As long as a listing has achieved its position naturally (not paid for), it is a “natural” or “organic [link:]” listing.
Natural Search – See Organic Search Results [link:].
Netscape – Originally a company that created a popular web browser by the same name, Netscape is now a social news site similar to Digg.com [link:].
Niche – A topic or subject which a website is focused on.
Search is a broad field, but as you drill down each niche consists of many smaller niches. An example of drilling down to a niche market:
- Search marketing, privacy considerations, legal issues, history of, future of, different types of vertical search, etc.
- Search Engine Optimization, Search Engine Marketing, Pay-Per-Click
- Link building, keyword research, reputation monitoring and management, viral marketing, SEO copywriting, Google AdWords, information architecture, etc.
Generally it is easier to compete in small, new, or underdeveloped niches than trying to dominate large verticals. As your brand [link:] and authority [link:] grow you can go after bigger markets.
Niche Directory – Directories which focus on or niche topics, specialist sectors, restricted regions, or single languages. One type of niche directory with a large number of sites in existence, is the shopping directory for example.
Nofollow – Attribute used to prevent a link from passing link authority. Commonly used on sites with user generated content, like in blog comments.
The code to use nofollow on a link appears like:
Nofollow can also be used in a robots meta tag to prevent a search engine from counting any outbound links on a page. This code would look like this
Noscript Tags – The tags are used to supplement Flash [link:], and can be a great way to introduce content to supplement 100% Flash websites, where there otherwise would be none. You have to be careful with how you use it, but if used in a white-hat, non-malicious manner it can be an effective way to optimize Flash websites.
Here is an example of what Noscript Tags look like:
<p>Your Supplemental Content</p>
Off-Site SEO – This refers to Search Engine Optimization tactics that do not involved any direct manipulation of your website’s files. It refers to things that happen away from your website that have a direct influence on how your site behaves in search engines.
Some examples of Off-Site SEO include:
- Link Building [link: ]
- Directory Submission [link: ]
- Local Search [link: ]
- Social Media [link: ]
On-Site SEO – This refers to Search Engine Optimization tactics that involve any direct edits that you might make to your website’s files. These edits and the way your website is structured have a direct influence on how your site behaves in search engines.
Some examples of On-Site SEO include:
- Editing Title Tags [link: ]
- Editing Meta Tags [link: ]
- Copywriting [link: ]
- Image Optimization [link: ]
One-Way Link – A One-way link is a term used among search engine optimizers referring to a specific type of link building method. It is a Hyperlink that points to a website without any reciprocal link; thus the link goes “one way” in direction.
It is suspected by many industry consultants that this type of link would be considered more natural in the eyes of search engines (and thus more valuable).
There are many theories on link building and one-way links verse reciprocal links. Google [link: ] is the company that has made this concept very popular with their PageRank [link: ] algorithm.
Open-Source – Software which is distributed with its source code such that developers can modify it as they see fit. On the web open source is a great strategy for quickly building immense exposure and mindshare.
A great example of open-source technology would be WordPress.
Opera – A fast standards based web browser.
Organic Results – Most major search engines have results that consist of paid ads and unpaid listings. The unpaid / algorithmic listings are called the organic search results (or “natural search results”). Organic search results are organized by relevancy, which is largely determined based on linkage data, page content, usage data, and historical domain and trust related data.
Most clicks on search results [link: ] are on the organic search results. Some studies have shown that 60-80% of clicks are on the organic search results.
Also read: Click Distribution & Percentages by SERP Rank
Outbound Link – A link from one website to another external site.
Some webmasters believe in link hoarding [link: ], however linking out to useful relevant related documents is an easy way to help search engines understand what your website is about.
If you reference other resources it also helps you build credibility and leverage the work of others without having to do everything yourself. Some webmasters track where their traffic comes from, so if you link to related websites they may be more likely to link back to your site.
Overture – The company which pioneered search marketing by selling targeted searches on a pay per click basis. Originally named GoTo, they were eventually bought out by Yahoo! and branded as Yahoo! Search Marketing.
Visit: Yahoo! Search Marketing
Overture Keyword Selector Tool – Popular keyword research tool, based largely on Yahoo! search statistics. Heavily skewed toward commercially oriented searches, also combines singular and plural versions of a keyword into a single version.
Visit: Overture Keyword Selector Tool
PageRank – A logarithmic scale based on link equity which estimates the importance of web documents.
As defined by Google “PageRank relies on the uniquely democratic nature of the web by using its vast link structure as an indicator of an individual page’s value.
In essence, Google interprets a link from page A to page B as a vote, by page A, for page B. But, Google looks at more than the sheer volume of votes, or links a page receives; it also analyzes the page that casts the vote. Votes cast by pages that are themselves “important” weigh more heavily and help to make other pages “important.”
Since PageRank is widely bartered, Google’s relevancy algorithms had to move away from relying on PageRank and place more emphasis on trusted links via algorithms such as TrustRank [link: ].
The PageRank formula is:
PR(A) = (1-d) + d (PR(T1)/C(T1) + … + PR(Tn)/C(Tn))
d= dampening factor (~0.85)
c = number of links on the page
PR(T1)/C(T1) = PageRank of page 1 divided by the total number of links on page 1, (transferred PageRank)
In text: For any given page A, the PageRank PR(A) is equal to the sum of the parsed partial PageRank given from each page pointing at it multiplied by the dampening factor plus one minus the dampening factor.
- The Anatomy of a Large-Scale Hypertextual Web Search Engine
- The PageRank Citation Ranking: Bringing Order to the Web
Page Title – See Title Tags [link: ]
Paid Inclusion – A method of allowing websites which pass editorial quality guidelines to buy relevant exposure.
Sometimes the payment is arranged through an affiliate or partner of the database company.
- Directories such as the Yahoo! Directory [link: ] and Business.com allow websites to be listed for a flat yearly cost.
- Yahoo! Search allows webmasters to pay for inclusion for a flat review fee and a category based cost per click.
Paid Listings – Directory listings where the results display only advertisers who have paid for inclusion, and possibly for position.
Paid Placement – Paid placement is a program in which advertisers’ listings are guaranteed to appear on a Results Page when particular Keywords are searched. The ranking of paid placement listings is determined by competitive bidding.
Unlike Paid Inclusion listings, paid placement listings are usually displayed separately from Natural Listings and are labeled as advertisements or sponsored links. Google and Yahoo! Search Marketing (formerly Overture) are two of the largest paid placement search networks.
See: Pay-Per-Click Advertising [link: ]
Pay-for-Performance – Payment structure where affiliated sales workers are paid commission for getting consumers to perform certain actions.
Publishers publishing contextual ads [link: ] are typically paid per ad click. Affiliate marketing programs [link: ] pay affiliates for conversions – leads, downloads, or sales.
Pay-Per-Click Advertising – Also known as Cost-Per-Click (CPC) [link:] or Pay-For-Performance [link:], cost-per-click is an advertising revenue system used by search engines and ad networks in which advertising companies pay an agreed amount for each click of their ads.
This Click-Through Rate [link:]-based payment structure is considered by some advertisers to be more cost-effective than the Cost-Per-Thousand payment structure, but it can at times lead to Click Fraud [link:].
Penalty – Search engines prevent some websites suspected of spamming [link: ] from ranking highly in the results by banning or penalizing them. These penalties may be automated algorithmically or manually applied.
If a site is penalized algorithmically the site may start ranking again after a certain period of time after the reason for being penalized is fixed. If a site is penalized manually the penalty may last an exceptionally long time or require contacting the search engine with a reinclusion [link: ] request to remedy.
Some sites are also filtered [link: ] for various reasons.
- Google -30 rank penalty – an example of a penalty
Permalink – A permalink, or permanent link, is a URL that points to a specific blog [link: ] or forum entry after it has passed from the front page to the archives. Because a permalink remains unchanged indefinitely, it is less susceptible to link rot.
Most modern weblogging and content-syndication software systems support such links (see: WordPress [link: ]). Other types of websites use the term permanent links, but the term permalink is most common within the blogosphere.
Personalization – Altering of the search results based on a person’s location, search history, content they recently viewed, or other factors relevant to them on a personal level. Personalized search results are most prevalent on Google [link: ].
PHP – Pending
Poison Word – Pending
Portal – Pending
Position – Pending
Quality Content – Pending
Quality Link – Pending
Query – Pending
Query Refinement – Pending
Rank – Pending
Real-Time Search – Pending
Re-Inclusion Request – Pending
Reciprocal Link – Pending
Redirect – Pending
Referrer – Pending
Relative Link – Pending
Relevancy – Pending
Reputation Management – Pending
Results Page – Pending
Return-On-Investment (ROI) – Pending
Rich-Snippets – Pending
Robots.txt File – Pending
RSS – Pending
Safari – Pending
Sandbox – Pending
SEO Specialist – Pending
Search Engine – Pending
Search Engine Friendly – Pending
Search Engine Marketing (SEM) – Pending
Search Engine Marketing Professional Organization (SEMPO) – Pending
Search Engine Optimization (SEO) – Pending
SEO Copywriting – Pending
Search Engine Results Page (SERP) – Pending
Search History – Pending
Search Terms – Pending
Semantic Web – Pending
Server – Pending
Server Logs – Pending
Siphoning – Pending
Sitelinks – On some search results where Google thinks one result is far more relevant than other results (like navigational or brand related searches) they may list numerous deep links to that site at the top of the search results.
Sitemap – Pending
Slug – Pending
Social Bookmarking – Pending
Social Media – Pending
Social Search – Pending
Spam – Pending
Spamming – Pending
Sphinn – Pending
Spider – Pending
Splash Page – Pending
Splog – Pending
Static Content – Pending
Status Code – Pending
Status Found – Pending
Status OK – Pending
Status Unauthorized – Pending
Stop Words – Pending
Submission – Pending
Tag Cloud – Pending
Taxonomy – Pending
Technorati – Pending
Term Frequency – Pending
Text-Link Ads – Pending
Title Tags – Pending
Toolbar – Pending
Trackback – Pending
Traffic – Pending
TrustRank – Pending
Twitter – Pending
Typepad – Pending
URI – Pending
URL – Pending
URL Rewrite – Pending
Usability – Pending
Vertical Search – Pending
Viral Marketing – Pending
Virtual Domain – Pending
Virtual Server – Pending
Vlog – Pending
W3C – Pending
Web – Pending
Web Analytics – Pending
Web CEO – Pending
Web Crawler – Also known as Spider or Robot, a crawler is a search engine program that “crawls” the web, collecting data, following links, making copies of new and updated sites, and storing URLs in the search engine’s Index. This allows search engines to provide faster and more up-to-date listings.
Examples of ‘good’ spiders are Googlebot, MSNbot and Yahoo Slurp, which search the web for pages to place in their respective search indexes. ‘Bad’ spiders are those that search for vulnerabilities in server configurations, web content to steal, etc.
Many non-traditional search companies have different spiders which perform other functions. For example, TurnItInBot searches for plagiarism. Spiders should obey the robot.txt protocol.
White-Hat Tactics – Pending
Wiki – Pending
Wikipedia – Pending
WordPress – Pending
WYSIWYG – Pending
XHTML – Pending
XML – Pending
XML Feed – Pending
Yahoo! – Pending
Yahoo! Answers – Pending
Yahoo! Directory – Pending
Yahoo! Local – Pending
Yahoo! Search Marketing – Pending
Yahoo! Site Explorer – Pending
Youtube – Pending
10-Pack – Pending
200-Status – Pending
301-Redirect – Pending
302-Redirect – Pending
404-Error – Pending
As I stated above, I did not come up with all of the definitions above – simply compiled the best available information and added my thoughts, as well as other relevant links and material in most cases. Below are my sources:
- SEOBook’s Glossary – From Aaron Wall, one of the most highly-recognized people in the SEO industry today.
- Submit Express’ SEO Dictionary
- Wikipedia – I used this for a few definitions. I know that Wikipedia is human-edited, but that doesn’t mean all the information there is crap.
Am I missing a term that you would like defined? Simply contact me and I’ll do my best to get it on this list!