Search Engine
Website that offers its visitors the ability to search the
content of numerous web pages on the Internet
google, yahoo, msn(bing), ask, aol
Google history:
Google was founded by Stanford
University students Larry
Page and Sergey Brin in 1998.
Yahoo! was founded
in 1994 by David Filo and Jerry Yang as a directory of websites. For many
years they outsourced their search service to other providers, but by the end
of 2002 they realized the importance and value of search and started
aggressively acquiring search companies.
Overture purchased AllTheWeb and AltaVista. Yahoo! purchased Inktomi (in
December 2002) and then consumed Overture (in July of
2003), and combined the technologies from the various search companies they
bought to make a new search engine. Yahoo! dumped Google in favor of their own
in house technology on
February 17th, 2004.
Search Engine Optimization (SEO)
Search Engine Optimization is a process or technique to increase a
visitor on a website in organic result of a search engine. SEO also work as a bridge between technology
and business.
Spider/Crawler/Robots
Search engine spider, also known as robots, are software
application that “crawl” the web looking for websites to “index”(add to their
database)
Crawling
Crawling is a search engine process of finding new web sites
or updating of old site and adding them to search engine database.
Index
A search engine’s database in which it store textual content
from every webpage that search engine’s spider visits.
Cache
Cache is a snapshot of a webpage or literally a copy of webpage when the search
engine spider visited last time on a webpage.
Search Engine Algorithm
Search engine algorithm is a set of rules,
or a unique formula, that the search engine uses to determine the significance
of a web page, and each search engine has its own set of rules.
·
Relevancy
·
Individual factors
·
Off page factor
Search Engine Commands
Show
the cached snapshot of a page
link:
List
pages which link to a page
related:
List
pages which Google consider to be related to another
info:
Find
one specific URL in the search database
define:
Show
Google's glossary definition for a term
site:
Restrict
a search to a single site
allintitle:
Restrict
a search so that all the keywords must appear in the title
intitle:
Restrict
a search so that some of the keywords must appear in the title
allintext:
Restrict
a search so that all of the keywords must appear in the body text
allinurl:
Restrict
a search so that all of the keywords must appear in the page address
inurl:
Restrict
a search so that some keywords must appear in the page address
OR
List
pages which have at least some of the keywords
+
Insist
that the search engine includes a given keyword in the search results
-
Insist
that the search engine omits pages which match a given keyword in the search
results
~
Enhance
a search to include synonyms for a given keyword
filetype: Restrict a
search to a given type of file
site:
Restrict
a search to a single site
link:
List
pages which link to a page
url:
Find
one specific URL in the search database
inurl:
Restrict
a search so that some keywords must appear in the page address
intitle:
Restrict
a search so that some of the keywords must appear in the title
+
Insist
that the search engine includes a given keyword in the search results
-
Insist
that the search engine omits pages which match a given keyword in the search
results
link:
List
pages which link to a page
LinkFromDomain:
List
pages which a domain links to
inurl
Restrict
a search so that some keywords must appear in the page address
inanchor:
Restrict
a search so that the keyword must appear in anchors tags on the page
intitle:
Restrict
a search so that some of the keywords must appear in the title
inbody:
Restrict
a search so that all of the keywords must appear in the body text
site:
Restrict
a search to a single site
url:
Find
one specific URL in the search database
+
Insist
that the search engine includes a given keyword in the search results
-
Insist
that the search engine omits pages which match a given keyword in the search
results
NOT
Insist
that the search engine omits pages which match a given keyword in the search
results
OR
List
pages which have at least some of the keywords
|
List
pages which have at least some of the keywords
""
Restrict
a search so that the keywords must appear consecutively in a phrase
()
Group
keywords together
Page rank
Page rank is an algorithm
used by search engine Goolge for link analysis and this Page rank was developed
by Larry Page (consequently the name
Page-rank) and later Sergey Brin at Stanford University. Google search engine assign
a numerical weighting or value to webpage that measure the importance of the
webpage, based on a scale or range of 0 to 10. Its mean Google assign a Page
rank 10 for the most popular websites and Page rank 0 to least one. One more
term associated normally with page rank is Toolbar Page rank and the value of
this Toolbar Page rank is always in whole number like 1,2,3,……10.
PageRank, a
trademark of Google search engine define the value of webpage and this Page
rank depend on other individual page's value indicating a link to webpage.
Exact information of this algorithm is reserved by Google.
How to Calculate the
Page rank?
First you must know two type of link. There is two type of like, one is inbound and
other is out bound link. Inbound links are the link that pointed your webpage
from other WebPages and out bound link are the link on your webpage that
pointed to other webpage.
Calculate Page Rank:
All of the inbound links are taken into relation. The equation that calculated page rank of
webpage’s is:
PR (A) = (1-d) +
d(PR(t1)/C(t1) + ... + PR(tn)/C(tn))
* t1 to t2 are
the link from other pages pointing to Page A.
* “C” is a number
of outbound links that the pages have.
* D stand for damping factor usually set to 0.85.
As per
this formula the conclusion is, that link from another site increases our
site's Page Rank.
TLD
The root of the tree has no name. All siblings of a domain must have unique
names. Children of a domain are called subdomains
of the parent.
ROOT
___________________|________________________
|
| | |
| | |
| |
COM EDU GOV MIL NET ORG INT US ARPA
(biz, info : should not)
The information is distributed among thousands of name servers. The root
name servers only have complete information about the top-level domains. No
servers have complete information about all domains, but the root name servers
have pointers to the servers for the second level domains. This gives them
complete access to the information about all domains by querying
its list of name servers.
Top-Level Domains
Descriptions of each main top-level domain that divide up the Internet Name
Space are given below.
Any commercial, for-profit business may register in the COM domain. This is
the most common type of domain being registered today. Although a full domain
name may be up to 63 characters, the InterNIC recommends 12 characters or less
in the portion preceding the "dot". Any alphanumeric character is
allowed. Hyphens are permitted, but only for interior characters (no name may
start or end with a hyphen).
Originally, the EDU domain was designed to be the top-level domain for all
educational institutions. However, due to the explosive growth of the Internet,
the InterNIC now limits registrations in the EDU domain to 4-year, degree
granting colleges and universities. All other educational institutions must
register in the US
domain. If you intend to register in the EDU domain, you must include a
brief sentence in your application affirming your organization is a 4-year
college or university.
This top-level domain is reserved for certain Federal Government Agencies
and only under specific circumstances. Details regarding these specifications
can be found in RFC 2146.
Any department or agency of the
US military may register a name in
the MIL domain. In your application, be sure to affirm you are a department or
agency of the
US
military.
The NET domain is intended to hold only the computers of network providers.
The appropriate entity for a NET domain is an Internet Service Provider or an
administrative organization providing Internet support functions (such as the
InterNIC).
The InterNIC previously defined the ORG domain for non-profit organizations
and other organizations that did not fit anywhere else. Currently, registering
in the ORG domain is restricted to only non-profit organizations. Be sure to
affirm you are a non-profit organization in your application.
The INT domain is used only for registering organizations established by
international treaties between governments or Internet infrastructure
databases. Applicants for the International Organization domain must comply
with the requirements found in RFC 1591.
This is the top-level country domain for the
United States. The US Domain
currently registers hosts of businesses, individuals, federal government
agencies, state government agencies, K12 schools, community colleges,
technical/vocational schools, private schools, libraries, museums, city and
county government agencies. Further information can be obtained at:
http://www.nic.us/
This was originally used during the ARPANET's transition from host tables to
DNS. All ARPANET hosts originally had host names under "arpa" so they
were easy to find. Later, they moved into various subdomains of the
organizational top-level domains.
Keyword Proximity: Keyword proximity refers to how close
two or more keywords are to each other. You will achieve higher rankings if you
place your keywords close together.
Keyword Prominence: Keyword prominence refers to how prominent
your keywords are within your web page. In simple English, this means how close
to the beginning of the web page, sentence, title, h tag, or
Meta
description your keywords are placed.
If at all possible, place your most important keyword at the very beginning
of the following:
- Title
- Description
- H1 Tag
- The body of the text
Keyword Density: Keyword density refers to the percentage
of keywords that can be found in the indexable text on a web page. To be
honest, I don’t believe there is a secret percentage or ratio. I will mention
that a conservative percentage would be anywhere from 2-4% and not to exceed
10%.
Regarding keyword density, I recommend:
- Keyword in the title
- Keyword in the description
- Keyword in the H1 tag
- A few variations of keywords sprinkled throughout the text, where it sounds
natural and makes sense
Keyword Frequency: Keyword frequency is the number of times
a particular keyword appears on a website. Common sense may indicate the more
times a keyword, or keyword phrase is found on a web page, the better the site
will rank in Google.
- Matt Cutts - Matt is an
amazingly friendly and absurdly accessible guy given his position as the
head of Google's search quality team.
Example Vertical Search Engines
General large scale web search is just one type of search. There are many
other types of search engines and information organization tools, for example
- the Yellow Pages
- Television program guides
- directories like DMOZ, the
Yahoo! Directory, LII, or specialty
directories
- encyclopedia type sites like Wikipedia
- large general structured
databases like Google Base
- shopping search like Froogle
- local search like Google Local
- news search like Yahoo! News
- blog search like Technorati
- tag search like Del.icio.us
- video search like YouTube
- photo search like Flickr
- meme trackers like Techmeme
- social communities like Digg
- social networks like MySpace
- some people may also rely on an individual content
channel or a group of them to find the most interesting things and deliver
it through daily updated content streams
On page Optimization
<html>
<head>
<title> Here is the title of the HTML document. </title>
<meta name="description" content="">
<meta name="keywords" content=" ">
<meta name="robots" content="noodp">
<meta name="robots" content="index, follow">
</head>
<body>
Here is the body of the HTML document.
</body>
</html>
HTML Hyperlink
<a href="sample.html">This link</a>
<A HREF="newwindow.html" TARGET="_blank">a new window</A>
HTML Image
<img src="images/aboutus.gif" width="111" height="45" border="0" alt="Web Development India" Title="Outsourcing Website Development India">
page emphasis
<b>bold</b>
<u>underline</u>
<i>italicized</i>
HEADER TAG
<h1>...<h6> tag
Introduction
"On Page" SEO simply refers to the
text and content on your web site pages. Basically editing your page and
content so search engine can find your web page.
On page SEO is making sure that your website
is as search engine friendly as possible. If your website is not
optimized then you have less chance of getting good results in the search
engines, here is a quick guide towards on page SEO
n Positive On Page SEO / White
Hat SEO Techniques
n Negative On Page / Black Hat
SEO Techniques
There are many factors that
Search Engine uses to rank pages in the search results (SERPs). Here we will
see the main On-Page SEO ranking factors.
§ URL
n Page Title
n Meta Tag
n Content
n Anchor text
n Heading Tag
n Image
Improve the structure of your URLs
The URL of a document should ideally be as descriptive and brief as
possible
Poor Practice
URLs like these can be confusing and unfriendly.
Good Practice
The highlighted words above could inform a user or search engine what the
target page is about before following the link
Good practices for URL
structure
• Use
words in URLs - URLs with words that are relevant
to your site's content and structure are friendlier for visitors.
Avoid:
• using lengthy URLs with unnecessary parameters and
session IDs
• choosing generic page names like
"page1.html"
• using excessive keywords like
"baseball-cards-baseball-cards-baseballcaerds.htm"
• Create
a simple directory structure - Use a
directory structure that organizes your content well and is easy for visitors
to know where they're at on your site.
Avoid:
• having deep nesting of subdirectories like
".../dir1/dir2/…/dir8/page.html"
• using directory names that have no relation to the
content in them
• Provide
one version of a URL to reach a document - To
prevent users from linking to one version of a URL and others linking to a
different version
Avoid:
• having pages from subdomains and the root
directory (e.g. "domain.com/
page.htm" and
"sub.domain.com/page.htm") access the same content
• mixing www. and non-www. versions of URLs in your
internal linking structure
• using odd capitalization of URLs (many users
expect lower-case URLs and
remember them better)
Create
unique, accurate page titles
A title tag tells both
users and search engines what the topic of a particular page is.
The title of the homepage for our baseball
card site, which lists the business name and three main focus areas
Good practices
for page title tags
o Accurately
describe the page's content - Choose a title that effectively communicates
the topic of the page's content.
n
Avoid:
o
choosing a title that has no relation to the content on the page
o
using default or vague titles like "Untitled" or "New
Page 1"
o Create
unique title tags for each page - Each of your pages should ideally have a
unique title tag, which helps engine know how the page is distinct from the
others on your site.
n
Avoid:
o
using a single title tag across all of your site's pages or a large
group of pages
o Use
brief, but descriptive titles - Titles can be both short and informative. If
the title is too long, engine will show only a portion of it in the search
result.
n
Avoid:
o
using extremely lengthy titles that are unhelpful to users
o
stuffing unneeded keywords in your title tags
META TAG
Make use of the description" meta tag
A page's
description meta tag gives search engines a summary of what the page is about.
Good
practices for description meta tags
o Accurately summarize the page's content - Write a
description that would both inform and interest users if they saw your
description meta tag as a snippet in a search result.
n Avoid:
o writing a
description meta tag that has no relation to the content on the page
o using generic
descriptions like "This is a webpage" or "Page about baseball
cards"
o filling the
description with only keywords
o copy and
pasting the entire content of the document into the description meta tag
o Use unique descriptions for each page - Having a
different description meta tag for each page helps both users and engine.
n Avoid:
o using a single
description meta tag across all / (more that one) page of your site's pages
Offer quality content and services
Creating
compelling and useful content will likely influence your website more than any
of the other factors discussed here. Users know good content when they see it
and will likely want to direct other users to it. This could be through blog
posts, social media services, email, forums, or other means.
Good practices for content
n
Write
easy-to-read text - Users enjoy content that is well written and easy to
follow.
o Avoid:
n writing
sloppy text with many spelling and grammatical mistakes
n embedding
text in images for textual content (users may want to copy and paste the text
and search engines can't read it)
n
Stay
organized around the topic - It's always beneficial to organize your
content so that visitors have a good sense of where one content topic begins
and another ends. Breaking your content up into logical chunks or divisions helps
users find the content they want faster.
n Avoid:
n dumping
large amounts of text on varying topics onto a page without
n paragraph,
subheading, or layout separation
n
Use
relevant language - Think about the words that a user might search for to
find a piece of your content. Users who know a lot about the topic might use
different keywords in their search queries than someone who is new to the
topic. For example, a long-time baseball fan might search for [nlcs], an
acronym for the National League Championship Series, while a new fan might use
a more general query like [baseball playoffs]. Anticipating these differences
in search behavior and accounting for them while writing your content.
n
Create
fresh, unique content - New content will not only keep your existing
visitor base coming back, but also bring in new visitors.
n Avoid:
n rehashing
(or even copying) existing content that will bring little extra value to users
n having
duplicate or near-duplicate versions of your content across your site
n
Offer
exclusive content or services - Consider creating a new, useful service
that no other site offers. You could also write an original piece of research,
break an exciting news story, or leverage your unique user base.
n
Create
content primarily for your users, not search engines - Designing your site
around your visitors' needs while making sure your site is easily accessible to
search engines usually produces positive results.
n Avoid:
n inserting
numerous unnecessary keywords aimed at search engines but are annoying or nonsensical
to users
n having
blocks of text like "frequent misspellings used to reach this page"
that add little value for users
n deceptively
hiding text from users, but displaying it to search engines
Write
better anchor text
This text tells users and engines
something about the page you're linking to.
Good
practices for anchor text
o Format links so they're easy to spot - Make it easy
for users to distinguish between regular text and the anchor text of your
links. Your content becomes less useful if users miss the links or accidentally
click them.
o Avoid:
n using CSS or
text styling that make links look just like regular text
o Think about anchor text for internal links too
- You may usually think about linking in terms of pointing to
outside websites, but paying more attention to the anchor text used for
internal links can help users and engine navigate your site better.
o Avoid:
n using
excessively keyword-filled or lengthy anchor text just for search engines
Use heading
tags appropriately
Heading are used to present structure on the page to users.
Good practices for heading tags
• Imagine you're writing an outline
• Use headings sparingly across the page
Avoid:
• excessively using heading tags throughout the page
• putting all of the page's text into a heading tag
• using heading tags only for styling text and not presenting structure
Optimize
your use of images
Text styles that cannot be indexed by
search engines include images. And if you used image in your site then give
proper name to image file and use useful alt text.
Why
use this attribute? If a user is viewing your site on a browser that doesn't
support images, or is using alternative technologies, such as a screen reader,
the contents of the alt attribute provide information about the picture.
Good
Practice
There are several different techniques known as "black hat" or
"unethical" On Page Techniques. Some SEO companies engage in these
type of activities and should be avoided. Sooner or later the search engines
will catch up to these unethical techniques and the likely result will be your
site being demoted or banned from the search engines.
Negative ON Page SEO/ Black Hat SEO Techniques
·
Avoid Using "hidden" or
invisible text on your page for the purpose of higher the words are "hidden" from the
surfer. search engine placement. For example the words/text for search phrase
"Widget" in the html, the font color has been set to White. The background
of the page is also white. Therefore the textual content is actually there,
however This is frowned upon by search engines and frequently results in your
site being penalized
·
Avoid Using Negative <div>
htmltags. Div tags, Div tags are division tags. Unscrupulous seo services may insert them
into your page with negative x/y coordinates to place content outside of the
visible page for the surfer, but the text itself is in the html page. The
search engine finds the keywords in the text, yet the surfer does not see it.
Again a technique to be avoided and not recommended under any circumstances
·
Avoid Cloaking or Sneaky Redirects. Cloaking refers to serving up 2
different types of content based on the visitor who is visiting. Is the visitor
a regular web surfer, serve up this page. Is the visitor a search engine
spider? Serve up this OTHER page specifically for the search engine spider. The
other page being served up is typically garbled textual content with no meaning
to a human, and is stuffed with various keywords and search phrases. Again this
technique is not recommended and will likely get your site penalized or banned
from search engines
·
Avoid duplicate content. Duplicate content means you create
one web site, with content on topic a, and then repeat the content over and
over again on multiple websites. In theory you could create one website,
achieve high ranking on it, and then clog up the search engines with the same
content duplicated on multiple domains. Again this is not recommended and
should be avoided
Dynamic Websites
Optimization
What are "Dynamic Websites"?
Dynamic websites are websites whose pages are generated on the fly. Unlike
static pages (primarily .htm/.html pages), dynamic pages are generated when an
user triggers an action through that particular page.
Here isa sample dynamic URL-
http://www.bbc.co.uk/cgi-bin/search/results.pl?q=%22dynamic+websites%22&tab=news
&go=homepage
As per the above example of www.bbc.co.uk, the dynamic part (i.e. the part) of
the URL which changes as per surfer request is the part after the question mark
(?)
What are the problems that search engines face in indexing Dynamic URLs?
1. Search engines often consider a dynamic URL as an infinite set of links.
2. Since dynamic URLs find maximum application in online shopping carts, there
is a possibility of incorporating a session id to a particular page. As session
ids of that particular page change, the search engine spider needs to index an
infinite number of copies of the same page, which is a Herculean task for them.
3. Proceeding with the same logic presented in point #2, indexing the same
dynamic page might overload the servers of the search engines and therefore
prevent the search engines to present with the most relevant information in the
fastest possible time.
Here is what Google says about indexing of dynamic websites-
Reasons your site may not be included: Your pages are dynamically generated. We
are able to index dynamically generated pages. However, because our web crawler
can easily overwhelm and crash sites serving dynamic content, we limit the
amount of dynamic pages we index. (Source - http://www.google.com/webmasters/)
What are the options that you have in order to make a search engine spider
index your Dynamic URLs?
1. Use of softwares - Exception Digital Enterprise Solutions
(http://www.xde.net ) offers a software which can change the dynamic URLs to
static ones. Named XQASP, it will remove the "?" in the Query String
and replace it with "/", thereby allowing the search engine spiders
to index the dynamic content.
Example -
http://www.my-online-store.com/books.asp?id=1190 will change to
http://www.my-online-store.com/books/1190.
The latter being a static URL, it can easily be indexed by the search engine
spiders.
2. Use of CGI/Perl scripts - One of the easiest ways to get your dynamic sites
indexed by search engines is using CGI/Perl scripts. Path_Info or Script_Name
is a variable in a dynamic application that contains the complete URL address
(including the query string information). In order to fix this problem, you'll
need to write a script that will pull all the information before the query
string and set the rest of the information equal to a variable. You can then
use this variable in your URL address.
Example - http://www.my-online-store.com/books.asp?id=1190
When you are using CGI/Perl scripts, the query part of the dynamic URL is
assigned a variable. So, in the above example "?id=1190" is assigned
a variable, say "A". The dynamic URL
http://www.my-online-store.com/coolpage.asp?id=1190 will change to
http://www.my-online-store.com/books/A through CGI/Perl scripts which can
easily be indexed by the search engines.
3. Re-configuring your web servers-
Apache Server - Apache has a rewrite module (mod_rewrite) that enables you to
turn URLs containing query strings into URLs that search engines can index.
This module however, isn't installed with Apache software by default, so you
need to check with your web hosting company for installation.
ColdFusion - You'll need to reconfigure ColdFusion on your server so that the
"?" in a query string is replaced with a '/' and pass the value to
the URL.
4. Creation of a Static Page linked to an array of dynamic Pages - This
approach is very effective, especially if you are the owner of a small online
store selling a few products online. Just create a static page linking to all
your dynamic pages. Optimize this static page for search engine rankings.
Include a link title for all the product categories, place appropriate
"alt" tag for the product images along with product description
containing highly popular keywords relevant to your business (you can conduct
keyword research for your site through http://www.wordtracker.com). Submit this
static page along with all the dynamic pages in various search engines,
conforming to the search engine submission guidelines.
How Amazon.com, Earth's Biggest Bookstore, coped with the issue of indexing of
dynamic URLs?
A search in Google for internet marketing books, yielded a result that takes
you directly to the appropriate dynamic page at Amazon -
http://www.amazon.com/exec/obidos/ISBN%3D0395683297/103-0475212-8205437.
Since the above URL does not contain any query strings, all search engines can
index Amazon.com's products. Amazon.com uses this method to get its product
selections indexed by search engines. This is very important for Amazon,
because being an online bookstore, it is very natural for them to adopt dynamic
URLs yet it was equally important for them to make their dynamic URLs search
engine index friendly.
Conclusion
Even a few years back, most of the major search engines did not index dynamic
URLs, thereby often preventing top search engine rankings for the online
stores. With Google starting to index dynamic URLs a few months ago, the
picture is going to change in the coming days. This is more so because Google's
numero uno position is currently being threatened by Microsoft's MSN
(developing its own search engine) and Yahoo! who recently acquired Overture,
the biggest player in the PPC Search Engine industry.
To read more articles click on
http://www.searchengine-optimization-guru.com/seo-whitepapers.html
===============================================================
Robots.txt Concept
User-agent: *
Disallow:
User-agent: Googlebot
Disallow: /
User-agent: Slurp
Disallow: /
User-agent: Teoma
Disallow: /
Disallow: /login/
.htaccess file
RewriteCond %{THE_REQUEST} ^[A-Z]{3,9}\ /index\.html\ HTTP/
RewriteRule ^index\.html$ http://www.abc.com/ [R=301,L]
XML Sitemap Generator
This checklist provides 50 simple ways for you to
make your website even better!
1. Use
.htaccess
to redirect the non-www version of your website to the www version.
2. Extend the registration of your domain for at
least five years.
3. Make sure that when a visitor clicks your website’s
logo, they are taken back to the homepage.
4. If you think your font is too small, make it
bigger!
5. Remove widgets that aren’t directly enhancing
your website.
6. Include an easy to find contact link, e-mail
address or phone number on every page of your website.
8. Fix or remove broken links (external and
internal).
9. Add ALT tags to all of your images.
10. If your website is using frames, get
rid of them!
11. If you don’t have a
robots.txt file,
create one.
12. Provide the same basic navigation menu on
every page of your website.
13. Use the same color for links throughout your
website.
14. Spell check your content.
15. Format your content so that it is easy to
scan.
16. Add at least one picture to every page of your
website.
17. If you don’t have an About page, create one.
18. Use your About page to show your brand’s
personality.
19. Include at least three pictures on your About
page.
20. If your website has pop-up windows, get
rid of them!
21. Link to relevant social media profiles from
your About page.
22. Add a search box near the top of your
website’s layout.
24. Place a link to your privacy policy at the
bottom of every page.
25. Use dashes in filenames instead of
underscores.
26. Add a link to Google Maps on your Contact
page.
27. Add internal links from one piece of content
to another.
28. Write a unique META
description for all of your main pages.
30. If you have a link exchange page, get
rid of it!
31. Add a site search box to your 404 page.
32. Create a sitemap for your website.
33. Add a link to your sitemap at the bottom of
every page of your website.
34. Add a copyright notification to the bottom of
every page of your website.
35. Underline your links throughout your website.
36. Turn off music that automatically plays.
37. Get rid of any flashing GIF animations.
38. Include your website logo at the top of every
page.
39. Register the .net and .org versions of your
domain and redirect them to your website.
40. If you don’t have a help or support section, create
one!
41. When you answer a question via e-mail, add the
answer to the help/support section.
42. Look at your contact forms and see if there is
any information you don’t have to request from your visitors.
43. Add your analytics code to every page of your
website.
44. Get rid of any pages that automatically resize
visitors’ browser window.
45. Create a
favicon
for your website.
46. Only use email address that include
@yourdomain.
47. Label each box of any forms on your website.
48. Show a confirmation any time a visitor
completes a significant action on your website.
49. Create one new piece of quality
content for your website every week.
50. Leave a comment with your suggestion for
the fiftieth tip!