Sunday, July 20, 2014

How to test mobile web page and application

When you are building a webpage or an application for mobile. It is required to test before you go live or send it to store repository. Testing gives you a good idea to find out errors in application. Once the errors are in front of you can fix them and give user a error free and reliable program.

You need to check it on different OS platform and different browsers. Many of designers ignores testing phase by which sometime the web page don’t appears to user with perfect layout. So it is mandatory to check each and every page of website.

Testing on tablets is most important for designers. Because tablets are growing every faster every years. Different touch screen tablets have different requirements.


For testing an application a programmer can use emulator for testing and debugging mobile apps. An emulator work as a real device on your computer. It provides all the frameworks that you need to test a mobile application there are many emulators over internet is free to download.

Saturday, July 12, 2014

Thursday, July 3, 2014

Tuesday, July 1, 2014

Important tips for seo (search engine optimization)

Search Engine

Website that offers its visitors the ability to search the content of numerous web pages on the Internet

google, yahoo, msn(bing), ask, aol

Google history:  Google was founded by Stanford University students Larry Page and Sergey Brin in 1998.

Yahoo! was founded in 1994 by David Filo and Jerry Yang as a directory of websites. For many years they outsourced their search service to other providers, but by the end of 2002 they realized the importance and value of search and started aggressively acquiring search companies.
Overture purchased AllTheWeb and AltaVista. Yahoo! purchased Inktomi (in December 2002) and then consumed Overture (in July of 2003), and combined the technologies from the various search companies they bought to make a new search engine. Yahoo! dumped Google in favor of their own in house technology on February 17th, 2004.
Search Engine Optimization (SEO)

Search Engine Optimization is a process or technique to increase a visitor on a website in organic result of a search engine.  SEO also work as a bridge between technology and business.


Spider/Crawler/Robots

Search engine spider, also known as robots, are software application that “crawl” the web looking for websites to “index”(add to their database)

Crawling

Crawling is a search engine process of finding new web sites or updating of old site and adding them to search engine database.


Index

A search engine’s database in which it store textual content from every webpage that search engine’s spider visits.

Cache

Cache is a snapshot of a webpage or literally a copy of webpage when the search engine spider visited last time on a webpage.


Search Engine Algorithm

Search engine algorithm is a set of rules, or a unique formula, that the search engine uses to determine the significance of a web page, and each search engine has its own set of rules.

·         Relevancy
·         Individual factors
·         Off page factor

 

Search Engine Commands

Show the cached snapshot of a page
link:
List pages which link to a page
related:
List pages which Google consider to be related to another
info:
Find one specific URL in the search database
define:
Show Google's glossary definition for a term
site:
Restrict a search to a single site
allintitle:
Restrict a search so that all the keywords must appear in the title
intitle:
Restrict a search so that some of the keywords must appear in the title
allintext:
Restrict a search so that all of the keywords must appear in the body text
allinurl:
Restrict a search so that all of the keywords must appear in the page address
inurl:
Restrict a search so that some keywords must appear in the page address
OR
List pages which have at least some of the keywords
+
Insist that the search engine includes a given keyword in the search results
-
Insist that the search engine omits pages which match a given keyword in the search results
~
Enhance a search to include synonyms for a given keyword

filetype: Restrict a search to a given type of file


Yahoo search commands summary

site:
Restrict a search to a single site
link:
List pages which link to a page
url:
Find one specific URL in the search database
inurl:
Restrict a search so that some keywords must appear in the page address
intitle:
Restrict a search so that some of the keywords must appear in the title
+
Insist that the search engine includes a given keyword in the search results
-
Insist that the search engine omits pages which match a given keyword in the search results

Live search commands

link:
List pages which link to a page
LinkFromDomain:
List pages which a domain links to

inurl
Restrict a search so that some keywords must appear in the page address
inanchor:
Restrict a search so that the keyword must appear in anchors tags on the page
intitle:
Restrict a search so that some of the keywords must appear in the title
inbody:
Restrict a search so that all of the keywords must appear in the body text
site:
Restrict a search to a single site
url:
Find one specific URL in the search database
+
Insist that the search engine includes a given keyword in the search results
-
Insist that the search engine omits pages which match a given keyword in the search results
NOT
Insist that the search engine omits pages which match a given keyword in the search results
OR
List pages which have at least some of the keywords
|
List pages which have at least some of the keywords
""
Restrict a search so that the keywords must appear consecutively in a phrase
()
Group keywords together




Page rank



Page rank is an algorithm used by search engine Goolge for link analysis and this Page rank was developed by Larry Page (consequently the name Page-rank) and later Sergey Brin at Stanford University. Google search engine assign a numerical weighting or value to webpage that measure the importance of the webpage, based on a scale or range of 0 to 10. Its mean Google assign a Page rank 10 for the most popular websites and Page rank 0 to least one. One more term associated normally with page rank is Toolbar Page rank and the value of this Toolbar Page rank is always in whole number like 1,2,3,……10.

PageRank, a trademark of Google search engine define the value of webpage and this Page rank depend on other individual page's value indicating a link to webpage. Exact information of this algorithm is reserved by Google.


How to Calculate the Page rank?

First you must know two type of link.  There is two type of like, one is inbound and other is out bound link. Inbound links are the link that pointed your webpage from other WebPages and out bound link are the link on your webpage that pointed to other webpage.

Calculate Page Rank:  All of the inbound links are taken into relation.  The equation that calculated page rank of webpage’s is:

PR (A) = (1-d) + d(PR(t1)/C(t1) + ... + PR(tn)/C(tn))

* t1 to t2 are the link from other pages pointing to Page A.
* “C” is a number of outbound links that the pages have.
* D stand for damping factor usually set to 0.85.

As per this formula the conclusion is, that link from another site increases our site's Page Rank.



TLD
The root of the tree has no name. All siblings of a domain must have unique names. Children of a domain are called subdomains of the parent.

                            ROOT

           ___________________|________________________
           |    |    |    |       |    |    |    |    |
          COM  EDU  GOV  MIL      NET  ORG  INT  US  ARPA

(biz, info : should not)
The information is distributed among thousands of name servers. The root name servers only have complete information about the top-level domains. No servers have complete information about all domains, but the root name servers have pointers to the servers for the second level domains. This gives them complete access to the information about all domains by querying its list of name servers.

Top-Level Domains

Descriptions of each main top-level domain that divide up the Internet Name Space are given below.

COM

Any commercial, for-profit business may register in the COM domain. This is the most common type of domain being registered today. Although a full domain name may be up to 63 characters, the InterNIC recommends 12 characters or less in the portion preceding the "dot". Any alphanumeric character is allowed. Hyphens are permitted, but only for interior characters (no name may start or end with a hyphen).

EDU

Originally, the EDU domain was designed to be the top-level domain for all educational institutions. However, due to the explosive growth of the Internet, the InterNIC now limits registrations in the EDU domain to 4-year, degree granting colleges and universities. All other educational institutions must register in the US domain. If you intend to register in the EDU domain, you must include a brief sentence in your application affirming your organization is a 4-year college or university.

GOV

This top-level domain is reserved for certain Federal Government Agencies and only under specific circumstances. Details regarding these specifications can be found in RFC 2146.

MIL

Any department or agency of the US military may register a name in the MIL domain. In your application, be sure to affirm you are a department or agency of the US military.

NET

The NET domain is intended to hold only the computers of network providers. The appropriate entity for a NET domain is an Internet Service Provider or an administrative organization providing Internet support functions (such as the InterNIC).

ORG

The InterNIC previously defined the ORG domain for non-profit organizations and other organizations that did not fit anywhere else. Currently, registering in the ORG domain is restricted to only non-profit organizations. Be sure to affirm you are a non-profit organization in your application.

INT

The INT domain is used only for registering organizations established by international treaties between governments or Internet infrastructure databases. Applicants for the International Organization domain must comply with the requirements found in RFC 1591.

US

This is the top-level country domain for the United States. The US Domain currently registers hosts of businesses, individuals, federal government agencies, state government agencies, K12 schools, community colleges, technical/vocational schools, private schools, libraries, museums, city and county government agencies. Further information can be obtained at:
http://www.nic.us/

ARPA

This was originally used during the ARPANET's transition from host tables to DNS. All ARPANET hosts originally had host names under "arpa" so they were easy to find. Later, they moved into various subdomains of the organizational top-level domains.


Keyword Proximity: Keyword proximity refers to how close two or more keywords are to each other. You will achieve higher rankings if you place your keywords close together.
Keyword Prominence: Keyword prominence refers to how prominent your keywords are within your web page. In simple English, this means how close to the beginning of the web page, sentence, title, h tag, or Meta description your keywords are placed.
If at all possible, place your most important keyword at the very beginning of the following:
- Title
- Description
- H1 Tag
- The body of the text
Keyword Density: Keyword density refers to the percentage of keywords that can be found in the indexable text on a web page. To be honest, I don’t believe there is a secret percentage or ratio. I will mention that a conservative percentage would be anywhere from 2-4% and not to exceed 10%.
Regarding keyword density, I recommend:
- Keyword in the title
- Keyword in the description
- Keyword in the H1 tag
- A few variations of keywords sprinkled throughout the text, where it sounds natural and makes sense
Keyword Frequency: Keyword frequency is the number of times a particular keyword appears on a website. Common sense may indicate the more times a keyword, or keyword phrase is found on a web page, the better the site will rank in Google.
  • Matt Cutts - Matt is an amazingly friendly and absurdly accessible guy given his position as the head of Google's search quality team.

Example Vertical Search Engines

General large scale web search is just one type of search. There are many other types of search engines and information organization tools, for example
  • the Yellow Pages
  • Television program guides
  • directories like DMOZ, the Yahoo! Directory, LII, or specialty directories
  • encyclopedia type sites like Wikipedia
  • large general structured databases like Google Base
  • shopping search like Froogle
  • local search like Google Local
  • news search like Yahoo! News
  • blog search like Technorati
  • tag search like Del.icio.us
  • video search like YouTube
  • photo search like Flickr
  • meme trackers like Techmeme
  • social communities like Digg
  • social networks like MySpace

  • some people may also rely on an individual content channel or a group of them to find the most interesting things and deliver it through daily updated content streams
On page Optimization
 
<html> 

<head> 

<title> Here is the title of the HTML document. </title> 

<meta name="description" content=""> 
<meta name="keywords" content=" ">
<meta name="robots" content="noodp">
<meta name="robots" content="index, follow">

</head>
<body>
Here is the body of the HTML document.
</body>
</html>


HTML Hyperlink
<a href="sample.html">This link</a>

<A HREF="newwindow.html" TARGET="_blank">a new window</A>


HTML Image

<img src="images/aboutus.gif" width="111" height="45" border="0" alt="Web Development India" Title="Outsourcing Website Development India">
page emphasis
<b>bold</b>
<u>underline</u>

<i>italicized</i>



HEADER TAG
 
<h1>...<h6> tag




Introduction
"On Page" SEO simply refers to the text and content on your web site pages. Basically editing your page and content so search engine can find your web page.
   On page SEO is making sure that your website is as search engine friendly as possible. If your website is not optimized then you have less chance of getting good results in the search engines, here is a quick guide towards on page SEO

TYPE of On PAGE SEO techniques

n  Positive On Page SEO / White Hat SEO Techniques
n  Negative On Page / Black Hat SEO Techniques

On PAGE SEO RANKING FACTORS

There are many factors that Search Engine uses to rank pages in the search results (SERPs). Here we will see the main On-Page SEO ranking factors.
§  URL
n  Page Title
n  Meta Tag
n  Content
n  Anchor text
n  Heading Tag
n  Image

URL

Improve the structure of your URLs
The URL of a document should ideally be as descriptive and brief as possible
Poor Practice

URLs like these can be confusing and unfriendly.
Good Practice

The highlighted words above could inform a user or search engine what the target page is about before following the link

Good practices for URL structure
Use words in URLs - URLs with words that are relevant to your site's content and structure are friendlier for visitors.
Avoid:
• using lengthy URLs with unnecessary parameters and session IDs
• choosing generic page names like "page1.html"
• using excessive keywords like "baseball-cards-baseball-cards-baseballcaerds.htm"
Create a simple directory structure - Use a directory structure that organizes your content well and is easy for visitors to know where they're at on your site.
Avoid:
• having deep nesting of subdirectories like ".../dir1/dir2/…/dir8/page.html"
• using directory names that have no relation to the content in them

Provide one version of a URL to reach a document - To prevent users from linking to one version of a URL and others linking to a different version
Avoid:
• having pages from subdomains and the root directory (e.g. "domain.com/
page.htm" and "sub.domain.com/page.htm") access the same content
• mixing www. and non-www. versions of URLs in your internal linking structure
• using odd capitalization of URLs (many users expect lower-case URLs and
remember them better)

page title

Create unique, accurate page titles
A title tag tells both users and search engines what the topic of a particular page is.

The title of the homepage for our baseball card site, which lists the business name and three main focus areas
Good practices for page title tags
o  Accurately describe the page's content - Choose a title that effectively communicates the topic of the page's content.
n  Avoid:
o  choosing a title that has no relation to the content on the page
o  using default or vague titles like "Untitled" or "New Page 1"
o  Create unique title tags for each page - Each of your pages should ideally have a unique title tag, which helps engine know how the page is distinct from the others on your site.
n  Avoid:
o  using a single title tag across all of your site's pages or a large group of pages
o  Use brief, but descriptive titles - Titles can be both short and informative. If the title is too long, engine will show only a portion of it in the search result.
n  Avoid:
o  using extremely lengthy titles that are unhelpful to users
o  stuffing unneeded keywords in your title tags

META TAG

 

Make use of the description" meta tag
A page's description meta tag gives search engines a summary of what the page is about.

Good practices for description meta tags
o  Accurately summarize the page's content - Write a description that would both inform and interest users if they saw your description meta tag as a snippet in a search result.
n  Avoid:
o  writing a description meta tag that has no relation to the content on the page
o  using generic descriptions like "This is a webpage" or "Page about baseball cards"
o  filling the description with only keywords
o  copy and pasting the entire content of the document into the description meta tag
o  Use unique descriptions for each page - Having a different description meta tag for each page helps both users and engine.
n   Avoid:
o  using a single description meta tag across all / (more that one) page of your site's pages

Content

Offer quality content and services

Creating compelling and useful content will likely influence your website more than any of the other factors discussed here. Users know good content when they see it and will likely want to direct other users to it. This could be through blog posts, social media services, email, forums, or other means.



Good practices for content
n  Write easy-to-read text - Users enjoy content that is well written and easy to follow.
o  Avoid:
n  writing sloppy text with many spelling and grammatical mistakes
n  embedding text in images for textual content (users may want to copy and paste the text and search engines can't read it)

n  Stay organized around the topic - It's always beneficial to organize your content so that visitors have a good sense of where one content topic begins and another ends. Breaking your content up into logical chunks or divisions helps users find the content they want faster.
n  Avoid:
n  dumping large amounts of text on varying topics onto a page without
n  paragraph, subheading, or layout separation

n  Use relevant language - Think about the words that a user might search for to find a piece of your content. Users who know a lot about the topic might use different keywords in their search queries than someone who is new to the topic. For example, a long-time baseball fan might search for [nlcs], an acronym for the National League Championship Series, while a new fan might use a more general query like [baseball playoffs]. Anticipating these differences in search behavior and accounting for them while writing your content.

n  Create fresh, unique content - New content will not only keep your existing visitor base coming back, but also bring in new visitors.
n  Avoid:
n  rehashing (or even copying) existing content that will bring little extra value to users
n  having duplicate or near-duplicate versions of your content across your site
n  Offer exclusive content or services - Consider creating a new, useful service that no other site offers. You could also write an original piece of research, break an exciting news story, or leverage your unique user base.

n  Create content primarily for your users, not search engines - Designing your site around your visitors' needs while making sure your site is easily accessible to search engines usually produces positive results.
n  Avoid:
n  inserting numerous unnecessary keywords aimed at search engines but are annoying or nonsensical to users
n  having blocks of text like "frequent misspellings used to reach this page" that add little value for users
n  deceptively hiding text from users, but displaying it to search engines

Anchor Text

Write better anchor text
This text tells users and engines something about the page you're linking to.

Good practices for anchor text
o  Format links so they're easy to spot - Make it easy for users to distinguish between regular text and the anchor text of your links. Your content becomes less useful if users miss the links or accidentally click them.
o  Avoid:
n  using CSS or text styling that make links look just like regular text

o  Think about anchor text for internal links too - You may usually think about linking in terms of pointing to outside websites, but paying more attention to the anchor text used for internal links can help users and engine navigate your site better.
o  Avoid:
n  using excessively keyword-filled or lengthy anchor text just for search engines

HEading Tag

Use heading tags appropriately
Heading are used to present structure on the page to users.
Good practices for heading tags
Imagine you're writing an outline
Use headings sparingly across the page
    Avoid:
• excessively using heading tags throughout the page
• putting all of the page's text into a heading tag
• using heading tags only for styling text and not presenting structure

Image

Optimize your use of images
Text styles that cannot be indexed by search engines include images. And if you used image in your site then give proper name to image file and use useful alt text.
Why use this attribute? If a user is viewing your site on a browser that doesn't support images, or is using alternative technologies, such as a screen reader, the contents of the alt attribute provide information about the picture.
Good Practice

Negative on page seo techniques

There are several different techniques known as "black hat" or "unethical" On Page Techniques. Some SEO companies engage in these type of activities and should be avoided. Sooner or later the search engines will catch up to these unethical techniques and the likely result will be your site being demoted or banned from the search engines.

Negative ON Page SEO/ Black Hat SEO Techniques       

·         Avoid Using "hidden" or invisible text on your page for the purpose of higher the words are "hidden" from the surfer. search engine placement. For example the words/text for search phrase "Widget" in the html, the font color has been set to White. The background of the page is also white. Therefore the textual content is actually there, however This is frowned upon by search engines and frequently results in your site being penalized
·         Avoid Using Negative <div> htmltags. Div tags, Div tags are division tags. Unscrupulous seo services may insert them into your page with negative x/y coordinates to place content outside of the visible page for the surfer, but the text itself is in the html page. The search engine finds the keywords in the text, yet the surfer does not see it. Again a technique to be avoided and not recommended under any circumstances
·         Avoid Cloaking or Sneaky Redirects. Cloaking refers to serving up 2 different types of content based on the visitor who is visiting. Is the visitor a regular web surfer, serve up this page. Is the visitor a search engine spider? Serve up this OTHER page specifically for the search engine spider. The other page being served up is typically garbled textual content with no meaning to a human, and is stuffed with various keywords and search phrases. Again this technique is not recommended and will likely get your site penalized or banned from search engines
·         Avoid duplicate content. Duplicate content means you create one web site, with content on topic a, and then repeat the content over and over again on multiple websites. In theory you could create one website, achieve high ranking on it, and then clog up the search engines with the same content duplicated on multiple domains. Again this is not recommended and should be avoided

Dynamic Websites Optimization
What are "Dynamic Websites"?

Dynamic websites are websites whose pages are generated on the fly. Unlike static pages (primarily .htm/.html pages), dynamic pages are generated when an user triggers an action through that particular page.
Here isa sample dynamic URL-
http://www.bbc.co.uk/cgi-bin/search/results.pl?q=%22dynamic+websites%22&tab=news &go=homepage
As per the above example of www.bbc.co.uk, the dynamic part (i.e. the part) of the URL which changes as per surfer request is the part after the question mark (?)

What are the problems that search engines face in indexing Dynamic URLs?

1. Search engines often consider a dynamic URL as an infinite set of links.

2. Since dynamic URLs find maximum application in online shopping carts, there is a possibility of incorporating a session id to a particular page. As session ids of that particular page change, the search engine spider needs to index an infinite number of copies of the same page, which is a Herculean task for them.

3. Proceeding with the same logic presented in point #2, indexing the same dynamic page might overload the servers of the search engines and therefore prevent the search engines to present with the most relevant information in the fastest possible time.
Here is what Google says about indexing of dynamic websites-
Reasons your site may not be included: Your pages are dynamically generated. We are able to index dynamically generated pages. However, because our web crawler can easily overwhelm and crash sites serving dynamic content, we limit the amount of dynamic pages we index. (Source - http://www.google.com/webmasters/)

What are the options that you have in order to make a search engine spider index your Dynamic URLs?

1. Use of softwares - Exception Digital Enterprise Solutions (http://www.xde.net ) offers a software which can change the dynamic URLs to static ones. Named XQASP, it will remove the "?" in the Query String and replace it with "/", thereby allowing the search engine spiders to index the dynamic content.

Example -
http://www.my-online-store.com/books.asp?id=1190 will change to
http://www.my-online-store.com/books/1190.

The latter being a static URL, it can easily be indexed by the search engine spiders.

2. Use of CGI/Perl scripts - One of the easiest ways to get your dynamic sites indexed by search engines is using CGI/Perl scripts. Path_Info or Script_Name is a variable in a dynamic application that contains the complete URL address (including the query string information). In order to fix this problem, you'll need to write a script that will pull all the information before the query string and set the rest of the information equal to a variable. You can then use this variable in your URL address.

Example - http://www.my-online-store.com/books.asp?id=1190

When you are using CGI/Perl scripts, the query part of the dynamic URL is assigned a variable. So, in the above example "?id=1190" is assigned a variable, say "A". The dynamic URL http://www.my-online-store.com/coolpage.asp?id=1190 will change to http://www.my-online-store.com/books/A through CGI/Perl scripts which can easily be indexed by the search engines.

3. Re-configuring your web servers-

Apache Server - Apache has a rewrite module (mod_rewrite) that enables you to turn URLs containing query strings into URLs that search engines can index. This module however, isn't installed with Apache software by default, so you need to check with your web hosting company for installation.

ColdFusion - You'll need to reconfigure ColdFusion on your server so that the "?" in a query string is replaced with a '/' and pass the value to the URL.

4. Creation of a Static Page linked to an array of dynamic Pages - This approach is very effective, especially if you are the owner of a small online store selling a few products online. Just create a static page linking to all your dynamic pages. Optimize this static page for search engine rankings. Include a link title for all the product categories, place appropriate "alt" tag for the product images along with product description containing highly popular keywords relevant to your business (you can conduct keyword research for your site through http://www.wordtracker.com). Submit this static page along with all the dynamic pages in various search engines, conforming to the search engine submission guidelines.
How Amazon.com, Earth's Biggest Bookstore, coped with the issue of indexing of dynamic URLs?

A search in Google for internet marketing books, yielded a result that takes you directly to the appropriate dynamic page at Amazon - http://www.amazon.com/exec/obidos/ISBN%3D0395683297/103-0475212-8205437.
Since the above URL does not contain any query strings, all search engines can index Amazon.com's products. Amazon.com uses this method to get its product selections indexed by search engines. This is very important for Amazon, because being an online bookstore, it is very natural for them to adopt dynamic URLs yet it was equally important for them to make their dynamic URLs search engine index friendly.

Conclusion
Even a few years back, most of the major search engines did not index dynamic URLs, thereby often preventing top search engine rankings for the online stores. With Google starting to index dynamic URLs a few months ago, the picture is going to change in the coming days. This is more so because Google's numero uno position is currently being threatened by Microsoft's MSN (developing its own search engine) and Yahoo! who recently acquired Overture, the biggest player in the PPC Search Engine industry.
To read more articles click on http://www.searchengine-optimization-guru.com/seo-whitepapers.html

===============================================================
Robots.txt Concept

User-agent: *
Disallow:
User-agent: Googlebot
Disallow: /
User-agent: Slurp
Disallow: /
User-agent: Teoma
Disallow: /
Disallow: /login/

.htaccess file

RewriteCond %{THE_REQUEST} ^[A-Z]{3,9}\ /index\.html\ HTTP/
RewriteRule ^index\.html$
http://www.abc.com/ [R=301,L]

XML Sitemap Generator


 

This checklist provides 50 simple ways for you to make your website even better!

1. Use .htaccess to redirect the non-www version of your website to the www version.
2. Extend the registration of your domain for at least five years.
3. Make sure that when a visitor clicks your website’s logo, they are taken back to the homepage.
4. If you think your font is too small, make it bigger!
5. Remove widgets that aren’t directly enhancing your website.
6. Include an easy to find contact link, e-mail address or phone number on every page of your website.
7. Use a CSS stylesheet to make your website printer friendly.
8. Fix or remove broken links (external and internal).
9. Add ALT tags to all of your images.
10. If your website is using frames, get rid of them!
11. If you don’t have a robots.txt file, create one.
12. Provide the same basic navigation menu on every page of your website.
13. Use the same color for links throughout your website.
14. Spell check your content.
15. Format your content so that it is easy to scan.
16. Add at least one picture to every page of your website.
17. If you don’t have an About page, create one.
18. Use your About page to show your brand’s personality.
19. Include at least three pictures on your About page.
20. If your website has pop-up windows, get rid of them!
21. Link to relevant social media profiles from your About page.
22. Add a search box near the top of your website’s layout.
23. Create a privacy policy page.
24. Place a link to your privacy policy at the bottom of every page.
25. Use dashes in filenames instead of underscores.
26. Add a link to Google Maps on your Contact page.
27. Add internal links from one piece of content to another.
28. Write a unique META description for all of your main pages.
29. Validate your source code and correct any errors.
30. If you have a link exchange page, get rid of it!
31. Add a site search box to your 404 page.
32. Create a sitemap for your website.
33. Add a link to your sitemap at the bottom of every page of your website.
34. Add a copyright notification to the bottom of every page of your website.
35. Underline your links throughout your website.
36. Turn off music that automatically plays.
37. Get rid of any flashing GIF animations.
38. Include your website logo at the top of every page.
39. Register the .net and .org versions of your domain and redirect them to your website.
40. If you don’t have a help or support section, create one!
41. When you answer a question via e-mail, add the answer to the help/support section.
42. Look at your contact forms and see if there is any information you don’t have to request from your visitors.
43. Add your analytics code to every page of your website.
44. Get rid of any pages that automatically resize visitors’ browser window.
45. Create a favicon for your website.
46. Only use email address that include @yourdomain.
47. Label each box of any forms on your website.
48. Show a confirmation any time a visitor completes a significant action on your website.
49. Create one new piece of quality content for your website every week.

50. Leave a comment with your suggestion for the fiftieth tip!