HTML Error Codes - From Google Webmaster Tools



Status codes that indicate a provisional response and require the requester to take action to continue.

100 (Continue) The requestor should continue with the request. The server returns this code to indicate that it has received the first part of a request and is waiting for the rest.
101 (Switching protocols) The requestor has asked the server to switch protocols and the server is acknowledging that it will do so.

2xx status codes

Status codes that indicate that the server successfully processed the request.

200 (Successful) The server successfully processed the request. Generally, this means that the server provided the requested page. If you see this status for your robots.txt file, it means that Googlebot retrieved it successfully.
201 (Created)
The request was successful and the server created a new resource.
202 (Accepted) The server has accepted the request, but hasn't yet processed it.
203 (Non-authoritative information) The server successfully processed the request, but is returning information that may be from another source.
204 (No content) The server successfully processed the request, but isn't returning any content.
205 (Reset content) The server successfully proccessed the request, but isn't returning any content. Unlike a 204 response, this response requires that the requestor reset the document view (for instance, clear a form for new input).
206 (Partial content) The server successfully processed a partial GET request.

3xx status codes

Further action is needed to fulfill the request. Often, these status codes are used for redirection. Google recommends that you use fewer than five redirects for each request. You can use Webmaster Tools to see if Googlebot is having trouble crawling your redirected pages. The Web crawl page under Diagnostics lists URLs that Googlebot was unable to crawl due to redirect errors.

300 (Multiple choices)
The server has several actions available based on the request. The server may choose an action based on the requestor (user agent) or the server may present a list so the requestor can choose an action.
301 (Moved permanently)
The requested page has been permanently moved to a new location. When the server returns this response (as a response to a GET or HEAD request), it automatically forwards the requestor to the new location. You should use this code to let Googlebot know that a page or site has permanently moved to a new location.
302 (Moved temporarily)
The server is currently responding to the request with a page from a different location, but the requestor should continue to use the original location for future requests. This code is similar to a 301 in that for a GET or HEAD request, it automatically forwards the requestor to a different location, but you shouldn't use it to tell the Googlebot that a page or site has moved because Googlebot will continue to crawl and index the original location.
303 (See other location) The server returns this code when the requestor should make a separate GET request to a different location to retrieve the response. For all requests other than a HEAD request, the server automatically forwards to the other location.
304 (Not modified)

The requested page hasn't been modified since the last request. When the server returns this response, it doesn't return the contents of the page.

You should configure your server to return this response (called the If-Modified-Since HTTP header) when a page hasn't changed since the last time the requestor asked for it. This saves you bandwidth and overhead because your server can tell Googlebot that a page hasn't changed since the last time it was crawled.

305 (Use proxy)
The requestor can only access the requested page using a proxy. When the server returns this response, it also indicates the proxy that the requestor should use.
307 (Temporary redirect) The server is currently responding to the request with a page from a different location, but the requestor should continue to use the original location for future requests. This code is similar to a 301 in that for a GET or HEAD request, it automatically forwards the requestor to a different location, but you shouldn't use it to tell the Googlebot that a page or site has moved because Googlebot will continue to crawl and index the original location.

4xx status codes

These status codes indicate that there was likely an error in the request which prevented the server from being able to process it.

400 (Bad request)

The server didn't understand the syntax of the request.
401 (authentication error) The page requires authentication. It's likely that you do not intend this page to be indexed. If this page is listed in your Sitemap, you can remove it. However, if you leave it in your Sitemap, we won't crawl or index it (although the page will continue to be listed here with this error).
403 (Forbidden)

The server is refusing the request. If you see that Googlebot received this status code when trying to crawl valid pages of your site (you can see this on the Web crawl page under Diagnostics in Google Webmaster Tools), it's possible that your server or host is blocking Googlebot's access.
404 (Not found)

The server can't find the requested page. For instance, the server often returns this code if the request is for a page that doesn't exist on the server.

If you don't have a robots.txt file on your site and see this status on the robots.txt page of the Diagnostic tab in Google Webmaster Tools, this is the correct status. However, if you do have a robots.txt file and you see this status, then your robots.txt file may be named incorrectly or in the wrong location. (It should be at the top-level of the domain and named robots.txt.)

If you see this status for URLs that Googlebot tried to crawl (on the HTTP errors page of the Diagnostic tab), then Googlebot likely followed an invalid link from another page (either an old link or a mistyped one).

405 (Method not allowed)

The method specified in the request is not allowed.

406 (Not acceptable)
The requested page can't respond with the content characteristics requested.
407 (Proxy authentication required) This status code is similar to 401 but specifies that the requestor has to authenticate using a proxy. When the server returns this response, it also indicates the proxy that the requestor should use.
408 (Request timeout) The server timed out waiting for the request.
409 (Conflict) The server encountered a conflict fulfilling the request. The server must include information about the conflict in the response. The server might return this code in response to a PUT request that conflicts with an earlier request, along with a list of differences between the requests.
410 (Gone) The server returns this response when the requested resource has been permanently removed. It is similar to a 404 (Not found) code, but is sometimes used in the place of a 404 for resources that used to exist but no longer do. If the resource has permanently moved, you should use a 301 to specify the resource's new location.
411 (Length required) The server won't accept the request without a valid Content-Length header field.
412 (Precondition failed) The server doesn't meet one of the preconditions that the requestor put on the request.
413 (Request entity too large) The server can't process the request because it is too large for the server to handle.
414 (Requested URI is too long) The requested URI (typically, a URL) is too long for the server to process.
415 (Unsupported media type) The request is in a format not support by the requested page.
416 (Requested range not satisfiable) The server returns this status code if the request is for a range not available for the page.
417 (Expectation failed) The server can't meet the requirements of the Expect request-header field.

5xx status codes

These status codes indicate that the server had an internal error when trying to process the request. These errors tend to be with the server itself, not with the request.

500 (Internal server error) The server encountered an error and can't fulfill the request.
501 (Not implemented) The server doesn't have the functionality to fulfill the request. For instance, the server might return this code when it doesn't recognize the request method.
502 (Bad gateway) The server was acting as a gateway or proxy and received an invalid response from the upstream server.
503 (Service unavailable) The server is currently unavailable (because it is overloaded or down for maintenance). Generally, this is a temporary state.
504 (Gateway timeout) The server was acting as a gateway or proxy and didn't receive a timely request from the upstream server.
505 (HTTP version not supported) The server doesn't support the HTTP protocol version used in the request.

Google Crawling - My WebPage is 1 Trillion+1 Indexed


1,000,000,000,000+1 is my Web Page
1 Trillion Webpages + 1 is my webpage (MYGoogleOnline)
Just for Fun

Google 1 Trillion Web Pages in its Index - Still Crawling

Google vs Internet

We know that Internet is huge. Google started its index in 1998 by 26 million pages and by 2000 Google reached the index of around 1 billion. Over the past eight years Google have seen a lot of content and information in the web. Recently Google engineers was astonishing about the web in recent internet revolution, they proudly announces that Google systems over reached a milestone of 1 trillion (as in 1,000,000,000,000) unique URL's in its index

Google Crawling for Informations

Google engineers explain how they crawl the above mentioned dream number. They started indexing the URL's with some initial pages and then they start follow the links of each pages they indexed, then they followed the links from the new pages they indexed and so on. To be frank they have more than 1 trillion pages but they are not sure that all the pages are unique and have only one URL because some of the pages have two or more URL's. So we avoid all those URL's and pages and reached 1 trillion and we are indexing billions of unique pages and unique links every day.

Infinite Internet Web Pages

To be open minded, web is infinite so we can’t crawl all the pages but we will try our level best to crawl all the unique pages.

They also admit that all the trillion pages are not unique because in Google calendar everyday it will show up new pages but it will mostly have same content so we can't avoid such criteria. But they are very much proud that they have the most comprehensive index of any search engines, and their goal is to crawl all the world information’s and display to searchers.

Some of the important points from Googlers:

“To keep up with this volume of information, Google systems have come a long way since the first set of web data Google processed to answer queries. Back then, we did everything in batches: one workstation could compute the PageRank graph on 26 million pages in a couple of hours, and that set of pages would be used as Google's index for a fixed period of time. Today, Google downloads the web continuously, collecting updated page information and re-processing the entire web-link graph several times per day. This graph of one trillion URLs is similar to a map made up of one trillion intersections. So multiple times every day, we do the computational equivalent of fully exploring every intersection of every road in the United States. Except it'd be a map about 50,000 times as big as the U.S., with 50,000 times as many roads and intersections.
As you can see, our distributed infrastructure allows applications to efficiently traverse a link graph with many trillions of connections, or quickly sort petabytes of data, just to prepare to answer the most important question: your next Google search.”


From this we can come to a conclusion that Google the leader of all search engines and it will go on to search for more information and index trillions and zillions of pages in its system for answering to the customer queries.

Search Engine Optimization Tips from Matt Cutts for SEO Freshers or Beginners

Matt Cutts says "One thing doesn't have to do with the other,"

1. Search term on the page is very important - Highlight top search keywords

"Think about what people are going to type in to try and find you," Cutts says. He tells of meeting a chiropractor from San Diego who complained that his site couldn't be found easily using Google search. The words "San Diego chiropractor" were listed nowhere on his site. "You have to make sure the keywords are on the page," Cutts says. If you're a San Diego doctor, Des Moines architect or Portland ad agency, best to let people know so immediately, at the top of your page.

2. Tags are very important especially meta tags for the pages

San Diego ChiropractorWhen creating websites, Internet coding language includes two key tags: title and description. Even if you don't know code, which is used to create pages, software programs such as Adobe's Dreamweaver have tools that let you fill them in in plain English (rather than "

If Cutts' chiropractor had properly tagged his Web page, a search would have returned something like this: "San Diego chiropractor. Local doctor serves San Diego community."

There's also a third tag, to add keywords, or search terms, but Cutts says Google doesn't put much weight in its rankings on that one.

3. Links from other websites is very important - External Links to your website

Google says it looks at more than 100 pieces of data to determine a site's ranking. But links are where it's at, once your search terms are clearly visible on your site and the title and description tags correctly marked.

In a nutshell: Google ranks sites based on popularity. If authoritative sites link to you, you must be good, and therefore you get to the top of the list. If you can't get top sites such as USATODAY.com or The New York Times to link to you, try your friends. And what if they don't have a site? They probably do. Read on.

4. Blogging for the website - Continuous blog post is necessary

Cutts says blogging is a great way to add links and start a conversation with customers and friends. It will cost you only time: Google's Blogger, WordPress and others offer free blogging tools. With a blog, you can link back to your site and offer links to others. It's also a great way to start building content, Cutts says.

5. Sign up for Google Free Services (Webmasters, Google Analytics)

Google's google.com/webmaster offers freebies to help get your site found. You can upload a text-based site map, which shows Google the pages of your site (create it at www.xml-sitemaps.com). Once that's done, you'll be registered with Google, where you can learn vital statistics — including who is linking to your site and how often Google "crawls" your site for updates.

Google's Local Business center (google.com/local/add) is the place for business owners to submit a site so it shows up in local searches, with a map attached. Savvy consumers who use Google for searches know that the first 10 non-advertising results often are from Google Maps, so if you have a business and haven't submitted it, you're losing out on potential customers.

Avoid Stuffing of any information

When weaving keywords into a main page, Cutts says, some zealous Web publishers will use the term over and over again. That's called "keyword stuffing." It's a big Google no-no that can have your site removed from the index.

"After you've said it two or three times, Google has a pretty good idea — 'OK, this page has something to do with this keyword,' " he says. "Just think about the two or three phrases you want to be known for and weave that in naturally."

For blogger newbies, Cutts knows that writing (for example, posting new material) doesn't always come easy. He suggests finding ideas by visiting social news sites such as Digg and StumbleUpon, to see what people are saying about your particular topic.

Aside from that, Cutts says, new material falls into the common-sense category: It's all about your business. "If I'm a plumber in Iowa, I may want to write about some of the strange things that happen to me on the job, or the five most common ways to fix a toilet," he says. "That kind of content can get really popular, and it's a great way to get links." Folks will post your piece on one of the social media sites. And with links comes higher Google rankings.

Finally, Cutts says, there is one big misconception about getting Google visibility that he wants to clear up: In order to be found at the top of Google's rankings, you do not also have to advertise.

Google Search Marketing: Google AdWords Vocabulary

Most important vocabulary in Google Adwords

• Ad copy
• AdSense
• AdWords
• Algorithms
• Cached
• Campaign
• Channel
• Contextual placement
• Display URL
• Distribution
• GAP
• Google Advertising Professional
• Google content network
• Google Network
• Google search button
• Google search network
• Headline
• Indexing
• Keyword
• Keyword targeted
• Marketing mix
• PageRank™
• Query
• Reach
• Relevance
• Return on Investment (ROI)
• Search Engine Marketing
• Search result
• Search results page
• SEM
• Similar Pages
• Sponsored Links
• Target
• Uniform Resource Locator (URL)
• Unique visitors
• Users

Google - Top Websites of Overall Internet Users June 2008

Google market share of the total internet users of United States is increasing month by month from it's top rivals Yahoo and Msn.

In the month of June World most lovable search engine is the hot favorite for all the U.S users below is the new statistical report



(Thanks: Nielsen Online)

According to
Nielsen Online

Google's internet users reached from 127.6 million in May to 128 million in June

Meanwhile, Microsoft dropped from its visitors rating a little from 123.3 million to 123 million while Yahoo has a huge change from last month of 115.6 million to 113.4 million in the month of June.

Leader if the amount of time spent on the top search search engines is Yahoo but it also lags from last month stating as below

Yahoo in the month of May it was 3 hours and 13 minutes and in the month of June 3 hours and 7 minutes

Microsoft in the month of May it was 2 hours and 16 minutes and in the month of June 2 hours and 12 minutes

Google in the month of May it was 1 hour and 49 minutes and in the month of June 1 hour and 50 minutes.

Google Affiliate Network - Double Click performics

We are happy to announce that Google has acquired DoubleClick. For more information related to the acquisition, Performics



Google Affiliate Network offers:

  • Industry-leading advertiser and publisher service
  • Rigorous network quality standards
  • Reliable conversion tracking
  • Detailed reporting
  • Automated payments to publishers
  • Simple link and creative delivery tools
  • Flexible commissions
  • Access to top brands on the web

Ready to learn more?


Google Time Search Engine Results is Getting Better for access

Google Timely search option is getting better and better and the crawling is very fast and more deep for the search results...




The Google Trifecta: Webmaster Tools, Analytics, Website Optimizer - Google Online Seminar

For the first time ever, the Google Webmaster Tools, Analytics, and Website Optimizer teams will join forces to hold a free online seminar. By using all three of these tools together, you can gain a comprehensive understanding of your pages' visibility on Google, how your visitors interact with your site, and what content will most effectively get them to take the desired action.

During this online seminar, team members will:

* Provide brief introductions to their products
* Highlight recent product updates and releases
* Discuss the benefits of using all three products together
* Answer common questions submitted by attendees

Upon registration, we also invite you to submit any specific topics you'd like them to cover. We look forward to seeing you there!

For more details visit the website: Google Trifecta

Google Announces Crawling of Flash!!!!!!!!!!! Hurray SEO

Improved Indexing of Flash from Google!!!!!!! SEO Guys Cheer up…

Google officially announced that they received received numerous requests to improve our indexing of Adobe Flash files. SO Ron Adler and Janis Stipins—software engineers on our indexing team provided useful informations related to the improvement in indexing flash.

Below are the Questions and answers from the Google indexing team software engineers:

The details are taken from Google webmaster central Blog


Q: Which Flash files can Google better index now?
We've improved our ability to index textual content in SWF files of all kinds. This includes Flash "gadgets" such as buttons or menus, self-contained Flash websites, and everything in between.

Q: What content can Google better index from these Flash files?
All of the text that users can see as they interact with your Flash file. If your website contains Flash, the textual content in your Flash files can be used when Google generates a snippet for your website. Also, the words that appear in your Flash files can be used to match query terms in Google searches.

In addition to finding and indexing the textual content in Flash files, we're also discovering URLs that appear in Flash files, and feeding them into our crawling pipeline—just like we do with URLs that appear in non-Flash webpages. For example, if your Flash application contains links to pages inside your website, Google may now be better able to discover and crawl more of your website.

Q: What about non-textual content, such as images?
At present, we are only discovering and indexing textual content in Flash files. If your Flash files only include images, we will not recognize or index any text that may appear in those images. Similarly, we do not generate any anchor text for Flash buttons which target some URL, but which have no associated text.

Also note that we do not index FLV files, such as the videos that play on YouTube, because these files contain no text elements.


Q: How does Google "see" the contents of a Flash file?
We've developed an algorithm that explores Flash files in the same way that a person would, by clicking buttons, entering input, and so on. Our algorithm remembers all of the text that it encounters along the way, and that content is then available to be indexed. We can't tell you all of the proprietary details, but we can tell you that the algorithm's effectiveness was improved by utilizing Adobe's new Searchable SWF library.


Q: What do I need to do to get Google to index the text in my Flash files?
Basically, you don't need to do anything. The improvements that we have made do not require any special action on the part of web designers or webmasters. If you have Flash content on your website, we will automatically begin to index it, up to the limits of our current technical ability (see next question).

That said, you should be aware that Google is now able to see the text that appears to visitors of your website. If you prefer Google to ignore your less informative content, such as a "copyright" or "loading" message, consider replacing the text within an image, which will make it effectively invisible to us.

Q: What are the current technical limitations of Google's ability to index Flash?
There are three main limitations at present, and we are already working on resolving them:

1. Googlebot does not execute some types of JavaScript. So if your web page loads a Flash file via JavaScript, Google may not be aware of that Flash file, in which case it will not be indexed.
2. We currently do not attach content from external resources that are loaded by your Flash files. If your Flash file loads an HTML file, an XML file, another SWF file, etc., Google will separately index that resource, but it will not yet be considered to be part of the content in your Flash file.
3. While we are able to index Flash in almost all of the languages found on the web, currently there are difficulties with Flash content written in bidirectional languages. Until this is fixed, we will be unable to index Hebrew language or Arabic language content from Flash files.

We're already making progress on these issues, so stay tuned!

Still there are some issues but they gave the assurance that they will rectify all the issues related to indexing of flash soon. I am sure that this will be a turnaround for the search engine giant and it is again making sure that they are the top most performers in the search engine field.

Thanks to Google for making it happen………………….


Sites.Google.com launched by Google

Create websites using Sites.google.com from Google

What is Google Sites?

Google Sites is an online application helps creating a web site as easy as editing a document.

Using the website we can share variety of information such as videos, calendars, presentations, attachments, and text and it is easy to view and share with a small group, their entire organization, or the world.

Google sites - A new application from google. So it will take some time to get name in the internet market.

The search engine gaint created this application with JotSpot which is now a part of Google.

Google Sites Help Center:

http://www.google.com/support/sites/

Google Sites Features are as follows:

  • Creating a new page is very simple as clicking a button
  • HTML knowledge is not required because it will similar to edit a document
  • They offer templates which can be edited and we can change the colors but we cannot insert the template as we need.
  • The entire layout is customizable so you can change the layout as needed.
  • We can upload files up to 10mb in size.
  • We can insert Google Calendar, Document, Picasa slides, Presentations, Spreadsheet, Video and more gadgets.
  • Sitemap is created automatically.
  • Navigation and recent activities will be displayed at the left or right side bar.
  • If you want to make the website private to your organization then you can select the option private.
  • We can create Web Pages, Dashboard, Announcements, File Cabinet and List.
  • We can put comments to all the postings or web pages we created.
  • Footer links are created automatically
  • URL normalization is present in the website.
  • We can request more number of contributors to the Google Sites.

SEO Point of View

In blogger we need to insert a code for unique title tags but here each page will be having separate title tags.

We can infuse the Google Analytics immediately.

Alt tags can be inserted for the images.

We can easily edit the documents and create internal links and external links from Google sites.

Linking structure is automatically generated in the form of Sitemap

Still lot of updates required.

URL Normalization is very good and we can use the keywords in the URL.

Announcements and Comments are very useful.

Domain is from Google so it is an added advantage to crawl the pages.

We can use this as a promotional way for business similar to blogger.

We can create this application for the Clients.

Suggestions for Google:

How we can edit and place the webmaster sitemap

We can have an option for infusing the template

Site elements should have more option for editing the pages.

The application seems like an updated version for GooglePages

This is a new application still not more information related to the usage of this application.

MY Google Online SEO Tips

SMO Conventions - Social Media Optimization

As now we have gained a generalized idea about SMO, let’s not overlook the rules that have to be followed while providing SMO services to the clients. These rules were first generated by Rohit Bhargava and then Jeremiah Owyang, Cameron Olthuis, Lee Baker, and Lee Oden joined the suit and developed it further.

  1. Increase your linkability- An optimized site for social media should have enhanced linkability of the content. You can add blogs, aggregate content or linkbait to achieve linkability.
  2. Make tagging and bookmarking easy- Use discretion and logic for tagging and bookmarking.
  3. Reward inbound links- List blogs which link back to you via permalinks, trackbacks or recently linking blogs (like the Yahoo & Google blogs do).
  4. Help your content travel- Let your portable content move freely over the social media.
  5. Encourage the mashup- Allow further use of your content and RSS feeds to create mashups for traffic generation.
  6. Be a User Resource, even if it doesn’t help you.
  7. Reward helpful and valuable users.
  8. Actively participate in the discussion of blogs and sites of others.
  9. Know how to target your audience.
  10. Create content.
  11. Be real.
  12. Don’t forget your roots, be humble.
  13. Don’t be afraid to try new things, stay fresh.
  14. Develop a SMO strategy.
  15. Choose your SMO tactics wisely. Be cognizant of what actions will influence the desired outcome with the most impact.
  16. Make SMO part of your process and best practices.
By
Rinki Singh Yadav