Microsoft adCenter Frequently Asked Questions (FAQs)
Tuesday, December 30, 2008
1
The right keywords for the right audience to capture the right customer
What is the adCenter Add-in for Excel with the Ad Intelligence module?
The adCenter Add-in is a keyword research and optimization tool, which enables users to rapidly build out or expand keyword lists and plan keyword strategy based on a range of important attributes such as relevance, historical cost, volume, geography and demographics. The adCenter Add-in provides the ability to easily research keywords to help you reach the right audience, and capture and convert the right customer.
This document contains:
I. External Frequently Asked Questions (FAQs) Contents
• About adCenter Add-in
• Getting Started
• How it Works
• Using the adCenter Add-in
• Support
• Troubleshooting
II. Glossary of Key Terms
I - External Frequently Asked Questions (FAQs) Contents
About adCenter Add-in
What is the Microsoft adCenter Add-in1. Beta?
Will the adCenter Add-in be available in other markets?2.
What features are found in the adCenter Add-in?3.
Why was the adCenter Add-in released in Beta?4.
Getting Started
What are the system requirements for the adCenter Add-in?5.
How do I download and install the adCenter Add-in?6.
I have an old version of the adCenter Add-in (formerly called adSage). How do I upgrade to the
7. new version?
I’ve installed the adCenter Add-in and now where do I find it?
8.
adCenter Add-inBeta for Excel
Frequently Asked
Questions (FAQs)
2
How it Works
9. How come I don’t always get results when I use the adCenter Add-in?
10. Under the Keyword Suggestion feature, what is the difference between the three algorithms shown
(Campaign Association, Contained and Similarity)?
11. What are the date ranges available for the data provided by the adCenter Add-in?
12. When using the Monthly Traffic feature, how does the adCenter Add-in forecast the traffic for the
three future months?
13. When using the Daily Traffic feature, why do the results not show search impressions more recent
than 3 to 5 days in the past?
14. When using the Monetization feature, how are the output ranges defined for the conditional
formatting of CTR and CPC?
15. When I set Maximum Suggested Keywords to 100 in Keyword Suggestion, why do I sometimes
receive less than this amount?
16. Why do I get duplicate terms in my results for Spiky Keywords?
17. How do I set the Geography function to show state, city, or country?
18. How do I set custom defaults for date ranges or other preferences for the various functions?
19. If I bid the CPCs recommended for Position 1, am I guaranteed the number one spot?
20. Please explain why the query number in Monthly Traffic for a month does not match the
impression data in Monetization?
21. Why does a lower position have a higher average CPC than a higher position?
22. Could my keyword impression count be different than Microsoft adCenter Add-in’s report?
23. Is content match included in impression counts?
24. Why is my search result displayed in a new sheet?
25. How do I compare the scores from two different functions?
26. Is adCenter Add-in information only derived from MSN data?
27. What is the taxonomy used for Keyword Categorization?
28. For results derived from the Monetization feature, why is it that sometimes the impression count is
vastly different for Positions 1, 2, and 3 for terms with high-bid density?
29. How are the impression numbers different between adCenter Add-in forecast data and the research
tab data in adCenter?
30. How accurate is monthly traffic forecasting and how is it done?
31. Are these keyword expansion tools based on actual search queries or keywords in adCenter?
32. Can keywords be classified into more than one category (e.g. “apple” can be both a fruit and a
technology brand)?
3
Using the Add-in
How can I use the adCenter Add-in to quickly generate relevant keywords to describe my web sites?33.
How can I use the adCenter Add-in to quickly create or expand a list of relevant keywords?34.
How can I use the adCenter Add-in to forecast keyword search traffic?35.
How can I use the adCenter Add-in to understand the seasonality of keyword searches?36.
How can I use the adCenter Add-in to understand the localization of keyword searches?37.
How can I use the adCenter Add-in to understand the demographics of keyword searches?38.
How can I use the adCenter Add-in to obtain KPIs for keywords, such as CTR, CPC, Impressions and Position?39.
Support
Where else can I learn more about the most recent updates to the adCenter Add-in?40.
Who do I contact for adCenter Add-in support?41.
Where can I find adCenter Add-in support information?42.
How do I provide feedback and how can I suggest new features for the adCenter Add-in? (Excel 2007)43.
Troubleshooting
Why can’t I find my adCenter Add-in tab in the Excel ribbon? (Excel 2007)44.
I’ve successfully installed the adCenter Add-in, but when I run it I get an error message regarding “invalid user
45. authorization.” How do I resolve this? (Excel 2007)
How do you Optimize your campaign with the adCenter Add-in Beta for Excel?46.
About Microsoft adCenter Add-in
1. What is the Microsoft adCenter Add-in?
The Microsoft adCenter Add-in is a no-cost keyword research and optimization tool that integrates into Excel.
2. Will the adCenter Add-in be available in other markets?
The adCenter Add-in is currently available for adCenter customers in the US market. However, the adCenter Add-in provides data that can be useful in markets other than the US. You are free to download and use the adCenter Add-in even if you conduct business outside of the US.
3. What features are found in the adCenter Add-in?
The adCenter Add-in offers a broad selection of powerful keyword expansion and campaign optimization tools:
Keyword Extraction:• extracts commonly used keywords from a given website address, which allows the user to see what keywords are in popular use on that site.
Keyword Suggestion: • enables quick generation of a large list of relevant keywords from an initial list of keywords entered.
Search Buzz:• returns the top keywords, by vertical market, for the given time period. Two keyword trend research options are offered:
Top Spiky Keywords:• returns the keywords whose trends are sudden and spiking (i.e. not overall popularity, but demonstrate flashing and surging popularity).
4
Top Frequent Keywords: • displays the keyword trends for the overall most popular keywords.
Monthly Traffic: • provides both Monthly Traffic and Daily Traffic.
Monthly Traffic:• allows the user to look at monthly historical and future, forecasted query counts for the given keywords.
Daily Traffic:• provides the user with daily query counts for the given keywords.
Keyword Categorization: • input keywords, and then the algorithm suggests a category based on the taxonomy.
Geographic: • provides location information of searchers for specific keywords (user can specify country, state or city) about the input keywords within a set time period.
Demographic: • provides demographic information on searchers (by gender, age group) about the given keywords within a set time period.
Monetization: • this feature has two sub-features:
Monetization: • allows users to gather monetization information for the given keyword(s). This
feature provides information about clicks, impressions, position, click through rate, cost per click, and match type • within a date range.
Vertical KPIs: • returns KPI’s (key performance indicators) through monetization information about selected verticals within a date range.
Advanced Algorithm: • enables you to set the parameters of all of the features to suit your business needs. Each feature can be set to meet personal preferences or needs.
4. Why was the adCenter Add-in released in Beta?
The adCenter Add-in was released in Beta as a way to gather feedback from customers. We hope that customers enjoy the tool and understand that since this is a Beta version additional functionality will be forthcoming.
Getting Started
5. What are the system requirements for the adCenter Add-in?
You will need the following on your PC:
adCenter Login•
Microsoft Office Excel 2003 or Excel 2007•
Internet access•
6. How do I download and install the adCenter Add-in?
The adCenter Add-in installation instructions and setup program is located at the following site:
http://advertising.microsoft.com/advertising/adCenter_addin (Excel 2007)
http://advertising.microsoft.com/adcenter-beta-pilot-signup (Excel 2003)
7. I have an old version of the adCenter Add-in (formerly called adSage). How do I upgrade to the new version?
To upgrade to the new version, first uninstall the old version and then install the new version. Please see the installation guide for detailed instructions.
8. I’ve installed the adCenter Add-in and now where do I find it?
After the adCenter Add-in is installed, it should automatically be available to you within Excel as a new tab called “Ad Intelligence,” located at the top of the page. Alternatively, the application (including Excel) can be launched by double-clicking a short cut adCenter Add-in icon on your desktop or by selecting the application from your program list, under the Windows Start Menu.
5
How it Works
9. How come I don’t always get results when I use the adCenter Add-in?
In order to ensure that the data is anonymous when aggregated, we limit results to those that meet a minimum bar activity. For example, we typically limit the search counts to queries that are above a minimum search count for data that is based upon results from Live Search. In the case of the Campaign Association algorithm (based on data from adCenter), if only a small pool of advertisers have bid on a keyword, then the resulting data will be suppressed.
10. Under the Keyword Suggestion feature, what is the difference between the three algorithms shown (Campaign Association, Contained and Similarity)?
Campaign Association: • keywords that have been bid upon in adCenter.
MicrosoftContained: • keywords that have been part of a search query in Live Search and contain the given term.
Similarity: • keywords that are similar in context to other keywords used on the World Wide Web.
11. What are the date ranges available for the data provided by the adCenter Add-in?
Several different features in the adCenter Add-in offer different options for historical data ranges:
Top Spiky Keywords: • past 1, 10, 30 or 60 day period or a custom range.
Top Frequent Keywords, Monetization and Vertical KPIs: • past day, week or 30 day period.
Monthly Traffic: • past data for 1, 3, 6, 12 or 24 month period forecasted for 1, 3, 6 or 12 months.
Daily Traffic, Geographic and Demographic: • past data for 1, 7, 10, 30 or 60 day period — or you may specify a time period.
12. When using the Monthly Traffic feature, how does the adCenter Add-in forecast the traffic for the future months?
Monthly Traffic can be forecasted forward 1, 3, 6, or 12 months into the future. The forecasted values are determined by an algorithm, which considers past keyword trends and seasonality.
13. When using the Daily Traffic feature, why do the results not show search impressions more recent than 3 to 5 days in the past?
To ensure data quality, the adCenter Add-in provides a sample set of data that typically commences 3 to 5 days prior to the current date.
14. When using the Monetization feature, how are the output ranges defined for the conditional formatting of CTR and CPC?
The result ranges specified in the tool can be found by selecting the Conditional Formatting menu in Excel, found under the Home menu.
15. When I set Maximum Suggested Keywords to 100 in Keyword Suggestion, why do I sometimes receive less than this amount?
Maximum Suggested Keywords provides an estimated set of keywords based upon all that are available. Sometimes there are fewer keywords available than the maximum number specified. You can modify the maximum number of results returned in your options.
16. Why do I get duplicate terms in my results for Spiky Keywords?
In our system, we detect and give starting and ending dates for each spiky term. If there are duplicate terms in the results, it means the same term has become a spiky term multiple times in the given time frame.
17. How do I set the Geography function to show state, city or country?
Click the Options or the Advanced Algorithm buttons and select Geography. From here, you can set your geographical level.
18. How do I set custom defaults for date ranges or other preferences for the various functions?
Click the Options or the Advanced Algorithm buttons and select the function you wish to customize.
6
19. If I bid the CPCs recommended for Position 1, am I guaranteed the number one spot?
The adCenter Add-in provides recommendations for bids based upon position; however, these recommendations are not conclusive. Variable factors impact placement such as the number of advertisers bidding, current bid prices and budget, term frequency, geographic and demographic targeting.
20. Please explain why the query number in Monthly Traffic for a month does not match the impression data in Monetization?
Monthly Traffic (and Daily Traffic) numbers are based on exact search query data. Monetization numbers are based on ad impressions which are further segmented by the match type chosen. The numbers will not match as a result of this.
21. Why does a lower position have a higher average CPC than a higher position?
CPC position can vary based on various factors, such as maximum CPC listed by the advertiser. For example, positions 1 through 3 could have a lower CPC than position 4 for one of your keywords. This happens when no ads were served for positions 1 through 3 for that keyword.
22. Could my keyword impression count be different than Microsoft adCenter Add-in’s report
Microsoft adCenter Add-in impressions for monthly data is based on exact search queries on Live.com. Monetization is based on keyword impressions by all advertisers. Microsoft adCenter keyword reports are based solely on keywords served. Microsoft adCenter data has also been filtered to remove invalid impressions.
23. Is content match included in impression counts?
Currently, content impressions are not included in Microsoft adCenter Add-in data.
24. Why is my search result displayed in a new sheet?
If you select more than 100 cells or if your selection is inside a data table, adCenter Add-in always creates a new sheet to keep the Excel sheet clean. You can also choose to always create a new sheet by clicking on Options and selecting Display Settings.
25. How do I compare the scores from two different functions?
Comparing scores from different functions is not meaningful since different functions have different algorithms that don’t use the same scales for their scores.
26. Is adCenter Add-in information only derived from MSN data?
The Traffic and Monetization features are derived from MSN data exclusively. With the exclusion of the traffic and monetization features our functions utilize data collected from crawling the Worldwide Web.
27. What is the taxonomy used for Keyword Categorization?
The Microsoft Contextual Taxonomy is used for this function.
28. For results derived from the Monetization feature, why is it that sometimes the impression count is vastly different for Positions 1, 2, and 3 for terms with high-bid density?
There are different match types for the same keyword and it is possible to have different impressions for keywords served.
For example, the query “red shoes” could result in the following keywords being served:
Position 1: red shoes
Position 2: shoes
Position 3: red shoes
A second search for “red shoes” could have results like this:
Position 1: shoes
Position 2: shoes
Position 3: red shoes
7
Monetization would show the following information:
keyword
position
impressions
red shoes
1
1
red shoes
2
0
red shoes
3
2
shoes
1
1
shoes
2
2
shoes
3
0
29. How are the impression numbers different between adCenter Add-in forecast data and the research tab data in adCenter?
The monthly forecast is based on an algorithm incorporating significant historical exact match search query data (not impression data). The Microsoft adCenter research tab provides projected numbers based on recent impression counts which could be different from the query counts based on match type. The projected data in adCenter and the forecasted traffic numbers in the Add-in will not match as a result of this.
30. How accurate is monthly traffic forecasting and how is it done?
The forecasting is based on several algorithms, which includes historical data. Forecasting and the other algorithms are checked regularly and updates are applied in order to try to make the best predictions possible. That said, historical trends may change over time and this would cause future results to deviate from past performance.
31. Are the keyword expansion tools (Keyword Wizard / Extraction / Suggestion) based on actual search queries or keywords in adCenter?
Keyword suggestions are based on Web data and are not limited to search queries.
32. Can keywords be classified into more than one category (e.g. “apple” can be both a fruit and a technology brand)?
Keywords can be in multiple categories. By default, the application provides multiple categories, sorted by confidence.
Using the adCenter Add-in
33. How can I use the adCenter Add-in to quickly generate relevant keywords to describe my web sites?
Relevant keywords can be quickly generated either by using the Keyword Wizard or by using the Keyword Extraction feature. The Keyword Extraction feature extracts commonly used keywords from the selected website addresses, which allows the user to see what keywords are in popular use on those sites.
34. How can I use the adCenter Add-in to quickly create or expand a list of relevant keywords?
Relevant keywords can be quickly generated either by using the Keyword Wizard or by using the Keyword Suggestion feature. The user simply enters a list of keywords into Excel and the Keyword Suggestion feature will generate relevant keywords based on one or more of the following three algorithms:
Campaign Association: • keywords are suggested based on advertisers’ keyword bidding behavior.
Contained:• keywords are suggested based on past user searches that contained those keywords.
Similarity: • keywords that are similar in context to other keywords used in the World Wide Web.
35. How can I use the adCenter Add-in to forecast keyword search traffic?
The keyword search traffic can be forecasted 1, 3, 6 or 12 months into the future, either through the Keyword Wizard or by selecting the Monthly Traffic feature. Both the historical queries (searches) for the past 12 months and the forecasted quotes for the upcoming 1, 3, 6 or 12 months are shown for each keyword.
36. How can I use the adCenter Add-in to understand the seasonality of keyword searches?
Viewing historical trends of keyword searches provides insight into the seasonality of keyword searches, which can be done by using the Monthly Traffic feature.
8
37. How can I use the adCenter Add-in to understand the localization of keyword searches?
Through the Geographic feature, you can view the number of searches for selected keywords by city, state or country, for up to 10 locations over various time periods. These settings can be temporarily or permanently changed through the Options feature or through the Advanced Algorithm feature.
38. How can I use the adCenter Add-in to understand the demographics of keyword searches?
You can view the breakdown of searches in percent by gender and age group through the use of the Demographics feature. It is possible to obtain this data for a range of time periods, which can be selected through the Advanced Algorithm or Options feature.
39. How can I use the adCenter Add-in to obtain KPIs for keywords, such as CTR, CPC, Impressions, and Position?
Keyword KPIs can be obtained through the Monetization feature.
Support
40. Where else can I learn more about the most recent updates to the adCenter Add-in?
You can visit the adCenter forum to learn more about the most recent updates to the adCenter Add-in.
41. Who do I contact for adCenter Add-in support?
If you have questions about Microsoft adCenter or the adCenter Add-in, please contact our support team by accessing the adCenter Add-in forum.
42. Where can I find adCenter Add-in support information?
You can find adCenter Add-in support information on the adCenter website.
43. How do I provide feedback and how can I suggest new features for the adCenter Add-in? (Excel 2007)
Premium: We value and appreciate your feedback. It is a critical component to the success of the adCenter Add-in. Please contact your account representative with any comments; however, feel free to provide feedback at any time by accessing the adCenter Add-in forum.
Troubleshooting
44. Why can’t I find my adCenter Add-in tab in the Excel ribbon? (Excel 2007)
First, please validate that you’ve successfully installed the Microsoft adCenter Add-in by looking in your Start menu, under All Programs, for the Microsoft adCenter Add-in for Excel 2.0. If you successfully find the Add-in installed, return to Excel, and open Excel Options as shown in Figure 1 below.
Figure1
9
Select Add-ins and then select COM Add-ins in the Manage box as shown in Figure 2.
After you’ve selected COM Add-ins, click Go. A dialog box called Com Add-ins will be displayed. Check adCenter Add-in and click the OK button.
**Workaround: If the above steps do not work, please uninstall the adCenter Add-in and reinstall.
45. I’ve successfully installed the adCenter Add-in, but when I run it I get an error message regarding “invalid user authorization.” How do I resolve this? (Excel 2007)
Please perform the following steps:
1. Click the Ad Intelligence tab to display the adCenter Add-in ribbon.
2. Click the Options button.
3. Under KSP Settings, click KSP Server.
4. Next, enter your username and password.
5. In the KSP Server dialog box, specify the Cloud Server address
https://kspv2.microsoft.com/ServiceContainer and click Test to verify your connection.
6. Click OK.
Figure2
10
46. How do you Optimize your campaign with the adCenter Add-in Beta for Excel?
The adCenter Add-inBeta is a keyword research and optimization tool, based in Excel and includes the Ad Intelligence model, which enables you to:
Easily and quickly build out or expand keyword lists•
Effectively plan keyword strategy based relevance, cost history, volume, demographics, geography, and more. •
Forecast monthly and daily keyword impressions and future trends.
•
Download the adCenter Add-inBeta today to start researching keywords to help reach the right audience—and capture and convert the right customer.
Microsoft adCenter Add-In Beta for Excel 2003 is here!
We now have the Add-in for Excel 2003 users. Download the latest version here and give it a try!
II. Glossary of Key Terms
API
Application programming interface: the Microsoft adCenter API enables you to programmatically access adCenter to create and manage advertising campaigns that are run on the Internet.
Beta
The adCenter Add-in will remain in Beta as we work to provide additional features and functionality. The product is fully functional and we’re excited to offer keyword optimization, search impression forecasting, monetization and other powerful features.
Category
The category provides the industry that the keyword or URL is affiliated with, based on our taxonomy.
Cloud Server
A technology that automatically distributes user queries to a farm of servers.
CPC
Cost-per-click: the advertiser only pays when the consumer clicks on their ad.
Crawl
Terminology used to describe how a search engine moves through the Internet by following links, in order to download web page content.
Confidence
This represents the confidence we have in a result generated by the tool, and ranges from 0 (not confident) to 1 (very confident).
CTR
Click Through Rate: the percentage of those clicking on a link out of the total number of those who see the link, which reflects the efficiency of the ad investment.
Impressions
Impressions in the adCenter Add-in are the number of instances that an ad was served on Live Search.
KPI
Key Performance Indicators: they are pre-defined, quantifiable measurements that reflect the critical success factors of a business or organization.
11
KSP
Keyword Services Platform: the platform from adCenter that allows third-party developers to build applications through standardized APIs to analyze keywords and websites and optimize campaigns.
Match type: Exact, Phrase, Broad
When using the Keyword Wizard to determine the Monetization of keywords, it is possible to select various match types:
Exact: the output will only return information for those keywords that you input, in the exact order, with no additional keywords and with no spelling variations.
Phrase: the output will reflect keywords resulting from when users search on the exact phrase and also when their search contains additional terms, as long as the keyword phrase is in exactly the same order.
Broad: matches are often less targeted than exact or phrase matches. The output returned will include any keywords that contain the same keywords you have input, in any order, and possibly along with other terms.
Position
This represents the position of the displayed ads on the page of the Live Search page results. There are eight possible positions on any Live Search page — three at the top of the page and five along the right hand side.
Search Query
A search performed on Live.com.
Spiky Index
The Spiky index is a relative measurement of how popular a keyword has become. The index has a relatively wide scale, which ensures that keywords with very large search impressions do not drown out keywords with a small amount of search impressions. The higher this index, the more the keyword has surged.
Spiky Keywords
Spiky keywords are keywords that have shown a recent surge in popularity, which may be short-lived. For example, Halloween is a typical spiky keyword, as its popularity surges around October each year.
SpikyStartDate
This represents the date when the keyword began to surge in popularity.
SpikyEndDate
This represents the date when the keyword’s surge in popularity ended.
Posted bySumedh at 2:26 AM 0 comments
Google's Search Engine Optimization Starter Guide
Sunday, December 28, 2008
Welcome to Google's Search Engine Optimization Starter Guide. This document first began as an
effort to help teams within Google, but we thought it'd be just as useful to webmasters that are new to
the topic of search engine optimization and wish to improve their sites' interaction with both users and
search engines. Although this guide won't tell you any secrets that'll automatically rank your site first
for queries in Google (sorry!), following the best practices outlined below will make it easier for search
engines to both crawl and index your content.
Search engine optimization is often about making small modifications to parts of your website. When
viewed individually, these changes might seem like incremental improvements, but when combined
with other optimizations, they could have a noticeable impact on your site's user experience and
performance in organic search results. You're likely already familiar with many of the topics in this
guide, because they're essential ingredients for any webpage, but you may not be making the most
out of them.
Search engine optimization affects only organic search results, not paid or "sponsored" results,
such as Google AdWords
Google's Search Engine Optimization Starter Guide, Version 1.1, published 13 November 2008
Even though this guide's title contains the words "search engine", we'd like to say that you should
base your optimization decisions first and foremost on what's best for the visitors of your site. They're
the main consumers of your content and are using search engines to find your work. Focusing too
hard on specific tweaks to gain ranking in the organic results of search engines may not deliver the
desired results. Search engine optimization is about putting your site's best foot forward when it
comes to visibility in search engines.
An example may help our explanations, so we've created a fictitious website to follow throughout the
guide. For each topic, we've fleshed out enough information about the site to illustrate the point being
covered. Here's some background information about the site we'll use:
• Website/business name: "Brandon's Baseball Cards"
• Domain name: brandonsbaseballcards.com
• Focus: Online-only baseball card sales, price guides, articles, and news content
• Size: Small, ~250 pages
Your site may be smaller or larger than this and offer vastly different content, but the optimization
topics we discussed below should apply to sites of all sizes and types.
We hope our guide gives you some fresh ideas on how to improve your website, and we'd love to
hear your questions, feedback, and success stories in the Google Webmaster Help Group.
Create unique, accurate page titles
A title tag tells both users and search engines what the topic of a particular page is. The
should be placed within the tag of the HTML document. Ideally, you should create a unique
title for each page on your site.
The title of the homepage for our baseball card site, which lists the business name and three
main focus areas
Google's Search Engine Optimization Starter Guide, Version 1.1, published 13 November 2008
If your document appears in a search results page, the contents of the title tag will usually appear in
the first line of the results (If you're unfamiliar with the different parts of a Google search result, you
might want to check out the anatomy of a search result video by Google engineer Matt Cutts, and this
helpful diagram of a Google search results page.) Words in the title are bolded if they appear in the
user's search query. This can help users recognize if the page is likely to be relevant to their search.
The title for your homepage can list the name of your website/business and could include other bits of
important information like the physical location of the business or maybe a few of its main focuses or
offerings.
A user performs the query [baseball cards]
Our homepage shows up as a result, with the title listed on the first line (notice that the query
terms the user searched for appear in bold)
If the user clicks the result and visits the page, the page's title will appear at the top of the
browser
Titles for deeper pages on your site should accurately describe the focus of that particular page and
also might include your site or business name.
A user performs the query [rarest baseball cards]
Google's Search Engine Optimization Starter Guide, Version 1.1, published 13 November 2008
A relevant, deeper page (its title is unique to the content of the page) on our site appears as a
result
Good practices for page title tags
• Accurately describe the page's content - Choose a title that effectively communicates the
topic of the page's content.
Avoid:
• choosing a title that has no relation to the content on the page
• using default or vague titles like "Untitled" or "New Page 1"
• Create unique title tags for each page - Each of your pages should ideally have a unique
title tag, which helps Google know how the page is distinct from the others on your site.
Avoid:
• using a single title tag across all of your site's pages or a large group of pages
• Use brief, but descriptive titles - Titles can be both short and informative. If the title is too
long, Google will show only a portion of it in the search result.
Avoid:
• using extremely lengthy titles that are unhelpful to users
• stuffing unneeded keywords in your title tags
Make use of the "description" meta tag
A page's description meta tag gives Google and other search engines a summary of what the page is
about. Whereas a page's title may be a few words or a phrase, a page's description meta tag might
be a sentence or two or a short paragraph. Google Webmaster Tools provides a handy content
analysis section that'll tell you about any description meta tags that are either too short, long, or
duplicated too many times (the same information is also shown for
the description meta tag is placed within the tag of your HTML document.
Google's Search Engine Optimization Starter Guide, Version 1.1, published 13 November 2008
The beginning of the description meta tag for our homepage, which gives a brief overview of
the site's offerings
Description meta tags are important because Google might use them as snippets for your pages.
Note that we say "might" because Google may choose to use a relevant section of your page's visible
text if it does a good job of matching up with a user's query. Alternatively, Google might use your site's
description in the Open Directory Project if your site is listed there (learn how to prevent search
engines from displaying ODP data). Adding description meta tags to each of your pages is always a
good practice in case Google cannot find a good selection of text to use in the snippet. The
Webmaster Central Blog has an informative post on improving snippets with better description meta
tags.
Snippets appear under a page's title and above a page's URL in a search result.
A user performs the query [baseball cards]
Our homepage appears as a result, with part of its description meta tag used as the snippet
Words in the snippet are bolded when they appear in the user's query. This gives the user clues about
whether the content on the page matches with what he or she is looking for. Below is another
example, this time showing a snippet from a description meta tag on a deeper page (which ideally has
its own unique description meta tag) containing an article.
Google's Search Engine Optimization Starter Guide, Version 1.1, published 13 November 2008
A user performs the query [rarest baseball cards]
One of our deeper pages, with its unique description meta tag used as the snippet, appears as
a result
Good practices for description meta tags
• Accurately summarize the page's content - Write a description that would both inform and
interest users if they saw your description meta tag as a snippet in a search result.
Avoid:
• writing a description meta tag that has no relation to the content on the page
• using generic descriptions like "This is a webpage" or "Page about baseball
cards"
• filling the description with only keywords
• copy and pasting the entire content of the document into the description meta
tag
• Use unique descriptions for each page - Having a different description meta tag for each
page helps both users and Google, especially in searches where users may bring up
multiple pages on your domain (e.g. searches using the site: operator). If your site has
thousands or even millions of pages, hand-crafting description meta tags probably isn't
feasible. In this case, you could automatically generate description meta tags based on
each page's content.
Avoid:
• using a single description meta tag across all of your site's pages or a large
group of pages
Improve the structure of your URLs
Creating descriptive categories and filenames for the documents on your website can not only help
you keep your site better organized, but it could also lead to better crawling of your documents by
Google's Search Engine Optimization Starter Guide, Version 1.1, published 13 November 2008
search engines. Also, it can create easier, "friendlier" URLs for those that want to link to your content.
Visitors may be intimidated by extremely long and cryptic URLs that contain few recognizable words.
A URL to a page on our baseball card site that a user might have a hard time with
URLs like these can be confusing and unfriendly. Users would have a hard time reciting the URL from
memory or creating a link to it. Also, users may believe that a portion of the URL is unnecessary,
especially if the URL shows many unrecognizable parameters. They might leave off a part, breaking
the link.
Some users might link to your page using the URL of that page as the anchor text. If your URL
contains relevant words, this provides users and search engines with more information about the
page than an ID or oddly named parameter would.
The highlighted words above could inform a user or search engine what the target page is
about before following the link
Lastly, remember that the URL to a document is displayed as part of a search result in Google, below
the document's title and snippet. Like the title and snippet, words in the URL on the search result
appear in bold if they appear in the user's query.
A user performs the query [baseball cards]
Our homepage appears as a result, with the URL listed under the title and snippet
Google's Search Engine Optimization Starter Guide, Version 1.1, published 13 November 2008
Below is another example showing a URL on our domain for a page containing an article about the
rarest baseball cards. The words in the URL might appeal to a search user more than an ID number
like "www.brandonsbaseballcards.com/article/102125/" would.
A user performs the query [rarest baseball cards]
A deeper page, with a URL that reflects the type of content found on it, appears as a result
Google is good at crawling all types of URL structures, even if they're quite complex, but spending the
time to make your URLs as simple as possible for both users and search engines can help. Some
webmasters try to achieve this by rewriting their dynamic URLs to static ones; while Google is fine
with this, we'd like to note that this is an advanced procedure and if done incorrectly, could cause
crawling issues with your site. To learn even more about good URL structure, we recommend this
Webmaster Help Center page on creating Google-friendly URLs.
Good practices for URL structure
• Use words in URLs - URLs with words that are relevant to your site's content and structure
are friendlier for visitors navigating your site. Visitors remember them better and might be
more willing to link to them.
Avoid:
• using lengthy URLs with unnecessary parameters and session IDs
• choosing generic page names like "page1.html"
• using excessive keywords like "baseball-cards-baseball-cards-baseballcards.
htm"
• Create a simple directory structure - Use a directory structure that organizes your content
well and is easy for visitors to know where they're at on your site. Try using your directory
structure to indicate the type of content found at that URL.
Avoid:
• having deep nesting of subdirectories like ".../dir1/dir2/dir3/dir4/dir5/dir6/
page.html"
• using directory names that have no relation to the content in them
Google's Search Engine Optimization Starter Guide, Version 1.1, published 13 November 2008
• Provide one version of a URL to reach a document - To prevent users from linking to one
version of a URL and others linking to a different version (this could split the reputation of
that content between the URLs), focus on using and referring to one URL in the structure
and internal linking of your pages. If you do find that people are accessing the same content
through multiple URLs, setting up a 301 redirect from non-preferred URLs to the dominant
URL is a good solution for this.
Avoid:
• having pages from subdomains and the root directory (e.g. "domain.com/
page.htm" and "sub.domain.com/page.htm") access the same content
• mixing www. and non-www. versions of URLs in your internal linking structure
• using odd capitalization of URLs (many users expect lower-case URLs and
remember them better)
Make your site easier to navigate
The navigation of a website is important in helping visitors quickly find the content they want. It can
also help search engines understand what content the webmaster thinks is important. Although
Google's search results are provided at a page level, Google also likes to have a sense of what role a
page plays in the bigger picture of the site.
All sites have a home or "root" page, which is usually the most frequented page on the site and the
starting place of navigation for many visitors. Unless your site has only a handful of pages, you should
think about how visitors will go from a general page (your root page) to a page containing more
specific content. Do you have enough pages around a specific topic area that it would make sense to
create a page describing these related pages (e.g. root page -> related topic listing -> specific topic)?
Do you have hundreds of different products that need to be classified under multiple category and
subcategory pages?
The directory structure for our small website on baseball cards
Google's Search Engine Optimization Starter Guide, Version 1.1, published 13 November 2008
A sitemap (lower-case) is a simple page on your site that displays the structure of your website, and
usually consists of a hierarchical listing of the pages on your site. Visitors may visit this page if they
are having problems finding pages on your site. While search engines will also visit this page, getting
good crawl coverage of the pages on your site, it's mainly aimed at human visitors.
An XML Sitemap (upper-case) file, which you can submit through Google's Webmaster Tools, makes
it easier for Google to discover the pages on your site. Using a Sitemap file is also one way (though
not guaranteed) to tell Google which version of a URL you'd prefer as the canonical one (e.g.
http://brandonsbaseballcards.com/ or http://www.brandonsbaseballcards.com/; more on what's a
preferred domain). Google helped create the open source Sitemap Generator script to help you
create a Sitemap file for your site. To learn more about Sitemaps, the Webmaster Help Center
provides a useful guide to Sitemap files.
Good practices for site navigation
• Create a naturally flowing hierarchy - Make it as easy as possible for users to go from
general content to the more specific content they want on your site. Add navigation pages
when it makes sense and effectively work these into your internal link structure.
Avoid:
• creating complex webs of navigation links, e.g. linking every page on your site
to every other page
• going overboard with slicing and dicing your content (it takes twenty clicks to
get to deep content)
• Use mostly text for navigation - Controlling most of the navigation from page to page on
your site through text links makes it easier for search engines to crawl and understand your
site. Many users also prefer this over other approaches, especially on some devices that
might not handle Flash or JavaScript.
Avoid:
• having a navigation based entirely on drop-down menus, images, or
animations (many, but not all, search engines can discover such links on a site,
but if a user can reach all pages on a site via normal text links, this will improve
the accessibility of your site; more on how Google deals with non-text files)
• Use "breadcrumb" navigation - A breadcrumb is a row of internal links at the top or bottom
of the page that allows visitors to quickly navigate back to a previous section or the root
page. Many breadcrumbs have the most general page (usually the root page) as the first,
left-most link and list the more specific sections out to the right.
Breadcrumb links appearing on a deeper article page on our site
Google's Search Engine Optimization Starter Guide, Version 1.1, published 13 November 2008
• Put an HTML sitemap page on your site, and use an XML Sitemap file - A simple
sitemap page with links to all of the pages or the most important pages (if you have
hundreds or thousands) on your site can be useful. Creating an XML Sitemap file for your
site helps ensure that search engines discover the pages on your site.
Avoid:
• letting your HTML sitemap page become out of date with broken links
• creating an HTML sitemap that simply lists pages without organizing them, for
example by subject
• Consider what happens when a user removes part of your URL - Some users might
navigate your site in odd ways, and you should anticipate this. For example, instead of using
the breadcrumb links on the page, a user might drop off a part of the URL in the hopes of
finding more general content. He or she might be visiting
http://www.brandonsbaseballcards.com/news/2008/upcoming-baseball-card-shows.htm, but
then enter http://www.brandonsbaseballcards.com/news/2008/ into the browser's address
bar, believing that this will show all news from 2008. Is your site prepared to show content in
this situation or will it give the user a 404 ("page not found" error)? What about moving up a
directory level to http://www.brandonsbaseballcards.com/news/?
• Have a useful 404 page - Users will occasionally come to a page that doesn't exist on your
site, either by following a broken link or typing in the wrong URL. Having a custom 404 page
that kindly guides users back to a working page on your site can greatly improve a user's
experience. Your 404 page should probably have a link back to your root page and could
also provide links to popular or related content on your site. Google provides a 404 widget
that you can embed in your 404 page to automatically populate it with many useful features.
You can also use Google Webmaster Tools to find the sources of URLs causing "not found"
errors.
Avoid:
• allowing your 404 pages to be indexed in search engines (make sure that your
webserver is configured to give a 404 HTTP status code when non-existent
pages are requested)
• providing only a vague message like "Not found", "404", or no 404 page at all
• using a design for your 404 pages that isn't consistent with the rest of your site
Offer quality content and services
Creating compelling and useful content will likely influence your website more than any of the other
factors discussed here. Users know good content when they see it and will likely want to direct other
users to it. This could be through blog posts, social media services, email, forums, or other means.
Organic or word-of-mouth buzz is what helps build your site's reputation with both users and Google,
and it rarely comes without quality content.
Google's Search Engine Optimization Starter Guide, Version 1.1, published 13 November 2008
A blogger finds a piece of your content, likes it, and then references it in a blog post
While the content you create could be on any topic imaginable, here are some recommended best
practices:
Good practices for content
• Write easy-to-read text - Users enjoy content that is well written and easy to follow.
Avoid:
• writing sloppy text with many spelling and grammatical mistakes
• embedding text in images for textual content (users may want to copy and
paste the text and search engines can't read it)
• Stay organized around the topic - It's always beneficial to organize your content so that
visitors have a good sense of where one content topic begins and another ends. Breaking
your content up into logical chunks or divisions helps users find the content they want faster.
Avoid:
• dumping large amounts of text on varying topics onto a page without
paragraph, subheading, or layout separation
• Use relevant language - Think about the words that a user might search for to find a piece
of your content. Users who know a lot about the topic might use different keywords in their
search queries than someone who is new to the topic. For example, a long-time baseball fan
might search for [nlcs], an acronym for the National League Championship Series, while a
new fan might use a more general query like [baseball playoffs]. Anticipating these
differences in search behavior and accounting for them while writing your content (using a
good mix of keyword phrases) could produce positive results. Google AdWords provides a
handy Keyword Tool that helps you discover new keyword variations and see the
approximate search volume for each keyword. Also, Google Webmaster Tools provides you
with the top search queries your site appears for and the ones that led the most users to
your site.
• Create fresh, unique content - New content will not only keep your existing visitor base
coming back, but also bring in new visitors.
Avoid:
• rehashing (or even copying) existing content that will bring little extra value to
users
Google's Search Engine Optimization Starter Guide, Version 1.1, published 13 November 2008
• having duplicate or near-duplicate versions of your content across your site
(more on duplicate content)
• Offer exclusive content or services - Consider creating a new, useful service that no other
site offers. You could also write an original piece of research, break an exciting news story,
or leverage your unique user base. Other sites may lack the resources or expertise to do
these things.
• Create content primarily for your users, not search engines - Designing your site around
your visitors' needs while making sure your site is easily accessible to search engines
usually produces positive results.
Avoid:
• inserting numerous unnecessary keywords aimed at search engines but are
annoying or nonsensical to users
• having blocks of text like "frequent misspellings used to reach this page" that
add little value for users
• deceptively hiding text from users, but displaying it to search engines
Write better anchor text
Anchor text is the clickable text that users will see as a result of a link, and is placed within the anchor
tag .
This anchor text accurately describes the content on one of our article pages
This text tells users and Google something about the page you're linking to. Links on your page may
be internal—pointing to other pages on your site—or external—leading to content on other sites. In
either of these cases, the better your anchor text is, the easier it is for users to navigate and for
Google to understand what the page you're linking to is about.
Good practices for anchor text
• Choose descriptive text - The anchor text you use for a link should provide at least a basic
idea of what the page linked to is about.
Avoid:
• writing generic anchor text like "page", "article", or "click here"
• using text that is off-topic or has no relation to the content of the page linked to
• using the page's URL as the anchor text in most cases (although there are
certainly legitimate uses of this, such as promoting or referencing a new
website's address)
Google's Search Engine Optimization Starter Guide, Version 1.1, published 13 November 2008
• Write concise text - Aim for short but descriptive text—usually a few words or a short
phrase.
Avoid:
• writing long anchor text, such as a lengthy sentence or short paragraph of text
• Format links so they're easy to spot - Make it easy for users to distinguish between
regular text and the anchor text of your links. Your content becomes less useful if users
miss the links or accidentally click them.
Avoid:
• using CSS or text styling that make links look just like regular text
• Think about anchor text for internal links too - You may usually think about linking in
terms of pointing to outside websites, but paying more attention to the anchor text used for
internal links can help users and Google navigate your site better.
Avoid:
• using excessively keyword-filled or lengthy anchor text just for search engines
• creating unnecessary links that don't help with the user's navigation of the site
Use heading tags appropriately
Heading tags (not to be confused with the HTML tag or HTTP headers) are used to present
structure on the page to users. There are six sizes of heading tags, beginning with
, the most
important, and ending with , the least important.
On a page containing a news story, we might put the name of our site into an tag and the
topic of the story into an tag
Since heading tags typically make text contained in them larger than normal text on the page, this is a
visual cue to users that this text is important and could help them understand something about the
type of content underneath the heading text. Multiple heading sizes used in order create a hierarchical
structure for your content, making it easier for users to navigate through your document.
Google's Search Engine Optimization Starter Guide, Version 1.1, published 13 November 2008
Good practices for heading tags
• Imagine you're writing an outline - Similar to writing an outline for a large paper, put some
thought into what the main points and sub-points of the content on the page will be and
decide where to use heading tags appropriately.
Avoid:
• placing text in heading tags that wouldn't be helpful in defining the structure of
the page
• using heading tags where other tags like and may be more
appropriate
• erratically moving from one heading tag size to another
• Use headings sparingly across the page - Use heading tags where it makes sense. Too
many heading tags on a page can make it hard for users to scan the content and determine
where one topic ends and another begins.
Avoid:
• excessively using heading tags throughout the page
• putting all of the page's text into a heading tag
• using heading tags only for styling text and not presenting structure
Optimize your use of images
Images may seem like a straightforward component of your site, but you can optimize your use of
them. All images can have a distinct filename and "alt" attribute, both of which you should take
advantage of.
The "alt" attribute allows you to specify alternative text for the image if it cannot be displayed for some
reason.
Our alt text here is a brief but accurate description of the image
Why use this attribute? If a user is viewing your site on a browser that doesn't support images, or is
using alternative technologies, such as a screen reader, the contents of the alt attribute provide
information about the picture.
Google's Search Engine Optimization Starter Guide, Version 1.1, published 13 November 2008
Our image wasn't displayed to the user for some reason, but at least the alt text was
Another reason is that if you're using an image as a link, the alt text for that image will be treated
similarly to the anchor text of a text link. However, we don't recommend using too many images for
links in your site's navigation when text links could serve the same purpose. Lastly, optimizing your
image filenames and alt text makes it easier for image search projects like Google Image Search to
better understand your images.
Good practices for images
• Use brief, but descriptive filenames and alt text - Like many of the other parts of the page
targeted for optimization, filenames and alt text (for ASCII languages) are best when they're
short, but descriptive.
Avoid:
• using generic filenames like "image1.jpg", "pic.gif", "1.jpg" when possible
(some sites with thousands of images might consider automating the naming of
images)
• writing extremely lengthy filenames
• stuffing keywords into alt text or copying and pasting entire sentences
• Supply alt text when using images as links - If you do decide to use an image as a link,
filling out its alt text helps Google understand more about the page you're linking to. Imagine
that you're writing anchor text for a text link.
Avoid:
• writing excessively long alt text that would be considered spammy
• using only image links for your site's navigation
• Store images in a directory of their own - Instead of having image files spread out in
numerous directories and subdirectories across your domain, consider consolidating your
images into a single directory (e.g. brandonsbaseballcards.com/images/). This simplifies the
path to your images.
Google's Search Engine Optimization Starter Guide, Version 1.1, published 13 November 2008
• Use commonly supported filetypes - Most browsers support JPEG, GIF, PNG, and BMP
image formats. It's also a good idea to have the extension of your filename match with the
filetype.
Make effective use of robots.txt
A "robots.txt" file tells search engines whether they can access and therefore crawl parts of your
site. This file, which must be named "robots.txt", is placed in the root directory of your site.
The address of our robots.txt file
All compliant search engine bots (denoted by the wildcard * symbol) shouldn't access and
crawl the content under /images/ or any URL whose path begins with /search
You may not want certain pages of your site crawled because they might not be useful to users if
found in a search engine's search results. If you do want to prevent search engines from crawling
your pages, Google Webmaster Tools has a friendly robots.txt generator to help you create this file.
Note that if your site uses subdomains and you wish to have certain pages not crawled on a particular
subdomain, you'll have to create a separate robots.txt file for that subdomain. For more information
on robots.txt, we suggest this Webmaster Help Center guide on using robots.txt files.
There are a handful of other ways to prevent content appearing in search results, such as adding
"NOINDEX" to your robots meta tag, using .htaccess to password protect directories, and using
Google Webmaster Tools to remove content that has already been crawled. Google engineer Matt
Cutts walks through the caveats of each URL blocking method in a helpful video.
Good practices for robots.txt
• Use more secure methods for sensitive content - You shouldn't feel comfortable using
robots.txt to block sensitive or confidential material. One reason is that search engines could
still reference the URLs you block (showing just the URL, no title or snippet) if there happen
to be links to those URLs somewhere on the Internet (like referrer logs). Also, non-compliant
or rogue search engines that don't acknowledge the Robots Exclusion Standard could
disobey the instructions of your robots.txt. Finally, a curious user could examine the
directories or subdirectories in your robots.txt file and guess the URL of the content that you
Google's Search Engine Optimization Starter Guide, Version 1.1, published 13 November 2008
don't want seen. Encrypting the content or password-protecting it with .htaccess are more
secure alternatives.
Avoid:
• allowing search result-like pages to be crawled (users dislike leaving one
search result page and landing on another search result page that doesn't add
significant value for them)
• allowing a large number of auto-generated pages with the same or only slightly
different content to be crawled: "Should these 100,000 near-duplicate pages
really be in a search engine's index?"
• allowing URLs created as a result of proxy services to be crawled
Be aware of rel="nofollow" for links
Setting the value of the "rel" attribute of a link to "nofollow" will tell Google that certain links on your
site shouldn't be followed or pass your page's reputation to the pages linked to. Nofollowing a link is
adding rel="nofollow" inside of the link's anchor tag.
If you link to a site that you don't trust and don't want to pass your site's reputation to, use
nofollow
When would this be useful? If your site has a blog with public commenting turned on, links within
those comments could pass your reputation to pages that you may not be comfortable vouching for.
Blog comment areas on pages are highly susceptible to comment spam. Nofollowing these useradded
links ensures that you're not giving your page's hard-earned reputation to a spammy site. Many
blogging software packages automatically nofollow user comments, but those that don't can most
likely be manually edited to do this. This advice also goes for other areas of your site that may involve
user-generated content, such as guestbooks, forums, shout-boards, referrer listings, etc. If you're
willing to vouch for links added by third parties (e.g. if a commenter is trusted on your site), then
there's no need to use nofollow on links; however, linking to sites that Google considers spammy can
affect the reputation of your own site. The Webmaster Help Center has more tips on avoiding
comment spam, like using CAPTCHAs and turning on comment moderation.
Google's Search Engine Optimization Starter Guide, Version 1.1, published 13 November 2008
A comment spammer leaves a message on one of our news posts, hoping to get some of our
site's reputation
Another use of nofollow is when you're writing content and wish to reference a website, but don't want
to pass your reputation on to it. For example, imagine that you're writing a blog post on the topic of
comment spamming and you want to call out a site that recently comment spammed your blog. You
want to warn others of the site, so you include the link to it in your content; however, you certainly
don't want to give the site some of your reputation from your link. This would be a good time to use
nofollow.
Lastly, if you're interested in nofollowing all of the links on a page, you can use "nofollow" in your
robots meta tag, which is placed inside the tag of that page's HTML. The Webmaster Central
Blog provides a helpful post on using the robots meta tag. This method is written as name="robots" content="nofollow">.
This nofollows all of the links on a page
Google's Search Engine Optimization Starter Guide, Version 1.1, published 13 November 2008
Promote your website in the right ways
While most of the links to your site will be gained gradually, as people discover your content through
search or other ways and link to it, Google understands that you'd like to let others know about the
hard work you've put into your content. Effectively promoting your new content will lead to faster
discovery by those who are interested in the same subject. As with most points covered in this
document, taking these recommendations to an extreme could actually harm the reputation of your
site.
Good practices for promoting your website
• Blog about new content or services - A blog post on your own site letting your visitor base
know that you added something new is a great way to get the word out about new content or
services. Other webmasters who follow your site or RSS feed could pick the story up as well.
• Don't forget about offline promotion - Putting effort into the offline promotion of your
company or site can also be rewarding. For example, if you have a business site, make sure
its URL is listed on your business cards, letterhead, posters, etc. You could also send out
recurring newsletters to clients through the mail letting them know about new content on the
company's website.
• Know about social media sites - Sites built around user interaction and sharing have made
it easier to match interested groups of people up with relevant content.
Avoid:
• attempting to promote each new, small piece of content you create; go for big,
interesting items
• involving your site in schemes where your content is artificially promoted to the
top of these services
• Add your business to Google's Local Business Center - If you run a local business,
adding its information to Google's Local Business Center will help you reach customers on
Google Maps and web search. The Webmaster Help Center has more tips on promoting
your local business.
• Reach out to those in your site's related community - Chances are, there are a number
of sites that cover topic areas similar to yours. Opening up communication with these sites is
usually beneficial. Hot topics in your niche or community could spark additional ideas for
content or building a good community resource.
Avoid:
• spamming link requests out to all sites related to your topic area
• purchasing links from another site with the aim of getting PageRank instead of
traffic
Google's Search Engine Optimization Starter Guide, Version 1.1, published 13 November 2008
Make use of free webmaster tools
Major search engines, including Google, provide free tools for webmasters. Google's Webmaster
Tools help webmasters better control how Google interacts with their websites and get useful
information from Google about their site. Using Webmaster Tools won't help your site get preferential
treatment; however, it can help you identify issues that, if addressed, can help your site perform better
in search results. With the service, webmasters can:
• see which parts of a site Googlebot had problems crawling
• upload an XML Sitemap file
• analyze and generate robots.txt files
• remove URLs already crawled by Googlebot
• specify the preferred domain
• identify issues with title and description meta tags
• understand the top searches used to reach a site
• get a glimpse at how Googlebot sees pages
• remove unwanted sitelinks that Google may use in results
• receive notification of quality guideline violations and file for a site reconsideration
Yahoo! (Yahoo! Site Explorer) and Microsoft (Live Search Webmaster Tools) also offer free tools for
webmasters.
Take advantage of web analytics services
If you've improved the crawling and indexing of your site using Google Webmasters Tools or other
services, you're probably curious about the traffic coming to your site. Web analytics programs like
Google Analytics are a valuable source of insight for this. You can use these to:
• get insight into how users reach and behave on your site
• discover the most popular content on your site
• measure the impact of optimizations you make to your site (e.g. did changing those title and
description meta tags improve traffic from search engines?)
For advanced users, the information an analytics package provides, combined with data from your
server log files, can provide even more comprehensive information about how visitors are interacting
with your documents (such as additional keywords that searchers might use to find your site).
Lastly, Google offers another tool called Google Website Optimizer that allows you to run experiments
to find what on-page changes will produce the best conversion rates with visitors. This, in combination
with Google Analytics and Google Webmaster Tools (see our video on using the "Google Trifecta"), is
a powerful way to begin improving your site.
Google's Search Engine Optimization Starter Guide, Version 1.1, published 13 November 2008
Helpful resources for webmasters
Google Webmaster Help Group - Have questions or feedback on our guide? Let us know
Google Webmaster Central Blog - Frequent posts by Googlers on how to improve your website
Google Webmaster Help Center - Filled with in-depth documentation on webmaster-related issues
Google Webmaster Tools - Optimize how Google interacts with your website
Google Webmaster Guidelines - Design, content, technical, and quality guidelines from Google
Google Analytics - Find the source of your visitors, what they're viewing, and benchmark changes
Google Website Optimizer - Run experiments on your pages to see what will work and what won't
Tips on Hiring an SEO - If you don't want to go at it alone, these tips should help you choose an
SEO company
Except as otherwise noted, the content of this document is licensed under the Creative Commons
Attribution 3.0 License.
Google's Search Engine Optimization Starter Guide, Version 1.1, published 13 November 2008
Posted bySumedh
at
8:27 PM
0
comments
WWW 2
Sunday, December 14, 2008
In the last lesson I told you that "URLs" is pronounced "earls." In fact, a lot of people still use the initials and call them "U-R-Ls." I personally prefer calling them "earls" because "earls" rhymes with "squirrels," but the choice of what you call them is completely up to you :)
Let's take a look at how a sample WWW browser works. There are three basic types of WWW browsers available: line-mode browsers, full screen browsers (like Lynx), and graphical browsers (like Mosaic).
Line-mode browsers are about as user un-friendly as you can get. This is hard to describe, but line-mode browsers work a little like FTP inasmuch as you type a command, get some information on your screen, type a new command, get some more information, and so on ...
A full screen browser puts a menu on your screen that looks a little like the Gopher menus that we saw in MAP18: Gopher (Part One). You move the cursor up and down the screen, select a highlighted link, press enter or return, and you are automatically taken to a new document or file (your fill screen browser may work differently than this, though).
Graphical browsers allow you to access not only text, but also pictures and sound (a.k.a. "hypermedia"). In fact, these pictures can be put in Web documents (a.k.a. Web pages), making that Web page look less like a Gopher menu and more like a page from a color magazine! Most graphical browsers use a mouse, and you point-and-click on a highlighted link to access it.
The URL for the sample Web page that we are going to use in this lesson is
http://ua1vm.ua.edu/~vmhttpd/index.html
and I am going to be using the UF WWW Browser for CMS to access this page. I'll talk a little more about how you can access a WWW browser in a few minutes, but I first want to show you what a basic Web page looks like. The UF WWW Browser for CMS is the browser that my service provider uses, and it is a full screen browser. The browser that you use -- if you can even access a WWW browser -- will probably look and work a little differently than what you will see in this example.
Finally, in real life my browser highlights the links by making them a different color than the rest of the text. There is no way for me to use different colors in this letter, so I have highlighted the links in this example by surrounding them with a (* *).
Just like I can access an item in a Gopher menu by pointing at it and selecting it, I can access a WWW link by pointing at it and selecting it.
Enough talk. Time for the example.
I access my provider's WWW browser, and the following appears on my screen:
Albert 1.2.0 (UF WWW Browser for CMS) Screen 1 of 2 (more screens)
Viewing=http://ua1vm.ua.edu/~vmhttpd/index.html
Title=UA1VM WWW Home Page
Welcome to The University of Alabama's CMS WWW Server
This CMS server is still under development. Any (*comments*)
or (*suggestions*) will be greatly appreciated. Thank you.
Gopher Sites:
- (*UA1VM CMS Gopher Server*)
- (*UA1IX AIX/370 Gopher Server*)
- (*RISC/6000 Gopher Server*)
- (*RICEVM1 CMS Gopher Server*)
Telnet Sessions:
- (*UA1VM.UA.EDU*)
- (*UA1IX.UA.EDU - Line Mode*)
- (*RISC.UA.EDU - Line Mode*)
WWW Sites:
- The University of Alabama Libraries (*WWW*)
- The University of Illinois at Urbana-Champaign (*WWW*)
- The Alabama Supercomputer Network (*WWW*)
- NASA Information Services via (*WWW*)
Leisure:
- (*Intertext Magazine*) - Electronic Fictional Magazine at The
University of Michigan
- (*Wiretap*) - a gopher to Cupertino, California
- (*NNR*) - UA1VM's Network News Reader
Other Neat Stuff:
- The University of Alabama Library's On-Line (*Card Catalog*)
- a (*map*) of The University of Alabama campus
... snip snip snip ...
COOL! I can select any of these links -- the words set apart from the rest of the text with a (* *) -- and be transported to that particular link.
>From this one Web page, I can access Gopher, telnet, and even other Web pages! (I can also access FTP, although this page does not show it).
We've seen a lot of Gopher and telnet recently. Let's take a look at another Web page. Since I used to be a Simulations Director at the United States Space Camp (did I forget to tell you that?), I'm going to move my cursor down to the (*WWW*) link next to "NASA Information Services", press enter, and see what happens:
Albert 1.2.0 (UF WWW Browser for CMS) Screen 1 of 2 (more screens)
Viewing=http://hypatia.gsfc.nasa.gov/NASA_homepage.html
Title=NASA Information Services via World Wide Web
National Aeronautics and Space Administration
NASA Logo
(*World Wide Web (WWW) information services*)
(*Hot Topics*) NASA news and subjects of public interest
(*NASA Strategic Plan*)
(*NASA Strategies, Policies, and Public Affairs*)
(*NASA Online Educational Resources*)
(*NASA Information Sources by Subject*)
... snip snip snip ...
This is certainly more interesting than SURAnet! ;) >From this Web page I can access OTHER Web pages, and from those Web pages I can access even MORE Web pages, and so on ...
Yesterday I told you that it is possible for you to connect directly to a specific Internet resource so long as you know the resource's URL. How you do that depends on the browser that you use.
For the line-mode browser at CERN, for example, the command to connect to a particular URL is
GO
replacing with the URL of the Internet resource that you want to access. In Lynx, you just select the "GO" link on the browser's start-up page; in most graphical browsers (like Mosaic), there is usually an "Open URL" option in one of the menus. (1) Before you can do this, however, you have to first access the Web. There are three ways that you can do this:
- Through a browser stored on your local Internet service provider's machine. Ask your provider if your site has a Web browser, and how you can access it.
- Through a browser stored on your own machine. Until recently, you had to have a SLIP or PPP connection to do this. Some providers -- providers who FLOODED my mailbox when I did not talk about the special Level 2.0002746278546723 access that they offer -- now allow you to store a Web browser on your own machine even though you only have Level 2 access.
If you do not have a SLIP or PPP connection, contact your provider BEFORE you store a Web reader on your own computer and double-check that your provider will support the browser. Most service providers can not support a Web browser unless you have a SLIP or PPP connection.
- Through a telnet connection to a publicly-accessible Web browser.
If you have a SLIP or a PPP connection, the WWW FAQ that I have stored on the University of Alabama's LISTSERV file server (WWW FAQ1) has a list of FTP sites where you can get specific Web browsers. (Do me a favor ... re-read that last sentence. Did you EVER think you would understand a sentence like that? Isn't this workshop COOL?!!)
If you do not have access to a Web browser through your site, you may still be able to access a Web browser through telnet. The following are a few of the public Web sites: (1)
telnet address comments
-------------- -----------------------
info.cern.ch No password is required. This is in
Switzerland, so U.S. users might be
better off using a closer browser.
www.cc.ukans.edu The "Lynx" full screen browser, which
requires a vt100 terminal. Login as www.
Does not allow users to "go" to arbitrary
URLs.
www.njit.edu Login as www. A full-screen browser at
the New Jersey Institute of Technology.
sun.uakom.cs Slovakia. Has a slow link, so only use
from nearby.
info.funet.fi Login as www. Offers several browsers,
including Lynx (goto option disabled
there too).
fserv.kfki.hu Hungary. Has a slow link, so use from
nearby. Login as www.
Once you are on the Web, it is possible for you to do keyword searches (much like the Veronica searches we did in MAP21: Veronica) using one of the Web's many search engines. One of the best Web search engines is the WebCrawler. The WebCrawler's URL is http://www.biotech.washington.edu/WebQuery.html
and the WebCrawler searches for documents whose title AND CONTENT match your keyword. Another Web search engine you probably will want to check out is the World Wide Web Worm. The Worm's URL is
http://www.cs.colorado.edu/home/mcbryan/WWWW.html
The Worm is not at thorough as the WebCrawler, but it is still a pretty competent search engine. Both of these search engines provide really good on-line help and instructions.
One last thing, and I am though for this lesson. Please remember:
- The "Web" is the collection of all of the files and information that can be accessed by a Web browser.
- Mosaic and Lynx and just BROWSERS that allow you to access the Web.
Posted bySumedh
at
11:51 PM
1 comments
Labels:
Internet
WWW
I wish I had six weeks just to talk about the World Wide Web (a.k.a. WWW or "the Web."). If you think Gopher is neat, wait until you start playing around on the Web :)
Unfortunately, I don't have six weeks to talk about the Web -- I only have two lessons. Because of this, we are going to go through the Web like Sherman went through Georgia (1).
That's the bad news. The good news is that there are a lot of REALLY good Web guides available, and I am even seriously considering developing my own Web workshop that I will offer late next year (2).
Until that time comes, however, let's talk about the BASICS of the Web.
In the last few lessons I showed you how most Gopher menus are linked together. We started out in the gopher.squirrel.com root menu, and eventually ended up at the SURAnet gopher menu. We were able to do this because the menus that we traveled through had links to menus and files that were located at other Gopher sites.
Because Gopher menus are linked together, a whole world of information is available to us with just a few keystrokes!
Imagine if we were able to take these links one step further. Instead of linking menus, we could link documents together. You could read one document, find a keyword in that document that really interests you, touch that keyword, and automatically be taken to a NEW document somewhere else in the world -- and this new document could even have links to OTHER documents around the world, and so on.
Sound too good to be true? It isn't, thanks to something called "hypertext." If you have ever played with Apple's HyperCard program or the "help" menus in the latest Microsoft packages, you have already experienced hypertext. You "select" a highlighted word -- usually by clicking on it with a mouse -- and you are taken into an entirely new document or help screen.
The World Wide Web is based on hypertext. It is possible for you to go roaming around the Web, bouncing from document to document, using nothing but the links in those documents!
Just as you can access Gopherspace through a Gopher server or client, you can access the Web through something called a "browser." A browser can read documents, fetch documents, access files by FTP, read Usenet newsgroups, telnet into remote sites, and even travel around Gopherspace. In short, everything that we have talked about from MAP08: Usenet to MAP22:Gophermail can be done using nothing but a Web browser!
The Web is able to accomplish all of this thanks to something called URLs ("earls") -- Universal Resource Locators. URLs list the exact location of any Internet resource.
If you think about it, giving every Internet resource a unique address is the hard part. Once you have given something an address, linking to it is pretty easy :)
What is really special about the Web is that the Web does all of this "behind the scenes." It is possible for you to bounce from one link to another without ever knowing the exact address of where you are, or even how you got there.
If you ever want to jump directly to a particular Internet resource, however, you are going to need to know a little bit more about URLs. Here are a few basic URLs:
file://wuarchive.wustl.edu/mirrors/msdos/graphics/gifkit.zip
ftp://wuarchive.wustl.edu/mirrors
http://info.cern.ch:80/default.html
news:alt.hypertext
telnet://dra.com
Gee ... those look a little like FTP addresses, don't they? The first part of an URL -- the stuff before the colon -- tells the browser how to access that particular file. For example, to access
ftp://wuarchive.wustl.edu/mirrors
your browser would use FTP. Most of the access methods are pretty straight-forward. Here is a list of some of the more common access methods that you are going to see listed in the first part of URLs:
method what it stands for
------ ---------------------------------------
ftp File Transfer Protocol
file File Transfer Protocol (same as ftp)
news Internet News Protocol (Usenet)
gopher Gopher
telnet Telnet
http Hypertext Transport Protocol
We've used all of these before, except for "http". If you ever see a URL with "http" at the beginning of it, that means that the file is a hypertext document (with hypertext links to other documents). The rest of a URL -- the stuff after the colon -- is the address of that particular file. In general, two slashes (//) after the colon indicates a machine name or address.
For example,
file://wuarchive.wustl.edu/mirrors/msdos/graphics/gifkit.zip
is the URL for an FTP file at wuarchive.wustl.edu, and http://info.cern.ch:80/default.html
is the URL for a hypertext document at info.cern.ch, port 80.
Posted bySumedh
at
11:50 PM
0
comments
Labels:
Internet
Q & A
- I'm subscribed to the 'squirrel' LISTSERV mailing list (SQUIRREL@UA1VM.UA.EDU). I just heard a story about how an energetic squirrel once crashed the mainframe at Mississippi State University during registration (a true story) and I want to share this information with everyone on the 'squirrel' mailing list. What address should I send my e-mail letter to if I want the letter to be distributed to everyone on the "squirrel" list?
- SQUIRREL@UA1VM.UA.EDU
Letters should be sent to the list address.
- I'm looking for a list of all known LISTSERV lists. What LISTSERV command should I use to get such a list?
- LISTS GLOBAL
BTW, you can send the lists global command to any LISTSERV address ... it is a universal LISTSERV command.
- Okay, I just got the "list of lists." I found a neat list called "navigate" but the list of lists only shows the address as NAVIGATE@UBVM. What address should I send my subscribe command to?
- LISTSERV@UBVM.BITNET
This was a tough question. You had to remember to convert the LIST@NODE address to LIST@NODE.BITNET. You also had to remember that LISTSERV commands have to be sent to the LISTSERV address. :)
- I want to subscribe to the alt.abuse.recovery newsgroup. I sent a e-mail letter to LISTSERV@UA1VM.UA.EDU with the command SUBSCRIBE ALT.ABUSE.RECOVERY JOE STUDENT, and I got back an error message saying that "alt.abuse.recovery" is unknown to the LISTSERV. What's going on?
- alt.abuse.recovery is a Usenet newsgroup; to subscribe to the group from a Usenet reader.
Remember, if the group's name has periods in it, it probably is a Usenet newsgroup.
- How can I unsubscribe from the Roadmap workshop?
- send an e-mail letter to LISTSERV@UA1VM.UA.EDU which says UNSUB * in the body of your letter.
PART TWO: TRUE OR FALSE
- FALSE
There is nothing wrong with giving my best friend my password. (Never give your password to ANYONE!)
- FALSE
WRITING IN ALL CAPS IS CONSIDERED TO BE GOOD NETIQUETTE. (Writing in all caps is ANNOYING.)
- FALSE
The word "squirrel" would be an excellent password. (Squirrel is in the dictionary. NEVER use a password that is in the dictionary.)
- FALSE
Letters to a LISTSERV list should be sent to the LISTSERV address. (Letters should be sent to the list address.)
- FALSE
LISTSERV commands should be sent to the list address. (Commands should be sent to the LISTSERV address.)
PART THREE: SHORT ANSWER
- One of my friends just e-mailed me a chain letter offering me good luck so long as I send the chain letter to ten additional people. Heck, there are several hundred people subscribed to each of the Usenet newsgroups and LISTSERV lists. Why can't I just forward the chain letter to these groups?
Chain letters violate the relevant posting rule (and are likely to turn you into flame-bait).
Also -- and this was not covered in the lessons -- most networks have Acceptable Usage policies that strictly prohibit chain letters.
If someone sends you a chain letter, forward the chain letter (along with an angry reply) to that person's postmaster.
- Patrick's dad said "DON'T include the entire contents of a previous posting in your reply(s)." Why can't I do this?
It is rude. When you re-post the entire contents of a previous posting, you are forcing the readers to reread something that they have already read.
The purpose of reposting is to remind everyone what was said, not to repeat what was said :)
- If someone spams a list that I am on, what should I do? Why shouldn't I just send my angry replies to the list?
Your best bet is to ignore it. If you are really angry, reply to the poster at his e-mail address, or complain to the poster's site administrator.
You should not send your replies to the list because the spammer is probably not even a member or the list, and, if he is, he probably does not care about what you think.
BONUS QUESTIONS
- How many National Football Championships has the University of Alabama's Crimson Tide football team won?
- 12
- Patrick Douglas Crispen is:
- a squirrel
Posted bySumedh
at
11:50 PM
0
comments
Labels:
Internet
FTP
I hope that you are having as much fun participating in this workshop as I am having teaching it! :)
In the last lesson we used a tool called "Telnet" which allowed us to access remote computers and run programs on those remote computers.
In the next few lessons we are going to use a tool called "File Transfer Protocol" (or "FTP") which will allow us to access remote computers and retrieve files from these computers. Actually, it is probably more accurate to say that we will be using "anonymous" FTP, but I'll explain that in a minute.
What sort of files are available through FTP? Well, "hundreds of systems connected to the Internet have file libraries, or archives, accessible to the public. Much of this consists of free or low-cost {computer} programs for virtually every make of computer. If you want a different communications program for your IBM, or feel like playing a new game on your Amiga, you'll {probably} be able to get it {using FTP}."
"But there are also libraries of documents as well. If you want a copy of a recent U.S. Supreme Court decision, you can {get a copy using FTP}. Copies of historical documents, from the Magna Carta to the Declaration of Independence are also yours for the asking ... You can also find song lyrics, poems, {and} even summaries of every {Animaniacs} episode ever made. You can also find extensive files detailing everything you could possibly want to know about the Net itself". (1)
Before we get too in-depth into FTP, I think that now is as good a time as any to quickly review the client/server relationship that I briefly mentioned in lesson three (MAP03: Levels of Internet Connectivity).
"Client" is just another word for a program running on your service provider's system that enables you to talk to, and get stuff from, distant computers. The client on your service provider's system takes your commands and turns them into requests for information from programs -- "servers" -- running on other, distant computers.
The client/server relationship allows you to TELNET into remote computers and run programs on those remote computers, and it also allows you to use FTP to get files from remote sites.
The client/server relationship is also what makes my job as an Internet trainer quite difficult. While all of the FTP clients out there do the exact same thing, they all use different commands to do it.
Fortunately, if you can understand the basics of what happens during an FTP session, the commands -- no matter what client you are using -- are pretty easy.
The basic steps in an FTP session are:
- Start-up your FTP client.
- Give your FTP client an address to connect to (and, just like TELNET, step one and two can be combined into one easy step).
- Identify yourself to the remote site.
- Give the remote site your password.
- Look around the directory for files.
- Change Directories.
- Set the transfer mode (optional).
- Get the file(s) you want.
- Quit.
The best way to understand what is going on is to look at a sample FTP session. The University of Alabama's connection to the Internet is through SURAnet (a large regional network), so I'm going to FTP to them. Before I do that, though, there are three things that I need to emphasize:
- FTP requires a heck of a lot of resources, both on your system and on the remote system that you are accessing. Because of this, FTP sessions should only be run during non-working hours (usually 6 PM to 6 AM local).
I realize that this constraint is going to be a difficult for those of you who are reading this at work, and who only have Internet (and FTP access) through your employer. However, as responsible Internet citizens we have to remember that the FTP sites are providing us FTP access out of the kindness of their hearts. If we take advantage of this kindness, and access various FTP sites during working hours, those FTP sites may decide to close their doors to the public ... and then EVERYONE loses.
- In light of what was said in #1, please do not flood SURAnet. In MAP15: FTPMAIL I will post a list of FTP sites that you can visit (much like the TELNET list I posted in MAP12: Telnet (Part Two)). Until I post that list, just sit back and enjoy the show :)
- Since this lesson is already over 100 lines long, I hope that you won't mind if I break this FTP lesson into two lessons. The first lesson is in this document, and I'll finish it in the next document. This will mean that we will end this lesson in the middle of our sample FTP session, but it will also mean that this lesson won't be the size of a small book :)
Starting an FTP session is pretty easy. For most of you, all you have to do to start-up your FTP client is type ftp
on you system's command line (or, if you are in a Windows or Mac environment, double-click on the FTP icon). >From there, you would give the client an FTP address to connect to.
Like TELNET, however, there is a way to combine these two steps into one, easy step. For most of you, to access your FTP client and give your client an address to hook up to, all you would have to do it type the command
ftp
and replace with the address of the FTP site that you want your client software to access. In our example, the SURAnet FTP address is ftp.sura.net, so I would type
ftp ftp.sura.net
to start an FTP session. (Note that the second "ftp" is part of SURAnet's FTP address. If I wanted to ftp to info.umd.edu, I would type "ftp info.umd.edu"; if I wanted to ftp to lcs.mit.edu, I would type "ftp lcs.mit.edu"). Once I hit the enter key, the following appears on my screen:
ftp ftp.sura.net
Connecting to ftp.sura.net 128.167.254.179, port 21
220 nic.sura.net FTP server (Version wu-2.4(1) Fri May 20 10:20:58
EDT 1994) ready.
USER (identify yourself to the host):
The second line tells me that my system is connecting to ftp.sura.net (and even gives me the IP number for ftp.sura.net), the third line is some automatic information from SURAnet, and the bottom line is asking me to log in. If I had an account on the SURAnet system, I would enter my SURAnet user ID. But, since I don't have an account on this system, I have to find another way to access the system. ;)
This is where the "anonymous" FTP I mentioned earlier comes in :) The other way to access some FTP sites -- at least those FTP sites that allow outside access -- is to use the userid "anonymous". By using the name "anonymous", you are telling that FTP site that you aren't a regular user of that site, but you would still like to access that FTP site, look around, and retrieve files.
So, where it says USER, I type the word
anonymous
hit enter, and cross my fingers. If SURAnet does not allow anonymous access, I'm about to find out :) >>>USER anonymous
331 Guest login ok, send your complete e-mail address as password.
Password:
COOL! Its going to let me in. All I have to do is give the site a password. Out of politeness to the FTP site, if you login as "anonymous", you need to use your full Internet address as your password. This helps the FTP site keep track of who has visited its site.
So, since it wants my password, and since the password for any anonymous FTP session is my full Internet address, I type
pcrispe1@ua1vm.ua.edu
(Stop laughing -- p-crispy-one is not funny!!). Once I hit enter, my screen fills with the following: >>>PASS ********
230- SURAnet ftp server running wuarchive experimental ftpd
230-
230-Welcome to the SURAnet ftp server. If you have any problems with
230-the server please mail the to systems@sura.net. If you do have problems,
230-please try using a dash (-) as the first character of your password
230- -- this will turn off the continuation messages that may be confusing
230-your ftp client.
...
230 Guest login ok, access restrictions apply.
Command:
Notice the line "Guest login ok, access restrictions apply." This means that the site has given me access, but I only have access to the files that are available to the general public. Okay ... now what? I've started-up my FTP client, I've given the client an FTP address to connect to, I've identified myself to the remote site (I told it that I am anonymous), and I've given the site my password.
Now it's time to see what sort of files and directories are around, and to get those files ... which we will do in the next lesson :)
Posted bySumedh
at
11:49 PM
0
comments
Labels:
Internet
Subscribe to:
Posts (Atom)
On a page containing a news story, we might put the name of our site into an
tag and the
topic of the story into an tag
Since heading tags typically make text contained in them larger than normal text on the page, this is a
visual cue to users that this text is important and could help them understand something about the
type of content underneath the heading text. Multiple heading sizes used in order create a hierarchical
structure for your content, making it easier for users to navigate through your document.
Google's Search Engine Optimization Starter Guide, Version 1.1, published 13 November 2008
Good practices for heading tags
• Imagine you're writing an outline - Similar to writing an outline for a large paper, put some
thought into what the main points and sub-points of the content on the page will be and
decide where to use heading tags appropriately.
Avoid:
• placing text in heading tags that wouldn't be helpful in defining the structure of
the page
• using heading tags where other tags like and may be more
appropriate
• erratically moving from one heading tag size to another
• Use headings sparingly across the page - Use heading tags where it makes sense. Too
many heading tags on a page can make it hard for users to scan the content and determine
where one topic ends and another begins.
Avoid:
• excessively using heading tags throughout the page
• putting all of the page's text into a heading tag
• using heading tags only for styling text and not presenting structure
Optimize your use of images
Images may seem like a straightforward component of your site, but you can optimize your use of
them. All images can have a distinct filename and "alt" attribute, both of which you should take
advantage of.
The "alt" attribute allows you to specify alternative text for the image if it cannot be displayed for some
reason.
Our alt text here is a brief but accurate description of the image
Why use this attribute? If a user is viewing your site on a browser that doesn't support images, or is
using alternative technologies, such as a screen reader, the contents of the alt attribute provide
information about the picture.
Google's Search Engine Optimization Starter Guide, Version 1.1, published 13 November 2008
Our image wasn't displayed to the user for some reason, but at least the alt text was
Another reason is that if you're using an image as a link, the alt text for that image will be treated
similarly to the anchor text of a text link. However, we don't recommend using too many images for
links in your site's navigation when text links could serve the same purpose. Lastly, optimizing your
image filenames and alt text makes it easier for image search projects like Google Image Search to
better understand your images.
Good practices for images
• Use brief, but descriptive filenames and alt text - Like many of the other parts of the page
targeted for optimization, filenames and alt text (for ASCII languages) are best when they're
short, but descriptive.
Avoid:
• using generic filenames like "image1.jpg", "pic.gif", "1.jpg" when possible
(some sites with thousands of images might consider automating the naming of
images)
• writing extremely lengthy filenames
• stuffing keywords into alt text or copying and pasting entire sentences
• Supply alt text when using images as links - If you do decide to use an image as a link,
filling out its alt text helps Google understand more about the page you're linking to. Imagine
that you're writing anchor text for a text link.
Avoid:
• writing excessively long alt text that would be considered spammy
• using only image links for your site's navigation
• Store images in a directory of their own - Instead of having image files spread out in
numerous directories and subdirectories across your domain, consider consolidating your
images into a single directory (e.g. brandonsbaseballcards.com/images/). This simplifies the
path to your images.
Google's Search Engine Optimization Starter Guide, Version 1.1, published 13 November 2008
• Use commonly supported filetypes - Most browsers support JPEG, GIF, PNG, and BMP
image formats. It's also a good idea to have the extension of your filename match with the
filetype.
Make effective use of robots.txt
A "robots.txt" file tells search engines whether they can access and therefore crawl parts of your
site. This file, which must be named "robots.txt", is placed in the root directory of your site.
The address of our robots.txt file
All compliant search engine bots (denoted by the wildcard * symbol) shouldn't access and
crawl the content under /images/ or any URL whose path begins with /search
You may not want certain pages of your site crawled because they might not be useful to users if
found in a search engine's search results. If you do want to prevent search engines from crawling
your pages, Google Webmaster Tools has a friendly robots.txt generator to help you create this file.
Note that if your site uses subdomains and you wish to have certain pages not crawled on a particular
subdomain, you'll have to create a separate robots.txt file for that subdomain. For more information
on robots.txt, we suggest this Webmaster Help Center guide on using robots.txt files.
There are a handful of other ways to prevent content appearing in search results, such as adding
"NOINDEX" to your robots meta tag, using .htaccess to password protect directories, and using
Google Webmaster Tools to remove content that has already been crawled. Google engineer Matt
Cutts walks through the caveats of each URL blocking method in a helpful video.
Good practices for robots.txt
• Use more secure methods for sensitive content - You shouldn't feel comfortable using
robots.txt to block sensitive or confidential material. One reason is that search engines could
still reference the URLs you block (showing just the URL, no title or snippet) if there happen
to be links to those URLs somewhere on the Internet (like referrer logs). Also, non-compliant
or rogue search engines that don't acknowledge the Robots Exclusion Standard could
disobey the instructions of your robots.txt. Finally, a curious user could examine the
directories or subdirectories in your robots.txt file and guess the URL of the content that you
Google's Search Engine Optimization Starter Guide, Version 1.1, published 13 November 2008
don't want seen. Encrypting the content or password-protecting it with .htaccess are more
secure alternatives.
Avoid:
• allowing search result-like pages to be crawled (users dislike leaving one
search result page and landing on another search result page that doesn't add
significant value for them)
• allowing a large number of auto-generated pages with the same or only slightly
different content to be crawled: "Should these 100,000 near-duplicate pages
really be in a search engine's index?"
• allowing URLs created as a result of proxy services to be crawled
Be aware of rel="nofollow" for links
Setting the value of the "rel" attribute of a link to "nofollow" will tell Google that certain links on your
site shouldn't be followed or pass your page's reputation to the pages linked to. Nofollowing a link is
adding rel="nofollow" inside of the link's anchor tag.
If you link to a site that you don't trust and don't want to pass your site's reputation to, use
nofollow
When would this be useful? If your site has a blog with public commenting turned on, links within
those comments could pass your reputation to pages that you may not be comfortable vouching for.
Blog comment areas on pages are highly susceptible to comment spam. Nofollowing these useradded
links ensures that you're not giving your page's hard-earned reputation to a spammy site. Many
blogging software packages automatically nofollow user comments, but those that don't can most
likely be manually edited to do this. This advice also goes for other areas of your site that may involve
user-generated content, such as guestbooks, forums, shout-boards, referrer listings, etc. If you're
willing to vouch for links added by third parties (e.g. if a commenter is trusted on your site), then
there's no need to use nofollow on links; however, linking to sites that Google considers spammy can
affect the reputation of your own site. The Webmaster Help Center has more tips on avoiding
comment spam, like using CAPTCHAs and turning on comment moderation.
Google's Search Engine Optimization Starter Guide, Version 1.1, published 13 November 2008
A comment spammer leaves a message on one of our news posts, hoping to get some of our
site's reputation
Another use of nofollow is when you're writing content and wish to reference a website, but don't want
to pass your reputation on to it. For example, imagine that you're writing a blog post on the topic of
comment spamming and you want to call out a site that recently comment spammed your blog. You
want to warn others of the site, so you include the link to it in your content; however, you certainly
don't want to give the site some of your reputation from your link. This would be a good time to use
nofollow.
Lastly, if you're interested in nofollowing all of the links on a page, you can use "nofollow" in your
robots meta tag, which is placed inside the tag of that page's HTML. The Webmaster Central
Blog provides a helpful post on using the robots meta tag. This method is written as name="robots" content="nofollow">.
This nofollows all of the links on a page
Google's Search Engine Optimization Starter Guide, Version 1.1, published 13 November 2008
Promote your website in the right ways
While most of the links to your site will be gained gradually, as people discover your content through
search or other ways and link to it, Google understands that you'd like to let others know about the
hard work you've put into your content. Effectively promoting your new content will lead to faster
discovery by those who are interested in the same subject. As with most points covered in this
document, taking these recommendations to an extreme could actually harm the reputation of your
site.
Good practices for promoting your website
• Blog about new content or services - A blog post on your own site letting your visitor base
know that you added something new is a great way to get the word out about new content or
services. Other webmasters who follow your site or RSS feed could pick the story up as well.
• Don't forget about offline promotion - Putting effort into the offline promotion of your
company or site can also be rewarding. For example, if you have a business site, make sure
its URL is listed on your business cards, letterhead, posters, etc. You could also send out
recurring newsletters to clients through the mail letting them know about new content on the
company's website.
• Know about social media sites - Sites built around user interaction and sharing have made
it easier to match interested groups of people up with relevant content.
Avoid:
• attempting to promote each new, small piece of content you create; go for big,
interesting items
• involving your site in schemes where your content is artificially promoted to the
top of these services
• Add your business to Google's Local Business Center - If you run a local business,
adding its information to Google's Local Business Center will help you reach customers on
Google Maps and web search. The Webmaster Help Center has more tips on promoting
your local business.
• Reach out to those in your site's related community - Chances are, there are a number
of sites that cover topic areas similar to yours. Opening up communication with these sites is
usually beneficial. Hot topics in your niche or community could spark additional ideas for
content or building a good community resource.
Avoid:
• spamming link requests out to all sites related to your topic area
• purchasing links from another site with the aim of getting PageRank instead of
traffic
Google's Search Engine Optimization Starter Guide, Version 1.1, published 13 November 2008
Make use of free webmaster tools
Major search engines, including Google, provide free tools for webmasters. Google's Webmaster
Tools help webmasters better control how Google interacts with their websites and get useful
information from Google about their site. Using Webmaster Tools won't help your site get preferential
treatment; however, it can help you identify issues that, if addressed, can help your site perform better
in search results. With the service, webmasters can:
• see which parts of a site Googlebot had problems crawling
• upload an XML Sitemap file
• analyze and generate robots.txt files
• remove URLs already crawled by Googlebot
• specify the preferred domain
• identify issues with title and description meta tags
• understand the top searches used to reach a site
• get a glimpse at how Googlebot sees pages
• remove unwanted sitelinks that Google may use in results
• receive notification of quality guideline violations and file for a site reconsideration
Yahoo! (Yahoo! Site Explorer) and Microsoft (Live Search Webmaster Tools) also offer free tools for
webmasters.
Take advantage of web analytics services
If you've improved the crawling and indexing of your site using Google Webmasters Tools or other
services, you're probably curious about the traffic coming to your site. Web analytics programs like
Google Analytics are a valuable source of insight for this. You can use these to:
• get insight into how users reach and behave on your site
• discover the most popular content on your site
• measure the impact of optimizations you make to your site (e.g. did changing those title and
description meta tags improve traffic from search engines?)
For advanced users, the information an analytics package provides, combined with data from your
server log files, can provide even more comprehensive information about how visitors are interacting
with your documents (such as additional keywords that searchers might use to find your site).
Lastly, Google offers another tool called Google Website Optimizer that allows you to run experiments
to find what on-page changes will produce the best conversion rates with visitors. This, in combination
with Google Analytics and Google Webmaster Tools (see our video on using the "Google Trifecta"), is
a powerful way to begin improving your site.
Google's Search Engine Optimization Starter Guide, Version 1.1, published 13 November 2008
Helpful resources for webmasters
Google Webmaster Help Group - Have questions or feedback on our guide? Let us know
Google Webmaster Central Blog - Frequent posts by Googlers on how to improve your website
Google Webmaster Help Center - Filled with in-depth documentation on webmaster-related issues
Google Webmaster Tools - Optimize how Google interacts with your website
Google Webmaster Guidelines - Design, content, technical, and quality guidelines from Google
Google Analytics - Find the source of your visitors, what they're viewing, and benchmark changes
Google Website Optimizer - Run experiments on your pages to see what will work and what won't
Tips on Hiring an SEO - If you don't want to go at it alone, these tips should help you choose an
SEO company
Except as otherwise noted, the content of this document is licensed under the Creative Commons
Attribution 3.0 License.
Google's Search Engine Optimization Starter Guide, Version 1.1, published 13 November 2008
Posted bySumedh
at
8:27 PM
0
comments
WWW 2
Sunday, December 14, 2008
In the last lesson I told you that "URLs" is pronounced "earls." In fact, a lot of people still use the initials and call them "U-R-Ls." I personally prefer calling them "earls" because "earls" rhymes with "squirrels," but the choice of what you call them is completely up to you :)
Let's take a look at how a sample WWW browser works. There are three basic types of WWW browsers available: line-mode browsers, full screen browsers (like Lynx), and graphical browsers (like Mosaic).
Line-mode browsers are about as user un-friendly as you can get. This is hard to describe, but line-mode browsers work a little like FTP inasmuch as you type a command, get some information on your screen, type a new command, get some more information, and so on ...
A full screen browser puts a menu on your screen that looks a little like the Gopher menus that we saw in MAP18: Gopher (Part One). You move the cursor up and down the screen, select a highlighted link, press enter or return, and you are automatically taken to a new document or file (your fill screen browser may work differently than this, though).
Graphical browsers allow you to access not only text, but also pictures and sound (a.k.a. "hypermedia"). In fact, these pictures can be put in Web documents (a.k.a. Web pages), making that Web page look less like a Gopher menu and more like a page from a color magazine! Most graphical browsers use a mouse, and you point-and-click on a highlighted link to access it.
The URL for the sample Web page that we are going to use in this lesson is
http://ua1vm.ua.edu/~vmhttpd/index.html
and I am going to be using the UF WWW Browser for CMS to access this page. I'll talk a little more about how you can access a WWW browser in a few minutes, but I first want to show you what a basic Web page looks like. The UF WWW Browser for CMS is the browser that my service provider uses, and it is a full screen browser. The browser that you use -- if you can even access a WWW browser -- will probably look and work a little differently than what you will see in this example.
Finally, in real life my browser highlights the links by making them a different color than the rest of the text. There is no way for me to use different colors in this letter, so I have highlighted the links in this example by surrounding them with a (* *).
Just like I can access an item in a Gopher menu by pointing at it and selecting it, I can access a WWW link by pointing at it and selecting it.
Enough talk. Time for the example.
I access my provider's WWW browser, and the following appears on my screen:
Albert 1.2.0 (UF WWW Browser for CMS) Screen 1 of 2 (more screens)
Viewing=http://ua1vm.ua.edu/~vmhttpd/index.html
Title=UA1VM WWW Home Page
Welcome to The University of Alabama's CMS WWW Server
This CMS server is still under development. Any (*comments*)
or (*suggestions*) will be greatly appreciated. Thank you.
Gopher Sites:
- (*UA1VM CMS Gopher Server*)
- (*UA1IX AIX/370 Gopher Server*)
- (*RISC/6000 Gopher Server*)
- (*RICEVM1 CMS Gopher Server*)
Telnet Sessions:
- (*UA1VM.UA.EDU*)
- (*UA1IX.UA.EDU - Line Mode*)
- (*RISC.UA.EDU - Line Mode*)
WWW Sites:
- The University of Alabama Libraries (*WWW*)
- The University of Illinois at Urbana-Champaign (*WWW*)
- The Alabama Supercomputer Network (*WWW*)
- NASA Information Services via (*WWW*)
Leisure:
- (*Intertext Magazine*) - Electronic Fictional Magazine at The
University of Michigan
- (*Wiretap*) - a gopher to Cupertino, California
- (*NNR*) - UA1VM's Network News Reader
Other Neat Stuff:
- The University of Alabama Library's On-Line (*Card Catalog*)
- a (*map*) of The University of Alabama campus
... snip snip snip ...
COOL! I can select any of these links -- the words set apart from the rest of the text with a (* *) -- and be transported to that particular link.
>From this one Web page, I can access Gopher, telnet, and even other Web pages! (I can also access FTP, although this page does not show it).
We've seen a lot of Gopher and telnet recently. Let's take a look at another Web page. Since I used to be a Simulations Director at the United States Space Camp (did I forget to tell you that?), I'm going to move my cursor down to the (*WWW*) link next to "NASA Information Services", press enter, and see what happens:
Albert 1.2.0 (UF WWW Browser for CMS) Screen 1 of 2 (more screens)
Viewing=http://hypatia.gsfc.nasa.gov/NASA_homepage.html
Title=NASA Information Services via World Wide Web
National Aeronautics and Space Administration
NASA Logo
(*World Wide Web (WWW) information services*)
(*Hot Topics*) NASA news and subjects of public interest
(*NASA Strategic Plan*)
(*NASA Strategies, Policies, and Public Affairs*)
(*NASA Online Educational Resources*)
(*NASA Information Sources by Subject*)
... snip snip snip ...
This is certainly more interesting than SURAnet! ;) >From this Web page I can access OTHER Web pages, and from those Web pages I can access even MORE Web pages, and so on ...
Yesterday I told you that it is possible for you to connect directly to a specific Internet resource so long as you know the resource's URL. How you do that depends on the browser that you use.
For the line-mode browser at CERN, for example, the command to connect to a particular URL is
GO
replacing with the URL of the Internet resource that you want to access. In Lynx, you just select the "GO" link on the browser's start-up page; in most graphical browsers (like Mosaic), there is usually an "Open URL" option in one of the menus. (1) Before you can do this, however, you have to first access the Web. There are three ways that you can do this:
- Through a browser stored on your local Internet service provider's machine. Ask your provider if your site has a Web browser, and how you can access it.
- Through a browser stored on your own machine. Until recently, you had to have a SLIP or PPP connection to do this. Some providers -- providers who FLOODED my mailbox when I did not talk about the special Level 2.0002746278546723 access that they offer -- now allow you to store a Web browser on your own machine even though you only have Level 2 access.
If you do not have a SLIP or PPP connection, contact your provider BEFORE you store a Web reader on your own computer and double-check that your provider will support the browser. Most service providers can not support a Web browser unless you have a SLIP or PPP connection.
- Through a telnet connection to a publicly-accessible Web browser.
If you have a SLIP or a PPP connection, the WWW FAQ that I have stored on the University of Alabama's LISTSERV file server (WWW FAQ1) has a list of FTP sites where you can get specific Web browsers. (Do me a favor ... re-read that last sentence. Did you EVER think you would understand a sentence like that? Isn't this workshop COOL?!!)
If you do not have access to a Web browser through your site, you may still be able to access a Web browser through telnet. The following are a few of the public Web sites: (1)
telnet address comments
-------------- -----------------------
info.cern.ch No password is required. This is in
Switzerland, so U.S. users might be
better off using a closer browser.
www.cc.ukans.edu The "Lynx" full screen browser, which
requires a vt100 terminal. Login as www.
Does not allow users to "go" to arbitrary
URLs.
www.njit.edu Login as www. A full-screen browser at
the New Jersey Institute of Technology.
sun.uakom.cs Slovakia. Has a slow link, so only use
from nearby.
info.funet.fi Login as www. Offers several browsers,
including Lynx (goto option disabled
there too).
fserv.kfki.hu Hungary. Has a slow link, so use from
nearby. Login as www.
Once you are on the Web, it is possible for you to do keyword searches (much like the Veronica searches we did in MAP21: Veronica) using one of the Web's many search engines. One of the best Web search engines is the WebCrawler. The WebCrawler's URL is http://www.biotech.washington.edu/WebQuery.html
and the WebCrawler searches for documents whose title AND CONTENT match your keyword. Another Web search engine you probably will want to check out is the World Wide Web Worm. The Worm's URL is
http://www.cs.colorado.edu/home/mcbryan/WWWW.html
The Worm is not at thorough as the WebCrawler, but it is still a pretty competent search engine. Both of these search engines provide really good on-line help and instructions.
One last thing, and I am though for this lesson. Please remember:
- The "Web" is the collection of all of the files and information that can be accessed by a Web browser.
- Mosaic and Lynx and just BROWSERS that allow you to access the Web.
Posted bySumedh
at
11:51 PM
1 comments
Labels:
Internet
WWW
I wish I had six weeks just to talk about the World Wide Web (a.k.a. WWW or "the Web."). If you think Gopher is neat, wait until you start playing around on the Web :)
Unfortunately, I don't have six weeks to talk about the Web -- I only have two lessons. Because of this, we are going to go through the Web like Sherman went through Georgia (1).
That's the bad news. The good news is that there are a lot of REALLY good Web guides available, and I am even seriously considering developing my own Web workshop that I will offer late next year (2).
Until that time comes, however, let's talk about the BASICS of the Web.
In the last few lessons I showed you how most Gopher menus are linked together. We started out in the gopher.squirrel.com root menu, and eventually ended up at the SURAnet gopher menu. We were able to do this because the menus that we traveled through had links to menus and files that were located at other Gopher sites.
Because Gopher menus are linked together, a whole world of information is available to us with just a few keystrokes!
Imagine if we were able to take these links one step further. Instead of linking menus, we could link documents together. You could read one document, find a keyword in that document that really interests you, touch that keyword, and automatically be taken to a NEW document somewhere else in the world -- and this new document could even have links to OTHER documents around the world, and so on.
Sound too good to be true? It isn't, thanks to something called "hypertext." If you have ever played with Apple's HyperCard program or the "help" menus in the latest Microsoft packages, you have already experienced hypertext. You "select" a highlighted word -- usually by clicking on it with a mouse -- and you are taken into an entirely new document or help screen.
The World Wide Web is based on hypertext. It is possible for you to go roaming around the Web, bouncing from document to document, using nothing but the links in those documents!
Just as you can access Gopherspace through a Gopher server or client, you can access the Web through something called a "browser." A browser can read documents, fetch documents, access files by FTP, read Usenet newsgroups, telnet into remote sites, and even travel around Gopherspace. In short, everything that we have talked about from MAP08: Usenet to MAP22:Gophermail can be done using nothing but a Web browser!
The Web is able to accomplish all of this thanks to something called URLs ("earls") -- Universal Resource Locators. URLs list the exact location of any Internet resource.
If you think about it, giving every Internet resource a unique address is the hard part. Once you have given something an address, linking to it is pretty easy :)
What is really special about the Web is that the Web does all of this "behind the scenes." It is possible for you to bounce from one link to another without ever knowing the exact address of where you are, or even how you got there.
If you ever want to jump directly to a particular Internet resource, however, you are going to need to know a little bit more about URLs. Here are a few basic URLs:
file://wuarchive.wustl.edu/mirrors/msdos/graphics/gifkit.zip
ftp://wuarchive.wustl.edu/mirrors
http://info.cern.ch:80/default.html
news:alt.hypertext
telnet://dra.com
Gee ... those look a little like FTP addresses, don't they? The first part of an URL -- the stuff before the colon -- tells the browser how to access that particular file. For example, to access
ftp://wuarchive.wustl.edu/mirrors
your browser would use FTP. Most of the access methods are pretty straight-forward. Here is a list of some of the more common access methods that you are going to see listed in the first part of URLs:
method what it stands for
------ ---------------------------------------
ftp File Transfer Protocol
file File Transfer Protocol (same as ftp)
news Internet News Protocol (Usenet)
gopher Gopher
telnet Telnet
http Hypertext Transport Protocol
We've used all of these before, except for "http". If you ever see a URL with "http" at the beginning of it, that means that the file is a hypertext document (with hypertext links to other documents). The rest of a URL -- the stuff after the colon -- is the address of that particular file. In general, two slashes (//) after the colon indicates a machine name or address.
For example,
file://wuarchive.wustl.edu/mirrors/msdos/graphics/gifkit.zip
is the URL for an FTP file at wuarchive.wustl.edu, and http://info.cern.ch:80/default.html
is the URL for a hypertext document at info.cern.ch, port 80.
Posted bySumedh
at
11:50 PM
0
comments
Labels:
Internet
Q & A
- I'm subscribed to the 'squirrel' LISTSERV mailing list (SQUIRREL@UA1VM.UA.EDU). I just heard a story about how an energetic squirrel once crashed the mainframe at Mississippi State University during registration (a true story) and I want to share this information with everyone on the 'squirrel' mailing list. What address should I send my e-mail letter to if I want the letter to be distributed to everyone on the "squirrel" list?
- SQUIRREL@UA1VM.UA.EDU
Letters should be sent to the list address.
- I'm looking for a list of all known LISTSERV lists. What LISTSERV command should I use to get such a list?
- LISTS GLOBAL
BTW, you can send the lists global command to any LISTSERV address ... it is a universal LISTSERV command.
- Okay, I just got the "list of lists." I found a neat list called "navigate" but the list of lists only shows the address as NAVIGATE@UBVM. What address should I send my subscribe command to?
- LISTSERV@UBVM.BITNET
This was a tough question. You had to remember to convert the LIST@NODE address to LIST@NODE.BITNET. You also had to remember that LISTSERV commands have to be sent to the LISTSERV address. :)
- I want to subscribe to the alt.abuse.recovery newsgroup. I sent a e-mail letter to LISTSERV@UA1VM.UA.EDU with the command SUBSCRIBE ALT.ABUSE.RECOVERY JOE STUDENT, and I got back an error message saying that "alt.abuse.recovery" is unknown to the LISTSERV. What's going on?
- alt.abuse.recovery is a Usenet newsgroup; to subscribe to the group from a Usenet reader.
Remember, if the group's name has periods in it, it probably is a Usenet newsgroup.
- How can I unsubscribe from the Roadmap workshop?
- send an e-mail letter to LISTSERV@UA1VM.UA.EDU which says UNSUB * in the body of your letter.
PART TWO: TRUE OR FALSE
- FALSE
There is nothing wrong with giving my best friend my password. (Never give your password to ANYONE!)
- FALSE
WRITING IN ALL CAPS IS CONSIDERED TO BE GOOD NETIQUETTE. (Writing in all caps is ANNOYING.)
- FALSE
The word "squirrel" would be an excellent password. (Squirrel is in the dictionary. NEVER use a password that is in the dictionary.)
- FALSE
Letters to a LISTSERV list should be sent to the LISTSERV address. (Letters should be sent to the list address.)
- FALSE
LISTSERV commands should be sent to the list address. (Commands should be sent to the LISTSERV address.)
PART THREE: SHORT ANSWER
- One of my friends just e-mailed me a chain letter offering me good luck so long as I send the chain letter to ten additional people. Heck, there are several hundred people subscribed to each of the Usenet newsgroups and LISTSERV lists. Why can't I just forward the chain letter to these groups?
Chain letters violate the relevant posting rule (and are likely to turn you into flame-bait).
Also -- and this was not covered in the lessons -- most networks have Acceptable Usage policies that strictly prohibit chain letters.
If someone sends you a chain letter, forward the chain letter (along with an angry reply) to that person's postmaster.
- Patrick's dad said "DON'T include the entire contents of a previous posting in your reply(s)." Why can't I do this?
It is rude. When you re-post the entire contents of a previous posting, you are forcing the readers to reread something that they have already read.
The purpose of reposting is to remind everyone what was said, not to repeat what was said :)
- If someone spams a list that I am on, what should I do? Why shouldn't I just send my angry replies to the list?
Your best bet is to ignore it. If you are really angry, reply to the poster at his e-mail address, or complain to the poster's site administrator.
You should not send your replies to the list because the spammer is probably not even a member or the list, and, if he is, he probably does not care about what you think.
BONUS QUESTIONS
- How many National Football Championships has the University of Alabama's Crimson Tide football team won?
- 12
- Patrick Douglas Crispen is:
- a squirrel
Posted bySumedh
at
11:50 PM
0
comments
Labels:
Internet
FTP
I hope that you are having as much fun participating in this workshop as I am having teaching it! :)
In the last lesson we used a tool called "Telnet" which allowed us to access remote computers and run programs on those remote computers.
In the next few lessons we are going to use a tool called "File Transfer Protocol" (or "FTP") which will allow us to access remote computers and retrieve files from these computers. Actually, it is probably more accurate to say that we will be using "anonymous" FTP, but I'll explain that in a minute.
What sort of files are available through FTP? Well, "hundreds of systems connected to the Internet have file libraries, or archives, accessible to the public. Much of this consists of free or low-cost {computer} programs for virtually every make of computer. If you want a different communications program for your IBM, or feel like playing a new game on your Amiga, you'll {probably} be able to get it {using FTP}."
"But there are also libraries of documents as well. If you want a copy of a recent U.S. Supreme Court decision, you can {get a copy using FTP}. Copies of historical documents, from the Magna Carta to the Declaration of Independence are also yours for the asking ... You can also find song lyrics, poems, {and} even summaries of every {Animaniacs} episode ever made. You can also find extensive files detailing everything you could possibly want to know about the Net itself". (1)
Before we get too in-depth into FTP, I think that now is as good a time as any to quickly review the client/server relationship that I briefly mentioned in lesson three (MAP03: Levels of Internet Connectivity).
"Client" is just another word for a program running on your service provider's system that enables you to talk to, and get stuff from, distant computers. The client on your service provider's system takes your commands and turns them into requests for information from programs -- "servers" -- running on other, distant computers.
The client/server relationship allows you to TELNET into remote computers and run programs on those remote computers, and it also allows you to use FTP to get files from remote sites.
The client/server relationship is also what makes my job as an Internet trainer quite difficult. While all of the FTP clients out there do the exact same thing, they all use different commands to do it.
Fortunately, if you can understand the basics of what happens during an FTP session, the commands -- no matter what client you are using -- are pretty easy.
The basic steps in an FTP session are:
- Start-up your FTP client.
- Give your FTP client an address to connect to (and, just like TELNET, step one and two can be combined into one easy step).
- Identify yourself to the remote site.
- Give the remote site your password.
- Look around the directory for files.
- Change Directories.
- Set the transfer mode (optional).
- Get the file(s) you want.
- Quit.
The best way to understand what is going on is to look at a sample FTP session. The University of Alabama's connection to the Internet is through SURAnet (a large regional network), so I'm going to FTP to them. Before I do that, though, there are three things that I need to emphasize:
- FTP requires a heck of a lot of resources, both on your system and on the remote system that you are accessing. Because of this, FTP sessions should only be run during non-working hours (usually 6 PM to 6 AM local).
I realize that this constraint is going to be a difficult for those of you who are reading this at work, and who only have Internet (and FTP access) through your employer. However, as responsible Internet citizens we have to remember that the FTP sites are providing us FTP access out of the kindness of their hearts. If we take advantage of this kindness, and access various FTP sites during working hours, those FTP sites may decide to close their doors to the public ... and then EVERYONE loses.
- In light of what was said in #1, please do not flood SURAnet. In MAP15: FTPMAIL I will post a list of FTP sites that you can visit (much like the TELNET list I posted in MAP12: Telnet (Part Two)). Until I post that list, just sit back and enjoy the show :)
- Since this lesson is already over 100 lines long, I hope that you won't mind if I break this FTP lesson into two lessons. The first lesson is in this document, and I'll finish it in the next document. This will mean that we will end this lesson in the middle of our sample FTP session, but it will also mean that this lesson won't be the size of a small book :)
Starting an FTP session is pretty easy. For most of you, all you have to do to start-up your FTP client is type ftp
on you system's command line (or, if you are in a Windows or Mac environment, double-click on the FTP icon). >From there, you would give the client an FTP address to connect to.
Like TELNET, however, there is a way to combine these two steps into one, easy step. For most of you, to access your FTP client and give your client an address to hook up to, all you would have to do it type the command
ftp
and replace with the address of the FTP site that you want your client software to access. In our example, the SURAnet FTP address is ftp.sura.net, so I would type
ftp ftp.sura.net
to start an FTP session. (Note that the second "ftp" is part of SURAnet's FTP address. If I wanted to ftp to info.umd.edu, I would type "ftp info.umd.edu"; if I wanted to ftp to lcs.mit.edu, I would type "ftp lcs.mit.edu"). Once I hit the enter key, the following appears on my screen:
ftp ftp.sura.net
Connecting to ftp.sura.net 128.167.254.179, port 21
220 nic.sura.net FTP server (Version wu-2.4(1) Fri May 20 10:20:58
EDT 1994) ready.
USER (identify yourself to the host):
The second line tells me that my system is connecting to ftp.sura.net (and even gives me the IP number for ftp.sura.net), the third line is some automatic information from SURAnet, and the bottom line is asking me to log in. If I had an account on the SURAnet system, I would enter my SURAnet user ID. But, since I don't have an account on this system, I have to find another way to access the system. ;)
This is where the "anonymous" FTP I mentioned earlier comes in :) The other way to access some FTP sites -- at least those FTP sites that allow outside access -- is to use the userid "anonymous". By using the name "anonymous", you are telling that FTP site that you aren't a regular user of that site, but you would still like to access that FTP site, look around, and retrieve files.
So, where it says USER, I type the word
anonymous
hit enter, and cross my fingers. If SURAnet does not allow anonymous access, I'm about to find out :) >>>USER anonymous
331 Guest login ok, send your complete e-mail address as password.
Password:
COOL! Its going to let me in. All I have to do is give the site a password. Out of politeness to the FTP site, if you login as "anonymous", you need to use your full Internet address as your password. This helps the FTP site keep track of who has visited its site.
So, since it wants my password, and since the password for any anonymous FTP session is my full Internet address, I type
pcrispe1@ua1vm.ua.edu
(Stop laughing -- p-crispy-one is not funny!!). Once I hit enter, my screen fills with the following: >>>PASS ********
230- SURAnet ftp server running wuarchive experimental ftpd
230-
230-Welcome to the SURAnet ftp server. If you have any problems with
230-the server please mail the to systems@sura.net. If you do have problems,
230-please try using a dash (-) as the first character of your password
230- -- this will turn off the continuation messages that may be confusing
230-your ftp client.
...
230 Guest login ok, access restrictions apply.
Command:
Notice the line "Guest login ok, access restrictions apply." This means that the site has given me access, but I only have access to the files that are available to the general public. Okay ... now what? I've started-up my FTP client, I've given the client an FTP address to connect to, I've identified myself to the remote site (I told it that I am anonymous), and I've given the site my password.
Now it's time to see what sort of files and directories are around, and to get those files ... which we will do in the next lesson :)
Posted bySumedh
at
11:49 PM
0
comments
Labels:
Internet
Subscribe to:
Posts (Atom)
Since heading tags typically make text contained in them larger than normal text on the page, this is a
visual cue to users that this text is important and could help them understand something about the
type of content underneath the heading text. Multiple heading sizes used in order create a hierarchical
structure for your content, making it easier for users to navigate through your document.
Google's Search Engine Optimization Starter Guide, Version 1.1, published 13 November 2008
Good practices for heading tags
• Imagine you're writing an outline - Similar to writing an outline for a large paper, put some
thought into what the main points and sub-points of the content on the page will be and
decide where to use heading tags appropriately.
Avoid:
• placing text in heading tags that wouldn't be helpful in defining the structure of
the page
• using heading tags where other tags like and may be more
appropriate
• erratically moving from one heading tag size to another
• Use headings sparingly across the page - Use heading tags where it makes sense. Too
many heading tags on a page can make it hard for users to scan the content and determine
where one topic ends and another begins.
Avoid:
• excessively using heading tags throughout the page
• putting all of the page's text into a heading tag
• using heading tags only for styling text and not presenting structure
Optimize your use of images
Images may seem like a straightforward component of your site, but you can optimize your use of
them. All images can have a distinct filename and "alt" attribute, both of which you should take
advantage of.
The "alt" attribute allows you to specify alternative text for the image if it cannot be displayed for some
reason.
Our alt text here is a brief but accurate description of the image
Why use this attribute? If a user is viewing your site on a browser that doesn't support images, or is
using alternative technologies, such as a screen reader, the contents of the alt attribute provide
information about the picture.
Google's Search Engine Optimization Starter Guide, Version 1.1, published 13 November 2008
Our image wasn't displayed to the user for some reason, but at least the alt text was
Another reason is that if you're using an image as a link, the alt text for that image will be treated
similarly to the anchor text of a text link. However, we don't recommend using too many images for
links in your site's navigation when text links could serve the same purpose. Lastly, optimizing your
image filenames and alt text makes it easier for image search projects like Google Image Search to
better understand your images.
Good practices for images
• Use brief, but descriptive filenames and alt text - Like many of the other parts of the page
targeted for optimization, filenames and alt text (for ASCII languages) are best when they're
short, but descriptive.
Avoid:
• using generic filenames like "image1.jpg", "pic.gif", "1.jpg" when possible
(some sites with thousands of images might consider automating the naming of
images)
• writing extremely lengthy filenames
• stuffing keywords into alt text or copying and pasting entire sentences
• Supply alt text when using images as links - If you do decide to use an image as a link,
filling out its alt text helps Google understand more about the page you're linking to. Imagine
that you're writing anchor text for a text link.
Avoid:
• writing excessively long alt text that would be considered spammy
• using only image links for your site's navigation
• Store images in a directory of their own - Instead of having image files spread out in
numerous directories and subdirectories across your domain, consider consolidating your
images into a single directory (e.g. brandonsbaseballcards.com/images/). This simplifies the
path to your images.
Google's Search Engine Optimization Starter Guide, Version 1.1, published 13 November 2008
• Use commonly supported filetypes - Most browsers support JPEG, GIF, PNG, and BMP
image formats. It's also a good idea to have the extension of your filename match with the
filetype.
Make effective use of robots.txt
A "robots.txt" file tells search engines whether they can access and therefore crawl parts of your
site. This file, which must be named "robots.txt", is placed in the root directory of your site.
The address of our robots.txt file
All compliant search engine bots (denoted by the wildcard * symbol) shouldn't access and
crawl the content under /images/ or any URL whose path begins with /search
You may not want certain pages of your site crawled because they might not be useful to users if
found in a search engine's search results. If you do want to prevent search engines from crawling
your pages, Google Webmaster Tools has a friendly robots.txt generator to help you create this file.
Note that if your site uses subdomains and you wish to have certain pages not crawled on a particular
subdomain, you'll have to create a separate robots.txt file for that subdomain. For more information
on robots.txt, we suggest this Webmaster Help Center guide on using robots.txt files.
There are a handful of other ways to prevent content appearing in search results, such as adding
"NOINDEX" to your robots meta tag, using .htaccess to password protect directories, and using
Google Webmaster Tools to remove content that has already been crawled. Google engineer Matt
Cutts walks through the caveats of each URL blocking method in a helpful video.
Good practices for robots.txt
• Use more secure methods for sensitive content - You shouldn't feel comfortable using
robots.txt to block sensitive or confidential material. One reason is that search engines could
still reference the URLs you block (showing just the URL, no title or snippet) if there happen
to be links to those URLs somewhere on the Internet (like referrer logs). Also, non-compliant
or rogue search engines that don't acknowledge the Robots Exclusion Standard could
disobey the instructions of your robots.txt. Finally, a curious user could examine the
directories or subdirectories in your robots.txt file and guess the URL of the content that you
Google's Search Engine Optimization Starter Guide, Version 1.1, published 13 November 2008
don't want seen. Encrypting the content or password-protecting it with .htaccess are more
secure alternatives.
Avoid:
• allowing search result-like pages to be crawled (users dislike leaving one
search result page and landing on another search result page that doesn't add
significant value for them)
• allowing a large number of auto-generated pages with the same or only slightly
different content to be crawled: "Should these 100,000 near-duplicate pages
really be in a search engine's index?"
• allowing URLs created as a result of proxy services to be crawled
Be aware of rel="nofollow" for links
Setting the value of the "rel" attribute of a link to "nofollow" will tell Google that certain links on your
site shouldn't be followed or pass your page's reputation to the pages linked to. Nofollowing a link is
adding rel="nofollow" inside of the link's anchor tag.
If you link to a site that you don't trust and don't want to pass your site's reputation to, use
nofollow
When would this be useful? If your site has a blog with public commenting turned on, links within
those comments could pass your reputation to pages that you may not be comfortable vouching for.
Blog comment areas on pages are highly susceptible to comment spam. Nofollowing these useradded
links ensures that you're not giving your page's hard-earned reputation to a spammy site. Many
blogging software packages automatically nofollow user comments, but those that don't can most
likely be manually edited to do this. This advice also goes for other areas of your site that may involve
user-generated content, such as guestbooks, forums, shout-boards, referrer listings, etc. If you're
willing to vouch for links added by third parties (e.g. if a commenter is trusted on your site), then
there's no need to use nofollow on links; however, linking to sites that Google considers spammy can
affect the reputation of your own site. The Webmaster Help Center has more tips on avoiding
comment spam, like using CAPTCHAs and turning on comment moderation.
Google's Search Engine Optimization Starter Guide, Version 1.1, published 13 November 2008
A comment spammer leaves a message on one of our news posts, hoping to get some of our
site's reputation
Another use of nofollow is when you're writing content and wish to reference a website, but don't want
to pass your reputation on to it. For example, imagine that you're writing a blog post on the topic of
comment spamming and you want to call out a site that recently comment spammed your blog. You
want to warn others of the site, so you include the link to it in your content; however, you certainly
don't want to give the site some of your reputation from your link. This would be a good time to use
nofollow.
Lastly, if you're interested in nofollowing all of the links on a page, you can use "nofollow" in your
robots meta tag, which is placed inside the tag of that page's HTML. The Webmaster Central
Blog provides a helpful post on using the robots meta tag. This method is written as name="robots" content="nofollow">.
This nofollows all of the links on a page
Google's Search Engine Optimization Starter Guide, Version 1.1, published 13 November 2008
Promote your website in the right ways
While most of the links to your site will be gained gradually, as people discover your content through
search or other ways and link to it, Google understands that you'd like to let others know about the
hard work you've put into your content. Effectively promoting your new content will lead to faster
discovery by those who are interested in the same subject. As with most points covered in this
document, taking these recommendations to an extreme could actually harm the reputation of your
site.
Good practices for promoting your website
• Blog about new content or services - A blog post on your own site letting your visitor base
know that you added something new is a great way to get the word out about new content or
services. Other webmasters who follow your site or RSS feed could pick the story up as well.
• Don't forget about offline promotion - Putting effort into the offline promotion of your
company or site can also be rewarding. For example, if you have a business site, make sure
its URL is listed on your business cards, letterhead, posters, etc. You could also send out
recurring newsletters to clients through the mail letting them know about new content on the
company's website.
• Know about social media sites - Sites built around user interaction and sharing have made
it easier to match interested groups of people up with relevant content.
Avoid:
• attempting to promote each new, small piece of content you create; go for big,
interesting items
• involving your site in schemes where your content is artificially promoted to the
top of these services
• Add your business to Google's Local Business Center - If you run a local business,
adding its information to Google's Local Business Center will help you reach customers on
Google Maps and web search. The Webmaster Help Center has more tips on promoting
your local business.
• Reach out to those in your site's related community - Chances are, there are a number
of sites that cover topic areas similar to yours. Opening up communication with these sites is
usually beneficial. Hot topics in your niche or community could spark additional ideas for
content or building a good community resource.
Avoid:
• spamming link requests out to all sites related to your topic area
• purchasing links from another site with the aim of getting PageRank instead of
traffic
Google's Search Engine Optimization Starter Guide, Version 1.1, published 13 November 2008
Make use of free webmaster tools
Major search engines, including Google, provide free tools for webmasters. Google's Webmaster
Tools help webmasters better control how Google interacts with their websites and get useful
information from Google about their site. Using Webmaster Tools won't help your site get preferential
treatment; however, it can help you identify issues that, if addressed, can help your site perform better
in search results. With the service, webmasters can:
• see which parts of a site Googlebot had problems crawling
• upload an XML Sitemap file
• analyze and generate robots.txt files
• remove URLs already crawled by Googlebot
• specify the preferred domain
• identify issues with title and description meta tags
• understand the top searches used to reach a site
• get a glimpse at how Googlebot sees pages
• remove unwanted sitelinks that Google may use in results
• receive notification of quality guideline violations and file for a site reconsideration
Yahoo! (Yahoo! Site Explorer) and Microsoft (Live Search Webmaster Tools) also offer free tools for
webmasters.
Take advantage of web analytics services
If you've improved the crawling and indexing of your site using Google Webmasters Tools or other
services, you're probably curious about the traffic coming to your site. Web analytics programs like
Google Analytics are a valuable source of insight for this. You can use these to:
• get insight into how users reach and behave on your site
• discover the most popular content on your site
• measure the impact of optimizations you make to your site (e.g. did changing those title and
description meta tags improve traffic from search engines?)
For advanced users, the information an analytics package provides, combined with data from your
server log files, can provide even more comprehensive information about how visitors are interacting
with your documents (such as additional keywords that searchers might use to find your site).
Lastly, Google offers another tool called Google Website Optimizer that allows you to run experiments
to find what on-page changes will produce the best conversion rates with visitors. This, in combination
with Google Analytics and Google Webmaster Tools (see our video on using the "Google Trifecta"), is
a powerful way to begin improving your site.
Google's Search Engine Optimization Starter Guide, Version 1.1, published 13 November 2008
Helpful resources for webmasters
Google Webmaster Help Group - Have questions or feedback on our guide? Let us know
Google Webmaster Central Blog - Frequent posts by Googlers on how to improve your website
Google Webmaster Help Center - Filled with in-depth documentation on webmaster-related issues
Google Webmaster Tools - Optimize how Google interacts with your website
Google Webmaster Guidelines - Design, content, technical, and quality guidelines from Google
Google Analytics - Find the source of your visitors, what they're viewing, and benchmark changes
Google Website Optimizer - Run experiments on your pages to see what will work and what won't
Tips on Hiring an SEO - If you don't want to go at it alone, these tips should help you choose an
SEO company
Except as otherwise noted, the content of this document is licensed under the Creative Commons
Attribution 3.0 License.
Google's Search Engine Optimization Starter Guide, Version 1.1, published 13 November 2008
Posted bySumedh at 8:27 PM 0 comments
WWW 2
Sunday, December 14, 2008
In the last lesson I told you that "URLs" is pronounced "earls." In fact, a lot of people still use the initials and call them "U-R-Ls." I personally prefer calling them "earls" because "earls" rhymes with "squirrels," but the choice of what you call them is completely up to you :)
Let's take a look at how a sample WWW browser works. There are three basic types of WWW browsers available: line-mode browsers, full screen browsers (like Lynx), and graphical browsers (like Mosaic).
Line-mode browsers are about as user un-friendly as you can get. This is hard to describe, but line-mode browsers work a little like FTP inasmuch as you type a command, get some information on your screen, type a new command, get some more information, and so on ...
A full screen browser puts a menu on your screen that looks a little like the Gopher menus that we saw in MAP18: Gopher (Part One). You move the cursor up and down the screen, select a highlighted link, press enter or return, and you are automatically taken to a new document or file (your fill screen browser may work differently than this, though).
Graphical browsers allow you to access not only text, but also pictures and sound (a.k.a. "hypermedia"). In fact, these pictures can be put in Web documents (a.k.a. Web pages), making that Web page look less like a Gopher menu and more like a page from a color magazine! Most graphical browsers use a mouse, and you point-and-click on a highlighted link to access it.
The URL for the sample Web page that we are going to use in this lesson is
- http://ua1vm.ua.edu/~vmhttpd/index.html
The UF WWW Browser for CMS is the browser that my service provider uses, and it is a full screen browser. The browser that you use -- if you can even access a WWW browser -- will probably look and work a little differently than what you will see in this example.
Finally, in real life my browser highlights the links by making them a different color than the rest of the text. There is no way for me to use different colors in this letter, so I have highlighted the links in this example by surrounding them with a (* *).
Just like I can access an item in a Gopher menu by pointing at it and selecting it, I can access a WWW link by pointing at it and selecting it.
Enough talk. Time for the example.
I access my provider's WWW browser, and the following appears on my screen:
Albert 1.2.0 (UF WWW Browser for CMS) Screen 1 of 2 (more screens)COOL!
Viewing=http://ua1vm.ua.edu/~vmhttpd/index.html
Title=UA1VM WWW Home Page
Welcome to The University of Alabama's CMS WWW Server
This CMS server is still under development. Any (*comments*)
or (*suggestions*) will be greatly appreciated. Thank you.
Gopher Sites:
- (*UA1VM CMS Gopher Server*)
- (*UA1IX AIX/370 Gopher Server*)
- (*RISC/6000 Gopher Server*)
- (*RICEVM1 CMS Gopher Server*)
Telnet Sessions:
- (*UA1VM.UA.EDU*)
- (*UA1IX.UA.EDU - Line Mode*)
- (*RISC.UA.EDU - Line Mode*)
WWW Sites:
- The University of Alabama Libraries (*WWW*)
- The University of Illinois at Urbana-Champaign (*WWW*)
- The Alabama Supercomputer Network (*WWW*)
- NASA Information Services via (*WWW*)
Leisure:
- (*Intertext Magazine*) - Electronic Fictional Magazine at The
University of Michigan
- (*Wiretap*) - a gopher to Cupertino, California
- (*NNR*) - UA1VM's Network News Reader
Other Neat Stuff:
- The University of Alabama Library's On-Line (*Card Catalog*)
- a (*map*) of The University of Alabama campus
... snip snip snip ...
I can select any of these links -- the words set apart from the rest of the text with a (* *) -- and be transported to that particular link.
>From this one Web page, I can access Gopher, telnet, and even other Web pages! (I can also access FTP, although this page does not show it).
We've seen a lot of Gopher and telnet recently. Let's take a look at another Web page. Since I used to be a Simulations Director at the United States Space Camp (did I forget to tell you that?), I'm going to move my cursor down to the (*WWW*) link next to "NASA Information Services", press enter, and see what happens:
Albert 1.2.0 (UF WWW Browser for CMS) Screen 1 of 2 (more screens)This is certainly more interesting than SURAnet! ;)
Viewing=http://hypatia.gsfc.nasa.gov/NASA_homepage.html
Title=NASA Information Services via World Wide Web
National Aeronautics and Space Administration
NASA Logo
(*World Wide Web (WWW) information services*)
(*Hot Topics*) NASA news and subjects of public interest
(*NASA Strategic Plan*)
(*NASA Strategies, Policies, and Public Affairs*)
(*NASA Online Educational Resources*)
(*NASA Information Sources by Subject*)
... snip snip snip ...
>From this Web page I can access OTHER Web pages, and from those Web pages I can access even MORE Web pages, and so on ...
Yesterday I told you that it is possible for you to connect directly to a specific Internet resource so long as you know the resource's URL. How you do that depends on the browser that you use.
For the line-mode browser at CERN, for example, the command to connect to a particular URL is
- GO
Before you can do this, however, you have to first access the Web. There are three ways that you can do this:
- Through a browser stored on your local Internet service provider's machine. Ask your provider if your site has a Web browser, and how you can access it.
- Through a browser stored on your own machine. Until recently, you had to have a SLIP or PPP connection to do this. Some providers -- providers who FLOODED my mailbox when I did not talk about the special Level 2.0002746278546723 access that they offer -- now allow you to store a Web browser on your own machine even though you only have Level 2 access.
If you do not have a SLIP or PPP connection, contact your provider BEFORE you store a Web reader on your own computer and double-check that your provider will support the browser. Most service providers can not support a Web browser unless you have a SLIP or PPP connection.
- Through a telnet connection to a publicly-accessible Web browser.
(Do me a favor ... re-read that last sentence. Did you EVER think you would understand a sentence like that? Isn't this workshop COOL?!!)
If you do not have access to a Web browser through your site, you may still be able to access a Web browser through telnet. The following are a few of the public Web sites: (1)
telnet address commentsOnce you are on the Web, it is possible for you to do keyword searches (much like the Veronica searches we did in MAP21: Veronica) using one of the Web's many search engines. One of the best Web search engines is the WebCrawler. The WebCrawler's URL is
-------------- -----------------------
info.cern.ch No password is required. This is in
Switzerland, so U.S. users might be
better off using a closer browser.
www.cc.ukans.edu The "Lynx" full screen browser, which
requires a vt100 terminal. Login as www.
Does not allow users to "go" to arbitrary
URLs.
www.njit.edu Login as www. A full-screen browser at
the New Jersey Institute of Technology.
sun.uakom.cs Slovakia. Has a slow link, so only use
from nearby.
info.funet.fi Login as www. Offers several browsers,
including Lynx (goto option disabled
there too).
fserv.kfki.hu Hungary. Has a slow link, so use from
nearby. Login as www.
- http://www.biotech.washington.edu/WebQuery.html
Another Web search engine you probably will want to check out is the World Wide Web Worm. The Worm's URL is
- http://www.cs.colorado.edu/home/mcbryan/WWWW.html
Both of these search engines provide really good on-line help and instructions.
One last thing, and I am though for this lesson. Please remember:
- The "Web" is the collection of all of the files and information that can be accessed by a Web browser.
- Mosaic and Lynx and just BROWSERS that allow you to access the Web.
Posted bySumedh at 11:51 PM 1 comments
Labels: Internet
WWW
I wish I had six weeks just to talk about the World Wide Web (a.k.a. WWW or "the Web."). If you think Gopher is neat, wait until you start playing around on the Web :)
Unfortunately, I don't have six weeks to talk about the Web -- I only have two lessons. Because of this, we are going to go through the Web like Sherman went through Georgia (1).
That's the bad news. The good news is that there are a lot of REALLY good Web guides available, and I am even seriously considering developing my own Web workshop that I will offer late next year (2).
Until that time comes, however, let's talk about the BASICS of the Web.
In the last few lessons I showed you how most Gopher menus are linked together. We started out in the gopher.squirrel.com root menu, and eventually ended up at the SURAnet gopher menu. We were able to do this because the menus that we traveled through had links to menus and files that were located at other Gopher sites.
Because Gopher menus are linked together, a whole world of information is available to us with just a few keystrokes!
Imagine if we were able to take these links one step further. Instead of linking menus, we could link documents together. You could read one document, find a keyword in that document that really interests you, touch that keyword, and automatically be taken to a NEW document somewhere else in the world -- and this new document could even have links to OTHER documents around the world, and so on.
Sound too good to be true? It isn't, thanks to something called "hypertext." If you have ever played with Apple's HyperCard program or the "help" menus in the latest Microsoft packages, you have already experienced hypertext. You "select" a highlighted word -- usually by clicking on it with a mouse -- and you are taken into an entirely new document or help screen.
The World Wide Web is based on hypertext. It is possible for you to go roaming around the Web, bouncing from document to document, using nothing but the links in those documents!
Just as you can access Gopherspace through a Gopher server or client, you can access the Web through something called a "browser." A browser can read documents, fetch documents, access files by FTP, read Usenet newsgroups, telnet into remote sites, and even travel around Gopherspace. In short, everything that we have talked about from MAP08: Usenet to MAP22:Gophermail can be done using nothing but a Web browser!
The Web is able to accomplish all of this thanks to something called URLs ("earls") -- Universal Resource Locators. URLs list the exact location of any Internet resource.
If you think about it, giving every Internet resource a unique address is the hard part. Once you have given something an address, linking to it is pretty easy :)
What is really special about the Web is that the Web does all of this "behind the scenes." It is possible for you to bounce from one link to another without ever knowing the exact address of where you are, or even how you got there.
If you ever want to jump directly to a particular Internet resource, however, you are going to need to know a little bit more about URLs. Here are a few basic URLs:
file://wuarchive.wustl.edu/mirrors/msdos/graphics/gifkit.zipGee ... those look a little like FTP addresses, don't they?
ftp://wuarchive.wustl.edu/mirrors
http://info.cern.ch:80/default.html
news:alt.hypertext
telnet://dra.com
The first part of an URL -- the stuff before the colon -- tells the browser how to access that particular file. For example, to access
- ftp://wuarchive.wustl.edu/mirrors
Most of the access methods are pretty straight-forward. Here is a list of some of the more common access methods that you are going to see listed in the first part of URLs:
method what it stands forWe've used all of these before, except for "http". If you ever see a URL with "http" at the beginning of it, that means that the file is a hypertext document (with hypertext links to other documents).
------ ---------------------------------------
ftp File Transfer Protocol
file File Transfer Protocol (same as ftp)
news Internet News Protocol (Usenet)
gopher Gopher
telnet Telnet
http Hypertext Transport Protocol
The rest of a URL -- the stuff after the colon -- is the address of that particular file. In general, two slashes (//) after the colon indicates a machine name or address.
For example,
- file://wuarchive.wustl.edu/mirrors/msdos/graphics/gifkit.zip
- http://info.cern.ch:80/default.html
Posted bySumedh at 11:50 PM 0 comments
Labels: Internet
Q & A
- I'm subscribed to the 'squirrel' LISTSERV mailing list (SQUIRREL@UA1VM.UA.EDU). I just heard a story about how an energetic squirrel once crashed the mainframe at Mississippi State University during registration (a true story) and I want to share this information with everyone on the 'squirrel' mailing list. What address should I send my e-mail letter to if I want the letter to be distributed to everyone on the "squirrel" list?
- SQUIRREL@UA1VM.UA.EDU
Letters should be sent to the list address.
- SQUIRREL@UA1VM.UA.EDU
- I'm looking for a list of all known LISTSERV lists. What LISTSERV command should I use to get such a list?
- LISTS GLOBAL
BTW, you can send the lists global command to any LISTSERV address ... it is a universal LISTSERV command.
- LISTS GLOBAL
- Okay, I just got the "list of lists." I found a neat list called "navigate" but the list of lists only shows the address as NAVIGATE@UBVM. What address should I send my subscribe command to?
- LISTSERV@UBVM.BITNET
This was a tough question. You had to remember to convert the LIST@NODE address to LIST@NODE.BITNET. You also had to remember that LISTSERV commands have to be sent to the LISTSERV address. :)
- LISTSERV@UBVM.BITNET
- I want to subscribe to the alt.abuse.recovery newsgroup. I sent a e-mail letter to LISTSERV@UA1VM.UA.EDU with the command SUBSCRIBE ALT.ABUSE.RECOVERY JOE STUDENT, and I got back an error message saying that "alt.abuse.recovery" is unknown to the LISTSERV. What's going on?
- alt.abuse.recovery is a Usenet newsgroup; to subscribe to the group from a Usenet reader.
Remember, if the group's name has periods in it, it probably is a Usenet newsgroup.
- alt.abuse.recovery is a Usenet newsgroup; to subscribe to the group from a Usenet reader.
- How can I unsubscribe from the Roadmap workshop?
- send an e-mail letter to LISTSERV@UA1VM.UA.EDU which says UNSUB * in the body of your letter.
PART TWO: TRUE OR FALSE
- FALSE
There is nothing wrong with giving my best friend my password.(Never give your password to ANYONE!)
- FALSE
WRITING IN ALL CAPS IS CONSIDERED TO BE GOOD NETIQUETTE.(Writing in all caps is ANNOYING.)
- FALSE
The word "squirrel" would be an excellent password.(Squirrel is in the dictionary. NEVER use a password that is in the dictionary.)
- FALSE
Letters to a LISTSERV list should be sent to the LISTSERV address.(Letters should be sent to the list address.)
- FALSE
LISTSERV commands should be sent to the list address.(Commands should be sent to the LISTSERV address.)
PART THREE: SHORT ANSWER
- One of my friends just e-mailed me a chain letter offering me good luck so long as I send the chain letter to ten additional people. Heck, there are several hundred people subscribed to each of the Usenet newsgroups and LISTSERV lists. Why can't I just forward the chain letter to these groups?
Chain letters violate the relevant posting rule (and are likely to turn you into flame-bait).
Also -- and this was not covered in the lessons -- most networks have Acceptable Usage policies that strictly prohibit chain letters.
If someone sends you a chain letter, forward the chain letter (along with an angry reply) to that person's postmaster.
- Patrick's dad said "DON'T include the entire contents of a previous posting in your reply(s)." Why can't I do this?
It is rude. When you re-post the entire contents of a previous posting, you are forcing the readers to reread something that they have already read.
The purpose of reposting is to remind everyone what was said, not to repeat what was said :)
- If someone spams a list that I am on, what should I do? Why shouldn't I just send my angry replies to the list?
Your best bet is to ignore it. If you are really angry, reply to the poster at his e-mail address, or complain to the poster's site administrator.
You should not send your replies to the list because the spammer is probably not even a member or the list, and, if he is, he probably does not care about what you think.
BONUS QUESTIONS
- How many National Football Championships has the University of Alabama's Crimson Tide football team won?
- 12
- Patrick Douglas Crispen is:
- a squirrel
Posted bySumedh at 11:50 PM 0 comments
Labels: Internet
FTP
I hope that you are having as much fun participating in this workshop as I am having teaching it! :)
In the last lesson we used a tool called "Telnet" which allowed us to access remote computers and run programs on those remote computers.
In the next few lessons we are going to use a tool called "File Transfer Protocol" (or "FTP") which will allow us to access remote computers and retrieve files from these computers. Actually, it is probably more accurate to say that we will be using "anonymous" FTP, but I'll explain that in a minute.
What sort of files are available through FTP? Well, "hundreds of systems connected to the Internet have file libraries, or archives, accessible to the public. Much of this consists of free or low-cost {computer} programs for virtually every make of computer. If you want a different communications program for your IBM, or feel like playing a new game on your Amiga, you'll {probably} be able to get it {using FTP}."
"But there are also libraries of documents as well. If you want a copy of a recent U.S. Supreme Court decision, you can {get a copy using FTP}. Copies of historical documents, from the Magna Carta to the Declaration of Independence are also yours for the asking ... You can also find song lyrics, poems, {and} even summaries of every {Animaniacs} episode ever made. You can also find extensive files detailing everything you could possibly want to know about the Net itself". (1)
Before we get too in-depth into FTP, I think that now is as good a time as any to quickly review the client/server relationship that I briefly mentioned in lesson three (MAP03: Levels of Internet Connectivity).
"Client" is just another word for a program running on your service provider's system that enables you to talk to, and get stuff from, distant computers. The client on your service provider's system takes your commands and turns them into requests for information from programs -- "servers" -- running on other, distant computers.
The client/server relationship allows you to TELNET into remote computers and run programs on those remote computers, and it also allows you to use FTP to get files from remote sites.
The client/server relationship is also what makes my job as an Internet trainer quite difficult. While all of the FTP clients out there do the exact same thing, they all use different commands to do it.
Fortunately, if you can understand the basics of what happens during an FTP session, the commands -- no matter what client you are using -- are pretty easy.
The basic steps in an FTP session are:
- Start-up your FTP client.
- Give your FTP client an address to connect to (and, just like TELNET, step one and two can be combined into one easy step).
- Identify yourself to the remote site.
- Give the remote site your password.
- Look around the directory for files.
- Change Directories.
- Set the transfer mode (optional).
- Get the file(s) you want.
- Quit.
Before I do that, though, there are three things that I need to emphasize:
- FTP requires a heck of a lot of resources, both on your system and on the remote system that you are accessing. Because of this, FTP sessions should only be run during non-working hours (usually 6 PM to 6 AM local).
I realize that this constraint is going to be a difficult for those of you who are reading this at work, and who only have Internet (and FTP access) through your employer. However, as responsible Internet citizens we have to remember that the FTP sites are providing us FTP access out of the kindness of their hearts. If we take advantage of this kindness, and access various FTP sites during working hours, those FTP sites may decide to close their doors to the public ... and then EVERYONE loses.
- In light of what was said in #1, please do not flood SURAnet. In MAP15: FTPMAIL I will post a list of FTP sites that you can visit (much like the TELNET list I posted in MAP12: Telnet (Part Two)). Until I post that list, just sit back and enjoy the show :)
- Since this lesson is already over 100 lines long, I hope that you won't mind if I break this FTP lesson into two lessons. The first lesson is in this document, and I'll finish it in the next document. This will mean that we will end this lesson in the middle of our sample FTP session, but it will also mean that this lesson won't be the size of a small book :)
- ftp
>From there, you would give the client an FTP address to connect to.
Like TELNET, however, there is a way to combine these two steps into one, easy step. For most of you, to access your FTP client and give your client an address to hook up to, all you would have to do it type the command
- ftp
In our example, the SURAnet FTP address is ftp.sura.net, so I would type
- ftp ftp.sura.net
Once I hit the enter key, the following appears on my screen:
ftp ftp.sura.netThe second line tells me that my system is connecting to ftp.sura.net (and even gives me the IP number for ftp.sura.net), the third line is some automatic information from SURAnet, and the bottom line is asking me to log in.
Connecting to ftp.sura.net 128.167.254.179, port 21
220 nic.sura.net FTP server (Version wu-2.4(1) Fri May 20 10:20:58
EDT 1994) ready.
USER (identify yourself to the host):
If I had an account on the SURAnet system, I would enter my SURAnet user ID. But, since I don't have an account on this system, I have to find another way to access the system. ;)
This is where the "anonymous" FTP I mentioned earlier comes in :) The other way to access some FTP sites -- at least those FTP sites that allow outside access -- is to use the userid "anonymous". By using the name "anonymous", you are telling that FTP site that you aren't a regular user of that site, but you would still like to access that FTP site, look around, and retrieve files.
So, where it says USER, I type the word
- anonymous
>>>USER anonymousCOOL! Its going to let me in. All I have to do is give the site a password.
331 Guest login ok, send your complete e-mail address as password.
Password:
Out of politeness to the FTP site, if you login as "anonymous", you need to use your full Internet address as your password. This helps the FTP site keep track of who has visited its site.
So, since it wants my password, and since the password for any anonymous FTP session is my full Internet address, I type
- pcrispe1@ua1vm.ua.edu
>>>PASS ********Notice the line "Guest login ok, access restrictions apply." This means that the site has given me access, but I only have access to the files that are available to the general public.
230- SURAnet ftp server running wuarchive experimental ftpd
230-
230-Welcome to the SURAnet ftp server. If you have any problems with
230-the server please mail the to systems@sura.net. If you do have problems,
230-please try using a dash (-) as the first character of your password
230- -- this will turn off the continuation messages that may be confusing
230-your ftp client.
...
230 Guest login ok, access restrictions apply.
Command:
Okay ... now what? I've started-up my FTP client, I've given the client an FTP address to connect to, I've identified myself to the remote site (I told it that I am anonymous), and I've given the site my password.
Now it's time to see what sort of files and directories are around, and to get those files ... which we will do in the next lesson :)
Posted bySumedh at 11:49 PM 0 comments
Labels: Internet