A comprehensive list of some online tutorials to learn using computers and Windows XP

All about Windows7

What is Windows 7?

The upcoming Windows desktop operating system following Windows Vista. It is currently in development at Microsoft.

When will Windows 7 ship?

Originally, a Microsoft VP confirmed that Windows 7 is expected to be released in January 2010, which puts it in the three-year period after the general availability of Windows Vista (which took place in January 2007.) However, more recent news claim to have Windows 7 released 6 months earlier, in June 3, 2009.

Why the name 'Windows 7'?

If the history of naming products at Microsoft has thought us anything, it's that until very soon before the release of the product, we will be dealing with a code-name. This means that Windows 7 is not the final name of the product. The reason behind the name is that Windows Vista is using NT Kernel version 6, and Windows 7 will presumably be labeled with NT Kernel version 7.

Is Windows 7 the same as Windows Vienna?

Yes; Windows 7 was previously named Windows Vienna (hence the name of this website) and before that, Windows Blackcomb. They are the same operating systems only with different names.

Is Windows 7 a major operating system release?

Since "major" is a relative term, and it can mean different things to different people, there is no straight answer to this one. Windows 7 will not break all compatibility with previous applications and hardware supported by Windows Vista for the sake of starting from scratch, despite original reports that claimed so. All the security hardening introduced in Vista will be found in Windows 7. Windows Vista serves as a foundation for upcoming Windows operating systems (such as Windows 7 and the already released Windows Server 2008.)

Are there any distinguished features of Windows 7?

While Microsoft is being careful at releasing details on the features of Windows 7, the released videos and screenshots show an improved Windows Explorer, WinFS storage technology (but under a different name), improved search functions (for the local system, networks and the Internet) and a revamped GUI designed by Julie Larson-Green and other members of the team responsible for the Office 2007 ribbon interface.

Will Windows 7 be released exclusively for 64-bit processors?

No, but it will be the last one to ship for 32-bit processors.

How many people are working on Windows 7?

According to Microsoft, over 2000 developers and 500 managers.

World's biggest scientific experiment LHC (Finding the Earth)

View from the surface during lowering of the first ATLAS small wheel into the tunnel on side C of the cavern. (Claudia Marcelloni, © CERN)

A welder works on the interconnection between two of the LHC's superconducting magnet systems, in the LHC tunnel. (Maximilien Brice, © CERN)

Transporting the ATLAS Magnet Toroid End-Cap A between building 180 to ATLAS point 1. (Claudia Marcelloni, © CERN)

View of the Computer Center during the installation of servers. (Maximilien Brice; Claudia Marcelloni, © CERN)

Aerial view of CERN and the surrounding region of Switzerland and France. Three rings are visible, the smaller (at lower right) shows the underground position of the Proton Synchrotron, the middle ring is the Super Proton Synchrotron (SPS) with a circumference of 7 km and the largest ring (27 km) is that of the former Large Electron and Positron collider (LEP) accelerator with part of Lake Geneva in the background. (© CERN)

View from the surface during lowering of the first ATLAS small wheel into the tunnel on side C of the cavern. (Claudia Marcelloni, © CERN)

View of the LHC cryo-magnet inside the tunnel. (Maximilien Brice, © CERN)

Assembly and installation of the ATLAS Hadronic endcap Liquid Argon Calorimeter. The ATLAS detector contains a series of ever-larger concentric cylinders around the central interaction point where the LHC's proton beams collide. (Roy Langstaff, © CERN)

View of the CMS (Compact Muon Solenoid) experiment Tracker Outer Barrel (TOB) in the cleaning room. The CMS is one of two general-purpose LHC experiments designed to explore the physics of the Terascale, the energy region where physicists believe they will find answers to the central questions at the heart of 21st-century particle physics. (Maximilien Brice, © CERN)

The Globe of Innovation in the morning. The wooden globe is a structure originally built for Switzerland's national exhibition, Expo'02, and is 40 meters wide, 27 meters tall. (Maximilien Brice; Claudia Marcelloni, © CERN)

Assembly and installation of the ATLAS Hadronic endcap Liquid Argon Calorimeter. The ATLAS detector contains a series of ever-larger concentric cylinders around the central interaction point where the LHC's proton beams collide. (Roy Langstaff, © CERN)

The ALICE Inner Tracking System during its transport in the experimental cavern and its insertion into the Time Projection Chamber (TPC). ALICE (A Large Ion Collider Experiment @ CERN) will study the physics of ultrahigh-energy proton-proton and lead-lead collisions and will explore conditions in the first instants of the universe, a few microseconds after the Big Bang. (Maximilien Brice, © CERN)

Google Chrome Privacy Settings and Concerns

Google is considered to be providing the most privacy on the Internet. Google chrome provides Incognito browsing or privacy browsing feature, which protects you from saving your browser history, cookie, etc. in the computer. This is really good.
Internet Explorer 8 beta 2 also offers private browsing feature. Currently I am using Zone Alarm ForceField in my laptop for privacy protection for bank transaction, etc.
At the same time, when opening a new chrome tab, it displays thumb images of the most visited websites, which you may not want to display, especially when your boss opens your browser! Google chrome also do not provide master password protection for saved passwords. These features/bugs do not really protect the privacy of the users.
Incognito or Stealth Browsing
The word Incognito refers to means “without being known or in disguise”. having one’s identity concealed, as under an assumed name, especially to avoid notice or formal attentions.
Incognito mode helps you browse in stealth mode. That means, the web pages you view won’t appear in your browser history or search history, and they won’t leave other traces, like cookies, on your computer after you close the incognito window.
Browsing in incognito mode keeps Google Chrome from storing information about the websites you’ve visited. The websites you visit may still have records of your visit. Any files saved to your computer will still remain on your computer.
This comes handy if you wish to visit your bank or any other financial website, etc.. from a friends laptop or from an Internet browsing center. You can use the incognito mode to make sure that your information is not tracked in the browser for later use.
To turn on the incognito mode, click on the Page menu and select New incognito window. A new window now opens with the incognito icon in the top left corner. You can continue browsing as normal in the other window. You can also right-click any link and select Open link in incognito window.
You can browse normally and in incognito mode at the same time by using separate windows.
Note that, any files you download or bookmarks you create, will be preserved. incognito mode does not protect you from websites that collect or share information about you, internet service providers or employers that track the pages you visit, malicious software that tracks your keystrokes in exchange for free smileys, surveillance by secret agents, etc..
Clear Most Visited Websites in New Tab
You might have noticed that, when you open a new tab in Chrome, the list of most visited websites appears. This list is controlled by the browsing history. If you need to delete just one item from the list, you cannot. Currently Chrome does not support removing just selected items from the list. You will need to clear the entire browsing history. Follow the Google help page to clear personal information.
Chrome is an Open Source, so if any of us have any concern on the privacy settings, any conspiracy, etc., they can inspect the code. To know more about how Google Chrome communicates with Google, read Matt’s blog.
No Master Password for Password Manager
You might remember that Firefox 3 has master password protection for all the saved password. Even if someone else get to your computer and browse using Firefox, you could sit back because your saved passwords are protected by master password.
The current beta release of Google Chrome, does not provide the master password protection. Chrome prompts to save password on successful login into a page. But when anyone can retrieve it by simply click Show Passwords button.
Click on the Tools -> Options -> Minor Tweaks tab ->Show saved passwords will show you the list of websites and usernames. Now you can click on the show password button on the window to see the password for the selected website. Unlike Firefox, the good thing is, Chrome displays only the password for the selected website rather than the whole list. I hope Google will have the master password in the next release

Google Pagerank Algorithm and How It Works

Page Rank is a topic much discussed by Search Engine Optimisation (SEO) experts. At the heart of PageRank is a mathematical formula that seems scary to look at but is actually fairly simple to understand.
Despite this many people seem to get it wrong! In particular “Chris Ridings of www.searchenginesystems.net” has written a paper entitled “PageRank Explained: Everything you've always wanted to know about PageRank”, pointed to by many people, that contains a fundamental mistake early on in the explanation! Unfortunately this means some of the recommendations in the paper are not quite accurate.
By showing code to correctly calculate real PageRank I hope to achieve several things in this response:
Clearly explain how PageRank is calculated.
Go through every example in Chris' paper, and add some more of my own, showing the correct PageRank for each diagram. By showing the code used to calculate each diagram I've opened myself up to peer review - mostly in an effort to make sure the examples are correct, but also because the code can help explain the PageRank calculations.
Describe some principles and observations on website design based on these correctly calculated examples.
Any good web designer should take the time to fully understand how PageRank really works - if you don't then your site's layout could be seriously hurting your Google listings!
[Note: I have nothing in particular against Chris. If I find any other papers on the subject I'll try to comment evenly]

How is PageRank Used?

PageRank is one of the methods Google uses to determine a page's relevance or importance. It is only one part of the story when it comes to the Google listing, but the other aspects are discussed elsewhere (and are ever changing) and PageRank is interesting enough to deserve a paper of its own.
PageRank is also displayed on the toolbar of your browser if you've installed the Google toolbar ( http://toolbar.google.com/ ). But the Toolbar PageRank only goes from 0 – 10 and seems to be something like a logarithmic scale:
Toolbar PageRank (log base 10)
Real PageRank

0 0 - 10
1 100 - 1,000
2 1,000 - 10,000
3 10,000 - 100,000
4 and so on...

We can't know the exact details of the scale because, as we'll see later, the maximum PR of all pages on the web changes every month when Google does its re-indexing! If we presume the scale is logarithmic (although there is only anecdotal evidence for this at the time of writing) then Google could simply give the highest actual PR page a toolbar PR of 10 and scale the rest appropriately.
Also the toolbar sometimes guesses! The toolbar often shows me a Toolbar PR for pages I've only just uploaded and cannot possibly be in the index yet!
What seems to be happening is that the toolbar looks at the URL of the page the browser is displaying and strips off everything down the last “/” (i.e. it goes to the “parent” page in URL terms). If Google has a Toolbar PR for that parent then it subtracts 1 and shows that as the Toolbar PR for this page. If there's no PR for the parent it goes to the parent's parent's page, but subtracting 2, and so on all the way up to the root of your site. If it can't find a Toolbar PR to display in this way, that is if it doesn't find a page with a real calculated PR, then the bar is greyed out.
Note that if the Toolbar is guessing in this way, the Actual PR of the page is 0 - though its PR will be calculated shortly after the Google spider first sees it.
PageRank says nothing about the content or size of a page, the language it's written in, or the text used in the anchor of a link!


I've started to use some technical terms and shorthand in this paper. Now's as good a time as any to define all the terms I'll use:
Shorthand for PageRank: the actual, real, page rank for each page as calculated by Google. As we'll see later this can range from 0.15 to billions.
Toolbar PR:
The PageRank displayed in the Google toolbar in your browser. This ranges from 0 to 10.
If page A links out to page B, then page B is said to have a “backlink” from page A.
That's enough of that, let's get back to the meat…
So what is PageRank?

In short PageRank is a “vote”, by all the other pages on the Web, about how important a page is. A link to a page counts as a vote of support. If there's no link there's no support (but it's an abstention from voting rather than a vote against the page).
Quoting from the original Google paper, PageRank is defined like this:
We assume page A has pages T1...Tn which point to it (i.e., are citations). The parameter d is a damping factor which can be set between 0 and 1. We usually set d to 0.85. There are more details about d in the next section. Also C(A) is defined as the number of links going out of page A. The PageRank of a page A is given as follows:
PR(A) = (1-d) + d (PR(T1)/C(T1) + ... + PR(Tn)/C(Tn))
Note that the PageRanks form a probability distribution over web pages, so the sum of all web pages' PageRanks will be one.
PageRank or PR(A) can be calculated using a simple iterative algorithm, and corresponds to the principal eigenvector of the normalized link matrix of the web.
but that's not too helpful so let's break it down into sections.
PR(Tn) - Each page has a notion of its own self-importance. That's “PR(T1)” for the first page in the web all the way up to “PR(Tn)” for the last page
C(Tn) - Each page spreads its vote out evenly amongst all of it's outgoing links. The count, or number, of outgoing links for page 1 is “C(T1)”, “C(Tn)” for page n, and so on for all pages.
PR(Tn)/C(Tn) - so if our page (page A) has a backlink from page “n” the share of the vote page A will get is “PR(Tn)/C(Tn)”
d(... - All these fractions of votes are added together but, to stop the other pages having too much influence, this total vote is “damped down” by multiplying it by 0.85 (the factor “d”)
(1 - d) - The (1 – d) bit at the beginning is a bit of probability math magic so the “ sum of all web pages' PageRanks will be one ”: it adds in the bit lost by the d(... . It also means that if a page has no links to it (no backlinks) even then it will still get a small PR of 0.15 (i.e. 1 – 0.85). (Aside: the Google paper says “the sum of all pages” but they mean the “the normalised sum” – otherwise known as “the average” to you and me.

How is PageRank Calculated?

This is where it gets tricky. The PR of each page depends on the PR of the pages pointing to it. But we won't know what PR those pages have until the pages pointing to them have their PR calculated and so on… And when you consider that page links can form circles it seems impossible to do this calculation!
But actually it's not that bad. Remember this bit of the Google paper:
PageRank or PR(A) can be calculated using a simple iterative algorithm, and corresponds to the principal eigenvector of the normalized link matrix of the web.
What that means to us is that we can just go ahead and calculate a page's PR without knowing the final value of the PR of the other pages . That seems strange but, basically, each time we run the calculation we're getting a closer estimate of the final value. So all we need to do is remember the each value we calculate and repeat the calculations lots of times until the numbers stop changing much.
Lets take the simplest example network: two pages, each pointing to the other:
Each page has one outgoing link (the outgoing count is 1, i.e. C(A) = 1 and C(B) = 1).
Guess 1
We don't know what their PR should be to begin with, so let's take a guess at 1.0 and do some calculations:
d = 0.85
PR(A) = (1 – d) + d(PR(B)/1)
PR(B) = (1 – d) + d(PR(A)/1)
PR(A) = 0.15 + 0.85 * 1 = 1
PR(B) = 0.15 + 0.85 * 1 = 1
Hmm, the numbers aren't changing at all! So it looks like we started out with a lucky guess!!!
Guess 2
No, that's too easy, maybe I got it wrong (and it wouldn't be the first time). Ok, let's start the guess at 0 instead and re-calculate:
PR(A) = 0.15 + 0.85 * 0 = 0.15

PR(B) = 0.15 + 0.85 * 0.15 = 0.2775
NB. we've already calculated a “next best guess” at PR(A) so we use it here
And again:
PR(A) = 0.15 + 0.85 * 0.2775 = 0.385875
PR(B) = 0.15 + 0.85 * 0.385875 = 0.47799375
And again
PR(A) = 0.15 + 0.85 * 0.47799375 = 0.5562946875
PR(B) = 0.15 + 0.85 * 0.5562946875 = 0.622850484375
and so on. The numbers just keep going up. But will the numbers stop increasing when they get to 1.0? What if a calculation over-shoots and goes above 1.0?
Guess 3
Well let's see. Let's start the guess at 40 each and do a few cycles:
PR(A) = 40 PR(B) = 40
First calculation
PR(A) = 0.15 + 0.85 * 40 = 34.25
PR(B) = 0.15 + 0.85 * 0.385875 = 29.1775
And again
PR(A) = 0.15 + 0.85 * 29.1775 = 24.950875
PR(B) = 0.15 + 0.85 * 24.950875 = 21.35824375
Yup, those numbers are heading down alright! It sure looks the numbers will get to 1.0 and stop
Here's the code used to calculate this example starting the guess at 0: Show the code Run the program
Principle: it doesn't matter where you start your guess, once the PageRank calculations have settled down, the “ normalized probability distribution ” (the average PageRank for all pages) will be 1.0

Getting the answer quicker

How many times do we need to repeat the calculation for big networks? That's a difficult question; for a network as large as the World Wide Web it can be many millions of iterations! The “damping factor” is quite subtle. If it's too high then it takes ages for the numbers to settle, if it's too low then you get repeated over-shoot, both above and below the average - the numbers just swing about the average like a pendulum and never settle down.
Also choosing the order of calculations can help. The answer will always come out the same no matter which order you choose, but some orders will get you there quicker than others.
I'm sure there's been several Master's Thesis on how to make this calculation as efficient as possible, but, in the examples below, I've used very simple code for clarity and roughly 20 to 40 iterations were needed!
Example 1
This is the first example Chris used in his paper.
I'm not going to repeat the calculations here, but you can see them by running the program (yes, if you click the link the program really is re-run to do the calculations for you)
Show the code Run the program

Site Maps

Site maps are useful in at least two ways:
If a user types in a bad URL most websites return a really unhelpful “404 – page not found” error page. This can be discouraging. Why not configure your server to return a page that shows an error has been made, but also gives the site map? This can help the user enormously
Linking to a site map on each page increases the number of internal links in the site, spreading the PR out and protecting you against your vote “donations”

Yup, those spam pages are pretty worthless but they sure add up!
Observation : it doesn't matter how many pages you have in your site, your average PR will always be 1.0 at best. But a hierarchical layout can strongly concentrate votes, and therefore the PR, into the home page!
This is a technique used by some disreputable sites (mostly adult content sites). But I can't advise this - if Google's robots decide you're doing this there's a good chance you'll be banned from Google! Disaster!
On the other hand there are at least two right ways to do this:
1. Be a Mega-site
Mega-sites, like http://news.bbc.co.uk/ have tens or hundreds of editors writing new content – i.e. new pages - all day long! Each one of those pages has rich, worthwile content of its own and a link back to its parent or the home page! That's why the Home page Toolbar PR of these sites is 9/10 and the rest of us just get pushed lower and lower by comparison…
Principle : Content Is King! There really is no substitute for lots of good content…
2. Give away something useful
http://www.phpbb.com/ has a Toolbar PR of 8/10 (at the time of writing) and it has no big money or marketing behind it! How can this be?
What the group has done is write a very useful bulletin board system that is becoming very popular on many websites. And at the bottom of every page, in every installation, is this HTML code:
Powered by phpBB
The administrator of each installation can remove that link, but most don't because they want to return the favour…
Can you imagine all those millions of pages giving a fraction of a vote to http://www.phpbb.com/ ? Wow!
Principle : Make it worth other people's while to use your content or tools. If your give-away is good enough other site admins will gladly give you a link back.
Principle : it's probably better to get lots (perhaps thousands) of links from sites with small PR than to spend any time or money desperately trying to get just the one link from a high PR page.

A Discussion on Averages

From the Brin and Page paper, the average Actual PR of all pages in the index is 1.0!
So if you add pages to a site you're building the total PR will go up by 1.0 for each page (but only if you link the pages together so the equation can work), but the average will remain the same.
If you want to concentrate the PR into one, or a few, pages then hierarchical linking will do that. If you want to average out the PR amongst the pages then "fully meshing" the site (lots of evenly distributed links) will do that - examples 5, 6, and 7 in my above. (NB. this is where Ridings' goes wrong, in his MiniRank model feedback loops will increase PR - indefinitely!)
Getting inbound links to your site is the only way to increase your site's average PR. How that PR is distributed amongst the pages on your site depends on the details of your internal linking and which of your pages are linked to.
If you give outbound links to other sites then your site's average PR will decrease (you're not keeping your vote "in house" as it were). Again the details of the decrease will depend on the details of the linking.
Given that the average of every page is 1.0 we can see that for every site that has an actual ranking in the millions (and there are some!) there must be lots and lots of sites who's Actual PR is below 1.0 (particularly because the absolute lowest Actual PR available is (1 - d)).
It may be that the Toolbar PR 1,2 correspond to Actual PR's lower than 1.0! E.g. the logbase for the Toolbar may be 10 but the Actual PR sequence could start quite low: 0.01, 0.1, 1, 10, 100, 1,000 etc...


PageRank is, in fact, very simple (apart from one scary looking formula). But when a simple calculation is applied hundreds (or billions) of times over the results can seem complicated.
PageRank is also only part of the story about what results get displayed high up in a Google listing. For example there's some evidence to suggest that Google is paying a lot of attention these days to the text in a link's anchor when deciding the relevance of a target page – perhaps more so than the page's PR…
PageRank is still part of the listings story though, so it's worth your while as a good designer to make sure you understand it correctly.


The original PageRank paper by Google's founders Sergey Brin and Lawrence Page - http://www-db.stanford.edu/~backrub/google.html
Chris Ridings' “PageRank Explained” paper which, as of April 2002 http://web.archive.org/web/*/http://www.goodlookingcooking.co.uk/PageRank.pdf , contains one major mistake/misunderstanding - http://www.goodlookingcooking.co.uk/PageRank.pdf
Phil Craven's PageRank Calculator (fortunately his figures agree with mine)
A detailed explanation of how Chris incorrectly altered the PageRank equation with his MiniRank model
An excellent discussion on chad-jams (including “pregnant chad”) by Douglas W. Jones - http://www.cs.uiowa.edu/~jones/cards/chad.html - I don't think many people know the United States' voting system is this flawed!!!
Discussion forums on this topic:
MarketPositionTalk - PageRank updates
SearchEngineForums - PR documents and calculator
WebmasterWorld - PR document and calculator

This Day in History

Today's Birthday

In the News

Quote of the Day