How to Install VMware Tools on CentOS 6.5

vmwaretoolsVMware Tools is one of important components in order to run virtual machine (VM) with excellent performance. It is a group of utilities that help to enhances the overall performance of the virtual machine’s guest operating system (OS) and improves management of the VM. Without the VMware Tools, guest OS performance will lacks some of the important functionality. Below steps shows how to install the VMware Tools on CentOS 6.5 :

1. Install the prerequisites into your CentOS 6.5.

[root@centos6 ~]# yum install perl gcc make kernel-headers kernel-devel -y

Example.

[root@centos6 ~]# yum install perl gcc make kernel-headers kernel-devel -y
Loaded plugins: fastestmirror
base                                                                         | 3.7 kB     00:00
http://ossm.utm.my/centos/6.5/os/x86_64/repodata/617a880b84a87969e04bfeebb1b96d5f441da42ee931d374fa6b6d57decd821e-primary.sqlite.bz2: [Errno 12] Timeout on http://ossm.utm.my/centos/6.5/os/x86_64/repodata/617a880b84a87969e04bfeebb1b96d5f441da42ee931d374fa6b6d57decd821e-primary.sqlite.bz2: (28, 'Operation too slow. Less than 1 bytes/sec transfered the last 30 seconds')
Trying other mirror.
base/primary_db                                                              | 4.4 MB     00:49
extras                                                                       | 3.4 kB     00:00
extras/primary_db                                                            |  18 kB     00:00
updates                                                                      | 3.4 kB     00:00
updates/primary_db                                                           |  65 kB     00:00
Setting up Install Process
Package 1:make-3.81-20.el6.x86_64 already installed and latest version
Resolving Dependencies
--> Running transaction check
---> Package gcc.x86_64 0:4.4.7-4.el6 will be installed
--> Processing Dependency: libgomp = 4.4.7-4.el6 for package: gcc-4.4.7-4.el6.x86_64
--> Processing Dependency: cpp = 4.4.7-4.el6 for package: gcc-4.4.7-4.el6.x86_64
--> Processing Dependency: glibc-devel >= 2.2.90-12 for package: gcc-4.4.7-4.el6.x86_64
--> Processing Dependency: cloog-ppl >= 0.15 for package: gcc-4.4.7-4.el6.x86_64
--> Processing Dependency: libgomp.so.1()(64bit) for package: gcc-4.4.7-4.el6.x86_64
---> Package kernel-devel.x86_64 0:2.6.32-431.el6 will be installed
---> Package kernel-headers.x86_64 0:2.6.32-431.el6 will be installed
---> Package perl.x86_64 4:5.10.1-136.el6 will be installed
--> Processing Dependency: perl-libs = 4:5.10.1-136.el6 for package: 4:perl-5.10.1-136.el6.x86_64
--> Processing Dependency: perl-libs for package: 4:perl-5.10.1-136.el6.x86_64
--> Processing Dependency: perl(version) for package: 4:perl-5.10.1-136.el6.x86_64
--> Processing Dependency: perl(Pod::Simple) for package: 4:perl-5.10.1-136.el6.x86_64
--> Processing Dependency: perl(Module::Pluggable) for package: 4:perl-5.10.1-136.el6.x86_64
--> Processing Dependency: libperl.so()(64bit) for package: 4:perl-5.10.1-136.el6.x86_64
--> Running transaction check
---> Package cloog-ppl.x86_64 0:0.15.7-1.2.el6 will be installed
--> Processing Dependency: libppl_c.so.2()(64bit) for package: cloog-ppl-0.15.7-1.2.el6.x86_64
--> Processing Dependency: libppl.so.7()(64bit) for package: cloog-ppl-0.15.7-1.2.el6.x86_64
---> Package cpp.x86_64 0:4.4.7-4.el6 will be installed
--> Processing Dependency: libmpfr.so.1()(64bit) for package: cpp-4.4.7-4.el6.x86_64
---> Package glibc-devel.x86_64 0:2.12-1.132.el6 will be installed
--> Processing Dependency: glibc-headers = 2.12-1.132.el6 for package: glibc-devel-2.12-1.132.el6.x86_64
--> Processing Dependency: glibc-headers for package: glibc-devel-2.12-1.132.el6.x86_64
---> Package libgomp.x86_64 0:4.4.7-4.el6 will be installed
---> Package perl-Module-Pluggable.x86_64 1:3.90-136.el6 will be installed
---> Package perl-Pod-Simple.x86_64 1:3.13-136.el6 will be installed
--> Processing Dependency: perl(Pod::Escapes) >= 1.04 for package: 1:perl-Pod-Simple-3.13-136.el6.x86_64
---> Package perl-libs.x86_64 4:5.10.1-136.el6 will be installed
---> Package perl-version.x86_64 3:0.77-136.el6 will be installed
--> Running transaction check
---> Package glibc-headers.x86_64 0:2.12-1.132.el6 will be installed
---> Package mpfr.x86_64 0:2.4.1-6.el6 will be installed
---> Package perl-Pod-Escapes.x86_64 1:1.04-136.el6 will be installed
---> Package ppl.x86_64 0:0.10.2-11.el6 will be installed
--> Finished Dependency Resolution

Dependencies Resolved

====================================================================================================
 Package                         Arch             Version                      Repository      Size
====================================================================================================
Installing:
 gcc                             x86_64           4.4.7-4.el6                  base            10 M
 kernel-devel                    x86_64           2.6.32-431.el6               base           8.7 M
 kernel-headers                  x86_64           2.6.32-431.el6               base           2.8 M
 perl                            x86_64           4:5.10.1-136.el6             base            10 M
Installing for dependencies:
 cloog-ppl                       x86_64           0.15.7-1.2.el6               base            93 k
 cpp                             x86_64           4.4.7-4.el6                  base           3.7 M
 glibc-devel                     x86_64           2.12-1.132.el6               base           978 k
 glibc-headers                   x86_64           2.12-1.132.el6               base           608 k
 libgomp                         x86_64           4.4.7-4.el6                  base           118 k
 mpfr                            x86_64           2.4.1-6.el6                  base           157 k
 perl-Module-Pluggable           x86_64           1:3.90-136.el6               base            40 k
 perl-Pod-Escapes                x86_64           1:1.04-136.el6               base            32 k
 perl-Pod-Simple                 x86_64           1:3.13-136.el6               base           212 k
 perl-libs                       x86_64           4:5.10.1-136.el6             base           578 k
 perl-version                    x86_64           3:0.77-136.el6               base            51 k
 ppl                             x86_64           0.10.2-11.el6                base           1.3 M

Transaction Summary
====================================================================================================
Install      16 Package(s)

Total download size: 40 M
Installed size: 98 M
Downloading Packages:
(1/16): cloog-ppl-0.15.7-1.2.el6.x86_64.rpm                                  |  93 kB     00:00
(2/16): cpp-4.4.7-4.el6.x86_64.rpm                                           | 3.7 MB     00:42
(3/16): gcc-4.4.7-4.el6.x86_64.rpm                                           |  10 MB     01:40
(4/16): glibc-devel-2.12-1.132.el6.x86_64.rpm                                | 978 kB     00:09
(5/16): glibc-headers-2.12-1.132.el6.x86_64.rpm                              | 608 kB     00:06
(6/16): kernel-devel-2.6.32-431.el6.x86_64.rpm                               | 8.7 MB     01:25
(7/16): kernel-headers-2.6.32-431.el6.x86_64.rpm                             | 2.8 MB     00:30
(8/16): libgomp-4.4.7-4.el6.x86_64.rpm                                       | 118 kB     00:01
(9/16): mpfr-2.4.1-6.el6.x86_64.rpm                                          | 157 kB     00:01
(10/16): perl-5.10.1-136.el6.x86_64.rpm                                      |  10 MB     01:35
(11/16): perl-Module-Pluggable-3.90-136.el6.x86_64.rpm                       |  40 kB     00:00
(12/16): perl-Pod-Escapes-1.04-136.el6.x86_64.rpm                            |  32 kB     00:00
(13/16): perl-Pod-Simple-3.13-136.el6.x86_64.rpm                             | 212 kB     00:01
(14/16): perl-libs-5.10.1-136.el6.x86_64.rpm                                 | 578 kB     00:05
(15/16): perl-version-0.77-136.el6.x86_64.rpm                                |  51 kB     00:00
(16/16): ppl-0.10.2-11.el6.x86_64.rpm                                        | 1.3 MB     00:11
----------------------------------------------------------------------------------------------------
Total                                                               103 kB/s |  40 MB     06:35
warning: rpmts_HdrFromFdno: Header V3 RSA/SHA1 Signature, key ID c105b9de: NOKEY
Retrieving key from file:///etc/pki/rpm-gpg/RPM-GPG-KEY-CentOS-6
Importing GPG key 0xC105B9DE:
 Userid : CentOS-6 Key (CentOS 6 Official Signing Key) <centos-6-key@centos.org>
 Package: centos-release-6-5.el6.centos.11.1.x86_64 (@anaconda-CentOS-201311272149.x86_64/6.5)
 From   : /etc/pki/rpm-gpg/RPM-GPG-KEY-CentOS-6
Running rpm_check_debug
Running Transaction Test
Transaction Test Succeeded
Running Transaction
  Installing : 1:perl-Pod-Escapes-1.04-136.el6.x86_64                                          1/16
  Installing : 4:perl-libs-5.10.1-136.el6.x86_64                                               2/16
  Installing : 1:perl-Pod-Simple-3.13-136.el6.x86_64                                           3/16
  Installing : 3:perl-version-0.77-136.el6.x86_64                                              4/16
  Installing : 1:perl-Module-Pluggable-3.90-136.el6.x86_64                                     5/16
  Installing : 4:perl-5.10.1-136.el6.x86_64                                                    6/16
  Installing : ppl-0.10.2-11.el6.x86_64                                                        7/16
  Installing : cloog-ppl-0.15.7-1.2.el6.x86_64                                                 8/16
  Installing : mpfr-2.4.1-6.el6.x86_64                                                         9/16
  Installing : cpp-4.4.7-4.el6.x86_64                                                         10/16
  Installing : libgomp-4.4.7-4.el6.x86_64                                                     11/16
  Installing : kernel-headers-2.6.32-431.el6.x86_64                                           12/16
  Installing : glibc-headers-2.12-1.132.el6.x86_64                                            13/16
  Installing : glibc-devel-2.12-1.132.el6.x86_64                                              14/16
  Installing : gcc-4.4.7-4.el6.x86_64                                                         15/16
  Installing : kernel-devel-2.6.32-431.el6.x86_64                                             16/16
  Verifying  : 1:perl-Module-Pluggable-3.90-136.el6.x86_64                                     1/16
  Verifying  : kernel-headers-2.6.32-431.el6.x86_64                                            2/16
  Verifying  : cpp-4.4.7-4.el6.x86_64                                                          3/16
  Verifying  : glibc-devel-2.12-1.132.el6.x86_64                                               4/16
  Verifying  : 1:perl-Pod-Escapes-1.04-136.el6.x86_64                                          5/16
  Verifying  : libgomp-4.4.7-4.el6.x86_64                                                      6/16
  Verifying  : 4:perl-5.10.1-136.el6.x86_64                                                    7/16
  Verifying  : mpfr-2.4.1-6.el6.x86_64                                                         8/16
  Verifying  : 4:perl-libs-5.10.1-136.el6.x86_64                                               9/16
  Verifying  : 1:perl-Pod-Simple-3.13-136.el6.x86_64                                          10/16
  Verifying  : kernel-devel-2.6.32-431.el6.x86_64                                             11/16
  Verifying  : gcc-4.4.7-4.el6.x86_64                                                         12/16
  Verifying  : 3:perl-version-0.77-136.el6.x86_64                                             13/16
  Verifying  : ppl-0.10.2-11.el6.x86_64                                                       14/16
  Verifying  : cloog-ppl-0.15.7-1.2.el6.x86_64                                                15/16
  Verifying  : glibc-headers-2.12-1.132.el6.x86_64                                            16/16

Installed:
  gcc.x86_64 0:4.4.7-4.el6                          kernel-devel.x86_64 0:2.6.32-431.el6
  kernel-headers.x86_64 0:2.6.32-431.el6            perl.x86_64 4:5.10.1-136.el6

Dependency Installed:
  cloog-ppl.x86_64 0:0.15.7-1.2.el6                   cpp.x86_64 0:4.4.7-4.el6
  glibc-devel.x86_64 0:2.12-1.132.el6                 glibc-headers.x86_64 0:2.12-1.132.el6
  libgomp.x86_64 0:4.4.7-4.el6                        mpfr.x86_64 0:2.4.1-6.el6
  perl-Module-Pluggable.x86_64 1:3.90-136.el6         perl-Pod-Escapes.x86_64 1:1.04-136.el6
  perl-Pod-Simple.x86_64 1:3.13-136.el6               perl-libs.x86_64 4:5.10.1-136.el6
  perl-version.x86_64 3:0.77-136.el6                  ppl.x86_64 0:0.10.2-11.el6

Complete!

2. Attach the vmware tools sofware from vSphere client. Assumed that the host is ESXi 4.1.
1

3. Mount the attached vmware tool software.
[root@centos6 tmp]# mount /dev/cdrom /mnt
[root@centos6 tmp]# cd /mnt
[root@centos6 tmp]# ls
VMwareTools-8.3.7-341836.tar.gz  yum.log

4. Extract the VMware tools

[root@centos6 tmp]# tar xzvf VMwareTools-8.3.7-341836.tar.gz

5. Start install by issue command “./vmware-install.pl”

[root@centos6 tmp]# cd vmware-tools-distrib
[root@centos6 vmware-tools-distrib]# ls
bin  doc  etc  FILES  INSTALL  installer  lib  vmware-install.pl
[root@centos6 vmware-tools-distrib]# ./vmware-install.pl

Full installation as below :

[root@centos6 vmware-tools-distrib]# ./vmware-install.pl
Creating a new VMware Tools installer database using the tar4 format.

Installing VMware Tools.

In which directory do you want to install the binary files?
[/usr/bin]

What is the directory that contains the init directories (rc0.d/ to rc6.d/)?
[/etc/rc.d]

What is the directory that contains the init scripts?
[/etc/rc.d/init.d]

In which directory do you want to install the daemon files?
[/usr/sbin]

In which directory do you want to install the library files?
[/usr/lib/vmware-tools]

The path "/usr/lib/vmware-tools" does not exist currently. This program is
going to create it, including needed parent directories. Is this what you want?
[yes]

In which directory do you want to install the documentation files?
[/usr/share/doc/vmware-tools]

The path "/usr/share/doc/vmware-tools" does not exist currently. This program
is going to create it, including needed parent directories. Is this what you
want? [yes]

The installation of VMware Tools 8.3.7 build-341836 for Linux completed
successfully. You can decide to remove this software from your system at any
time by invoking the following command: "/usr/bin/vmware-uninstall-tools.pl".

Before running VMware Tools for the first time, you need to configure it by
invoking the following command: "/usr/bin/vmware-config-tools.pl". Do you want
this program to invoke the command for you now? [yes]

Initializing...

Stopping VMware Tools services in the virtual machine:
   Guest operating system daemon:                          [  OK  ]
   Virtual Printing daemon:                                [  OK  ]
   Unmounting HGFS shares:                                 [  OK  ]
   Guest filesystem driver:                                [  OK  ]

The module vmmemctl has already been installed on this system by another
installer or package and will not be modified by this installer.  Use the flag
--clobber-kernel-modules=vmmemctl to override.

Found a compatible pre-built module for vmhgfs.  Installing it...

Found a compatible pre-built module for vmxnet.  Installing it...

Found a compatible pre-built module for vmblock.  Installing it...

Found a compatible pre-built module for vmci.  Installing it...

Found a compatible pre-built module for vsock.  Installing it...

The module vmxnet3 has already been installed on this system by another
installer or package and will not be modified by this installer.  Use the flag
--clobber-kernel-modules=vmxnet3 to override.

The module pvscsi has already been installed on this system by another
installer or package and will not be modified by this installer.  Use the flag
--clobber-kernel-modules=pvscsi to override.

No X install found.

Creating a new initrd boot image for the kernel.
   Checking acpi hot plug                                  [  OK  ]
Starting VMware Tools services in the virtual machine:
   Switching to guest configuration:                       [  OK  ]
   VM communication interface:                             [  OK  ]
   VM communication interface socket family:               [  OK  ]
   Guest operating system daemon:                          [  OK  ]
   Virtual Printing daemon:                                [  OK  ]

The configuration of VMware Tools 8.3.7 build-341836 for Linux for this running
kernel completed successfully.

You must restart your X session before any mouse or graphics changes take
effect.

You can now run VMware Tools by invoking the following command:
"/usr/bin/vmware-toolbox" during an X server session.

To enable advanced X features (e.g., guest resolution fit, drag and drop, and
file and text copy/paste), you will need to do one (or more) of the following:
1. Manually start /usr/bin/vmware-user
2. Log out and log back into your desktop session; and,
3. Restart your X session.

Enjoy,

--the VMware team

/sbin/restorecon:  Warning no default label for /tmp/vmware-block-restore0/tmp_file

Linode VPS – Lish SSH Gateway

We’re rolling out a new Lish SSH gateway that simplifies Lish by introducing a single place to access Lish for all of your Linodes, regardless of their host or datacenter. It eliminates the need for per-Linode Lish passwords and SSH keys. Instead, the Lish gateway uses your existing Linode Manager credentials for authentication. You will also notice a new “Lish SSH Keys” field in the My Profile section of the Linode Manager, where you can submit SSH public keys to authenticate yourself to these new Lish gateway boxes.

To provide a little background, Lish is the Linode Shell. It provides you with the ability to issue reboot and shutdown jobs, check the status of your Linode, and most importantly, access and interact with the console of your running Linode. Lish is an out-of-band console, which means you can access it even when your Linode’s networking is disabled.

Previously, each Linode required its own Lish SSH username, password, and SSH keys. Access to Lish was via a direct SSH connection to your Linode’s host machine. In the coming weeks, we will be completely removing host access from the public Internet, and as such the old Lish access methods will no longer function.

Lish-via-SSH into hosts will cease to function on Friday, May 10, 2013 1:00PM EDT.  As such, please adjust any scripts or aliases to utilize the new Lish gateway.

When you log in to the new Lish gateway, you’ll see a list of your Linodes and their locations, as shown below:

$ ssh caker@lish-newark.linode.com 
Linodes located in this data center:
linode2345           Newark, NJ
linode3456           Newark, NJ
linode4567           Newark, NJ

Linodes located in other data centers:
sandbox              Dallas, TX
linode5678           Dallas, TX
[caker@lish-newark.linode.com]#

Then, at the command prompt, you can enter the name of the Linode to which you want to make your Lish connection. In the example shown above, you could enter “linode2345? to access the Lish console for linode2345. Once you’re on a specific Linode, Lish will work like it always has. When you exit linode2345?s Lish, you’ll be taken back to the gateway menu.

You can also do tricks like this to bypass the menu all together:

$ ssh -t caker@lish-newark.linode.com linode2345

And like this to send commands directly to a Linode’s Lish:

$ ssh -t caker@lish-newark.linode.com linode2345 logview

We’ve set up Lish gateways in all six data centers. You can use any gateway to get to any Linode, but we recommend using the one geographically closest to you or your Linodes. Here are the Lish gateway boxes:

  • lish-tokyo.linode.com
  • lish-fremont.linode.com
  • lish-dallas.linode.com
  • lish-atlanta.linode.com
  • lish-newark.linode.com
  • lish-london.linode.com

The Lish gateway boxes are accessible via both IPv4 and IPv6. The Ajax method of connecting to your Linode’s Lish is unaffected by these changes.

Enjoy!
-Chris

Click here for full Story

Linode NextGen: RAM Upgrade

This is the third and final post in a series about Linode: NextGen. The first post in the series focused on network upgrades and the second post focused on host hardware. This post announces yet another upgrade, and discusses the upgrade procedure and availability.

We’re doubling the RAM on all of our plans. This upgrade is available to existing and new customers. New Linodes will automatically be created with the new resources. Existing Linodes will need to go through the Upgrade Queue to receive the upgrades.

The new Linode plans lineup is now the following:

PlanRAMDiskXFERCPUPrice
Linode 1G1 GB24 GB2 TB8 cores (1x priority)$20 / mo
Linode 2G2 GB48 GB4 TB8 cores (2x priority)$40 / mo
Linode 4G4 GB96 GB8 TB8 cores (4x priority)$80 / mo
Linode 8G8 GB192 GB16 TB8 cores (8x priority)$160 / mo
Linode 16G16 GB384 GB20 TB8 cores (16x priority)$320 / mo
Linode 24G24 GB576 GB20 TB8 cores (24x priority)$480 / mo
Linode 32G32 GB768 GB20 TB8 cores (32x priority)$640 / mo
Linode 40G40 GB960 GB20 TB8 cores (40x priority)$800 / mo

Upgrade Queue

Here’s how to get the upgrade for your existing Linode: Log into the Linode Manager and view your Linode’s Dashboard, where you’ll have a new “Upgrade Available” box on the right-hand side. This links to a page describing the upgrade process, which is very simple. Simply click the button and your Linode will enter the Upgrade Queue. While in the queue, your Linode can remain booted.

Once it’s your Linode’s turn in the queue, your Linode will be shut down, upgraded, and migrated to another host. The migration will take about 1 minute per GB of disk images. After the migration has completed, your Linode will be returned to its last state (booted or shutdown) – but with the new RAM!

Full disclosure: the new plans are $0.05 more expensive per month. We did this to get rid of the legacy $19.95, $39.95, $59.95, etc pricing model in favor of a simpler $20, $40, $60 model. The upgrade is not mandatory, so if you’re not down with the 5 cent increase you can keep your existing resources and pricing.

Upgrade Availability

We’ll be enabling the upgrade by data center very soon, with the exception of Fremont which may take another week or two – we’ll be explaining more on Fremont in another post.

Fremont, CA: TBD
Dallas, TX: Upgrades are available
Atlanta, GA: Upgrades are available
Newark, NJ: Upgrades are available
London, UK: Upgrades are available
Tokyo, JP: Upgrades are available

Check back regularly for updates for your data center.

Linode NextGen Recap

This has been a great couple of weeks for Linode and our customers. We’ve spent millions improving our network, a fleet refresh with new hardware and 8 core Linodes, and now this: doubling your RAM without doubling the price. Enjoy!

-Chris

Click here for full Story

Linode Network Upgrades and 8 Cores Xen Instances

Linode Network Upgrade

Early of March 2013, Linode has performed network upgrade that improves throughput, decrease latency and
add redundancy at their host layer. They are upgrading overall network in all six datacenters in the US, Europe, and Asia-Pacific. With zero downtime, Linode customer will get outbound network cap increased 5 times and amount of outbound transfer increased 10 times as below :

Linode 512 upgraded from 200GB to 2000GB (2TB)
Linode 1G upgraded from 400GB to 4000GB (4TB)
Linode 2G upgraded from 800GB to 8000GB (8TB)
Linode 4G upgraded from 1600GB to 16000GB (16TB)
Linode 8G upgraded from 2000GB to 20000GB (20TB)

8 Cores Xen instances

8coreIn 18th March 2013, Linode has upgrading new “NextGen’ host hardware specification, CPUs and a fleet refresh. Linode has upgrades all Linodes to 8 cores Xen instances. Customer will require to schedule a downtime for reboot and then get doubling in computing power. Customer will get faster websites, services, and processing at the same price points. Another things that Linode need to improve is the drive speed. Will they go for SSDs in the future ?

VMware vs. Microsoft: Server Virtualization Showdown Heats Up

VMware and Microsoft are increasingly at each other’s throat these days, and that can mean only one thing — each company sees the other as a real threat when it comes to competing in the server virtualization and private cloud computing markets.

VMware has long dominated the server virtualization technology market, and even though the second iteration of Microsoft’s Hyper-V hypervisor is much better than the first one, VMware still looks like top dog. So while the company has had to suffer the constant sniping of Microsoft, including Microsoft’s risible attempts to ridicule “VM Limited” with its Tad Talks videos, VMware has been content to rise above it all and simply say nothing to this point.

But with Windows Server 2012 and the third iteration of Hyper-V just over the horizon, things are changing. Virtually Speaking – RoundedAt Microsoft’s TechEd in San Francisco this week, Jeff Woolsey, the company’s server virtualization head honcho, boasted that the new Hyper-V will be formidable.

“The guys at VMware claim that they can deliver up to 300,000 IOPS from a single VM. With Windows Server 2012, we’re delivering 985,000 IOPS from a single virtual machine, more than three times [the IOPS claims of VMware],” Woolsey said. “We can go much, much higher, but this is as fast as the hardware will go.”

Sooner or later it was inevitable that VMware would stop ignoring Microsoft and fight back (just as it was inevitable that Microsoft would eventually stop ignoring Apple and fight back in that arena). That’s because VMware’s platform is generally perceived to be much more expensive than Microsoft’s, and once Hyper-V 3 is available there will be fewer and fewer reasons to pay the VMware “tax”.

Click here for full Story