Storage Analysis of VMware View Composer

Tuesday, December 02, 2008 Category : 3

Can I turn 16TB of storage for 1000 VDI users into 619GB, let me show you how it’s actually done. The release today of VMware View Manager 3 brings to market the long anticipated thin provisioning of storage for virtual desktops. Previewed in 2007 as SVI (Scalable Virtual Images) what does this now released View Composer linked clone technology look like under the hood? How much storage will it actually use?

Here is the diagram presented on page 94 of the View Manager Administration Guide (http://www.vmware.com/pdf/viewmanager3_admin_guide.pdf).



This diagram as presented is a conceptual view of the storage. The important logical elements to note here are


  • Parent VM. This is the standard virtual machine you use to create and maintain your various versions of your image. It can have various versions as different snapshots.

  • Replica. The replica is a copy of a specific version of the Parent VM. That is its one of your snapshots states of the Parent. The key thing here is that the disk in the replica is a thin provisioned copy of the parent disk and snapshot.

  • Clones. The diagram shows two clones. The clones are an instance of a replica for a particular VM. For its disk the clone uses the thin provisioned disk in the replica plus its own snapshot to provide the disk. Changes the clone makes to its disk are isolated in the clones snapshot and the replica disk remains untouched and shared by all the clones.


The diagram in the manual is not a great representation, here is my own one that adds some needed detail. Of course there is more complexity, but we can handle that.



What are we looking at here.


  • Each box is a directory in your Datastore. You have one directory for your Parent VM (what I have labeled base image), one for each replica, a special one called a source and one for each clone (which I have labeled user).

  • The Parent VM (base image) in blue is your standard VM. Notice that the C: disk is thick provisioned as is normal in ESX. A 15G disk will consume 15G in your base image. Then you have your snapshots. Notice that the C: and snapshot 0002 have been combined logically, as this is our view of the disk from the VM.

  • Using the Add Desktop wizard in the View Administrator you can create a pool of desktops based on a snapshot from a ParentVM. As part of the process you have to choose a VM and one of its snapshots. When this is done a unique replica is created. This process is marked as (1) on the diagram. Here a copy of the machine is performed, into a new directory however the disk is thin provisioned. If our original disk was 15G yet only 2G was consumed, the disk in the replica will only by 2G. This process can take a short period of time as the data copies, but it is a once off process. This thin provisioned disk is the master disk that all of the clone VMs will use as their base. You can make changes to the parent VM, and the replica can not be harmed.

  • What is not shown in the documentation is that a source directory is also created. This source directory is unique to the replica and contains all of the files required to make a clone. These files are essentially your standard VM files with an empty snapshot of the disk in the replica. It is my thinking that clones are created by making copies of the files in this directory. This is why the cloning process is very very fast, all of the work in the background is mostly done. My testing shows under 60 seconds to deploy a new clone. Again this creation of the source directory is a once off process.

  • The clone (labeled user) directories are created once for each VM in the pool as required/directed by your pool configuration. The directory name is based on the naming convention given at pool creation. Here we have the files required to run the VM instance. The two important files are firstly the snapshot file which is based of the thin provisioned disk in the replica directory. This is where all of the writes for the VM will be stored, so this file will grow over time (until you Recomposition, Refresh or Rebalance the VM). The diagram tries to shows how the C: drive is made of the combination of the thin disk from the replica directory and the local snapshot file in the clone machine. A separate thin provisioned disk is also created for the user D: drive. This is where the quickprep and user data is stored. This user D drive will grow over time as data is put there, it can’t cant be shrunk.


There you have it, the storage layout of View Composer. What does it look like in reality? Here are some screen snippets.


Datastore Directories.

Here is the directories in the Datastore. You can see there is one Parent VM (XP_VMMAST), one replica directory with its matching source directory, and 3 clone directories XPROD 1 through 3.


Parent VM directory files

These are the files in the parent VM, all the usual suspects and the disk is 15G thick provisioned.


Replica directory files

These are the files in the replica directory. Notice the disk has shrunk to only 2G as its thin provisioned and there is our snaptshot which does not really get used.


Source directory files

These are the files in the source directory. They are pristine and clean, ready for use as a clone. Notice the vmdk file, its based on the replica name, a special kind of snapshot.



Clone VM directory files

These are the files in the actual VM clone directory, one directory for each provisioned desktop. Notice that the vmdk file has grown. This is the growth after booting windows the first time and letting it settle, 50M. Notice there are two more files here, one is the user D disk, which is persistent but thin provisioned, its grown to 23M in size. There is also a vswp file as the machine is booted, otherwise if suspended it would be the vmss.

There you have it. For this test of a 15G machine with just over 2.1G used, what would the storage look like for 1000 users. We will leave the user space aside, we need to cater for that either way. We just want to compare the old method with the new View Composer.

Parent VM is 16G.

Replica and source is 2.1G

1000 machines including swap is 600G

Grand storage space for 1000 users is 619GB.

Compare that to one week ago when it was 16TB, that’s some saving. Of course these figures a little extreme, we now have 1000 users running off a single 2.1G thin provisioned disk, its going to need a lot of spindles to deliver the IOPS required.

Exciting times. We are all going to see how View plays out over the next six months. There is some great architectural work to be done in designing for implementation.

Rodos (Rod Haywood)
Senior Consulting Architect - Virtualisation
Alphawest Services, Sydney Australia

Searchable VMware Technical Resource Documents Listing

Category : 0

VMware admins are always looking for quick reference to whitepapers and best practices. How do you do that?

Well there are some great sites which list links such as VMware-land. However the most popular document in VMTN (for non desktop products) is the "VMware Technical Resource Documents Listing" at http://communities.vmware.com/docs/DOC-2590

In a single page you will find the title, abstract and link to the PDF for every VMware published technical note. There are currently 195 documents listed.

The source for the listing is the VMware technical resources document list, however this has no usable search function, hence why this alternate listing was originally created. The excellent thing about the document list page is within seconds you can quick search with your browser for any keyword. Go on give it a try, jump over and search for a keyword that interests you. You may be surprised to find a paper that you never knew existed.

In the 11 months since I created this page it has had over 10,000 hits. This is dwarfed by some of the Fusion documents, one of which is over 33,000. VMware, its probably time to improved the search facility of the technical resource list. It would be great to be able to search within the documents, until then, at least we have a usable alternative.

Rodos

What to buy your VMware geek for Christmas?

Sunday, November 30, 2008 Category : 0

Wondering what to get the special VMware geek in your life for Christmas? Wondering what to put on your Amazon wish list so your mother does not buy you an André Rieu DVD. Then read on for a suggestion ..

VMware Infrastructure 3: Advanced Technical Design Guide and Advanced Operations Guide
by Scott Herold, Ron Oglesby, Mike Laverick

“Every VMware Architect should buy and read this book to assist their understanding of designing for VI3. By helping you understand the important design decisions in building a virtual infrastructure, and importantly showing the pro’s and con’s, you will be better informed as you make your architectural decisions. This book will help you think, rather than prescript an answer and that’s a good thing.”
Rodney Haywood
National Senior Consulting Architect for Virtualisation
Alphawest - VMware Premier and VAC Partner
Sydney, Australia

As a person who has spent a few years earning a living architecting many big and small VMware infrastructures I was keen to I get my hands on a copy of the “Advanced Technical Design Guide” by Scott Herold, Ron Oglesby and Mike Laverick. It had just the words that inspire me; “Advanced” which hopefully meant it had some meat to it and “Technical Design”, so something that might not be marketing or just another manual. I was also keen to validate my own practices and hopefully pick up on some new ideas or improvements.


One of my first activities at VMworld in September was therefore to head to the bookshop and get my hands on a copy. They sold out fast. It has taken me a while to get through the 352 pages, so here is the high level of what I think. In keeping with the books methodology we will have to put the answer into advantages and disadvantages.


Advantages
  • Each of the elements of a virtual infrastructure is explained and the design options detailed. The key element here is that rather than providing wrote answers or saying it depends, the options available are compared, maybe an opinion given and then a list of advantages and disadvantages. This is exactly what the reader needs and it was refreshing to see, especially as it was consistently used throughout the chapters as appropriate. Each organizations requirements and situation is different and this approach helps a designer think through the options and come up with their own answer right for them.
  • It’s written by people who know what they are talking about and have a passion for it. The scattered stories and examples give you a good feel that you are in the hands of people who have trodden this path before you. 
  • It’s great to have a book that is entertaining, as there are comments and geek humor buried in many pages. For example in the chapter on Business Continuity there is discussion on protecting Active Directory Controllers, it states “Unless you only have one Active Directory server in your environment, in which case you should close this book and proceed to beat yourself with it until you pass out, it will always be easier to deploy a new server and follow the proper steps to create a new domain controller and allow the …”. 
  • It covers the topics you need to address; it has breadth combined with sufficient depth.
Disadvantages
  • The books is dated already, being based on prior version of ESX. Often whilst reading you think, but that’s not true anymore, or that has changed. A book of this nature can’t but help suffer from this problem but it does feel like quite a period of time has transpired between writing and publishing. However I don’t think this is a significant problem, which I will speak too in my recommendations further on.
  • The editing is to a poor standard. There is a high frequency of grammar and spelling errors that really should not be there, it gets a bit distracting. The authors are not professional writers, but isn’t that why you have a copy editor?
  • The diagrams could do with some improvement and consistency. Some are barely readable because of the black and white printing, dark backgrounds that most likely looked wonderful on a computer screen have no contrast and look like a dark blob (p121). Other diagrams could be made conceptually a lot clearer, such as the networking ones in Chapter 6, where I certainly had to look twice to understand how port groups were being represented. 
  • I was expecting to see some worked out check lists or outlines for guidance but there are few. There is some details in the chapter on managing your environment but you feel you might want some more, maybe as an appendix. 
If you follow my recommendation to purchase this book, here are some things to consider as you read:
  • Focus on the principles presented and verify any specific features. As the software versions have changed you will need to validate current features and functions with their most recent best practices. So focus on the concepts, the why and the reasoning.
  • Do some subsequent reading on a topic if you are going to make a decision. Quickly skim the relevant component of the VMware documentation, along with some online resources, the best of which will be vendor specific best practice guides, VMTN and the key bloggers.
  • Read with a pencil and mark bits to come back and review. There is no point reading without taking any action. If you mark areas to come back and review you can focus on reading and understanding. Once you have finished, flip through and pick some areas to do some further research on and put some plans it place to update your infrastructure, next design or start a new project to make some improvements.
Now to some more detailed comments on a few areas. Some of these are relevant to you if you are thinking of purchasing a copy, whereas others assume you have read the book.
  • In the Chapter 5 - Storage there is a nice diagram showing the parts of a storage fabric. In referring to the diagram the chapter states “If following the best practices of not only ESX storage configurations, but also the best practices defined for storage infrastructures, your environment should be laid out similarly to the example illustrate in Figure 5 – 1.




    Now when you look at this diagram what jumps out at you in regards to being “best practice”. You got it in one, no redundant paths from the fibre channel switches the storage processors. As a reference on "best practice" in a “Advanced Technical Design” book this should not be so. For details on this topic see the relevant tip write up at the VI Team blog. 
  • On pages 256 and 257 is the best analogy of how shares work that I have certainly ever heard. It involves screaming children and trying to convince your teenage daughter that a life of drunken debauchery might have appeal but it won’t lead to a prosperous career in IT. I won’t spoil it by repeating it here, you can enjoy it yourself when your copy arrives.
  • Disaster Recovery is a topic I spend quite a bit of time in, which is covered in Chapter 10. I was a little surprised to read the following on page 323, “First, it takes a lot of planning and preparation to do right, and second, it’s never cheap when done correctly. DR with ESX server is no different. We have all heard the sales pitch as to how easy DR is with VMware, but is it really easy or cheaper? Maybe it’s easier; however, I am not sure about cheaper.” I disagree. No matter how you play it, DR with VMware is cheaper. You need less hardware full stop. Sure SAN replication is not cheap, you can do expensive DR configurations with VMware like you can in a physical world, but that’s not a comparison of with versus without VMware. 
  • In regards to Disaster Recovery I would have expected to see some reference to VMware licensing for DR sites and some techniques or guidelines, around this as this is a frequently asked question. 
  • Ignore the statement on page 318 that you can not run VCB on the same server as VC, this information is out of date. There is a good tip about having a VCB host for each VMware cluster as clusters also serve as shared storage boundaries. However some of the arguments are no longer valid, such as guaranteeing different LUN IDs if you don’t, because VCB now compares VMFS signatures (or NAA IDs for RDMs).
  • There are only five pages on the topic of Physical to Virtual migrations. I think this is big enough of a topic to deserve more space. Yes, the book is about the VI infrastructure but anyone reading the book is going to have to deal with the large project of how to move 20 or 500 machines which currently exist into this new environment. There is much more that could be said and some methodologies, guidelines and check lists could be very useful to new entrants to the space.
  • The sizing techniques presented through the book are brilliant. This is probably the element that I liked the most as you really get to see and learn from the experience of the authors. Since reading this I have certainly enhanced my sizing analysis methodology to include some of the ideas presented. 
  • It was great (and a small relief) to see so many of the recommendations in my own practice validated. We can’t all live in a depends world, you sometimes need to give a recommendation and just state there are edge cases which may be exceptions. Mixing workloads, server sizing, boot from SAN, Vizioncore backups, network configuration, service console software, avoiding silos, when and when not to use an RDM, LUN sizing; all had good alignment, which is not really that surprising as I suspect most experienced architects will be quite similar on these issues.
  • No mention is made of UPS software or how to shutdown for power outages, this is often a question that people may turn to the book for and could be covered. 
  • I thought the details on host failure calculations for HA left out some details which I consider meet the “advanced” label. I would have liked to at least seen reference to slot calculations. I have had many whiteboard sessions trying to “educate” people about such HA calculations as presented in the book. If VMware improved their documentation this may not be such a problem.
  • Likewise I think the handling of split brain and isolation response is weak. Some good details on advanced configuration options would be welcome here. I also disagree with the conclusion about the “Leave Powered On” isolation response. Unlike the author I have seen and heard of many corruptions due to forced shutdown. It’s not specifically a VMware issue but a guest OS issue, however it is caused by VMware. Many times I have seen someone take out the network switches due to a failed firmware upgrade or the like and cause every ESX host go into split brain mode, resulting in the killing of all of the machines in the cluster. Given the fact that most sites will do as the book recommends and build reliable and redundant networks, the occurrence of a split brain is most likely a major network issue affecting all servers rather than just one, so leaving powered on is in my opinion a better option. If it’s a single host that has failed, your monitoring system should pick it up quickly enough for action to be taken. Of course if the host actually fails you still have the desired result, as the ESX host is off the network and the locks on the shared storage are released, hence the VMs will be restarted as you want. 
Note that you actually get two books in one. You also get the Advanced Operations Guide. In this review I have focused on the Design Guide, which is the first book or half. Rather than repeating all the details of the book, head over to its Amazon page.

There you have it; the final verdict is a buy. I think the fact that I have actually picked it up a few times and quickly read a small section whilst thinking about a design related issue shows the overall usefulness of the book. Get yourself a Christmas present or tell your mum, otherwise you may be stuck with André Rieu!

Its time for an open carrier network

Saturday, November 29, 2008 0

I can remember back 10 years ago, in the days when I was a Internet geek and the Australian government was selling of the national carrier, Telstra. At the time everyone in the industry was surprised and amazed that there was no split between the infrastructure and the retail arms. After all, us tax payers had funded the building of the network, now we were selling it off.

Here is a comment from the [Oz-ISP] mailing list that I was part of at the time.

If, like me, you belive that the only way there can be true competition in the telecommunications industry, is for the cabling and infrastructure to remain in Govt hands and be available for all carriers and CSP's to use then you must make your voice heard TODAY. This could be your last chance. Please contact your local senator and the leaders of the parties in the senate and tell them your views.
Many years later the government is now going to fund a new National Broadband Network (NBN). Again the issue has been raised regarding separation of the infrastructure and access from the services and retail. With a good wholesale model everyone can access and add value.

However yet again Telstra are not playing the game. A good analysis can be found in this article NBN: Government must now enforce open network principles. I think its time for me to write to as well as ring my local member. I am not sure we can rely on our politicians to do the right thing this time around either.

Disclaimer : I have recently started working for the other major Australian carrier who is bidding in a consortium on the NBN, and who is also promoting reasonable wholesale access. I can't help it if they get it.

Virtualise XenApp

Saturday, November 01, 2008 Category : 0

Eventually the world catches up with you. A post by an attendee at the Citrix Summit 2008 reports the recommendation to virtualise your XenApp loads.

Virtualizing XenApp is a concept to get on board with in general, but it’s also logical to assume Citrix can best support and optimize their products when they are used together.

The message is that virtualizing existing and future implementations of XenApp (or Presentation Server) on XenServer can
  • Reduce physical server count
  • Increase availability
  • Increase flexibility
  • Improve performance
I have been recommending doing this under VMware for a while, and often with a lot of push back. Of course one must be sensible about it, architect appropriately and there will always be the edge cases where it does not make sense. However in principle to me there has always been a strong case.

It's good to see that Citrix are now seeing this as good officially too. We can leave the debate as to what's the best hypervisor to run it on aside.

Reminds me of those technical people who use to tell me they would never virtualise a Microsoft Server application such as Exchange or SQL. Those same people are now running it in production happily and love the benefits.

Powered by Blogger.