<?xml version="1.0" encoding="UTF-8"?><rss version="2.0"
	xmlns:content="http://purl.org/rss/1.0/modules/content/"
	xmlns:wfw="http://wellformedweb.org/CommentAPI/"
	xmlns:dc="http://purl.org/dc/elements/1.1/"
	xmlns:atom="http://www.w3.org/2005/Atom"
	xmlns:sy="http://purl.org/rss/1.0/modules/syndication/"
	xmlns:slash="http://purl.org/rss/1.0/modules/slash/"
	>

<channel>
	<title>Andrew Morgan &#187; VMware View</title>
	<atom:link href="http://andrewmorgan.ie/category/vmware-view/feed/" rel="self" type="application/rss+xml" />
	<link>http://andrewmorgan.ie</link>
	<description>Grumpy ramblings</description>
	<lastBuildDate>Fri, 30 Jun 2017 09:24:25 +0000</lastBuildDate>
	<language>en-US</language>
		<sy:updatePeriod>hourly</sy:updatePeriod>
		<sy:updateFrequency>1</sy:updateFrequency>
	<generator>http://wordpress.org/?v=4.0</generator>
	<item>
		<title>ThinIO facts and figures, Part 4: Storage design and dangerous assumptions.   </title>
		<link>http://andrewmorgan.ie/2014/10/thinio-facts-and-figures-part-4-storage-design-and-dangerous-assumptions/</link>
		<comments>http://andrewmorgan.ie/2014/10/thinio-facts-and-figures-part-4-storage-design-and-dangerous-assumptions/#comments</comments>
		<pubDate>Thu, 30 Oct 2014 20:53:13 +0000</pubDate>
		<dc:creator><![CDATA[andyjmorgan]]></dc:creator>
				<category><![CDATA[Remote Desktop Services (RDS)]]></category>
		<category><![CDATA[Server Based Computing]]></category>
		<category><![CDATA[ThinIO]]></category>
		<category><![CDATA[ThinScale]]></category>
		<category><![CDATA[VDI in a Box]]></category>
		<category><![CDATA[Virtual Desktop Infrastructure]]></category>
		<category><![CDATA[VMware View]]></category>
		<category><![CDATA[Windows]]></category>
		<category><![CDATA[XenApp]]></category>
		<category><![CDATA[XenDesktop]]></category>
		<category><![CDATA[End User Computing]]></category>
		<category><![CDATA[IOPS]]></category>
		<category><![CDATA[SBC]]></category>
		<category><![CDATA[VDI]]></category>
		<category><![CDATA[VMware Horizon]]></category>
		<category><![CDATA[xenapp]]></category>

		<guid isPermaLink="false">http://andrewmorgan.ie/?p=3222</guid>
		<description><![CDATA[Welcome back to this blog series discussing our new product ThinIO. Please find the below three earlier articles in this series: ThinIO facts and figures, Part 1: VDI and Ram caching. ThinIO facts and figures, Part 2: The Bootstorm chestnut. ThinIO facts and figures, Part 3: RDS and Ram caching. In the final blog post in this series, we’re going to discuss storage design and a frequent problem face when sizing storage. Lets get right into it: “Designing for average, [&#8230;]]]></description>
				<content:encoded><![CDATA[<p><img class="alignright" src="http://andrewmorgan.ie/wp-content/uploads/2014/04/logo.png" alt="" width="216" height="41" />Welcome back to this blog series discussing our new product ThinIO. Please find the below three earlier articles in this series:</p>
<ul>
<li><a href="http://andrewmorgan.ie/2014/10/thinio-facts-and-figures-part-1-vdi-and-ram-caching/">ThinIO facts and figures, Part 1: VDI and Ram caching.</a></li>
<li><a href="http://andrewmorgan.ie/2014/10/thinio-facts-and-figures-part-2-the-bootstorm-chestnut/">ThinIO facts and figures, Part 2: The Bootstorm chestnut.</a></li>
<li><a href="http://andrewmorgan.ie/2014/10/thinio-facts-and-figures-part-3-rds-and-ram-caching/">ThinIO facts and figures, Part 3: RDS and Ram caching.</a></li>
</ul>
<p>In the final blog post in this series, we’re going to discuss storage design and a frequent problem face when sizing storage. Lets get right into it:</p>
<p><span id="more-3222"></span></p>
<h3><strong>“Designing for average, is designing for failure”</strong></h3>
<p><a href="http://andrewmorgan.ie/wp-content/uploads/2014/10/image0011.png"><img class="aligncenter size-large wp-image-3223" src="http://andrewmorgan.ie/wp-content/uploads/2014/10/image0011-1024x626.png" alt="image001" width="625" height="382" /></a></p>
<p style="text-align: center;"><em>Peak IOPS:1015, Average IOPS: 78</em></p>
<p>A frequent mistake I see customers and consultants alike make is taking an average of a sizing requirement and using that as a baseline for sizing environments.</p>
<p>Looking at the figures produced from our internal load tests, we saw just an average of 78 IOPS required on write from a vanilla server without ThinIO to provide storage from this XenApp server.</p>
<p>Now frequently, people will take a figure like that, throw in 25% for growth and bob’s your uncle, ‘order the hardware Mr SAN man’. When I have questioned them about that assumption, they’ll often respond “oh it will be a bit slow if theres contention but it’ll even itself out”.</p>
<p><strong>Right? Wrong. </strong></p>
<p><strong>Things don’t go slow when they are over subscribed, they stop.</strong></p>
<p>Don’t take my word for it! Lets do some simple theoretical math:<img class="alignright  wp-image-3227" src="http://andrewmorgan.ie/wp-content/uploads/2014/10/calculator-150x150.png" alt="calculator" width="98" height="98" /></p>
<p>If you take a storage device and allocate 100 IOPS to this machine, what’s going to happen when a peak like 1000 IOPS is requested? <strong>A lot of queuing.</strong><strong> </strong></p>
<p>In theory, keeping to the 100 IOPS figure, that 1 second burst IO is now taking over 10 seconds to satisfy (1000 / 10).</p>
<p>But it gets worse, all subsequent IO that is requested after that spike occurred is going to also be haulted waiting for this task to occur.</p>
<p>Assuming you’re now mid spike and 10 seconds later the request is finished&#8230; taking your average figure, you now have 10 seconds worth of 100 IO’s per second potentially queued up behind…</p>
<p>Low and behold another login occurs and? <strong>STOP.</strong> Storage timeouts, twirly whirlies, application crashes, hour glasses and the good old <strong>“logins are slow”.</strong></p>
<h2><strong>Oh ok, So how do I size to accommodate this?</strong><strong> </strong></h2>
<p>Well you’re between a rock and a hard place aren’t you. You can’t tell users when to login, the price tag of a SAN sized for peak activity + 20% is going to cost you more than your entire desktop estate. And as you can see, it’s never safe to assume it will run slow.<strong> </strong></p>
<h3><strong>Buying into shared storage is a tricky business</strong><strong> <img class="alignright  wp-image-3228" src="http://andrewmorgan.ie/wp-content/uploads/2014/10/119460-117532-150x150.jpg" alt="119460-117532" width="115" height="115" /></strong></h3>
<p>Storage is expensive. Very expensive. It always annoys me when you hear vendors in this space refering to themselves as ‘reassuringly expensive’. To me this directly translates to ‘We can charge what we want, so we will and you can be reassured we feel the price tag is worth it.’</p>
<p>Storage was never written with desktop workloads in mind, it was written for ‘steady state’ server workloads and was in the progress of going the way of the ‘dodo’ (extinct) up until that first release of vMotion requiring shared storage, which some say was the saving of the market.</p>
<p>Many vendors are going with software or hardware intelligent tiering. This is a great feature, but the real question to ask is how frequently data is moved from the hot tier to lower tier? Press your vendor on this as they more than likely wont know! Microsoft storage spaces is a prime example of this with a really poor optimisation process of just once a day!</p>
<p>Then ask yourself what happens when a base image update occurs and changes the disk layout of the base golden image? Further, stateless technologies from the bigger vendors delete the differencing disk on restart, can you be sure the new disk is going to end up in the smaller, faster SSD or RAM tier? Or is data up there already in contention?</p>
<h3><strong>RAM is far less tricky<img class="alignright size-thumbnail wp-image-3224" src="http://andrewmorgan.ie/wp-content/uploads/2014/10/ram_icon-1-150x150.jpg" alt="ram_icon (1)" width="150" height="150" /></strong></h3>
<p>RAM is commodity, available in abundance and throughout every virtual desktop project I’ve architected and deployed, you run out of CPU resources in a ‘fully loaded’ host way before you will run out of RAM. RAM has no running maintenance cost, Ram is an upfront CAPEX cost and requires little to no maintenance.</p>
<p>The beauty of what ThinIO does with the little resources you assign it, is turn that <strong>desktop workload</strong> into a healthier and happier <strong>server workload</strong> of minimal burst IO and a low steady state IO requirement.</p>
<p><a href="http://andrewmorgan.ie/wp-content/uploads/2014/10/image0033.png"><img class="alignright size-large wp-image-3225" src="http://andrewmorgan.ie/wp-content/uploads/2014/10/image0033-1024x626.png" alt="image003" width="625" height="382" /></a></p>
<p style="text-align: center;"><em>Note the peak of just 40.5 IOPS and average IOPS of less than 2.</em></p>
<p>with as little as just 200mb cache for each of the 10 users logging in, within an aggressive 3 minute window, we reduced the peak from 1000 to 40. That’s a <strong>96% reduction</strong> in burst IO.</p>
<h3><strong>With ThinIO, you:</strong></h3>
<ul>
<li>reduce your exposure to massive IO spikes.</li>
<li>Improve user logon times.</li>
<li>significantly reduce your daily IOPS run rate.</li>
<li>Increase user productivity by spending less time waiting for the storage.</li>
<li>Commit to nothing up front, test it and see how well it works. If you are happy, then buy in.</li>
</ul>
<h3><strong>Lots of Intelligence baked in:</strong></h3>
<p>ThinIO is acutely aware of key operating system events that will cause these kind of spikes and react accordingly to reduce the spikes in IOPS created. ThinIO constantly watches the behavior and IO pattern of the storage and tunes itself accordingly.</p>
<p><a href="http://andrewmorgan.ie/wp-content/uploads/2014/10/image0053.png"><img class="aligncenter wp-image-3229 size-full" src="http://andrewmorgan.ie/wp-content/uploads/2014/10/image0053.png" alt="image005" width="1024" height="468" /></a></p>
<p>Unlike other technologies, ThinIO is a true caching and performance solution. We do not move useful data in and out of the cache on demand when cache availability is contencious. We track patterns and frequency of block access to respond accordingly, delivering all the benefits we have mentioned, even with the tiniest cache, without EVER reducing the capability of the storage when overwhelmed.</p>
<p>And on the opposite side of the scale, when underworked, we leverage our cache to deliver deeper read savings as above.</p>
<p>ThinIO also has a powerful API and PowerShell interface to allow you to report and interact with the cache on demand.</p>
<h2><strong>Wrap up:</strong></h2>
<p>And with the end of the series looming, allow me to finish on some easy points:</p>
<p><strong> ThinIO Allows you to:</strong></p>
<ul>
<li>size your SAN outside of the Lamborghini category &amp; price tag for your desktop estate.</li>
<li>rapidly achieve far deeper density on your current hardware when you are feeling the Pinch.</li>
<li>guarantee a level of performance by assigning cache per VM, disallowing other users to steal or hamper caching resources.</li>
<li>Improve user experience and login times immediately.</li>
<li>Reduce the impact of boot storms and similar IO storm scenarios.</li>
</ul>
<p>No other vendor can offer as quick a turn around time with their product. ThinIO installs in seconds and offers a huge range of compatibility.</p>
<p><strong>One more thing:</strong></p>
<p><a href="http://thinscaletechnology.com/download-thinio/"><img class="aligncenter" src="http://thinscaletechnology.com/wp-content/uploads/Download-ThinIO.jpg" alt="" width="300" height="116" /></a></p>
<p>In case you missed ThinIO’s launch day at <a href="http://www.e2evc.com/home/" target="_blank">E2EVC </a>Barcelona, <strong>ThinIO is now in GA</strong>, available from our website and production ready! More marketing to follow! But grab your copy now and get playing!</p>
<p>&nbsp;</p>
<p>&nbsp;</p>
]]></content:encoded>
			<wfw:commentRss>http://andrewmorgan.ie/2014/10/thinio-facts-and-figures-part-4-storage-design-and-dangerous-assumptions/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
		</item>
		<item>
		<title>ThinIO facts and figures, Part 2: The Bootstorm chestnut.</title>
		<link>http://andrewmorgan.ie/2014/10/thinio-facts-and-figures-part-2-the-bootstorm-chestnut/</link>
		<comments>http://andrewmorgan.ie/2014/10/thinio-facts-and-figures-part-2-the-bootstorm-chestnut/#comments</comments>
		<pubDate>Sat, 18 Oct 2014 14:30:00 +0000</pubDate>
		<dc:creator><![CDATA[andyjmorgan]]></dc:creator>
				<category><![CDATA[Citrix]]></category>
		<category><![CDATA[ThinIO]]></category>
		<category><![CDATA[VMware]]></category>
		<category><![CDATA[VMware View]]></category>
		<category><![CDATA[Windows]]></category>
		<category><![CDATA[XenApp]]></category>
		<category><![CDATA[XenDesktop]]></category>
		<category><![CDATA[IOPS]]></category>
		<category><![CDATA[RDS]]></category>
		<category><![CDATA[Remote Desktop services]]></category>
		<category><![CDATA[VDI]]></category>
		<category><![CDATA[VMware Horizon View]]></category>
		<category><![CDATA[xenapp]]></category>

		<guid isPermaLink="false">http://andrewmorgan.ie/?p=3186</guid>
		<description><![CDATA[Welcome back! This blog post is part of a number of posts in advance of our upcoming release, for reference you can find part one below: ThinIO facts and figures, Part 1: VDI and Ram caching. Getting right to it: In this industry when somebody says ‘boot storms!&#8217; &#8211; most of us will respond with: Boot storms are a well documented, boring problem and have many solutions available from vendors and hypervisors alike. Most solutions today rely on a &#8216;shared memory&#8217; [&#8230;]]]></description>
				<content:encoded><![CDATA[<p><img class="alignright wp-image-2865 " src="http://andrewmorgan.ie/wp-content/uploads/2014/04/logo.png" alt="logo" width="174" height="33" />Welcome back! This blog post is part of a number of posts in advance of our upcoming release, for reference you can find part one below:</p>
<ul>
<li class="entry-title"><a href="http://andrewmorgan.ie/2014/10/thinio-facts-and-figures-part-1-vdi-and-ram-caching/" target="_blank">ThinIO facts and figures, Part 1: VDI and Ram caching.</a></li>
</ul>
<h2>Getting right to it:</h2>
<p>In this industry when somebody says ‘boot storms!&#8217; &#8211; most of us will respond with:</p>
<p><a href="http://andrewmorgan.ie/wp-content/uploads/2014/10/image0022.png"><img class="aligncenter  wp-image-3195" src="http://andrewmorgan.ie/wp-content/uploads/2014/10/image0022.png" alt="image002" width="163" height="119" /></a></p>
<p>Boot storms are a well documented, boring problem and have many solutions available from vendors and hypervisors alike. Most solutions today rely on a &#8216;shared memory&#8217; storage area to cache &#8216;on boot&#8217;, in theory caching only one startup or one pattern in order to then serve it back to the proceeding desktops to boot.</p>
<p>But why are boot storms an issue? While working on ThinIO we had the unique ability to really dive into the Windows boot process and analyse why boot storms cause the damage they do and in this post we thought we’d share our findings to better document the issue.</p>
<p><span id="more-3186"></span></p>
<h2>Boot data:</h2>
<p>Taking a typical windows 7 boot, to the login screen and idling until all services have started, the data traversing from disk to VM is relatively small. in our testing we found an average of Just 500-600 mb of data is read during this process, and write data barely registers at between 20 and 30mb.</p>
<p><a href="http://andrewmorgan.ie/wp-content/uploads/2014/10/image0031.png"><img class="aligncenter size-full wp-image-3189" src="http://andrewmorgan.ie/wp-content/uploads/2014/10/image0031.png" alt="image003" width="867" height="515" /></a></p>
<p>But hey, what gives? Taking such low data throughput, why is boot such a contenscious issue? Have I been misled with marketing and vendor nonsense?</p>
<h2><strong>The IO chestnut:</strong></h2>
<p>Sadly no, it’s the way windows requests this data, but don’t take my word for it…. Behold, the incredible mess that is the Windows boot process!</p>
<p><a href="http://andrewmorgan.ie/wp-content/uploads/2014/10/image0051.png"><img class="aligncenter size-full wp-image-3190" src="http://andrewmorgan.ie/wp-content/uploads/2014/10/image0051.png" alt="image005" width="867" height="532" /></a></p>
<p>Yep, that’s right, in the time Windows requested roughly 600mb of data, it sent down an astounding 70 thousand IO’s in the space of 2-3 minutes!</p>
<h2><strong>Math time:</strong></h2>
<p>Now if you were to take these figures as they stand, you would take 70,000 IO’s divide this into 560mb and you’d probably end up with an average of about 8k of data requested per IO… You’d be wrong.</p>
<p><em>As my good buddy Conor Scolard would say, ‘when you Assume, you make an ass out of you and me’.</em></p>
<p>To better understand the bounderies of Windows, Windows requests IO’s between the minimum of 512 bytes all the way up the spectrum later in the boot process to 128k and above. But it requests these blocks sparcely, on demand, and not just once per sector, the same blocks are frequently accessed.</p>
<p>The net result of this causes absolute havok on the storage:</p>
<p><a href="http://andrewmorgan.ie/wp-content/uploads/2014/10/image007.png"><img class="aligncenter size-full wp-image-3191" src="http://andrewmorgan.ie/wp-content/uploads/2014/10/image007.png" alt="image007" width="867" height="473" /></a></p>
<p>The crux of the issue is, for each one of these IO’s, the storage provider needs to compute the block data requested, seek the data out, then return it.</p>
<p>But 70,000 of these IO operations for a meagre 600mb of data is madness and you can now see exactly why boot storms were labelled as such for those early adopters who had their hands burned by this fact finding mission.</p>
<p><em>I’ll mitigate this issue by just booting my VM’s at night!</em></p>
<p>I’m sure you will! I would also love to see your face if a number of users happen to restart their desktops during the day, cascading 70,000 IO’s per desktop to the storage in a 2 minute window, per desktop!</p>
<h2><strong>Bootstorming IS an issue.</strong></h2>
<p>Now, knowing all this, it makes sense as to why storage and hypervisors alike are using a cache of ram.</p>
<h2><strong>But how does ThinIO fit in here? With Read Ahead of course!</strong></h2>
<p>Knowing the Windows boot process as intimate as only a technology like ThinIO can, there are many, many optimisations we can make to this process.</p>
<p>We can both speed the boot process up and also massively reduce the storage requirement while in VM, without any fancy caching mechanism!</p>
<p>With ThinIO’s read ahead technology, we can deliver just shy of an 80% boot IO reduction with nothing other than having our technology in the virtual machine:</p>
<p><a href="http://andrewmorgan.ie/wp-content/uploads/2014/10/image009.png"><img class="aligncenter size-full wp-image-3192" src="http://andrewmorgan.ie/wp-content/uploads/2014/10/image009.png" alt="image009" width="867" height="513" /></a></p>
<p>Taking a ThinIO averaged test and overlaying it to a baseline averaged test, it’s clear just how much impact this technology can have:</p>
<p><a href="http://andrewmorgan.ie/wp-content/uploads/2014/10/image011.png"><img class="aligncenter size-full wp-image-3193" src="http://andrewmorgan.ie/wp-content/uploads/2014/10/image011.png" alt="image011" width="867" height="473" /></a></p>
<p>&nbsp;</p>
<h2><strong>Wrap up:</strong></h2>
<p>So there you have it, with ThinIO, a simple, in VM solution, not only can you seriously reduce your IO footprint, boost user performance and achieve greater storage density per virtual machine, you also can also massively negate the impact a booting VM has on your storage.</p>
<p>If you would like a chance to test ThinIO pre-release, find access to the public beta below. Thank you for your time and happy testing!</p>
<p><a href="http://thinscaletechnology.com/download-thinio/" target="_blank"><img class="aligncenter wp-image-3171 size-medium" src="http://andrewmorgan.ie/wp-content/uploads/2014/10/Download-ThinIO-Beta-300x101.jpg" alt="Download-ThinIO-Beta" width="300" height="101" /></a></p>
<p>&nbsp;</p>
<p>&nbsp;</p>
]]></content:encoded>
			<wfw:commentRss>http://andrewmorgan.ie/2014/10/thinio-facts-and-figures-part-2-the-bootstorm-chestnut/feed/</wfw:commentRss>
		<slash:comments>2</slash:comments>
		</item>
		<item>
		<title>ThinIO facts and figures, Part 1: VDI and Ram caching.</title>
		<link>http://andrewmorgan.ie/2014/10/thinio-facts-and-figures-part-1-vdi-and-ram-caching/</link>
		<comments>http://andrewmorgan.ie/2014/10/thinio-facts-and-figures-part-1-vdi-and-ram-caching/#comments</comments>
		<pubDate>Mon, 13 Oct 2014 12:00:33 +0000</pubDate>
		<dc:creator><![CDATA[andyjmorgan]]></dc:creator>
				<category><![CDATA[Server Based Computing]]></category>
		<category><![CDATA[ThinIO]]></category>
		<category><![CDATA[ThinScale]]></category>
		<category><![CDATA[Virtual Desktop Infrastructure]]></category>
		<category><![CDATA[VMware View]]></category>
		<category><![CDATA[Horizon View]]></category>
		<category><![CDATA[IOPS]]></category>
		<category><![CDATA[VDI]]></category>
		<category><![CDATA[Vmware]]></category>

		<guid isPermaLink="false">http://andrewmorgan.ie/?p=3168</guid>
		<description><![CDATA[As we draw ever closer to ThinIO’s big day, I thought I’d put a blog post together talking about the RAM caching, statistics, facts and figures we’ve baked into version 1 to deliver some really kick ass performance improvements with even the smallest of allocations of cache per VM. Test, test, review and tune. Rinse and repeat! We’ve spent months load testing, tuning, fixing and retesting ThinIO. And for the first time we’re going to start talking about the dramatic [&#8230;]]]></description>
				<content:encoded><![CDATA[<p>A<img class="alignright wp-image-2865 " src="http://andrewmorgan.ie/wp-content/uploads/2014/04/logo.png" alt="logo" width="216" height="41" />s we draw ever closer to ThinIO’s big day, I thought I’d put a blog post together talking about the RAM caching, statistics, facts and figures we’ve baked into version 1 to deliver some really kick ass performance improvements with even the smallest of allocations of cache per VM.</p>
<h2>Test, test, review and tune. Rinse and repeat!</h2>
<p>We’ve spent months load testing, tuning, fixing and retesting ThinIO. And for the first time we’re going to start talking about the dramatic results ThinIO can have on storage scalability and user perceived performance.</p>
<p>During our extensive testing cycles, we’ve covered:</p>
<p>• Horizon View<br />
• Citrix XenDesktop<br />
• Microsoft RDS</p>
<p>We’ve been seeing very similar, if not identical results when testing against pools on the following storage types too:</p>
<p>• XenServer SR<br />
• VMFS<br />
• NFS<br />
• Microsoft Clustered Shared Volumes</p>
<p><span id="more-3168"></span></p>
<p>For reference the statistics we’re sharing today are based on VDI via VMware Horizon View 6, these figures are averages across at least three independent tests. All details of the tests that exported these results are covered below.</p>
<h2>Testing details:<img class="alignright  wp-image-3174" src="http://andrewmorgan.ie/wp-content/uploads/2014/10/Logo_Login_VSI_Transparent-300x50.png" alt="Logo_Login_VSI_Transparent" width="174" height="29" /></h2>
<p>The VM’s we tested in this particular workload are as follows:</p>
<p>• <strong>Testing Method:</strong> Login VSI 4.1 Medium Workload<br />
• <strong>Operating System:</strong> Windows 7 x64 SP1<br />
• <strong>System ram:</strong> 3gb<br />
• <strong>vCPU:</strong> 2<br />
• <strong>ThinIO cache:</strong> 350mb<br />
•<strong> Technology:  </strong>VMware Horizon View 6<br />
• <strong>Test runtime:</strong> 1 hour*<br />
• <strong>Statistic sample period:</strong> 5 seconds.</p>
<p>With that out of the way, lets jump right in!</p>
<h2>Storage IO:</h2>
<p>The number of IO’s per second is crucially important when dealing with storage; many, many small IO’s sent to sparse locations on disk are a killer to storage technologies, only made worse by certain file systems.</p>
<p>As a storage acceleration and negation technology, were extremely happy with the IO’s reduction we see on the storage:</p>
<p><a href="http://andrewmorgan.ie/wp-content/uploads/2014/10/image002.png" target="_blank"><img class="aligncenter wp-image-3173 size-large" src="http://andrewmorgan.ie/wp-content/uploads/2014/10/image002-1024x691.png" alt="image002" width="625" height="421" /></a></p>
<p>Even with just 350 MB of ram as a cache, we achieve phenomenal IO reduction.</p>
<h2>Storage MB / Sec:</h2>
<p>But IO’s are just one part of the puzzle, what about the size of the data requests being sent to the storage?</p>
<p>A true solution to take the pressure off the SAN, improve user performance, and increase storage density needs to tackle both the IO and the <strong>throughput</strong> problem.</p>
<p><a href="http://andrewmorgan.ie/wp-content/uploads/2014/10/image003.png" target="_blank"><img class="aligncenter size-large wp-image-3172" src="http://andrewmorgan.ie/wp-content/uploads/2014/10/image003-1024x648.png" alt="image003" width="625" height="395" /></a></p>
<p>As you can see above, with just 350mb we’re very good at it!</p>
<h2>Side by Side Comparison:</h2>
<p>So rounded figures are fine so long as the data is trustworthy, but here’s a real preview laid bare for your analysis.</p>
<p>Here’s a direct comparison on NFS and VMFS of taking a standard load test IO pattern and comparing it to an identical test with ThinIO installed in the VM:</p>
<h3 style="text-align: center;">VMFS:</h3>
<p><a href="http://andrewmorgan.ie/wp-content/uploads/2014/10/image004.png" target="_blank"><img class="aligncenter wp-image-3169 size-large" src="http://andrewmorgan.ie/wp-content/uploads/2014/10/image004-1024x466.png" alt="image004" width="625" height="284" /></a></p>
<h3 style="text-align: center;">NFS:</h3>
<p><a href="http://andrewmorgan.ie/wp-content/uploads/2014/10/image005.png" target="_blank"><img class="aligncenter wp-image-3170 size-large" src="http://andrewmorgan.ie/wp-content/uploads/2014/10/image005-1024x473.png" alt="image005" width="625" height="288" /></a></p>
<p>As you can imagine, we’re extremely proud of what we can achieve with as little as 350mb per desktop.</p>
<p>The beauty of our approach is simplicity, your users can see this benefit not in a matter of weeks, days or even hours. ThinIO can be up and running in minutes, delivering reduced login times, storage acceleration and providing a far deeper density on your current storage.</p>
<h2>Wrap up:</h2>
<p>So there you have it, we’ll be adding additional blog posts in the coming days looking at Remote Desktop Services / XenApp, intelligent cache management built in and our Read Ahead technology, so check back. In the mean time, if you would like a chance to test ThinIO pre-release, find access to the public beta below. Thank you for your time and happy testing!</p>
<p><a href="http://thinscaletechnology.com/download-thinio/" target="_blank"><img class="aligncenter wp-image-3171 size-medium" src="http://andrewmorgan.ie/wp-content/uploads/2014/10/Download-ThinIO-Beta-300x101.jpg" alt="Download-ThinIO-Beta" width="300" height="101" /></a></p>
<p>* Eight-hour figures and complete statistics are also available, we have nothing to hide and we’d encourage you to get in touch with the ThinScale team and we’ll share them with you.</p>
]]></content:encoded>
			<wfw:commentRss>http://andrewmorgan.ie/2014/10/thinio-facts-and-figures-part-1-vdi-and-ram-caching/feed/</wfw:commentRss>
		<slash:comments>7</slash:comments>
		</item>
		<item>
		<title>HDXWatcher and PCOIPWatcher &#8211; Realtime, easy virtual desktop traffic reporting.</title>
		<link>http://andrewmorgan.ie/2014/02/hdxwatcher-and-pcoipwatcher-realtime-easy-virtual-desktop-traffic-reporting/</link>
		<comments>http://andrewmorgan.ie/2014/02/hdxwatcher-and-pcoipwatcher-realtime-easy-virtual-desktop-traffic-reporting/#comments</comments>
		<pubDate>Mon, 24 Feb 2014 14:30:48 +0000</pubDate>
		<dc:creator><![CDATA[andyjmorgan]]></dc:creator>
				<category><![CDATA[VDI in a Box]]></category>
		<category><![CDATA[VMware]]></category>
		<category><![CDATA[VMware View]]></category>
		<category><![CDATA[XenApp]]></category>
		<category><![CDATA[XenDesktop]]></category>
		<category><![CDATA[Citrix]]></category>
		<category><![CDATA[View]]></category>
		<category><![CDATA[Vmware]]></category>
		<category><![CDATA[xenapp]]></category>

		<guid isPermaLink="false">http://andrewmorgan.ie/?p=2829</guid>
		<description><![CDATA[When checking the bandwidth requirement of multimedia sites, checking how much additional bandwidth video conferencing is going to require or even troubleshooting WAN capacity issues, it&#8217;s extremely useful to have a visible interpretation of realtime bandwidth consumption from your virtual desktop. I wrote a tool quite some time ago called watcher2 while troubleshooting a similar issue. I finally took the time to refactor that tool for use with XenApp 6.5 , XenDesktop and VMware View and they are finally available [&#8230;]]]></description>
				<content:encoded><![CDATA[<p><a href="/wp-content/uploads/2013/09/logo.png"><img class="alignright size-full wp-image-2784" src="/wp-content/uploads/2013/09/logo.png" alt="logo" width="88" height="89" /></a>When checking the bandwidth requirement of multimedia sites, checking how much additional bandwidth video conferencing is going to require or even troubleshooting WAN capacity issues, it&#8217;s extremely useful to have a visible interpretation of realtime bandwidth consumption from your virtual desktop.</p>
<p>I wrote a tool quite some time ago called watcher2 while troubleshooting a similar issue. I finally took the time to refactor that tool for use with XenApp 6.5 , XenDesktop and VMware View and they are finally available to download! Both watcher utilities also include a latency counter which was a request that came in over and over.</p>
<p>HDX and PCOIP watcher by default dock to the top of the screen and can be moved left or right as below:</p>
<p><img class="aligncenter size-full wp-image-2830" src="/wp-content/uploads/2014/02/hdx-watcher-docked.png" alt="hdx watcher docked" width="310" height="52" /></p>
<p><img class="aligncenter size-full wp-image-2832" src="/wp-content/uploads/2014/02/pcoip-watcher-docked.png" alt="pcoip watcher docked" width="281" height="33" /></p>
<p>They can now also be completely un docked:</p>
<p><img class="aligncenter size-full wp-image-2831" src="/wp-content/uploads/2014/02/hdx-watcher.png" alt="hdx watcher" width="294" height="72" /></p>
<p><img class="aligncenter size-full wp-image-2833" src="/wp-content/uploads/2014/02/pcoip-watcher-undocked.png" alt="pcoip watcher undocked" width="289" height="65" /><span id="more-2829"></span></p>
<p><strong>How do they work?</strong></p>
<p>The tool finds your username in the performance monitor counters for session bandwidth, once it finds this entry it reads your performance monitor data once every second and reports on it.</p>
<p>In the case of PCOIP watcher, it reads the PCOIP counters from performance monitor.</p>
<p><strong>what do the values mean?</strong></p>
<p><em>All values are in either Kilobits per second or Megabits per second.</em></p>
<p><strong>In</strong> = Traffic from the client to the virtual, this may spike during large copy / paste jobs,web cams or copying data from a usb key to the session:<br />
<strong>Out</strong> = Traffic from the virtual desktop to the client, mainly audio or video traffic causes this to spike.<br />
<b>Latency = </b>The delay between your client and the virtual desktop.</p>
<p><strong>Can I Configure it?</strong></p>
<p>Two thresholds are available, a yellow warning and a red warning, currently . These default values can be written to:</p>
<p>&#8220;HKEY_local_machine\software\ThinScaleTech\HDXmonitor&#8221;<br />
&#8220;HKEY_current_user\software\ThinScaleTech\HDXmonitor&#8221;</p>
<p>Dword: YellowWarning = 300 (decimal)</p>
<p>Dword: RedWarning = 600 (decimal)</p>
<p><strong>Do they have any dependencies?</strong></p>
<p>.net framework 3.5</p>
<p>if you are running XenApp 6.5 or XenDesktop 5.6, ensure you have the latest hot-fixes installed or the counters may be incorrect.</p>
<p><strong>How do I launch it?</strong></p>
<p>Allow the user to run it manually, or place the executable in their start-up folder or login script.</p>
<p><strong>Where Can I download it?</strong></p>
<p><a href="https://app.box.com/s/gfftbwrcy216jos80jxz" target="_blank">Here:</a></p>
<p><strong>What&#8217;s coming next:</strong></p>
<ul>
<li>Native Microsoft RDP Counters.</li>
<li>Realtime graphs and recording.</li>
<li>source code is available on request.</li>
</ul>
]]></content:encoded>
			<wfw:commentRss>http://andrewmorgan.ie/2014/02/hdxwatcher-and-pcoipwatcher-realtime-easy-virtual-desktop-traffic-reporting/feed/</wfw:commentRss>
		<slash:comments>25</slash:comments>
		</item>
	</channel>
</rss>
