With VMworld 2014 in the United States fast approaching, I have been working on building out my schedule based on my personal objectives and checking the popular blogger sites for their recommendations. In that spirit, I thought I would share the sessions I am most excited about this year in San Francisco.
Last year was my first year at VMworld and I focused on the Hands-on-Labs (HoLs) and generic sessions to better understand the VMware ecosystem. This year I am focused on three primary topics:
- VMware NSX
- Openstack|Docker|Containers with VMware
- VMware VSAN
Here are the sessions I am focused on:
- SEC1746 NSX Distributed Firewall Deep Dive
- NET1966 Operational Best Practices for VMware NSX
- NET1949 VMware NSX for Docker, Containers & Mesos
- SDDC3350 VMware and Docker — Better Together
- SDDC2370 Why Openstack runs best with the vCloud suite
- STO1279 Virtual SAN Architecture Deep Dive
- STO1424 Massively Scaling Virtual SAN implementations
In addition to that, I am also excited for my own sessions at VMworld this year around Hadoop , VMware BDE and building a Hadoop-as-a-Service!
- VAPP1428 Hadoop-as-a-Service: Utilizing VMware Cloud Automation Center and Big Data Extensions at Adobe (Monday & Wednesday sessions)
Excited for the week to get kicked off and see all the exciting things coming to our virtualized world.
Not specifically related to Hadoop or Big Data Extensions, but I came across this bug tonight. There is a KB article on the VMware website (here), but the syntax it lists is incorrect.
The error I was seeing on the VM console was “vmsvc [warning] [guestinfo] RecordRoutingInfo: Unable to collect IPv4 routing table” immediately after it brought eth0 online. The workaround to fix the issue, beyond upgrading arping in the OS, is to add the following line in the virtual machine .vmx file:
rtc.diffFromUTC = “0”
The quotes are missing from the VMware knowledge base article and are indeed necessary to fix the issue and get the virtual machine past this point in the boot process.
Working on a specific use-case at work has required that I modify the Chef recipe templates for mapred-site.xml and yarn-site.xml to configure the memory allocations correctly. The container sizes themselves will depend on the size of VMs you are creating, and BDE has some generic settings by default, but again with each workload being different it is necessary to tune these parameters just as you would with a physical Hadoop cluster.
The virtual machines within this compute-only (Isilon-backed HDFS + NameNode) cluster utilized the ‘Medium’ sized node within BDE. That means:
- 2 vCPU
- 7.5GB RAM
- 100GB drives
The specific YARN and MapReduce settings I have used to take advantage of the total memory allocated to the cluster was:
155 <% else %>
161 <!-- <property> -->
162 <!-- <name>mapred.child.ulimit</name> -->
163 <!-- <value><%= node[:hadoop][:java_child_ulimit] %></value> -->
164 <!-- </property> -->
167 <description>MapReduce map memory, in MB</description>
173 <description>MapReduce map java options</description>
179 <description>MapReduce reduce memory, in MB</description>
185 <description>MapReduce reduce java options</description>
191 <description>MapReduce task IO sort, in MB</description>
196 <% end %>
73 <description>Amount of physical memory, in MB, that can be allocated
74 for containers.</description>
76 <!-- <value><%= node[:yarn][:nm_resource_mem] %></value> -->
81 <description>The amount of memory the MR AppMaster needs.</description>
83 <!-- <value><%= node[:yarn][:am_resource_mem] %></value> -->
88 <description>Scheduler minimum memory, in MB, that can be allocated.</description>
94 <description>Scheduler maximum memory, in MB, that can be allocated.</description>
100 <description>Application master options</description>
127 <description>Disable the vmem check that is turned on by default in Yarn.</description>
Again, mileage will vary depending on your Hadoop workload, but these configuration settings should allow you to utilize the majority of the memory resources within a cluster deployed with the ‘Medium’ sized nodes within BDE.
I used the following articles as guidelines when tuning my cluster, along with trial and error.