VMware Software Manager: The Good…The Bad…The Alternative!

In this post, I am going to discuss a little, “not-so-well-known” utility, called VMware Software Manager.  This little “beast” was first released as v1.0 back on 2015-03-12, and its most current release, v1.5, came out on 2016-08-25.  So as you can see, it’s been quite a while since this tool has seen a new update release.  The problem now is that this utility seems to have been forgotten and/or neglected by VMware, but I will get into more of that a little later.  Let’s start off with the positive stuff.

The Good:

The VMware Software Manager allowed valid VMware-account holders the ability to download various software, such as ESXi or vCenter server, quickly and easily.  The installation was a breeze, the interface was clean and downloading software was effortless.  It basically just worked flawlessly!  The new software was readily available for download shortly after it was announced/released since this utility would read configuration files to see what software is available from the VMware repository and then provided that software for download.

So, having said that, what could possibly be wrong with this thing?  Let’s continue…

The Bad:

My main gripe with this utility was that support was only community-based, so if you had issues, you could forget about raising an official SR with support.  You had to rely patiently on the VMTN community and hope that users were knowledgeable enough and willing to help out.  Not saying anything bad about the community though, it’s a great forum full of some really smart people and it provides a wealth of information.

As I mentioned earlier, when this thing worked…it just plain WORKED!  Then, vSphere 6.0U3A was released with some bad/corrupt or missing files and upon launching the utility, it would simply hang at login or error out due to the missing files not being available! (Shhhh!..someone never updated the configuration files… ¯\_(ツ)_/¯  )  There is a thread over on the VMTN community that myself and others have contributed to and another describing this issue and a workaround method to ultimately get you to log in again.  This involved looking at log files to find which missing file was causing the error, then editing the configuration file to remove the missing files, and it was a major pain!

Lastly, it’s become extremely out of date and it seems as if VMware has completed forgotten about and/or decided to neglect and give up on it as there has not been any new software available for download in at least 6 months.  It’s also possible that the original responsible for this thing disbanded or moved onto other roles…who really knows?  Luckily, a friend and fellow community/vExpert member has provided a solution!

The Alternative:

Fellow vExpert, Edward Haletky aka Texiwill, has created a Linux-based port of the utility, titled “VSM“, which he updates almost daily to add new software and simply improve the appliance and it runs on an RHEL type distro like CentOS.  You can hit the ground running with this appliance in about 30 mins or less.  I have been fortunate enough to serve as a beta tester for him and have been doing so since he released v0.95, just shortly after the initial public launch.  At the time of this writing, the most recent release was v3.7.7 but has now been updated to the newest 4.5.3.  Screenshots may reflect previous versions.

Well…How do you get it?

In this post, I am going to cover how to install this bad-boy on a CentOS 7 minimal installation on ESXi using NFS, and on VMware Workstation leveraging the “Shared Folders” feature using VMHGFS.  Another vExpert, Michael White, has a similar post on setting up this appliance and using SMB/CIFS for storing the downloaded software.  Let’s get to it, shall we!


  • CentOS 7 x64 (Minimal) w/ open-vm-tools installed
  • a non-root user account
  • NFS storage or Local Storage (if running this on VMware Workstation/Fusion)

I will assume that you already have a basic installation of CentOS 7 running, or know how to set up a minimal installation, so that process is out of scope for this post.  

If you plan on running this on VMware Workstation/Fusion, run the following command to install the prerequisites needed for installing VMware Tools to get the vmhgfs driver for Shared Folders support.

sudo yum install -y perl gcc binutils make fuse kernel-headers kernel-devel net-tools policycoreutils-python

Otherwise, if you plan on using NFS to store the software, run the following command to install the NFS utilities.

sudo yum install -y nfs-utils

VMware Tools Installation:

Note: Enables vmhgfs driver support for Shared Folders on VMware Workstation or VMware Fusion

I will be installing the latest version of VMware Tools, version 10.2.0.  This can be obtained here.  Extract the .zip and attach the linux.iso to your CentOS VM in Workstation or Fusion.  Once connected, do the following.

Make a directory to mount the cdrom to.

sudo mkdir /mnt/cdrom

Mount the cdrom to this new directory.

sudo mount /dev/cdrom /mnt/cdrom

With the cdrom now mounted, let take a look at whats on the .iso

sudo ls /mnt/cdrom

We can see the compressed file that we will need to extract

Extract/Uncompress this file by running the following.  This is going to extract the file to my current working directory, $HOME

sudo tar zxpf /mnt/cdrom/VMwareTools-10.2.0-7253323.tar.gz

When that completes, we can see what was extracted by running


At this point, we are done with the .iso and can unmount it by running.

sudo umount /dev/cdrom

Now, let’s navigate to the extracted folder and see what we’ve got.

cd vmware-tools-distrib/

The Perl script is what we’re looking for here to install VMware Tools.  This will uninstall “open-vm-tools” if that is already installed on this machine.  Let’s go ahead and run it. 

sudo ./

Proceed to answer all the questions asked and select all of the defaults until the installation completes.

(Optional:) If you’d prefer to bypass the questions and force-install with all defaults run the following instead

sudo ./ --force-install --default

This completes the VMware Tools installation and not the vmhgfs driver is installed.  Onto the good stuff!

VSM Installation:

Texiwill has provided an installation script on his GitHub page which will take care of the installation and its prerequisites. This is the preferred installation method so I’d advise that you download this script, and run it.  You’ll first have to make the script “executable” by running

chmod +x

For those who prefer the manual approach, you’ll first need to install “wget“, then run the following commands to install the utility.

Note: please change timezone location to your respective location

sudo yum install -y wget
mkdir aac-base
cd aac-base
wget -O aac-base.install
chmod +x aac-base.install
./aac-base.install -u America/New_York
sudo ./aac-base.install -i vsm America/New_York

Congrats!  You’ve successfully installed the utility.  Now to configure it for use.

By default, the VSM script will save the config file (.vsmrc) to “$HOME“.  It will also save the index.html and credstore files in “/tmp/vsm“.  I highly recommend you create an alternate directory to store these files in if you are using a system that has multiple users or change it to save them in the users $HOME directory.  Again, this is optional and can be skipped, but if you’d like to do so, run the following

mkdir /tmp/my_vsm

VSM Initial Configuration:

The last step before you can use the VSM script utility is to configure it.  To list the all available parameters, execute the script with the help parameter. -h

Based on the help info, the required parameters to set are:

  • -v | –vsmdir VSMDirectory (this will be the “my_vsm” directory we created earlier)
  • –repo repopath (this will be the path to the directory we’ll create to save our download)

When first executing the script utility, you will be required to enter valid MyVMware credentials which will be saved to the configuration file.

Let’s go ahead and create our repo directories first.  Feel free to use my examples or anything else you’d like.  We will also use the same steps mentioned about to take ownership of these directories and then mount our physical storage to the newly created mount point directory. 

Note: Be sure to enable “Shared Folders” on Workstation or Fusion and select the shared folder to mount it to the VM.  Commands may be different from screenshots as I’ve updated some commands.


sudo mkdir -pvm 755 /mnt/vmhgfs/depot/content
sudo chown -R <non-root_username>.<non-root_username> /mnt/vmhgfs

For NFS:

sudo mkdir -pvm 755 /mnt/nfs/depot/content
sudo chown -R <non-root_username>.<non-root_username> /mnt/nfs

With our directories ready, let’s go ahead and mount our storage to them!


Note: I have mounted the following shared folder to the VM to use for this command.

sudo mount -t fuse.vmhgfs-fuse .host:/content /mnt/vmhgfs/depot/content -o allow_other

For NFS:

sudo mount -t nfs <IP_address>:/<volume#>/<path_to_share> /mnt/nfs/depot/content

For NFS 4.1 (optional):

sudo mount -t nfs4 -o vers=4,minorversion=1 <IP_address>:/<volume#>/<path_to_share> /mnt/nfs/depot/content

Now, with our storage mounted we can run the VSM script and configure for use.

Note: You will be prompted to enter your VMware credentials unless you supply the “-u” and “-p” parameters

For VMHGFS: -y -v /tmp/my_vsm --repo /mnt/vmhgfs/depot/content --save

For NFS: -y -v /tmp/my_vsm --repo /mnt/nfs/depot/content --save

That about does it!  The appliance is ready to use and you can navigate through the menus to find your desired software. 


To make your mount points persistent through reboots, edit the “fstab” file with your editor of choice.  I prefer to use vi or vim, but many may choose to use nano instead.

sudo vi /etc/fstab

Then, add the following line and save the file.


.host:/content /mnt/vmhgfs/depot/content    fuse.vmhgfs-fuse        allow_other     0 0

For NFS:

<IP_address>:/<volume#>/<path_to_share> /mnt/nfs/depot/content    nfs      soft,bg,rsize=8192,wsize=8192       0 0

For NFS 4.1 (optional):

<IP_address>:/<volume#>/<path_to_share> /mnt/nfs/depot/content    nfs      nfsvers=4.1,soft,bg,rsize=8192,wsize=8192       0 0

Updates are released quite frequently.  To update VSM to the latest version, you can run the following commands manually or add them to a shell script within “/etc/cron.daily” which will run around 3 AM.

Note: please change timezone location to your respective location

cd /home/<user_name>/aac-base; ./aac-base.install -u; ./aac-base.install -i vsm America/New_York

Additionally, if you’d like to keep your favorite download repository up-to-date, edit crontab

Note:  This requires that you first “mark” a repository by using that menu option

crontab -e

Then add the following line to it and save.  This will run daily at 6 AM.

0 6 * * * /usr/local/bin/ -y -mr -c --favorite

My Preference:

I tend to use the following command when running VSM to get missing suites and packages from MyVMware.  If you do not use the “-m” parameter, it will only pull the same software available in the original VMware Software Manager tool from VMware.  The “-mr” parameter resets the “MyVMware” software info and implies “-m” hence why this is also used in the crontab line above.  The “-c” tells VSM to generate sha256sum checks against each downloaded file. -y -mr -c

Note: Depending on what you downloaded, there may be certain .txt files that will fail the checksum.  This is expected and can be safely ignored.

Well, I hope you’ve found this post useful and I thank you for reading!  Special thanks again to Texiwill for making this awesome utility, as well as Mike White for posting his similar article using SMB/CIFS.




  • 03.23.18 – Updated to reflect changes in manual install commands for VSM v4.0.2, and cron.daily entry since cron runs as root so no need to use sudo
  • 04.12.18 – Added TZ (Time Zone) setting to manual install commands and modified /etc/fstab command for NFS mounts
  • 04.15.18 – Added “-c” parameter reference
  • 01.17.19 – Added some command syntaxes for NFS 4.1

vSphere…Synology…NFS v4.1

| 31/12/2017 | Tags: , , , ,

Welcome, and thanks for visiting my blog!

In this post, I am going to cover how to enable NFS v4.1 on a Synology device and then mount and NFS v4.1 datastore in VMware vSphere 6.5.  By default, Synology devices support NFS v4 natively, and although they can also support NFS v4.1, it is not enabled.  Well, not to worry because I am going to show you just how to enable the feature on your device.

NFS v4 and v4.1 have been around for quite a few years but it has not taken off then way NFS v3 did way back when.  There were some major flaws pointed out with NFSv4 so NFSv4.1 was created to rectify those flaws, and VMware was one of the first major companies to adopt and support the new Network File System.  But unless your storage device supported the newer NFS versions, you would be stuck mounting NFSv3 volumes by default.

In this demo, I will be using my new replacement Synology DiskStation DS415+ and my homelab “datacenter” running the latest version vSphere 6.5.  So let’s jump right in!

Using a terminal application like PuTTY, connect to your Synology device via SSH using an admin user account.  This can be the default “admin” account and any new user account with Administrator privileges.  Once connected enter the following command to change the directory:

cd /usr/syno/etc/rc.sysv

Once in this directory, run the following command (enter the account password if prompted):

sudo cat /proc/fs/nfsd/versions

This will show us the current NFS version currently enabled and supported by the Synology device. 

We can see that all versions prior to 4.1 have a “+” sign next to them and 4.1 has a “-” sign next to it.  Let’s change that!

In order to change this, we will need to edit a shell ( file using “vi”.  Run the following command to open the file with VI Editor:

sudo vi

This will open the shell file and will place the cursor at Line 1, Character 1 as depicted in the following screenshot.  

Navigate down to line 90 using the down arrow and you will see the following line of text.

This is where the magic happens!  To edit the file now, press the “I” key on your keyboard to initiate an “Insert” then add the following to the end of the text so the line looks like the following screenshot.

-V 4.1

To commit and save this change, first press the Esc key.  Next type the following command and hit “Enter” to write and then quit vi editor.




Next, we need to restart the NFS service.  To do so enter the following command:

sudo ./ restart


If we again run the following command, we will see that there is now a “+” sign next to 4.1.  Hooray!

sudo cat /proc/fs/nfsd/versions


Now that we have enabled NFSv4.1 functionality on your storage device, let’s go ahead and mount an NFS volume to our hosts in vSphere.

I have enabled NFS and NFS v4 support then created the following shares with assigned permissions on my device, and am going to mount the ISOs share first in this example by issuing a command via PowerCLI.  We can also see that I do not have any NFS mounts currently in my environment

I’ve launched PowerCLI and connected to my vCenter Server using the Connect-VIServer cmdlet then issued the following command:

Get-VMHost | New-Datastore -Nfs -FileSystemVersion '4.1' -Name SYN-NFS04-ISOs -Path "/volume3/NFS04-ISOs" -NfsHost DS415 -ReadOnly

*Note:* an important argument here in the “-FileSystemVersion”.  If I do not specify the version, it will assume version 3.0 by default.

If I go back and look at my datastores via the Web Client, I can see that my new NFS 4.1 datastore has been mounted to each one of my ESXi hosts. Nice!

*Bonus:* If I’d like to easily remove this datastore from all of my hosts, I can issue the following command via PowerCLI.

Get-VMHost | Remove-Datastore -Datastore SYN-NFS04-ISOs -Confirm:$false

Now I can see that the host has been removed successfully!

Well, that about wraps this one up.  I hope that this has been useful and informative for you and I’d like to thank you for reading!  Until next time!


Homelab Makeover 2.0

Hello and first off, thank so much for visiting my blog!  If you have followed any part of my “Homelab” series, you will be familiar with the components that make up my home “Datacenter”.  If not, take some time to catch up on those posts!

In this post, I am quickly going to cover my lab makeover as I decided to get some new equipment and redo a bunch of my networking.  So without any further hesitation, let’s get to it!

Beginning with my networking equipment, I wanted to move my Cisco SG300-10 out of my home network enclosure cabinet and into my Navepoint rack enclosure.  But then I realized I would have to replace that switch with another to feed the rest of my homes connections.  Currently, I am using Ubiquiti’s UniFi equipment for my home networking and since I’m already running Ubiquiti gear, I figured I would purchase a few more of their 8-port switches to do the job so that I can manage those devices from a “single-pane-of-glass” via the controller.  So I went ahead and purchased 2 US-8 switches, in which 1 will feed the home networking and the other will extend to the lab primarily serving as a trunk for my VLANs to reach the labs Cisco switches.

So now, my UniFi network consists of:

On to the lab network…

The US-8-LAB switch connects to my SG300-10 which I’ve configured 2-ports as a LAG “Trunk” between the switches for VLAN traffic, 2-ports as another LAG “Trunk” connection to the SG300-52 switch, and the others as “Access” ports which connect to the IPMI interfaces of my servers.  The IPMI connections were previously on my SG300-52 switch.  On to the SG300-52 switch, I have configured all of my ESXi management ports, vMotion ports, iSCSI & NFS ports, VSAN ports, and data ports for my servers, along with a few LAG connections which connect to my storage devices, and a few which connect my UPS and ATS/PDU units.  I also configured an additional LAG “Trunk” which connects to a Netgear Prosafe GS108T that I had laying around.  I’ve dedicated that switch and it’sports for my ex-gaming PC turned “DEV” ESXi host.  Eventually, that host will be decommissioned when I add a new host to my rack enclosure.

So now, my lab network consists of:

Now for the storage devices.  Previously, I was running my lab VMs using a Synology DS415+ storage unit via NFS mounts.  This was all fine and dandy, except for the fact that it would randomly shut itself down for no apparent reason, leading to eventual corruption of my VMs.  I got tired of spending hours trying to recover my machines and eventually discovered that my device was plagued by the Intel ATOM C2000 CPU issue described here.  I then reached out to Synology and they quickly responded and issued an immediate RMA of the device.  Again this was fine, but where was I going to move my VMs and data too?  I didn’t have another storage device with an ample amount of free space to accommodate all my data, so I decided to bite the bullet and pick up a brand new Synology RS815+ which I could now mount in my rack enclosure.  I also scooped up some 1TB SSDs from their compatibility matrix to populate the drive bays.  The difference here is that with the new RackStation, I opted to configure my LUNs via iSCSI instead of NFS like I had previously done with the DiskStation.  Once set up and connected, I vMotion’d all of my machines to the new device, and disconnected the DS415+ while I waited for the replacement device to arrive.  That replacement unit eventually came, so I swapped my SSD’s from the old unit into the new unit and fired it back up.  I will eventually recreate some NFS mounts and reconnect them to the vSphere environment.

Now, my lab storage consists of:

Finally, the cabinet.  I became rather displeased with the amount of space I had with my Navepoint 9U 450mm enclosure.  The case itself was great, but I just needed some more room in the event I needed to un-rack a server or do anything else in there.  Also, I started to do some “forward-thinking” about eventual future expansion, and the current 9U enclosure was no longer going to suffice.  I decided to upgrade to a new Navepoint 18U 600mm enclosure, and now I have plenty of room for all of my equipment and future expansion.  After relocating my servers to the new rack enclosure, I now have the following equipment mounted in the rack and, still, have room for growth.

  • 2 x Cat6 keystone patch panels
  • 2 x Cisco SG300 switches
  • 4 x Supermicro servers
  • 1 x Synology storage unit
  • 1 x UPS
  • 1 x ATS/PDU
  • 1 x CyberPower Surge power strip (in the event I need to plug-in some other stuff)

Thanks for stopping by!  Please do leave some comments as feedback is always appreciated!  Until next time!



Create a Synology VM with XPEnology

| 30/04/2016 | Tags: , ,

Create a Synology VM with XPEnology


I’m a huge fan of Synology NAS systems, but I must say, they do often put a gaping hole in your wallet.  Well, fortunately the folks over at XPEnology have created an alternative way for us to create your own Synology devices, whether it be deployed on a bare-metal system or as a virtual machine.  I currently own a few Synology NAS devices, but I love having the ability to spin up a working VM version quickly and with ease, for use in my nested lab environments.


In this post, I am going to show you how to create your very own Synology VM on VMware ESXi, Workstation, and Fusion hypervisors.  As I mentioned previously, you can also deploy this onto a bare-metal system, but since I do not have a spare system to test this with, I will not cover that deployment.  So without any further hesitation, let’s get to it!


  • VMware ESXi
  • VMware Workstation
  • VMware Fusion
  • XPEnology files
    • DSM 5.x PAT file – (please note that DSM 6.0 remains unsupported with XPEnology at this time)
    • XPEnoboot 5.x VMDK file
    • Synology Assistant (optional but recommended)
    • NFS Plug-in for VMware VAAI (optional but recommended)
    • Open-VM-tools (optional but recommended)

The type of Synology system that XPEnology emulates is a DS3615xs 12-bay unit.  Let’s begin by heading over to the XPEnology website and grabbing what we need.  At the time of this writing, the current version of DSM is the newly released DSM 5.2-5967, but the XPEnology team has not yet created an updated bootloader (XPEnoboot) that supports this version.  The current stable version is DSM 5.2.-5644 Update 5 (and can be manually updated to Update 8 at a later time).  The current stable version of XPEnoboot is 5.2-5644.5 so I’ve downloaded the following files…


Now the fun begins, and I will start with the VMware ESXi deployment.  I will be deploying on the newest version of ESXi 6.0 Update 2 via the vSphere Web Client.  For those of you using the vSphere C# client, the directions are the same just the interface is different.


ESXi 6.0 Update 2

Part 1:

Create a new virtual machine


Give it a name and select a folder to place the VM in


Select a compute resource


Select a storage location


Select the proper Compatibility version


Select the following Guest OS settings


Customize the VM with the following settings, remove the floppy, and select the appropriate network 


Finish and let it build the “Shell” VM 



Part 2:

Now that the VM has been built, we need to add the XPEnboot VMDK that we downloaded earlier to the VM

Browse to the datastore where you created the VM and upload the XPEnoboot VMDK 


Once uploaded, we need to add the disk to the VM 

2016-04-29_22-29-15 2016-04-29_22-29-562016-04-29_22-30-58

At this point, I also like to add an additional SCSI disk which will be used as an NFS volume, or iSCSI LUN.  You can make this any size you like but for the purposes of this post, I’m just going to add a simple 50GB Thin Provisioned disk. The 16GB disk we created with the VM will be used to install application packages (i.e. – VMware Tools, etc.)

Keep in mind that this is a 12-bay device, so you can technically add 10 more disks to fill all the drive bays


Now power-on the VM and open the console window.  Keep an eye on the window for the IP address assigned to the VM.  You can connect to this IP using a web browser instead of using the Synology Assistant to detect and connect to it.


Open you favorite web browser and type in the IP address of the VM, and hit enter.  This will connect you to the VM’s Synology Web UI

Click the Set up button, then click the Manual Install link and browse for the DSM .pat file we downloaded earlier. Then click the Install Now button

2016-04-29_23-03-382016-04-29_23-04-27 2016-04-29_23-06-12

This will prompt you that it will erase all data on all disks, including the XPEnoboot disk we uploaded and attached earlier.  Note, I have experienced failures at this point and received a message that the .pat file may be corrupt.  In the even this also happens to you, please use this alternative XPEnology download link.


Once complete, the VM will be rebooted.  But since we erased all of the disks during installation, the VM will fail to boot properly and this is expected

2016-04-29_23-17-15 2016-04-29_23-17-36

Power off the VM and access the datastore that the VM is stored on.  Delete the XPEnoboot VMDK then re-upload the VMDK that we originally downloaded

2016-04-29_23-19-41 2016-04-29_22-27-40

Power on the VM and select the Install/Upgrade option


When it finishes it boot, this time you will notice that it will not display the IP address we saw the last time.  The IP should have remained the same but I’m going to use the Synology Assistant to detect it and help me connect since my web browser was not connecting to the same address

2016-04-29_23-25-50 2016-04-29_23-44-17

Now we are presented with the Web UI screen and we can login with admin and a blank password


Click next and give the Synology VM a name and change the admin user password or enter a new username and password then click Next

2016-04-29_23-48-13 2016-04-29_23-50-20

In order to prevent updates from installing automatically and possibly breaking the boot up of the Synology VM, I chose “Download DSM updates but let me choose whether to install them” and clicked next


I also chose to skip the “Set up QuickConnect”


And….You’re Done!  You now have a fully functional Synology virtual machine.  You can feel free to add additional disks or what have you.


At this point, I like to go ahead and install the Open-VM-Tools package so that I can use the VMware tools to gracefully power off the VM as needed.  To do so, you will need to have downloaded the package from the XPEnology website.

Open Package Center, then click the Settings button.


On the General tab, set the Trust Level to Any Publisher and click OK


Back in the Package Center, click Manual Install.  Browse to the package and click Next


Since we have yet to create a volume, it will prompt you to click OK to launch Storage Manager and create a volume to install the package on.


In Storage Manager, click Volume the click Create.  Keep the default of “Quick”, Next, select Disk 3, Next, OK to the warning message, Yes for disk check, Next, Apply….Done!

2016-04-30_0-08-57 2016-04-30_0-11-08 2016-04-30_0-12-02 2016-04-30_0-13-10 2016-04-30_0-14-49 2016-04-30_0-15-13

Now, repeat step 3 above, click Apply, and you will have VMware tools installed.


Wow, that seemed like an awful lot of steps, but it really wasn’t all that much.  In the next sections, I will just go over the deployment of the “Shell VMs”  and the steps to add the XPEnoboot VMDK to the VMs in VMware Workstation and VMware Fusion as the rest of the post-boot setup steps are the same.


VMware Workstation Pro 12.1.1

Part 1:

Create the Shell VM…

2016-04-30_1-07-28 2016-04-30_1-07-43 2016-04-30_1-08-01 2016-04-30_1-08-17 2016-04-30_1-10-34 2016-04-30_1-10-46 2016-04-30_1-11-05 2016-04-30_1-11-47 2016-04-30_1-12-14 2016-04-30_1-12-27 2016-04-30_1-12-39 2016-04-30_1-13-17 2016-04-30_1-13-40 2016-04-30_1-15-42

Part 2: 

Add the XPEnoboot VMDK to the VM…


2016-04-30_1-18-32 2016-04-30_1-18-46 2016-04-30_1-19-02 2016-04-30_1-19-14 2016-04-30_1-19-54 2016-04-30_1-21-00


Power on the VM and continue with same process outlined above for ESXi deployment.  Use the Synology Assistant to find and connect as the VM may not display the IP address right away.


VMware Fusion Pro 8.1.1

Part 1:

Create the Shell VM…

2016-04-30_09-21-41 2016-04-30_09-22-04 2016-04-30_09-22-36 2016-04-30_09-22-58 2016-04-30_09-24-14

Part 2:

Add the XPEnoboot VMDK to the VM…

2016-04-30_09-25-28 2016-04-30_09-26-35 2016-04-30_09-27-02 2016-04-30_09-29-08 2016-04-30_09-30-49 2016-04-30_09-31-51 2016-04-30_09-32-41 2016-04-30_09-33-14 2016-04-30_09-34-58 2016-04-30_09-36-10 2016-04-30_09-36-45 2016-04-30_09-37-39

Power on the VM and continue with same process outlined above for ESXi deployment.  Use the Synology Assistant to find and connect as the VM may not display the IP address right away.


Now you should all feel confident in creating your own Synology VM systems for whatever your use case may be and hopefully you will all love using Synology products as much as I do.  If you would like the true experience, I’d recommend purchasing a real system.  But in the interim, this will give you a way to create your own and play around with it.

Keep in mind that when updates are released, it is always wise to first check on the XPEnology site and/or forum to see if the updates break anything as sometimes they will cause the boot process to break since the XPEnoboot cannot support it yet.  Usually you are safe to update as long as the update is within the same DSM version, but consult the site/forum first as I am not responsible for any upgrade impacts.

Well, I hope that you all have enjoyed this read/guide and come back for more!

And giving credit where it is due, I’d like to give a shout out to Erik Bussink as his guide is what inspired me to write this updated guide for all of you!