Follow along with the video below to see how to install our site as a web app on your home screen.
Note: This feature currently requires accessing the site using the built-in Safari browser.
1. VMotion
2. High availability
3. Suspend and resume
4. Record and replay
5. Fault tolerance
6. Memory overcommitment and page sharing
7. Hot add/remove of virtual devices
8. No Snapshot backup
thanks for the tip!
Question for all you other all-in-one devotees: what is the best way to back up the OpenIndiana VM that hosts your ZFS SAN when you have used PCI hardware passthrough of the HBA? Using VMDirectPath I/O has the trade off of getting better performance by eliminating the virtual I/O overhead at the cost of preventing the use of the following VMWare capabilities:
Code:1. VMotion 2. High availability 3. Suspend and resume 4. Record and replay 5. Fault tolerance 6. Memory overcommitment and page sharing 7. Hot add/remove of virtual devices 8. No Snapshot backup
For those us using the free version of ESXi 5, we lose the first five of that list anyway. It is the loss of VMWare Snapshot backup that causes our particular problem. We were trying out Veeam Backup and Replication (free version) to back up all our VMs before upgrading the server to ESXi 5.1 and discovered that backing up the OpenIndiana VM failed due to no VMware snapshot capability.
So I'm curious how the experts out there back up their OI VM (assuming they don't have a second ZFS installation where they could do ZFS send).
Thanks in advance,
--peter
Hi guys ...
I'm having a problem managing automatic snapshots through the napp-it UI.
Jobs > snap > create autosnap job
I'm configuring everything as I'd like it to be and when I hit the submit button the page reloads, but it's not creating/displaying the job in the task list at the top of the page.
At first I thought this meant that it wasn't creating the task for some reason, however looking at my list of snapshots now, I can see that the jobs ARE being run, but they're still NOT being displayed at all on the autosnap page. This means that I've got no way of modifying/deleting these jobs through the napp-it UI.
When I thought that the jobs weren't being created properly, I tried configuring them multiple times, and now I've ended up with identical jobs that are creating a whole bunch of duplicate snapshots.
1. Does anyone know where napp-it actually saves the autosnap configuration when a job is created? In the short term, I want to just manually edit the jobs to remove the duplicate entries.
2. (This is aimed mainly at _Gea) Any ideas why this is happening in the first place, or perhaps more importantly, how I might fix it?
If it's useful at all, I'm running Solaris 11 (fully updated) and was originally using napp-it 0.8k. When I noticed the problem I tried updating napp-it to the latest version 0.8l3, but I'm still having the same problem.
You should not look on all-in-one and its storage VM like a regular VM. If you use it like you would use a separate SAN (iSCSI or NFS) you are well.
There is also no VMotion/ HA scenario, because you need real hardware/disc access for performance
You should also avoid memory overcommitment with ZFS and with most modern OS's. If you assign RAM to them, they will use it - especially with ZFS and its ARC read-caching.
If you need snapshots of the ZFS OS itself, you have them on ZFS level as boot environments - much better than those on ESXi level
If you need disaster recovery for the SAN VM, you can do on ESXi local datastore level.
(I use hardware Raid-1 enclosures for ESXi and local datastores for that - and remove the mirror after rebuild)
But the storage-VM is not critical.
You can reinstall within half an hour and import the datapool - similar to ESXi.
Gea,
We do have the datastore for the OI VM mirrored. We do not over-commit memory. And we do not need VMotion (at least for the virtual SAN). So how would one backup a boot environment? That's all we probably need. I think what you are saying is that you can copy the BE to offline storage and later could import the BE somehow. Is this different from a zfs import?
Question for all you other all-in-one devotees: what is the best way to back up the OpenIndiana VM that hosts your ZFS SAN when you have used PCI hardware passthrough of the HBA? Using VMDirectPath I/O has the trade off of getting better performance by eliminating the virtual I/O overhead at the cost of preventing the use of the following VMWare capabilities:
Code:1. VMotion 2. High availability 3. Suspend and resume 4. Record and replay 5. Fault tolerance 6. Memory overcommitment and page sharing 7. Hot add/remove of virtual devices 8. No Snapshot backup
For those us using the free version of ESXi 5, we lose the first five of that list anyway. It is the loss of VMWare Snapshot backup that causes our particular problem. We were trying out Veeam Backup and Replication (free version) to back up all our VMs before upgrading the server to ESXi 5.1 and discovered that backing up the OpenIndiana VM failed due to no VMware snapshot capability.
So I'm curious how the experts out there back up their OI VM (assuming they don't have a second ZFS installation where they could do ZFS send).
Thanks in advance,
--peter
This may be possible
http://docs.oracle.com/cd/E19963-01/html/821-1448/ghzvz.html#ghzur
For myself, I hot-unplug a mirrorred boot disk to have working bootdisks (ESXi+OI) in case of problems. Disks are so cheap and any efforts in restoring a BE are too complicated in case of problems where ou need a quick and easy solution to restore ESXi and OI or prepare a update of them.
I'm curious how you were able to install ESXi onto a mirrored array?
_Gea,
for all napp-it+Solaris 11 users: is it a good idea to hold for a while our update enthusiasm for Solaris 11.1 ?
@ Gea
I'm now running OI + Napp-it for more than a year and am now thinking of virtualizing my server!
I have all the needed hardware : SM X9SCM-F, 16GB Ram, E-1230, Norco 4224 case and 3 IBM M1015 flashed HBA cards.
I've been reading your AIO-tutorial and was wondering if I can put all OS : ESXi, Win7,OI and eventually 1 more on a mirrored SSD data store? Or would it better be to use the onboard sata connectors on mobo?
Also I need to export my 2 pools in ZFS and detache all my data drives, right? Cuz ESXi deletes them?
ty
I was planing on using m1015 or maybe m5016 for same purpose (SSD raid1 for esxi, OI and the rest of VMs), so what speed am I to expect if I go with NFS shared VM datastore (on SSD ofcourse) as you suggest?The idea behind all-in-one
- install ESXi on USB or on a sata disk
- use the sata disk as local datastore and install OI on it
- pass-through your SAS controllers to OI
- autostart OI with ESXI and share your pools via NFS and SMB
OI acts as a normal NAS or SAN filer, similar to a barebone server
- mount the NFS share from ESXi
- put your other VM's to this shared NFS storage to have the same NAS/NAS features like with separate barebone server
but all with one box and very fast connectivity up to several GBit (you do not need fast SAN switches, all traffic is ESXi internal in software)
In this scenario, you just need to import your current pools in OI just like you would do with any other pool-move to a new server.
I think it is, but you have to "soft" remove it in CLI i guess, not just unplug it from computer
MAtej
Don't really see an option in Napp-it to remove the disks, only to swap them!
Look for disk errors or timeouts AND check the SMART status of the drives.. I've ran into similar issues and both cases have been a drive that was starting to be flaky.
In the last instance this happened there were no timeouts reported by ZFS, but the drive had a number of reallocated sectors. Replacing the drive fixed all the problems I was having.
Riley
I am using Seagate higher end raid class drives (I buy Constellation ES line for my replacements). I have ST31000340NS and ST31000524NS models right now. I do have one drive removed as it is failed (waiting on the replacement to get here). In addtion I've already replaced 2 other drives in the last year. I don't see any errors on the pool currently (outside of the failed drive of course). Is there somewhere else I should look for errors or timeouts? I looked in /var/adm/messages to see if I could see anything but I think I just saw references to the failed drive. Also I can boot to the Seagate tools if I need to but is there any way to check the SMART status in Solaris 11?
wheelz,
...
One last question relating to your NIC's... you're having a lot of ports there. Do you do Link Aggregation of some sort? If so, try disabling that and just use one wire at a time... Gigabit is usually not easily fully used from a bandwidth point of view. Might help to narrow down the cause.
Please also let us know what OS you run - all in one, standalone, OI, Solaris?
I primarily use iSCSI. I am seeing huge latencies from both ESXi hosts and a Windows host (like 30 seconds or worse). What is strange is it seems to have to build up to it. At first I don't see it after a reboot, but after I have transferred approximately 42 GB to it (suspiciously close to my RAM). After that I get huge latencies that freeze everything up. On the Windows side if I try a 10 GB transfer, it starts at ~130MB/s but then drops down to as low as 10MB/s and many times dies. When it dies I have to disconnect from the iSCSI host and reconnect to get the drive back. Sometimes when it doesn't die it will start going fast (~80MB/s) again after a number of minutes.
On the VMware side the VMs just grind to a halt (even when I only have one VM running) and typically is a crash. Local transfers didn't seem to have a problem though a little slower than ~100MB/s as I was transferring from the slower OS drives. I've done network captures, disabled TSO offloading but then I noticed that when latencies hit from iSCSI then a local transfer dropped too??? Then I tried some limited testing of transfers from CIFs share but that didn't seem to be affected when run by itself.
I'm about to pull out precious hair that I need to keep as much as possible.Does anyone have any ideas?
Hi Gea,
Can you please tell me how/what napp-it uses to send emails?
I want to use the settings that are already there for my NUT (UPS) alerts rather than re-invent the wheel if possible.
Thankyou
Paul
use Net::SMTP;
my $mailer = Net::SMTP->new($server) || &mess("could not connect smtp-server $server");
if (($user ne "") && ($pw ne "")) {
$mailer->auth ($user,$pw);
}
$mailer->mail($from);
$mailer->to($t. "\n") || &mess("send error to $t ");
$mailer->data();
$mailer->datasend("From: $from\n");
$mailer->datasend("To: $to\n");
$mailer->datasend("Subject: $sub\n\n");
$mailer->datasend("$text\n");
$mailer->dataend();
$mailer->quit;
Hoping someone in this thread or maybe Gea can give me a hand.
I have an OI+nappit all-in-one setup. One of my VM guests is a Server 2012 acting as a Domain Controller. I was able to get the OI to join the domain and everything was working great. When I started working on handling power outages and doing graceful shutdowns I noticed that the OI machine had issues on reboot finding the DC. Course the DC would not be up waiting for OI but I assumed that once the DC was up and running OI would be able to find it. That does not seem to be the case. I have to remove the OI box from the DC then re-join the domain from the OI box. Below are some of the errors I see on the OI box:
These happen at bootup before the DC is up, expected but it does not seem idmap ever trys again.s
I wonder how quickly you can reproduce your problem. If it is easy enough to reproduce you could isolate if it is a disk problem or some other problem by using 1-2 spare SAS/SATA ports to plug in 1-2 extra drives that you think have no issues (could be different model/make even). Create a new single disk or mirrored pool with just these drives and then you can share this with iSCSI. Move a test VM onto it and test it to try and reproduce your problem. If it still happens then it is not your main pool disks. Also maybe as a second test try removing one of your ZIL cache drives from its mirror and add it to this mini pool to make sure its not this causing your problems. Not sure if testing moving the L2ARC is worth it or not as well.
One great thing about this kind of testing is you can do it in parallel meaning your main pool can be online and still going and in use while you do this testing. Might be interesting to see if your main pool faults again does the second mini pool have any problems if you test it at that point. This may be an easier way to test it if it is not easy to make it fault on demand.
Michael