mirror of
				https://github.com/xcat2/xcat-core.git
				synced 2025-10-30 19:02:27 +00:00 
			
		
		
		
	This patch refactor the interface to get xcatmaster and tftpserver attributes in the subvars subroutine to improve the performance. partial-issue: #3958
		
			
				
	
	
		
			5385 lines
		
	
	
		
			217 KiB
		
	
	
	
		
			Perl
		
	
	
	
	
	
			
		
		
	
	
			5385 lines
		
	
	
		
			217 KiB
		
	
	
	
		
			Perl
		
	
	
	
	
	
| package xCAT_plugin::esx;
 | |
| 
 | |
| use strict;
 | |
| use warnings;
 | |
| use xCAT::Table;
 | |
| use xCAT::Utils;
 | |
| use xCAT::TableUtils;
 | |
| use xCAT::ServiceNodeUtils;
 | |
| use xCAT::TZUtils;
 | |
| use Time::HiRes qw (sleep);
 | |
| use xCAT::Template;
 | |
| use xCAT::MsgUtils;
 | |
| use xCAT::SvrUtils;
 | |
| use xCAT::NodeRange;
 | |
| use xCAT::Common;
 | |
| use xCAT::VMCommon;
 | |
| use POSIX "WNOHANG";
 | |
| use Getopt::Long;
 | |
| use Thread qw(yield);
 | |
| use POSIX qw(WNOHANG nice);
 | |
| use File::Path qw/mkpath rmtree/;
 | |
| use File::Temp qw/tempdir/;
 | |
| use File::Copy;
 | |
| use Fcntl qw/:flock/;
 | |
| use IO::Socket;    #Need name resolution
 | |
| use Scalar::Util qw/looks_like_number/;
 | |
| 
 | |
| #use Data::Dumper;
 | |
| Getopt::Long::Configure("bundling");
 | |
| Getopt::Long::Configure("pass_through");
 | |
| my @cpiopid;
 | |
| our @ISA = 'xCAT::Common';
 | |
| 
 | |
| 
 | |
| #in xCAT, the lifetime of a process ends on every request
 | |
| #therefore, the lifetime of assignments to these glabals as architected
 | |
| #is to be cleared on every request
 | |
| #my %esx_comm_pids;
 | |
| my %limbonodes;  #nodes in limbo during a forced migration due to missing parent
 | |
| my %hyphash; #A data structure to hold hypervisor-wide variables (i.e. the current resource pool, virtual machine folder, connection object
 | |
| my %vcenterhash; #A data structure to reflect the state of vcenter connectivity to hypervisors
 | |
| my %vmhash;      #store per vm info of interest
 | |
| my %clusterhash;
 | |
| my %hypready; #A structure for hypervisor readiness to be tracked before proceeding to normal operations
 | |
| my %running_tasks;     #A struct to track this processes
 | |
| my $output_handler;    #Pointer to the function to drive results to client
 | |
| my $executerequest;
 | |
| my $usehostnamesforvcenter;
 | |
| my %tablecfg;          #to hold the tables
 | |
| my %hostrefbynode;
 | |
| my $currkey;
 | |
| my $requester;
 | |
| my $viavcenter;
 | |
| my $viavcenterbyhyp;
 | |
| my $vcenterautojoin    = 1;
 | |
| my $datastoreautomount = 1;
 | |
| my $vcenterforceremove = 0;       #used in rmhypervisor
 | |
| my $reconfigreset      = 1;
 | |
| my $vmwaresdkdetect    = eval {
 | |
|     require VMware::VIRuntime;
 | |
|     VMware::VIRuntime->import();
 | |
|     1;
 | |
| };
 | |
| my %lockhandles;
 | |
| 
 | |
| sub recursion_copy {
 | |
|     my $source      = shift;
 | |
|     my $destination = shift;
 | |
|     my $dirhandle;
 | |
|     opendir($dirhandle, $source);
 | |
|     my $entry;
 | |
|     foreach $entry (readdir($dirhandle)) {
 | |
|         if ($entry eq '.' or $entry eq '..') { next; }
 | |
|         my $tempsource      = "$source/$entry";
 | |
|         my $tempdestination = "$destination/$entry";
 | |
|         if (-d $tempsource) {
 | |
|             unless (-d $tempdestination) { mkdir $tempdestination or die "failure creating directory $tempdestination, $!"; }
 | |
|             recursion_copy($tempsource, $tempdestination);
 | |
|         } else {
 | |
|             copy($tempsource, $tempdestination) or die "failed copy from $tempsource to $tempdestination, $!";
 | |
|         }
 | |
|     }
 | |
| }
 | |
| 
 | |
| sub lockbyname {
 | |
|     my $name = shift;
 | |
|     my $lckh;
 | |
|     mkpath("/tmp/xcat/locks/");
 | |
|     while (-e "/tmp/xcat/locks/$name") { sleep 1; }
 | |
|     open($lockhandles{$name}, ">>", "/tmp/xcat/locks/$name");
 | |
|     flock($lockhandles{$name}, LOCK_EX);
 | |
| }
 | |
| 
 | |
| sub unlockbyname {
 | |
|     my $name = shift;
 | |
|     unlink("/tmp/xcat/locks/$name");
 | |
|     close($lockhandles{$name});
 | |
| }
 | |
| 
 | |
| my %guestidmap = (
 | |
|     "rhel.6.*"   => "rhel6_",
 | |
|     "rhel.5.*"   => "rhel5_",
 | |
|     "rhel4.*"    => "rhel4_",
 | |
|     "centos6.*"  => "rhel6_",
 | |
|     "centos5.*"  => "rhel5_",
 | |
|     "centos4.*"  => "rhel4_",
 | |
|     "sles12.*"   => "sles12_",
 | |
|     "sles11.*"   => "sles11_",
 | |
|     "sles10.*"   => "sles10_",
 | |
|     "win2k8"     => "winLonghorn",
 | |
|     "win2k8r2"   => "windows7Server",
 | |
|     "win2012"    => "windows8Server",
 | |
|     "hyperv2012" => "windows8Server",
 | |
|     "esix5.*"    => "vmkernel5",
 | |
|     "esix4.*"    => "vmkernel",
 | |
|     "win8"       => "windows8_",
 | |
|     "win7"       => "windows7_",
 | |
|     "win2k3"     => "winNetStandard",
 | |
|     "imagex"     => "winNetStandard",
 | |
|     "boottarget" => "otherLinux"
 | |
| 
 | |
|       #otherGuest, otherGuest64, otherLinuxGuest, otherLinux64Guest
 | |
| );
 | |
| 
 | |
| sub handled_commands {
 | |
|     return {
 | |
|         copycd         => 'esx',
 | |
|         mknetboot      => "nodetype:os=(esxi.*)",
 | |
|         mkinstall      => "nodetype:os=(esxi[56].*)",
 | |
|         rpower         => 'nodehm:power,mgt',
 | |
|         esxiready      => "esx",
 | |
|         rsetboot       => 'nodehm:power,mgt',
 | |
|         rmigrate       => 'nodehm:power,mgt',
 | |
|         formatdisk     => "nodetype:os=(esxi.*)",
 | |
|         rescansan      => "nodetype:os=(esxi.*)",
 | |
|         mkvm           => 'nodehm:mgt',
 | |
|         rmvm           => 'nodehm:mgt',
 | |
|         clonevm        => 'nodehm:mgt',
 | |
|         createvcluster => 'esx',
 | |
|         lsvcluster     => 'esx',
 | |
|         rmvcluster     => 'esx',
 | |
|         rinv           => 'nodehm:mgt',
 | |
|         chvm           => 'nodehm:mgt',
 | |
|         rshutdown      => "nodetype:os=(esxi.*)",
 | |
|         lsvm           => [ 'hypervisor:type', 'nodetype:os=(esx.*)' ],
 | |
|         rmhypervisor   => [ 'hypervisor:type', 'nodetype:os=(esx.*)' ],
 | |
|         chhypervisor   => [ 'hypervisor:type', 'nodetype:os=(esx.*)' ],
 | |
| 
 | |
|         #lsvm => 'nodehm:mgt', not really supported yet
 | |
|     };
 | |
| }
 | |
| 
 | |
| 
 | |
| 
 | |
| 
 | |
| 
 | |
| sub preprocess_request {
 | |
|     my $request  = shift;
 | |
|     my $callback = shift;
 | |
|     if ($request->{command}->[0] eq 'createvcluster' or $request->{command}->[0] eq 'lsvcluster' or $request->{command}->[0] eq 'rmvcluster') {
 | |
|         return [$request];
 | |
|     }
 | |
| 
 | |
|     #if already preprocessed, go straight to request
 | |
|     if ((defined($request->{_xcatpreprocessed}))
 | |
|         && ($request->{_xcatpreprocessed}->[0] == 1))
 | |
|     {
 | |
|         return [$request];
 | |
|     }
 | |
| 
 | |
|     my $username  = 'root';
 | |
|     my $password  = '';
 | |
|     my $vusername = "Administrator";
 | |
|     my $vpassword = "";
 | |
| 
 | |
|     unless ($request and $request->{command} and $request->{command}->[0]) { return; }
 | |
| 
 | |
|     if ($request->{command}->[0] eq 'copycd')
 | |
|     {    #don't farm out copycd
 | |
|         return [$request];
 | |
|     } elsif ($request->{command}->[0] eq 'mknetboot'
 | |
|         or $request->{command}->[0] eq 'mkinstall') {
 | |
|         return [$request];
 | |
|     }
 | |
|     xCAT::Common::usage_noderange($request, $callback);
 | |
| 
 | |
|     if ($request->{_xcatpreprocessed} and $request->{_xcatpreprocessed}->[0] == 1) { return [$request]; }
 | |
| 
 | |
|     # exit if preprocesses
 | |
|     my @requests;
 | |
| 
 | |
|     my $noderange;
 | |
|     my $command = $request->{command}->[0];
 | |
|     if ($request->{node}) {
 | |
|         $noderange = $request->{node};    # array ref
 | |
|     } elsif ($command eq "esxiready") {
 | |
|         my $node;
 | |
|         ($node) = noderange($request->{'_xcat_clienthost'}->[0]);
 | |
|         $noderange = [$node];
 | |
|         $request->{node} = $noderange;
 | |
|     }
 | |
| 
 | |
|     my $extraargs    = $request->{arg};
 | |
|     my @exargs       = ($request->{arg});
 | |
|     my %hyp_hash     = ();
 | |
|     my %cluster_hash = ();
 | |
| 
 | |
|     # Get nodes from mp table and assign nodes to mp hash.
 | |
|     my $passtab = xCAT::Table->new('passwd');
 | |
|     my $tmp;
 | |
|     if ($passtab) {
 | |
|         ($tmp) = $passtab->getAttribs({ 'key' => 'vmware' }, 'username', 'password');
 | |
|         if (defined($tmp)) {
 | |
|             $username = $tmp->{username};
 | |
|             $password = $tmp->{password};
 | |
|         }
 | |
|         ($tmp) = $passtab->getAttribs({ 'key' => 'vcenter' }, 'username', 'password');
 | |
|         if (defined($tmp)) {
 | |
|             $vusername = $tmp->{username};
 | |
|             $vpassword = $tmp->{password};
 | |
|         }
 | |
|     }
 | |
| 
 | |
|     my $vmtab = xCAT::Table->new("vm");
 | |
|     unless ($vmtab) {
 | |
|         $callback->({ data => ["Cannot open vm table"] });
 | |
|         $request = {};
 | |
|         return;
 | |
|     }
 | |
| 
 | |
|     my $vmtabhash = $vmtab->getNodesAttribs($noderange, [ 'host', 'migrationdest' ]);
 | |
|     foreach my $node (@$noderange) {
 | |
|         if ($command eq "rmhypervisor" or $command eq 'lsvm' or $command eq 'esxiready' or $command eq 'rshutdown' or $command eq "chhypervisor" or $command eq "formatdisk" or $command eq 'rescansan') {
 | |
|             $hyp_hash{$node}{nodes} = [$node];
 | |
|         } else {
 | |
|             my $ent = $vmtabhash->{$node}->[0];
 | |
|             if (defined($ent->{host})) {
 | |
|                 push @{ $hyp_hash{ $ent->{host} }{nodes} }, $node;
 | |
|             } elsif (defined($ent->{migrationdest})) {
 | |
|                 $cluster_hash{ $ent->{migrationdest} }->{nodes}->{$node} = 1;
 | |
|             } else {
 | |
|                 xCAT::SvrUtils::sendmsg([ 1, ": no host or cluster defined for guest" ], $callback, $node);
 | |
|             }
 | |
|         }
 | |
|     }
 | |
| 
 | |
|     # find service nodes for the MMs
 | |
|     # build an individual request for each service node
 | |
|     my $service = "xcat";
 | |
|     my @hyps    = keys(%hyp_hash);
 | |
|     my %targethyps;
 | |
|     if ($command eq 'rmigrate' and (scalar @{$extraargs} >= 1)) {
 | |
|         @ARGV = @{$extraargs};
 | |
|         my $offline;
 | |
|         my $junk;
 | |
|         GetOptions(
 | |
|             "f" => \$offline,
 | |
|             "s=s" => \$junk #wo don't care about it, but suck up nfs:// targets so they don't get added
 | |
|         );
 | |
|         my $dsthyp = $ARGV[0];
 | |
|         if ($dsthyp) {
 | |
|             push @hyps, $dsthyp;
 | |
|             $targethyps{$dsthyp} = 1;
 | |
|         }
 | |
|     }
 | |
| 
 | |
|     #TODO: per hypervisor table password lookup
 | |
|     my @allnodes;
 | |
|     push @allnodes, @hyps;
 | |
|     push @allnodes, @$noderange;
 | |
|     my $sn = xCAT::ServiceNodeUtils->get_ServiceNode(\@allnodes, $service, "MN");
 | |
| 
 | |
|     #vmtabhash was from when we had vm.host do double duty for hypervisor data
 | |
|     #$vmtabhash = $vmtab->getNodesAttribs(\@hyps,['host']);
 | |
|     #We now use hypervisor fields to be unambiguous
 | |
|     my $hyptab     = xCAT::Table->new('hypervisor');
 | |
|     my $hyptabhash = {};
 | |
|     if ($hyptab) {
 | |
|         $hyptabhash = $hyptab->getNodesAttribs(\@hyps, ['mgr']);
 | |
|     }
 | |
| 
 | |
| 
 | |
|     # build each request for each service node
 | |
|     foreach my $snkey (keys %$sn) {
 | |
|         my $reqcopy = {%$request};
 | |
|         $reqcopy->{'_xcatdest'} = $snkey;
 | |
|         $reqcopy->{_xcatpreprocessed}->[0] = 1;
 | |
|         my $hyps1    = $sn->{$snkey};
 | |
|         my @moreinfo = ();
 | |
|         my @nodes    = ();
 | |
|         foreach (@$hyps1) { #This preserves the constructed data to avoid redundant table lookup
 | |
|             my $cfgdata;
 | |
|             if (not $targethyps{$_} and not $hyp_hash{$_}) {    #a vm, skip it
 | |
|                 next;
 | |
|             } elsif ($hyp_hash{$_}{nodes}) {
 | |
|                 push @nodes, @{ $hyp_hash{$_}{nodes} };
 | |
|                 $cfgdata = "[$_][" . join(',', @{ $hyp_hash{$_}{nodes} }) . "][$username][$password][$vusername][$vpassword]"; #TODO: not use vm.host?
 | |
|             } else {
 | |
|                 $cfgdata = "[$_][][$username][$password][$vusername][$vpassword]"; #TODO: not use vm.host?
 | |
|             }
 | |
|             if (defined $hyptabhash->{$_}->[0]->{mgr}) {
 | |
|                 $cfgdata .= "[" . $hyptabhash->{$_}->[0]->{mgr} . "]";
 | |
|             } else {
 | |
|                 $cfgdata .= "[]";
 | |
|             }
 | |
|             push @moreinfo, $cfgdata; #"[$_][".join(',',@{$hyp_hash{$_}{nodes}})."][$username][$password]";
 | |
|         }
 | |
|         foreach (keys %cluster_hash) {
 | |
|             my $cluster;
 | |
|             my $vcenter;
 | |
|             if (/@/) {
 | |
|                 ($cluster, $vcenter) = split /@/, $_, 2;
 | |
|             } else {
 | |
|                 die "TODO: implement default vcenter (for now, user, do vm.migratiodest=cluster" . '@' . "vcentername)";
 | |
|             }
 | |
|             push @moreinfo, "[CLUSTER:$cluster][" . join(',', keys %{ $cluster_hash{$_}->{nodes} }) . "][$username][$password][$vusername][$vpassword][$vcenter]";
 | |
|         }
 | |
|         if (scalar @nodes) {
 | |
|             $reqcopy->{node} = \@nodes;
 | |
|         }
 | |
| 
 | |
|         #print "nodes=@nodes\n";
 | |
|         $reqcopy->{moreinfo} = \@moreinfo;
 | |
|         push @requests, $reqcopy;
 | |
|     }
 | |
|     return \@requests;
 | |
| }
 | |
| 
 | |
| 
 | |
| 
 | |
| sub process_request {
 | |
| 
 | |
|     #$SIG{INT} = $SIG{TERM} = sub{
 | |
|     #	foreach (keys %esx_comm_pids){
 | |
|     #		kill 2,$_;
 | |
|     #	}
 | |
|     #	exit 0;
 | |
|     #};
 | |
| 
 | |
|     my $request = shift;
 | |
|     $output_handler = shift;
 | |
|     $executerequest = shift;
 | |
|     if ($request->{_xcat_authname}->[0]) {
 | |
|         $requester = $request->{_xcat_authname}->[0];
 | |
|     }
 | |
|     %vcenterhash = (); #A data structure to reflect the state of vcenter connectivity to hypervisors
 | |
|     my $level    = shift;
 | |
|     my $distname = undef;
 | |
|     my $arch     = undef;
 | |
|     my $path     = undef;
 | |
|     my $command  = $request->{command}->[0];
 | |
| 
 | |
|     #The first segment is fulfilling the role of this plugin as
 | |
|     #a hypervisor provisioning plugin (akin to anaconda, windows, sles plugins)
 | |
|     if ($command eq 'copycd') {
 | |
|         return copycd($request, $executerequest);
 | |
|     } elsif ($command eq 'mkinstall') {
 | |
|         return mkinstall($request, $executerequest);
 | |
|     } elsif ($command eq 'mknetboot') {
 | |
|         return mknetboot($request, $executerequest);
 | |
|     }
 | |
| 
 | |
|     #From here on out, code for managing guests under VMware
 | |
|     #Detect whether or not the VMware SDK is available on this specific system
 | |
|     unless ($vmwaresdkdetect) {
 | |
|         $vmwaresdkdetect = eval {
 | |
|             require VMware::VIRuntime;
 | |
|             VMware::VIRuntime->import();
 | |
|             1;
 | |
|         };
 | |
|     }
 | |
|     unless ($vmwaresdkdetect) {
 | |
|         xCAT::SvrUtils::sendmsg([ 1, "VMWare SDK required for operation, but not installed" ], $output_handler);
 | |
|         return;
 | |
|     }
 | |
|     if ($command eq 'createvcluster') {
 | |
|         create_new_cluster($request);
 | |
|         return;
 | |
|     }
 | |
|     if ($command eq 'lsvcluster') {
 | |
|         list_clusters($request);
 | |
|         return;
 | |
|     }
 | |
|     if ($command eq 'rmvcluster') {
 | |
|         remove_cluster($request);
 | |
|         return;
 | |
|     }
 | |
| 
 | |
|     my $moreinfo;
 | |
|     my $noderange;
 | |
|     if ($request->{node}) {
 | |
|         $noderange = $request->{node};    # array ref
 | |
|     } elsif ($command eq "esxiready") {
 | |
|         my $node;
 | |
|         ($node) = noderange($request->{'_xcat_clienthost'}->[0]);
 | |
|         $noderange = [$node];
 | |
|     }
 | |
|     xCAT::VMCommon::grab_table_data($noderange, \%tablecfg, $output_handler);
 | |
|     my @exargs;
 | |
|     unless ($command) {
 | |
|         return;                           # Empty request
 | |
|     }
 | |
|     if (ref($request->{arg})) {
 | |
|         @exargs = @{ $request->{arg} };
 | |
|     } else {
 | |
|         @exargs = ($request->{arg});
 | |
|     }
 | |
| 
 | |
|     #pdu commands will be handled in the pdu plugin
 | |
|     if ($command eq "rpower" and grep(/^pduon|pduoff|pdureset|pdustat$/, @exargs)) {
 | |
|         return;
 | |
|     }
 | |
|     #my $sitetab = xCAT::Table->new('site');
 | |
|     #if($sitetab){
 | |
|     #(my $ref) = $sitetab->getAttribs({key => 'usehostnamesforvcenter'}, 'value');
 | |
|     my @entries = xCAT::TableUtils->get_site_attribute("usehostnamesforvcenter");
 | |
|     my $t_entry = $entries[0];
 | |
|     if (defined($t_entry)) {
 | |
|         $usehostnamesforvcenter = $t_entry;
 | |
|     }
 | |
| 
 | |
|     #($ref) = $sitetab->getAttribs({key => 'vcenterautojoin'}, 'value');
 | |
|     @entries = xCAT::TableUtils->get_site_attribute("vcenterautojoin");
 | |
|     $t_entry = $entries[0];
 | |
|     if (defined($t_entry)) {
 | |
|         $vcenterautojoin = $t_entry;
 | |
|         if ($vcenterautojoin =~ /^n/ or $vcenterautojoin =~ /^dis/) {
 | |
|             $vcenterautojoin = 0;
 | |
|         }
 | |
|     }
 | |
| 
 | |
|     #($ref) = $sitetab->getAttribs({key => 'vmwaredatastoreautomount'}, 'value');
 | |
|     @entries = xCAT::TableUtils->get_site_attribute("vmwaredatastoreautomount");
 | |
|     $t_entry = $entries[0];
 | |
|     if (defined($t_entry)) {
 | |
|         $datastoreautomount = $t_entry;
 | |
|         if ($datastoreautomount =~ /^n/ or $datastoreautomount =~ /^dis/) {
 | |
|             $datastoreautomount = 0;
 | |
|         }
 | |
|     }
 | |
| 
 | |
|     #($ref) = $sitetab->getAttribs({key => 'vmwarereconfigonpower'},'value');
 | |
|     @entries = xCAT::TableUtils->get_site_attribute("vmwarereconfigonpower");
 | |
|     $t_entry = $entries[0];
 | |
|     if (defined($t_entry)) {
 | |
|         $reconfigreset = $t_entry;
 | |
|         if ($reconfigreset =~ /^(n|d)/i) {    #if no or disable, skip it
 | |
|             $reconfigreset = 0;
 | |
|         }
 | |
|     }
 | |
| 
 | |
|     #	}
 | |
| 
 | |
| 
 | |
|     if ($request->{moreinfo}) { $moreinfo = $request->{moreinfo}; }
 | |
|     else { $moreinfo = build_more_info($noderange, $output_handler); }
 | |
|     foreach my $info (@$moreinfo) {
 | |
|         $info =~ /^\[(.*?)\]\[(.*?)\]\[(.*?)\]\[(.*?)\]\[(.*?)\]\[(.*?)\]\[(.*?)\]/;
 | |
|         my $hyp        = $1;
 | |
|         my @nodes      = split(',', $2);
 | |
|         my $username   = $3;
 | |
|         my $password   = $4;
 | |
|         my $tmpvcname  = $7;
 | |
|         my $tmpvcuname = $5;
 | |
|         my $tmpvcpass  = $6;
 | |
|         if ($hyp =~ /^CLUSTER:/) {    #a cluster, not a host.
 | |
|             $hyp =~ s/^CLUSTER://;
 | |
|             $clusterhash{$hyp}->{vcenter}->{name}     = $tmpvcname;
 | |
|             $clusterhash{$hyp}->{vcenter}->{username} = $tmpvcuname;
 | |
|             $clusterhash{$hyp}->{vcenter}->{password} = $tmpvcpass;
 | |
|             foreach (@nodes) {
 | |
|                 $clusterhash{$hyp}->{nodes}->{$_} = 1;
 | |
|             }
 | |
|             next;
 | |
|         }
 | |
|         $hyphash{$hyp}->{vcenter}->{name}     = $tmpvcname;
 | |
|         $hyphash{$hyp}->{vcenter}->{username} = $tmpvcuname;
 | |
|         $hyphash{$hyp}->{vcenter}->{password} = $tmpvcpass;
 | |
|         $hyphash{$hyp}->{username}            = $username;     # $nodeid;
 | |
|         $hyphash{$hyp}->{password}            = $password;     # $nodeid;
 | |
|         unless ($hyphash{$hyp}->{vcenter}->{password}) {
 | |
|             $hyphash{$hyp}->{vcenter}->{password} = "";
 | |
|         }
 | |
|         my $ent;
 | |
|         for (my $i = 0 ; $i < @nodes ; $i++) {
 | |
|             if ($command eq 'rmigrate' and grep /-f/, @exargs) { #offline migration,
 | |
|                 $hyphash{$hyp}->{offline} = 1; #if it is migrate and it has nodes, it is a source hypervisor apt to be offline
 | |
|                                                #this will hint to relevant code to operate under the assumption of a
 | |
|                                                #downed hypervisor source
 | |
|                   #note this will make dangerous assumptions, it will make a very minimal attempt
 | |
|                   #to operate normally, but really should only be called if the source is down and
 | |
|                   #fenced (i.e. storage, network, or turned off and stateless
 | |
|             }
 | |
|             my $node = $nodes[$i];
 | |
| 
 | |
|             #my $nodeid = $ids[$i];
 | |
|             $hyphash{$hyp}->{nodes}->{$node} = 1;    # $nodeid;
 | |
|         }
 | |
|     }
 | |
|     my $hyptab = xCAT::Table->new('hypervisor', create => 0);
 | |
|     if ($hyptab) {
 | |
|         my @hyps = keys %hyphash;
 | |
|         $tablecfg{hypervisor} = $hyptab->getNodesAttribs(\@hyps, [ 'mgr', 'netmap', 'defaultnet', 'cluster', 'preferdirect', 'datacenter' ]);
 | |
|     }
 | |
|     my $hoststab = xCAT::Table->new('hosts', create => 0);
 | |
|     if ($hoststab) {
 | |
|         my @hyps = keys %hyphash;
 | |
|         $tablecfg{hosts} = $hoststab->getNodesAttribs(\@hyps, ['hostnames']);
 | |
| 
 | |
|     }
 | |
| 
 | |
|     #my $children = 0;
 | |
|     #my $vmmaxp = 84;
 | |
|     #$SIG{CHLD} = sub { my $cpid; while ($cpid = waitpid(-1, WNOHANG) > 0) { delete $esx_comm_pids{$cpid}; $children--; } };
 | |
|     $viavcenter = 0;
 | |
|     if ($command eq 'rmigrate' or $command eq 'rmhypervisor') { #Only use vcenter when required, fewer prereqs
 | |
|         $viavcenter = 1;
 | |
|     }
 | |
|     if ($command eq 'rmhypervisor' and grep /-f/, @exargs) { #force remove of hypervisor
 | |
|         $vcenterforceremove = 1;
 | |
|     }
 | |
|     my $keytab = xCAT::Table->new('prodkey');
 | |
|     if ($keytab) {
 | |
|         my @hypes = keys %hyphash;
 | |
|         $tablecfg{prodkey} = $keytab->getNodesAttribs(\@hypes, [qw/product key/]);
 | |
|     }
 | |
|     my $hyp;
 | |
|     my %needvcentervalidation;
 | |
|     my $cluster;
 | |
|     foreach $cluster (keys %clusterhash) {
 | |
|         my $vcenter = $clusterhash{$cluster}->{vcenter}->{name};
 | |
|         unless ($vcenterhash{$vcenter}->{conn}) {
 | |
|             eval {
 | |
|                 $vcenterhash{$vcenter}->{conn} = Vim->new(service_url => "https://$vcenter/sdk");
 | |
|                 $vcenterhash{$vcenter}->{conn}->login(user_name => $clusterhash{$cluster}->{vcenter}->{username},
 | |
|                     password => $clusterhash{$cluster}->{vcenter}->{password});
 | |
|             };
 | |
|             if ($@) {
 | |
|                 $vcenterhash{$vcenter}->{conn} = undef;
 | |
|                 xCAT::SvrUtils::sendmsg([ 1, "Unable to reach $vcenter vCenter server to manage cluster $cluster: $@" ], $output_handler);
 | |
|                 next;
 | |
|             }
 | |
|             my $clusternode;
 | |
|         }
 | |
|         $clusterhash{$cluster}->{conn} = $vcenterhash{$vcenter}->{conn};
 | |
|         foreach my $clusternode (keys %{ $clusterhash{$cluster}->{nodes} }) {
 | |
|             $vmhash{$clusternode}->{conn} = $vcenterhash{$vcenter}->{conn};
 | |
|         }
 | |
|     }
 | |
|     foreach $hyp (sort(keys %hyphash)) {
 | |
| 
 | |
|         #if($pid == 0){
 | |
|         if ($viavcenter or (defined $tablecfg{hypervisor}->{$hyp}->[0]->{mgr} and not $tablecfg{hypervisor}->{$hyp}->[0]->{preferdirect})) {
 | |
|             $viavcenterbyhyp->{$hyp} = 1;
 | |
|             $hypready{$hyp} = 0; #This hypervisor requires a flag be set to signify vCenter sanenes before proceeding
 | |
|             my $vcenter = $hyphash{$hyp}->{vcenter}->{name};
 | |
|             unless ($vcenterhash{$vcenter}->{conn}) {
 | |
|                 eval {
 | |
|                     $vcenterhash{$vcenter}->{conn} =
 | |
|                       Vim->new(service_url => "https://$vcenter/sdk");
 | |
|                     $vcenterhash{$vcenter}->{conn}->login(
 | |
|                         user_name => $hyphash{$hyp}->{vcenter}->{username},
 | |
|                         password  => $hyphash{$hyp}->{vcenter}->{password}
 | |
|                     );
 | |
|                 };
 | |
|                 if ($@) {
 | |
|                     $vcenterhash{$vcenter}->{conn} = undef;
 | |
|                     xCAT::SvrUtils::sendmsg([ 1, "Unable to reach $vcenter vCenter server to manage $hyp: $@" ], $output_handler);
 | |
|                     next;
 | |
|                 }
 | |
|             }
 | |
|             my $hypnode;
 | |
|             foreach $hypnode (keys %{ $hyphash{$hyp}->{nodes} }) {
 | |
|                 $vmhash{$hypnode}->{conn} = $vcenterhash{ $hyphash{$hyp}->{vcenter}->{name} }->{conn};
 | |
|             }
 | |
|             $hyphash{$hyp}->{conn} = $vcenterhash{ $hyphash{$hyp}->{vcenter}->{name} }->{conn};
 | |
|             $hyphash{$hyp}->{vcenter}->{conn} = $vcenterhash{ $hyphash{$hyp}->{vcenter}->{name} }->{conn};
 | |
|             $needvcentervalidation{$hyp} = $vcenter;
 | |
|             $vcenterhash{$vcenter}->{allhyps}->{$hyp} = 1;
 | |
|         } else {
 | |
|             eval {
 | |
|                 $hyphash{$hyp}->{conn} = Vim->new(service_url => "https://$hyp/sdk");
 | |
|                 $hyphash{$hyp}->{conn}->login(user_name => $hyphash{$hyp}->{username}, password => $hyphash{$hyp}->{password});
 | |
|             };
 | |
|             if ($@) {
 | |
|                 $hyphash{$hyp}->{conn} = undef;
 | |
|                 xCAT::SvrUtils::sendmsg([ 1, "Unable to reach $hyp to perform operation due to $@" ], $output_handler);
 | |
|                 $hypready{$hyp} = -1;
 | |
|                 next;
 | |
|             }
 | |
|             my $localnode;
 | |
|             foreach $localnode (keys %{ $hyphash{$hyp}->{nodes} }) {
 | |
|                 $vmhash{$localnode}->{conn} = $hyphash{$hyp}->{conn};
 | |
|             }
 | |
|             validate_licenses($hyp);
 | |
|         }
 | |
| 
 | |
|         #}else{
 | |
|         #	$esx_comm_pids{$pid} = 1;
 | |
|         #}
 | |
|     }
 | |
|     foreach $hyp (keys %needvcentervalidation) {
 | |
|         my $vcenter = $needvcentervalidation{$hyp};
 | |
|         if (not defined $vcenterhash{$vcenter}->{hostviews}) {
 | |
|             populate_vcenter_hostviews($vcenter);
 | |
|         }
 | |
|         if (validate_vcenter_prereqs($hyp, \&declare_ready, {
 | |
|                     hyp     => $hyp,
 | |
|                     vcenter => $vcenter
 | |
|                 }) eq "failed") {
 | |
|             $hypready{$hyp} = -1;
 | |
|         }
 | |
|     }
 | |
|     while (grep { $_ == 0 } values %hypready) {
 | |
|         wait_for_tasks();
 | |
|         sleep(1); #We'll check back in every second.  Unfortunately, we have to poll since we are in web service land
 | |
|     }
 | |
|     my @badhypes;
 | |
|     if (grep { $_ == -1 } values %hypready) {
 | |
|         foreach (keys %hypready) {
 | |
|             if ($hypready{$_} == -1) {
 | |
|                 unless ($hyphash{$_}->{offline}) {
 | |
|                     push @badhypes, $_;
 | |
|                 }
 | |
|                 my @relevant_nodes = sort (keys %{ $hyphash{$_}->{nodes} });
 | |
|                 my $sadhypervisor  = $_;
 | |
|                 foreach (@relevant_nodes) {
 | |
|                     if ($command eq "rmigrate" and grep /-f/, @exargs) { $limbonodes{$_} = $needvcentervalidation{$sadhypervisor}; } else {
 | |
|                         xCAT::SvrUtils::sendmsg([ 1, ": hypervisor unreachable" ], $output_handler, $_);
 | |
|                     }
 | |
|                     if ($command eq "rpower" and grep /stat/, @exargs) { $limbonodes{$_} = $needvcentervalidation{$sadhypervisor}; } #try to stat power anyway through vcenter of interest...
 | |
|                 }
 | |
|                 delete $hyphash{$_};
 | |
|             }
 | |
|         }
 | |
|         if (@badhypes) {
 | |
|             xCAT::SvrUtils::sendmsg([ 1, ": The following hypervisors failed to become ready for the operation: " . join(',', @badhypes) ], $output_handler);
 | |
|         }
 | |
|     }
 | |
|     do_cmd($command, @exargs);
 | |
|     foreach (@badhypes) { delete $hyphash{$_}; }
 | |
|     foreach my $vm (sort(keys %vmhash)) {
 | |
|         $vmhash{$vm}->{conn}->logout();
 | |
|     }
 | |
| }
 | |
| 
 | |
| sub validate_licenses {
 | |
|     my $hyp = shift;
 | |
|     my $conn = $hyphash{$hyp}->{conn}; #This can't possibly be called via a cluster stack, so hyphash is appropriate here
 | |
|     unless ($tablecfg{prodkey}->{$hyp}) {    #if no license specified, no-op
 | |
|         return;
 | |
|     }
 | |
|     my $hv = get_hostview(hypname => $hyp, conn => $conn, properties => [ 'configManager', 'name' ]);
 | |
|     my $lm = $conn->get_view(mo_ref => $hv->configManager->licenseManager);
 | |
|     my @licenses;
 | |
|     foreach (@{ $lm->licenses }) {
 | |
|         push @licenses, uc($_->licenseKey);
 | |
|     }
 | |
|     my @newlicenses;
 | |
|     foreach (@{ $tablecfg{prodkey}->{$hyp} }) {
 | |
|         if (defined($_->{product}) and $_->{product} eq 'esx') {
 | |
|             my $key = uc($_->{key});
 | |
|             unless (grep /$key/, @licenses) {
 | |
|                 push @newlicenses, $key;
 | |
|             }
 | |
|         }
 | |
|     }
 | |
|     foreach (@newlicenses) {
 | |
|         $lm->UpdateLicense(licenseKey => $_);
 | |
|     }
 | |
| }
 | |
| 
 | |
| sub do_cmd {
 | |
|     my $command = shift;
 | |
|     my @exargs  = @_;
 | |
|     if ($command eq 'esxiready') {
 | |
|         return;
 | |
|     }
 | |
|     if ($command eq 'rpower') {
 | |
|         generic_vm_operation([ 'config.name', 'config.guestId', 'config.hardware.memoryMB', 'config.hardware.numCPU', 'runtime.powerState', 'runtime.host' ], \&power, @exargs);
 | |
|     } elsif ($command eq 'rmvm') {
 | |
|         generic_vm_operation([ 'config.name', 'runtime.powerState', 'runtime.host' ], \&rmvm, @exargs);
 | |
|     } elsif ($command eq 'rsetboot') {
 | |
|         generic_vm_operation([ 'config.name', 'runtime.host' ], \&setboot, @exargs);
 | |
|     } elsif ($command eq 'rinv') {
 | |
|         generic_vm_operation([ 'config.name', 'config', 'runtime.host', 'layoutEx' ], \&inv, @exargs);
 | |
|     } elsif ($command eq 'formatdisk') {
 | |
|         generic_hyp_operation(\&formatdisk, @exargs);
 | |
|     } elsif ($command eq 'rescansan') {
 | |
|         generic_hyp_operation(\&rescansan, @exargs);
 | |
|     } elsif ($command eq 'rmhypervisor') {
 | |
|         generic_hyp_operation(\&rmhypervisor, @exargs);
 | |
|     } elsif ($command eq 'rshutdown') {
 | |
|         generic_hyp_operation(\&rshutdown, @exargs);
 | |
|     } elsif ($command eq 'chhypervisor') {
 | |
|         generic_hyp_operation(\&chhypervisor, @exargs);
 | |
|     } elsif ($command eq 'lsvm') {
 | |
|         generic_hyp_operation(\&lsvm, @exargs);
 | |
|     } elsif ($command eq 'clonevm') {
 | |
|         generic_hyp_operation(\&clonevms, @exargs);
 | |
|     } elsif ($command eq 'mkvm') {
 | |
|         generic_hyp_operation(\&mkvms, @exargs);
 | |
|     } elsif ($command eq 'chvm') {
 | |
|         generic_vm_operation([ 'config.name', 'config', 'runtime.host' ], \&chvm, @exargs);
 | |
| 
 | |
|         #generic_hyp_operation(\&chvm,@exargs);
 | |
|     } elsif ($command eq 'rmigrate') { #Technically, on a host view, but vcenter path is 'weirder'
 | |
|         generic_hyp_operation(\&migrate, @exargs);
 | |
|     }
 | |
|     wait_for_tasks();
 | |
|     if ($command eq 'clonevm') {  #TODO: unconditional, remove mkvms hosted copy
 | |
|         my @dhcpnodes;
 | |
|         foreach (keys %{ $tablecfg{dhcpneeded} }) {
 | |
|             push @dhcpnodes, $_;
 | |
|             delete $tablecfg{dhcpneeded}->{$_};
 | |
|         }
 | |
|         unless ($::XCATSITEVALS{'dhcpsetup'} and ($::XCATSITEVALS{'dhcpsetup'} =~ /^n/i or $::XCATSITEVALS{'dhcpsetup'} =~ /^d/i or $::XCATSITEVALS{'dhcpsetup'} eq '0')) {
 | |
|             $executerequest->({ command => ['makedhcp'], node => \@dhcpnodes });
 | |
|         }
 | |
|     }
 | |
| }
 | |
| 
 | |
| #inventory request for esx
 | |
| sub inv {
 | |
|     my %args = @_;
 | |
|     my $node = $args{node};
 | |
|     my $hyp  = $args{hyp};
 | |
|     if (not defined $args{vmview}) {    #attempt one refresh
 | |
|         $args{vmview} = $vmhash{$node}->{conn}->find_entity_view(view_type => 'VirtualMachine', properties => [ 'config.name', 'runtime.powerState' ], filter => { name => $node });
 | |
|         if (not defined $args{vmview}) {
 | |
|             xCAT::SvrUtils::sendmsg([ 1, "VM does not appear to exist" ], $output_handler, $node);
 | |
|             return;
 | |
|         }
 | |
|     }
 | |
|     if (not $args{vmview}->{config}) {
 | |
|         xCAT::SvrUtils::sendmsg([ 1, "VM is in an invalid state" ], $output_handler, $node);
 | |
|         return;
 | |
|     }
 | |
| 
 | |
|     @ARGV = @{ $args{exargs} };
 | |
|     require Getopt::Long;
 | |
|     my $tableUpdate;
 | |
|     my $rc = GetOptions(
 | |
|         't' => \$tableUpdate,
 | |
|     );
 | |
|     $SIG{__WARN__} = 'DEFAULT';
 | |
| 
 | |
|     if (@ARGV > 1) {
 | |
|         xCAT::SvrUtils::sendmsg("Invalid arguments:  @ARGV", $output_handler);
 | |
|         return;
 | |
|     }
 | |
| 
 | |
|     if (!$rc) {
 | |
|         return;
 | |
|     }
 | |
| 
 | |
|     my $vmview = $args{vmview};
 | |
|     my $moref  = $vmview->{mo_ref}->value;
 | |
|     xCAT::SvrUtils::sendmsg("Managed Object Reference: $moref", $output_handler, $node);
 | |
|     my $uuid = $vmview->config->uuid;
 | |
|     $uuid =~ s/(..)(..)(..)(..)-(..)(..)-(..)(..)/$4$3$2$1-$6$5-$8$7/;
 | |
|     xCAT::SvrUtils::sendmsg("UUID/GUID:  $uuid", $output_handler, $node);
 | |
|     my $cpuCount = $vmview->config->hardware->numCPU;
 | |
|     xCAT::SvrUtils::sendmsg("CPUs:  $cpuCount", $output_handler, $node);
 | |
|     my $memory = $vmview->config->hardware->memoryMB;
 | |
|     xCAT::SvrUtils::sendmsg("Memory:  $memory MB", $output_handler, $node);
 | |
|     my %updatehash = (cpus => $cpuCount, memory => $memory);
 | |
| 
 | |
| 
 | |
|     my $devices = $vmview->config->hardware->device;
 | |
|     my $label;
 | |
|     my $size;
 | |
|     my $fileName;
 | |
|     my $device;
 | |
|     if ($tableUpdate and $hyp) {
 | |
|         validate_datastore_prereqs([$node], $hyp); #need datastoremaps to verify names...
 | |
|     }
 | |
|     my %vmstorageurls;
 | |
|     foreach $device (@$devices) {
 | |
|         $label = $device->deviceInfo->label;
 | |
| 
 | |
|         if ($label =~ /^Hard disk/) {
 | |
|             $label .= " (d" . $device->controllerKey . ":" . $device->unitNumber . ")";
 | |
|             $size     = $device->capacityInKB / 1024;
 | |
|             $fileName = $device->backing->fileName;
 | |
|             $output_handler->({
 | |
|                     node => {
 | |
|                         name => $node,
 | |
|                         data => {
 | |
|                             desc     => $label,
 | |
|                             contents => "$size MB @ $fileName"
 | |
|                           }
 | |
|                       }
 | |
|             });
 | |
| 
 | |
|             #if ($tableUpdate) {
 | |
|             #		$fileName =~ /\[([^\]]+)\]/;
 | |
|             #	$vmstorageurls{$hyphash{$hyp}->{datastoreurlmap}->{$1}}=1;
 | |
|             #}
 | |
|         } elsif ($label =~ /Network/) {
 | |
|             xCAT::SvrUtils::sendmsg("$label: " . $device->macAddress, $output_handler, $node);
 | |
|         }
 | |
|     }
 | |
|     if ($tableUpdate) {
 | |
|         my $cfgdatastore;
 | |
|         foreach (@{ $vmview->layoutEx->file }) {
 | |
| 
 | |
|             #TODO, track ALL layoutEx->file....
 | |
|             if ($_->type eq 'config') {
 | |
|                 $_->name =~ /\[([^\]]+)\]/;
 | |
|                 $cfgdatastore = $hyphash{$hyp}->{datastoreurlmap}->{$1};
 | |
|                 last;
 | |
|             }
 | |
|         }
 | |
|         my $cfgkey;
 | |
|         if ($tablecfg{vm}->{$node}->[0]->{cfgstore}) { #check the config file explicitly, ignore the rest
 | |
|             $cfgkey = 'cfgstore';
 | |
|         } elsif ($tablecfg{vm}->{$node}->[0]->{storage}) { #check the config file explicitly, ignore the rest
 | |
|             $cfgkey = 'storage';
 | |
|         }
 | |
|         my $configuration = $tablecfg{vm}->{$node}->[0]->{$cfgkey}; #TODO: prune urls that map to no layoutEx->file entries anymore
 | |
|         my $configappend  = $configuration;
 | |
|         $configappend =~ s/^[^,=]*//;
 | |
|         $tablecfg{vm}->{$node}->[0]->{$cfgkey} =~ m!nfs://([^/]+)/!;
 | |
|         my $tablecfgserver = $1;
 | |
|         my $cfgserver      = inet_aton($tablecfgserver);
 | |
|         if ($cfgserver) {
 | |
|             $cfgserver = inet_ntoa($cfgserver); #get the IP address (TODO: really need to wrap getaddrinfo this handily...
 | |
|             my $cfgurl = $tablecfg{vm}->{$node}->[0]->{$cfgkey};
 | |
|             $cfgurl =~ s/$tablecfgserver/$cfgserver/;
 | |
|             if ($cfgurl ne $cfgdatastore) {
 | |
|                 $updatehash{$cfgkey} = $cfgdatastore . $configappend;
 | |
|             }
 | |
|         }
 | |
|     }
 | |
|     if ($tableUpdate) {
 | |
|         my $vm = xCAT::Table->new('vm', -create => 1);
 | |
|         $vm->setNodeAttribs($node, \%updatehash);
 | |
|     }
 | |
| 
 | |
| }
 | |
| 
 | |
| 
 | |
| #changes the memory, number of cpus and device size
 | |
| #can also add,resize and remove disks
 | |
| sub chvm {
 | |
|     my %args = @_;
 | |
|     my $node = $args{node};
 | |
|     my $hyp  = $args{hyp};
 | |
|     if (not defined $args{vmview}) {    #attempt one refresh
 | |
|         $args{vmview} = $vmhash{$node}->{conn}->find_entity_view(view_type => 'VirtualMachine',
 | |
|             properties => [ 'config.name', 'runtime.powerState' ],
 | |
|             filter => { name => $node });
 | |
|         if (not defined $args{vmview}) {
 | |
|             xCAT::SvrUtils::sendmsg([ 1, "VM does not appear to exist" ], $output_handler, $node);
 | |
|             return;
 | |
|         }
 | |
|     }
 | |
|     @ARGV = @{ $args{exargs} };
 | |
|     my @deregister;
 | |
|     my @purge;
 | |
|     my @add;
 | |
|     my %resize;
 | |
|     my $cpuCount;
 | |
|     my $memory;
 | |
|     my $vmview = $args{vmview};
 | |
| 
 | |
|     require Getopt::Long;
 | |
|     $SIG{__WARN__} = sub {
 | |
|         xCAT::SvrUtils::sendmsg([ 1, "Could not parse options, " . shift() ], $output_handler);
 | |
|     };
 | |
|     my @otherparams;
 | |
|     my $cdrom;
 | |
|     my $eject;
 | |
|     my $rc = GetOptions(
 | |
|         "d=s"               => \@deregister,
 | |
|         "p=s"               => \@purge,
 | |
|         "a=s"               => \@add,
 | |
|         "o=s"               => \@otherparams,
 | |
|         "resize=s%"         => \%resize,
 | |
|         "optical|cdrom|c=s" => \$cdrom,
 | |
|         "eject"             => \$eject,
 | |
|         "cpus=s"            => \$cpuCount,
 | |
|         "mem=s"             => \$memory
 | |
|     );
 | |
|     $SIG{__WARN__} = 'DEFAULT';
 | |
| 
 | |
|     if (@ARGV) {
 | |
|         xCAT::SvrUtils::sendmsg("Invalid arguments:  @ARGV", $output_handler);
 | |
|         return;
 | |
|     }
 | |
| 
 | |
|     if (!$rc) {
 | |
|         return;
 | |
|     }
 | |
| 
 | |
|     #use Data::Dumper;
 | |
|     #xCAT::SvrUtils::sendmsg("dereg = ".Dumper(\@deregister));
 | |
|     #xCAT::SvrUtils::sendmsg("purge = ".Dumper(\@purge));
 | |
|     #xCAT::SvrUtils::sendmsg("add = ".Dumper(\@add));
 | |
|     #xCAT::SvrUtils::sendmsg("resize = ".Dumper(\%resize));
 | |
|     #xCAT::SvrUtils::sendmsg("cpus = $cpuCount");
 | |
|     #xCAT::SvrUtils::sendmsg("mem = ".getUnits($memory,"K",1024));
 | |
| 
 | |
| 
 | |
|     my %conargs;
 | |
|     if ($cpuCount) {
 | |
|         if ($cpuCount =~ /^\+(\d+)/) {
 | |
|             $cpuCount = $vmview->config->hardware->numCPU + $1;
 | |
|         } elsif ($cpuCount =~ /^-(\d+)/) {
 | |
|             $cpuCount = $vmview->config->hardware->numCPU - $1;
 | |
|         }
 | |
|         $conargs{numCPUs} = $cpuCount;
 | |
|     }
 | |
| 
 | |
|     if ($memory) {
 | |
|         if ($memory =~ /^\+(.+)/) {
 | |
|             $conargs{memoryMB} = $vmview->config->hardware->memoryMB + getUnits($1, "M", 1048576);
 | |
|         } elsif ($memory =~ /^-(\d+)/) {
 | |
|             $conargs{memoryMB} = $vmview->config->hardware->memoryMB - getUnits($1, "M", 1048576);
 | |
|         } else {
 | |
|             $conargs{memoryMB} = getUnits($memory, "M", 1048576);
 | |
|         }
 | |
|     }
 | |
| 
 | |
|     my $disk;
 | |
|     my $devices = $vmview->config->hardware->device;
 | |
|     my $label;
 | |
|     my $device;
 | |
|     my $cmdLabel;
 | |
|     my $newSize;
 | |
|     my @devChanges;
 | |
| 
 | |
|     if (@deregister) {
 | |
|         for $disk (@deregister) {
 | |
|             $device = getDiskByLabel($disk, $devices);
 | |
|             unless ($device) {
 | |
|                 xCAT::SvrUtils::sendmsg([ 1, "Disk:  $disk does not exist" ], $output_handler, $node);
 | |
|                 return;
 | |
|             }
 | |
| 
 | |
|             #xCAT::SvrUtils::sendmsg(Dumper($device));
 | |
|             push @devChanges, VirtualDeviceConfigSpec->new(
 | |
|                 device    => $device,
 | |
|                 operation => VirtualDeviceConfigSpecOperation->new('remove'));
 | |
| 
 | |
|         }
 | |
|     }
 | |
| 
 | |
|     if (@purge) {
 | |
|         for $disk (@purge) {
 | |
|             $device = getDiskByLabel($disk, $devices);
 | |
|             unless ($device) {
 | |
|                 xCAT::SvrUtils::sendmsg([ 1, "Disk:  $disk does not exist" ], $output_handler, $node);
 | |
|                 return;
 | |
|             }
 | |
| 
 | |
|             #xCAT::SvrUtils::sendmsg(Dumper($device));
 | |
|             push @devChanges, VirtualDeviceConfigSpec->new(
 | |
|                 device    => $device,
 | |
|                 operation => VirtualDeviceConfigSpecOperation->new('remove'),
 | |
|                 fileOperation => VirtualDeviceConfigSpecFileOperation->new('destroy'));
 | |
| 
 | |
|         }
 | |
| 
 | |
|     }
 | |
| 
 | |
|     if (@add) {
 | |
|         my $addSizes = join(',', @add);
 | |
|         my $scsiCont;
 | |
|         my $scsiUnit;
 | |
|         my $ideCont;
 | |
|         my $ideUnit;
 | |
|         my $label;
 | |
|         my $idefull  = 0;
 | |
|         my $scsifull = 0;
 | |
|         foreach $device (@$devices) {
 | |
|             $label = $device->deviceInfo->label;
 | |
|             if ($label =~ /^SCSI controller/) {
 | |
|                 my $tmpu = getAvailUnit($device->{key}, $devices, maxnum => 15);
 | |
|                 if ($tmpu > 0) {
 | |
|                     $scsiCont = $device;
 | |
|                     $scsiUnit = $tmpu;
 | |
|                 } else {
 | |
|                     $scsifull = 1;
 | |
|                 }
 | |
| 
 | |
|                 #ignore scsiControllers that are full, problem still remains if trying to add across two controllers in one go
 | |
|             }
 | |
|             if ($label =~ /^IDE/ and not $ideCont) {
 | |
|                 my $tmpu = getAvailUnit($device->{key}, $devices, maxnum => 1);
 | |
|                 if ($tmpu >= 0) {
 | |
|                     $ideCont = $device;
 | |
|                     $ideUnit = $tmpu;
 | |
|                 } elsif ($device->{key} == 201) {
 | |
|                     $idefull = 1;
 | |
|                 }
 | |
|             }
 | |
|         }
 | |
|         unless ($hyphash{$hyp}->{datastoremap}) { validate_datastore_prereqs([], $hyp); }
 | |
|         push @devChanges, create_storage_devs($node, $hyphash{$hyp}->{datastoremap}, $addSizes, $scsiCont, $scsiUnit, $ideCont, $ideUnit, $devices, idefull => $idefull, scsifull => $scsifull);
 | |
|     }
 | |
| 
 | |
|     if ($cdrom or $eject) {
 | |
|         my $opticalbackingif;
 | |
|         my $opticalconnectable;
 | |
|         if ($cdrom) {
 | |
|             my $storageurl;
 | |
|             if ($cdrom =~ m!://!) {
 | |
|                 $storageurl = $cdrom;
 | |
|                 $storageurl =~ s!/[^/]*\z!!;
 | |
|                 unless (validate_datastore_prereqs([], $hyp, { $storageurl => [$node] })) {
 | |
|                     xCAT::SvrUtils::sendmsg([ 1, "Unable to find/mount datastore holding $cdrom" ], $output_handler, $node);
 | |
|                     return;
 | |
|                 }
 | |
|                 $cdrom =~ s!.*/!!;
 | |
|             } else {
 | |
|                 $storageurl = $tablecfg{vm}->{$node}->[0]->{storage};
 | |
|                 $storageurl =~ s/=.*//;
 | |
|                 $storageurl =~ s/.*,//;
 | |
|                 $storageurl =~ s/\/\z//;
 | |
|             }
 | |
|             $opticalbackingif = VirtualCdromIsoBackingInfo->new(fileName => "[" . $hyphash{$hyp}->{datastoremap}->{$storageurl} . "] $cdrom");
 | |
|             $opticalconnectable = VirtualDeviceConnectInfo->new(startConnected => 1, allowGuestControl => 1, connected => 1);
 | |
|         } elsif ($eject) {
 | |
|             $opticalbackingif = VirtualCdromRemoteAtapiBackingInfo->new(deviceName => "");
 | |
|             $opticalconnectable = VirtualDeviceConnectInfo->new(startConnected => 0, allowGuestControl => 1, connected => 0);
 | |
|         }
 | |
|         my $oldcd;
 | |
|         foreach my $dev (@$devices) {
 | |
|             if ($dev->deviceInfo->label eq "CD/DVD drive 1") {
 | |
|                 $oldcd = $dev;
 | |
|                 last;
 | |
|             }
 | |
|         }
 | |
|         unless ($oldcd) {
 | |
|             if ($cdrom) {
 | |
|                 xCAT::SvrUtils::sendmsg([ 1, "Unable to find Optical drive in VM to insert ISO image" ], $output_handler, $node);
 | |
|             } else {
 | |
|                 xCAT::SvrUtils::sendmsg([ 1, "Unable to find Optical drive in VM to perform eject" ], $output_handler, $node);
 | |
|             }
 | |
|             return;
 | |
|         }
 | |
|         my $newDevice = VirtualCdrom->new(backing => $opticalbackingif,
 | |
|             key           => $oldcd->key,
 | |
|             controllerKey => 201,
 | |
|             unitNumber    => 0,
 | |
|             connectable   => $opticalconnectable,
 | |
|         );
 | |
|         push @devChanges, VirtualDeviceConfigSpec->new(
 | |
|             device    => $newDevice,
 | |
|             operation => VirtualDeviceConfigSpecOperation->new('edit'));
 | |
| 
 | |
|     }
 | |
|     if (%resize) {
 | |
|         while (my ($key, $value) = each(%resize)) {
 | |
|             my @drives = split(/,/, $key);
 | |
|             for my $device (@drives) {
 | |
|                 my $disk = $device;
 | |
|                 $device = getDiskByLabel($disk, $devices);
 | |
|                 unless ($device) {
 | |
|                     xCAT::SvrUtils::sendmsg([ 1, "Disk:  $disk does not exist" ], $output_handler, $node);
 | |
|                     return;
 | |
|                 }
 | |
|                 if ($value =~ /^\+(.+)/) {
 | |
|                     $value = $device->capacityInKB + getUnits($1, "G", 1024);
 | |
|                 } else {
 | |
|                     $value = getUnits($value, "G", 1024);
 | |
|                 }
 | |
|                 my $newDevice = VirtualDisk->new(deviceInfo => $device->deviceInfo,
 | |
|                     key           => $device->key,
 | |
|                     controllerKey => $device->controllerKey,
 | |
|                     unitNumber    => $device->unitNumber,
 | |
|                     deviceInfo    => $device->deviceInfo,
 | |
|                     backing       => $device->backing,
 | |
|                     capacityInKB  => $value);
 | |
|                 push @devChanges, VirtualDeviceConfigSpec->new(
 | |
|                     device    => $newDevice,
 | |
|                     operation => VirtualDeviceConfigSpecOperation->new('edit'));
 | |
|             }
 | |
|         }
 | |
| 
 | |
|     }
 | |
|     if (@devChanges) {
 | |
|         $conargs{deviceChange} = \@devChanges;
 | |
|     }
 | |
|     if (@otherparams) {
 | |
|         my $key;
 | |
|         my $value;
 | |
|         my @optionvals;
 | |
|         foreach (@otherparams) {
 | |
|             ($key, $value) = split /=/;
 | |
|             unless ($key) {
 | |
|                 xCAT::SvrUtils::sendmsg([ 1, "Invalid format for other parameter specification" ], $output_handler, $node);
 | |
|                 return;
 | |
|             }
 | |
|             if ($value) {
 | |
|                 push @optionvals, OptionValue->new(key => $key, value => $value);
 | |
|             } else {
 | |
|                 push @optionvals, OptionValue->new(key => $key); #the api doc says this is *supposed* to delete a key, don't think it works though, e.g. http://communities.vmware.com/message/1602644
 | |
|             }
 | |
|         }
 | |
|         $conargs{extraConfig} = \@optionvals;
 | |
|     }
 | |
| 
 | |
|     my $reconfigspec = VirtualMachineConfigSpec->new(%conargs);
 | |
| 
 | |
|     #xCAT::SvrUtils::sendmsg("reconfigspec = ".Dumper($reconfigspec));
 | |
|     my $task = $vmview->ReconfigVM_Task(spec => $reconfigspec);
 | |
|     $running_tasks{$task}->{task}     = $task;
 | |
|     $running_tasks{$task}->{callback} = \&chvm_task_callback;
 | |
|     $running_tasks{$task}->{hyp}      = $hyp;
 | |
|     $running_tasks{$task}->{data} = { node => $node, successtext => "node successfully changed", cpus => $cpuCount, mem => $memory };
 | |
| 
 | |
| }
 | |
| 
 | |
| sub getUsedUnits {
 | |
|     my $contKey = shift;
 | |
|     my $devices = shift;
 | |
|     my %usedids;
 | |
|     $usedids{7} = 1;
 | |
|     $usedids{'7'} = 1; #TODO: figure out which of these is redundant, the string or the number variant
 | |
|     for my $device (@$devices) {
 | |
|         if ($device->{controllerKey} eq $contKey) {
 | |
|             $usedids{ $device->{unitNumber} } = 1;
 | |
|         }
 | |
|     }
 | |
|     return \%usedids;
 | |
| }
 | |
| 
 | |
| sub getAvailUnit {
 | |
|     my $contKey = shift;
 | |
|     my $devices = shift;
 | |
|     my %args    = @_;
 | |
|     my $maxunit = -1;
 | |
|     if (defined $args{maxnum}) {
 | |
|         $maxunit = $args{maxnum};
 | |
|     }
 | |
|     my %usedids;
 | |
|     $usedids{7} = 1;
 | |
|     $usedids{'7'} = 1; #TODO: figure out which of these is redundant, the string or the number variant
 | |
|     for my $device (@$devices) {
 | |
|         if ($device->{controllerKey} eq $contKey) {
 | |
|             $usedids{ $device->{unitNumber} } = 1;
 | |
|         }
 | |
|     }
 | |
|     my $highestUnit = 0;
 | |
|     while ($usedids{$highestUnit}) {
 | |
|         if ($highestUnit == $maxunit) {
 | |
|             return -1;
 | |
|         }
 | |
|         $highestUnit++;
 | |
|     }
 | |
|     return $highestUnit;
 | |
| }
 | |
| 
 | |
| #given a device list from a vm and a label for a hard disk, returns the device object
 | |
| sub getDiskByLabel {
 | |
|     my $cmdLabel = shift;
 | |
|     my $devices  = shift;
 | |
|     my $device;
 | |
|     my $label;
 | |
| 
 | |
|     $cmdLabel = commandLabel($cmdLabel);
 | |
|     foreach $device (@$devices) {
 | |
|         $label = $device->deviceInfo->label;
 | |
| 
 | |
|         if ($cmdLabel eq $label) {
 | |
|             return $device;
 | |
|         } elsif (($label =~ /^Hard disk/) and ($cmdLabel =~ /^d(.*)/)) {
 | |
|             my $desc = $1;
 | |
|             if ($desc =~ /(.*):(.*)/) {    #specific
 | |
|                 my $controller = $1;
 | |
|                 my $unit       = $2;
 | |
|                 if ($device->unitNumber == $unit and $device->controllerKey == $controller) {
 | |
|                     return $device;
 | |
|                 }
 | |
|             } elsif ($desc =~ /\d+/ and $device->unitNumber == $desc) { #not specific
 | |
|                 return $device;
 | |
|             }
 | |
|         }
 | |
| 
 | |
|     }
 | |
|     return undef;
 | |
| }
 | |
| 
 | |
| #takes a label for a hard disk and prepends "Hard disk " if it's not there already
 | |
| sub commandLabel {
 | |
|     my $label = shift;
 | |
|     if (($label =~ /^Hard disk/) or ($label =~ /^d\d+/)) {
 | |
|         return $label;
 | |
|     }
 | |
|     return "Hard disk " . $label;
 | |
| }
 | |
| 
 | |
| #this function will check pending task status
 | |
| sub process_tasks {
 | |
|     foreach (keys %running_tasks) {
 | |
|         my $curcon;
 | |
|         if (defined $running_tasks{$_}->{conn}) {
 | |
|             $curcon = $running_tasks{$_}->{conn};
 | |
|         } elsif ($running_tasks{$_}->{hyp}) {
 | |
|             $curcon = $hyphash{ $running_tasks{$_}->{hyp} }->{conn};
 | |
|         } elsif ($running_tasks{$_}->{vm}) {
 | |
|             $curcon = $vmhash{ $running_tasks{$_}->{vm} }->{conn};
 | |
|         } elsif ($running_tasks{$_}->{cluster}) {
 | |
|             $curcon = $clusterhash{ $running_tasks{$_}->{cluster} }->{conn};
 | |
|         } else {
 | |
|             use Carp qw/confess/;
 | |
|             confess "This stack trace indicates a cluster unfriendly path";
 | |
|         }
 | |
|         my $curt = $curcon->get_view(mo_ref => $running_tasks{$_}->{task});
 | |
|         my $state = $curt->info->state->val;
 | |
|         unless ($state eq 'running' or $state eq 'queued') {
 | |
|             $running_tasks{$_}->{callback}->($curt, $running_tasks{$_}->{data});
 | |
|             delete $running_tasks{$_};
 | |
|         }
 | |
|         if ($state eq 'running' and not $running_tasks{$_}->{questionasked}) { # and $curt->info->progress == 95) { #This is unfortunate, there should be a 'state' to indicate a question is blocking
 | |
|              #however there isn't, so if we see something running at 95%, we just manually see if a question blocked the rest
 | |
|             my $vm;
 | |
|             $@ = "";
 | |
|             eval {
 | |
|                 $vm = $curcon->get_view(mo_ref => $curt->info->entity);
 | |
|             };
 | |
|             if ($@) { $vm = 0; }
 | |
|             if ($vm and $vm->{summary} and $vm->summary->{runtime} and $vm->summary->runtime->{question} and $vm->summary->runtime->question) {
 | |
|                 $running_tasks{$_}->{questionasked} = 1;
 | |
|                 $running_tasks{$_}->{callback}->($curt, $running_tasks{$_}->{data}, $vm->summary->runtime->question, $vm);
 | |
|             }
 | |
|         }
 | |
|     }
 | |
| }
 | |
| 
 | |
| #this function is a barrier to ensure prerequisites are met
 | |
| sub wait_for_tasks {
 | |
|     while (scalar keys %running_tasks) {
 | |
|         process_tasks;
 | |
|         sleep(1); #We'll check back in every second.  Unfortunately, we have to poll since we are in web service land
 | |
|     }
 | |
| }
 | |
| 
 | |
| sub connecthost_callback {
 | |
|     my $task  = shift;
 | |
|     my $args  = shift;
 | |
|     my $hv    = $args->{hostview};
 | |
|     my $state = $task->info->state->val;
 | |
|     if ($state eq "success") {
 | |
|         $hypready{ $args->{hypname} } = 1;    #declare readiness
 | |
|         enable_vmotion(hypname => $args->{hypname}, hostview => $args->{hostview}, conn => $args->{conn});
 | |
|         $vcenterhash{ $args->{vcenter} }->{goodhyps}->{ $args->{hypname} } = 1;
 | |
|         if (defined $args->{depfun}) { #If a function is waiting for the host connect to go valid, call it
 | |
|             $args->{depfun}->($args->{depargs});
 | |
|         }
 | |
|         return;
 | |
|     }
 | |
|     my $thumbprint;
 | |
|     eval {
 | |
|         $thumbprint = $task->{info}->error->fault->thumbprint;
 | |
|     };
 | |
|     if ($thumbprint) { #was an unknown certificate error, retry and accept the unknown certificate
 | |
|         $args->{connspec}->{sslThumbprint} = $task->info->error->fault->thumbprint;
 | |
|         my $task;
 | |
|         if (defined $args->{hostview}) {    #It was a reconnect request
 | |
|             $task = $hv->ReconnectHost_Task(cnxSpec => $args->{connspec});
 | |
|         } elsif (defined $args->{foldview}) {    #was an add host request
 | |
|             $task = $args->{foldview}->AddStandaloneHost_Task(spec => $args->{connspec}, addConnected => 1);
 | |
|         } elsif (defined $args->{cluster}) { #was an add host to cluster request
 | |
|             $task = $args->{cluster}->AddHost_Task(spec => $args->{connspec}, asConnected => 1);
 | |
|         }
 | |
|         $running_tasks{$task}->{task}     = $task;
 | |
|         $running_tasks{$task}->{callback} = \&connecthost_callback;
 | |
|         $running_tasks{$task}->{conn}     = $args->{conn};
 | |
|         $running_tasks{$task}->{data} = $args; #{ conn_spec=>$connspec,hostview=>$hv,hypname=>$args->{hypname},vcenter=>$args->{vcenter} };
 | |
|     } elsif ($state eq 'error') {
 | |
|         my $error = $task->info->error->localizedMessage;
 | |
|         if (defined($task->info->error->fault->faultMessage)) { #Only in 4.0, support of 3.5 must be careful?
 | |
|             foreach (@{ $task->info->error->fault->faultMessage }) {
 | |
|                 $error .= $_->message;
 | |
|             }
 | |
|         }
 | |
|         xCAT::SvrUtils::sendmsg([ 1, $error ], $output_handler);    #,$node);
 | |
|         $hypready{ $args->{hypname} } = -1; #Impossible for this hypervisor to ever be ready
 | |
|         $vcenterhash{ $args->{vcenter} }->{badhyps}->{ $args->{hypname} } = 1;
 | |
|     }
 | |
| }
 | |
| 
 | |
| sub delhost_callback { #only called in rmhypervisor -f case during validate vcenter phase
 | |
|     my $task  = shift;
 | |
|     my $args  = shift;
 | |
|     my $hv    = $args->{hostview};
 | |
|     my $state = $task->info->state->val;
 | |
|     if ($state eq "success") {
 | |
|         xCAT::SvrUtils::sendmsg("removed", $output_handler, $args->{hypname});
 | |
|         $hypready{ $args->{hypname} } = -1; #Impossible for this hypervisor to ever be ready
 | |
|         $vcenterhash{ $args->{vcenter} }->{badhyps}->{ $args->{hypname} } = 1;
 | |
|     } elsif ($state eq 'error') {
 | |
|         my $error = $task->info->error->localizedMessage;
 | |
|         if (defined($task->info->error->fault->faultMessage)) { #Only in 4.0, support of 3.5 must be careful?
 | |
|             foreach (@{ $task->info->error->fault->faultMessage }) {
 | |
|                 $error .= $_->message;
 | |
|             }
 | |
|         }
 | |
|         xCAT::SvrUtils::sendmsg([ 1, $error ], $output_handler);    #,$node);
 | |
|         $hypready{ $args->{hypname} } = -1; #Impossible for this hypervisor to ever be ready
 | |
|         $vcenterhash{ $args->{vcenter} }->{badhyps}->{ $args->{hypname} } = 1;
 | |
|     }
 | |
| }
 | |
| 
 | |
| sub get_clusterview {
 | |
|     my %args      = @_;
 | |
|     my $clustname = $args{clustname};
 | |
|     my %subargs   = (
 | |
|         view_type => 'ClusterComputeResource',
 | |
|     );
 | |
|     if ($args{properties}) {
 | |
|         $subargs{properties} = $args{properties};
 | |
|     }
 | |
|     $subargs{filter} = { name => $clustname };
 | |
|     my $view = $args{conn}->find_entity_view(%subargs);
 | |
|     return $view;
 | |
| 
 | |
|     #foreach (@{$args{conn}->find_entity_views(%subargs)}) {
 | |
|     #   if ($_->name eq "$clustname") {
 | |
|     #       return $_;
 | |
|     #       last;
 | |
|     #   }
 | |
|     #}
 | |
| }
 | |
| 
 | |
| sub get_hostview {
 | |
|     my %args    = @_;
 | |
|     my $host    = $args{hypname};
 | |
|     my %subargs = (
 | |
|         view_type => 'HostSystem',
 | |
|     );
 | |
|     if ($args{properties}) {
 | |
|         $subargs{properties} = $args{properties};
 | |
|     }
 | |
|     my @addrs = gethostbyname($host);
 | |
|     my $ip;
 | |
|     my $name;
 | |
|     my $aliases;
 | |
|     if ($addrs[4]) {
 | |
|         $ip = inet_ntoa($addrs[4]);
 | |
|         ($name, $aliases) = gethostbyaddr($addrs[4], AF_INET);    #TODO: IPv6
 | |
|     } else {
 | |
|         ($ip, $name, $aliases) = ($host, $host, "");
 | |
|     }
 | |
|     my @matchvalues = ($host, $ip, $name);
 | |
|     foreach (split /\s+/, $aliases) {
 | |
|         push @matchvalues, $_;
 | |
|     }
 | |
|     my $view;
 | |
|     $subargs{filter} = { 'name' => qr/$host(?:\.|\z)/ };
 | |
|     $view = $args{conn}->find_entity_view(%subargs);
 | |
|     if ($view) { return $view; }
 | |
|     foreach (@matchvalues) {
 | |
|         $subargs{filter} = { 'name' => qr/$_(?:\.|\z)/ };
 | |
|         $view = $args{conn}->find_entity_view(%subargs);
 | |
|         if ($view) { return $view; }
 | |
|     }
 | |
|     $subargs{filter} = { 'name' => qr/localhost(?:\.|\z)/ };
 | |
|     $view = $args{conn}->find_entity_view(%subargs);
 | |
|     if ($view) { return $view; }
 | |
|     return undef; #rest of function should be obsoleted, going to run with that assumption for 2.5 at least
 | |
| 
 | |
|     #    $subargs{filter}={'name' =~ qr/.*/};
 | |
|     #   foreach (@{$args{conn}->find_entity_views(%subargs)}) {
 | |
|     #      my $view = $_;
 | |
|     #      if ($_->name =~ /$host(?:\.|\z)/ or $_->name =~ /localhost(?:\.|\z)/ or grep { $view->name =~ /$_(?:\.|\z)/ } @matchvalues) {
 | |
|     #          return $view;
 | |
|     #          last;
 | |
|     #      }
 | |
|     #   }
 | |
| }
 | |
| 
 | |
| sub enable_vmotion {
 | |
| 
 | |
|     #TODO: vmware 3.x semantics too?  this is 4.0...
 | |
|     my %args = @_;
 | |
|     unless ($args{hostview}) {
 | |
|         $args{hostview} = get_hostview(conn => $args{conn}, hypname => $args{hypname}, properties => [ 'configManager', 'name' ]);
 | |
|     }
 | |
|     my $nicmgr = $args{conn}->get_view(mo_ref => $args{hostview}->configManager->virtualNicManager);
 | |
|     my $qnc = $nicmgr->QueryNetConfig(nicType => "vmotion");
 | |
|     if ($qnc->{selectedVnic}) {
 | |
|         return 1;
 | |
|     } else {
 | |
|         my $vniccount = scalar @{ $qnc->candidateVnic };
 | |
|         if ($vniccount == 1 or ($vniccount == 2 and $qnc->candidateVnic->[1]->spec->ip->ipAddress =~ /^169.254/)) { #There is only one possible path, use it
 | |
|             $nicmgr->SelectVnicForNicType(nicType => "vmotion", device => $qnc->candidateVnic->[0]->device);
 | |
|             return 1;
 | |
|         } else {
 | |
|             xCAT::SvrUtils::sendmsg([ 1, "TODO: use configuration to pick the nic " . $args{hypname} ], $output_handler);
 | |
|         }
 | |
|         return 0;
 | |
|     }
 | |
| }
 | |
| 
 | |
| sub mkvm_callback {
 | |
|     my $task = shift;
 | |
|     my $args = shift;
 | |
|     my $node = $args->{node};
 | |
|     my $hyp  = $args->{hyp};
 | |
|     if ($task->info->state->val eq 'error') {
 | |
|         my $error = $task->info->error->localizedMessage;
 | |
|         xCAT::SvrUtils::sendmsg([ 1, $error ], $output_handler, $node);
 | |
|     }
 | |
| }
 | |
| 
 | |
| sub relay_vmware_err {
 | |
|     my $task      = shift;
 | |
|     my $extratext = shift;
 | |
|     my @nodes     = @_;
 | |
|     my $error     = $task->info->error->localizedMessage;
 | |
|     if (defined($task->info->error->fault->faultMessage)) { #Only in 4.0, support of 3.5 must be careful?
 | |
|         foreach (@{ $task->info->error->fault->faultMessage }) {
 | |
|             $error .= $_->message;
 | |
|         }
 | |
|     }
 | |
|     if (@nodes) {
 | |
|         foreach (@nodes) {
 | |
|             xCAT::SvrUtils::sendmsg([ 1, $extratext . $error ], $output_handler, $_);
 | |
|         }
 | |
|     } else {
 | |
|         xCAT::SvrUtils::sendmsg([ 1, $extratext . $error ], $output_handler);
 | |
|     }
 | |
| }
 | |
| 
 | |
| sub relocate_callback {
 | |
|     my $task  = shift;
 | |
|     my $parms = shift;
 | |
|     my $state = $task->info->state->val;
 | |
|     if ($state eq 'success') {
 | |
|         my $vmtab = xCAT::Table->new('vm');    #TODO: update vm.storage?
 | |
|         my $prevloc = $tablecfg{vm}->{ $parms->{node} }->[0]->{storage};
 | |
|         my $model;
 | |
|         ($prevloc, $model) = split /=/, $prevloc;
 | |
|         my $target = $parms->{target};
 | |
|         if ($model) {
 | |
|             $target .= "=$model";
 | |
|         }
 | |
|         $vmtab->setNodeAttribs($parms->{node}, { storage => $target });
 | |
|         xCAT::SvrUtils::sendmsg(":relocated to to " . $parms->{target}, $output_handler, $parms->{node});
 | |
|     } else {
 | |
|         relay_vmware_err($task, "Relocating to " . $parms->{target} . " ", $parms->{node});
 | |
|     }
 | |
| }
 | |
| 
 | |
| sub migrate_ok {  #look like a successful migrate, callback for registering a vm
 | |
|     my %args  = @_;
 | |
|     my $vmtab = xCAT::Table->new('vm');
 | |
|     $vmtab->setNodeAttribs($args{nodes}->[0], { host => $args{target} });
 | |
|     xCAT::SvrUtils::sendmsg("migrated to " . $args{target}, $output_handler, $args{nodes}->[0]);
 | |
| }
 | |
| 
 | |
| sub migrate_callback {
 | |
|     my $task  = shift;
 | |
|     my $parms = shift;
 | |
|     my $state = $task->info->state->val;
 | |
|     if (not $parms->{skiptodeadsource} and $state eq 'success') {
 | |
|         my $vmtab = xCAT::Table->new('vm');
 | |
|         $vmtab->setNodeAttribs($parms->{node}, { host => $parms->{target} });
 | |
|         xCAT::SvrUtils::sendmsg("migrated to " . $parms->{target}, $output_handler, $parms->{node});
 | |
|     } elsif ($parms->{offline}) {    #try a forceful RegisterVM instead
 | |
|         my $target = $parms->{target};
 | |
|         my $hostview = $hyphash{$target}->{conn}->find_entity_view(view_type => 'VirtualMachine', properties => ['config.name'], filter => { name => $parms->{node} });
 | |
|         if ($hostview) { #this means vcenter still has it in inventory, but on a dead node...
 | |
|              #unfortunately, vcenter won't give up the old one until we zap the dead hypervisor
 | |
|              #also unfortunately, it doesn't make it easy to find said hypervisor..
 | |
|             $hostview = $hyphash{ $parms->{src} }->{conn}->get_view(mo_ref => $hyphash{ $parms->{src} }->{deletionref});
 | |
|             $task                             = $hostview->Destroy_Task();
 | |
|             $running_tasks{$task}->{task}     = $task;
 | |
|             $running_tasks{$task}->{callback} = \&migrate_callback;
 | |
|             $running_tasks{$task}->{conn} = $hyphash{$target}->{vcenter}->{conn};
 | |
|             $running_tasks{$task}->{data} = { offline => 1, src => $parms->{src}, node => $parms->{node}, target => $target, skiptodeadsource => 1 };
 | |
|         } else {    #it is completely gone, attempt a register_vm strategy
 | |
|             register_vm($target, $parms->{node}, undef, \&migrate_ok, { nodes => [ $parms->{node} ], target => $target, }, "failonerror");
 | |
|         }
 | |
|     } else {
 | |
|         relay_vmware_err($task, "Migrating to " . $parms->{target} . " ", $parms->{node});
 | |
|     }
 | |
| }
 | |
| 
 | |
| sub poweron_task_callback {
 | |
|     my $task   = shift;
 | |
|     my $parms  = shift;
 | |
|     my $q      = shift;                     #question if blocked
 | |
|     my $vm     = shift;                     #path to answer questions if asked
 | |
|     my $state  = $task->info->state->val;
 | |
|     my $node   = $parms->{node};
 | |
|     my $intent = $parms->{successtext};
 | |
|     if ($state eq 'success') {
 | |
|         xCAT::SvrUtils::sendmsg($intent, $output_handler, $node);
 | |
|     } elsif ($state eq 'error') {
 | |
|         relay_vmware_err($task, "", $node);
 | |
|     } elsif ($q and $q->text =~ /^msg.uuid.altered:/ and ($q->choice->choiceInfo->[0]->summary eq 'Cancel' and ($q->choice->choiceInfo->[0]->key eq '0'))) { #make sure it is what is what we have seen it to be
 | |
|         if ($parms->{forceon} and $q->choice->choiceInfo->[1]->summary =~ /I (_)?moved it/ and $q->choice->choiceInfo->[1]->key eq '1') { #answer the question as 'moved'
 | |
|             $vm->AnswerVM(questionId => $q->id, answerChoice => '1');
 | |
|         } else {
 | |
|             $vm->AnswerVM(questionId => $q->id, answerChoice => '0');
 | |
|             xCAT::SvrUtils::sendmsg([ 1, "Failure powering on VM, it mismatched against the hypervisor.  If positive VM is not running on another hypervisor, use -f to force VM on" ], $output_handler, $node);
 | |
|         }
 | |
|     } elsif ($q) {
 | |
|         if ($q->choice->choiceInfo->[0]->summary eq 'Cancel') {
 | |
|             xCAT::SvrUtils::sendmsg([ 1, ":Cancelling due to unexpected question executing task: " . $q->text ], $output_handler, $node);
 | |
|         } else {
 | |
|             xCAT::SvrUtils::sendmsg([ 1, ":Task hang due to unexpected question executing task, need to use VMware tools to clean up the mess for now: " . $q->text ], $output_handler, $node);
 | |
|         }
 | |
|     }
 | |
| 
 | |
| }
 | |
| 
 | |
| sub chvm_task_callback {
 | |
|     my $task   = shift;
 | |
|     my $parms  = shift;
 | |
|     my $state  = $task->info->state->val;
 | |
|     my $node   = $parms->{node};
 | |
|     my $intent = $parms->{successtext};
 | |
|     if ($state eq 'success') {
 | |
|         my $updatehash;
 | |
|         if ($parms->{cpus} and $tablecfg{vm}->{$node}->[0]->{cpus}) { #need to update
 | |
|             $updatehash->{cpus} = $parms->{cpus};
 | |
|         }
 | |
|         if ($parms->{mem} and $tablecfg{vm}->{$node}->[0]->{memory}) { #need to update
 | |
|             $updatehash->{memory} = $parms->{mem};
 | |
|         }
 | |
|         if ($updatehash) {
 | |
|             my $vmtab = xCAT::Table->new('vm', -create => 1);
 | |
|             $vmtab->setNodeAttribs($node, $updatehash);
 | |
|         }
 | |
|         xCAT::SvrUtils::sendmsg($intent, $output_handler, $node);
 | |
|     } elsif ($state eq 'error') {
 | |
|         relay_vmware_err($task, "", $node);
 | |
|     }
 | |
| }
 | |
| 
 | |
| sub generic_task_callback {
 | |
|     my $task   = shift;
 | |
|     my $parms  = shift;
 | |
|     my $state  = $task->info->state->val;
 | |
|     my $node   = $parms->{node};
 | |
|     my $intent = $parms->{successtext};
 | |
|     if ($state eq 'success') {
 | |
|         xCAT::SvrUtils::sendmsg($intent, $output_handler, $node);
 | |
|     } elsif ($state eq 'error') {
 | |
|         relay_vmware_err($task, "", $node);
 | |
|     }
 | |
| }
 | |
| 
 | |
| sub migrate {
 | |
|     my %args    = @_;
 | |
|     my @nodes   = @{ $args{nodes} };
 | |
|     my $hyp     = $args{hyp};
 | |
|     my $vcenter = $hyphash{$hyp}->{vcenter}->{name};
 | |
|     my $datastoredest;
 | |
|     my $offline;
 | |
|     @ARGV = @{ $args{exargs} };
 | |
|     unless (GetOptions(
 | |
|             's=s' => \$datastoredest,
 | |
|             'f'   => \$offline,
 | |
|         )) {
 | |
|         xCAT::SvrUtils::sendmsg([ 1, "Error parsing arguments" ], $output_handler);
 | |
|         return;
 | |
|     }
 | |
|     my $target = $hyp;    #case for storage migration
 | |
|     if ($datastoredest) { $datastoredest =~ s/=.*//; } #remove =scsi and similar if specified
 | |
|     if ($datastoredest and scalar @ARGV) {
 | |
|         xCAT::SvrUtils::sendmsg([ 1, "Unable to mix storage migration and processing of arguments " . join(' ', @ARGV) ], $output_handler);
 | |
|         return;
 | |
|     } elsif (@ARGV) {
 | |
|         $target = shift @ARGV;
 | |
|         if (@ARGV) {
 | |
|             xCAT::SvrUtils::sendmsg([ 1, "Unrecognized arguments " . join(' ', @ARGV) ], $output_handler);
 | |
|             return;
 | |
|         }
 | |
|     } elsif ($datastoredest) {                         #storage migration only
 | |
|         unless (validate_datastore_prereqs([], $hyp, { $datastoredest => \@nodes })) {
 | |
|             xCAT::SvrUtils::sendmsg([ 1, "Unable to find/mount target datastore $datastoredest" ], $output_handler);
 | |
|             return;
 | |
|         }
 | |
|         foreach (@nodes) {
 | |
|             my $hostview = $hyphash{$hyp}->{conn}->find_entity_view(view_type => 'VirtualMachine', properties => ['config.name'], filter => { name => $_ });
 | |
|             my $relocatspec = VirtualMachineRelocateSpec->new(
 | |
|                 datastore => $hyphash{$hyp}->{datastorerefmap}->{$datastoredest},
 | |
|             );
 | |
|             my $task = $hostview->RelocateVM_Task(spec => $relocatspec);
 | |
|             $running_tasks{$task}->{task}     = $task;
 | |
|             $running_tasks{$task}->{callback} = \&relocate_callback;
 | |
|             $running_tasks{$task}->{hyp}      = $args{hyp};
 | |
|             $running_tasks{$task}->{data} = { node => $_, target => $datastoredest };
 | |
|             process_tasks; #check for tasks needing followup actions before the task is forgotten (VMWare's memory is fairly short at times
 | |
|         }
 | |
|         return;
 | |
|     }
 | |
|     if ((not $offline and $vcenterhash{$vcenter}->{badhyps}->{$hyp}) or $vcenterhash{$vcenter}->{badhyps}->{$target}) {
 | |
|         xCAT::SvrUtils::sendmsg([ 1, "Unable to migrate " . join(',', @nodes) . " to $target due to inability to validate vCenter connectivity" ], $output_handler);
 | |
|         return;
 | |
|     }
 | |
|     if (($offline or $vcenterhash{$vcenter}->{goodhyps}->{$hyp}) and $vcenterhash{$vcenter}->{goodhyps}->{$target}) {
 | |
|         unless (validate_datastore_prereqs(\@nodes, $target)) {
 | |
|             xCAT::SvrUtils::sendmsg([ 1, "Unable to verify storage state on target system" ], $output_handler);
 | |
|             return;
 | |
|         }
 | |
|         unless (validate_network_prereqs(\@nodes, $target)) {
 | |
|             xCAT::SvrUtils::sendmsg([ 1, "Unable to verify target network state" ], $output_handler);
 | |
|             return;
 | |
|         }
 | |
|         my $dstview = get_hostview(conn => $hyphash{$target}->{conn}, hypname => $target, properties => [ 'name', 'parent' ]);
 | |
|         unless ($hyphash{$target}->{pool}) {
 | |
|             $hyphash{$target}->{pool} = $hyphash{$target}->{conn}->get_view(mo_ref => $dstview->parent, properties => ['resourcePool'])->resourcePool;
 | |
|         }
 | |
|         foreach (@nodes) {
 | |
|             process_tasks; #check for tasks needing followup actions before the task is forgotten (VMWare's memory is fairly short at times
 | |
|             my $srcview = $hyphash{$target}->{conn}->find_entity_view(view_type => 'VirtualMachine', properties => ['config.name'], filter => { name => $_ });
 | |
|             if ($offline and not $srcview) { #we have a request to resurrect the dead..
 | |
|                 register_vm($target, $_, undef, \&migrate_ok, { nodes => [$_], exargs => $args{exargs}, target => $target, hyp => $args{hyp}, offline => $offline }, "failonerror");
 | |
|                 return;
 | |
|             } elsif (not $srcview) {
 | |
|                 $srcview = $hyphash{$hyp}->{conn}->find_entity_view(view_type => 'VirtualMachine', properties => ['config.name'], filter => { name => $_ });
 | |
|             }
 | |
|             unless ($srcview) {
 | |
|                 xCAT::SvrUtils::sendmsg([ 1, "Unable to locate node in vCenter" ], $output_handler, $_);
 | |
|                 next;
 | |
|             }
 | |
| 
 | |
|             my $task = $srcview->MigrateVM_Task(
 | |
|                 host     => $dstview,
 | |
|                 pool     => $hyphash{$target}->{pool},
 | |
|                 priority => VirtualMachineMovePriority->new('highPriority'));
 | |
|             $running_tasks{$task}->{task}     = $task;
 | |
|             $running_tasks{$task}->{callback} = \&migrate_callback;
 | |
|             $running_tasks{$task}->{hyp}      = $args{hyp};
 | |
|             $running_tasks{$task}->{data} = { node => $_, src => $hyp, target => $target, offline => $offline };
 | |
|         }
 | |
|     } else {
 | |
| 
 | |
|         #xCAT::SvrUtils::sendmsg("Waiting for BOTH to be 'good'");
 | |
|         return;    #One of them is still 'pending'
 | |
|     }
 | |
| }
 | |
| 
 | |
| 
 | |
| sub reconfig_callback {
 | |
|     my $task = shift;
 | |
|     my $args = shift;
 | |
| 
 | |
|     #$args->{reconfig_args}->{vmview}->update_view_data();
 | |
|     delete $args->{reconfig_args}->{vmview}; #Force a reload of the view, update_view_data seems to not work as advertised..
 | |
|     $args->{reconfig_fun}->(%{ $args->{reconfig_args} });
 | |
| }
 | |
| 
 | |
| sub repower {    #Called to try power again after power down for reconfig
 | |
|     my $task    = shift;
 | |
|     my $args    = shift;
 | |
|     my $powargs = $args->{power_args};
 | |
|     $powargs->{pretendop} = 1;
 | |
| 
 | |
|     #$powargs->{vmview}->update_view_data();
 | |
|     delete $powargs->{vmview}; #Force a reload of the view, update_view_data seems to not work as advertised..
 | |
|     power(%$powargs);
 | |
| }
 | |
| 
 | |
| sub retry_rmvm {
 | |
|     my $task  = shift;
 | |
|     my $args  = shift;
 | |
|     my $node  = $args->{node};
 | |
|     my $state = $task->info->state->val;
 | |
|     if ($state eq "success") {
 | |
| 
 | |
|         #$Data::Dumper::Maxdepth=2;
 | |
|         delete $args->{args}->{vmview};
 | |
|         rmvm(%{ $args->{args} });
 | |
|     } elsif ($state eq 'error') {
 | |
|         relay_vmware_err($task, "", $node);
 | |
|     }
 | |
| }
 | |
| 
 | |
| sub rmvm {
 | |
|     my %args = @_;
 | |
|     my $node = $args{node};
 | |
|     my $hyp  = $args{hyp};
 | |
|     if (not defined $args{vmview}) {    #attempt one refresh
 | |
|         $args{vmview} = $vmhash{$node}->{conn}->find_entity_view(view_type => 'VirtualMachine', properties => [ 'config.name', 'runtime.powerState' ], filter => { name => $node });
 | |
|         if (not defined $args{vmview}) {
 | |
|             xCAT::SvrUtils::sendmsg([ 1, "VM does not appear to exist" ], $output_handler, $node);
 | |
|             return;
 | |
|         }
 | |
|     }
 | |
|     @ARGV = @{ $args{exargs} };
 | |
|     require Getopt::Long;
 | |
|     my $forceremove;
 | |
|     my $purge;
 | |
|     GetOptions(
 | |
|         'f' => \$forceremove,
 | |
|         'p' => \$purge,
 | |
|     );
 | |
|     my $task;
 | |
|     unless ($args{vmview}->{'runtime.powerState'}->val eq 'poweredOff') {
 | |
|         if ($forceremove) {
 | |
|             $task = $args{vmview}->PowerOffVM_Task();
 | |
|             $running_tasks{$task}->{task} = $task;
 | |
|             $running_tasks{$task}->{callback} = \&retry_rmvm,
 | |
|               $running_tasks{$task}->{hyp}    = $args{hyp};
 | |
|             $running_tasks{$task}->{vm} = $node;
 | |
|             $running_tasks{$task}->{data} = { node => $node, args => \%args };
 | |
|             return;
 | |
|         } else {
 | |
|             xCAT::SvrUtils::sendmsg([ 1, "Cannot rmvm active guest (use -f argument to force)" ], $output_handler, $node);
 | |
|             return;
 | |
|         }
 | |
|     }
 | |
|     if ($purge) {
 | |
|         $task = $args{vmview}->Destroy_Task();
 | |
|         $running_tasks{$task}->{data} = { node => $node, successtext => 'purged' };
 | |
|         $running_tasks{$task}->{task}     = $task;
 | |
|         $running_tasks{$task}->{callback} = \&generic_task_callback;
 | |
|         $running_tasks{$task}->{hyp} = $args{hyp};    #$hyp_conns->{$hyp};
 | |
|         $running_tasks{$task}->{vm}  = $node;
 | |
|     } else {
 | |
|         $task = $args{vmview}->UnregisterVM();
 | |
|     }
 | |
| }
 | |
| 
 | |
| 
 | |
| 
 | |
| sub getreconfigspec {
 | |
|     my %args    = @_;
 | |
|     my $node    = $args{node};
 | |
|     my $vmview  = $args{view};
 | |
|     my $currid  = $args{view}->{'config.guestId'};
 | |
|     my $rightid = getguestid($node);
 | |
|     my %conargs;
 | |
|     my $reconfigneeded = 0;
 | |
|     if ($currid ne $rightid) {
 | |
|         $reconfigneeded = 1;
 | |
|         $conargs{guestId} = $rightid;
 | |
|     }
 | |
|     my $newmem;
 | |
|     if ($tablecfg{vm}->{$node}->[0]->{memory} and $newmem = getUnits($tablecfg{vm}->{$node}->[0]->{memory}, "M", 1048576)) {
 | |
|         my $currmem = $vmview->{'config.hardware.memoryMB'};
 | |
|         if ($newmem ne $currmem) {
 | |
|             $conargs{memoryMB} = $newmem;
 | |
|             $reconfigneeded = 1;
 | |
|         }
 | |
|     }
 | |
|     my $newcpus;
 | |
|     if ($tablecfg{vm}->{$node}->[0]->{cpus} and $newcpus = $tablecfg{vm}->{$node}->[0]->{cpus}) {
 | |
|         my $currncpu = $vmview->{'config.hardware.numCPU'};
 | |
|         if ($newcpus ne $currncpu) {
 | |
|             $conargs{numCPUs} = $newcpus;
 | |
|             $reconfigneeded = 1;
 | |
|         }
 | |
|     }
 | |
|     if ($reconfigneeded) {
 | |
|         return VirtualMachineConfigSpec->new(%conargs);
 | |
|     } else {
 | |
|         return 0;
 | |
|     }
 | |
| }
 | |
| 
 | |
| #This routine takes a single node, managing vmv instance, and task tracking hash to submit a power on request
 | |
| sub power {
 | |
|     my %args = @_;
 | |
|     my $node = $args{node};
 | |
|     my $hyp  = $args{hyp};
 | |
|     my $pretendop = $args{pretendop}; #to pretend a system was on for reset or boot when we have to turn it off internally for reconfig
 | |
|     if (not defined $args{vmview}) {  #attempt one refresh
 | |
|         $args{vmview} = $vmhash{$node}->{conn}->find_entity_view(view_type => 'VirtualMachine', properties => [ 'config.name', 'config.guestId', 'config.hardware.memoryMB', 'config.hardware.numCPU', 'runtime.powerState' ], filter => { name => $node });
 | |
| 
 | |
|         #vmview not existing now is not an issue, this function
 | |
|         #is designed to handle that and correct if reasonably possible
 | |
|         #comes into play particularly in a stateless context
 | |
|     }
 | |
|     @ARGV = @{ $args{exargs} };    #for getoptions;
 | |
|     my $forceon;
 | |
|     require Getopt::Long;
 | |
|     GetOptions(
 | |
|         'force|f' => \$forceon,
 | |
|     );
 | |
|     my $subcmd = $ARGV[0];
 | |
|     my $intent = "";
 | |
|     my $task;
 | |
|     my $currstat;
 | |
| 
 | |
|     if ($args{vmview}) {
 | |
|         $currstat = $args{vmview}->{'runtime.powerState'}->val;
 | |
|         if (grep /$subcmd/, qw/on reset boot/) {
 | |
|             my $reconfigspec;
 | |
|             if ($reconfigreset and ($reconfigspec = getreconfigspec(node => $node, view => $args{vmview}))) {
 | |
|                 if ($currstat eq 'poweredOff') {
 | |
| 
 | |
|                     #xCAT::SvrUtils::sendmsg("Correcting guestId because $currid and $rightid are not the same...");#DEBUG
 | |
|                     my $task = $args{vmview}->ReconfigVM_Task(spec => $reconfigspec);
 | |
|                     $running_tasks{$task}->{task}     = $task;
 | |
|                     $running_tasks{$task}->{callback} = \&reconfig_callback;
 | |
|                     $running_tasks{$task}->{hyp}      = $args{hyp};
 | |
|                     $running_tasks{$task}->{vm}       = $node;
 | |
|                     $running_tasks{$task}->{data} = { node => $node, reconfig_fun => \&power, reconfig_args => \%args };
 | |
|                     return;
 | |
|                 } elsif (grep /$subcmd/, qw/reset boot/) { #going to have to do a 'cycle' and present it up normally..
 | |
|                         #xCAT::SvrUtils::sendmsg("DEBUG: forcing a cycle");
 | |
|                     $task = $args{vmview}->PowerOffVM_Task();
 | |
|                     $running_tasks{$task}->{task}     = $task;
 | |
|                     $running_tasks{$task}->{callback} = \&repower;
 | |
|                     $running_tasks{$task}->{hyp}      = $args{hyp};
 | |
|                     $running_tasks{$task}->{vm}       = $node;
 | |
|                     $running_tasks{$task}->{data} = { node => $node, power_args => \%args };
 | |
|                     return;    #we have to wait
 | |
|                 }
 | |
| 
 | |
|                 #TODO: fixit
 | |
|                 #xCAT::SvrUtils::sendmsg("I see vm has $currid and I want it to be $rightid");
 | |
|             }
 | |
|         }
 | |
|     } else {
 | |
|         $currstat = 'off';
 | |
|     }
 | |
|     if ($currstat eq 'poweredOff') {
 | |
|         $currstat = 'off';
 | |
|     } elsif ($currstat eq 'poweredOn') {
 | |
|         $currstat = 'on';
 | |
|     } elsif ($currstat eq 'suspended') {
 | |
|         $currstat = 'suspend';
 | |
|     }
 | |
|     if ($subcmd =~ /^stat/) {
 | |
|         xCAT::SvrUtils::sendmsg($currstat, $output_handler, $node);
 | |
|         return;
 | |
|     }
 | |
|     if ($subcmd =~ /boot/) {
 | |
|         $intent = "$currstat ";
 | |
|         if ($currstat eq 'on' or $args{pretendop}) {
 | |
|             $intent = "on ";
 | |
|             $subcmd = 'reset';
 | |
|         } else {
 | |
|             $subcmd = 'on';
 | |
|         }
 | |
|     }
 | |
|     if ($subcmd =~ /on/) {
 | |
|         if ($currstat eq 'off' or $currstat eq 'suspend') {
 | |
|             if (not $args{vmview}) { #We are asking to turn on a system the hypervisor
 | |
|                                      #doesn't know, attempt to register it first
 | |
|                 register_vm($hyp, $node, undef, \&power, \%args);
 | |
|                 return;    #We'll pick it up on the retry if it gets registered
 | |
|             }
 | |
|             eval {
 | |
|                 if ($hyp) {
 | |
|                     $task = $args{vmview}->PowerOnVM_Task(host => $hyphash{$hyp}->{hostview});
 | |
|                 } else {
 | |
|                     $task = $args{vmview}->PowerOnVM_Task(); #DRS may have it's way with me
 | |
|                 }
 | |
|             };
 | |
|             if ($@) {
 | |
|                 xCAT::SvrUtils::sendmsg([ 1, ":" . $@ ], $output_handler, $node);
 | |
|                 return;
 | |
|             }
 | |
|             $running_tasks{$task}->{task}     = $task;
 | |
|             $running_tasks{$task}->{callback} = \&poweron_task_callback;
 | |
|             $running_tasks{$task}->{hyp} = $args{hyp};    #$hyp_conns->{$hyp};
 | |
|             $running_tasks{$task}->{vm}  = $node;
 | |
|             $running_tasks{$task}->{data} = { node => $node, successtext => $intent . 'on', forceon => $forceon };
 | |
|         } else {
 | |
|             xCAT::SvrUtils::sendmsg($currstat, $output_handler, $node);
 | |
|         }
 | |
|     } elsif ($subcmd =~ /softoff/) {
 | |
|         if ($currstat eq 'on') {
 | |
|             $args{vmview}->ShutdownGuest();
 | |
|             xCAT::SvrUtils::sendmsg("softoff", $output_handler, $node);
 | |
|         } else {
 | |
|             xCAT::SvrUtils::sendmsg($currstat, $output_handler, $node);
 | |
|         }
 | |
|     } elsif ($subcmd =~ /off/) {
 | |
|         if ($currstat eq 'on') {
 | |
|             $task = $args{vmview}->PowerOffVM_Task();
 | |
|             $running_tasks{$task}->{task} = $task;
 | |
|             $running_tasks{$task}->{callback} = \&generic_task_callback;
 | |
|             $running_tasks{$task}->{hyp}      = $args{hyp};
 | |
|             $running_tasks{$task}->{vm}       = $node;
 | |
|             $running_tasks{$task}->{data} = { node => $node, successtext => 'off' };
 | |
|         } else {
 | |
|             xCAT::SvrUtils::sendmsg($currstat, $output_handler, $node);
 | |
|         }
 | |
|     } elsif ($subcmd =~ /suspend/) {
 | |
|         if ($currstat eq 'on') {
 | |
|             $task                             = $args{vmview}->SuspendVM_Task();
 | |
|             $running_tasks{$task}->{task}     = $task;
 | |
|             $running_tasks{$task}->{callback} = \&generic_task_callback;
 | |
|             $running_tasks{$task}->{hyp}      = $args{hyp};
 | |
|             $running_tasks{$task}->{vm}       = $node;
 | |
|             $running_tasks{$task}->{data} = { node => $node, successtext => 'suspend' };
 | |
|         } else {
 | |
|             xCAT::SvrUtils::sendmsg("off", $output_handler, $node);
 | |
|         }
 | |
|     } elsif ($subcmd =~ /reset/) {
 | |
|         if ($currstat eq 'on') {
 | |
|             $task                             = $args{vmview}->ResetVM_Task();
 | |
|             $running_tasks{$task}->{task}     = $task;
 | |
|             $running_tasks{$task}->{callback} = \&generic_task_callback;
 | |
|             $running_tasks{$task}->{hyp}      = $args{hyp};
 | |
|             $running_tasks{$task}->{vm}       = $node;
 | |
|             $running_tasks{$task}->{data} = { node => $node, successtext => $intent . 'reset' };
 | |
|         } elsif ($args{pretendop}) {    #It is off, but pretend it was on
 | |
|             eval {
 | |
|                 if ($hyp) {
 | |
|                     $task = $args{vmview}->PowerOnVM_Task(host => $hyphash{$hyp}->{hostview});
 | |
|                 } else {
 | |
|                     $task = $args{vmview}->PowerOnVM_Task();    #allow DRS
 | |
|                 }
 | |
|             };
 | |
|             if ($@) {
 | |
|                 xCAT::SvrUtils::sendmsg([ 1, ":" . $@ ], $output_handler, $node);
 | |
|                 return;
 | |
|             }
 | |
|             $running_tasks{$task}->{task}     = $task;
 | |
|             $running_tasks{$task}->{callback} = \&generic_task_callback;
 | |
|             $running_tasks{$task}->{hyp}      = $args{hyp};
 | |
|             $running_tasks{$task}->{vm}       = $node;
 | |
|             $running_tasks{$task}->{data} = { node => $node, successtext => $intent . 'reset' };
 | |
|         } else {
 | |
|             xCAT::SvrUtils::sendmsg($currstat, $output_handler, $node);
 | |
|         }
 | |
|     }
 | |
| }
 | |
| 
 | |
| sub generic_vm_operation { #The general form of firing per-vm requests to ESX hypervisor
 | |
|     my $properties = shift; #The relevant properties to the general task, MUST INCLUDE config.name
 | |
|     my $function = shift; #The function to actually run against the right VM view
 | |
|     my @exargs   = @_;    #Store the rest to pass on
 | |
|     my $hyp;
 | |
|     my $vmviews;
 | |
|     my %vcviews; #views populated once per vcenter server for improved performance
 | |
|     my $node;
 | |
|     foreach $hyp (keys %hyphash) {
 | |
| 
 | |
|         if ($viavcenterbyhyp->{$hyp}) {
 | |
|             foreach $node (keys %{ $hyphash{$hyp}->{nodes} }) {
 | |
|                 $vcenterhash{ $hyphash{$hyp}->{vcenter}->{name} }->{vms}->{$node} = 1;
 | |
|             }
 | |
|         }
 | |
|     }
 | |
|     foreach (keys %limbonodes) {
 | |
|         $vcenterhash{ $limbonodes{$_} }->{vms}->{$_} = 1;
 | |
|     }
 | |
|     my $cluster;
 | |
|     foreach $cluster (keys %clusterhash) {
 | |
|         foreach $node (keys %{ $clusterhash{$cluster}->{nodes} }) {
 | |
|             $vcenterhash{ $clusterhash{$cluster}->{vcenter}->{name} }->{vms}->{$node} = 1;
 | |
|         }
 | |
|     }
 | |
|     my $currentvcenter;
 | |
|     my %foundlimbo;
 | |
|     foreach $currentvcenter (keys %vcenterhash) {
 | |
| 
 | |
|         #retrieve all vm views in one gulp
 | |
|         my $vmsearchstring = join(")|(", keys %{ $vcenterhash{$currentvcenter}->{vms} });
 | |
|         $vmsearchstring = '^((' . $vmsearchstring . '))(\z|\.)';
 | |
|         my $regex = qr/$vmsearchstring/;
 | |
|         $vcviews{$currentvcenter} = $vcenterhash{$currentvcenter}->{conn}->find_entity_views(view_type => 'VirtualMachine', properties => $properties, filter => { 'config.name' => $regex });
 | |
|         foreach (@{ $vcviews{$currentvcenter} }) {
 | |
|             my $node = $_->{'config.name'};
 | |
|             unless (defined $tablecfg{vm}->{$node}) {
 | |
|                 $node =~ s/\..*//;    #try the short name;
 | |
|             }
 | |
|             if (defined $tablecfg{vm}->{$node}) { #see if the host pointer requires a refresh
 | |
|                 my $hostref = $hostrefbynode{$node};
 | |
|                 if ($hostref and $hostref eq $_->{'runtime.host'}->value) { next; } #the actual host reference  matches the one that we got when populating hostviews based on what the table had to say #TODO: does this mean it is buggy if we want to mkvm/rmigrate/etc if the current vm.host is wrong and the noderange doesn't have something on the right hostview making us not get it in the
 | |
|                                                                                     #mass request?  Or is it just slower because it hand gets host views?
 | |
|                 my $host = $vcenterhash{$currentvcenter}->{conn}->get_view(mo_ref => $_->{'runtime.host'}, properties => ['summary.config.name']);
 | |
|                 $host = $host->{'summary.config.name'};
 | |
|                 my $shost = $host;
 | |
|                 $shost =~ s/\..*//;
 | |
| 
 | |
|                 #time to figure out which of these is a node
 | |
|                 my @nodes = noderange("$host,$shost");
 | |
|                 my $vmtab = xCAT::Table->new("vm", -create => 1);
 | |
|                 unless ($vmtab) {
 | |
|                     die "Error opening vm table";
 | |
|                 }
 | |
|                 if ($nodes[0]) {
 | |
|                     if ($limbonodes{$node}) { $foundlimbo{$node} = $currentvcenter; }
 | |
|                     $vmtab->setNodeAttribs($node, { host => $nodes[0] });
 | |
|                 }    #else {
 | |
|                      #  $vmtab->setNodeAttribs($node,{host=>$host});
 | |
|                      #}
 | |
|             }
 | |
|         }
 | |
|     }
 | |
|     foreach my $lnode (keys %foundlimbo) {
 | |
|         $vmviews = $vcviews{ $foundlimbo{$lnode} };
 | |
|         my %mgdvms;    #sort into a hash for convenience
 | |
|         foreach (@$vmviews) {
 | |
|             $mgdvms{ $_->{'config.name'} } = $_;
 | |
|         }
 | |
|         $function->(
 | |
|             node   => $lnode,
 | |
|             vm     => $lnode,
 | |
|             vmview => $mgdvms{$node},
 | |
|             exargs => \@exargs
 | |
|         );
 | |
|     }
 | |
|     my @entitylist;
 | |
|     push @entitylist, keys %hyphash;
 | |
|     push @entitylist, keys %clusterhash;
 | |
|     foreach my $entity (@entitylist) {
 | |
|         if ($hyphash{$entity}) {
 | |
|             $hyp = $entity;    #save some retyping...
 | |
|             if ($viavcenterbyhyp->{$hyp}) {
 | |
|                 $vmviews = $vcviews{ $hyphash{$hyp}->{vcenter}->{name} }
 | |
|             } else {
 | |
|                 $vmviews = [];
 | |
|                 my $node;
 | |
|                 foreach $node (sort (keys %{ $hyphash{$hyp}->{nodes} })) {
 | |
|                     push @{$vmviews}, $hyphash{$hyp}->{conn}->find_entity_view(view_type => 'VirtualMachine', properties => $properties, filter => { 'config.name' => qr/^$node/ });
 | |
|                 }
 | |
| 
 | |
|                 #$vmviews = $hyphash{$hyp}->{conn}->find_entity_views(view_type => 'VirtualMachine',properties=>$properties);
 | |
|             }
 | |
|             my %mgdvms;        #sort into a hash for convenience
 | |
|             foreach (@$vmviews) {
 | |
|                 $mgdvms{ $_->{'config.name'} } = $_;
 | |
|             }
 | |
|             my $node;
 | |
|             foreach $node (sort (keys %{ $hyphash{$hyp}->{nodes} })) {
 | |
|                 $function->(
 | |
|                     node   => $node,
 | |
|                     hyp    => $hyp,
 | |
|                     vmview => $mgdvms{$node},
 | |
|                     exargs => \@exargs
 | |
|                 );
 | |
|                 process_tasks; #check for tasks needing followup actions before the task is forgotten (VMWare's memory is fairly short at times
 | |
|             }
 | |
|         } else {    #a cluster.
 | |
|             $vmviews = $vcviews{ $clusterhash{$entity}->{vcenter}->{name} };
 | |
|             my %mgdvms;    #sort into a hash for convenience
 | |
|             foreach (@$vmviews) {
 | |
|                 $mgdvms{ $_->{'config.name'} } = $_;
 | |
|             }
 | |
|             my $node;
 | |
|             foreach $node (sort (keys %{ $clusterhash{$entity}->{nodes} })) {
 | |
|                 $function->(
 | |
|                     node    => $node,
 | |
|                     cluster => $entity,
 | |
|                     vm      => $node,
 | |
|                     vmview  => $mgdvms{$node},
 | |
|                     exargs  => \@exargs
 | |
|                 );
 | |
|             }
 | |
|         }
 | |
|     }
 | |
| }
 | |
| 
 | |
| sub generic_hyp_operation { #The general form of firing per-hypervisor requests to ESX hypervisor
 | |
|     my $function = shift; #The function to actually run against the right VM view
 | |
|     my @exargs   = @_;    #Store the rest to pass on
 | |
|     my $hyp;
 | |
|     if (scalar keys %limbonodes) { #we are in forced migration with dead sources, try to register them
 | |
|         @ARGV = @exargs;
 | |
|         my $datastoredest;
 | |
|         my $offline;
 | |
|         unless (GetOptions(
 | |
|                 's=s' => \$datastoredest,
 | |
|                 'f'   => \$offline,
 | |
|             )) {
 | |
|             xCAT::SvrUtils::sendmsg([ 1, "Error parsing arguments" ], $output_handler);
 | |
|             return;
 | |
|         }
 | |
|         if ($datastoredest) {
 | |
|             xCAT::SvrUtils::sendmsg([ 1, "Storage migration impossible with dead hypervisor, must be migrated to live hypervisor first" ], $output_handler);
 | |
|             return;
 | |
|         } elsif (@ARGV) {
 | |
|             my $target = shift @ARGV;
 | |
|             if (@ARGV) {
 | |
|                 xCAT::SvrUtils::sendmsg([ 1, "Unrecognized arguments " . join(' ', @ARGV) ], $output_handler);
 | |
|                 return;
 | |
|             }
 | |
|             foreach (keys %limbonodes) {
 | |
|                 register_vm($target, $_, undef, \&migrate_ok, { nodes => [$_], target => $target, }, "failonerror");
 | |
|             }
 | |
|         } else {    #storage migration only
 | |
|             xCAT::SvrUtils::sendmsg([ 1, "No target hypervisor specified" ], $output_handler);
 | |
| 
 | |
|         }
 | |
|     }
 | |
|     foreach $hyp (keys %hyphash) {
 | |
|         process_tasks; #check for tasks needing followup actions before the task is forgotten (VMWare's memory is fairly short at times
 | |
|         my @relevant_nodes = sort (keys %{ $hyphash{$hyp}->{nodes} });
 | |
|         unless (scalar @relevant_nodes) {
 | |
|             next;
 | |
|         }
 | |
|         $function->(
 | |
|             nodes  => \@relevant_nodes,
 | |
|             hyp    => $hyp,
 | |
|             exargs => \@exargs
 | |
|         );
 | |
| 
 | |
|         #my $vmviews = $hyp_conns->{$hyp}->find_entity_views(view_type => 'VirtualMachine',properties=>['runtime.powerState','config.name']);
 | |
|         #my %mgdvms; #sort into a hash for convenience
 | |
|         #foreach (@$vmviews) {
 | |
|         # $mgdvms{$_->{'config.name'}} = $_;
 | |
|         #}
 | |
|         #my $node;
 | |
|         #foreach $node (sort (keys %{$hyp_hash->{$hyp}->{nodes}})){
 | |
|         #    $function->($node,$mgdvms{$node},$taskstotrack,$callback,@exargs);
 | |
|         #REMINDER FOR RINV TO COME
 | |
|         #     foreach (@nothing) { #@{$mgdvms{$node}->config->hardware->device}) {
 | |
|         #           if (defined $_->{macAddress}) {
 | |
|         #                  print "\nFound a mac: ".$_->macAddress."\n";
 | |
|         #               }
 | |
|         #           }
 | |
|         #        }
 | |
|     }
 | |
|     foreach $hyp (keys %clusterhash) { #clonevm, mkvm, rmigrate could land here in clustered mode with DRS/HA
 | |
|         process_tasks;
 | |
|         my @relevant_nodes = sort (keys %{ $clusterhash{$hyp}->{nodes} });
 | |
|         unless (scalar @relevant_nodes) {
 | |
|             next;
 | |
|         }
 | |
|         $function->(nodes => \@relevant_nodes, cluster => $hyp, exargs => \@exargs, conn => $clusterhash{$hyp}->{conn});
 | |
|     }
 | |
| }
 | |
| 
 | |
| sub rmhypervisor_disconnected {
 | |
|     my $task  = shift;
 | |
|     my $parms = shift;
 | |
|     my $node  = $parms->{node};
 | |
|     my $hyp   = $node;
 | |
|     my $state = $task->info->state->val;
 | |
|     if ($state eq 'success') {
 | |
|         my $task = $hyphash{$hyp}->{hostview}->Destroy_Task();
 | |
|         $running_tasks{$task}->{data} = { node => $node, successtext => 'removed' };
 | |
|         $running_tasks{$task}->{task}     = $task;
 | |
|         $running_tasks{$task}->{callback} = \&generic_task_callback;
 | |
|         $running_tasks{$task}->{hyp}      = $hyp;
 | |
|     } elsif ($state eq 'error') {
 | |
|         relay_vmware_err($task, "", $node);
 | |
|     }
 | |
| }
 | |
| 
 | |
| sub rmhypervisor_inmaintenance {
 | |
|     my $task   = shift;
 | |
|     my $parms  = shift;
 | |
|     my $state  = $task->info->state->val;
 | |
|     my $node   = $parms->{node};
 | |
|     my $intent = $parms->{successtext};
 | |
|     if ($state eq 'success') {
 | |
|         my $hyp  = $parms->{node};
 | |
|         my $task = $hyphash{$hyp}->{hostview}->DisconnectHost_Task();
 | |
|         $running_tasks{$task}->{task}     = $task;
 | |
|         $running_tasks{$task}->{callback} = \&rmhypervisor_disconnected;
 | |
|         $running_tasks{$task}->{hyp}      = $hyp;
 | |
|         $running_tasks{$task}->{data}     = { node => $hyp };
 | |
|     } elsif ($state eq 'error') {
 | |
|         relay_vmware_err($task, "", $node);
 | |
|     }
 | |
| }
 | |
| 
 | |
| sub lsvm {
 | |
|     my %args = @_;
 | |
|     my $hyp  = $args{hyp};
 | |
|     $hyphash{$hyp}->{hostview} = get_hostview(hypname => $hyp, conn => $hyphash{$hyp}->{conn}); #,properties=>['config','configManager']);
 | |
|     use Data::Dumper;
 | |
|     my $vms = $hyphash{$hyp}->{hostview}->vm;
 | |
|     unless ($vms) {
 | |
|         return;
 | |
|     }
 | |
|     foreach (@$vms) {
 | |
|         my $vmv = $hyphash{$hyp}->{conn}->get_view(mo_ref => $_);
 | |
|         xCAT::SvrUtils::sendmsg($vmv->name, $output_handler, $hyp);
 | |
|     }
 | |
|     return;
 | |
| }
 | |
| 
 | |
| sub chhypervisor {
 | |
|     my %args = @_;
 | |
|     @ARGV = @{ $args{exargs} };    #for getoptions;
 | |
|     my $maintenance;
 | |
|     my $online;
 | |
|     my $stat;
 | |
|     my $vlanaddspec;
 | |
|     my $vlanremspec;
 | |
|     require Getopt::Long;
 | |
|     GetOptions(
 | |
|         'maintenance|m' => \$maintenance,
 | |
|         'online|o'      => \$online,
 | |
|         'show|s'        => \$stat,
 | |
|         'show|s'        => \$stat,
 | |
|         'addvlan=s'     => \$vlanaddspec,
 | |
|         'removevlan=s'  => \$vlanremspec,
 | |
|     );
 | |
|     my $hyp = $args{hyp};
 | |
|     $hyphash{$hyp}->{hostview} = get_hostview(hypname => $hyp, conn => $hyphash{$hyp}->{conn}); #,properties=>['config','configManager']);
 | |
|     if ($maintenance) {
 | |
|         if (defined $hyphash{$hyp}->{hostview}) {
 | |
|             my $task = $hyphash{$hyp}->{hostview}->EnterMaintenanceMode_Task(timeout => 0);
 | |
|             $running_tasks{$task}->{task}     = $task;
 | |
|             $running_tasks{$task}->{callback} = \&generic_task_callback;
 | |
|             $running_tasks{$task}->{hyp}      = $args{hyp};
 | |
|             $running_tasks{$task}->{data} = { node => $hyp, successtext => "hypervisor in maintenance mode" };
 | |
|         }
 | |
|     } elsif ($online) {
 | |
|         if (defined $hyphash{$hyp}->{hostview}) {
 | |
|             my $task = $hyphash{$hyp}->{hostview}->ExitMaintenanceMode_Task(timeout => 0);
 | |
|             $running_tasks{$task}->{task}     = $task;
 | |
|             $running_tasks{$task}->{callback} = \&generic_task_callback;
 | |
|             $running_tasks{$task}->{hyp}      = $args{hyp};
 | |
|             $running_tasks{$task}->{data} = { node => $hyp, successtext => "hypervisor online" };
 | |
|         }
 | |
|     } elsif ($stat) {
 | |
|         if (defined $hyphash{$hyp}->{hostview}) {
 | |
|             if ($hyphash{$hyp}->{hostview}->runtime->inMaintenanceMode) {
 | |
|                 xCAT::SvrUtils::sendmsg("hypervisor in maintenance mode", $output_handler, $hyp);
 | |
|             } else {
 | |
|                 xCAT::SvrUtils::sendmsg("hypervisor online", $output_handler, $hyp);
 | |
|             }
 | |
|         }
 | |
|     } elsif ($vlanaddspec) {
 | |
|         fixup_hostportgroup($vlanaddspec, $hyp);
 | |
|     } elsif ($vlanremspec) {
 | |
|         fixup_hostportgroup($vlanremspec, $hyp, action => 'remove');
 | |
|     }
 | |
|     return;
 | |
| }
 | |
| 
 | |
| sub rshutdown {    #TODO: refactor with next function too
 | |
|     my %args = @_;
 | |
|     my $hyp  = $args{hyp};
 | |
|     $hyphash{$hyp}->{hostview} = get_hostview(hypname => $hyp, conn => $hyphash{$hyp}->{conn}); #,properties=>['config','configManager']);
 | |
|     if (defined $hyphash{$hyp}->{hostview}) {
 | |
|         my $task = $hyphash{$hyp}->{hostview}->EnterMaintenanceMode_Task(timeout => 0);
 | |
|         $running_tasks{$task}->{task}     = $task;
 | |
|         $running_tasks{$task}->{callback} = \&rshutdown_inmaintenance;
 | |
|         $running_tasks{$task}->{hyp}      = $args{hyp};
 | |
|         $running_tasks{$task}->{data}     = { node => $hyp };
 | |
|     }
 | |
|     return;
 | |
| }
 | |
| 
 | |
| sub rshutdown_inmaintenance {
 | |
|     my $task   = shift;
 | |
|     my $parms  = shift;
 | |
|     my $state  = $task->info->state->val;
 | |
|     my $node   = $parms->{node};
 | |
|     my $intent = $parms->{successtext};
 | |
|     if ($state eq 'success') {
 | |
|         my $hyp = $parms->{node};
 | |
|         if (defined $hyphash{$hyp}->{hostview}) {
 | |
|             my $task = $hyphash{$hyp}->{hostview}->ShutdownHost_Task(force => 0);
 | |
|             $running_tasks{$task}->{task}     = $task;
 | |
|             $running_tasks{$task}->{callback} = \&generic_task_callback;
 | |
|             $running_tasks{$task}->{hyp}      = $hyp;
 | |
|             $running_tasks{$task}->{data} = { node => $hyp, successtext => "shutdown initiated" };
 | |
|         }
 | |
|     } elsif ($state eq 'error') {
 | |
|         relay_vmware_err($task, "", $node);
 | |
|     }
 | |
|     return;
 | |
| }
 | |
| 
 | |
| sub rescansan {
 | |
|     my %args = @_;
 | |
|     my $hyp  = $args{hyp};
 | |
|     my $hostview = get_hostview(hypname => $hyp, conn => $hyphash{$hyp}->{conn}, properties => [ 'config', 'configManager' ]);
 | |
|     if (defined $hostview) {
 | |
|         my $hdss = $hostview->{vim}->get_view(mo_ref => $hostview->configManager->storageSystem);
 | |
|         $hdss->RescanAllHba();
 | |
|         $hdss->RescanVmfs();
 | |
|     }
 | |
| }
 | |
| 
 | |
| sub formatdisk {
 | |
|     my %args = @_;
 | |
|     my $hyp  = $args{hyp};
 | |
|     $hyphash{$hyp}->{hostview} = get_hostview(hypname => $hyp, conn => $hyphash{$hyp}->{conn}, properties => [ 'config', 'configManager' ]);
 | |
|     @ARGV = @{ $args{exargs} };
 | |
|     my $nid;
 | |
|     my $name;
 | |
|     GetOptions(
 | |
|         'id=s'   => \$nid,
 | |
|         'name=s' => \$name,
 | |
|     );
 | |
|     my $hostview = $hyphash{$hyp}->{hostview};
 | |
|     if (defined $hyphash{$hyp}->{hostview}) {
 | |
|         my $hdss = $hostview->{vim}->get_view(mo_ref => $hostview->configManager->storageSystem);
 | |
|         $hdss->RescanAllHba();
 | |
|         my $dss = $hostview->{vim}->get_view(mo_ref => $hostview->configManager->datastoreSystem);
 | |
|         my $diskList = $dss->QueryAvailableDisksForVmfs();
 | |
|         foreach my $disk (@$diskList) {
 | |
|             foreach my $id (@{ $disk->{descriptor} }) {
 | |
|                 if (lc($id->{id}) eq lc('naa.' . $nid)) {
 | |
|                     my $options = $dss->QueryVmfsDatastoreCreateOptions(devicePath => $disk->devicePath);
 | |
|                     @$options[0]->spec->vmfs->volumeName($name);
 | |
|                     my $newDatastore = $dss->CreateVmfsDatastore(spec => @$options[0]->spec);
 | |
|                 }
 | |
|             }
 | |
|         }
 | |
| 
 | |
|     }
 | |
|     return;
 | |
| }
 | |
| 
 | |
| sub rmhypervisor {
 | |
|     my %args = @_;
 | |
|     my $hyp  = $args{hyp};
 | |
|     $hyphash{$hyp}->{hostview} = get_hostview(hypname => $hyp, conn => $hyphash{$hyp}->{conn}); #,properties=>['config','configManager']);
 | |
|     if (defined $hyphash{$hyp}->{hostview}) {
 | |
|         my $task = $hyphash{$hyp}->{hostview}->EnterMaintenanceMode_Task(timeout => 0);
 | |
|         $running_tasks{$task}->{task}     = $task;
 | |
|         $running_tasks{$task}->{callback} = \&rmhypervisor_inmaintenance;
 | |
|         $running_tasks{$task}->{hyp}      = $args{hyp};
 | |
|         $running_tasks{$task}->{data}     = { node => $hyp };
 | |
|     }
 | |
|     return;
 | |
| }
 | |
| 
 | |
| sub clonevms {
 | |
|     my %args    = @_;
 | |
|     my $nodes   = $args{nodes};
 | |
|     my $hyp     = $args{hyp};
 | |
|     my $cluster = $args{cluster};
 | |
|     @ARGV = @{ $args{exargs} };    #for getoptions;
 | |
|     my $base;
 | |
|     my $force;
 | |
|     my $detach;
 | |
|     my $specialize;
 | |
|     my $target;
 | |
|     require Getopt::Long;
 | |
|     GetOptions(
 | |
|         'b=s'        => \$base,
 | |
|         'f'          => \$force,
 | |
|         'd'          => \$detach,
 | |
|         'specialize' => \$specialize,
 | |
|         't=s'        => \$target,
 | |
|     );
 | |
|     if ($base and $target) {
 | |
|         foreach my $node (@$nodes) {
 | |
|             xCAT::SvrUtils::sendmsg([ 1, "Cannot specify both base (-b) and target (-t)" ], $output_handler, $node);
 | |
|         }
 | |
|         return;
 | |
|     }
 | |
|     unless ($base or $target) {
 | |
|         foreach my $node (@$nodes) {
 | |
|             xCAT::SvrUtils::sendmsg([ 1, "Must specify one of base (-b) or target (-t)" ], $output_handler, $node);
 | |
|         }
 | |
|         return;
 | |
|     }
 | |
|     if ($target and (scalar @{$nodes} != 1)) {
 | |
|         foreach my $node (@$nodes) {
 | |
|             xCAT::SvrUtils::sendmsg([ 1, "Cannot specify mulitple nodes to create a master from" ], $output_handler, $node);
 | |
|         }
 | |
|         return;
 | |
|     }
 | |
|     if ($hyp) {
 | |
|         $hyphash{$hyp}->{hostview} = get_hostview(hypname => $hyp, conn => $hyphash{$hyp}->{conn});
 | |
|     }
 | |
|     my $newdatastores;
 | |
|     my $mastername;
 | |
|     my $url;
 | |
|     my $masterref;
 | |
|     if ($base) {    #if base, we need to pull in the target datastores
 | |
|         my $mastertab = xCAT::Table->new('vmmaster');
 | |
|         $masterref = $mastertab->getNodeAttribs($base, [qw/storage os arch profile storagemodel nics specializeparameters/]);
 | |
|         unless ($masterref) {
 | |
|             foreach my $node (@$nodes) {
 | |
|                 xCAT::SvrUtils::sendmsg([ 1, "Cannot find master $base in vmmaster table" ], $output_handler, $node);
 | |
|             }
 | |
|             return;
 | |
|         }
 | |
|         $newdatastores->{ $masterref->{storage} } = []; #make sure that the master datastore is mounted...
 | |
|         foreach (@$nodes) {
 | |
|             my $url;
 | |
|             if ($tablecfg{vm}->{$_}->[0]->{storage}) {
 | |
|                 $url = $tablecfg{vm}->{$_}->[0]->{storage};
 | |
|                 $url =~ s/=.*//;
 | |
|             } else {
 | |
|                 $url = $masterref->{storage};
 | |
|             }
 | |
|             unless ($url) { die "Shouldn't be possible"; }
 | |
|             if (ref $newdatastores->{$_}) {
 | |
|                 push @{ $newdatastores->{$url} }, $_;
 | |
|             } else {
 | |
|                 $newdatastores->{$url} = [$_];
 | |
|             }
 | |
|         }
 | |
|     } elsif ($target) {
 | |
|         if ($url =~ m!/!) {
 | |
|             $url = $target;
 | |
|             $url =~ s!/([^/]*)\z!!;
 | |
|             $mastername = $1;
 | |
|         } else {
 | |
|             $url = $tablecfg{vm}->{ $nodes->[0] }->[0]->{storage};
 | |
|             $url =~ s/.*\|//;
 | |
|             $url =~ s/=(.*)//;
 | |
|             $url =~ s/,.*//;
 | |
|             $mastername = $target
 | |
|         }
 | |
|         $newdatastores->{$url} = [ $nodes->[0] ];
 | |
|     }
 | |
|     if ($hyp) {
 | |
|         unless (validate_datastore_prereqs($nodes, $hyp, $newdatastores)) {
 | |
|             return;
 | |
|         }
 | |
|     } else {    #need to build datastore map for cluster
 | |
|         refreshclusterdatastoremap($cluster);
 | |
|     }
 | |
|     sortoutdatacenters(nodes => $nodes, hyp => $hyp, cluster => $cluster);
 | |
|     if ($target) {
 | |
|         return promote_vm_to_master(node => $nodes->[0], target => $target, force => $force, detach => $detach, cluster => $cluster, hyp => $hyp, url => $url, mastername => $mastername);
 | |
|     } elsif ($base) {
 | |
|         return clone_vms_from_master(nodes => $nodes, base => $base, detach => $detach, cluster => $cluster, hyp => $hyp, mastername => $base, masterent => $masterref, specialize => $specialize);
 | |
|     }
 | |
| }
 | |
| 
 | |
| sub sortoutdatacenters { #figure out all the vmfolders for all the nodes passed in
 | |
|     my %args    = @_;
 | |
|     my $nodes   = $args{nodes};
 | |
|     my $hyp     = $args{hyp};
 | |
|     my $cluster = $args{cluster};
 | |
|     my %nondefaultdcs;
 | |
|     my $deffolder;
 | |
|     my $conn;
 | |
|     if ($hyp) {
 | |
| 
 | |
|         unless (defined $hyphash{$hyp}->{vmfolder}) {
 | |
|             $hyphash{$hyp}->{vmfolder} = $hyphash{$hyp}->{conn}->get_view(mo_ref => $hyphash{$hyp}->{conn}->find_entity_view(view_type => 'Datacenter', properties => ['vmFolder'])->vmFolder);
 | |
|         }
 | |
|         $conn      = $hyphash{$hyp}->{conn};
 | |
|         $deffolder = $hyphash{$hyp}->{vmfolder};
 | |
|     } else {    #clustered
 | |
|         unless (defined $clusterhash{$cluster}->{vmfolder}) {
 | |
|             $clusterhash{$cluster}->{vmfolder} = $clusterhash{$cluster}->{conn}->get_view(mo_ref => $clusterhash{$cluster}->{conn}->find_entity_view(view_type => 'Datacenter', properties => ['vmFolder'])->vmFolder);
 | |
|         }
 | |
|         $deffolder = $clusterhash{$cluster}->{vmfolder};
 | |
|         $conn      = $clusterhash{$cluster}->{conn};
 | |
|     }
 | |
|     foreach (@$nodes) {
 | |
|         if ($tablecfg{vm}->{$_}->[0]->{datacenter}) {
 | |
|             $nondefaultdcs{ $tablecfg{vm}->{$_}->[0]->{datacenter} }->{$_} = 1;
 | |
|         } else {
 | |
|             $vmhash{$_}->{vmfolder} = $deffolder;
 | |
|         }
 | |
|     }
 | |
|     my $datacenter;
 | |
|     foreach $datacenter (keys %nondefaultdcs) {
 | |
|         my $vmfolder = $conn->get_view(mo_ref => $conn->find_entity_view(view_type => 'Datacenter', properties => ['vmFolder'], filter => { name => $datacenter })->vmFolder, filter => { name => $datacenter });
 | |
|         foreach (keys %{ $nondefaultdcs{$datacenter} }) {
 | |
|             $vmhash{$_}->{vmfolder} = $vmfolder;
 | |
|         }
 | |
|     }
 | |
| }
 | |
| 
 | |
| sub clone_vms_from_master {
 | |
|     my %args       = @_;
 | |
|     my $mastername = $args{mastername};
 | |
|     my $specialize = $args{specialize};
 | |
|     my $hyp        = $args{hyp};
 | |
|     my $cluster    = $args{cluster};
 | |
|     my $regex      = qr/^$mastername\z/;
 | |
|     my @nodes      = @{ $args{nodes} };
 | |
|     my $node;
 | |
|     my $conn;
 | |
| 
 | |
|     if ($hyp) {
 | |
|         $conn = $hyphash{$hyp}->{conn};
 | |
|     } else {
 | |
|         $conn = $clusterhash{$cluster}->{conn};
 | |
|     }
 | |
|     my $masterviews = $conn->find_entity_views(view_type => 'VirtualMachine', filter => { 'config.name' => $regex });
 | |
|     if (scalar(@$masterviews) != 1) {
 | |
|         foreach $node (@nodes) {
 | |
|             xCAT::SvrUtils::sendmsg([ 1, "Unable to find master $mastername in VMWare infrastructure" ], $output_handler, $node);
 | |
|         }
 | |
|         return;
 | |
|     }
 | |
|     my $masterview = $masterviews->[0];
 | |
|     my $masterent  = $args{masterent};
 | |
|     my $ostype;
 | |
|     foreach $node (@nodes) {
 | |
|         my $destination = $tablecfg{vm}->{$node}->[0]->{storage};
 | |
|         my $nodetypeent;
 | |
|         my $vment;
 | |
| 
 | |
|         $ostype = $masterent->{'os'};
 | |
|         foreach (qw/os arch profile/) {
 | |
|             $nodetypeent->{$_} = $masterent->{$_};
 | |
|         }
 | |
|         foreach (qw/storagemodel nics/) {
 | |
|             $vment->{$_} = $masterent->{$_};
 | |
|         }
 | |
|         $vment->{master} = $args{mastername};
 | |
|         unless ($destination) {
 | |
|             $destination = $masterent->{storage};
 | |
|             $vment->{storage} = $destination;
 | |
|         }
 | |
|         $destination =~ s/=.*//;
 | |
|         my $placement_resources = get_placement_resources(hyp => $hyp, cluster => $cluster, destination => $destination);
 | |
|         my $pool             = $placement_resources->{pool};
 | |
|         my $dstore           = $placement_resources->{datastore};
 | |
|         my %relocatespecargs = (
 | |
|             datastore => $dstore, #$hyphash{$hyp}->{datastorerefmap}->{$destination},
 | |
|             pool      => $pool,
 | |
| 
 | |
|             #diskMoveType=>"createNewChildDiskBacking", #fyi, requires a snapshot, which isn't compatible with templates, moveChildMostDiskBacking would potentially be fine, but either way is ha incopmatible and limited to 8, arbitrary limitations hard to work around...
 | |
|         );
 | |
|         unless ($args{detach}) {
 | |
|             $relocatespecargs{diskMoveType} = "createNewChildDiskBacking";
 | |
|         }
 | |
|         if ($hyp) { $relocatespecargs{host} = $hyphash{$hyp}->{hostview} }
 | |
|         my $relocatespec  = VirtualMachineRelocateSpec->new(%relocatespecargs);
 | |
|         my %clonespecargs = (
 | |
|             location => $relocatespec,
 | |
|             template => 0,
 | |
|             powerOn  => 0
 | |
|         );
 | |
|         unless ($args{detach}) {
 | |
|             $clonespecargs{snapshot} = $masterview->snapshot->currentSnapshot;
 | |
|         }
 | |
|         if ($specialize) {
 | |
|             my %custargs;
 | |
|             if ($masterent->{specializeparameters}) { %custargs = (parameters => $masterent->{specializeparameters}); }
 | |
|             $clonespecargs{customization} = make_customization_spec($node, ostype => $ostype, %custargs);
 | |
|         }
 | |
|         my $clonespec = VirtualMachineCloneSpec->new(%clonespecargs);
 | |
|         my $vmfolder  = $vmhash{$node}->{vmfolder};
 | |
|         my $task = $masterview->CloneVM_Task(folder => $vmfolder, name => $node, spec => $clonespec);
 | |
|         $running_tasks{$task}->{data} = { node => $node, conn => $conn, successtext => 'Successfully cloned from ' . $args{mastername},
 | |
|             mastername => $args{mastername}, nodetypeent => $nodetypeent, vment => $vment,
 | |
|             hyp => $args{hyp},
 | |
|         };
 | |
|         $running_tasks{$task}->{task}     = $task;
 | |
|         $running_tasks{$task}->{callback} = \&clone_task_callback;
 | |
|         $running_tasks{$task}->{hyp} = $args{hyp};    #$hyp_conns->{$hyp};
 | |
|         $running_tasks{$task}->{vm}  = $node;         #$hyp_conns->{$hyp};
 | |
|     }
 | |
| }
 | |
| 
 | |
| sub make_customization_spec {
 | |
|     my $node = shift;
 | |
|     my %args = @_;
 | |
|     my $password;
 | |
|     my $wintimezone;
 | |
| 
 | |
|     #map of number to strings can be found at
 | |
|     #http://osman-shener-en.blogspot.com/2008/02/unattendedtxt-time-zone-index.html
 | |
|     my $fullname = "Unspecified User";
 | |
|     my $orgName  = "Unspecified Organization";
 | |
|     if ($::XCATSITEVALS{winfullname}) { $fullname = $::XCATSITEVALS{winfullname}; }
 | |
|     if ($::XCATSITEVALS{winorgname}) { $orgName = $::XCATSITEVALS{winorgname}; }
 | |
|     my @runonce = ();    #to be read in from postscripts table
 | |
|     $wintimezone = xCAT::TZUtils::get_wintimezonenum();
 | |
|     my $ptab = xCAT::Table->new('postscripts', -create => 0);
 | |
| 
 | |
|     if ($ptab) {
 | |
|         my $psent = $ptab->getNodeAttribs($node, [qw/postscripts postbootscripts/]);
 | |
|         if ($psent and $psent->{postscripts}) {
 | |
|             push @runonce, split /,/, $psent->{postscripts};
 | |
|         }
 | |
|         if ($psent and $psent->{postbootscripts}) {
 | |
|             push @runonce, split /,/, $psent->{postbootscripts};
 | |
|         }
 | |
|     }
 | |
|     $ptab = xCAT::Table->new('passwd', -create => 0);
 | |
|     unless ($ptab) {
 | |
|         die "passwd table needed";
 | |
|     }
 | |
|     my ($passent) = $ptab->getAttribs({ "key" => "system", username => "Administrator" }, 'password');
 | |
|     unless ($passent) {
 | |
|         die "need passwd table entry for system account Administrator";
 | |
|     }
 | |
|     $password = $passent->{password};
 | |
|     my %lfpd;
 | |
|     if ($args{ostype} and $args{ostype} =~ /win2k3/) {
 | |
|         %lfpd = (
 | |
|             licenseFilePrintData => CustomizationLicenseFilePrintData->new(
 | |
|                 autoMode => CustomizationLicenseDataMode->new(
 | |
|                     'perSeat'
 | |
|                   )
 | |
|               )
 | |
|         );
 | |
|     }
 | |
|     my %runonce;
 | |
|     if (scalar @runonce) {    #skip section if no postscripts or postbootscripts
 | |
|         %runonce = (
 | |
|             guiRunOnce => CustomizationGuiRunOnce->new(
 | |
|                 commandList => \@runonce,
 | |
|               )
 | |
|         );
 | |
|     }
 | |
|     my %autologonargs = (autoLogon => 0, autoLogonCount => 1,);
 | |
|     if ($args{parameters} and $args{parameters} =~ /autoLogonCount=([^,]*)/i) {
 | |
|         my $count = $1;
 | |
|         if ($count) {
 | |
|             $autologonargs{autoLogon}      = 1;
 | |
|             $autologonargs{autoLogonCount} = $count;
 | |
|         }
 | |
|     }
 | |
|     my $identity = CustomizationSysprep->new(
 | |
|         %runonce,
 | |
|         %lfpd,
 | |
|         guiUnattended => CustomizationGuiUnattended->new(
 | |
|             %autologonargs,
 | |
|             password => CustomizationPassword->new(
 | |
|                 plainText => 1,
 | |
|                 value     => $password,
 | |
|             ),
 | |
|             timeZone => $wintimezone,
 | |
|         ),
 | |
|         identification => get_customizedidentification(),
 | |
|         userData       => CustomizationUserData->new(
 | |
|             computerName => CustomizationFixedName->new(name => $node),
 | |
|             fullName     => $fullname,
 | |
|             orgName      => $orgName,
 | |
|             productId    => "",
 | |
|         ),
 | |
|     );
 | |
|     my $options = CustomizationWinOptions->new(changeSID => 1, deleteAccounts => 0);
 | |
|     my $customizationspec = CustomizationSpec->new(
 | |
|         globalIPSettings => CustomizationGlobalIPSettings->new(),
 | |
|         identity         => $identity,
 | |
|         nicSettingMap    => [
 | |
|             CustomizationAdapterMapping->new(adapter => CustomizationIPSettings->new(ip => CustomizationDhcpIpGenerator->new()))
 | |
|         ],
 | |
|         options => $options,
 | |
|     );
 | |
|     return $customizationspec;
 | |
| 
 | |
| }
 | |
| 
 | |
| sub get_customizedidentification {
 | |
| 
 | |
|     #for now, just do a 'TBD' workgroup.  VMWare not supporting joining without domain admin password is rather unfortunate
 | |
|     return CustomizationIdentification->new(
 | |
|         joinWorkgroup => "TBD",
 | |
|     );
 | |
| }
 | |
| 
 | |
| 
 | |
| 
 | |
| 
 | |
| sub get_placement_resources {
 | |
|     my %args = @_;
 | |
|     my $pool;
 | |
|     my $dstore;
 | |
|     my $hyp         = $args{hyp};
 | |
|     my $cluster     = $args{cluster};
 | |
|     my $destination = $args{destination};
 | |
|     if ($hyp) {
 | |
|         unless (defined $hyphash{$hyp}->{pool}) {
 | |
|             $hyphash{$hyp}->{pool} = $hyphash{$hyp}->{conn}->get_view(mo_ref => $hyphash{$hyp}->{hostview}->parent, properties => ['resourcePool'])->resourcePool;
 | |
|         }
 | |
|         $pool = $hyphash{$hyp}->{pool};
 | |
|         if ($destination) { $dstore = $hyphash{$hyp}->{datastorerefmap}->{$destination} }
 | |
|     } else {    #clustered...
 | |
|         unless (defined $clusterhash{$cluster}->{pool}) {
 | |
|             my $cview = get_clusterview(clustname => $cluster, conn => $clusterhash{$cluster}->{conn});
 | |
|             $clusterhash{$cluster}->{pool} = $cview->resourcePool;
 | |
|         }
 | |
|         $pool = $clusterhash{$cluster}->{pool};
 | |
|         if ($destination) { $dstore = $clusterhash{$cluster}->{datastorerefmap}->{$destination} }
 | |
|     }
 | |
|     return {
 | |
|         pool      => $pool,
 | |
|         datastore => $dstore,
 | |
|       }
 | |
| }
 | |
| 
 | |
| sub clone_task_callback {
 | |
|     my $task   = shift;
 | |
|     my $parms  = shift;
 | |
|     my $state  = $task->info->state->val;
 | |
|     my $node   = $parms->{node};
 | |
|     my $conn   = $parms->{conn};
 | |
|     my $intent = $parms->{successtext};
 | |
|     if ($state eq 'success') {
 | |
| 
 | |
|         #xCAT::SvrUtils::sendmsg($intent, $output_handler,$node);
 | |
|         my $nodetype = xCAT::Table->new('nodetype', -create => 1);
 | |
|         my $vm       = xCAT::Table->new('vm',       -create => 1);
 | |
|         $vm->setAttribs({ node => $node }, $parms->{vment});
 | |
| 
 | |
|         $nodetype->setAttribs({ node => $node }, $parms->{nodetypeent});
 | |
|         foreach (keys %{ $parms->{vment} }) {
 | |
|             $tablecfg{vm}->{$node}->[0]->{$_} = $parms->{vment}->{$_};
 | |
|         }
 | |
| 
 | |
|         my @networks = split /,/, $tablecfg{vm}->{$node}->[0]->{nics};
 | |
|         my @macs = xCAT::VMCommon::getMacAddresses(\%tablecfg, $node, scalar @networks);
 | |
| 
 | |
|         #now with macs, change all macs in the vm to match our generated macs
 | |
|         my $regex = qr/^$node(\z|\.)/;
 | |
| 
 | |
|         #have to do an expensive pull of the vm view, since it is brand new
 | |
|         my $nodeviews = $conn->find_entity_views(view_type => 'VirtualMachine', filter => { 'config.name' => $regex });
 | |
|         unless (scalar @$nodeviews == 1) { die "this should be impossible"; }
 | |
|         my $vpdtab = xCAT::Table->new('vpd', -create => 1);
 | |
|         $vpdtab->setAttribs({ node => $node }, { uuid => $nodeviews->[0]->config->uuid });
 | |
|         my $ndev;
 | |
|         my @devstochange;
 | |
|         foreach $ndev (@{ $nodeviews->[0]->config->hardware->device }) {
 | |
|             unless ($ndev->{macAddress}) { next; }    #not an ndev
 | |
|             $ndev->{macAddress}  = shift @macs;
 | |
|             $ndev->{addressType} = "manual";
 | |
|             push @devstochange, VirtualDeviceConfigSpec->new(
 | |
|                 device    => $ndev,
 | |
|                 operation => VirtualDeviceConfigSpecOperation->new('edit'));
 | |
|         }
 | |
|         if (@devstochange) {
 | |
|             my $reconfigspec = VirtualMachineConfigSpec->new(deviceChange => \@devstochange);
 | |
|             my $task = $nodeviews->[0]->ReconfigVM_Task(spec => $reconfigspec);
 | |
|             $running_tasks{$task}->{task}     = $task;
 | |
|             $running_tasks{$task}->{callback} = \&generic_task_callback;
 | |
|             $running_tasks{$task}->{hyp}      = $parms->{hyp};
 | |
|             $running_tasks{$task}->{conn}     = $parms->{conn};
 | |
|             $running_tasks{$task}->{data} = { node => $node, successtext => $intent };
 | |
|         }
 | |
| 
 | |
| 
 | |
|     } elsif ($state eq 'error') {
 | |
|         relay_vmware_err($task, "", $node);
 | |
|     }
 | |
| }
 | |
| 
 | |
| sub promote_vm_to_master {
 | |
|     my %args    = @_;
 | |
|     my $node    = $args{node};
 | |
|     my $hyp     = $args{hyp};
 | |
|     my $cluster = $args{cluster};
 | |
|     my $regex   = qr/^$node(\z|\.)/;
 | |
|     my $conn;
 | |
|     if ($hyp) {
 | |
|         $conn = $hyphash{$hyp}->{conn};
 | |
|     } else {
 | |
|         $conn = $clusterhash{$cluster}->{conn};
 | |
|     }
 | |
|     my $nodeviews = $conn->find_entity_views(view_type => 'VirtualMachine', filter => { 'config.name' => $regex });
 | |
|     if (scalar(@$nodeviews) != 1) {
 | |
|         xCAT::SvrUtils::sendmsg([ 1, "Cannot find $node in VMWare infrastructure" ], $output_handler, $node);
 | |
|         return;
 | |
|     }
 | |
|     my $nodeview = shift @$nodeviews;
 | |
|     my $dstore;
 | |
|     if ($hyp) {
 | |
|         $dstore = $hyphash{$hyp}->{datastorerefmap}->{ $args{url} },
 | |
|     } else {
 | |
|         $dstore = $clusterhash{$cluster}->{datastorerefmap}->{ $args{url} },
 | |
|     }
 | |
|     my $relocatespec = VirtualMachineRelocateSpec->new(
 | |
|         datastore => $dstore,
 | |
|     );
 | |
|     my $clonespec = VirtualMachineCloneSpec->new(
 | |
|         location => $relocatespec,
 | |
|         template => 0, #can't go straight to template, need to clone, then snap, then templatify
 | |
|         powerOn  => 0
 | |
|     );
 | |
| 
 | |
|     my $vmfolder = $vmhash{$node}->{vmfolder};
 | |
|     my $task = $nodeview->CloneVM_Task(folder => $vmfolder, name => $args{mastername}, spec => $clonespec);
 | |
|     $running_tasks{$task}->{data} = { node => $node, hyp => $args{hyp}, conn => $conn, successtext => 'Successfully copied to ' . $args{mastername}, mastername => $args{mastername}, url => $args{url} };
 | |
|     $running_tasks{$task}->{task}     = $task;
 | |
|     $running_tasks{$task}->{callback} = \&promote_task_callback;
 | |
|     $running_tasks{$task}->{hyp} = $args{hyp};    #$hyp_conns->{$hyp};
 | |
|     $running_tasks{$task}->{vm}  = $node;
 | |
| }
 | |
| 
 | |
| sub promote_task_callback {
 | |
|     my $task   = shift;
 | |
|     my $parms  = shift;
 | |
|     my $state  = $task->info->state->val;
 | |
|     my $node   = $parms->{node};
 | |
|     my $intent = $parms->{successtext};
 | |
|     if ($state eq 'success') { #now, we have to make one snapshot for linked clones
 | |
|         my $mastername = $parms->{mastername};
 | |
|         my $regex      = qr/^$mastername\z/;
 | |
|         my $masterviews = $parms->{conn}->find_entity_views(view_type => 'VirtualMachine', filter => { 'config.name' => $regex });
 | |
|         unless (scalar @$masterviews == 1) {
 | |
|             die "Impossible";
 | |
|         }
 | |
|         my $masterview = $masterviews->[0];
 | |
|         my $task = $masterview->CreateSnapshot_Task(name => "xcatsnap", memory => "false", quiesce => "false");
 | |
|         $parms->{masterview}              = $masterview;
 | |
|         $running_tasks{$task}->{data}     = $parms;
 | |
|         $running_tasks{$task}->{task}     = $task;
 | |
|         $running_tasks{$task}->{callback} = \&promotesnap_task_callback;
 | |
|         $running_tasks{$task}->{hyp} = $parms->{hyp};    #$hyp_conns->{$hyp};
 | |
|         $running_tasks{$task}->{vm}  = $parms->{node};
 | |
| 
 | |
|         #xCAT::SvrUtils::sendmsg($intent, $output_handler,$node);
 | |
|     } elsif ($state eq 'error') {
 | |
|         relay_vmware_err($task, "", $node);
 | |
|     }
 | |
| }
 | |
| 
 | |
| sub promotesnap_task_callback {
 | |
|     my $task   = shift;
 | |
|     my $parms  = shift;
 | |
|     my $state  = $task->info->state->val;
 | |
|     my $node   = $parms->{node};
 | |
|     my $intent = $parms->{successtext};
 | |
|     if ($state eq 'success') {
 | |
|         $parms->{masterview}->MarkAsTemplate;    #time to be a template
 | |
|         xCAT::SvrUtils::sendmsg($intent, $output_handler, $node);
 | |
|         my $mastertabentry = {
 | |
|             originator => $requester,
 | |
|             vintage    => scalar(localtime),
 | |
|             storage    => $parms->{url},
 | |
|         };
 | |
|         foreach (qw/os arch profile/) {
 | |
|             if (defined($tablecfg{nodetype}->{$node}->[0]->{$_})) {
 | |
|                 $mastertabentry->{$_} = $tablecfg{nodetype}->{$node}->[0]->{$_};
 | |
|             }
 | |
|         }
 | |
|         foreach (qw/storagemodel nics/) {
 | |
|             if (defined($tablecfg{vm}->{$node}->[0]->{$_})) {
 | |
|                 $mastertabentry->{$_} = $tablecfg{vm}->{$node}->[0]->{$_};
 | |
|             }
 | |
|         }
 | |
|         my $vmmastertab = xCAT::Table->new('vmmaster', -create => 1);
 | |
|         my $date = scalar(localtime);
 | |
|         $vmmastertab->setAttribs({ name => $parms->{mastername} }, $mastertabentry);
 | |
|     } elsif ($state eq 'error') {
 | |
|         relay_vmware_err($task, "", $node);
 | |
|     }
 | |
| }
 | |
| 
 | |
| sub mkvms {
 | |
|     my %args    = @_;
 | |
|     my $nodes   = $args{nodes};
 | |
|     my $hyp     = $args{hyp};
 | |
|     my $cluster = $args{cluster};
 | |
|     @ARGV = @{ $args{exargs} };    #for getoptions;
 | |
|     my $disksize;
 | |
|     require Getopt::Long;
 | |
|     my $cpuCount;
 | |
|     my $memory;
 | |
|     GetOptions(
 | |
|         'size|s=s' => \$disksize,
 | |
|         "cpus=s"   => \$cpuCount,
 | |
|         "mem=s"    => \$memory
 | |
|     );
 | |
|     my $node;
 | |
|     my $conn;
 | |
|     if ($hyp) {
 | |
|         $hyphash{$hyp}->{hostview} = get_hostview(hypname => $hyp, conn => $hyphash{$hyp}->{conn}); #,properties=>['config','configManager']);
 | |
|         unless (validate_datastore_prereqs($nodes, $hyp)) {
 | |
|             return;
 | |
|         }
 | |
|         $conn = $hyphash{$hyp}->{conn};
 | |
|     } else {
 | |
|         refreshclusterdatastoremap($cluster);
 | |
|         $conn = $clusterhash{$cluster}->{conn};
 | |
|     }
 | |
|     sortoutdatacenters(nodes => $nodes, hyp => $hyp, cluster => $cluster);
 | |
|     my $placement_resources = get_placement_resources(hyp => $hyp, cluster => $cluster);
 | |
| 
 | |
|     #$hyphash{$hyp}->{pool} = $hyphash{$hyp}->{conn}->get_view(mo_ref=>$hyphash{$hyp}->{hostview}->parent,properties=>['resourcePool'])->resourcePool;
 | |
|     my $cfg;
 | |
|     foreach $node (@$nodes) {
 | |
|         process_tasks; #check for tasks needing followup actions before the task is forgotten (VMWare's memory is fairly short at times
 | |
|         if ($conn->find_entity_view(view_type => "VirtualMachine", filter => { name => $node })) {
 | |
|             xCAT::SvrUtils::sendmsg([ 1, "Virtual Machine already exists" ], $output_handler, $node);
 | |
|             next;
 | |
|         } else {
 | |
|             register_vm($hyp, $node, $disksize, undef, undef, undef, cpus => $cpuCount, memory => $memory, cluster => $cluster);
 | |
|         }
 | |
|     }
 | |
|     my @dhcpnodes;
 | |
|     foreach (keys %{ $tablecfg{dhcpneeded} }) {
 | |
|         push @dhcpnodes, $_;
 | |
|         delete $tablecfg{dhcpneeded}->{$_};
 | |
|     }
 | |
|     unless ($::XCATSITEVALS{'dhcpsetup'} and ($::XCATSITEVALS{'dhcpsetup'} =~ /^n/i or $::XCATSITEVALS{'dhcpsetup'} =~ /^d/i or $::XCATSITEVALS{'dhcpsetup'} eq '0')) {
 | |
|         $executerequest->({ command => ['makedhcp'], node => \@dhcpnodes });
 | |
|     }
 | |
| }
 | |
| 
 | |
| sub setboot {
 | |
|     my %args = @_;
 | |
|     my $node = $args{node};
 | |
|     my $hyp  = $args{hyp};
 | |
|     if (not defined $args{vmview}) {    #attempt one refresh
 | |
|         $args{vmview} = $vmhash{$node}->{conn}->find_entity_view(view_type => 'VirtualMachine', properties => ['config.name'], filter => { name => $node });
 | |
|         if (not defined $args{vmview}) {
 | |
|             xCAT::SvrUtils::sendmsg([ 1, "VM does not appear to exist" ], $output_handler, $node);
 | |
|             return;
 | |
|         }
 | |
|     }
 | |
|     my $bootorder = ${ $args{exargs} }[0];
 | |
| 
 | |
|     #NOTE: VMware simply does not currently seem to allow programatically changing the boot
 | |
|     #order like other virtualization solutions supported by xCAT.
 | |
|     #This doesn't behave quite like any existing mechanism:
 | |
|     #vm.bootorder was meant to take the place of system nvram, vmware imitates that unfortunate aspect of bare metal too well..
 | |
|     #rsetboot was created to describe the ipmi scenario of a transient boot device, this is persistant *except* for setup, which is not
 | |
|     #rbootseq was meant to be entirely persistant and ordered.
 | |
|     #rsetboot is picked, the usage scenario matches about as good as I could think of
 | |
|     my $reconfigspec;
 | |
|     if ($bootorder =~ /setup/) {
 | |
|         unless ($bootorder eq 'setup') {
 | |
|             xCAT::SvrUtils::sendmsg([ 1, "rsetboot parameter may not contain 'setup' with other items, assuming vm.bootorder is just 'setup'" ], $output_handler, $node);
 | |
|         }
 | |
|         $reconfigspec = VirtualMachineConfigSpec->new(
 | |
|             bootOptions => VirtualMachineBootOptions->new(enterBIOSSetup => 1),
 | |
|         );
 | |
|     } else {
 | |
|         $bootorder    = "allow:" . $bootorder;
 | |
|         $reconfigspec = VirtualMachineConfigSpec->new(
 | |
|             bootOptions => VirtualMachineBootOptions->new(enterBIOSSetup => 0),
 | |
|             extraConfig => [ OptionValue->new(key => 'bios.bootDeviceClasses', value => $bootorder) ]
 | |
|         );
 | |
|     }
 | |
|     my $task = $args{vmview}->ReconfigVM_Task(spec => $reconfigspec);
 | |
|     $running_tasks{$task}->{task}     = $task;
 | |
|     $running_tasks{$task}->{callback} = \&generic_task_callback;
 | |
|     $running_tasks{$task}->{hyp}      = $args{hyp};
 | |
|     $running_tasks{$task}->{vm}       = $args{node};
 | |
|     $running_tasks{$task}->{data} = { node => $node, successtext => ${ $args{exargs} }[0] };
 | |
| }
 | |
| 
 | |
| sub register_vm {    #Attempt to register existing instance of a VM
 | |
|     my $hyp         = shift;
 | |
|     my $node        = shift;
 | |
|     my $disksize    = shift;
 | |
|     my $blockedfun  = shift; #a pointer to a blocked function to call on success
 | |
|     my $blockedargs = shift; #hash reference to call blocked function with
 | |
|     my $failonerr   = shift;
 | |
|     my %args = @_; #ok, went overboard with positional arguments, from now on, named arguments
 | |
|     my $task;
 | |
| 
 | |
|     if ($hyp) {
 | |
|         validate_network_prereqs([ keys %{ $hyphash{$hyp}->{nodes} } ], $hyp);
 | |
|         unless (defined $hyphash{$hyp}->{datastoremap} or validate_datastore_prereqs([ keys %{ $hyphash{$hyp}->{nodes} } ], $hyp)) {
 | |
|             die "unexpected condition";
 | |
|         }
 | |
|     } else {
 | |
|         scan_cluster_networks($args{cluster});
 | |
|     }
 | |
| 
 | |
|     sortoutdatacenters(nodes => [$node], hyp => $hyp, cluster => $args{cluster});
 | |
|     my $placement_resources = get_placement_resources(hyp => $hyp, cluster => $args{cluster});
 | |
| 
 | |
|     # Try to add an existing VM to the machine folder
 | |
|     my $success = eval {
 | |
|         if ($hyp) {
 | |
|             $task = $vmhash{$node}->{vmfolder}->RegisterVM_Task(path => getcfgdatastore($node, $hyphash{$hyp}->{datastoremap}) . " /$node/$node.vmx", name => $node, pool => $hyphash{$hyp}->{pool}, host => $hyphash{$hyp}->{hostview}, asTemplate => 0);
 | |
|         } else {
 | |
|             $task = $vmhash{$node}->{vmfolder}->RegisterVM_Task(path => getcfgdatastore($node, $clusterhash{ $args{cluster} }->{datastoremap}) . " /$node/$node.vmx", name => $node, pool => $placement_resources->{pool}, asTemplate => 0);
 | |
|         }
 | |
|     };
 | |
| 
 | |
|     # if we couldn't add it then it means it wasn't created yet.  So we create it.
 | |
|     my $cluster = $args{cluster};
 | |
|     if ($@ or not $success) {
 | |
| 
 | |
|         #if (ref($@) eq 'SoapFault') {
 | |
|         # if (ref($@->detail) eq 'NotFound') {
 | |
|         register_vm_callback(undef, {
 | |
|                 node        => $node,
 | |
|                 disksize    => $disksize,
 | |
|                 blockedfun  => $blockedfun,
 | |
|                 blockedargs => $blockedargs,
 | |
|                 errregister => $failonerr,
 | |
|                 cpus        => $args{cpus},
 | |
|                 memory      => $args{memory},
 | |
|                 hyp         => $hyp,
 | |
|                 cluster     => $cluster,
 | |
|         });
 | |
|     }
 | |
|     if ($task) {
 | |
|         $running_tasks{$task}->{task}     = $task;
 | |
|         $running_tasks{$task}->{callback} = \®ister_vm_callback;
 | |
|         $running_tasks{$task}->{hyp}      = $hyp;
 | |
|         $running_tasks{$task}->{cluster}  = $cluster;
 | |
|         $running_tasks{$task}->{data}     = {
 | |
|             node        => $node,
 | |
|             disksize    => $disksize,
 | |
|             blockedfun  => $blockedfun,
 | |
|             blockedargs => $blockedargs,
 | |
|             errregister => $failonerr,
 | |
|             cpus        => $args{cpus},
 | |
|             memory      => $args{memory},
 | |
|             hyp         => $hyp,
 | |
|             cluster     => $cluster,
 | |
|         };
 | |
|     }
 | |
| }
 | |
| 
 | |
| sub register_vm_callback {
 | |
|     my $task = shift;
 | |
|     my $args = shift;
 | |
|     if (not $task or $task->info->state->val eq 'error') { #TODO: fail for 'rpower' flow, mkvm is too invasive in VMWare to be induced by 'rpower on'
 | |
|         if (not defined $args->{blockedfun}) {
 | |
|             mknewvm($args->{node}, $args->{disksize}, $args->{hyp}, $args);
 | |
|         } elsif ($args->{errregister}) {
 | |
|             relay_vmware_err($task, "", $args->{node});
 | |
|         } else {
 | |
|             xCAT::SvrUtils::sendmsg([ 1, "mkvm must be called before use of this function" ], $output_handler, $args->{node});
 | |
|         }
 | |
|     } elsif (defined $args->{blockedfun}) { #If there is a blocked function, call it here)
 | |
|         $args->{blockedfun}->(%{ $args->{blockedargs} });
 | |
|     }
 | |
| }
 | |
| 
 | |
| 
 | |
| sub getURI {
 | |
|     my $method   = shift;
 | |
|     my $location = shift;
 | |
|     my $uri      = '';
 | |
| 
 | |
|     if ($method =~ /nfs/) {
 | |
| 
 | |
|         (my $server, my $path) = split /\//, $location, 2;
 | |
|         $server =~ s/:$//;    #tolerate habitual colons
 | |
|         my $servern = inet_aton($server);
 | |
|         unless ($servern) {
 | |
|             xCAT::SvrUtils::sendmsg([ 1, "could not resolve '$server' to an address from vm.storage/vm.cfgstore" ], $output_handler);
 | |
|         }
 | |
|         $server = inet_ntoa($servern);
 | |
|         $uri    = "nfs://$server/$path";
 | |
|     } elsif ($method =~ /vmfs/) {
 | |
|         (my $name, undef) = split /\//, $location, 2;
 | |
|         $name =~ s/:$//;      #remove a : if someone put it in for some reason.
 | |
|         $uri = "vmfs://$name";
 | |
|     } else {
 | |
|         xCAT::SvrUtils::sendmsg([ 1, "Unsupported VMware Storage Method: $method.  Please use 'vmfs or nfs'" ], $output_handler);
 | |
|     }
 | |
| 
 | |
|     return $uri;
 | |
| }
 | |
| 
 | |
| 
 | |
| sub getcfgdatastore {
 | |
|     my $node         = shift;
 | |
|     my $dses         = shift;
 | |
|     my $cfgdatastore = $tablecfg{vm}->{$node}->[0]->{cfgstore};
 | |
|     unless ($cfgdatastore) {
 | |
|         $cfgdatastore = $tablecfg{vm}->{$node}->[0]->{storage};
 | |
| 
 | |
|         #TODO: if multiple drives are specified, make sure to split this out
 | |
|         #DONE: I believe the regex after this conditional takes care of that case already..
 | |
|     }
 | |
|     $cfgdatastore =~ s/=.*//;
 | |
|     (my $method, my $location) = split /:\/\//, $cfgdatastore, 2;
 | |
|     my $uri = $cfgdatastore;
 | |
|     unless ($dses->{$uri}) {    #don't call getURI if map works out fine already
 | |
|         $uri = getURI($method, $location);
 | |
|     }
 | |
|     $cfgdatastore = "[" . $dses->{$uri} . "]";
 | |
| 
 | |
|     #$cfgdatastore =~ s/,.*$//; #these two lines of code were kinda pointless
 | |
|     #$cfgdatastore =~ s/\/$//;
 | |
|     return $cfgdatastore;
 | |
| }
 | |
| 
 | |
| 
 | |
| sub mknewvm {
 | |
|     my $node      = shift;
 | |
|     my $disksize  = shift;
 | |
|     my $hyp       = shift;
 | |
|     my $otherargs = shift;
 | |
|     my $cluster   = $otherargs->{cluster};
 | |
|     my $placement_resources = get_placement_resources(hyp => $hyp, cluster => $cluster);
 | |
|     my $pool = $placement_resources->{pool};
 | |
|     my $cfg;
 | |
| 
 | |
|     if ($hyp) {
 | |
|         $cfg = build_cfgspec($node, $hyphash{$hyp}->{datastoremap}, $hyphash{$hyp}->{nets}, $disksize, $hyp, $otherargs);
 | |
|     } else {    #cluster based..
 | |
|         $cfg = build_cfgspec($node, $clusterhash{$cluster}->{datastoremap}, $clusterhash{$cluster}->{nets}, $disksize, $hyp, $otherargs);
 | |
|     }
 | |
|     my $task;
 | |
|     if ($hyp) {
 | |
|         $task = $vmhash{$node}->{vmfolder}->CreateVM_Task(config => $cfg, pool => $hyphash{$hyp}->{pool}, host => $hyphash{$hyp}->{hostview});
 | |
|     } else {
 | |
|         $task = $vmhash{$node}->{vmfolder}->CreateVM_Task(config => $cfg, pool => $pool); #drs away
 | |
|     }
 | |
|     $running_tasks{$task}->{task}     = $task;
 | |
|     $running_tasks{$task}->{callback} = \&mkvm_callback;
 | |
|     $running_tasks{$task}->{hyp}      = $hyp;
 | |
|     $running_tasks{$task}->{cluster}  = $cluster;
 | |
|     $running_tasks{$task}->{data} = { hyp => $hyp, cluster => $cluster, node => $node };
 | |
| }
 | |
| 
 | |
| 
 | |
| sub getUnits {
 | |
|     my $amount  = shift;
 | |
|     my $defunit = shift;
 | |
|     my $divisor = shift;
 | |
|     unless ($amount) { return; }
 | |
|     unless ($divisor) {
 | |
|         $divisor = 1;
 | |
|     }
 | |
|     if ($amount =~ /(\D)$/) {    #If unitless, add unit
 | |
|         $defunit = $1;
 | |
|         chop $amount;
 | |
|     }
 | |
|     if ($defunit =~ /k/i) {
 | |
|         return $amount * 1024 / $divisor;
 | |
|     } elsif ($defunit =~ /m/i) {
 | |
|         return $amount * 1048576 / $divisor;
 | |
|     } elsif ($defunit =~ /g/i) {
 | |
|         return $amount * 1073741824 / $divisor;
 | |
|     }
 | |
| }
 | |
| 
 | |
| 
 | |
| sub getguestid {
 | |
|     my $osfound = 0;
 | |
|     my $node    = shift;
 | |
|     if ($tablecfg{vm}->{$node}->[0]->{guestostype}) { #if admin wants to skip derivation from nodetype.os value, let em
 | |
|         return $tablecfg{vm}->{$node}->[0]->{guestostype};
 | |
|     }
 | |
|     my $nodeos   = $tablecfg{nodetype}->{$node}->[0]->{os};
 | |
|     my $nodearch = $tablecfg{nodetype}->{$node}->[0]->{arch};
 | |
|     foreach (keys %guestidmap) {
 | |
|         if (defined($nodeos) and $nodeos =~ /$_/) {
 | |
|             if ($nodearch eq 'x86_64' and $_ !~ /vmkernel/) {
 | |
|                 $nodeos = $guestidmap{$_} . "64Guest";
 | |
|             } else {
 | |
|                 $nodeos = $guestidmap{$_};
 | |
|                 $nodeos =~ s/_$//;
 | |
|                 $nodeos .= "Guest";
 | |
|             }
 | |
|             $osfound = 1;
 | |
|             last;
 | |
|         }
 | |
|     }
 | |
|     unless ($osfound) {
 | |
|         if (defined($nodearch) and $nodearch eq 'x86_64') {
 | |
|             $nodeos = "otherGuest64";
 | |
|         } else {
 | |
|             $nodeos = "otherGuest";
 | |
|         }
 | |
|     }
 | |
|     return $nodeos;
 | |
| }
 | |
| 
 | |
| sub build_cfgspec {
 | |
|     my $node      = shift;
 | |
|     my $dses      = shift;    #map to match vm table to datastore names
 | |
|     my $netmap    = shift;
 | |
|     my $disksize  = shift;
 | |
|     my $hyp       = shift;
 | |
|     my $otherargs = shift;
 | |
|     my $memory;
 | |
|     my $ncpus;
 | |
|     my $updatehash;
 | |
| 
 | |
|     if ($otherargs->{memory}) {
 | |
|         $memory = getUnits($otherargs->{memory}, "M", 1048576);
 | |
|         if ($tablecfg{vm}->{$node}->[0]->{memory}) {
 | |
|             $updatehash->{memory} = $memory;
 | |
|         }
 | |
|     } elsif ($tablecfg{vm}->{$node}->[0]->{memory}) {
 | |
|         $memory = getUnits($tablecfg{vm}->{$node}->[0]->{memory}, "M", 1048576);
 | |
|     } else {
 | |
|         $memory = 512;
 | |
|     }
 | |
|     if ($otherargs->{cpus}) {
 | |
|         $ncpus = $otherargs->{cpus};
 | |
|         if ($tablecfg{vm}->{$node}->[0]->{cpus}) {
 | |
|             $updatehash->{cpus} = $ncpus;
 | |
|         }
 | |
|     } elsif ($tablecfg{vm}->{$node}->[0]->{cpus}) {
 | |
|         $ncpus = $tablecfg{vm}->{$node}->[0]->{cpus};
 | |
|     } else {
 | |
|         $ncpus = 1;
 | |
|     }
 | |
|     if ($updatehash) {
 | |
|         my $vmtab = xCAT::Table->new('vm', -create => 1);
 | |
|         $vmtab->setNodeAttribs($node, $updatehash);
 | |
|     }
 | |
|     my @devices;
 | |
|     $currkey = 0;
 | |
|     my $opticalbacking = VirtualCdromRemoteAtapiBackingInfo->new(deviceName => "");
 | |
|     my $opticalconnectable = VirtualDeviceConnectInfo->new(startConnected => 0, allowGuestControl => 1, connected => 0);
 | |
|     my $optical = VirtualCdrom->new(controllerKey => 201,
 | |
|         connectable => $opticalconnectable,
 | |
|         backing     => $opticalbacking,
 | |
|         key         => $currkey++,
 | |
|         unitNumber  => 0,);
 | |
|     push @devices, VirtualDeviceConfigSpec->new(device => $optical, operation => VirtualDeviceConfigSpecOperation->new('add'));
 | |
|     push @devices, create_storage_devs($node, $dses, $disksize);
 | |
|     push @devices, create_nic_devs($node, $netmap, $hyp);
 | |
| 
 | |
|     #my $cfgdatastore = $tablecfg{vm}->{$node}->[0]->{storage}; #TODO: need a new cfglocation field in case of stateless guest?
 | |
|     #$cfgdatastore =~ s/,.*$//;
 | |
|     #$cfgdatastore =~ s/\/$//;
 | |
|     #$cfgdatastore = "[".$dses->{$cfgdatastore}."]";
 | |
|     my $cfgdatastore = getcfgdatastore($node, $dses);
 | |
|     my $vfiles = VirtualMachineFileInfo->new(vmPathName => $cfgdatastore);
 | |
| 
 | |
|     #my $nodeos = $tablecfg{nodetype}->{$node}->[0]->{os};
 | |
|     #my $nodearch = $tablecfg{nodetype}->{$node}->[0]->{arch};
 | |
|     my $nodeos = getguestid($node);    #nodeos=>$nodeos,nodearch=>$nodearch);
 | |
|     my $uuid;
 | |
|     if ($tablecfg{vpd}->{$node}->[0]->{uuid}) {
 | |
|         $uuid = $tablecfg{vpd}->{$node}->[0]->{uuid};
 | |
|     } else {
 | |
|         if ($tablecfg{mac}->{$node}->[0]->{mac}) { #a uuidv1 is possible, generate that for absolute uniqueness guarantee
 | |
|             my $mac = $tablecfg{mac}->{$node}->[0]->{mac};
 | |
|             $mac =~ s/\|.*//;
 | |
|             $mac =~ s/!.*//;
 | |
|             $uuid = xCAT::Utils::genUUID(mac => $mac);
 | |
|         } else {
 | |
|             $uuid = xCAT::Utils::genUUID();
 | |
|         }
 | |
| 
 | |
|         my $vpdtab = xCAT::Table->new('vpd');
 | |
|         $vpdtab->setNodeAttribs($node, { uuid => $uuid });
 | |
|     }
 | |
|     $uuid =~ s/^(..)(..)(..)(..)-(..)(..)-(..)(..)/$4$3$2$1-$6$5-$8$7/;
 | |
|     my @optionvals;
 | |
|     if ($tablecfg{vm}->{$node}->[0]->{othersettings}) {
 | |
|         my $key;
 | |
|         my $value;
 | |
|         foreach (split /;/, $tablecfg{vm}->{$node}->[0]->{othersettings}) {
 | |
|             ($key, $value) = split /=/;
 | |
|             if ($value) {
 | |
|                 push @optionvals, OptionValue->new(key => $key, value => $value);
 | |
|             } else {
 | |
|                 push @optionvals, OptionValue->new(key => $key);
 | |
|             }
 | |
|         }
 | |
|     }
 | |
|     my %specargs = (
 | |
|         name         => $node,
 | |
|         files        => $vfiles,
 | |
|         guestId      => $nodeos,
 | |
|         memoryMB     => $memory,
 | |
|         numCPUs      => $ncpus,
 | |
|         deviceChange => \@devices,
 | |
|         uuid         => $uuid,
 | |
|     );
 | |
|     if (@optionvals) {
 | |
|         $specargs{extraConfig} = \@optionvals;
 | |
|     }
 | |
|     return VirtualMachineConfigSpec->new(%specargs);
 | |
| }
 | |
| 
 | |
| sub create_nic_devs {
 | |
|     my $node     = shift;
 | |
|     my $netmap   = shift;
 | |
|     my $hyp      = shift;
 | |
|     my @networks = split /,/, $tablecfg{vm}->{$node}->[0]->{nics};
 | |
|     my @devs;
 | |
|     my $idx = 0;
 | |
|     my @macs = xCAT::VMCommon::getMacAddresses(\%tablecfg, $node, scalar @networks);
 | |
|     my $connprefs = VirtualDeviceConnectInfo->new(
 | |
|         allowGuestControl => 1,
 | |
|         connected         => 0,
 | |
|         startConnected    => 1
 | |
|     );
 | |
|     my $model = $tablecfg{vm}->{$node}->[0]->{nicmodel};
 | |
|     unless ($model) {
 | |
|         $model = 'e1000';
 | |
|     }
 | |
|     foreach (@networks) {
 | |
|         my $pgname = $_;
 | |
|         if ($hyp) {
 | |
|             $pgname = $hyphash{$hyp}->{pgnames}->{$_};
 | |
|         }
 | |
|         s/.*://;
 | |
|         my $hadspecmodel = 0;
 | |
|         if (m/=/) {
 | |
|             $hadspecmodel = 1;
 | |
|             s/=(.*)$//;
 | |
|         }
 | |
|         my $tmpmodel = $model;
 | |
|         if ($hadspecmodel) { $tmpmodel = $1; }
 | |
|         my $netname = $_;
 | |
|         my $backing = VirtualEthernetCardNetworkBackingInfo->new(
 | |
|             network    => $netmap->{$pgname},
 | |
|             deviceName => $pgname,
 | |
|         );
 | |
|         my %newcardargs = (
 | |
|             key              => 0,             #3, #$currkey++,
 | |
|             backing          => $backing,
 | |
|             addressType      => "manual",
 | |
|             macAddress       => shift @macs,
 | |
|             connectable      => $connprefs,
 | |
|             wakeOnLanEnabled => 1,             #TODO: configurable in tables?
 | |
|         );
 | |
|         my $newcard;
 | |
|         if ($tmpmodel eq 'e1000') {
 | |
|             $newcard = VirtualE1000->new(%newcardargs);
 | |
|         } elsif ($tmpmodel eq 'vmxnet3') {
 | |
|             $newcard = VirtualVmxnet3->new(%newcardargs);
 | |
|         } elsif ($tmpmodel eq 'pcnet32') {
 | |
|             $newcard = VirtualPCNet32->new(%newcardargs);
 | |
|         } elsif ($tmpmodel eq 'vmxnet2') {
 | |
|             $newcard = VirtualVmxnet2->new(%newcardargs);
 | |
|         } elsif ($tmpmodel eq 'vmxnet') {
 | |
|             $newcard = VirtualVmxnet->new(%newcardargs);
 | |
|         } else {
 | |
|             xCAT::SvrUtils::sendmsg([ 1, "$tmpmodel not a recognized nic type, falling back to e1000 (vmxnet3, e1000, pcnet32, vmxnet2, vmxnet are recognized" ], $output_handler, $node);
 | |
|             $newcard = VirtualE1000->new(%newcardargs);
 | |
|         }
 | |
| 
 | |
|         push @devs, VirtualDeviceConfigSpec->new(device => $newcard,
 | |
|             operation => VirtualDeviceConfigSpecOperation->new('add'));
 | |
|         $idx++;
 | |
|     }
 | |
|     return @devs;
 | |
|     die "Stop running for test";
 | |
| }
 | |
| 
 | |
| sub create_storage_devs {
 | |
|     my $node              = shift;
 | |
|     my $sdmap             = shift;
 | |
|     my $sizes             = shift;
 | |
|     my @sizes             = split /[,:]/, $sizes;
 | |
|     my $existingScsiCont  = shift;
 | |
|     my $scsiUnit          = shift;
 | |
|     my $existingIdeCont   = shift;
 | |
|     my $ideUnit           = shift;
 | |
|     my $devices           = shift;
 | |
|     my %args              = @_;
 | |
|     my $scsicontrollerkey = 0;
 | |
|     my $idecontrollerkey = 200; #IDE 'controllers' exist at 200 and 201 invariably, with no flexibility?
 | |
|       #Cannot find documentation that declares this absolute, but attempts to do otherwise
 | |
|       #lead in failure, also of note, these are single-channel controllers, so two devs per controller
 | |
| 
 | |
|     my $backingif;
 | |
|     my @devs;
 | |
|     my $havescsidevs       = 0;
 | |
|     my $disktype           = 'ide';
 | |
|     my $ideunitnum         = 0;
 | |
|     my $scsiunitnum        = 0;
 | |
|     my $havescsicontroller = 0;
 | |
|     my %usedideunits;
 | |
|     my %usedscsiunits = (7 => 1, '7' => 1);
 | |
| 
 | |
|     if (defined $existingScsiCont) {
 | |
|         $havescsicontroller = 1;
 | |
|         $scsicontrollerkey  = $existingScsiCont->{key};
 | |
|         $scsiunitnum        = $scsiUnit;
 | |
|         %usedscsiunits      = %{ getUsedUnits($scsicontrollerkey, $devices) };
 | |
|     }
 | |
|     if (defined $existingIdeCont) {
 | |
|         $idecontrollerkey = $existingIdeCont->{key};
 | |
|         $ideunitnum       = $ideUnit;
 | |
|         %usedideunits     = %{ getUsedUnits($idecontrollerkey, $devices) };
 | |
|     }
 | |
|     my $unitnum;
 | |
|     my %disktocont;
 | |
|     my $dev;
 | |
|     my @storelocs = split /,/, $tablecfg{vm}->{$node}->[0]->{storage};
 | |
|     my $globaldisktype = $tablecfg{vm}->{$node}->[0]->{storagemodel};
 | |
|     unless ($globaldisktype) { $globaldisktype = 'ide'; }
 | |
| 
 | |
|     #number of devices is the larger of the specified sizes (TODO: masters) or storage pools to span
 | |
|     my $numdevs = (scalar @storelocs > scalar @sizes ? scalar @storelocs : scalar @sizes);
 | |
|     my $controllertype = 'scsi';
 | |
|     while ($numdevs-- > 0) {
 | |
|         my $storeloc = shift @storelocs;
 | |
|         unless (scalar @storelocs) { @storelocs = ($storeloc); } #allow reuse of one cfg specified pool for multiple devs
 | |
|         my $disksize = shift @sizes;
 | |
|         unless (scalar @sizes) { @sizes = ($disksize); } #if we emptied the array, stick the last entry back on to allow it to specify all remaining disks
 | |
|         $disksize = getUnits($disksize, 'G', 1024);
 | |
|         $disktype = $globaldisktype;
 | |
|         if ($storeloc =~ /=/) {
 | |
|             ($storeloc, $disktype) = split /=/, $storeloc;
 | |
|         }
 | |
|         if ($disktype eq 'ide' and $args{idefull}) {
 | |
|             xCAT::SvrUtils::sendmsg([ 1, "VM is at capacity for IDE devices, a drive was not added" ], $output_handler, $node);
 | |
|             return;
 | |
|         } elsif (($disktype eq 'scsi' or $disktype eq 'sas' or $disktype eq 'pvscsi') and $args{scsifull}) {
 | |
|             xCAT::SvrUtils::sendmsg([ 1, "SCSI Controller at capacity, a drive was not added" ], $output_handler, $node);
 | |
|             return;
 | |
|         }
 | |
| 
 | |
|         $storeloc =~ s/\/$//;
 | |
|         (my $method, my $location) = split /:\/\//, $storeloc, 2;
 | |
|         my $uri = $storeloc;
 | |
|         unless ($sdmap->{$uri}) { #don't call getURI if map works out fine already
 | |
|             $uri = getURI($method, $location);
 | |
|         }
 | |
| 
 | |
|         #(my $server,my $path) = split/\//,$location,2;
 | |
|         #$server =~ s/:$//; #tolerate habitual colons
 | |
|         #my $servern = inet_aton($server);
 | |
|         #unless ($servern) {
 | |
|         #    xCAT::SvrUtils::sendmsg([1,"could not resolve '$server' to an address from vm.storage"]);
 | |
|         #    return;
 | |
|         #}
 | |
|         #$server = inet_ntoa($servern);
 | |
|         #my $uri = "nfs://$server/$path";
 | |
|         $backingif = VirtualDiskFlatVer2BackingInfo->new(diskMode => 'persistent',
 | |
|             thinProvisioned => 1,
 | |
|             fileName        => "[" . $sdmap->{$uri} . "]");
 | |
|         if ($disktype eq 'ide' and $idecontrollerkey == 1 and $ideunitnum == 0) { #reserve a spot for CD
 | |
|             $ideunitnum = 1;
 | |
|         } elsif ($disktype eq 'ide' and $ideunitnum == 2) { #go from current to next ide 'controller'
 | |
|             $idecontrollerkey++;
 | |
|             $ideunitnum = 0;
 | |
|         }
 | |
|         unless ($disktype eq 'ide') {
 | |
|             push @{ $disktocont{$scsicontrollerkey} }, $currkey;
 | |
|         }
 | |
|         my $controllerkey;
 | |
|         if ($disktype eq 'ide') {
 | |
|             $controllerkey = $idecontrollerkey;
 | |
|             $unitnum       = 0;
 | |
|             while ($usedideunits{$unitnum}) {
 | |
|                 $unitnum++;
 | |
|             }
 | |
|             if ($unitnum == 2) {
 | |
|                 $idecontrollerkey++;
 | |
|                 $ideunitnum    = 1;
 | |
|                 $unitnum       = 1;
 | |
|                 $controllerkey = $idecontrollerkey;
 | |
|             }
 | |
|             $usedideunits{$unitnum} = 1;
 | |
|         } else {
 | |
|             $controllertype = $disktype;
 | |
|             $controllerkey  = $scsicontrollerkey;
 | |
|             $unitnum        = 0;
 | |
|             while ($usedscsiunits{$unitnum}) {
 | |
|                 $unitnum++;
 | |
|             }
 | |
|             $usedscsiunits{$unitnum} = 1;
 | |
|             $havescsidevs = 1;
 | |
|         }
 | |
| 
 | |
|         $dev = VirtualDisk->new(backing => $backingif,
 | |
|             controllerKey => $controllerkey,
 | |
|             key           => $currkey++,
 | |
|             unitNumber    => $unitnum,
 | |
|             capacityInKB  => $disksize);
 | |
|         push @devs, VirtualDeviceConfigSpec->new(device => $dev,
 | |
|             fileOperation => VirtualDeviceConfigSpecFileOperation->new('create'),
 | |
|             operation => VirtualDeviceConfigSpecOperation->new('add'));
 | |
|     }
 | |
| 
 | |
|     #It *seems* that IDE controllers are not subject to require creation, so we skip it
 | |
|     if ($havescsidevs and not $havescsicontroller) { #need controllers to attach the disks to
 | |
|         foreach (0 .. $scsicontrollerkey) {
 | |
|             if ($controllertype eq 'scsi') {
 | |
|                 $dev = VirtualLsiLogicController->new(key => $_,
 | |
|                     device    => \@{ $disktocont{$_} },
 | |
|                     sharedBus => VirtualSCSISharing->new('noSharing'),
 | |
|                     busNumber => $_);
 | |
|             } elsif ($controllertype eq 'sas') {
 | |
|                 $dev = VirtualLsiLogicSASController->new(key => $_,
 | |
|                     device    => \@{ $disktocont{$_} },
 | |
|                     sharedBus => VirtualSCSISharing->new('noSharing'),
 | |
|                     busNumber => $_);
 | |
|             } elsif ($controllertype eq 'pvscsi') {
 | |
|                 $dev = ParaVirtualSCSIController->new(key => $_,
 | |
|                     device    => \@{ $disktocont{$_} },
 | |
|                     sharedBus => VirtualSCSISharing->new('noSharing'),
 | |
|                     busNumber => $_);
 | |
|             }
 | |
| 
 | |
|             push @devs, VirtualDeviceConfigSpec->new(device => $dev,
 | |
|                 operation => VirtualDeviceConfigSpecOperation->new('add'));
 | |
| 
 | |
|         }
 | |
|     }
 | |
|     return @devs;
 | |
| 
 | |
|     #    my $ctlr = VirtualIDEController->new(
 | |
| }
 | |
| 
 | |
| sub declare_ready {
 | |
|     my %args = %{ shift() };
 | |
|     $hypready{ $args{hyp} } = 1;
 | |
| }
 | |
| 
 | |
| sub populate_vcenter_hostviews {
 | |
|     my $vcenter = shift;
 | |
|     my @hypervisors;
 | |
|     my %nametohypmap;
 | |
|     my $iterations = 1;
 | |
|     if ($usehostnamesforvcenter and $usehostnamesforvcenter !~ /no/i) {
 | |
|         $iterations = 2;    #two passes possible
 | |
|         my $hyp;
 | |
|         foreach $hyp (keys %{ $vcenterhash{$vcenter}->{allhyps} }) {
 | |
| 
 | |
|             if ($tablecfg{hosts}->{$hyp}->[0]->{hostnames}) {
 | |
|                 $nametohypmap{ $tablecfg{hosts}->{$hyp}->[0]->{hostnames} } = $hyp;
 | |
|             }
 | |
|         }
 | |
|         @hypervisors = keys %nametohypmap;
 | |
|     } else {
 | |
|         @hypervisors = keys %{ $vcenterhash{$vcenter}->{allhyps} };
 | |
|     }
 | |
|     while ($iterations and scalar(@hypervisors)) {
 | |
|         my $hosts = join(")|(", @hypervisors);
 | |
|         $hosts = '^((' . $hosts . '))(\z|\.)';
 | |
|         my $search = qr/$hosts/;
 | |
|         my @hypviews = @{ $vcenterhash{$vcenter}->{conn}->find_entity_views(view_type => 'HostSystem', properties => [ 'summary.config.name', 'summary.runtime.connectionState', 'runtime.inMaintenanceMode', 'parent', 'configManager', 'summary.host' ], filter => { 'summary.config.name' => $search }) };
 | |
|         foreach (@hypviews) {
 | |
|             my $hypname = $_->{'summary.config.name'};
 | |
|             my $hypv    = $_;
 | |
|             my $hyp;
 | |
|             if ($vcenterhash{$vcenter}->{allhyps}->{$hypname}) { #simplest case, config.name is exactly the same as node name
 | |
|                 $vcenterhash{$vcenter}->{hostviews}->{$hypname} = $_;
 | |
|                 $hyp = $hypname;
 | |
|             } elsif ($nametohypmap{$hypname}) { #second case, there is a name mapping this to a real name
 | |
|                 $vcenterhash{$vcenter}->{hostviews}->{ $nametohypmap{$hypname} } = $_;
 | |
|                 $hyp = $nametohypmap{$hypname};
 | |
|             } else { #name as-is doesn't work, start stripping domain and hope for the best
 | |
|                 $hypname =~ s/\..*//;
 | |
|                 if ($vcenterhash{$vcenter}->{allhyps}->{$hypname}) { #shortname is a node
 | |
|                     $vcenterhash{$vcenter}->{hostviews}->{$hypname} = $_;
 | |
|                     $hyp = $hypname;
 | |
|                 } elsif ($nametohypmap{$hypname}) {    #alias for node
 | |
|                     $vcenterhash{$vcenter}->{hostviews}->{ $nametohypmap{$hypname} } = $_;
 | |
|                     $hyp = $nametohypmap{$hypname};
 | |
|                 }
 | |
|             }
 | |
|             foreach my $nodename (keys %{ $hyphash{$hyp}->{nodes} }) {
 | |
|                 $hostrefbynode{$nodename} = $hypv->{'summary.host'}->value;
 | |
|             }
 | |
|         }
 | |
|         $iterations--;
 | |
|         @hypervisors = ();
 | |
|         if ($usehostnamesforvcenter and $usehostnamesforvcenter !~ /no/i) { #check for hypervisors by native node name if missed above
 | |
|             foreach my $hyp (keys %{ $vcenterhash{$vcenter}->{allhyps} }) {
 | |
|                 unless ($vcenterhash{$vcenter}->{hostviews}->{$hyp}) {
 | |
|                     push @hypervisors, $hyp;
 | |
|                 }
 | |
|             }
 | |
|         }
 | |
|     }
 | |
| }
 | |
| 
 | |
| sub create_new_cluster {
 | |
|     my $req = shift;
 | |
|     @ARGV = @{ $req->{arg} };
 | |
|     my $vcenter;
 | |
|     my $password;
 | |
|     my $user;
 | |
|     my $datacenter;
 | |
|     GetOptions(
 | |
|         'vcenter=s'    => \$vcenter,
 | |
|         'password=s'   => \$password,
 | |
|         'datacenter=s' => \$datacenter,
 | |
|         'username=s'   => \$user,
 | |
|     );
 | |
|     my $clustername = shift @ARGV;
 | |
|     my $conn = Vim->new(service_url => "https://$vcenter/sdk");
 | |
|     $conn->login(user_name => $user, password => $password);
 | |
|     if ($datacenter) {
 | |
|         $datacenter = $conn->find_entity_view(view_type => 'Datacenter', properties => ['hostFolder'], filter => { name => $datacenter });
 | |
|         unless ($datacenter) {
 | |
|             xCAT::SvrUtils::sendmsg([ 1, ": Unable to find requested datacenter" ], $output_handler);
 | |
|             return;
 | |
|         }
 | |
|     } else {
 | |
|         $datacenter = $conn->find_entity_view(view_type => 'Datacenter', properties => ['hostFolder']);
 | |
|     }
 | |
|     my $hfolder = $conn->get_view(mo_ref => $datacenter->hostFolder);
 | |
|     my $cfgspec = ClusterConfigSpecEx->new();
 | |
|     $hfolder->CreateClusterEx(name => $clustername, spec => $cfgspec);
 | |
| }
 | |
| 
 | |
| sub remove_cluster {
 | |
|     my $req = shift;
 | |
|     @ARGV = @{ $req->{arg} };
 | |
|     my $vcenter;
 | |
|     my $user;
 | |
|     my $password;
 | |
|     my $clustername;
 | |
|     GetOptions(
 | |
|         'vcenter=s'  => \$vcenter,
 | |
|         'password=s' => \$password,
 | |
|         'username=s' => \$user,
 | |
|     );
 | |
|     $clustername = shift @ARGV;
 | |
|     my $conn = Vim->new(service_url => "https://$vcenter/sdk");
 | |
|     $conn->login(user_name => $user, password => $password);
 | |
| 
 | |
|     #  $clustview = $hyphash{$hyp}->{vcenter}->{conn}->find_entity_view(view_type => 'Datacenter', properties=>['hostFolder'],filter=>{name=>$tablecfg{hypervisor}->{$hyp}->[0]->{datacenter}});
 | |
|     #my $conn = Vim->new(service_url=>"https://$vcenter/sdk");
 | |
|     $conn->login(user_name => $user, password => $password);
 | |
|     my $clustview = $conn->find_entity_view(view_type => 'ClusterComputeResource', filter => { name => $clustername });
 | |
|     my $task = $clustview->Destroy_Task();
 | |
|     my $done = 0;
 | |
|     while (not $done) {
 | |
|         my $curt = $conn->get_view(mo_ref => $task);
 | |
|         my $state = $curt->info->state->val;
 | |
|         unless ($state eq 'running' or $state eq 'queued') {
 | |
|             $done = 1;
 | |
|         }
 | |
|     }
 | |
| }
 | |
| 
 | |
| 
 | |
| sub list_clusters {
 | |
|     my $req = shift;
 | |
|     @ARGV = @{ $req->{arg} };
 | |
|     my $vcenter;
 | |
|     my $password;
 | |
|     my $user;
 | |
|     my $datacenter;
 | |
|     GetOptions(
 | |
|         'vcenter=s'    => \$vcenter,
 | |
|         'password=s'   => \$password,
 | |
|         'datacenter=s' => \$datacenter,
 | |
|         'username=s'   => \$user,
 | |
|     );
 | |
|     my $clustername = shift @ARGV;
 | |
|     my $conn = Vim->new(service_url => "https://$vcenter/sdk");
 | |
|     $conn->login(user_name => $user, password => $password);
 | |
|     use Data::Dumper;
 | |
|     my $clustviews = $conn->find_entity_views(view_type => 'ClusterComputeResource');
 | |
|     foreach (@$clustviews) {
 | |
|         xCAT::SvrUtils::sendmsg($_->{name}, $output_handler);
 | |
|     }
 | |
|     return;
 | |
| }
 | |
| 
 | |
| sub validate_vcenter_prereqs { #Communicate with vCenter and ensure this host is added correctly to a vCenter instance when an operation requires it
 | |
|     my $hyp     = shift;
 | |
|     my $depfun  = shift;
 | |
|     my $depargs = shift;
 | |
|     my $vcenter = $hyphash{$hyp}->{vcenter}->{name};
 | |
|     unless ($hyphash{$hyp}->{vcenter}->{conn}) {
 | |
|         eval {
 | |
|             $hyphash{$hyp}->{vcenter}->{conn} = Vim->new(service_url => "https://$vcenter/sdk");
 | |
|             $hyphash{$hyp}->{vcenter}->{conn}->login(user_name => $hyphash{$hyp}->{vcenter}->{username}, password => $hyphash{$hyp}->{vcenter}->{password});
 | |
|         };
 | |
|         if ($@) {
 | |
|             $hyphash{$hyp}->{vcenter}->{conn} = undef;
 | |
|         }
 | |
|     }
 | |
|     unless ($hyphash{$hyp}->{vcenter}->{conn}) {
 | |
|         xCAT::SvrUtils::sendmsg([ 1, ": Unable to reach vCenter server managing $hyp" ], $output_handler);
 | |
|         return undef;
 | |
|     }
 | |
| 
 | |
| 
 | |
|     my $foundhyp;
 | |
|     my $name = $hyp;
 | |
|     if ($usehostnamesforvcenter and $usehostnamesforvcenter !~ /no/i) {
 | |
|         if ($tablecfg{hosts}->{$hyp}->[0]->{hostnames}) {
 | |
|             $name = $tablecfg{hosts}->{$hyp}->[0]->{hostnames};
 | |
|         }
 | |
|     }
 | |
|     my $connspec = HostConnectSpec->new(
 | |
|         hostName => $name,
 | |
|         password => $hyphash{$hyp}->{password},
 | |
|         userName => $hyphash{$hyp}->{username},
 | |
|         force    => 1,
 | |
|     );
 | |
|     my $hview;
 | |
|     $hview = $vcenterhash{$vcenter}->{hostviews}->{$hyp};
 | |
|     if ($hview) {
 | |
|         if ($hview->{'summary.config.name'} =~ /^$hyp(?:\.|\z)/ or $hview->{'summary.config.name'} =~ /^$name(?:\.|\z)/) { #Looks good, call the dependent function after declaring the state of vcenter to hypervisor as good
 | |
|             if ($hview->{'summary.runtime.connectionState'}->val eq 'connected') {
 | |
|                 if ($vcenterautojoin) { #admin has requested manual vcenter management, don't mess with vmotion settings
 | |
|                     enable_vmotion(hypname => $hyp, hostview => $hview, conn => $hyphash{$hyp}->{vcenter}->{conn});
 | |
|                 }
 | |
|                 $vcenterhash{$vcenter}->{goodhyps}->{$hyp} = 1;
 | |
|                 $depfun->($depargs);
 | |
|                 if ($hview->parent->type eq 'ClusterComputeResource') { #if it is in a cluster, we can directly remove it
 | |
|                     $hyphash{$hyp}->{deletionref} = $hview->{mo_ref};
 | |
|                 } elsif ($hview->parent->type eq 'ComputeResource') { #For some reason, we must delete the container instead
 | |
|                     $hyphash{$hyp}->{deletionref} = $hview->{parent}; #save off a reference to delete hostview off just in case
 | |
|                 }
 | |
| 
 | |
| 
 | |
|                 return 1;
 | |
|             } elsif ($vcenterautojoin or $vcenterforceremove) { #if allowed autojoin and the current view seems corrupt, throw it away and rejoin
 | |
|                 my $ref_to_delete;
 | |
|                 if ($hview->parent->type eq 'ClusterComputeResource') { #We are allowed to specifically kill a host in a cluster
 | |
|                     $ref_to_delete = $hview->{mo_ref};
 | |
|                 } elsif ($hview->parent->type eq 'ComputeResource') { #For some reason, we must delete the container instead
 | |
|                     $ref_to_delete = $hview->{parent};
 | |
|                 }
 | |
|                 my $task = $hyphash{$hyp}->{vcenter}->{conn}->get_view(mo_ref => $ref_to_delete)->Destroy_Task();
 | |
|                 $running_tasks{$task}->{task} = $task;
 | |
|                 if ($vcenterautojoin) {
 | |
|                     $running_tasks{$task}->{callback} = \&addhosttovcenter;
 | |
|                 } elsif ($vcenterforceremove) {
 | |
|                     $running_tasks{$task}->{callback} = \&delhost_callback;
 | |
|                 }
 | |
|                 $running_tasks{$task}->{conn} = $hyphash{$hyp}->{vcenter}->{conn};
 | |
|                 $running_tasks{$task}->{data} = { depfun => $depfun, depargs => $depargs, conn => $hyphash{$hyp}->{vcenter}->{conn}, connspec => $connspec, hostview => $hview, hypname => $hyp, vcenter => $vcenter };
 | |
|                 return undef;
 | |
| 
 | |
|                 #The rest would be shorter/ideal, but seems to be confused a lot by stateless
 | |
|                 #Maybe in a future VMWare technology level the following would work better
 | |
|                 #than it does today
 | |
|                 #               my $task = $hview_->ReconnectHost_Task(cnxSpec=>$connspec);
 | |
|                 #               my $task = $hview->DisconnectHost_Task();
 | |
|                 #               $running_tasks{$task}->{task} = $task;
 | |
|                 #               $running_tasks{$task}->{callback} = \&disconnecthost_callback;
 | |
|                 #               $running_tasks{$task}->{conn} = $hyphash{$hyp}->{vcenter}->{conn};
 | |
|                 #               $running_tasks{$task}->{data} = { depfun => $depfun, depargs => $depargs, conn=>  $hyphash{$hyp}->{vcenter}->{conn}, connspec=>$connspec,hostview=>$hview,hypname=>$hyp,vcenter=>$vcenter };
 | |
|                 #ADDHOST
 | |
|             } else {
 | |
|                 if ($hyphash{$hyp}->{offline}) {
 | |
|                     xCAT::SvrUtils::sendmsg(": Failed to communicate with $hyp, vCenter reports it as in inventory but not connected and xCAT is set to not autojoin", $output_handler);
 | |
|                 } else {
 | |
|                     xCAT::SvrUtils::sendmsg([ 1, ": Failed to communicate with $hyp, vCenter reports it as in inventory but not connected and xCAT is set to not autojoin" ], $output_handler);
 | |
|                 }
 | |
|                 $hyphash{$hyp}->{conn} = undef;
 | |
|                 return "failed";
 | |
|             }
 | |
|         }
 | |
|     }
 | |
|     unless ($vcenterautojoin) {
 | |
|         if ($hyphash{$hyp}->{offline}) {
 | |
|             xCAT::SvrUtils::sendmsg(": Failed to communicate with $hyp, vCenter does not have it in inventory and xCAT is set to not autojoin", $output_handler);
 | |
|         } else {
 | |
|             xCAT::SvrUtils::sendmsg([ 1, ": Failed to communicate with $hyp, vCenter does not have it in inventory and xCAT is set to not autojoin" ], $output_handler);
 | |
|         }
 | |
|         $hyphash{$hyp}->{conn} = undef;
 | |
|         return "failed";
 | |
|     }
 | |
| 
 | |
|     #If still in function, haven't found any likely host entries, make a new one
 | |
|     unless ($hyphash{$hyp}->{offline}) {
 | |
|         eval {
 | |
|             $hyphash{$hyp}->{conn} = Vim->new(service_url => "https://$hyp/sdk"); #Direct connect to install/check licenses
 | |
|             $hyphash{$hyp}->{conn}->login(user_name => $hyphash{$hyp}->{username}, password => $hyphash{$hyp}->{password});
 | |
|         };
 | |
|         if ($@) {
 | |
|             xCAT::SvrUtils::sendmsg([ 1, ": Failed to communicate with $hyp due to $@" ], $output_handler);
 | |
|             $hyphash{$hyp}->{conn} = undef;
 | |
|             return "failed";
 | |
|         }
 | |
|         validate_licenses($hyp);
 | |
|     }
 | |
|     addhosttovcenter(undef, {
 | |
|             depfun   => $depfun,
 | |
|             depargs  => $depargs,
 | |
|             conn     => $hyphash{$hyp}->{vcenter}->{conn},
 | |
|             connspec => $connspec,
 | |
|             hypname  => $hyp,
 | |
|             vcenter  => $vcenter,
 | |
|     });
 | |
| }
 | |
| 
 | |
| sub addhosttovcenter {
 | |
|     my $task     = shift;
 | |
|     my $args     = shift;
 | |
|     my $hyp      = $args->{hypname};
 | |
|     my $depfun   = $args->{depfun};
 | |
|     my $depargs  = $args->{depargs};
 | |
|     my $connspec = $args->{connspec};
 | |
|     my $vcenter  = $args->{vcenter};
 | |
|     if ($task) {
 | |
|         my $state = $task->info->state->val;
 | |
|         if ($state eq 'error') {
 | |
|             die;
 | |
|         }
 | |
|     }
 | |
|     if ($hyphash{ $args->{hypname} }->{offline}) {    #let it stay offline
 | |
|         $hypready{ $args->{hypname} } = 1;            #declare readiness
 | |
|           #enable_vmotion(hypname=>$args->{hypname},hostview=>$args->{hostview},conn=>$args->{conn});
 | |
|         $vcenterhash{ $args->{vcenter} }->{goodhyps}->{ $args->{hypname} } = 1;
 | |
|         if (defined $args->{depfun}) { #If a function is waiting for the host connect to go valid, call it
 | |
|             $args->{depfun}->($args->{depargs});
 | |
|         }
 | |
|         return;
 | |
|     }
 | |
|     if ($tablecfg{hypervisor}->{$hyp}->[0]->{cluster}) {
 | |
|         my $cluster = get_clusterview(clustname => $tablecfg{hypervisor}->{$hyp}->[0]->{cluster}, conn => $hyphash{$hyp}->{vcenter}->{conn});
 | |
|         unless ($cluster) {
 | |
|             xCAT::SvrUtils::sendmsg([ 1, $tablecfg{hypervisor}->{$hyp}->[0]->{cluster} . " is not a known cluster to the vCenter server." ], $output_handler);
 | |
|             $hypready{$hyp} = -1;      #Declare impossiblility to be ready
 | |
|             return;
 | |
|         }
 | |
|         $task = $cluster->AddHost_Task(spec => $connspec, asConnected => 1);
 | |
|         $running_tasks{$task}->{task}     = $task;
 | |
|         $running_tasks{$task}->{callback} = \&connecthost_callback;
 | |
|         $running_tasks{$task}->{conn}     = $hyphash{$hyp}->{vcenter}->{conn};
 | |
|         $running_tasks{$task}->{data} = { depfun => $depfun, depargs => $depargs, conn => $hyphash{$hyp}->{vcenter}->{conn}, connspec => $connspec, cluster => $cluster, hypname => $hyp, vcenter => $vcenter };
 | |
|     } else {
 | |
|         my $datacenter = validate_datacenter_prereqs($hyp);
 | |
|         unless ($datacenter) { return; }
 | |
|         my $hfolder = $datacenter->hostFolder; #$hyphash{$hyp}->{vcenter}->{conn}->find_entity_view(view_type=>'Datacenter',properties=>['hostFolder'])->hostFolder;
 | |
|         $hfolder = $hyphash{$hyp}->{vcenter}->{conn}->get_view(mo_ref => $hfolder);
 | |
|         $task = $hfolder->AddStandaloneHost_Task(spec => $connspec, addConnected => 1);
 | |
|         $running_tasks{$task}->{task}     = $task;
 | |
|         $running_tasks{$task}->{callback} = \&connecthost_callback;
 | |
|         $running_tasks{$task}->{conn}     = $hyphash{$hyp}->{vcenter}->{conn};
 | |
|         $running_tasks{$task}->{data} = { depfun => $depfun, depargs => $depargs, conn => $hyphash{$hyp}->{vcenter}->{conn}, connspec => $connspec, foldview => $hfolder, hypname => $hyp, vcenter => $vcenter };
 | |
|     }
 | |
| 
 | |
|     #print Dumper @{$hyphash{$hyp}->{vcenter}->{conn}->find_entity_views(view_type=>'HostSystem',properties=>['runtime.connectionState'])};
 | |
| }
 | |
| 
 | |
| sub validate_datacenter_prereqs {
 | |
|     my ($hyp) = @_;
 | |
| 
 | |
|     my $datacenter;
 | |
|     if ($tablecfg{hypervisor}->{$hyp}->[0]->{datacenter}) {
 | |
|         $datacenter = $hyphash{$hyp}->{vcenter}->{conn}->find_entity_view(view_type => 'Datacenter', properties => ['hostFolder'], filter => { name => $tablecfg{hypervisor}->{$hyp}->[0]->{datacenter} });
 | |
|         unless ($datacenter) {
 | |
|             xCAT::SvrUtils::sendmsg([ 1, ": Unable to find requested datacenter (hypervisor.datacenter for $hyp is " . $tablecfg{hypervisor}->{$hyp}->[0]->{datacenter} . ")" ], $output_handler);
 | |
|             return;
 | |
|         }
 | |
|     } else {
 | |
|         $datacenter = $hyphash{$hyp}->{vcenter}->{conn}->find_entity_view(view_type => 'Datacenter', properties => ['hostFolder']);
 | |
|     }
 | |
| 
 | |
|     if (!defined $datacenter) {
 | |
|         my $vconn = $hyphash{$hyp}->{vcenter}->{conn};
 | |
|         my $root_folder = $vconn->get_view(mo_ref => $vconn->get_service_content()->rootFolder);
 | |
|         $root_folder->CreateDatacenter(name => 'xcat-datacenter');
 | |
|         $datacenter = $hyphash{$hyp}->{vcenter}->{conn}->find_entity_view(view_type => 'Datacenter', properties => ['hostFolder']);
 | |
|     }
 | |
| 
 | |
|     return $datacenter;
 | |
| }
 | |
| 
 | |
| 
 | |
| 
 | |
| sub get_default_switch_for_hypervisor {
 | |
| 
 | |
|     #This will make sure the default, implicit switch is in order in accordance
 | |
|     #with the configuration.  If nothing specified, it just spits out vSwitch0
 | |
|     #if something specified, make sure it exists
 | |
|     #if it doesn't exist, and the syntax explains how to build it, build it
 | |
|     #return undef if something is specified, doesn't exist, and lacks instruction
 | |
|     my $hyp       = shift;
 | |
|     my $defswitch = 'vSwitch0';
 | |
|     my $switchmembers;
 | |
|     if ($tablecfg{hypervisor}->{$hyp}->[0]->{defaultnet}) {
 | |
|         $defswitch = $tablecfg{hypervisor}->{$hyp}->[0]->{defaultnet};
 | |
|         ($defswitch, $switchmembers) = split /=/, $defswitch, 2;
 | |
|         my $vswitch;
 | |
|         my $hostview = $hyphash{$hyp}->{hostview};
 | |
|         foreach $vswitch (@{ $hostview->config->network->vswitch }) {
 | |
|             if ($vswitch->name eq $defswitch) {
 | |
|                 return $defswitch;
 | |
|             }
 | |
|         }
 | |
| 
 | |
|         #If still here, means we need to build the switch
 | |
|         unless ($switchmembers) { return undef; } #No hope, no idea how to make it
 | |
|         return create_vswitch($hyp, $defswitch, split(/&/, $switchmembers));
 | |
|     } else {
 | |
|         return 'vSwitch0';
 | |
|     }
 | |
| }
 | |
| 
 | |
| sub get_switchname_for_portdesc {
 | |
| 
 | |
|     #Thisk function will examine all current switches to find or create a switch to match the described requirement
 | |
|     my $hyp      = shift;
 | |
|     my $portdesc = shift;
 | |
|     my $description;    #actual name to use for the virtual switch
 | |
|     if ($tablecfg{hypervisor}->{$hyp}->[0]->{netmap}) {
 | |
|         foreach (split /,/, $tablecfg{hypervisor}->{$hyp}->[0]->{netmap}) {
 | |
|             if (/^$portdesc=/) {
 | |
|                 ($description, $portdesc) = split /=/, $_, 2;
 | |
|                 last;
 | |
|             }
 | |
|         }
 | |
|     } else {
 | |
|         $description = 'vsw' . $portdesc;
 | |
|     }
 | |
|     unless ($description) {
 | |
|         xCAT::SvrUtils::sendmsg([ 1, ": Invalid format for hypervisor.netmap detected for $hyp" ], $output_handler);
 | |
|         return undef;
 | |
|     }
 | |
|     my %requiredports;
 | |
|     my %portkeys;
 | |
|     foreach (split /&/, $portdesc) {
 | |
|         $requiredports{$_} = 1;
 | |
|     }
 | |
| 
 | |
|     my $hostview = $hyphash{$hyp}->{hostview};
 | |
|     unless ($hostview) {
 | |
|         $hyphash{$hyp}->{hostview} = get_hostview(hypname => $hyp, conn => $hyphash{$hyp}->{conn}); #,properties=>['config','configManager']);  #clustered can't run here, hyphash conn reference good
 | |
|         $hostview = $hyphash{$hyp}->{hostview};
 | |
|     }
 | |
|     foreach (@{ $hostview->config->network->pnic }) {
 | |
|         if ($requiredports{ $_->device }) {    #We establish lookups both ways
 | |
|             $portkeys{ $_->key } = $_->device;
 | |
|             delete $requiredports{ $_->device };
 | |
|         }
 | |
|     }
 | |
|     if (keys %requiredports) {
 | |
|         xCAT::SvrUtils::sendmsg([ 1, ":Unable to locate the following nics on $hyp: " . join(',', keys %requiredports) ], $output_handler);
 | |
|         return undef;
 | |
|     }
 | |
|     my $foundmatchswitch;
 | |
|     my $cfgmismatch = 0;
 | |
|     my $vswitch;
 | |
|     foreach $vswitch (@{ $hostview->config->network->vswitch }) {
 | |
|         $cfgmismatch = 0;    #new switch, no sign of mismatch
 | |
|         foreach (@{ $vswitch->pnic }) {
 | |
|             if ($portkeys{$_}) {
 | |
|                 $foundmatchswitch = $vswitch->name;
 | |
|                 delete $requiredports{ $portkeys{$_} };
 | |
|                 delete $portkeys{$_};
 | |
|             } else {
 | |
|                 $cfgmismatch = 1; #If this turns out to have anything, it is bad
 | |
|             }
 | |
|         }
 | |
|         if ($foundmatchswitch) { last; }
 | |
|     }
 | |
|     if ($foundmatchswitch) {
 | |
|         if ($cfgmismatch) {
 | |
|             xCAT::SvrUtils::sendmsg([ 1, ": Aggregation mismatch detected, request nic is aggregated with a nic not requested" ], $output_handler);
 | |
|             return undef;
 | |
|         }
 | |
|         unless (keys %portkeys) {
 | |
|             return $foundmatchswitch;
 | |
|         }
 | |
|         die "TODO: add physical nics to aggregation if requested";
 | |
|     } else {
 | |
|         return create_vswitch($hyp, $description, values %portkeys);
 | |
|     }
 | |
|     die "impossible occurance";
 | |
|     return undef;
 | |
| }
 | |
| 
 | |
| sub create_vswitch {
 | |
|     my $hyp         = shift;
 | |
|     my $description = shift;
 | |
|     my @ports       = @_;
 | |
|     my $vswitch     = HostVirtualSwitchBondBridge->new(
 | |
|         nicDevice => \@ports
 | |
|     );
 | |
|     my $vswspec = HostVirtualSwitchSpec->new(
 | |
|         bridge   => $vswitch,
 | |
|         mtu      => 1500,
 | |
|         numPorts => 64
 | |
|     );
 | |
|     my $hostview = $hyphash{$hyp}->{hostview};
 | |
|     my $netman = $hyphash{$hyp}->{conn}->get_view(mo_ref => $hostview->configManager->networkSystem); #can't run in clustered mode, fine path..
 | |
|     $netman->AddVirtualSwitch(
 | |
|         vswitchName => $description,
 | |
|         spec        => $vswspec
 | |
|     );
 | |
|     return $description;
 | |
| }
 | |
| 
 | |
| sub scan_cluster_networks {
 | |
|     my $cluster = shift;
 | |
|     use Data::Dumper;
 | |
|     my $conn = $clusterhash{$cluster}->{conn};
 | |
|     my $cview = get_clusterview(clustname => $cluster, conn => $conn);
 | |
|     if (defined $cview->{network}) {
 | |
|         foreach (@{ $cview->network }) {
 | |
|             my $nvw = $conn->get_view(mo_ref => $_);
 | |
|             if (defined $nvw->name) {
 | |
|                 $clusterhash{$cluster}->{nets}->{ $nvw->name } = $_;
 | |
|             }
 | |
|         }
 | |
|     }
 | |
| }
 | |
| 
 | |
| sub fixup_hostportgroup {
 | |
|     my $vlanspec = shift;
 | |
|     my $hyp      = shift;
 | |
|     my %args     = @_;
 | |
|     my $action   = 'add';
 | |
|     if ($args{action}) { $action = $args{action} }
 | |
|     my $hostview      = $hyphash{$hyp}->{hostview};
 | |
|     my $switchsupport = 0;
 | |
|     eval {
 | |
|         require xCAT::SwitchHandler;
 | |
|         $switchsupport = 1;
 | |
|     };
 | |
|     my $hypconn = $hyphash{$hyp}->{conn}; #this function can't work in clustered mode anyway, so this is appropriote.
 | |
|     my $vldata  = $vlanspec;
 | |
|     my $switchname = get_default_switch_for_hypervisor($hyp);
 | |
|     my $pgname;
 | |
|     $vldata =~ s/=.*//;    #TODO specify nic model with <blah>=model
 | |
|     if ($vldata =~ /:/) {  #The config specifies a particular path in some way
 | |
|         $vldata =~ s/(.*)://;
 | |
|         $switchname = get_switchname_for_portdesc($hyp, $1);
 | |
|         $pgname = $switchname . "-" . $vldata;
 | |
|     } else { #Use the default vswitch per table config to connect this through, use the same name we did before to maintain compatibility
 | |
|         $pgname = $vldata;
 | |
|     }
 | |
|     my $netsys;
 | |
|     $hyphash{$hyp}->{pgnames}->{$vlanspec} = $pgname;
 | |
|     my $policy = HostNetworkPolicy->new();
 | |
|     unless ($hyphash{$hyp}->{nets}->{$pgname}) {
 | |
|         my $vlanid;
 | |
|         if (looks_like_number($vldata)) {
 | |
|             $vlanid = $vldata;
 | |
|         } elsif ($vldata =~ /trunk/) {
 | |
|             $vlanid = 4095;
 | |
|         } elsif ($vldata =~ /vl(an)?(\d+)$/) {
 | |
|             $vlanid = $2;
 | |
|         } else {
 | |
|             $vlanid = 0;
 | |
|         }
 | |
|         if ($vlanid > 0 and $vlanid < 4095 and $switchsupport) {
 | |
|             my $switchtab = xCAT::Table->new("switch", -create => 0);
 | |
|             if ($switchtab) {
 | |
|                 my $swent = $switchtab->getNodeAttribs($hyp, [qw/switch port/]);
 | |
|                 if ($swent and $swent->{'switch'}) {
 | |
|                     my $swh = new xCAT::SwitchHandler->new($swent->{'switch'});
 | |
|                     my @vlids = $swh->get_vlan_ids();
 | |
|                     if ($action eq 'add') {
 | |
|                         unless (grep { $_ eq $vlanid } @vlids) {
 | |
|                             $swh->create_vlan($vlanid);
 | |
|                         }
 | |
|                         $swh->add_ports_to_vlan($vlanid, $swent->{'port'});
 | |
|                     } elsif ($action eq 'remove') {
 | |
|                         $swh->remove_ports_from_vlan($vlanid, $swent->{'port'});
 | |
|                     }
 | |
|                 }
 | |
|             }
 | |
|         }
 | |
|         my $hostgroupdef = HostPortGroupSpec->new(
 | |
|             name        => $pgname,
 | |
|             vlanId      => $vlanid,
 | |
|             policy      => $policy,
 | |
|             vswitchName => $switchname
 | |
|         );
 | |
|         unless ($netsys) {
 | |
|             $netsys = $hyphash{$hyp}->{conn}->get_view(mo_ref => $hostview->configManager->networkSystem);
 | |
|         }
 | |
|         if ($action eq 'remove') {
 | |
|             $netsys->RemovePortGroup(pgName => $pgname);
 | |
|             return;
 | |
|         } elsif ($action eq 'add') {
 | |
|             $netsys->AddPortGroup(portgrp => $hostgroupdef);
 | |
|         }
 | |
| 
 | |
|         #$hyphash{$hyp}->{nets}->{$netname}=1;
 | |
|         while ((not defined $hyphash{$hyp}->{nets}->{$pgname}) and sleep 1) { #we will only sleep if we know something will be waiting for
 | |
|             $hostview->update_view_data(); #pull in changes induced by previous activity
 | |
|             if (defined $hostview->{network}) { #We load the new object references
 | |
|                 foreach (@{ $hostview->network }) {
 | |
|                     my $nvw = $hypconn->get_view(mo_ref => $_);
 | |
|                     if (defined $nvw->name) {
 | |
|                         $hyphash{$hyp}->{nets}->{ $nvw->name } = $_;
 | |
|                     }
 | |
|                 }
 | |
|             }
 | |
|         }
 | |
|     }
 | |
| }
 | |
| 
 | |
| sub validate_network_prereqs {
 | |
|     my $nodes = shift;
 | |
|     my $hyp   = shift;
 | |
|     my $hypconn = $hyphash{$hyp}->{conn}; #this function can't work in clustered mode anyway, so this is appropriote.
 | |
|     my $hostview = $hyphash{$hyp}->{hostview};
 | |
|     if ($hostview) {
 | |
|         $hostview->update_view_data(); #pull in changes induced by previous activity
 | |
|     } else {
 | |
|         $hyphash{$hyp}->{hostview} = get_hostview(hypname => $hyp, conn => $hyphash{$hyp}->{conn}); #,properties=>['config','configManager','network']);
 | |
|         $hostview = $hyphash{$hyp}->{hostview};
 | |
|     }
 | |
|     my $node;
 | |
|     my $method;
 | |
|     my $location;
 | |
|     if (defined $hostview->{network}) {
 | |
|         foreach (@{ $hostview->network }) {
 | |
|             my $nvw = $hypconn->get_view(mo_ref => $_);
 | |
|             if (defined $nvw->name) {
 | |
|                 $hyphash{$hyp}->{nets}->{ $nvw->name } = $_;
 | |
|             }
 | |
|         }
 | |
|     }
 | |
|     foreach $node (@$nodes) {
 | |
|         my @networks = split /,/, $tablecfg{vm}->{$node}->[0]->{nics};
 | |
|         foreach (@networks) {
 | |
|             fixup_hostportgroup($_, $hyp);
 | |
|         }
 | |
|     }
 | |
|     return 1;
 | |
| 
 | |
| }
 | |
| 
 | |
| sub refreshclusterdatastoremap {
 | |
|     my $cluster = shift;
 | |
|     my $conn    = $clusterhash{$cluster}->{conn};
 | |
|     my $cview   = get_clusterview(clustname => $cluster, conn => $conn);
 | |
|     if (defined $cview->{datastore}) {
 | |
|         foreach (@{ $cview->datastore }) {
 | |
|             my $dsv = $conn->get_view(mo_ref => $_);
 | |
|             if (defined $dsv->info->{nas}) {
 | |
|                 if ($dsv->info->nas->type eq 'NFS') {
 | |
|                     my $mnthost = $dsv->info->nas->remoteHost;
 | |
| 
 | |
|                     #        my $mnthost = inet_aton($dsv->info->nas->remoteHost);
 | |
|                     #        if ($mnthost) {
 | |
|                     #            $mnthost = inet_ntoa($mnthost);
 | |
|                     #        } else {
 | |
|                     #            $mnthost = $dsv->info->nas->remoteHost;
 | |
|                     #            xCAT::SvrUtils::sendmsg([1,"Unable to resolve VMware specified host '".$dsv->info->nas->remoteHost."' to an address, problems may occur"], $output_handler);
 | |
|                     #        }
 | |
|                     $clusterhash{$cluster}->{datastoremap}->{ "nfs://" . $mnthost . $dsv->info->nas->remotePath } = $dsv->info->name;
 | |
|                     $clusterhash{$cluster}->{datastoreurlmap}->{ $dsv->info->name } = "nfs://" . $mnthost . $dsv->info->nas->remotePath; #save off a suitable URL if needed
 | |
|                     $clusterhash{$cluster}->{datastorerefmap}->{ "nfs://" . $mnthost . $dsv->info->nas->remotePath } = $_;
 | |
|                 }    #TODO: care about SMB
 | |
|             } elsif (defined $dsv->info->{vmfs}) {
 | |
|                 my $name = $dsv->info->vmfs->name;
 | |
|                 $clusterhash{$cluster}->{datastoremap}->{ "vmfs://" . $name } = $dsv->info->name;
 | |
|                 $clusterhash{$cluster}->{datastoreurlmap}->{ $dsv->info->name } = "vmfs://" . $name;
 | |
|                 $clusterhash{$cluster}->{datastorerefmap}->{ "vmfs://" . $name } = $_;
 | |
|             }
 | |
|         }
 | |
|     }
 | |
| 
 | |
|     #that's... about it... not doing any of the fancy mounting and stuff, if you do it cluster style, you are on your own.  It's simply too terrifying to try to fixup
 | |
|     #a whole cluster instead of chasing one host, a whole lot slower.  One would hope vmware would've done this, but they don't
 | |
| }
 | |
| 
 | |
| sub validate_datastore_prereqs {
 | |
|     my $hyp = $_[1];
 | |
|     lockbyname($hyp . ".datastores");
 | |
|     $@ = "";
 | |
|     my $rc;
 | |
|     eval { $rc = validate_datastore_prereqs_inlock(@_); };
 | |
|     unlockbyname($hyp . ".datastores");
 | |
|     if ($@) { die $@; }
 | |
|     return $rc;
 | |
| }
 | |
| 
 | |
| sub validate_datastore_prereqs_inlock {
 | |
|     my $nodes = shift;
 | |
|     my $hyp   = shift;
 | |
|     my $newdatastores = shift; # a hash reference of URLs to afflicted nodes outside of table space
 | |
|     my $hypconn  = $hyphash{$hyp}->{conn};
 | |
|     my $hostview = $hyphash{$hyp}->{hostview};
 | |
|     unless ($hostview) {
 | |
|         $hyphash{$hyp}->{hostview} = get_hostview(hypname => $hyp, conn => $hypconn); #,properties=>['config','configManager']);
 | |
|         $hostview = $hyphash{$hyp}->{hostview};
 | |
|     }
 | |
|     my $node;
 | |
|     my $method;
 | |
|     my $location;
 | |
| 
 | |
|     # get all of the datastores that are currently available on this node.
 | |
|     # and put them into a hash
 | |
|     if (defined $hostview->{datastore}) {    # only iterate if it exists
 | |
|         foreach (@{ $hostview->datastore }) {
 | |
|             my $dsv = $hypconn->get_view(mo_ref => $_);
 | |
|             if (defined $dsv->info->{nas}) {
 | |
|                 if ($dsv->info->nas->type eq 'NFS') {
 | |
|                     my $mnthost = inet_aton($dsv->info->nas->remoteHost);
 | |
|                     if ($mnthost) {
 | |
|                         $mnthost = inet_ntoa($mnthost);
 | |
|                     } else {
 | |
|                         $mnthost = $dsv->info->nas->remoteHost;
 | |
|                         xCAT::SvrUtils::sendmsg([ 1, "Unable to resolve VMware specified host '" . $dsv->info->nas->remoteHost . "' to an address, problems may occur" ], $output_handler);
 | |
|                     }
 | |
|                     $hyphash{$hyp}->{datastoremap}->{ "nfs://" . $mnthost . $dsv->info->nas->remotePath } = $dsv->info->name;
 | |
|                     $hyphash{$hyp}->{datastoreurlmap}->{ $dsv->info->name } = "nfs://" . $mnthost . $dsv->info->nas->remotePath;
 | |
|                     $hyphash{$hyp}->{datastorerefmap}->{ "nfs://" . $mnthost . $dsv->info->nas->remotePath } = $_;
 | |
|                 }    #TODO: care about SMB
 | |
|             } elsif (defined $dsv->info->{vmfs}) {
 | |
|                 my $name = $dsv->info->vmfs->name;
 | |
|                 $hyphash{$hyp}->{datastoremap}->{ "vmfs://" . $name } = $dsv->info->name;
 | |
|                 $hyphash{$hyp}->{datastoreurlmap}->{ $dsv->info->name } = "vmfs://" . $name;
 | |
|                 $hyphash{$hyp}->{datastorerefmap}->{ "vmfs://" . $name } = $_;
 | |
|             }
 | |
|         }
 | |
|     }
 | |
|     my $refresh_names = 0;
 | |
| 
 | |
|     # now go through the nodes and make sure that we have matching datastores.
 | |
|     # E.g.: if its NFS, then mount it (if not mounted)
 | |
|     # E.g.: if its VMFS, then create it if not created already.  Note:  VMFS will persist on
 | |
|     # machine reboots, unless its destroyed by being overwritten.
 | |
|     foreach $node (@$nodes) {
 | |
|         my @storage = split /,/, $tablecfg{vm}->{$node}->[0]->{storage};
 | |
|         if ($tablecfg{vm}->{$node}->[0]->{cfgstore}) {
 | |
|             push @storage, $tablecfg{vm}->{$node}->[0]->{cfgstore};
 | |
|         }
 | |
|         foreach (@storage) { #TODO: merge this with foreach loop below.  Here we could build onto $newdatastores instead, for faster operation at scale
 | |
|             s/=.*//;         #remove device type information from configuration
 | |
|             s/\/$//; #Strip trailing slash if specified, to align to VMware semantics
 | |
|             if (/:\/\//) {
 | |
|                 ($method, $location) = split /:\/\//, $_, 2;
 | |
|                 if ($method =~ /nfs/) {
 | |
| 
 | |
|                     # go through and see if NFS is mounted, if not, then mount it.
 | |
|                     (my $server, my $path) = split /\//, $location, 2;
 | |
|                     $server =~ s/:$//; #remove a : if someone put it in out of nfs mount habit
 | |
|                     my $servern = inet_aton($server);
 | |
|                     unless ($servern) {
 | |
|                         xCAT::SvrUtils::sendmsg([ 1, ": Unable to resolve '$server' to an address, check vm.cfgstore/vm.storage" ], $output_handler);
 | |
|                         return 0;
 | |
|                     }
 | |
|                     $server = inet_ntoa($servern);
 | |
|                     my $uri = "nfs://$server/$path";
 | |
|                     unless ($hyphash{$hyp}->{datastoremap}->{$uri}) { #If not already there, must mount it
 | |
|                         unless ($datastoreautomount) {
 | |
|                             xCAT::SvrUtils::sendmsg([ 1, ": $uri is not currently accessible at the given location and automount is disabled in site table" ], $output_handler, $node);
 | |
|                             return 0;
 | |
|                         }
 | |
|                         $refresh_names = 1;
 | |
|                         ($hyphash{$hyp}->{datastoremap}->{$uri}, $hyphash{$hyp}->{datastorerefmap}->{$uri}) = mount_nfs_datastore($hostview, $location);
 | |
|                         $hyphash{$hyp}->{datastoreurlmap}->{ $hyphash{$hyp}->{datastoremap}->{$uri} } = $uri;
 | |
|                     }
 | |
|                 } elsif ($method =~ /vmfs/) {
 | |
|                     (my $name, undef) = split /\//, $location, 2;
 | |
|                     $name =~ s/:$//; #remove a : if someone put it in for some reason.
 | |
|                     my $uri = "vmfs://$name";
 | |
| 
 | |
|                     # check and see if this vmfs is on the node.
 | |
|                     unless ($hyphash{$hyp}->{datastoremap}->{$uri}) { #If not already there, try creating it.
 | |
|                         unless ($datastoreautomount) {
 | |
|                             xCAT::SvrUtils::sendmsg([ 1, ": $uri is not currently accessible at the given location and automount is disabled in site table" ], $output_handler, $node);
 | |
|                             return 0;
 | |
|                         }
 | |
|                         $refresh_names = 1;
 | |
|                         ($hyphash{$hyp}->{datastoremap}->{$uri}, $hyphash{$hyp}->{datastorerefmap}->{$uri}) = create_vmfs_datastore($hostview, $name, $hyp);
 | |
|                         unless ($hyphash{hyp}->{datastoremap}->{$uri}) { return 0; }
 | |
|                         $hyphash{$hyp}->{datastoreurlmap}->{ $hyphash{$hyp}->{datastoremap}->{$uri} } = $uri;
 | |
|                     }
 | |
|                 } else {
 | |
|                     xCAT::SvrUtils::sendmsg([ 1, ": $method is unsupported at this time (nfs would be)" ], $output_handler, $node);
 | |
|                     return 0;
 | |
|                 }
 | |
|             } else {
 | |
|                 xCAT::SvrUtils::sendmsg([ 1, ": $_ not supported storage specification for ESX plugin,\n\t'nfs://<server>/<path>'\n\t\tor\n\t'vmfs://<vmfs>'\n only currently supported vm.storage supported for ESX at the moment" ], $output_handler, $node);
 | |
|                 return 0;
 | |
|             }    #TODO: raw device mapping, VMFS via iSCSI, VMFS via FC?
 | |
|         }
 | |
|     }
 | |
| 
 | |
|     # newdatastores are for migrations or changing vms.
 | |
|     # TODO: make this work for VMFS.  Right now only NFS.
 | |
|     if (ref $newdatastores) {
 | |
|         foreach (keys %$newdatastores) {
 | |
|             my $origurl = $_;
 | |
|             s/\/$//; #Strip trailing slash if specified, to align to VMware semantics
 | |
|             if (/:\/\//) {
 | |
|                 ($method, $location) = split /:\/\//, $_, 2;
 | |
|                 if ($method =~ /nfs/) {
 | |
|                     (my $server, my $path) = split /\//, $location, 2;
 | |
|                     $server =~ s/:$//; #remove a : if someone put it in out of nfs mount habit
 | |
|                     my $servern = inet_aton($server);
 | |
|                     unless ($servern) {
 | |
|                         xCAT::SvrUtils::sendmsg([ 1, ": Unable to resolve '$server' to an address, check vm.cfgstore/vm.storage" ], $output_handler);
 | |
|                         return 0;
 | |
|                     }
 | |
|                     $server = inet_ntoa($servern);
 | |
|                     my $uri = "nfs://$server/$path";
 | |
|                     unless ($method =~ /nfs/) {
 | |
|                         foreach (@{ $newdatastores->{$_} }) {
 | |
|                             xCAT::SvrUtils::sendmsg([ 1, ": $method is unsupported at this time (nfs would be)" ], $output_handler, $_);
 | |
|                         }
 | |
|                         return 0;
 | |
|                     }
 | |
|                     unless ($hyphash{$hyp}->{datastoremap}->{$uri}) { #If not already there, must mount it
 | |
|                         unless ($datastoreautomount) {
 | |
|                             xCAT::SvrUtils::sendmsg([ 1, ":) $uri is not currently accessible at the given location and automount is disabled in site table" ], $output_handler, $node);
 | |
|                             return 0;
 | |
|                         }
 | |
|                         $refresh_names = 1;
 | |
|                         ($hyphash{$hyp}->{datastoremap}->{$uri}, $hyphash{$hyp}->{datastorerefmap}->{$uri}) = mount_nfs_datastore($hostview, $location);
 | |
|                     }
 | |
|                     $hyphash{$hyp}->{datastoreurlmap}->{ $hyphash{$hyp}->{datastoremap}->{$uri} } = $uri;
 | |
|                     $hyphash{$hyp}->{datastoremap}->{$origurl} = $hyphash{$hyp}->{datastoremap}->{$uri}; #we track both the uri xCAT expected and the one vCenter actually ended up with
 | |
|                     $hyphash{$hyp}->{datastorerefmap}->{$origurl} = $hyphash{$hyp}->{datastorerefmap}->{$uri};
 | |
|                 } elsif ($method =~ /vmfs/) {
 | |
|                     (my $name, undef) = split /\//, $location, 2;
 | |
|                     $name =~ s/:$//; #remove a : if someone put it in for some reason.
 | |
|                     my $uri = "vmfs://$name";
 | |
|                     unless ($hyphash{$hyp}->{datastoremap}->{$uri}) { #If not already there, it should be!
 | |
|                         unless ($datastoreautomount) {
 | |
|                             xCAT::SvrUtils::sendmsg([ 1, ": $uri is not currently accessible at the given location and automount is disabled in site table" ], $output_handler, $node);
 | |
|                             return 0;
 | |
|                         }
 | |
|                         $refresh_names = 1;
 | |
|                         ($hyphash{$hyp}->{datastoremap}->{$uri}, $hyphash{$hyp}->{datastorerefmap}->{$uri}) = create_vmfs_datastore($hostview, $name, $hyp);
 | |
|                         unless ($hyphash{hyp}->{datastoremap}->{$uri}) { return 0; }
 | |
|                     }
 | |
|                     $hyphash{$hyp}->{datastoreurlmap}->{ $hyphash{$hyp}->{datastoremap}->{$uri} } = $uri;
 | |
|                     $hyphash{$hyp}->{datastoremap}->{$origurl} = $hyphash{$hyp}->{datastoremap}->{$uri};
 | |
|                     $hyphash{$hyp}->{datastorerefmap}->{$origurl} = $hyphash{$hyp}->{datastorerefmap}->{$uri};
 | |
|                 } else {
 | |
|                     print "$method: not NFS and not VMFS here!\n";
 | |
|                 }
 | |
|             } else {
 | |
|                 my $datastore = $_;
 | |
|                 foreach my $ds (@{ $newdatastores->{$_} }) {
 | |
|                     xCAT::SvrUtils::sendmsg([ 1, ": $datastore not supported storage specification for ESX plugin, 'nfs://<server>/<path>' only currently supported vm.storage supported for ESX at the moment" ], $output_handler, $ds);
 | |
|                 }
 | |
|                 return 0;
 | |
|             } #TODO: raw device mapping, VMFS via iSCSI, VMFS via FC, VMFS on same local drive?
 | |
|         }
 | |
|     }
 | |
|     if ($refresh_names) { #if we are in a vcenter context, vmware can rename a datastore behind our backs immediately after adding
 | |
|         $hostview->update_view_data();
 | |
|         if (defined $hostview->{datastore}) {    # only iterate if it exists
 | |
|             foreach (@{ $hostview->datastore }) {
 | |
|                 my $dsv = $hypconn->get_view(mo_ref => $_);
 | |
|                 if (defined $dsv->info->{nas}) {
 | |
|                     if ($dsv->info->nas->type eq 'NFS') {
 | |
|                         my $mnthost = inet_aton($dsv->info->nas->remoteHost);
 | |
|                         if ($mnthost) {
 | |
|                             $mnthost = inet_ntoa($mnthost);
 | |
|                         } else {
 | |
|                             $mnthost = $dsv->info->nas->remoteHost;
 | |
|                             xCAT::SvrUtils::sendmsg([ 1, "Unable to resolve VMware specified host '" . $dsv->info->nas->remoteHost . "' to an address, problems may occur" ], $output_handler);
 | |
|                         }
 | |
|                         $hyphash{$hyp}->{datastoremap}->{ "nfs://" . $mnthost . $dsv->info->nas->remotePath } = $dsv->info->name;
 | |
|                         $hyphash{$hyp}->{datastoreurlmap}->{ $dsv->info->name } = "nfs://" . $mnthost . $dsv->info->nas->remotePath;
 | |
|                         $hyphash{$hyp}->{datastorerefmap}->{ "nfs://" . $mnthost . $dsv->info->nas->remotePath } = $_;
 | |
|                     }    #TODO: care about SMB
 | |
|                 }    #TODO: care about VMFS
 | |
|             }
 | |
|         }
 | |
|     }
 | |
|     return 1;
 | |
| }
 | |
| 
 | |
| sub getlabel_for_datastore {
 | |
|     my $method   = shift;
 | |
|     my $location = shift;
 | |
| 
 | |
|     $location =~ s/\//_/g;
 | |
|     $location = $method . '_' . $location;
 | |
| 
 | |
|     #VMware has a 42 character limit, we will start mangling to get under 42.
 | |
|     #Will try to preserve as much informative detail as possible, hence several conditionals instead of taking the easy way out
 | |
|     if (length($location) > 42) {
 | |
|         $location =~ s/nfs_//; #Ditch unique names for different protocols to the same path, seems unbelievably unlikely
 | |
|     }
 | |
|     if (length($location) > 42) {
 | |
|         $location =~ s/\.//g; #Next, ditch host delimiter, it is unlikely that hosts will have unique names if their dots are removed
 | |
|     }
 | |
|     if (length($location) > 42) {
 | |
|         $location =~ s/_//g; #Next, ditch path delimiter, it is unlikely that two paths will happen to look the same without delimiters
 | |
|     }
 | |
|     if (length($location) > 42) {    #finally, replace the middle with ellipsis
 | |
|         substr($location, 20, -20, '..');
 | |
|     }
 | |
|     return $location;
 | |
| }
 | |
| 
 | |
| sub mount_nfs_datastore {
 | |
|     my $hostview = shift;
 | |
|     my $location = shift;
 | |
|     my $server;
 | |
|     my $path;
 | |
|     unless ($datastoreautomount) {
 | |
|         die "automount of VMware datastores is disabled in site configuration, not continuing";
 | |
|     }
 | |
|     ($server, $path) = split /\//, $location, 2;
 | |
|     $location = getlabel_for_datastore('nfs', $location);
 | |
| 
 | |
|     my $nds = HostNasVolumeSpec->new(accessMode => 'readWrite',
 | |
|         remoteHost => $server,
 | |
|         localPath  => $location,
 | |
|         remotePath => "/" . $path);
 | |
|     my $dsmv = $hostview->{vim}->get_view(mo_ref => $hostview->configManager->datastoreSystem);
 | |
| 
 | |
|     my $dsref;
 | |
|     eval {
 | |
|         $dsref = $dsmv->CreateNasDatastore(spec => $nds);
 | |
|     };
 | |
| 
 | |
|     if ($@) {
 | |
|         die "$@" unless $@ =~ m/Fault detail: DuplicateNameFault/;
 | |
| 
 | |
|         die "esx plugin: a datastore was discovered with the same name referring to a different nominatum- cannot continue\n$@"
 | |
|           unless &match_nfs_datastore($server, "/$path", $hostview->{vim});
 | |
|     }
 | |
| 
 | |
|     return ($location, $dsref);
 | |
| }
 | |
| 
 | |
| # create a VMFS data store on a node so that VMs can live locally instead of NFS
 | |
| sub create_vmfs_datastore {
 | |
|     my $hostview = shift;    # VM object
 | |
|     my $name     = shift;    # name of storage we wish to create.
 | |
|     my $hyp      = shift;
 | |
|     unless ($datastoreautomount) {
 | |
|         die "automount of VMware datastores is disabled in site configuration, not continuing";
 | |
|     }
 | |
| 
 | |
|     # call some VMware API here to create
 | |
|     my $hdss = $hostview->{vim}->get_view(mo_ref => $hostview->configManager->datastoreSystem);
 | |
| 
 | |
|     my $diskList = $hdss->QueryAvailableDisksForVmfs();
 | |
|     my $count = scalar(@$diskList); # get the number of disks available for formatting.
 | |
|     unless ($count > 0) {
 | |
| 
 | |
|         #die "No disks are available to create VMFS volume for $name";
 | |
|         $output_handler->({ error => ["No disks are available on $hyp to create VMFS volume for $name"], errorcode => 1 });
 | |
|         return 0;
 | |
|     }
 | |
|     foreach my $disk (@$diskList) {
 | |
|         my $options = $hdss->QueryVmfsDatastoreCreateOptions(devicePath => $disk->devicePath);
 | |
|         @$options[0]->spec->vmfs->volumeName($name);
 | |
|         my $newDatastore = $hdss->CreateVmfsDatastore(spec => @$options[0]->spec);
 | |
| 
 | |
|         #return $newDatastore;
 | |
|         # create it on the first disk we see.
 | |
|         return ($name, $newDatastore);
 | |
|     }
 | |
|     return 0;
 | |
| }
 | |
| 
 | |
| 
 | |
| 
 | |
| sub build_more_info {
 | |
|     die("TODO: fix this function if called");
 | |
|     print "Does this acually get called????**********************************\n";
 | |
|     my $noderange = shift;
 | |
|     my $callback  = shift;
 | |
|     my $vmtab     = xCAT::Table->new("vm");
 | |
|     my @moreinfo  = ();
 | |
|     unless ($vmtab) {
 | |
|         $callback->({ data => ["Cannot open mp table"] });
 | |
|         return @moreinfo;
 | |
|     }
 | |
|     my %mpa_hash = ();
 | |
|     foreach my $node (@$noderange) {
 | |
|         my $ent = $vmtab->getNodeAttribs($node, [ 'mpa', 'id' ]);
 | |
|         if (defined($ent->{mpa})) { push @{ $mpa_hash{ $ent->{mpa} }{nodes} }, $node; }
 | |
|         else {
 | |
|             $callback->({ data => ["no mpa defined for node $node"] });
 | |
|             return @moreinfo;
 | |
|         }
 | |
|         if (defined($ent->{id})) { push @{ $mpa_hash{ $ent->{mpa} }{ids} }, $ent->{id}; }
 | |
|         else { push @{ $mpa_hash{ $ent->{mpa} }{ids} }, ""; }
 | |
|     }
 | |
| 
 | |
|     foreach (keys %mpa_hash) {
 | |
|         push @moreinfo, "\[$_\]\[" . join(',', @{ $mpa_hash{$_}{nodes} }) . "\]\[" . join(',', @{ $mpa_hash{$_}{ids} }) . "\]";
 | |
| 
 | |
|     }
 | |
| 
 | |
|     return \@moreinfo;
 | |
| }
 | |
| 
 | |
| sub copycd {
 | |
|     my $request  = shift;
 | |
|     my $doreq    = shift;
 | |
|     my $distname = "";
 | |
|     my $path;
 | |
|     my $arch;
 | |
|     my $darch;
 | |
|     my $installroot;
 | |
|     $installroot = "/install";
 | |
| 
 | |
|     #my $sitetab = xCAT::Table->new('site');
 | |
|     #if($sitetab){
 | |
|     #(my $ref) = $sitetab->getAttribs({key => 'installdir'}, 'value');
 | |
|     my @entries = xCAT::TableUtils->get_site_attribute("installdir");
 | |
|     my $t_entry = $entries[0];
 | |
|     if (defined($t_entry)) {
 | |
|         $installroot = $t_entry;
 | |
|     }
 | |
| 
 | |
|     #}
 | |
|     @ARGV = @{ $request->{arg} };
 | |
|     my $includeupdate = 0;
 | |
|     GetOptions(
 | |
|         'n=s' => \$distname,
 | |
|         'a=s' => \$arch,
 | |
|         'm=s' => \$path,
 | |
|         's'   => \$includeupdate
 | |
|     );
 | |
| 
 | |
|     # run a few tests to see if the copycds should use this plugin
 | |
|     unless ($path) {
 | |
| 
 | |
|         # can't use us cause we need a path and you didn't provide one!
 | |
|         return;
 | |
|     }
 | |
|     if ($distname and $distname !~ /^esx/) {
 | |
| 
 | |
|         # we're for esx, so if you didn't specify that its not us!
 | |
|         return;
 | |
|     }
 | |
|     my $found = 0;
 | |
| 
 | |
|     if (-r $path . "/README" and -r $path . "/build_number" and -d $path . "/VMware" and -r $path . "/packages.xml") { #We have a probable new style ESX media
 | |
|         open(LINE, $path . "/packages.xml");
 | |
|         my $product;
 | |
|         my $version;
 | |
|         while (<LINE>) {
 | |
|             if (/roductLineId>([^<]*)<\/Prod/) {
 | |
|                 $product = $1;
 | |
|             }
 | |
|             if (/ersion>([^<]*)<\/version/) {
 | |
|                 $version = $1;
 | |
|                 $version =~ s/\.0$//;
 | |
|             }
 | |
|             if (/arch>([^>]*)<\/arch/) {
 | |
|                 unless ($darch and $darch =~ /x86_64/) { #prefer to be characterized as x86_64
 | |
|                     $darch = $1;
 | |
|                     $arch  = $1;
 | |
|                 }
 | |
| 
 | |
|             }
 | |
|         }
 | |
|         close(LINE);
 | |
|         if ($product and $version) {
 | |
|             unless ($distname) { $distname = $product . $version; }
 | |
|             $found = 1;
 | |
|         }
 | |
|     } elsif (-r $path . "/README" and -r $path . "/open_source_licenses.txt" and -d $path . "/VMware") { #Candidate to be ESX 3.5
 | |
|         open(LINE, $path . "/README");
 | |
|         while (<LINE>) {
 | |
|             if (/VMware ESX Server 3.5\s*$/) {
 | |
|                 $darch = 'x86';
 | |
|                 $arch  = 'x86';
 | |
|                 unless ($distname) { $distname = 'esx3.5'; }
 | |
|                 $found = 1;
 | |
|                 last;
 | |
|             }
 | |
|         }
 | |
|         close(LINE);
 | |
|     } elsif (-r $path . "/README.txt" and -r $path . "/vmkernel.gz") {
 | |
| 
 | |
|         # its an esxi dvd!
 | |
|         # if we got here its probably ESX they want to copy
 | |
|         my $line;
 | |
|         my $darch;
 | |
|         open(LINE, $path . "/README.txt") or die "couldn't open!";
 | |
|         while ($line = <LINE>) {
 | |
|             chomp($line);
 | |
|             if ($line =~ /VMware ESXi(?: version)? 4\.(\d+)/) {
 | |
|                 $darch = "x86_64";
 | |
|                 unless ($distname) {
 | |
|                     $distname = "esxi4";
 | |
|                     if ($1) {
 | |
|                         $distname .= '.' . $1;
 | |
|                     }
 | |
|                 }
 | |
|                 $found = 1;
 | |
|                 if ($arch and $arch ne $darch) {
 | |
|                     xCAT::SvrUtils::sendmsg([ 1, "Requested distribution architecture $arch, but media is $darch" ], $output_handler);
 | |
|                     return;
 | |
|                 }
 | |
|                 $arch = $darch;
 | |
|                 last;    # we found our distro!  end this loop madness.
 | |
|             }
 | |
|         }
 | |
|         close(LINE);
 | |
|         unless ($found) {
 | |
|             xCAT::SvrUtils::sendmsg([ 1, "I don't recognize this VMware ESX DVD" ], $output_handler);
 | |
|             return;      # doesn't seem to be a valid DVD or CD
 | |
|         }
 | |
|     } elsif (-r $path . "/vmkernel.gz" and -r $path . "/isolinux.cfg") {
 | |
|         open(LINE, $path . "/isolinux.cfg");
 | |
|         while (<LINE>) {
 | |
|             if (/ThinESX Installer/) {
 | |
|                 $darch = 'x86';
 | |
|                 $arch  = 'x86';
 | |
|                 unless ($distname) { $distname = 'esxi3.5'; }
 | |
|                 $found = 1;
 | |
|                 last;
 | |
|             }
 | |
|         }
 | |
|         close(LINE);
 | |
|     } elsif (-r $path . "/upgrade/metadata.xml") {
 | |
|         open(LINE, $path . "/upgrade/metadata.xml");
 | |
|         my $detectdistname;
 | |
|         while (<LINE>) {
 | |
|             if (/esxVersion>([^<]*)</) {
 | |
|                 my $version = $1;
 | |
|                 while ($version =~ /\.0$/) {
 | |
|                     $version =~ s/\.0$//;
 | |
|                 }
 | |
|                 $darch          = "x86_64";
 | |
|                 $arch           = "x86_64";
 | |
|                 $detectdistname = 'esxi' . $version;
 | |
|                 $found          = 1;
 | |
|             } elsif (/esxRelease>([^<]*)</) {
 | |
|                 unless ($includeupdate) {
 | |
|                     next;
 | |
|                 }
 | |
|                 my $release = $1;
 | |
|                 while ($release =~ /\.0$/) {
 | |
|                     $release =~ s/\.0$//;
 | |
|                 }
 | |
|                 unless ($release ne "0") {
 | |
|                     next;
 | |
|                 }
 | |
|                 $detectdistname .= '_' . $release;
 | |
|             }
 | |
|         }
 | |
|         unless ($distname) { $distname = $detectdistname; }
 | |
|     } elsif (-r $path . "/vmware-esx-base-readme") {
 | |
|         open(LINE, $path . "/vmware-esx-base-readme");
 | |
|         while (<LINE>) {
 | |
|             if (/VMware ESXi 5\.0/) {
 | |
|                 $darch = "x86_64";
 | |
|                 $arch  = "x86_64";
 | |
|                 unless ($distname) { $distname = 'esxi5'; }
 | |
|                 $found = 1;
 | |
|                 last;
 | |
|             }
 | |
|             if (/VMware ESXi 5\.1/) {
 | |
|                 $darch = "x86_64";
 | |
|                 $arch  = "x86_64";
 | |
|                 unless ($distname) { $distname = 'esxi5.1'; }
 | |
|                 $found = 1;
 | |
|                 last;
 | |
|             }
 | |
|             if (/VMware ESXi 5\.5/) {
 | |
|                 $darch = "x86_64";
 | |
|                 $arch  = "x86_64";
 | |
|                 unless ($distname) { $distname = 'esxi5.5'; }
 | |
|                 $found = 1;
 | |
|                 last;
 | |
|             }
 | |
|         }
 | |
|     }
 | |
| 
 | |
|     unless ($found) { return; }               #not our media
 | |
|     if ($::XCATSITEVALS{osimagerequired}) {
 | |
|         my ($nohaveimages, $errstr) = xCAT::SvrUtils->update_tables_with_templates($distname, $arch, "", "", checkonly => 1);
 | |
|         if ($nohaveimages) {
 | |
|             $output_handler->({ error => "No Templates found to support $distname($arch)", errorcode => 2 });
 | |
|         }
 | |
|     }
 | |
| 
 | |
|     xCAT::SvrUtils::sendmsg("Copying media to $installroot/$distname/$arch/", $output_handler);
 | |
|     my $omask = umask 0022;
 | |
|     mkpath("$installroot/$distname/$arch");
 | |
|     umask $omask;
 | |
|     my $rc;
 | |
|     my $reaped = 0;
 | |
|     $SIG{INT} = $SIG{TERM} = sub {
 | |
|         foreach (@cpiopid) {
 | |
|             kill 2, $_;
 | |
|         }
 | |
|         if ($path) {
 | |
|             chdir("/");
 | |
|             system("umount $path");
 | |
|         }
 | |
|     };
 | |
|     my $KID;
 | |
|     chdir $path;
 | |
|     my $numFiles = `find . -print | wc -l`;
 | |
|     my $child = open($KID, "|-");
 | |
|     unless (defined $child)
 | |
|     {
 | |
|         xCAT::SvrUtils::sendmsg([ 1, "Media copy operation fork failure" ], $output_handler);
 | |
|         return;
 | |
|     }
 | |
|     if ($child)
 | |
|     {
 | |
|         push @cpiopid, $child;
 | |
|         my @finddata = `find .`;
 | |
|         for (@finddata)
 | |
|         {
 | |
|             print $KID $_;
 | |
|         }
 | |
|         close($KID);
 | |
|         $rc = $?;
 | |
|     }
 | |
|     else
 | |
|     {
 | |
|         nice 10;
 | |
|         my $c = "nice -n 20 cpio -vdump $installroot/$distname/$arch";
 | |
|         my $k2 = open(PIPE, "$c 2>&1 |") ||
 | |
|           xCAT::SvrUtils::sendmsg([ 1, "Media copy operation fork failure" ], $output_handler);
 | |
|         push @cpiopid, $k2;
 | |
|         my $copied = 0;
 | |
|         my ($percent, $fout);
 | |
|         while (<PIPE>) {
 | |
|             next if /^cpio:/;
 | |
|             $percent = $copied / $numFiles;
 | |
|             $fout = sprintf "%0.2f%%", $percent * 100;
 | |
|             $output_handler->({ sinfo => "$fout" });
 | |
|             ++$copied;
 | |
|         }
 | |
|         exit;
 | |
|     }
 | |
| 
 | |
|     # let everyone read it
 | |
|     #chdir "/tmp";
 | |
|     chmod 0755, "$installroot/$distname/$arch";
 | |
|     if ($distname =~ /esxi[56]/) { #going to tweak boot.cfg for install and default stateless case
 | |
|         if (!-r "$installroot/$distname/$arch/boot.cfg.stateless") {
 | |
|             copy("$installroot/$distname/$arch/boot.cfg", "$installroot/$distname/$arch/boot.cfg.stateless");
 | |
|             my $bootcfg;
 | |
|             open($bootcfg, "<", "$installroot/$distname/$arch/boot.cfg");
 | |
|             my @bootcfg = <$bootcfg>;
 | |
|             close($bootcfg);
 | |
|             foreach (@bootcfg) { #no point in optimizing trivial, infrequent code, readable this way
 | |
|                 s!kernel=/!kernel=!;      # remove leading /
 | |
|                 s!modules=/!modules=!;    #remove leading /
 | |
|                 s!--- /!--- !g;           #remove all the 'absolute' slashes
 | |
|             }
 | |
|             open($bootcfg, ">", "$installroot/$distname/$arch/boot.cfg.install");
 | |
|             foreach (@bootcfg) {
 | |
|                 if (/^modules=/ and $_ !~ /xcatmod.tgz/ and not $::XCATSITEVALS{xcatesximoddisable}) {
 | |
|                     chomp();
 | |
|                     s! *\z! --- xcatmod.tgz\n!;
 | |
|                 }
 | |
|                 print $bootcfg $_;
 | |
|             }
 | |
|             close($bootcfg);
 | |
|             foreach (@bootcfg) { #no point in optimizing trivial, infrequent code, readable this way
 | |
|                 s/runweasel//;        #don't run the installer in stateless mode
 | |
|                 s!--- imgdb.tgz!!;    #don't need the imgdb for stateless
 | |
|                 s!--- imgpayld.tgz!!; #don't need the boot payload since we aren't installing
 | |
|                 s!--- tools.t00!!; #tools could be useful, but for now skip the memory requirement
 | |
|                 s!--- weaselin.i00!!; #and also don't need the weasel install images if... not installing
 | |
| 
 | |
|                 if (/^modules=/ and $_ !~ /xcatmod.tgz/ and not $::XCATSITEVALS{xcatesximoddisable}) {
 | |
|                     chomp();
 | |
|                     s! *\z! --- xcatmod.tgz\n!;
 | |
|                 }
 | |
|                 s!Loading ESXi installer!xCAT is loading ESXi stateless!;
 | |
|             }
 | |
|             open($bootcfg, ">", "$installroot/$distname/$arch/boot.cfg.stateless");
 | |
|             foreach (@bootcfg) {
 | |
|                 print $bootcfg $_;
 | |
|             }
 | |
|             close($bootcfg);
 | |
|             if (grep /LSIProvi.v00/, @bootcfg and !-r "$installroot/$distname/$arch/LSIProvi.v00" and -r "$installroot/$distname/$arch/lsiprovi.v00") { #there is media with LSIProv.v00 expected, but the install media was mal-constructed, fix it
 | |
|                 move("$installroot/$distname/$arch/lsiprovi.v00", "$installroot/$distname/$arch/LSIProvi.v00");
 | |
|             }
 | |
|         }
 | |
|     }
 | |
| 
 | |
|     if ($rc != 0) {
 | |
|         xCAT::SvrUtils::sendmsg([ 1, "Media copy operation failed, status $rc" ], $output_handler);
 | |
|     } else {
 | |
|         xCAT::SvrUtils::sendmsg("Media copy operation successful", $output_handler);
 | |
|         my @ret = xCAT::SvrUtils->update_tables_with_templates($distname, $arch);
 | |
|         if ($ret[0] != 0) {
 | |
|             xCAT::SvrUtils::sendmsg("Error when updating the osimage tables: " . $ret[1], $output_handler);
 | |
|         }
 | |
| 
 | |
| 
 | |
|     }
 | |
| }
 | |
| 
 | |
| sub makecustomizedmod {
 | |
|     my $osver = shift;
 | |
|     my $dest  = shift;
 | |
|     if ($::XCATSITEVALS{xcatesximoddisable}) { return 1; }
 | |
|     my $modname;
 | |
|     if ($osver =~ /esxi4/) { #want more descriptive name,but don't break esxi4 setups.
 | |
|         $modname = "mod.tgz";
 | |
| 
 | |
|         # if it already exists, do not overwrite it because it may be someone
 | |
|         # else's custom image
 | |
|         if (-f "$dest/$modname") { return 1; }
 | |
|     } else {
 | |
|         $modname = "xcatmod.tgz";
 | |
|     }
 | |
|     my $passtab = xCAT::Table->new('passwd');
 | |
|     my $tmp;
 | |
|     my $password;
 | |
|     if ($passtab) {
 | |
|         ($tmp) = $passtab->getAttribs({ 'key' => 'vmware' }, 'username', 'password');
 | |
|         if (defined($tmp)) {
 | |
|             $password = $tmp->{password};
 | |
|         }
 | |
|     }
 | |
|     unless ($password) {
 | |
|         xCAT::SvrUtils::sendmsg([ 1, ": Unable to find a password entry for esxi in passwd table" ], $output_handler);
 | |
|         return 0;
 | |
|     }
 | |
|     mkpath("/tmp/xcat");
 | |
|     my $tempdir = tempdir("/tmp/xcat/esxmodcustXXXXXXXX");
 | |
|     my $shadow;
 | |
|     mkpath($tempdir . "/etc/");
 | |
|     my $oldmask = umask(0077);
 | |
|     open($shadow, ">", $tempdir . "/etc/shadow");
 | |
|     $password = crypt($password, '$1$' . xCAT::Utils::genpassword(8));
 | |
|     my $dayssince1970 = int(time() / 86400);    #Be truthful about /etc/shadow
 | |
|     my @otherusers = qw/nobody nfsnobody dcui daemon/;
 | |
| 
 | |
|     if ($osver =~ /esxi4/) {
 | |
|         push @otherusers, "vimuser";
 | |
|     } elsif ($osver =~ /esxi[56]/) {
 | |
|         push @otherusers, "vpxuser";
 | |
|     }
 | |
|     print $shadow "root:$password:$dayssince1970:0:99999:7:::\n";
 | |
|     foreach (@otherusers) {
 | |
|         print $shadow "$_:*:$dayssince1970:0:99999:7:::\n";
 | |
|     }
 | |
|     close($shadow);
 | |
|     umask($oldmask);
 | |
|     if ($osver =~ /esxi4/ and -e "$::XCATROOT/share/xcat/netboot/esxi/38.xcat-enableipv6") {
 | |
|         mkpath($tempdir . "/etc/vmware/init/init.d");
 | |
|         copy("$::XCATROOT/share/xcat/netboot/esxi/38.xcat-enableipv6", $tempdir . "/etc/vmware/init/init.d/38.xcat-enableipv6");
 | |
|     } elsif ($osver =~ /esxi[56]/ and -e "$::XCATROOT/share/xcat/netboot/esxi/xcat-ipv6.json") {
 | |
|         mkpath($tempdir . "/usr/libexec/jumpstart/plugins/");
 | |
|         copy("$::XCATROOT/share/xcat/netboot/esxi/xcat-ipv6.json", $tempdir . "/usr/libexec/jumpstart/plugins/xcat-ipv6.json");
 | |
|     }
 | |
|     if ($osver =~ /esxi4/ and -e "$::XCATROOT/share/xcat/netboot/esxi/47.xcat-networking") {
 | |
|         copy("$::XCATROOT/share/xcat/netboot/esxi/47.xcat-networking", $tempdir . "/etc/vmware/init/init.d/47.xcat-networking");
 | |
|     } elsif ($osver =~ /esxi[56]/ and -e "$::XCATROOT/share/xcat/netboot/esxi/39.ipv6fixup") {
 | |
|         mkpath($tempdir . "/etc/init.d");
 | |
|         copy("$::XCATROOT/share/xcat/netboot/esxi/39.ipv6fixup", $tempdir . "/etc/init.d/39.ipv6fixup");
 | |
|         chmod(0755, "$tempdir/etc/init.d/39.ipv6fixup");
 | |
|     }
 | |
|     if ($osver =~ /esxi[56]/ and -e "$::XCATROOT/share/xcat/netboot/esxi/48.esxifixup") {
 | |
|         mkpath($tempdir . "/etc/init.d");
 | |
|         copy("$::XCATROOT/share/xcat/netboot/esxi/48.esxifixup", $tempdir . "/etc/init.d/48.esxifixup");
 | |
|         chmod(0755, "$tempdir/etc/init.d/48.esxifixup");
 | |
|     }
 | |
|     if ($osver =~ /esxi5/ and -e "$::XCATROOT/share/xcat/netboot/esxi/99.esxiready") {
 | |
|         mkpath($tempdir . "/etc/init.d");
 | |
|         copy("$::XCATROOT/share/xcat/netboot/esxi/99.esxiready", $tempdir . "/etc/init.d/99.esxiready");
 | |
|         chmod(0755, "$tempdir/etc/init.d/99.esxiready");
 | |
|     }
 | |
|     if (-e "$::XCATROOT/share/xcat/netboot/esxi/xcatsplash") {
 | |
|         mkpath($tempdir . "/etc/vmware/");
 | |
|         copy("$::XCATROOT/share/xcat/netboot/esxi/xcatsplash", $tempdir . "/etc/vmware/welcome");
 | |
|     }
 | |
|     my $dossh = 0;
 | |
|     if (-r "/root/.ssh/id_rsa.pub") {
 | |
|         $dossh = 1;
 | |
|         my $umask = umask(0077); #don't remember if dropbear is picky, but just in case
 | |
|         if ($osver =~ /esxi4/) { #esxi4 used more typical path
 | |
|             mkpath($tempdir . "/.ssh");
 | |
|             copy("/root/.ssh/id_rsa.pub", $tempdir . "/.ssh/authorized_keys");
 | |
|         } elsif ($osver =~ /esxi5/) {    #weird path to keys
 | |
|             mkpath($tempdir . "/etc/ssh/keys-root");
 | |
|             copy("/root/.ssh/id_rsa.pub", $tempdir . "/etc/ssh/keys-root/authorized_keys");
 | |
|         }
 | |
|         umask($umask);
 | |
|     }
 | |
|     my $tfile;
 | |
|     mkpath($tempdir . "/var/run/vmware");
 | |
|     open $tfile, ">", $tempdir . "/var/run/vmware/show-tech-support-login";
 | |
|     close($tfile);
 | |
| 
 | |
|     #TODO: auto-enable ssh and request boot-time customization rather than on-demand?
 | |
|     require Cwd;
 | |
|     my $dir = Cwd::cwd();
 | |
|     chdir($tempdir);
 | |
|     if (-e "$dest/$modname") {
 | |
|         unlink("$dest/$modname");
 | |
|     }
 | |
|     if ($dossh and $osver =~ /esxi4/) {
 | |
|         system("tar czf $dest/$modname * .ssh");
 | |
|     } else {
 | |
|         system("tar czf $dest/$modname *");
 | |
|     }
 | |
|     chdir($dir);
 | |
|     rmtree($tempdir);
 | |
|     return 1;
 | |
| }
 | |
| 
 | |
| sub getplatform {
 | |
|     my $os = shift;
 | |
|     if ($os =~ /esxi/) {
 | |
|         return "esxi";
 | |
|     }
 | |
|     return $os;
 | |
| }
 | |
| 
 | |
| sub esxi_kickstart_from_template {
 | |
|     my %args       = @_;
 | |
|     my $installdir = "/install";
 | |
|     if ($::XCATSITEVALS{installdir}) { $installdir = $::XCATSITEVALS{installdir}; }
 | |
|     my $plat = getplatform($args{os});
 | |
|     my $template = xCAT::SvrUtils::get_tmpl_file_name("$installdir/custom/install/$plat", $args{profile}, $args{os}, $args{arch}, $args{os});
 | |
|     unless ($template) {
 | |
|         $template = xCAT::SvrUtils::get_tmpl_file_name("$::XCATROOT/share/xcat/install/$plat", $args{profile}, $args{os}, $args{arch}, $args{os});
 | |
|     }
 | |
|     my $tmperr;
 | |
|     if (-r "$template") {
 | |
|         $tmperr = xCAT::Template->subvars($template, "$installdir/autoinst/" . $args{node}, $args{node}, undef, undef, undef, $args{tmpl_hash});
 | |
|     } else {
 | |
|         $tmperr = "Unable to find template in /install/custom/install/$plat or $::XCATROOT/share/xcat/install/$plat (for $args{profile}/$args{os}/$args{arch} combination)";
 | |
|     }
 | |
|     if ($tmperr) {
 | |
|         xCAT::SvrUtils::sendmsg([ 1, $tmperr ], $output_handler, $args{node});
 | |
|     }
 | |
| 
 | |
| }
 | |
| 
 | |
| sub mkinstall {
 | |
|     return mkcommonboot("install", @_);
 | |
| }
 | |
| 
 | |
| sub mknetboot {
 | |
|     return mkcommonboot("stateless", @_);
 | |
| }
 | |
| 
 | |
| sub merge_esxi5_append {
 | |
|     my $tmpl    = shift;
 | |
|     my $append  = shift;
 | |
|     my $outfile = shift;
 | |
|     my $in;
 | |
|     my $out;
 | |
|     open($in,  "<", $tmpl);
 | |
|     open($out, ">", $outfile);
 | |
|     my $line;
 | |
| 
 | |
|     while ($line = <$in>) {
 | |
|         if ($line =~ /kernelopt=/) {
 | |
|             chomp($line);
 | |
|             $line .= $append . "\n";
 | |
| 
 | |
|             #if ($line =~ /modules=b.b00/) {
 | |
|             #	$line =~ s/modules=b.b00/modules=b.b00 $append/;
 | |
|         }
 | |
|         unless ($line =~ /^prefix=/) {
 | |
|             print $out $line;
 | |
|         }
 | |
|     }
 | |
| }
 | |
| 
 | |
| sub mkcommonboot {
 | |
|     my $bootmode      = shift;
 | |
|     my $req           = shift;
 | |
|     my $doreq         = shift;
 | |
|     my $globaltftpdir = "/tftpboot";
 | |
|     my @nodes         = @{ $req->{node} };
 | |
|     my $ostab         = xCAT::Table->new('nodetype');
 | |
| 
 | |
|     #my $sitetab  = xCAT::Table->new('site');
 | |
|     my $bptab = xCAT::Table->new('bootparams', -create => 1);
 | |
|     my $installroot = "/install";
 | |
| 
 | |
|     #if ($sitetab){
 | |
|     #(my $ref) = $sitetab->getAttribs({key => 'installdir'}, 'value');
 | |
|     my @entries = xCAT::TableUtils->get_site_attribute("installdir");
 | |
|     my $t_entry = $entries[0];
 | |
|     if (defined($t_entry)) {
 | |
|         $installroot = $t_entry;
 | |
|     }
 | |
| 
 | |
|     #($ref) = $sitetab->getAttribs({key => 'tftpdir'}, 'value');
 | |
|     @entries = xCAT::TableUtils->get_site_attribute("tftpdir");
 | |
|     $t_entry = $entries[0];
 | |
|     if (defined($t_entry)) {
 | |
|         $globaltftpdir = $t_entry;
 | |
|     }
 | |
| 
 | |
|     #}
 | |
|     my %donetftp = ();
 | |
| 
 | |
|     my $bpadds = $bptab->getNodesAttribs(\@nodes, ['addkcmdline']);
 | |
|     my $nodehmtab = xCAT::Table->new('nodehm', -create => 0);
 | |
|     my $serialconfig;
 | |
|     if ($nodehmtab) {
 | |
|         $serialconfig = $nodehmtab->getNodesAttribs(\@nodes, [ 'serialport', 'serialspeed' ]);
 | |
|     }
 | |
|     my $restab = xCAT::Table->new('noderes', -create => 0);
 | |
|     my $resents;
 | |
|     if ($restab) {
 | |
|         $resents = $restab->getNodesAttribs(\@nodes, [ 'tftpdir', 'nfsserver', 'xcatmaster', 'tftpserver']);
 | |
|     }
 | |
| 
 | |
|     my %tablecolumnsneededforaddkcmdline;
 | |
|     my %nodesubdata;
 | |
|     foreach my $key (keys %$bpadds) { #First, we identify all needed table.columns needed to aggregate database call
 | |
|         my $add = $bpadds->{$key}->[0]->{addkcmdline};
 | |
| 
 | |
|         next if !defined $add;
 | |
| 
 | |
|         while ($add =~ /#NODEATTRIB:([^:#]+):([^:#]+)#/) {
 | |
|             push @{ $tablecolumnsneededforaddkcmdline{$1} }, $2;
 | |
|             $add =~ s/#NODEATTRIB:([^:#]+):([^:#]+)#//;
 | |
|         }
 | |
|     }
 | |
|     foreach my $table (keys %tablecolumnsneededforaddkcmdline) {
 | |
|         my $tab = xCAT::Table->new($table, -create => 0);
 | |
|         if ($tab) {
 | |
|             $nodesubdata{$table} = $tab->getNodesAttribs(\@nodes, $tablecolumnsneededforaddkcmdline{$table});
 | |
|         }
 | |
|     }
 | |
| 
 | |
|     my $osents = $ostab->getNodesAttribs(\@nodes, [ 'os', 'arch', 'profile' ]);
 | |
|     foreach my $node (@nodes) {
 | |
|         my $ent     = $osents->{$node}->[0];
 | |
|         my $arch    = $ent->{'arch'};
 | |
|         my $profile = $ent->{'profile'};
 | |
|         my $osver   = $ent->{'os'};
 | |
|         my $tftpdir;
 | |
|         my $ksserver;
 | |
|         my %tmpl_hash;
 | |
| 
 | |
|         if ($resents and $resents->{$node}->[0]->{nfsserver}) {
 | |
|             $ksserver = $resents->{$node}->[0]->{nfsserver};
 | |
|         } else {
 | |
|             $ksserver = '!myipfn!';
 | |
|         }
 | |
| 
 | |
|         if ($resents and $resents->{$node}->[0]->{tftpdir}) {
 | |
|             $tftpdir = $resents->{$node}->[0]->{tftpdir};
 | |
|         } else {
 | |
|             $tftpdir = $globaltftpdir;
 | |
|         }
 | |
|         if ($resents and $resents->{$node}->[0]->{xcatmaster} ) {
 | |
|             $tmpl_hash{"xcatmaster"} = $resents->{$node}->[0]->{xcatmaster};
 | |
|         }
 | |
|         if ($resents and $resents->{$node}->[0]->{tftpserver}) {
 | |
|             $tmpl_hash{"tftpserver"} = $resents->{$node}->[0]->{tftpserver};
 | |
|         }
 | |
| 
 | |
|         #if($arch ne 'x86'){
 | |
|         #	xCAT::SvrUtils::sendmsg([1,"VMware ESX hypervisors are x86, please change the nodetype.arch value to x86 instead of $arch for $node before proceeding:
 | |
|         #e.g: nodech $node nodetype.arch=x86\n"]);
 | |
|         #	return;
 | |
|         #}
 | |
|         # first make sure copycds was done:
 | |
|         my $custprofpath = $profile;
 | |
|         unless ($custprofpath =~ /^\//) { #If profile begins with a /, assume it already is a path
 | |
|             $custprofpath = $installroot . "/custom/install/$osver/$arch/$profile";
 | |
|             unless (-d $custprofpath) {
 | |
|                 $custprofpath = $installroot . "/custom/install/esxi/$arch/$profile";
 | |
|             }
 | |
|         }
 | |
|         unless (
 | |
|             -r "$custprofpath/vmkboot.gz"
 | |
|             or -r "$custprofpath/b.z"
 | |
|             or -r "$custprofpath/mboot.c32"
 | |
|             or -r "$custprofpath/install.tgz"
 | |
|             or -r "$installroot/$osver/$arch/mboot.c32"
 | |
|             or -r "$installroot/$osver/$arch/install.tgz") {
 | |
|             xCAT::SvrUtils::sendmsg([ 1, "Please run copycds first for $osver or create custom image in $custprofpath/" ], $output_handler);
 | |
|         }
 | |
| 
 | |
|         my @reqmods = qw/vmkboot.gz vmk.gz sys.vgz cim.vgz/; #Required modules for an image to be considered complete
 | |
|         if (-r "$custprofpath/b.z") { #if someone hand extracts from imagedd, a different name scheme is used
 | |
|             @reqmods = qw/b.z k.z s.z c.z/;
 | |
|         }
 | |
|         my %mods;
 | |
|         foreach (@reqmods) {
 | |
|             $mods{$_} = 1;
 | |
|         }
 | |
|         my $shortprofname = $profile;
 | |
|         $shortprofname =~ s/\/\z//;
 | |
|         $shortprofname =~ s/.*\///;
 | |
|         mkpath("$tftpdir/xcat/netboot/$osver/$arch/$shortprofname/");
 | |
|         my $havemod = 0;
 | |
|         unless ($donetftp{ $osver, $arch, $profile, $tftpdir }) {
 | |
|             my $srcdir = "$installroot/$osver/$arch";
 | |
|             my $dest   = "$tftpdir/xcat/netboot/$osver/$arch/$shortprofname";
 | |
|             cpNetbootImages($osver, $srcdir, $dest, $custprofpath, \%mods, bootmode => $bootmode);
 | |
|             if ($havemod = makecustomizedmod($osver, $dest)) {
 | |
|                 push @reqmods, "mod.tgz";
 | |
|                 $mods{"mod.tgz"} = 1;
 | |
|             }
 | |
|             if ($osver =~ /esxi4/ and -r "$::XCATROOT/share/xcat/netboot/syslinux/mboot.c32") { #prefer xCAT patched mboot.c32 with BOOTIF for mboot
 | |
|                 copy("$::XCATROOT/share/xcat/netboot/syslinux/mboot.c32", $dest);
 | |
|             } elsif (-r "$custprofpath/mboot.c32") {
 | |
|                 copy("$custprofpath/mboot.c32", $dest);
 | |
|             } elsif (-r "$srcdir/mboot.c32") {
 | |
|                 copy("$srcdir/mboot.c32", $dest);
 | |
|             }
 | |
|             if (-f "$srcdir/efiboot.img") {
 | |
|                 copy("$srcdir/efiboot.img", $dest);
 | |
|                 print("$srcdir/efi");
 | |
|                 mkpath("$dest/efi");
 | |
|                 recursion_copy("$srcdir/efi", "$dest/efi");
 | |
|             }
 | |
|             $donetftp{ $osver, $arch, $profile, $tftpdir } = 1;
 | |
|         }
 | |
|         my $tp = "xcat/netboot/$osver/$arch/$shortprofname";
 | |
|         my $kernel;
 | |
|         my $kcmdline;
 | |
|         my $append;
 | |
|         my $shortappend;
 | |
|         if ($osver =~ /esxi4/) {
 | |
|             my $bail = 0;
 | |
|             foreach (@reqmods) {
 | |
|                 unless (-r "$tftpdir/$tp/$_") {
 | |
|                     xCAT::SvrUtils::sendmsg([ 1, "$_ is missing from the target destination, ensure that either copycds has been run or that $custprofpath contains this file" ], $output_handler);
 | |
|                     $bail = 1; #only flag to bail, present as many messages as possible to user
 | |
|                 }
 | |
|             }
 | |
|             if ($bail) { #if the above loop detected one or more failures, bail out
 | |
|                 return;
 | |
|             }
 | |
| 
 | |
|             # now make <HEX> file entry stuff
 | |
|             $kernel = "$tp/mboot.c32";
 | |
|             my $prepend;
 | |
|             if ($reqmods[0] eq "vmkboot.gz") {
 | |
|                 $prepend = "$tp/vmkboot.gz";
 | |
|                 delete $mods{"vmkboot.gz"};
 | |
|                 $append = " --- $tp/vmk.gz";
 | |
|                 delete $mods{"vmk.gz"};
 | |
|                 $append .= " --- $tp/sys.vgz";
 | |
|                 delete $mods{"sys.vgz"};
 | |
|                 $append .= " --- $tp/cim.vgz";
 | |
|                 delete $mods{"cim.vgz"};
 | |
|             } else {    #the single letter style
 | |
|                 $prepend = "$tp/b.z";
 | |
|                 delete $mods{"b.z"};
 | |
|                 $append = " --- $tp/k.z";
 | |
|                 delete $mods{"k.z"};
 | |
|                 $append .= " --- $tp/s.z";
 | |
|                 delete $mods{"s.z"};
 | |
|                 $append .= " --- $tp/c.z";
 | |
|                 delete $mods{"c.z"};
 | |
|             }
 | |
| 
 | |
|             if ($mods{"mod.tgz"}) {
 | |
|                 $append .= " --- $tp/mod.tgz";
 | |
|                 delete $mods{"mod.tgz"};
 | |
|             }
 | |
|             foreach (keys %mods) {
 | |
|                 $append .= " --- $tp/$_";
 | |
|             }
 | |
|             if (defined $bpadds->{$node}->[0]->{addkcmdline}) {
 | |
|                 my $modules;
 | |
|                 ($kcmdline, $modules) = split /---/, $bpadds->{$node}->[0]->{addkcmdline}, 2;
 | |
|                 $kcmdline =~ s/#NODEATTRIB:([^:#]+):([^:#]+)#/$nodesubdata{$1}->{$node}->[0]->{$2}/eg;
 | |
|                 if ($modules) {
 | |
|                     $append .= " --- " . $modules;
 | |
|                 }
 | |
|                 $prepend .= " " . $kcmdline;
 | |
|             }
 | |
|             $append = $prepend . $append;
 | |
|         }
 | |
|         elsif ($osver =~ /esxi[56]/) {    #do a more straightforward thing..
 | |
|             $kernel = "$tp/mboot.c32";
 | |
|             if (-r "$tftpdir/$tp/boot.cfg.$bootmode.tmpl") { #so much for straightforward..
 | |
|                 $shortappend = "-c $tp/boot.cfg.$bootmode.$node";
 | |
|             } else {
 | |
|                 $append = "-c $tp/boot.cfg.$bootmode";
 | |
|             }
 | |
|             $append .= " xcatd=$ksserver:3001";
 | |
|             if ($bootmode eq "install") {
 | |
|                 $append .= " ks=http://$ksserver/install/autoinst/$node";
 | |
|                 esxi_kickstart_from_template(node => $node, os => $osver, arch => $arch, profile => $profile, tmpl_hash => \%tmpl_hash);
 | |
|             }
 | |
|             if ($bootmode ne "install" and $serialconfig->{$node}) { #don't do it for install, installer croaks currently
 | |
|                 my $comport = 1;
 | |
|                 if (defined $serialconfig->{$node}->[0]->{serialport}) {
 | |
|                     $comport = $serialconfig->{$node}->[0]->{serialport} + 1;
 | |
|                     $append .= " -S $comport tty2port=com$comport";
 | |
|                 }
 | |
|                 if (defined $serialconfig->{$node}->[0]->{serialspeed}) {
 | |
|                     $append .= " -s " . $serialconfig->{$node}->[0]->{serialspeed} . " com" . $comport . "_baud=" . $serialconfig->{$node}->[0]->{serialspeed};
 | |
|                 }
 | |
|             }
 | |
|             if (defined $bpadds->{$node}->[0]->{addkcmdline}) {
 | |
|                 $append .= " " . $bpadds->{$node}->[0]->{addkcmdline};
 | |
|                 $append =~ s/#NODEATTRIB:([^:#]+):([^:#]+)#/$nodesubdata{$1}->{$node}->[0]->{$2}/eg;
 | |
|             }
 | |
|         }
 | |
|         if ($shortappend) { #esxi5 user desiring to put everything in one boot config file. . .
 | |
|             merge_esxi5_append("$tftpdir/$tp/boot.cfg.$bootmode.tmpl", $append, "$tftpdir/$tp/boot.cfg.$bootmode.$node");
 | |
|             $append = $shortappend;
 | |
|         }
 | |
|         $output_handler->({ node => [ { name => [$node], '_addkcmdlinehandled' => [1] } ] });
 | |
| 
 | |
| 
 | |
| 
 | |
|         my $bootparams = ${$req->{bootparams}};
 | |
|         $bootparams->{$node}->[0]->{kernel} = $kernel;
 | |
|         $bootparams->{$node}->[0]->{kcmdline} = $append;
 | |
|         $bootparams->{$node}->[0]->{initrd} = "";
 | |
|     }    # end of node loop
 | |
| 
 | |
| }
 | |
| 
 | |
| # this is where we extract the netboot images out of the copied ISO image
 | |
| sub cpNetbootImages {
 | |
|     my $osver        = shift;
 | |
|     my $srcDir       = shift;
 | |
|     my $destDir      = shift;
 | |
|     my $overridedir  = shift;
 | |
|     my $modulestoadd = shift;
 | |
|     my %parmargs     = @_;
 | |
|     my $bootmode     = "stateless";
 | |
|     if ($parmargs{bootmode}) { $bootmode = $parmargs{bootmode} }
 | |
|     my $tmpDir = "/tmp/xcat.$$";
 | |
| 
 | |
|     if ($osver =~ /esxi4/) {
 | |
| 
 | |
|         # we don't want to go through this all the time, so if its already
 | |
|         # there we're not going to extract:
 | |
|         unless (-r "$destDir/vmk.gz"
 | |
|             and -r "$destDir/vmkboot.gz"
 | |
|             and -r "$destDir/sys.vgz"
 | |
|             and -r "$destDir/cim.vgz"
 | |
|             and -r "$destDir/cimstg.tgz"
 | |
|           ) {
 | |
|             if (-r "$srcDir/image.tgz") { #it still may work without image.tgz if profile customization has everything replaced
 | |
|                 mkdir($tmpDir);
 | |
|                 chdir($tmpDir);
 | |
|                 xCAT::SvrUtils::sendmsg("extracting netboot files from OS image.  This may take about a minute or two...hopefully you have ~1GB free in your /tmp dir\n", $output_handler);
 | |
|                 my $cmd = "tar zxf $srcDir/image.tgz";
 | |
|                 if (system($cmd)) {
 | |
|                     xCAT::SvrUtils::sendmsg([ 1, "Unable to extract $srcDir/image.tgz\n" ], $output_handler);
 | |
|                 }
 | |
| 
 | |
|                 # this has the big image and may take a while.
 | |
|                 # this should now create:
 | |
|                 # /tmp/xcat.1234/usr/lib/vmware/installer/VMware-VMvisor-big-164009-x86_64.dd.bz2 or some other version.  We need to extract partition 5 from it.
 | |
|                 system("bunzip2 $tmpDir/usr/lib/vmware/installer/*bz2");
 | |
|                 xCAT::SvrUtils::sendmsg("finished extracting, now copying files...\n", $output_handler);
 | |
| 
 | |
|                 # now we need to get partition 5 which has the installation goods in it.
 | |
|                 my $scmd = "fdisk -lu $tmpDir/usr/lib/vmware/installer/*dd 2>&1 | grep dd5 | awk '{print \$2}'";
 | |
|                 my $sector = `$scmd`;
 | |
|                 chomp($sector);
 | |
|                 my $offset = $sector * 512;
 | |
|                 mkdir "/mnt/xcat";
 | |
|                 my $mntcmd = "mount $tmpDir/usr/lib/vmware/installer/*dd  /mnt/xcat -o loop,offset=$offset";
 | |
|                 if (system($mntcmd)) {
 | |
|                     xCAT::SvrUtils::sendmsg([ 1, "unable to mount partition 5 of the ESX netboot image to /mnt/xcat" ], $output_handler);
 | |
|                     return;
 | |
|                 }
 | |
| 
 | |
|                 if (!-d $destDir) {
 | |
|                     if (-e $destDir) {
 | |
|                         xCAT::SvrUtils::sendmsg([ 1, "Could not copy netboot contents to $destDir, it exists but is not currently a directory" ], $output_handler);
 | |
|                         return;
 | |
|                     }
 | |
|                     mkpath($destDir);
 | |
|                 }
 | |
| 
 | |
|                 if (system("cp /mnt/xcat/* $destDir/")) {
 | |
|                     xCAT::SvrUtils::sendmsg([ 1, "Could not copy netboot contents to $destDir" ], $output_handler);
 | |
|                     chdir("/");
 | |
|                     system("umount /mnt/xcat");
 | |
|                     return;
 | |
|                 }
 | |
|                 chdir("/tmp");
 | |
|                 system("umount /mnt/xcat");
 | |
|                 system("rm -rf $tmpDir");
 | |
|             } elsif (-r "$srcDir/cim.vgz" and -r "$srcDir/vmkernel.gz" and -r "$srcDir/vmkboot.gz" and -r "$srcDir/sys.vgz") {
 | |
|                 use File::Basename;
 | |
|                 if (!-d $destDir) {
 | |
|                     mkpath($destDir);
 | |
|                 }
 | |
| 
 | |
|                 #In ESXI 4.1, the above breaks, this seems to work, much simpler too
 | |
|                 foreach ("$srcDir/cim.vgz", "$srcDir/vmkernel.gz", "$srcDir/vmkboot.gz", "$srcDir/sys.vgz", "$srcDir/sys.vgz") {
 | |
|                     my $mod = scalar fileparse($_);
 | |
|                     if ($mod =~ /vmkernel.gz/) {
 | |
|                         copy($_, "$destDir/vmk.gz") or xCAT::SvrUtils::sendmsg([ 1, "Could not copy netboot contents from $_ to $destDir/$mod" ], $output_handler);
 | |
|                     } else {
 | |
|                         copy($_, "$destDir/$mod") or xCAT::SvrUtils::sendmsg([ 1, "Could not copy netboot contents from $_ to $destDir/$mod" ], $output_handler);
 | |
|                     }
 | |
|                 }
 | |
| 
 | |
|             }
 | |
|         }
 | |
| 
 | |
|         #this is the override directory if there is one, otherwise it's actually the default dir
 | |
|         if (-d $overridedir) {
 | |
|             mkdir($overridedir);
 | |
|         }
 | |
| 
 | |
|         #Copy over all modules
 | |
|         use File::Basename;
 | |
|         foreach (glob "$overridedir/*") {
 | |
|             my $mod = scalar fileparse($_);
 | |
|             if ($mod =~ /gz\z/ and $mod !~ /pkgdb.tgz/ and $mod !~ /vmkernel.gz/) {
 | |
|                 $modulestoadd->{$mod} = 1;
 | |
|                 copy($_, "$destDir/$mod") or xCAT::SvrUtils::sendmsg([ 1, "Could not copy netboot contents from $overridedir to $destDir" ], $output_handler);
 | |
|             } elsif ($mod =~ /vmkernel.gz/) {
 | |
|                 $modulestoadd->{"vmk.gz"} = 1;
 | |
|                 copy($_, "$destDir/vmk.gz") or xCAT::SvrUtils::sendmsg([ 1, "Could not copy netboot contents from $overridedir to $destDir" ], $output_handler);
 | |
|             }
 | |
|         }
 | |
| 
 | |
|     } elsif ($osver =~ /esxi[56]/) {    #we need boot.cfg.stateles
 | |
|         my @filestocopy = ("boot.cfg.$bootmode");
 | |
|         if (-r "$srcDir/boot.cfg.$bootmode" or -r "$overridedir/boot.cfg.$bootmode") {
 | |
|             @filestocopy = ("boot.cfg.$bootmode");
 | |
|         } elsif (-r "$srcDir/boot.cfg.$bootmode.tmpl" or -r "$overridedir/boot.cfg.$bootmode.tmpl") {
 | |
|             @filestocopy = ("boot.cfg.$bootmode.tmpl");
 | |
|         } else {
 | |
|             xCAT::SvrUtils::sendmsg([ 1, "$srcDir is missing boot.cfg.$bootmode file required for $bootmode boot" ], $output_handler);
 | |
|             return;
 | |
|         }
 | |
|         my $statelesscfg;
 | |
|         if (-r "$overridedir/boot.cfg.$bootmode.tmpl") {
 | |
|             open($statelesscfg, "<", "$overridedir/boot.cfg.$bootmode.tmpl");
 | |
|             @filestocopy = ("boot.cfg.$bootmode.tmpl");
 | |
|         } elsif (-r "$overridedir/boot.cfg.$bootmode") {
 | |
|             open($statelesscfg, "<", "$overridedir/boot.cfg.$bootmode");
 | |
|         } elsif (-r "$srcDir/boot.cfg.$bootmode.tmpl") {
 | |
|             @filestocopy = ("boot.cfg.$bootmode.tmpl");
 | |
|             open($statelesscfg, "<", "$srcDir/boot.cfg.$bootmode.tmpl");
 | |
|         } elsif (-r "$srcDir/boot.cfg.$bootmode") {
 | |
|             open($statelesscfg, "<", "$srcDir/boot.cfg.$bootmode");
 | |
|         } else {
 | |
|             die "boot.cfg.$bootmode was missing from $srcDir???";
 | |
|         }
 | |
|         my @statelesscfg = <$statelesscfg>;
 | |
| 
 | |
|         foreach (@statelesscfg) { #search for files specified by the boot cfg and pull them in
 | |
|             if (/^kernel=(.*)/) {
 | |
|                 push @filestocopy, $1;
 | |
|             } elsif (/^modules=(.*)/) {
 | |
|                 foreach (split / --- /, $1) {
 | |
|                     s/^\s*//;
 | |
|                     s/\s.*//;
 | |
|                     push @filestocopy, $_;
 | |
|                 }
 | |
|             }
 | |
|         }
 | |
| 
 | |
|         #now that we have a list, do the copy (mostly redundant, but PXE needs them tftp accessible)
 | |
|         foreach (@filestocopy) {
 | |
|             chomp;
 | |
|             s/ *\z//;
 | |
|             my $mod = scalar fileparse($_);
 | |
|             if (-r "$overridedir/$mod") {
 | |
|                 copyIfNewer("$overridedir/$mod", "$destDir/$mod") or xCAT::SvrUtils::sendmsg([ 1, "Could not copy netboot contents from $overridedir/$mod to $destDir/$mod, $!" ], $output_handler);
 | |
|             } elsif (-r "$srcDir/$mod") {
 | |
|                 copyIfNewer($srcDir . "/" . $mod, "$destDir/$mod") or xCAT::SvrUtils::sendmsg([ 1, "Could not copy netboot contents from $srcDir/$mod to $destDir/$mod, $!" ], $output_handler);
 | |
|             } elsif ($mod ne "xcatmod.tgz") {
 | |
|                 xCAT::SvrUtils::sendmsg([ 1, "Could not copy netboot contents from $srcDir/$mod to $destDir/$mod, $srcDir/$mod not found" ], $output_handler);
 | |
|             }
 | |
|         }
 | |
|     } else {
 | |
|         xCAT::SvrUtils::sendmsg([ 1, "VMware $osver is not supported for netboot" ], $output_handler);
 | |
|     }
 | |
| 
 | |
| }
 | |
| 
 | |
| sub copyIfNewer {
 | |
|     my $source = shift;
 | |
|     my $dest   = shift;
 | |
|     if (!-e $dest or -C $source < -C $dest) {
 | |
|         return copy($source, $dest);
 | |
|     }
 | |
|     return 1;
 | |
| }
 | |
| 
 | |
| # compares nfs target described by parameters to every share mounted by target hypervisor
 | |
| # returns 1 if matching datastore is present and 0 otherwise
 | |
| sub match_nfs_datastore {
 | |
|     my ($host, $path, $hypconn) = @_;
 | |
| 
 | |
|     die "esx plugin bug: no host provided for match_datastore" unless defined $host;
 | |
|     die "esx plugin bug: no path provided for match_datastore" unless defined $path;
 | |
| 
 | |
|     my @ip;
 | |
| 
 | |
|     eval {
 | |
|         if ($host =~ m/\b(?:(?:25[0-5]|2[0-4][0-9]|[01]?[0-9][0-9]?)\.){3}(?:25[0-5]|2[0-4][0-9]|[01]?[0-9][0-9]?)\//) {
 | |
|             use Socket;
 | |
| 
 | |
|             @ip = ($host);
 | |
|             $host = gethostbyaddr(inet_aton($host, AF_INET), AF_INET);
 | |
|         } else {
 | |
|             use Socket;
 | |
| 
 | |
|             (undef, undef, undef, undef, @ip) = gethostbyname($host);
 | |
| 
 | |
|             my @ip_ntoa = ();
 | |
|             foreach (@ip) {
 | |
|                 push(@ip_ntoa, inet_ntoa($_));
 | |
|             }
 | |
|             @ip = @ip_ntoa;
 | |
|         }
 | |
| 
 | |
|     };
 | |
| 
 | |
|     if ($@) {
 | |
|         die "error while resolving datastore host: $@\n";
 | |
|     }
 | |
| 
 | |
|     my %viewcrit = (
 | |
|         view_type  => 'HostSystem',
 | |
|         properties => ['config.fileSystemVolume'],
 | |
|     );
 | |
| 
 | |
|     my $dsviews = $hypconn->find_entity_views(%viewcrit);
 | |
| 
 | |
|     foreach (@$dsviews) {
 | |
|         foreach my $mount (@{ $_->get_property('config.fileSystemVolume.mountInfo') }) {
 | |
|             next unless $mount->{'volume'}{'type'} eq 'NFS';
 | |
| 
 | |
|             my $hostMatch = 0;
 | |
|           HOSTMATCH: foreach (@ip, $host) {
 | |
|                 next HOSTMATCH unless $mount->{'volume'}{'remoteHost'} eq $_;
 | |
| 
 | |
|                 $hostMatch = 1;
 | |
|                 last HOSTMATCH;
 | |
|             }
 | |
|             next unless $hostMatch;
 | |
| 
 | |
|             next unless $mount->{'volume'}{'remotePath'} eq $path;
 | |
| 
 | |
|             return 1;
 | |
|         }
 | |
|     }
 | |
| 
 | |
|     return 0;
 | |
| }
 | |
| 
 | |
| 1;
 | |
| 
 | |
| # vi: set ts=4 sw=4 filetype=perl:
 |