5238 lines
205 KiB
Perl
5238 lines
205 KiB
Perl
package xCAT_plugin::esx;
|
|
|
|
use strict;
|
|
use warnings;
|
|
use xCAT::Table;
|
|
use xCAT::Utils;
|
|
use xCAT::TableUtils;
|
|
use xCAT::ServiceNodeUtils;
|
|
use xCAT::TZUtils;
|
|
use Time::HiRes qw (sleep);
|
|
use xCAT::Template;
|
|
use xCAT::MsgUtils;
|
|
use xCAT::SvrUtils;
|
|
use xCAT::NodeRange;
|
|
use xCAT::Common;
|
|
use xCAT::VMCommon;
|
|
use POSIX "WNOHANG";
|
|
use Getopt::Long;
|
|
use Thread qw(yield);
|
|
use POSIX qw(WNOHANG nice);
|
|
use File::Path qw/mkpath rmtree/;
|
|
use File::Temp qw/tempdir/;
|
|
use File::Copy;
|
|
use Fcntl qw/:flock/;
|
|
use IO::Socket; #Need name resolution
|
|
use Scalar::Util qw/looks_like_number/;
|
|
#use Data::Dumper;
|
|
Getopt::Long::Configure("bundling");
|
|
Getopt::Long::Configure("pass_through");
|
|
my @cpiopid;
|
|
our @ISA = 'xCAT::Common';
|
|
|
|
|
|
#in xCAT, the lifetime of a process ends on every request
|
|
#therefore, the lifetime of assignments to these glabals as architected
|
|
#is to be cleared on every request
|
|
#my %esx_comm_pids;
|
|
my %limbonodes; #nodes in limbo during a forced migration due to missing parent
|
|
my %hyphash; #A data structure to hold hypervisor-wide variables (i.e. the current resource pool, virtual machine folder, connection object
|
|
my %vcenterhash; #A data structure to reflect the state of vcenter connectivity to hypervisors
|
|
my %vmhash; #store per vm info of interest
|
|
my %clusterhash;
|
|
my %hypready; #A structure for hypervisor readiness to be tracked before proceeding to normal operations
|
|
my %running_tasks; #A struct to track this processes
|
|
my $output_handler; #Pointer to the function to drive results to client
|
|
my $executerequest;
|
|
my $usehostnamesforvcenter;
|
|
my %tablecfg; #to hold the tables
|
|
my %hostrefbynode;
|
|
my $currkey;
|
|
my $requester;
|
|
my $viavcenter;
|
|
my $viavcenterbyhyp;
|
|
my $vcenterautojoin=1;
|
|
my $datastoreautomount=1;
|
|
my $vcenterforceremove=0; #used in rmhypervisor
|
|
my $reconfigreset=1;
|
|
my $vmwaresdkdetect = eval {
|
|
require VMware::VIRuntime;
|
|
VMware::VIRuntime->import();
|
|
1;
|
|
};
|
|
my %lockhandles;
|
|
|
|
sub recursion_copy {
|
|
my $source = shift;
|
|
my $destination = shift;
|
|
my $dirhandle;
|
|
opendir($dirhandle,$source);
|
|
my $entry;
|
|
foreach $entry (readdir($dirhandle)) {
|
|
if ($entry eq '.' or $entry eq '..') { next; }
|
|
my $tempsource = "$source/$entry";
|
|
my $tempdestination = "$destination/$entry";
|
|
if ( -d $tempsource ) {
|
|
unless (-d $tempdestination) { mkdir $tempdestination or die "failure creating directory $tempdestination, $!"; }
|
|
recursion_copy($tempsource,$tempdestination);
|
|
} else {
|
|
copy($tempsource,$tempdestination) or die "failed copy from $tempsource to $tempdestination, $!";
|
|
}
|
|
}
|
|
}
|
|
sub lockbyname {
|
|
my $name = shift;
|
|
my $lckh;
|
|
mkpath("/tmp/xcat/locks/");
|
|
while (-e "/tmp/xcat/locks/$name") { sleep 1; }
|
|
open($lockhandles{$name},">>","/tmp/xcat/locks/$name");
|
|
flock($lockhandles{$name},LOCK_EX);
|
|
}
|
|
sub unlockbyname {
|
|
my $name = shift;
|
|
unlink("/tmp/xcat/locks/$name");
|
|
close($lockhandles{$name});
|
|
}
|
|
|
|
my %guestidmap = (
|
|
"rhel.6.*" => "rhel6_",
|
|
"rhel.5.*" => "rhel5_",
|
|
"rhel4.*" => "rhel4_",
|
|
"centos6.*" => "rhel6_",
|
|
"centos5.*" => "rhel5_",
|
|
"centos4.*" => "rhel4_",
|
|
"sles12.*" => "sles12_",
|
|
"sles11.*" => "sles11_",
|
|
"sles10.*" => "sles10_",
|
|
"win2k8" => "winLonghorn",
|
|
"win2k8r2" => "windows7Server",
|
|
"win2012" => "windows8Server",
|
|
"hyperv2012" => "windows8Server",
|
|
"esix5.*" => "vmkernel5",
|
|
"esix4.*" => "vmkernel",
|
|
"win8" => "windows8_",
|
|
"win7" => "windows7_",
|
|
"win2k3" => "winNetStandard",
|
|
"imagex" => "winNetStandard",
|
|
"boottarget" => "otherLinux"
|
|
#otherGuest, otherGuest64, otherLinuxGuest, otherLinux64Guest
|
|
);
|
|
|
|
sub handled_commands{
|
|
return {
|
|
copycd => 'esx',
|
|
mknetboot => "nodetype:os=(esxi.*)",
|
|
mkinstall => "nodetype:os=(esxi5.*)",
|
|
rpower => 'nodehm:power,mgt',
|
|
esxiready => "esx",
|
|
rsetboot => 'nodehm:power,mgt',
|
|
rmigrate => 'nodehm:power,mgt',
|
|
formatdisk => "nodetype:os=(esxi.*)",
|
|
rescansan => "nodetype:os=(esxi.*)",
|
|
mkvm => 'nodehm:mgt',
|
|
rmvm => 'nodehm:mgt',
|
|
clonevm => 'nodehm:mgt',
|
|
createvcluster => 'esx',
|
|
lsvcluster => 'esx',
|
|
rmvcluster => 'esx',
|
|
rinv => 'nodehm:mgt',
|
|
chvm => 'nodehm:mgt',
|
|
rshutdown => "nodetype:os=(esxi.*)",
|
|
lsvm => ['hypervisor:type','nodetype:os=(esx.*)'],
|
|
rmhypervisor => ['hypervisor:type','nodetype:os=(esx.*)'],
|
|
chhypervisor => ['hypervisor:type','nodetype:os=(esx.*)'],
|
|
#lsvm => 'nodehm:mgt', not really supported yet
|
|
};
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
sub preprocess_request {
|
|
my $request = shift;
|
|
my $callback = shift;
|
|
if ($request->{command}->[0] eq 'createvcluster' or $request->{command}->[0] eq 'lsvcluster' or $request->{command}->[0] eq 'rmvcluster') {
|
|
return [$request];
|
|
}
|
|
#if already preprocessed, go straight to request
|
|
if ( (defined($request->{_xcatpreprocessed}))
|
|
&& ($request->{_xcatpreprocessed}->[0] == 1))
|
|
{
|
|
return [$request];
|
|
}
|
|
|
|
my $username = 'root';
|
|
my $password = '';
|
|
my $vusername = "Administrator";
|
|
my $vpassword = "";
|
|
|
|
unless ($request and $request->{command} and $request->{command}->[0]) { return; }
|
|
|
|
if ($request->{command}->[0] eq 'copycd')
|
|
{ #don't farm out copycd
|
|
return [$request];
|
|
}elsif($request->{command}->[0] eq 'mknetboot'
|
|
or $request->{command}->[0] eq 'mkinstall'){
|
|
return [$request];
|
|
}
|
|
xCAT::Common::usage_noderange($request,$callback);
|
|
|
|
if ($request->{_xcatpreprocessed} and $request->{_xcatpreprocessed}->[0] == 1) { return [$request]; }
|
|
# exit if preprocesses
|
|
my @requests;
|
|
|
|
my $noderange;
|
|
my $command = $request->{command}->[0];
|
|
if ($request->{node}) {
|
|
$noderange = $request->{node}; # array ref
|
|
} elsif ($command eq "esxiready") {
|
|
my $node;
|
|
($node) = noderange($request->{'_xcat_clienthost'}->[0]);
|
|
$noderange = [$node];
|
|
$request->{node} = $noderange;
|
|
}
|
|
|
|
my $extraargs = $request->{arg};
|
|
my @exargs=($request->{arg});
|
|
my %hyp_hash = ();
|
|
my %cluster_hash=();
|
|
|
|
# Get nodes from mp table and assign nodes to mp hash.
|
|
my $passtab = xCAT::Table->new('passwd');
|
|
my $tmp;
|
|
if ($passtab) {
|
|
($tmp) = $passtab->getAttribs({'key'=>'vmware'},'username','password');
|
|
if (defined($tmp)) {
|
|
$username = $tmp->{username};
|
|
$password = $tmp->{password};
|
|
}
|
|
($tmp) = $passtab->getAttribs({'key'=>'vcenter'},'username','password');
|
|
if (defined($tmp)) {
|
|
$vusername = $tmp->{username};
|
|
$vpassword = $tmp->{password};
|
|
}
|
|
}
|
|
|
|
my $vmtab = xCAT::Table->new("vm");
|
|
unless($vmtab){
|
|
$callback->({data=>["Cannot open vm table"]});
|
|
$request = {};
|
|
return;
|
|
}
|
|
|
|
my $vmtabhash = $vmtab->getNodesAttribs($noderange,['host','migrationdest']);
|
|
foreach my $node (@$noderange){
|
|
if ($command eq "rmhypervisor" or $command eq 'lsvm' or $command eq 'esxiready' or $command eq 'rshutdown' or $command eq "chhypervisor" or $command eq "formatdisk" or $command eq 'rescansan') {
|
|
$hyp_hash{$node}{nodes} = [$node];
|
|
} else {
|
|
my $ent = $vmtabhash->{$node}->[0];
|
|
if(defined($ent->{host})) {
|
|
push @{$hyp_hash{$ent->{host}}{nodes}}, $node;
|
|
} elsif (defined($ent->{migrationdest})) {
|
|
$cluster_hash{$ent->{migrationdest}}->{nodes}->{$node}=1;
|
|
} else {
|
|
xCAT::SvrUtils::sendmsg([1,": no host or cluster defined for guest"], $callback,$node);
|
|
}
|
|
}
|
|
}
|
|
|
|
# find service nodes for the MMs
|
|
# build an individual request for each service node
|
|
my $service = "xcat";
|
|
my @hyps=keys(%hyp_hash);
|
|
my %targethyps;
|
|
if ($command eq 'rmigrate' and (scalar @{$extraargs} >= 1)) {
|
|
@ARGV=@{$extraargs};
|
|
my $offline;
|
|
my $junk;
|
|
GetOptions(
|
|
"f" => \$offline,
|
|
"s=s" => \$junk #wo don't care about it, but suck up nfs:// targets so they don't get added
|
|
);
|
|
my $dsthyp = $ARGV[0];
|
|
if ($dsthyp) {
|
|
push @hyps,$dsthyp;
|
|
$targethyps{$dsthyp}=1;
|
|
}
|
|
}
|
|
#TODO: per hypervisor table password lookup
|
|
my @allnodes;
|
|
push @allnodes,@hyps;
|
|
push @allnodes,@$noderange;
|
|
my $sn = xCAT::ServiceNodeUtils->get_ServiceNode(\@allnodes, $service, "MN");
|
|
#vmtabhash was from when we had vm.host do double duty for hypervisor data
|
|
#$vmtabhash = $vmtab->getNodesAttribs(\@hyps,['host']);
|
|
#We now use hypervisor fields to be unambiguous
|
|
my $hyptab = xCAT::Table->new('hypervisor');
|
|
my $hyptabhash={};
|
|
if ($hyptab) {
|
|
$hyptabhash = $hyptab->getNodesAttribs(\@hyps,['mgr']);
|
|
}
|
|
|
|
|
|
# build each request for each service node
|
|
foreach my $snkey (keys %$sn){
|
|
my $reqcopy = {%$request};
|
|
$reqcopy->{'_xcatdest'} = $snkey;
|
|
$reqcopy->{_xcatpreprocessed}->[0] = 1;
|
|
my $hyps1=$sn->{$snkey};
|
|
my @moreinfo=();
|
|
my @nodes=();
|
|
foreach (@$hyps1) { #This preserves the constructed data to avoid redundant table lookup
|
|
my $cfgdata;
|
|
if (not $targethyps{$_} and not $hyp_hash{$_}) { #a vm, skip it
|
|
next;
|
|
} elsif ($hyp_hash{$_}{nodes}) {
|
|
push @nodes, @{$hyp_hash{$_}{nodes}};
|
|
$cfgdata = "[$_][".join(',',@{$hyp_hash{$_}{nodes}})."][$username][$password][$vusername][$vpassword]"; #TODO: not use vm.host?
|
|
} else {
|
|
$cfgdata = "[$_][][$username][$password][$vusername][$vpassword]"; #TODO: not use vm.host?
|
|
}
|
|
if (defined $hyptabhash->{$_}->[0]->{mgr}) {
|
|
$cfgdata .= "[". $hyptabhash->{$_}->[0]->{mgr}."]";
|
|
} else {
|
|
$cfgdata .= "[]";
|
|
}
|
|
push @moreinfo, $cfgdata; #"[$_][".join(',',@{$hyp_hash{$_}{nodes}})."][$username][$password]";
|
|
}
|
|
foreach (keys %cluster_hash) {
|
|
my $cluster;
|
|
my $vcenter;
|
|
if (/@/) {
|
|
($cluster,$vcenter) = split /@/,$_,2;
|
|
} else {
|
|
die "TODO: implement default vcenter (for now, user, do vm.migratiodest=cluster".'@'."vcentername)";
|
|
}
|
|
push @moreinfo,"[CLUSTER:$cluster][".join(',',keys %{$cluster_hash{$_}->{nodes}})."][$username][$password][$vusername][$vpassword][$vcenter]";
|
|
}
|
|
if (scalar @nodes) {
|
|
$reqcopy->{node} = \@nodes;
|
|
}
|
|
#print "nodes=@nodes\n";
|
|
$reqcopy->{moreinfo}=\@moreinfo;
|
|
push @requests, $reqcopy;
|
|
}
|
|
return \@requests;
|
|
}
|
|
|
|
|
|
|
|
sub process_request {
|
|
#$SIG{INT} = $SIG{TERM} = sub{
|
|
# foreach (keys %esx_comm_pids){
|
|
# kill 2,$_;
|
|
# }
|
|
# exit 0;
|
|
#};
|
|
|
|
my $request = shift;
|
|
$output_handler = shift;
|
|
$executerequest = shift;
|
|
if ($request->{_xcat_authname}->[0]) {
|
|
$requester=$request->{_xcat_authname}->[0];
|
|
}
|
|
%vcenterhash = ();#A data structure to reflect the state of vcenter connectivity to hypervisors
|
|
my $level = shift;
|
|
my $distname = undef;
|
|
my $arch = undef;
|
|
my $path = undef;
|
|
my $command = $request->{command}->[0];
|
|
#The first segment is fulfilling the role of this plugin as
|
|
#a hypervisor provisioning plugin (akin to anaconda, windows, sles plugins)
|
|
if($command eq 'copycd'){
|
|
return copycd($request,$executerequest);
|
|
}elsif($command eq 'mkinstall'){
|
|
return mkinstall($request,$executerequest);
|
|
}elsif($command eq 'mknetboot'){
|
|
return mknetboot($request,$executerequest);
|
|
}
|
|
#From here on out, code for managing guests under VMware
|
|
#Detect whether or not the VMware SDK is available on this specific system
|
|
unless ($vmwaresdkdetect) {
|
|
$vmwaresdkdetect = eval {
|
|
require VMware::VIRuntime;
|
|
VMware::VIRuntime->import();
|
|
1;
|
|
};
|
|
}
|
|
unless ($vmwaresdkdetect) {
|
|
xCAT::SvrUtils::sendmsg([1,"VMWare SDK required for operation, but not installed"], $output_handler);
|
|
return;
|
|
}
|
|
if ($command eq 'createvcluster') {
|
|
create_new_cluster($request);
|
|
return;
|
|
}
|
|
if ($command eq 'lsvcluster') {
|
|
list_clusters($request);
|
|
return;
|
|
}
|
|
if ($command eq 'rmvcluster') {
|
|
remove_cluster($request);
|
|
return;
|
|
}
|
|
|
|
my $moreinfo;
|
|
my $noderange;
|
|
if ($request->{node}) {
|
|
$noderange = $request->{node}; # array ref
|
|
} elsif ($command eq "esxiready") {
|
|
my $node;
|
|
($node) = noderange($request->{'_xcat_clienthost'}->[0]);
|
|
$noderange = [$node];
|
|
}
|
|
xCAT::VMCommon::grab_table_data($noderange,\%tablecfg,$output_handler);
|
|
my @exargs;
|
|
unless($command){
|
|
return; # Empty request
|
|
}
|
|
if (ref($request->{arg})) {
|
|
@exargs = @{$request->{arg}};
|
|
} else {
|
|
@exargs = ($request->{arg});
|
|
}
|
|
#my $sitetab = xCAT::Table->new('site');
|
|
#if($sitetab){
|
|
#(my $ref) = $sitetab->getAttribs({key => 'usehostnamesforvcenter'}, 'value');
|
|
my @entries = xCAT::TableUtils->get_site_attribute("usehostnamesforvcenter");
|
|
my $t_entry = $entries[0];
|
|
if ( defined($t_entry) ) {
|
|
$usehostnamesforvcenter = $t_entry;
|
|
}
|
|
#($ref) = $sitetab->getAttribs({key => 'vcenterautojoin'}, 'value');
|
|
@entries = xCAT::TableUtils->get_site_attribute("vcenterautojoin");
|
|
$t_entry = $entries[0];
|
|
if ( defined($t_entry) ) {
|
|
$vcenterautojoin = $t_entry;
|
|
if ($vcenterautojoin =~ /^n/ or $vcenterautojoin =~ /^dis/) {
|
|
$vcenterautojoin=0;
|
|
}
|
|
}
|
|
#($ref) = $sitetab->getAttribs({key => 'vmwaredatastoreautomount'}, 'value');
|
|
@entries = xCAT::TableUtils->get_site_attribute("vmwaredatastoreautomount");
|
|
$t_entry = $entries[0];
|
|
if ( defined($t_entry) ) {
|
|
$datastoreautomount = $t_entry;
|
|
if ($datastoreautomount =~ /^n/ or $datastoreautomount =~ /^dis/) {
|
|
$datastoreautomount=0;
|
|
}
|
|
}
|
|
#($ref) = $sitetab->getAttribs({key => 'vmwarereconfigonpower'},'value');
|
|
@entries = xCAT::TableUtils->get_site_attribute("vmwarereconfigonpower");
|
|
$t_entry = $entries[0];
|
|
if ( defined($t_entry) ) {
|
|
$reconfigreset=$t_entry;
|
|
if ($reconfigreset =~ /^(n|d)/i) { #if no or disable, skip it
|
|
$reconfigreset=0;
|
|
}
|
|
}
|
|
|
|
# }
|
|
|
|
|
|
if ($request->{moreinfo}) { $moreinfo=$request->{moreinfo}; }
|
|
else { $moreinfo=build_more_info($noderange,$output_handler);}
|
|
foreach my $info (@$moreinfo) {
|
|
$info=~/^\[(.*?)\]\[(.*?)\]\[(.*?)\]\[(.*?)\]\[(.*?)\]\[(.*?)\]\[(.*?)\]/;
|
|
my $hyp=$1;
|
|
my @nodes=split(',', $2);
|
|
my $username = $3;
|
|
my $password = $4;
|
|
my $tmpvcname=$7;
|
|
my $tmpvcuname=$5;
|
|
my $tmpvcpass=$6;
|
|
if ($hyp =~ /^CLUSTER:/) { #a cluster, not a host.
|
|
$hyp =~ s/^CLUSTER://;
|
|
$clusterhash{$hyp}->{vcenter}->{name} = $tmpvcname;
|
|
$clusterhash{$hyp}->{vcenter}->{username} = $tmpvcuname;
|
|
$clusterhash{$hyp}->{vcenter}->{password} = $tmpvcpass;
|
|
foreach (@nodes) {
|
|
$clusterhash{$hyp}->{nodes}->{$_}=1;
|
|
}
|
|
next;
|
|
}
|
|
$hyphash{$hyp}->{vcenter}->{name} = $tmpvcname;
|
|
$hyphash{$hyp}->{vcenter}->{username} = $tmpvcuname;
|
|
$hyphash{$hyp}->{vcenter}->{password} = $tmpvcpass;
|
|
$hyphash{$hyp}->{username}=$username;# $nodeid;
|
|
$hyphash{$hyp}->{password}=$password;# $nodeid;
|
|
unless ($hyphash{$hyp}->{vcenter}->{password}) {
|
|
$hyphash{$hyp}->{vcenter}->{password} = "";
|
|
}
|
|
my $ent;
|
|
for (my $i=0; $i<@nodes; $i++){
|
|
if ($command eq 'rmigrate' and grep /-f/, @exargs) { #offline migration,
|
|
$hyphash{$hyp}->{offline} = 1; #if it is migrate and it has nodes, it is a source hypervisor apt to be offline
|
|
#this will hint to relevant code to operate under the assumption of a
|
|
#downed hypervisor source
|
|
#note this will make dangerous assumptions, it will make a very minimal attempt
|
|
#to operate normally, but really should only be called if the source is down and
|
|
#fenced (i.e. storage, network, or turned off and stateless
|
|
}
|
|
my $node = $nodes[$i];
|
|
#my $nodeid = $ids[$i];
|
|
$hyphash{$hyp}->{nodes}->{$node}=1;# $nodeid;
|
|
}
|
|
}
|
|
my $hyptab = xCAT::Table->new('hypervisor',create=>0);
|
|
if ($hyptab) {
|
|
my @hyps = keys %hyphash;
|
|
$tablecfg{hypervisor} = $hyptab->getNodesAttribs(\@hyps,['mgr','netmap','defaultnet','cluster','preferdirect','datacenter']);
|
|
}
|
|
my $hoststab = xCAT::Table->new('hosts',create=>0);
|
|
if ($hoststab) {
|
|
my @hyps = keys %hyphash;
|
|
$tablecfg{hosts} = $hoststab->getNodesAttribs(\@hyps,['hostnames']);
|
|
|
|
}
|
|
|
|
#my $children = 0;
|
|
#my $vmmaxp = 84;
|
|
#$SIG{CHLD} = sub { my $cpid; while ($cpid = waitpid(-1, WNOHANG) > 0) { delete $esx_comm_pids{$cpid}; $children--; } };
|
|
$viavcenter = 0;
|
|
if ($command eq 'rmigrate' or $command eq 'rmhypervisor') { #Only use vcenter when required, fewer prereqs
|
|
$viavcenter = 1;
|
|
}
|
|
if ($command eq 'rmhypervisor' and grep /-f/, @exargs) { #force remove of hypervisor
|
|
$vcenterforceremove=1;
|
|
}
|
|
my $keytab = xCAT::Table->new('prodkey');
|
|
if ($keytab) {
|
|
my @hypes = keys %hyphash;
|
|
$tablecfg{prodkey} = $keytab->getNodesAttribs(\@hypes,[qw/product key/]);
|
|
}
|
|
my $hyp;
|
|
my %needvcentervalidation;
|
|
my $cluster;
|
|
foreach $cluster (keys %clusterhash) {
|
|
my $vcenter = $clusterhash{$cluster}->{vcenter}->{name};
|
|
unless ($vcenterhash{$vcenter}->{conn}) {
|
|
eval {
|
|
$vcenterhash{$vcenter}->{conn} = Vim->new(service_url=>"https://$vcenter/sdk");
|
|
$vcenterhash{$vcenter}->{conn}->login(user_name => $clusterhash{$cluster}->{vcenter}->{username},
|
|
password => $clusterhash{$cluster}->{vcenter}->{password});
|
|
};
|
|
if ($@) {
|
|
$vcenterhash{$vcenter}->{conn} = undef;
|
|
xCAT::SvrUtils::sendmsg([1,"Unable to reach $vcenter vCenter server to manage cluster $cluster: $@"], $output_handler);
|
|
next;
|
|
}
|
|
my $clusternode;
|
|
}
|
|
$clusterhash{$cluster}->{conn}=$vcenterhash{$vcenter}->{conn};
|
|
foreach my $clusternode (keys %{$clusterhash{$cluster}->{nodes}}) {
|
|
$vmhash{$clusternode}->{conn}=$vcenterhash{$vcenter}->{conn};
|
|
}
|
|
}
|
|
foreach $hyp (sort(keys %hyphash)){
|
|
#if($pid == 0){
|
|
if ($viavcenter or (defined $tablecfg{hypervisor}->{$hyp}->[0]->{mgr} and not $tablecfg{hypervisor}->{$hyp}->[0]->{preferdirect})) {
|
|
$viavcenterbyhyp->{$hyp}=1;
|
|
$hypready{$hyp} = 0; #This hypervisor requires a flag be set to signify vCenter sanenes before proceeding
|
|
my $vcenter = $hyphash{$hyp}->{vcenter}->{name};
|
|
unless ($vcenterhash{$vcenter}->{conn}) {
|
|
eval {
|
|
$vcenterhash{$vcenter}->{conn} =
|
|
Vim->new(service_url=>"https://$vcenter/sdk");
|
|
$vcenterhash{$vcenter}->{conn}->login(
|
|
user_name => $hyphash{$hyp}->{vcenter}->{username},
|
|
password => $hyphash{$hyp}->{vcenter}->{password}
|
|
);
|
|
};
|
|
if ($@) {
|
|
$vcenterhash{$vcenter}->{conn} = undef;
|
|
xCAT::SvrUtils::sendmsg([1,"Unable to reach $vcenter vCenter server to manage $hyp: $@"], $output_handler);
|
|
next;
|
|
}
|
|
}
|
|
my $hypnode;
|
|
foreach $hypnode (keys %{$hyphash{$hyp}->{nodes}}) {
|
|
$vmhash{$hypnode}->{conn}=$vcenterhash{$hyphash{$hyp}->{vcenter}->{name}}->{conn};
|
|
}
|
|
$hyphash{$hyp}->{conn} = $vcenterhash{$hyphash{$hyp}->{vcenter}->{name}}->{conn};
|
|
$hyphash{$hyp}->{vcenter}->{conn} = $vcenterhash{$hyphash{$hyp}->{vcenter}->{name}}->{conn};
|
|
$needvcentervalidation{$hyp}=$vcenter;
|
|
$vcenterhash{$vcenter}->{allhyps}->{$hyp}=1;
|
|
} else {
|
|
eval {
|
|
$hyphash{$hyp}->{conn} = Vim->new(service_url=>"https://$hyp/sdk");
|
|
$hyphash{$hyp}->{conn}->login(user_name=>$hyphash{$hyp}->{username},password=>$hyphash{$hyp}->{password});
|
|
};
|
|
if ($@) {
|
|
$hyphash{$hyp}->{conn} = undef;
|
|
xCAT::SvrUtils::sendmsg([1,"Unable to reach $hyp to perform operation due to $@"], $output_handler);
|
|
$hypready{$hyp} = -1;
|
|
next;
|
|
}
|
|
my $localnode;
|
|
foreach $localnode (keys %{$hyphash{$hyp}->{nodes}}) {
|
|
$vmhash{$localnode}->{conn}=$hyphash{$hyp}->{conn};
|
|
}
|
|
validate_licenses($hyp);
|
|
}
|
|
#}else{
|
|
# $esx_comm_pids{$pid} = 1;
|
|
#}
|
|
}
|
|
foreach $hyp (keys %needvcentervalidation) {
|
|
my $vcenter = $needvcentervalidation{$hyp};
|
|
if (not defined $vcenterhash{$vcenter}->{hostviews}) {
|
|
populate_vcenter_hostviews($vcenter);
|
|
}
|
|
if (validate_vcenter_prereqs($hyp, \&declare_ready, {
|
|
hyp=>$hyp,
|
|
vcenter=>$vcenter
|
|
}) eq "failed") {
|
|
$hypready{$hyp} = -1;
|
|
}
|
|
}
|
|
while (grep { $_ == 0 } values %hypready) {
|
|
wait_for_tasks();
|
|
sleep (1); #We'll check back in every second. Unfortunately, we have to poll since we are in web service land
|
|
}
|
|
my @badhypes;
|
|
if (grep { $_ == -1 } values %hypready) {
|
|
foreach (keys %hypready) {
|
|
if ($hypready{$_} == -1) {
|
|
unless ($hyphash{$_}->{offline}) {
|
|
push @badhypes,$_;
|
|
}
|
|
my @relevant_nodes = sort (keys %{$hyphash{$_}->{nodes}});
|
|
my $sadhypervisor=$_;
|
|
foreach (@relevant_nodes) {
|
|
if ($command eq "rmigrate" and grep /-f/,@exargs) { $limbonodes{$_}=$needvcentervalidation{$sadhypervisor}; } else {
|
|
xCAT::SvrUtils::sendmsg([1,": hypervisor unreachable"], $output_handler,$_);
|
|
}
|
|
if ($command eq "rpower" and grep /stat/,@exargs) { $limbonodes{$_}=$needvcentervalidation{$sadhypervisor}; } #try to stat power anyway through vcenter of interest...
|
|
}
|
|
delete $hyphash{$_};
|
|
}
|
|
}
|
|
if (@badhypes) {
|
|
xCAT::SvrUtils::sendmsg([1,": The following hypervisors failed to become ready for the operation: ".join(',',@badhypes)], $output_handler);
|
|
}
|
|
}
|
|
do_cmd($command,@exargs);
|
|
foreach (@badhypes) { delete $hyphash{$_}; }
|
|
foreach my $vm (sort(keys %vmhash)){
|
|
$vmhash{$vm}->{conn}->logout();
|
|
}
|
|
}
|
|
|
|
sub validate_licenses {
|
|
my $hyp = shift;
|
|
my $conn = $hyphash{$hyp}->{conn}; #This can't possibly be called via a cluster stack, so hyphash is appropriate here
|
|
unless ($tablecfg{prodkey}->{$hyp}) { #if no license specified, no-op
|
|
return;
|
|
}
|
|
my $hv = get_hostview(hypname=>$hyp,conn=>$conn,properties=>['configManager','name']);
|
|
my $lm = $conn->get_view(mo_ref=>$hv->configManager->licenseManager);
|
|
my @licenses;
|
|
foreach (@{$lm->licenses}) {
|
|
push @licenses,uc($_->licenseKey);
|
|
}
|
|
my @newlicenses;
|
|
foreach (@{$tablecfg{prodkey}->{$hyp}}) {
|
|
if (defined($_->{product}) and $_->{product} eq 'esx') {
|
|
my $key = uc($_->{key});
|
|
unless (grep /$key/,@licenses) {
|
|
push @newlicenses,$key;
|
|
}
|
|
}
|
|
}
|
|
foreach (@newlicenses) {
|
|
$lm->UpdateLicense(licenseKey=>$_);
|
|
}
|
|
}
|
|
|
|
sub do_cmd {
|
|
my $command = shift;
|
|
my @exargs = @_;
|
|
if ($command eq 'esxiready') {
|
|
return;
|
|
}
|
|
if ($command eq 'rpower') {
|
|
generic_vm_operation(['config.name','config.guestId','config.hardware.memoryMB','config.hardware.numCPU','runtime.powerState','runtime.host'],\&power,@exargs);
|
|
} elsif ($command eq 'rmvm') {
|
|
generic_vm_operation(['config.name','runtime.powerState','runtime.host'],\&rmvm,@exargs);
|
|
} elsif ($command eq 'rsetboot') {
|
|
generic_vm_operation(['config.name','runtime.host'],\&setboot,@exargs);
|
|
} elsif ($command eq 'rinv') {
|
|
generic_vm_operation(['config.name','config','runtime.host','layoutEx'],\&inv,@exargs);
|
|
} elsif ($command eq 'formatdisk') {
|
|
generic_hyp_operation(\&formatdisk,@exargs);
|
|
} elsif ($command eq 'rescansan') {
|
|
generic_hyp_operation(\&rescansan,@exargs);
|
|
} elsif ($command eq 'rmhypervisor') {
|
|
generic_hyp_operation(\&rmhypervisor,@exargs);
|
|
} elsif ($command eq 'rshutdown') {
|
|
generic_hyp_operation(\&rshutdown,@exargs);
|
|
} elsif ($command eq 'chhypervisor') {
|
|
generic_hyp_operation(\&chhypervisor,@exargs);
|
|
} elsif ($command eq 'lsvm') {
|
|
generic_hyp_operation(\&lsvm,@exargs);
|
|
} elsif ($command eq 'clonevm') {
|
|
generic_hyp_operation(\&clonevms,@exargs);
|
|
} elsif ($command eq 'mkvm') {
|
|
generic_hyp_operation(\&mkvms,@exargs);
|
|
} elsif ($command eq 'chvm') {
|
|
generic_vm_operation(['config.name','config','runtime.host'],\&chvm,@exargs);
|
|
#generic_hyp_operation(\&chvm,@exargs);
|
|
} elsif ($command eq 'rmigrate') { #Technically, on a host view, but vcenter path is 'weirder'
|
|
generic_hyp_operation(\&migrate,@exargs);
|
|
}
|
|
wait_for_tasks();
|
|
if ($command eq 'clonevm') { #TODO: unconditional, remove mkvms hosted copy
|
|
my @dhcpnodes;
|
|
foreach (keys %{$tablecfg{dhcpneeded}}) {
|
|
push @dhcpnodes,$_;
|
|
delete $tablecfg{dhcpneeded}->{$_};
|
|
}
|
|
unless ($::XCATSITEVALS{'dhcpsetup'} and ($::XCATSITEVALS{'dhcpsetup'} =~ /^n/i or $::XCATSITEVALS{'dhcpsetup'} =~ /^d/i or $::XCATSITEVALS{'dhcpsetup'} eq '0')) {
|
|
$executerequest->({command=>['makedhcp'],node=>\@dhcpnodes});
|
|
}
|
|
}
|
|
}
|
|
|
|
#inventory request for esx
|
|
sub inv {
|
|
my %args = @_;
|
|
my $node = $args{node};
|
|
my $hyp = $args{hyp};
|
|
if (not defined $args{vmview}) { #attempt one refresh
|
|
$args{vmview} = $vmhash{$node}->{conn}->find_entity_view(view_type => 'VirtualMachine',properties=>['config.name','runtime.powerState'],filter=>{name=>$node});
|
|
if (not defined $args{vmview}) {
|
|
xCAT::SvrUtils::sendmsg([1,"VM does not appear to exist"], $output_handler,$node);
|
|
return;
|
|
}
|
|
}
|
|
if (not $args{vmview}->{config}) {
|
|
xCAT::SvrUtils::sendmsg([1,"VM is in an invalid state"], $output_handler,$node);
|
|
return;
|
|
}
|
|
|
|
@ARGV= @{$args{exargs}};
|
|
require Getopt::Long;
|
|
my $tableUpdate;
|
|
my $rc = GetOptions(
|
|
't' => \$tableUpdate,
|
|
);
|
|
$SIG{__WARN__} = 'DEFAULT';
|
|
|
|
if(@ARGV > 1) {
|
|
xCAT::SvrUtils::sendmsg("Invalid arguments: @ARGV", $output_handler);
|
|
return;
|
|
}
|
|
|
|
if(!$rc) {
|
|
return;
|
|
}
|
|
|
|
my $vmview = $args{vmview};
|
|
my $moref = $vmview->{mo_ref}->value;
|
|
xCAT::SvrUtils::sendmsg("Managed Object Reference: $moref", $output_handler,$node);
|
|
my $uuid = $vmview->config->uuid;
|
|
$uuid =~ s/(..)(..)(..)(..)-(..)(..)-(..)(..)/$4$3$2$1-$6$5-$8$7/;
|
|
xCAT::SvrUtils::sendmsg("UUID/GUID: $uuid", $output_handler,$node);
|
|
my $cpuCount = $vmview->config->hardware->numCPU;
|
|
xCAT::SvrUtils::sendmsg("CPUs: $cpuCount", $output_handler,$node);
|
|
my $memory = $vmview->config->hardware->memoryMB;
|
|
xCAT::SvrUtils::sendmsg("Memory: $memory MB", $output_handler,$node);
|
|
my %updatehash = ( cpus => $cpuCount, memory=>$memory);
|
|
|
|
|
|
my $devices = $vmview->config->hardware->device;
|
|
my $label;
|
|
my $size;
|
|
my $fileName;
|
|
my $device;
|
|
if ($tableUpdate and $hyp) {
|
|
validate_datastore_prereqs([$node],$hyp); #need datastoremaps to verify names...
|
|
}
|
|
my %vmstorageurls;
|
|
foreach $device (@$devices) {
|
|
$label = $device->deviceInfo->label;
|
|
|
|
if($label =~ /^Hard disk/) {
|
|
$label .= " (d".$device->controllerKey.":".$device->unitNumber.")";
|
|
$size = $device->capacityInKB / 1024;
|
|
$fileName = $device->backing->fileName;
|
|
$output_handler->({
|
|
node=>{
|
|
name=>$node,
|
|
data=>{
|
|
desc=>$label,
|
|
contents=>"$size MB @ $fileName"
|
|
}
|
|
}
|
|
});
|
|
#if ($tableUpdate) {
|
|
# $fileName =~ /\[([^\]]+)\]/;
|
|
# $vmstorageurls{$hyphash{$hyp}->{datastoreurlmap}->{$1}}=1;
|
|
#}
|
|
} elsif ($label =~ /Network/) {
|
|
xCAT::SvrUtils::sendmsg("$label: ".$device->macAddress, $output_handler,$node);
|
|
}
|
|
}
|
|
if ($tableUpdate) {
|
|
my $cfgdatastore;
|
|
foreach (@{$vmview->layoutEx->file}) {
|
|
#TODO, track ALL layoutEx->file....
|
|
if ($_->type eq 'config') {
|
|
$_->name =~ /\[([^\]]+)\]/;
|
|
$cfgdatastore = $hyphash{$hyp}->{datastoreurlmap}->{$1};
|
|
last;
|
|
}
|
|
}
|
|
my $cfgkey;
|
|
if ($tablecfg{vm}->{$node}->[0]->{cfgstore}) { #check the config file explicitly, ignore the rest
|
|
$cfgkey='cfgstore';
|
|
} elsif ($tablecfg{vm}->{$node}->[0]->{storage}) { #check the config file explicitly, ignore the rest
|
|
$cfgkey='storage';
|
|
}
|
|
my $configuration = $tablecfg{vm}->{$node}->[0]->{$cfgkey}; #TODO: prune urls that map to no layoutEx->file entries anymore
|
|
my $configappend = $configuration;
|
|
$configappend =~ s/^[^,=]*//;
|
|
$tablecfg{vm}->{$node}->[0]->{$cfgkey} =~ m!nfs://([^/]+)/!;
|
|
my $tablecfgserver =$1;
|
|
my $cfgserver = inet_aton($tablecfgserver);
|
|
if ($cfgserver) {
|
|
$cfgserver = inet_ntoa($cfgserver); #get the IP address (TODO: really need to wrap getaddrinfo this handily...
|
|
my $cfgurl = $tablecfg{vm}->{$node}->[0]->{$cfgkey};
|
|
$cfgurl =~ s/$tablecfgserver/$cfgserver/;
|
|
if ($cfgurl ne $cfgdatastore) {
|
|
$updatehash{$cfgkey} = $cfgdatastore.$configappend;
|
|
}
|
|
}
|
|
}
|
|
if($tableUpdate){
|
|
my $vm=xCAT::Table->new('vm',-create=>1);
|
|
$vm->setNodeAttribs($node,\%updatehash);
|
|
}
|
|
|
|
}
|
|
|
|
|
|
#changes the memory, number of cpus and device size
|
|
#can also add,resize and remove disks
|
|
sub chvm {
|
|
my %args = @_;
|
|
my $node = $args{node};
|
|
my $hyp = $args{hyp};
|
|
if (not defined $args{vmview}) { #attempt one refresh
|
|
$args{vmview} = $vmhash{$node}->{conn}->find_entity_view(view_type => 'VirtualMachine',
|
|
properties=>['config.name','runtime.powerState'],
|
|
filter=>{name=>$node});
|
|
if (not defined $args{vmview}) {
|
|
xCAT::SvrUtils::sendmsg([1,"VM does not appear to exist"], $output_handler,$node);
|
|
return;
|
|
}
|
|
}
|
|
@ARGV= @{$args{exargs}};
|
|
my @deregister;
|
|
my @purge;
|
|
my @add;
|
|
my %resize;
|
|
my $cpuCount;
|
|
my $memory;
|
|
my $vmview = $args{vmview};
|
|
|
|
require Getopt::Long;
|
|
$SIG{__WARN__} = sub {
|
|
xCAT::SvrUtils::sendmsg([1,"Could not parse options, ".shift()], $output_handler);
|
|
};
|
|
my @otherparams;
|
|
my $cdrom;
|
|
my $eject;
|
|
my $rc = GetOptions(
|
|
"d=s" => \@deregister,
|
|
"p=s" => \@purge,
|
|
"a=s" => \@add,
|
|
"o=s" => \@otherparams,
|
|
"resize=s%" => \%resize,
|
|
"optical|cdrom|c=s" => \$cdrom,
|
|
"eject" => \$eject,
|
|
"cpus=s" => \$cpuCount,
|
|
"mem=s" => \$memory
|
|
);
|
|
$SIG{__WARN__} = 'DEFAULT';
|
|
|
|
if(@ARGV) {
|
|
xCAT::SvrUtils::sendmsg("Invalid arguments: @ARGV", $output_handler);
|
|
return;
|
|
}
|
|
|
|
if(!$rc) {
|
|
return;
|
|
}
|
|
|
|
#use Data::Dumper;
|
|
#xCAT::SvrUtils::sendmsg("dereg = ".Dumper(\@deregister));
|
|
#xCAT::SvrUtils::sendmsg("purge = ".Dumper(\@purge));
|
|
#xCAT::SvrUtils::sendmsg("add = ".Dumper(\@add));
|
|
#xCAT::SvrUtils::sendmsg("resize = ".Dumper(\%resize));
|
|
#xCAT::SvrUtils::sendmsg("cpus = $cpuCount");
|
|
#xCAT::SvrUtils::sendmsg("mem = ".getUnits($memory,"K",1024));
|
|
|
|
|
|
my %conargs;
|
|
if($cpuCount) {
|
|
if ($cpuCount =~ /^\+(\d+)/) {
|
|
$cpuCount = $vmview->config->hardware->numCPU+$1;
|
|
} elsif ($cpuCount =~ /^-(\d+)/) {
|
|
$cpuCount = $vmview->config->hardware->numCPU-$1;
|
|
}
|
|
$conargs{numCPUs} = $cpuCount;
|
|
}
|
|
|
|
if($memory) {
|
|
if ($memory =~ /^\+(.+)/) {
|
|
$conargs{memoryMB} = $vmview->config->hardware->memoryMB + getUnits($1,"M",1048576);
|
|
} elsif ($memory =~ /^-(\d+)/) {
|
|
$conargs{memoryMB} = $vmview->config->hardware->memoryMB - getUnits($1,"M",1048576);
|
|
} else {
|
|
$conargs{memoryMB} = getUnits($memory, "M", 1048576);
|
|
}
|
|
}
|
|
|
|
my $disk;
|
|
my $devices = $vmview->config->hardware->device;
|
|
my $label;
|
|
my $device;
|
|
my $cmdLabel;
|
|
my $newSize;
|
|
my @devChanges;
|
|
|
|
if(@deregister) {
|
|
for $disk (@deregister) {
|
|
$device = getDiskByLabel($disk, $devices);
|
|
unless($device) {
|
|
xCAT::SvrUtils::sendmsg([1,"Disk: $disk does not exist"], $output_handler,$node);
|
|
return;
|
|
}
|
|
#xCAT::SvrUtils::sendmsg(Dumper($device));
|
|
push @devChanges, VirtualDeviceConfigSpec->new(
|
|
device => $device,
|
|
operation => VirtualDeviceConfigSpecOperation->new('remove'));
|
|
|
|
}
|
|
}
|
|
|
|
if(@purge) {
|
|
for $disk (@purge) {
|
|
$device = getDiskByLabel($disk, $devices);
|
|
unless($device) {
|
|
xCAT::SvrUtils::sendmsg([1,"Disk: $disk does not exist"], $output_handler,$node);
|
|
return;
|
|
}
|
|
#xCAT::SvrUtils::sendmsg(Dumper($device));
|
|
push @devChanges, VirtualDeviceConfigSpec->new(
|
|
device => $device,
|
|
operation => VirtualDeviceConfigSpecOperation->new('remove'),
|
|
fileOperation => VirtualDeviceConfigSpecFileOperation->new('destroy'));
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if(@add) {
|
|
my $addSizes = join(',',@add);
|
|
my $scsiCont;
|
|
my $scsiUnit;
|
|
my $ideCont;
|
|
my $ideUnit;
|
|
my $label;
|
|
my $idefull=0;
|
|
my $scsifull=0;
|
|
foreach $device (@$devices) {
|
|
$label = $device->deviceInfo->label;
|
|
if($label =~ /^SCSI controller/) {
|
|
my $tmpu=getAvailUnit($device->{key},$devices,maxnum=>15);
|
|
if ($tmpu > 0) {
|
|
$scsiCont = $device;
|
|
$scsiUnit=$tmpu;
|
|
} else {
|
|
$scsifull=1;
|
|
}
|
|
#ignore scsiControllers that are full, problem still remains if trying to add across two controllers in one go
|
|
}
|
|
if($label =~ /^IDE/ and not $ideCont) {
|
|
my $tmpu=getAvailUnit($device->{key},$devices,maxnum=>1);
|
|
if ($tmpu >= 0) {
|
|
$ideCont = $device;
|
|
$ideUnit = $tmpu;
|
|
} elsif ($device->{key} == 201) {
|
|
$idefull=1;
|
|
}
|
|
}
|
|
}
|
|
unless ($hyphash{$hyp}->{datastoremap}) { validate_datastore_prereqs([],$hyp); }
|
|
push @devChanges, create_storage_devs($node,$hyphash{$hyp}->{datastoremap},$addSizes,$scsiCont,$scsiUnit,$ideCont,$ideUnit,$devices,idefull=>$idefull,scsifull=>$scsifull);
|
|
}
|
|
|
|
if ($cdrom or $eject) {
|
|
my $opticalbackingif;
|
|
my $opticalconnectable;
|
|
if ($cdrom) {
|
|
my $storageurl;
|
|
if ($cdrom =~ m!://!) {
|
|
$storageurl=$cdrom;
|
|
$storageurl =~ s!/[^/]*\z!!;
|
|
unless (validate_datastore_prereqs([],$hyp,{$storageurl=>[$node]})) {
|
|
xCAT::SvrUtils::sendmsg([1,"Unable to find/mount datastore holding $cdrom"], $output_handler,$node);
|
|
return;
|
|
}
|
|
$cdrom =~ s!.*/!!;
|
|
} else {
|
|
$storageurl = $tablecfg{vm}->{$node}->[0]->{storage};
|
|
$storageurl =~ s/=.*//;
|
|
$storageurl =~ s/.*,//;
|
|
$storageurl =~ s/\/\z//;
|
|
}
|
|
$opticalbackingif = VirtualCdromIsoBackingInfo->new( fileName => "[".$hyphash{$hyp}->{datastoremap}->{$storageurl}."] $cdrom");
|
|
$opticalconnectable = VirtualDeviceConnectInfo->new(startConnected=>1,allowGuestControl=>1,connected=>1);
|
|
} elsif ($eject) {
|
|
$opticalbackingif=VirtualCdromRemoteAtapiBackingInfo->new(deviceName=>"");
|
|
$opticalconnectable=VirtualDeviceConnectInfo->new(startConnected=>0,allowGuestControl=>1,connected=>0);
|
|
}
|
|
my $oldcd;
|
|
foreach my $dev (@$devices) {
|
|
if ($dev->deviceInfo->label eq "CD/DVD drive 1") {
|
|
$oldcd=$dev;
|
|
last;
|
|
}
|
|
}
|
|
unless ($oldcd) {
|
|
if ($cdrom) {
|
|
xCAT::SvrUtils::sendmsg([1,"Unable to find Optical drive in VM to insert ISO image"], $output_handler,$node);
|
|
} else {
|
|
xCAT::SvrUtils::sendmsg([1,"Unable to find Optical drive in VM to perform eject"], $output_handler,$node);
|
|
}
|
|
return;
|
|
}
|
|
my $newDevice = VirtualCdrom->new(backing => $opticalbackingif,
|
|
key=>$oldcd->key,
|
|
controllerKey=>201,
|
|
unitNumber=>0,
|
|
connectable=>$opticalconnectable,
|
|
);
|
|
push @devChanges, VirtualDeviceConfigSpec->new(
|
|
device => $newDevice,
|
|
operation => VirtualDeviceConfigSpecOperation->new('edit'));
|
|
|
|
}
|
|
if(%resize) {
|
|
while( my ($key, $value) = each(%resize) ) {
|
|
my @drives = split(/,/, $key);
|
|
for my $device ( @drives ) {
|
|
my $disk = $device;
|
|
$device = getDiskByLabel($disk, $devices);
|
|
unless($device) {
|
|
xCAT::SvrUtils::sendmsg([1,"Disk: $disk does not exist"], $output_handler,$node);
|
|
return;
|
|
}
|
|
if ($value =~ /^\+(.+)/) {
|
|
$value = $device->capacityInKB + getUnits($1,"G",1024);
|
|
} else {
|
|
$value = getUnits($value, "G", 1024);
|
|
}
|
|
my $newDevice = VirtualDisk->new(deviceInfo => $device->deviceInfo,
|
|
key => $device->key,
|
|
controllerKey => $device->controllerKey,
|
|
unitNumber => $device->unitNumber,
|
|
deviceInfo => $device->deviceInfo,
|
|
backing => $device->backing,
|
|
capacityInKB => $value);
|
|
push @devChanges, VirtualDeviceConfigSpec->new(
|
|
device => $newDevice,
|
|
operation => VirtualDeviceConfigSpecOperation->new('edit'));
|
|
}
|
|
}
|
|
|
|
}
|
|
if(@devChanges) {
|
|
$conargs{deviceChange} = \@devChanges;
|
|
}
|
|
if (@otherparams) {
|
|
my $key;
|
|
my $value;
|
|
my @optionvals;
|
|
foreach (@otherparams) {
|
|
($key,$value) = split /=/;
|
|
unless ($key) {
|
|
xCAT::SvrUtils::sendmsg([1,"Invalid format for other parameter specification"], $output_handler,$node);
|
|
return;
|
|
}
|
|
if ($value) {
|
|
push @optionvals,OptionValue->new(key=>$key,value=>$value);
|
|
} else {
|
|
push @optionvals,OptionValue->new(key=>$key); #the api doc says this is *supposed* to delete a key, don't think it works though, e.g. http://communities.vmware.com/message/1602644
|
|
}
|
|
}
|
|
$conargs{extraConfig} = \@optionvals;
|
|
}
|
|
|
|
my $reconfigspec = VirtualMachineConfigSpec->new(%conargs);
|
|
|
|
#xCAT::SvrUtils::sendmsg("reconfigspec = ".Dumper($reconfigspec));
|
|
my $task = $vmview->ReconfigVM_Task(spec=>$reconfigspec);
|
|
$running_tasks{$task}->{task} = $task;
|
|
$running_tasks{$task}->{callback} = \&chvm_task_callback;
|
|
$running_tasks{$task}->{hyp} = $hyp;
|
|
$running_tasks{$task}->{data} = { node => $node, successtext => "node successfully changed",cpus=>$cpuCount,mem=>$memory };
|
|
|
|
}
|
|
|
|
sub getUsedUnits {
|
|
my $contKey = shift;
|
|
my $devices = shift;
|
|
my %usedids;
|
|
$usedids{7}=1;
|
|
$usedids{'7'}=1; #TODO: figure out which of these is redundant, the string or the number variant
|
|
for my $device (@$devices) {
|
|
if($device->{controllerKey} eq $contKey) {
|
|
$usedids{$device->{unitNumber}}=1;
|
|
}
|
|
}
|
|
return \%usedids;
|
|
}
|
|
sub getAvailUnit {
|
|
my $contKey = shift;
|
|
my $devices = shift;
|
|
my %args = @_;
|
|
my $maxunit=-1;
|
|
if (defined $args{maxnum}) {
|
|
$maxunit=$args{maxnum};
|
|
}
|
|
my %usedids;
|
|
$usedids{7}=1;
|
|
$usedids{'7'}=1; #TODO: figure out which of these is redundant, the string or the number variant
|
|
for my $device (@$devices) {
|
|
if($device->{controllerKey} eq $contKey) {
|
|
$usedids{$device->{unitNumber}}=1;
|
|
}
|
|
}
|
|
my $highestUnit=0;
|
|
while ($usedids{$highestUnit}) {
|
|
if ($highestUnit == $maxunit) {
|
|
return -1;
|
|
}
|
|
$highestUnit++;
|
|
}
|
|
return $highestUnit;
|
|
}
|
|
|
|
#given a device list from a vm and a label for a hard disk, returns the device object
|
|
sub getDiskByLabel {
|
|
my $cmdLabel = shift;
|
|
my $devices = shift;
|
|
my $device;
|
|
my $label;
|
|
|
|
$cmdLabel = commandLabel($cmdLabel);
|
|
foreach $device (@$devices) {
|
|
$label = $device->deviceInfo->label;
|
|
|
|
if($cmdLabel eq $label) {
|
|
return $device;
|
|
} elsif (($label =~ /^Hard disk/) and ($cmdLabel =~ /^d(.*)/)) {
|
|
my $desc = $1;
|
|
if ($desc =~ /(.*):(.*)/) {#specific
|
|
my $controller=$1;
|
|
my $unit=$2;
|
|
if ($device->unitNumber == $unit and $device->controllerKey == $controller) {
|
|
return $device;
|
|
}
|
|
} elsif ($desc =~ /\d+/ and $device->unitNumber == $desc) { #not specific
|
|
return $device;
|
|
}
|
|
}
|
|
|
|
}
|
|
return undef;
|
|
}
|
|
|
|
#takes a label for a hard disk and prepends "Hard disk " if it's not there already
|
|
sub commandLabel {
|
|
my $label = shift;
|
|
if(($label =~ /^Hard disk/) or ($label =~ /^d\d+/)) {
|
|
return $label;
|
|
}
|
|
return "Hard disk ".$label;
|
|
}
|
|
|
|
#this function will check pending task status
|
|
sub process_tasks {
|
|
foreach (keys %running_tasks) {
|
|
my $curcon;
|
|
if (defined $running_tasks{$_}->{conn}) {
|
|
$curcon = $running_tasks{$_}->{conn};
|
|
} elsif ($running_tasks{$_}->{hyp}) {
|
|
$curcon = $hyphash{$running_tasks{$_}->{hyp}}->{conn};
|
|
} elsif ($running_tasks{$_}->{vm}) {
|
|
$curcon = $vmhash{$running_tasks{$_}->{vm}}->{conn};
|
|
} elsif ($running_tasks{$_}->{cluster}) {
|
|
$curcon = $clusterhash{$running_tasks{$_}->{cluster}}->{conn};
|
|
} else {
|
|
use Carp qw/confess/;
|
|
confess "This stack trace indicates a cluster unfriendly path";
|
|
}
|
|
my $curt = $curcon->get_view(mo_ref=>$running_tasks{$_}->{task});
|
|
my $state = $curt->info->state->val;
|
|
unless ($state eq 'running' or $state eq 'queued') {
|
|
$running_tasks{$_}->{callback}->($curt,$running_tasks{$_}->{data});
|
|
delete $running_tasks{$_};
|
|
}
|
|
if ($state eq 'running' and not $running_tasks{$_}->{questionasked}) { # and $curt->info->progress == 95) { #This is unfortunate, there should be a 'state' to indicate a question is blocking
|
|
#however there isn't, so if we see something running at 95%, we just manually see if a question blocked the rest
|
|
my $vm;
|
|
$@="";
|
|
eval {
|
|
$vm = $curcon->get_view(mo_ref=>$curt->info->entity);
|
|
};
|
|
if ($@) { $vm = 0; }
|
|
if ($vm and $vm->{summary} and $vm->summary->{runtime} and $vm->summary->runtime->{question} and $vm->summary->runtime->question) {
|
|
$running_tasks{$_}->{questionasked}=1;
|
|
$running_tasks{$_}->{callback}->($curt,$running_tasks{$_}->{data},$vm->summary->runtime->question,$vm);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
#this function is a barrier to ensure prerequisites are met
|
|
sub wait_for_tasks {
|
|
while (scalar keys %running_tasks) {
|
|
process_tasks;
|
|
sleep (1); #We'll check back in every second. Unfortunately, we have to poll since we are in web service land
|
|
}
|
|
}
|
|
|
|
sub connecthost_callback {
|
|
my $task = shift;
|
|
my $args = shift;
|
|
my $hv = $args->{hostview};
|
|
my $state = $task->info->state->val;
|
|
if ($state eq "success") {
|
|
$hypready{$args->{hypname}}=1; #declare readiness
|
|
enable_vmotion(hypname=>$args->{hypname},hostview=>$args->{hostview},conn=>$args->{conn});
|
|
$vcenterhash{$args->{vcenter}}->{goodhyps}->{$args->{hypname}} = 1;
|
|
if (defined $args->{depfun}) { #If a function is waiting for the host connect to go valid, call it
|
|
$args->{depfun}->($args->{depargs});
|
|
}
|
|
return;
|
|
}
|
|
my $thumbprint;
|
|
eval {
|
|
$thumbprint = $task->{info}->error->fault->thumbprint;
|
|
};
|
|
if ($thumbprint) { #was an unknown certificate error, retry and accept the unknown certificate
|
|
$args->{connspec}->{sslThumbprint}=$task->info->error->fault->thumbprint;
|
|
my $task;
|
|
if (defined $args->{hostview}) {#It was a reconnect request
|
|
$task = $hv->ReconnectHost_Task(cnxSpec=>$args->{connspec});
|
|
} elsif (defined $args->{foldview}) {#was an add host request
|
|
$task = $args->{foldview}->AddStandaloneHost_Task(spec=>$args->{connspec},addConnected=>1);
|
|
} elsif (defined $args->{cluster}) {#was an add host to cluster request
|
|
$task = $args->{cluster}->AddHost_Task(spec=>$args->{connspec},asConnected=>1);
|
|
}
|
|
$running_tasks{$task}->{task} = $task;
|
|
$running_tasks{$task}->{callback} = \&connecthost_callback;
|
|
$running_tasks{$task}->{conn} = $args->{conn};
|
|
$running_tasks{$task}->{data} = $args; #{ conn_spec=>$connspec,hostview=>$hv,hypname=>$args->{hypname},vcenter=>$args->{vcenter} };
|
|
} elsif ($state eq 'error') {
|
|
my $error = $task->info->error->localizedMessage;
|
|
if (defined ($task->info->error->fault->faultMessage)) { #Only in 4.0, support of 3.5 must be careful?
|
|
foreach(@{$task->info->error->fault->faultMessage}) {
|
|
$error.=$_->message;
|
|
}
|
|
}
|
|
xCAT::SvrUtils::sendmsg([1,$error], $output_handler); #,$node);
|
|
$hypready{$args->{hypname}} = -1; #Impossible for this hypervisor to ever be ready
|
|
$vcenterhash{$args->{vcenter}}->{badhyps}->{$args->{hypname}} = 1;
|
|
}
|
|
}
|
|
|
|
sub delhost_callback { #only called in rmhypervisor -f case during validate vcenter phase
|
|
my $task = shift;
|
|
my $args = shift;
|
|
my $hv = $args->{hostview};
|
|
my $state = $task->info->state->val;
|
|
if ($state eq "success") {
|
|
xCAT::SvrUtils::sendmsg("removed", $output_handler,$args->{hypname});
|
|
$hypready{$args->{hypname}} = -1; #Impossible for this hypervisor to ever be ready
|
|
$vcenterhash{$args->{vcenter}}->{badhyps}->{$args->{hypname}} = 1;
|
|
} elsif ($state eq 'error') {
|
|
my $error = $task->info->error->localizedMessage;
|
|
if (defined ($task->info->error->fault->faultMessage)) { #Only in 4.0, support of 3.5 must be careful?
|
|
foreach(@{$task->info->error->fault->faultMessage}) {
|
|
$error.=$_->message;
|
|
}
|
|
}
|
|
xCAT::SvrUtils::sendmsg([1,$error], $output_handler); #,$node);
|
|
$hypready{$args->{hypname}} = -1; #Impossible for this hypervisor to ever be ready
|
|
$vcenterhash{$args->{vcenter}}->{badhyps}->{$args->{hypname}} = 1;
|
|
}
|
|
}
|
|
|
|
sub get_clusterview {
|
|
my %args = @_;
|
|
my $clustname = $args{clustname};
|
|
my %subargs = (
|
|
view_type=>'ClusterComputeResource',
|
|
);
|
|
if ($args{properties}) {
|
|
$subargs{properties}=$args{properties};
|
|
}
|
|
$subargs{filter}={name=>$clustname};
|
|
my $view = $args{conn}->find_entity_view(%subargs);
|
|
return $view;
|
|
#foreach (@{$args{conn}->find_entity_views(%subargs)}) {
|
|
# if ($_->name eq "$clustname") {
|
|
# return $_;
|
|
# last;
|
|
# }
|
|
#}
|
|
}
|
|
|
|
sub get_hostview {
|
|
my %args = @_;
|
|
my $host = $args{hypname};
|
|
my %subargs = (
|
|
view_type=>'HostSystem',
|
|
);
|
|
if ($args{properties}) {
|
|
$subargs{properties}=$args{properties};
|
|
}
|
|
my @addrs = gethostbyname($host);
|
|
my $ip;
|
|
my $name;
|
|
my $aliases;
|
|
if ($addrs[4]) {
|
|
$ip=inet_ntoa($addrs[4]);
|
|
($name, $aliases) = gethostbyaddr($addrs[4],AF_INET); #TODO: IPv6
|
|
} else {
|
|
($ip,$name,$aliases) = ($host,$host,"");
|
|
}
|
|
my @matchvalues = ($host,$ip,$name);
|
|
foreach (split /\s+/,$aliases) {
|
|
push @matchvalues,$_;
|
|
}
|
|
my $view;
|
|
$subargs{filter}={'name' => qr/$host(?:\.|\z)/};
|
|
$view = $args{conn}->find_entity_view(%subargs);
|
|
if ($view) { return $view; }
|
|
foreach (@matchvalues) {
|
|
$subargs{filter}={'name' => qr/$_(?:\.|\z)/};
|
|
$view = $args{conn}->find_entity_view(%subargs);
|
|
if ($view) { return $view; }
|
|
}
|
|
$subargs{filter}={'name' => qr/localhost(?:\.|\z)/};
|
|
$view = $args{conn}->find_entity_view(%subargs);
|
|
if ($view) { return $view; }
|
|
return undef; #rest of function should be obsoleted, going to run with that assumption for 2.5 at least
|
|
# $subargs{filter}={'name' =~ qr/.*/};
|
|
# foreach (@{$args{conn}->find_entity_views(%subargs)}) {
|
|
# my $view = $_;
|
|
# if ($_->name =~ /$host(?:\.|\z)/ or $_->name =~ /localhost(?:\.|\z)/ or grep { $view->name =~ /$_(?:\.|\z)/ } @matchvalues) {
|
|
# return $view;
|
|
# last;
|
|
# }
|
|
# }
|
|
}
|
|
sub enable_vmotion {
|
|
#TODO: vmware 3.x semantics too? this is 4.0...
|
|
my %args = @_;
|
|
unless ($args{hostview}) {
|
|
$args{hostview} = get_hostview(conn=>$args{conn},hypname=>$args{hypname},properties=>['configManager','name']);
|
|
}
|
|
my $nicmgr=$args{conn}->get_view(mo_ref=>$args{hostview}->configManager->virtualNicManager);
|
|
my $qnc = $nicmgr->QueryNetConfig(nicType=>"vmotion");
|
|
if ($qnc->{selectedVnic}) {
|
|
return 1;
|
|
} else {
|
|
my $vniccount=scalar @{$qnc->candidateVnic};
|
|
if ($vniccount==1 or ($vniccount==2 and $qnc->candidateVnic->[1]->spec->ip->ipAddress =~ /^169.254/)) { #There is only one possible path, use it
|
|
$nicmgr->SelectVnicForNicType(nicType=>"vmotion",device=>$qnc->candidateVnic->[0]->device);
|
|
return 1;
|
|
} else {
|
|
xCAT::SvrUtils::sendmsg([1,"TODO: use configuration to pick the nic ".$args{hypname}], $output_handler);
|
|
}
|
|
return 0;
|
|
}
|
|
}
|
|
sub mkvm_callback {
|
|
my $task = shift;
|
|
my $args = shift;
|
|
my $node = $args->{node};
|
|
my $hyp = $args->{hyp};
|
|
if ($task->info->state->val eq 'error') {
|
|
my $error = $task->info->error->localizedMessage;
|
|
xCAT::SvrUtils::sendmsg([1,$error], $output_handler,$node);
|
|
}
|
|
}
|
|
|
|
sub relay_vmware_err {
|
|
my $task = shift;
|
|
my $extratext = shift;
|
|
my @nodes = @_;
|
|
my $error = $task->info->error->localizedMessage;
|
|
if (defined ($task->info->error->fault->faultMessage)) { #Only in 4.0, support of 3.5 must be careful?
|
|
foreach(@{$task->info->error->fault->faultMessage}) {
|
|
$error.=$_->message;
|
|
}
|
|
}
|
|
if (@nodes) {
|
|
foreach (@nodes) {
|
|
xCAT::SvrUtils::sendmsg([1,$extratext.$error], $output_handler,$_);
|
|
}
|
|
}else {
|
|
xCAT::SvrUtils::sendmsg([1,$extratext.$error], $output_handler);
|
|
}
|
|
}
|
|
|
|
sub relocate_callback {
|
|
my $task = shift;
|
|
my $parms = shift;
|
|
my $state = $task->info->state->val;
|
|
if ($state eq 'success') {
|
|
my $vmtab = xCAT::Table->new('vm'); #TODO: update vm.storage?
|
|
my $prevloc = $tablecfg{vm}->{$parms->{node}}->[0]->{storage};
|
|
my $model;
|
|
($prevloc,$model) = split /=/,$prevloc;
|
|
my $target = $parms->{target};
|
|
if ($model) {
|
|
$target.="=$model";
|
|
}
|
|
$vmtab->setNodeAttribs($parms->{node},{storage=>$target});
|
|
xCAT::SvrUtils::sendmsg(":relocated to to ".$parms->{target}, $output_handler,$parms->{node});
|
|
} else {
|
|
relay_vmware_err($task,"Relocating to ".$parms->{target}." ",$parms->{node});
|
|
}
|
|
}
|
|
sub migrate_ok { #look like a successful migrate, callback for registering a vm
|
|
my %args = @_;
|
|
my $vmtab = xCAT::Table->new('vm');
|
|
$vmtab->setNodeAttribs($args{nodes}->[0],{host=>$args{target}});
|
|
xCAT::SvrUtils::sendmsg("migrated to ".$args{target}, $output_handler,$args{nodes}->[0]);
|
|
}
|
|
sub migrate_callback {
|
|
my $task = shift;
|
|
my $parms = shift;
|
|
my $state = $task->info->state->val;
|
|
if (not $parms->{skiptodeadsource} and $state eq 'success') {
|
|
my $vmtab = xCAT::Table->new('vm');
|
|
$vmtab->setNodeAttribs($parms->{node},{host=>$parms->{target}});
|
|
xCAT::SvrUtils::sendmsg("migrated to ".$parms->{target}, $output_handler,$parms->{node});
|
|
} elsif($parms->{offline}) { #try a forceful RegisterVM instead
|
|
my $target = $parms->{target};
|
|
my $hostview = $hyphash{$target}->{conn}->find_entity_view(view_type=>'VirtualMachine',properties=>['config.name'],filter=>{name=>$parms->{node}});
|
|
if ($hostview) { #this means vcenter still has it in inventory, but on a dead node...
|
|
#unfortunately, vcenter won't give up the old one until we zap the dead hypervisor
|
|
#also unfortunately, it doesn't make it easy to find said hypervisor..
|
|
$hostview = $hyphash{$parms->{src}}->{conn}->get_view(mo_ref=>$hyphash{$parms->{src}}->{deletionref});
|
|
$task = $hostview->Destroy_Task();
|
|
$running_tasks{$task}->{task} = $task;
|
|
$running_tasks{$task}->{callback} = \&migrate_callback;
|
|
$running_tasks{$task}->{conn} = $hyphash{$target}->{vcenter}->{conn};
|
|
$running_tasks{$task}->{data} = { offline=>1, src=>$parms->{src}, node=>$parms->{node}, target=>$target, skiptodeadsource=>1 };
|
|
} else { #it is completely gone, attempt a register_vm strategy
|
|
register_vm($target,$parms->{node},undef,\&migrate_ok,{ nodes => [$parms->{node}], target=>$target, },"failonerror");
|
|
}
|
|
} else {
|
|
relay_vmware_err($task,"Migrating to ".$parms->{target}." ",$parms->{node});
|
|
}
|
|
}
|
|
|
|
sub poweron_task_callback {
|
|
my $task = shift;
|
|
my $parms = shift;
|
|
my $q = shift; #question if blocked
|
|
my $vm = shift; #path to answer questions if asked
|
|
my $state = $task->info->state->val;
|
|
my $node = $parms->{node};
|
|
my $intent = $parms->{successtext};
|
|
if ($state eq 'success') {
|
|
xCAT::SvrUtils::sendmsg($intent, $output_handler,$node);
|
|
} elsif ($state eq 'error') {
|
|
relay_vmware_err($task,"",$node);
|
|
} elsif ($q and $q->text =~ /^msg.uuid.altered:/ and ($q->choice->choiceInfo->[0]->summary eq 'Cancel' and ($q->choice->choiceInfo->[0]->key eq '0'))) { #make sure it is what is what we have seen it to be
|
|
if ($parms->{forceon} and $q->choice->choiceInfo->[1]->summary =~ /I (_)?moved it/ and $q->choice->choiceInfo->[1]->key eq '1') { #answer the question as 'moved'
|
|
$vm->AnswerVM(questionId=>$q->id,answerChoice=>'1');
|
|
} else {
|
|
$vm->AnswerVM(questionId=>$q->id,answerChoice=>'0');
|
|
xCAT::SvrUtils::sendmsg([1,"Failure powering on VM, it mismatched against the hypervisor. If positive VM is not running on another hypervisor, use -f to force VM on"], $output_handler,$node);
|
|
}
|
|
} elsif ($q) {
|
|
if ($q->choice->choiceInfo->[0]->summary eq 'Cancel') {
|
|
xCAT::SvrUtils::sendmsg([1,":Cancelling due to unexpected question executing task: ".$q->text], $output_handler,$node);
|
|
} else {
|
|
xCAT::SvrUtils::sendmsg([1,":Task hang due to unexpected question executing task, need to use VMware tools to clean up the mess for now: ".$q->text], $output_handler,$node);
|
|
}
|
|
}
|
|
|
|
}
|
|
sub chvm_task_callback {
|
|
my $task = shift;
|
|
my $parms = shift;
|
|
my $state = $task->info->state->val;
|
|
my $node = $parms->{node};
|
|
my $intent = $parms->{successtext};
|
|
if ($state eq 'success') {
|
|
my $updatehash;
|
|
if ($parms->{cpus} and $tablecfg{vm}->{$node}->[0]->{cpus}) { #need to update
|
|
$updatehash->{cpus}=$parms->{cpus};
|
|
}
|
|
if ($parms->{mem} and $tablecfg{vm}->{$node}->[0]->{memory}) { #need to update
|
|
$updatehash->{memory}=$parms->{mem};
|
|
}
|
|
if ($updatehash) {
|
|
my $vmtab = xCAT::Table->new('vm',-create=>1);
|
|
$vmtab->setNodeAttribs($node,$updatehash);
|
|
}
|
|
xCAT::SvrUtils::sendmsg($intent, $output_handler,$node);
|
|
} elsif ($state eq 'error') {
|
|
relay_vmware_err($task,"",$node);
|
|
}
|
|
}
|
|
sub generic_task_callback {
|
|
my $task = shift;
|
|
my $parms = shift;
|
|
my $state = $task->info->state->val;
|
|
my $node = $parms->{node};
|
|
my $intent = $parms->{successtext};
|
|
if ($state eq 'success') {
|
|
xCAT::SvrUtils::sendmsg($intent, $output_handler,$node);
|
|
} elsif ($state eq 'error') {
|
|
relay_vmware_err($task,"",$node);
|
|
}
|
|
}
|
|
|
|
sub migrate {
|
|
my %args = @_;
|
|
my @nodes = @{$args{nodes}};
|
|
my $hyp = $args{hyp};
|
|
my $vcenter = $hyphash{$hyp}->{vcenter}->{name};
|
|
my $datastoredest;
|
|
my $offline;
|
|
@ARGV=@{$args{exargs}};
|
|
unless (GetOptions(
|
|
's=s' => \$datastoredest,
|
|
'f' => \$offline,
|
|
)) {
|
|
xCAT::SvrUtils::sendmsg([1,"Error parsing arguments"], $output_handler);
|
|
return;
|
|
}
|
|
my $target=$hyp; #case for storage migration
|
|
if ($datastoredest) { $datastoredest =~ s/=.*//; }#remove =scsi and similar if specified
|
|
if ($datastoredest and scalar @ARGV) {
|
|
xCAT::SvrUtils::sendmsg([1,"Unable to mix storage migration and processing of arguments ".join(' ',@ARGV)], $output_handler);
|
|
return;
|
|
} elsif (@ARGV) {
|
|
$target=shift @ARGV;
|
|
if (@ARGV) {
|
|
xCAT::SvrUtils::sendmsg([1,"Unrecognized arguments ".join(' ',@ARGV)], $output_handler);
|
|
return;
|
|
}
|
|
} elsif ($datastoredest) { #storage migration only
|
|
unless (validate_datastore_prereqs([],$hyp,{$datastoredest=>\@nodes})) {
|
|
xCAT::SvrUtils::sendmsg([1,"Unable to find/mount target datastore $datastoredest"], $output_handler);
|
|
return;
|
|
}
|
|
foreach (@nodes) {
|
|
my $hostview = $hyphash{$hyp}->{conn}->find_entity_view(view_type=>'VirtualMachine',properties=>['config.name'],filter=>{name=>$_});
|
|
my $relocatspec = VirtualMachineRelocateSpec->new(
|
|
datastore=>$hyphash{$hyp}->{datastorerefmap}->{$datastoredest},
|
|
);
|
|
my $task = $hostview->RelocateVM_Task(spec=>$relocatspec);
|
|
$running_tasks{$task}->{task} = $task;
|
|
$running_tasks{$task}->{callback} = \&relocate_callback;
|
|
$running_tasks{$task}->{hyp} = $args{hyp};
|
|
$running_tasks{$task}->{data} = { node => $_, target=>$datastoredest };
|
|
process_tasks; #check for tasks needing followup actions before the task is forgotten (VMWare's memory is fairly short at times
|
|
}
|
|
return;
|
|
}
|
|
if ((not $offline and $vcenterhash{$vcenter}->{badhyps}->{$hyp}) or $vcenterhash{$vcenter}->{badhyps}->{$target}) {
|
|
xCAT::SvrUtils::sendmsg([1,"Unable to migrate ".join(',',@nodes)." to $target due to inability to validate vCenter connectivity"], $output_handler);
|
|
return;
|
|
}
|
|
if (($offline or $vcenterhash{$vcenter}->{goodhyps}->{$hyp}) and $vcenterhash{$vcenter}->{goodhyps}->{$target}) {
|
|
unless (validate_datastore_prereqs(\@nodes,$target)) {
|
|
xCAT::SvrUtils::sendmsg([1,"Unable to verify storage state on target system"], $output_handler);
|
|
return;
|
|
}
|
|
unless (validate_network_prereqs(\@nodes,$target)) {
|
|
xCAT::SvrUtils::sendmsg([1,"Unable to verify target network state"], $output_handler);
|
|
return;
|
|
}
|
|
my $dstview = get_hostview(conn=>$hyphash{$target}->{conn},hypname=>$target,properties=>['name','parent']);
|
|
unless ($hyphash{$target}->{pool}) {
|
|
$hyphash{$target}->{pool} = $hyphash{$target}->{conn}->get_view(mo_ref=>$dstview->parent,properties=>['resourcePool'])->resourcePool;
|
|
}
|
|
foreach (@nodes) {
|
|
process_tasks; #check for tasks needing followup actions before the task is forgotten (VMWare's memory is fairly short at times
|
|
my $srcview = $hyphash{$target}->{conn}->find_entity_view(view_type=>'VirtualMachine',properties=>['config.name'],filter=>{name=>$_});
|
|
if ($offline and not $srcview) { #we have a request to resurrect the dead..
|
|
register_vm($target,$_,undef,\&migrate_ok,{ nodes => [$_], exargs => $args{exargs}, target=>$target, hyp => $args{hyp}, offline => $offline },"failonerror");
|
|
return;
|
|
} elsif (not $srcview) {
|
|
$srcview = $hyphash{$hyp}->{conn}->find_entity_view(view_type=>'VirtualMachine',properties=>['config.name'],filter=>{name=>$_});
|
|
}
|
|
unless ($srcview) {
|
|
xCAT::SvrUtils::sendmsg([1,"Unable to locate node in vCenter"], $output_handler,$_);
|
|
next;
|
|
}
|
|
|
|
my $task = $srcview->MigrateVM_Task(
|
|
host=>$dstview,
|
|
pool=>$hyphash{$target}->{pool},
|
|
priority=>VirtualMachineMovePriority->new('highPriority'));
|
|
$running_tasks{$task}->{task} = $task;
|
|
$running_tasks{$task}->{callback} = \&migrate_callback;
|
|
$running_tasks{$task}->{hyp} = $args{hyp};
|
|
$running_tasks{$task}->{data} = { node => $_, src=>$hyp, target=>$target, offline => $offline };
|
|
}
|
|
} else {
|
|
#xCAT::SvrUtils::sendmsg("Waiting for BOTH to be 'good'");
|
|
return; #One of them is still 'pending'
|
|
}
|
|
}
|
|
|
|
|
|
sub reconfig_callback {
|
|
my $task = shift;
|
|
my $args = shift;
|
|
#$args->{reconfig_args}->{vmview}->update_view_data();
|
|
delete $args->{reconfig_args}->{vmview}; #Force a reload of the view, update_view_data seems to not work as advertised..
|
|
$args->{reconfig_fun}->(%{$args->{reconfig_args}});
|
|
}
|
|
|
|
sub repower { #Called to try power again after power down for reconfig
|
|
my $task = shift;
|
|
my $args = shift;
|
|
my $powargs=$args->{power_args};
|
|
$powargs->{pretendop}=1;
|
|
#$powargs->{vmview}->update_view_data();
|
|
delete $powargs->{vmview}; #Force a reload of the view, update_view_data seems to not work as advertised..
|
|
power(%$powargs);
|
|
}
|
|
|
|
sub retry_rmvm {
|
|
my $task = shift;
|
|
my $args = shift;
|
|
my $node = $args->{node};
|
|
my $state = $task->info->state->val;
|
|
if ($state eq "success") {
|
|
#$Data::Dumper::Maxdepth=2;
|
|
delete $args->{args}->{vmview};
|
|
rmvm(%{$args->{args}});
|
|
} elsif ($state eq 'error') {
|
|
relay_vmware_err($task,"",$node);
|
|
}
|
|
}
|
|
|
|
sub rmvm {
|
|
my %args = @_;
|
|
my $node = $args{node};
|
|
my $hyp = $args{hyp};
|
|
if (not defined $args{vmview}) { #attempt one refresh
|
|
$args{vmview} = $vmhash{$node}->{conn}->find_entity_view(view_type => 'VirtualMachine',properties=>['config.name','runtime.powerState'],filter=>{name=>$node});
|
|
if (not defined $args{vmview}) {
|
|
xCAT::SvrUtils::sendmsg([1,"VM does not appear to exist"], $output_handler,$node);
|
|
return;
|
|
}
|
|
}
|
|
@ARGV= @{$args{exargs}};
|
|
require Getopt::Long;
|
|
my $forceremove;
|
|
my $purge;
|
|
GetOptions(
|
|
'f' => \$forceremove,
|
|
'p' => \$purge,
|
|
);
|
|
my $task;
|
|
unless ($args{vmview}->{'runtime.powerState'}->val eq 'poweredOff') {
|
|
if ($forceremove) {
|
|
$task = $args{vmview}->PowerOffVM_Task();
|
|
$running_tasks{$task}->{task} = $task;
|
|
$running_tasks{$task}->{callback} = \&retry_rmvm,
|
|
$running_tasks{$task}->{hyp} = $args{hyp};
|
|
$running_tasks{$task}->{vm} = $node;
|
|
$running_tasks{$task}->{data} = { node => $node, args=>\%args };
|
|
return;
|
|
} else {
|
|
xCAT::SvrUtils::sendmsg([1,"Cannot rmvm active guest (use -f argument to force)"], $output_handler,$node);
|
|
return;
|
|
}
|
|
}
|
|
if ($purge) {
|
|
$task = $args{vmview}->Destroy_Task();
|
|
$running_tasks{$task}->{data} = { node => $node, successtext => 'purged' };
|
|
$running_tasks{$task}->{task} = $task;
|
|
$running_tasks{$task}->{callback} = \&generic_task_callback;
|
|
$running_tasks{$task}->{hyp} = $args{hyp}; #$hyp_conns->{$hyp};
|
|
$running_tasks{$task}->{vm} = $node;
|
|
} else {
|
|
$task = $args{vmview}->UnregisterVM();
|
|
}
|
|
}
|
|
|
|
|
|
|
|
sub getreconfigspec {
|
|
my %args = @_;
|
|
my $node = $args{node};
|
|
my $vmview = $args{view};
|
|
my $currid=$args{view}->{'config.guestId'};
|
|
my $rightid=getguestid($node);
|
|
my %conargs;
|
|
my $reconfigneeded=0;
|
|
if ($currid ne $rightid) {
|
|
$reconfigneeded=1;
|
|
$conargs{guestId}=$rightid;
|
|
}
|
|
my $newmem;
|
|
if ($tablecfg{vm}->{$node}->[0]->{memory} and $newmem = getUnits($tablecfg{vm}->{$node}->[0]->{memory},"M",1048576)) {
|
|
my $currmem = $vmview->{'config.hardware.memoryMB'};
|
|
if ($newmem ne $currmem) {
|
|
$conargs{memoryMB} = $newmem;
|
|
$reconfigneeded=1;
|
|
}
|
|
}
|
|
my $newcpus;
|
|
if ($tablecfg{vm}->{$node}->[0]->{cpus} and $newcpus = $tablecfg{vm}->{$node}->[0]->{cpus}) {
|
|
my $currncpu = $vmview->{'config.hardware.numCPU'};
|
|
if ($newcpus ne $currncpu) {
|
|
$conargs{numCPUs} = $newcpus;
|
|
$reconfigneeded=1;
|
|
}
|
|
}
|
|
if ($reconfigneeded) {
|
|
return VirtualMachineConfigSpec->new(%conargs);
|
|
} else {
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
#This routine takes a single node, managing vmv instance, and task tracking hash to submit a power on request
|
|
sub power {
|
|
my %args = @_;
|
|
my $node = $args{node};
|
|
my $hyp = $args{hyp};
|
|
my $pretendop = $args{pretendop}; #to pretend a system was on for reset or boot when we have to turn it off internally for reconfig
|
|
if (not defined $args{vmview}) { #attempt one refresh
|
|
$args{vmview} = $vmhash{$node}->{conn}->find_entity_view(view_type => 'VirtualMachine',properties=>['config.name','config.guestId','config.hardware.memoryMB','config.hardware.numCPU','runtime.powerState'],filter=>{name=>$node});
|
|
#vmview not existing now is not an issue, this function
|
|
#is designed to handle that and correct if reasonably possible
|
|
#comes into play particularly in a stateless context
|
|
}
|
|
@ARGV = @{$args{exargs}}; #for getoptions;
|
|
my $forceon;
|
|
require Getopt::Long;
|
|
GetOptions(
|
|
'force|f' => \$forceon,
|
|
);
|
|
my $subcmd = $ARGV[0];
|
|
my $intent="";
|
|
my $task;
|
|
my $currstat;
|
|
|
|
if ($args{vmview}) {
|
|
$currstat = $args{vmview}->{'runtime.powerState'}->val;
|
|
if (grep /$subcmd/,qw/on reset boot/) {
|
|
my $reconfigspec;
|
|
if ($reconfigreset and ($reconfigspec = getreconfigspec(node=>$node,view=>$args{vmview}))) {
|
|
if ($currstat eq 'poweredOff') {
|
|
#xCAT::SvrUtils::sendmsg("Correcting guestId because $currid and $rightid are not the same...");#DEBUG
|
|
my $task = $args{vmview}->ReconfigVM_Task(spec=>$reconfigspec);
|
|
$running_tasks{$task}->{task} = $task;
|
|
$running_tasks{$task}->{callback} = \&reconfig_callback;
|
|
$running_tasks{$task}->{hyp} = $args{hyp};
|
|
$running_tasks{$task}->{vm} = $node;
|
|
$running_tasks{$task}->{data} = { node => $node, reconfig_fun=>\&power, reconfig_args=>\%args };
|
|
return;
|
|
} elsif (grep /$subcmd/,qw/reset boot/) { #going to have to do a 'cycle' and present it up normally..
|
|
#xCAT::SvrUtils::sendmsg("DEBUG: forcing a cycle");
|
|
$task = $args{vmview}->PowerOffVM_Task();
|
|
$running_tasks{$task}->{task} = $task;
|
|
$running_tasks{$task}->{callback} = \&repower;
|
|
$running_tasks{$task}->{hyp} = $args{hyp};
|
|
$running_tasks{$task}->{vm} = $node;
|
|
$running_tasks{$task}->{data} = { node => $node, power_args=>\%args};
|
|
return; #we have to wait
|
|
}
|
|
#TODO: fixit
|
|
#xCAT::SvrUtils::sendmsg("I see vm has $currid and I want it to be $rightid");
|
|
}
|
|
}
|
|
} else {
|
|
$currstat = 'off';
|
|
}
|
|
if ($currstat eq 'poweredOff') {
|
|
$currstat = 'off';
|
|
} elsif ($currstat eq 'poweredOn') {
|
|
$currstat = 'on';
|
|
} elsif ($currstat eq 'suspended') {
|
|
$currstat = 'suspend';
|
|
}
|
|
if ($subcmd =~ /^stat/) {
|
|
xCAT::SvrUtils::sendmsg($currstat, $output_handler,$node);
|
|
return;
|
|
}
|
|
if ($subcmd =~ /boot/) {
|
|
$intent = "$currstat ";
|
|
if ($currstat eq 'on' or $args{pretendop}) {
|
|
$intent = "on ";
|
|
$subcmd = 'reset';
|
|
} else {
|
|
$subcmd = 'on';
|
|
}
|
|
}
|
|
if ($subcmd =~ /on/) {
|
|
if ($currstat eq 'off' or $currstat eq 'suspend') {
|
|
if (not $args{vmview}) { #We are asking to turn on a system the hypervisor
|
|
#doesn't know, attempt to register it first
|
|
register_vm($hyp,$node,undef,\&power,\%args);
|
|
return; #We'll pick it up on the retry if it gets registered
|
|
}
|
|
eval {
|
|
if ($hyp) {
|
|
$task = $args{vmview}->PowerOnVM_Task(host=>$hyphash{$hyp}->{hostview});
|
|
} else {
|
|
$task = $args{vmview}->PowerOnVM_Task(); #DRS may have it's way with me
|
|
}
|
|
};
|
|
if ($@) {
|
|
xCAT::SvrUtils::sendmsg([1,":".$@], $output_handler,$node);
|
|
return;
|
|
}
|
|
$running_tasks{$task}->{task} = $task;
|
|
$running_tasks{$task}->{callback} = \&poweron_task_callback;
|
|
$running_tasks{$task}->{hyp} = $args{hyp}; #$hyp_conns->{$hyp};
|
|
$running_tasks{$task}->{vm} = $node;
|
|
$running_tasks{$task}->{data} = { node => $node, successtext => $intent.'on', forceon=>$forceon };
|
|
} else {
|
|
xCAT::SvrUtils::sendmsg($currstat, $output_handler,$node);
|
|
}
|
|
} elsif ($subcmd =~ /softoff/) {
|
|
if ($currstat eq 'on') {
|
|
$args{vmview}->ShutdownGuest();
|
|
xCAT::SvrUtils::sendmsg("softoff", $output_handler,$node);
|
|
} else {
|
|
xCAT::SvrUtils::sendmsg($currstat, $output_handler,$node);
|
|
}
|
|
} elsif ($subcmd =~ /off/) {
|
|
if ($currstat eq 'on') {
|
|
$task = $args{vmview}->PowerOffVM_Task();
|
|
$running_tasks{$task}->{task} = $task;
|
|
$running_tasks{$task}->{callback} = \&generic_task_callback;
|
|
$running_tasks{$task}->{hyp} = $args{hyp};
|
|
$running_tasks{$task}->{vm} = $node;
|
|
$running_tasks{$task}->{data} = { node => $node, successtext => 'off' };
|
|
} else {
|
|
xCAT::SvrUtils::sendmsg($currstat, $output_handler,$node);
|
|
}
|
|
} elsif ($subcmd =~ /suspend/) {
|
|
if ($currstat eq 'on') {
|
|
$task = $args{vmview}->SuspendVM_Task();
|
|
$running_tasks{$task}->{task} = $task;
|
|
$running_tasks{$task}->{callback} = \&generic_task_callback;
|
|
$running_tasks{$task}->{hyp} = $args{hyp};
|
|
$running_tasks{$task}->{vm} = $node;
|
|
$running_tasks{$task}->{data} = { node => $node, successtext => 'suspend' };
|
|
} else {
|
|
xCAT::SvrUtils::sendmsg("off", $output_handler,$node);
|
|
}
|
|
} elsif ($subcmd =~ /reset/) {
|
|
if ($currstat eq 'on') {
|
|
$task = $args{vmview}->ResetVM_Task();
|
|
$running_tasks{$task}->{task} = $task;
|
|
$running_tasks{$task}->{callback} = \&generic_task_callback;
|
|
$running_tasks{$task}->{hyp} = $args{hyp};
|
|
$running_tasks{$task}->{vm} = $node;
|
|
$running_tasks{$task}->{data} = { node => $node, successtext => $intent.'reset' };
|
|
} elsif ($args{pretendop}) { #It is off, but pretend it was on
|
|
eval {
|
|
if ($hyp) {
|
|
$task = $args{vmview}->PowerOnVM_Task(host=>$hyphash{$hyp}->{hostview});
|
|
} else {
|
|
$task = $args{vmview}->PowerOnVM_Task(); #allow DRS
|
|
}
|
|
};
|
|
if ($@) {
|
|
xCAT::SvrUtils::sendmsg([1,":".$@], $output_handler,$node);
|
|
return;
|
|
}
|
|
$running_tasks{$task}->{task} = $task;
|
|
$running_tasks{$task}->{callback} = \&generic_task_callback;
|
|
$running_tasks{$task}->{hyp} = $args{hyp};
|
|
$running_tasks{$task}->{vm} = $node;
|
|
$running_tasks{$task}->{data} = { node => $node, successtext => $intent.'reset' };
|
|
} else {
|
|
xCAT::SvrUtils::sendmsg($currstat, $output_handler,$node);
|
|
}
|
|
}
|
|
}
|
|
sub generic_vm_operation { #The general form of firing per-vm requests to ESX hypervisor
|
|
my $properties = shift; #The relevant properties to the general task, MUST INCLUDE config.name
|
|
my $function = shift; #The function to actually run against the right VM view
|
|
my @exargs = @_; #Store the rest to pass on
|
|
my $hyp;
|
|
my $vmviews;
|
|
my %vcviews; #views populated once per vcenter server for improved performance
|
|
my $node;
|
|
foreach $hyp (keys %hyphash) {
|
|
if ($viavcenterbyhyp->{$hyp}) {
|
|
foreach $node (keys %{$hyphash{$hyp}->{nodes}}){
|
|
$vcenterhash{$hyphash{$hyp}->{vcenter}->{name}}->{vms}->{$node}=1;
|
|
}
|
|
}
|
|
}
|
|
foreach (keys %limbonodes) {
|
|
$vcenterhash{$limbonodes{$_}}->{vms}->{$_}=1;
|
|
}
|
|
my $cluster;
|
|
foreach $cluster (keys %clusterhash) {
|
|
foreach $node (keys %{$clusterhash{$cluster}->{nodes}}) {
|
|
$vcenterhash{$clusterhash{$cluster}->{vcenter}->{name}}->{vms}->{$node}=1;
|
|
}
|
|
}
|
|
my $currentvcenter;
|
|
my %foundlimbo;
|
|
foreach $currentvcenter (keys %vcenterhash) {
|
|
#retrieve all vm views in one gulp
|
|
my $vmsearchstring = join(")|(",keys %{$vcenterhash{$currentvcenter}->{vms}});
|
|
$vmsearchstring = '^(('.$vmsearchstring.'))(\z|\.)';
|
|
my $regex = qr/$vmsearchstring/;
|
|
$vcviews{$currentvcenter} = $vcenterhash{$currentvcenter}->{conn}->find_entity_views(view_type => 'VirtualMachine',properties=>$properties,filter=>{'config.name'=>$regex});
|
|
foreach (@{$vcviews{$currentvcenter}}) {
|
|
my $node = $_->{'config.name'};
|
|
unless (defined $tablecfg{vm}->{$node}) {
|
|
$node =~ s/\..*//; #try the short name;
|
|
}
|
|
if (defined $tablecfg{vm}->{$node}) { #see if the host pointer requires a refresh
|
|
my $hostref = $hostrefbynode{$node};
|
|
if ($hostref and $hostref eq $_->{'runtime.host'}->value) { next; } #the actual host reference matches the one that we got when populating hostviews based on what the table had to say #TODO: does this mean it is buggy if we want to mkvm/rmigrate/etc if the current vm.host is wrong and the noderange doesn't have something on the right hostview making us not get it in the
|
|
#mass request? Or is it just slower because it hand gets host views?
|
|
my $host = $vcenterhash{$currentvcenter}->{conn}->get_view(mo_ref=>$_->{'runtime.host'},properties=>['summary.config.name']);
|
|
$host = $host->{'summary.config.name'};
|
|
my $shost = $host;
|
|
$shost =~ s/\..*//;
|
|
#time to figure out which of these is a node
|
|
my @nodes = noderange("$host,$shost");
|
|
my $vmtab = xCAT::Table->new("vm",-create=>1);
|
|
unless($vmtab){
|
|
die "Error opening vm table";
|
|
}
|
|
if ($nodes[0]) {
|
|
if ($limbonodes{$node}) { $foundlimbo{$node}=$currentvcenter; }
|
|
$vmtab->setNodeAttribs($node,{host=>$nodes[0]});
|
|
} #else {
|
|
# $vmtab->setNodeAttribs($node,{host=>$host});
|
|
#}
|
|
}
|
|
}
|
|
}
|
|
foreach my $lnode (keys %foundlimbo) {
|
|
$vmviews= $vcviews{$foundlimbo{$lnode}};
|
|
my %mgdvms; #sort into a hash for convenience
|
|
foreach (@$vmviews) {
|
|
$mgdvms{$_->{'config.name'}} = $_;
|
|
}
|
|
$function->(
|
|
node=>$lnode,
|
|
vm=>$lnode,
|
|
vmview=>$mgdvms{$node},
|
|
exargs=>\@exargs
|
|
);
|
|
}
|
|
my @entitylist;
|
|
push @entitylist,keys %hyphash;
|
|
push @entitylist,keys %clusterhash;
|
|
foreach my $entity (@entitylist) {
|
|
if ($hyphash{$entity}) {
|
|
$hyp=$entity; #save some retyping...
|
|
if ($viavcenterbyhyp->{$hyp}) {
|
|
$vmviews= $vcviews{$hyphash{$hyp}->{vcenter}->{name}}
|
|
} else {
|
|
$vmviews = [];
|
|
my $node;
|
|
foreach $node (sort (keys %{$hyphash{$hyp}->{nodes}})){
|
|
push @{$vmviews},$hyphash{$hyp}->{conn}->find_entity_view(view_type => 'VirtualMachine',properties=>$properties,filter=>{'config.name'=>qr/^$node/});
|
|
}
|
|
#$vmviews = $hyphash{$hyp}->{conn}->find_entity_views(view_type => 'VirtualMachine',properties=>$properties);
|
|
}
|
|
my %mgdvms; #sort into a hash for convenience
|
|
foreach (@$vmviews) {
|
|
$mgdvms{$_->{'config.name'}} = $_;
|
|
}
|
|
my $node;
|
|
foreach $node (sort (keys %{$hyphash{$hyp}->{nodes}})){
|
|
$function->(
|
|
node=>$node,
|
|
hyp=>$hyp,
|
|
vmview=>$mgdvms{$node},
|
|
exargs=>\@exargs
|
|
);
|
|
process_tasks; #check for tasks needing followup actions before the task is forgotten (VMWare's memory is fairly short at times
|
|
}
|
|
} else { #a cluster.
|
|
$vmviews= $vcviews{$clusterhash{$entity}->{vcenter}->{name}};
|
|
my %mgdvms; #sort into a hash for convenience
|
|
foreach (@$vmviews) {
|
|
$mgdvms{$_->{'config.name'}} = $_;
|
|
}
|
|
my $node;
|
|
foreach $node (sort (keys %{$clusterhash{$entity}->{nodes}})){
|
|
$function->(
|
|
node=>$node,
|
|
cluster=>$entity,
|
|
vm=>$node,
|
|
vmview=>$mgdvms{$node},
|
|
exargs=>\@exargs
|
|
);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
sub generic_hyp_operation { #The general form of firing per-hypervisor requests to ESX hypervisor
|
|
my $function = shift; #The function to actually run against the right VM view
|
|
my @exargs = @_; #Store the rest to pass on
|
|
my $hyp;
|
|
if (scalar keys %limbonodes) { #we are in forced migration with dead sources, try to register them
|
|
@ARGV=@exargs;
|
|
my $datastoredest;
|
|
my $offline;
|
|
unless (GetOptions(
|
|
's=s' => \$datastoredest,
|
|
'f' => \$offline,
|
|
)) {
|
|
xCAT::SvrUtils::sendmsg([1,"Error parsing arguments"], $output_handler);
|
|
return;
|
|
}
|
|
if ($datastoredest) {
|
|
xCAT::SvrUtils::sendmsg([1,"Storage migration impossible with dead hypervisor, must be migrated to live hypervisor first"], $output_handler);
|
|
return;
|
|
} elsif (@ARGV) {
|
|
my $target=shift @ARGV;
|
|
if (@ARGV) {
|
|
xCAT::SvrUtils::sendmsg([1,"Unrecognized arguments ".join(' ',@ARGV)], $output_handler);
|
|
return;
|
|
}
|
|
foreach (keys %limbonodes) {
|
|
register_vm($target,$_,undef,\&migrate_ok,{ nodes => [$_], target=>$target, },"failonerror");
|
|
}
|
|
} else { #storage migration only
|
|
xCAT::SvrUtils::sendmsg([1,"No target hypervisor specified"], $output_handler);
|
|
|
|
}
|
|
}
|
|
foreach $hyp (keys %hyphash) {
|
|
process_tasks; #check for tasks needing followup actions before the task is forgotten (VMWare's memory is fairly short at times
|
|
my @relevant_nodes = sort (keys %{$hyphash{$hyp}->{nodes}});
|
|
unless (scalar @relevant_nodes) {
|
|
next;
|
|
}
|
|
$function->(
|
|
nodes => \@relevant_nodes,
|
|
hyp => $hyp,
|
|
exargs => \@exargs
|
|
);
|
|
#my $vmviews = $hyp_conns->{$hyp}->find_entity_views(view_type => 'VirtualMachine',properties=>['runtime.powerState','config.name']);
|
|
#my %mgdvms; #sort into a hash for convenience
|
|
#foreach (@$vmviews) {
|
|
# $mgdvms{$_->{'config.name'}} = $_;
|
|
#}
|
|
#my $node;
|
|
#foreach $node (sort (keys %{$hyp_hash->{$hyp}->{nodes}})){
|
|
# $function->($node,$mgdvms{$node},$taskstotrack,$callback,@exargs);
|
|
#REMINDER FOR RINV TO COME
|
|
# foreach (@nothing) { #@{$mgdvms{$node}->config->hardware->device}) {
|
|
# if (defined $_->{macAddress}) {
|
|
# print "\nFound a mac: ".$_->macAddress."\n";
|
|
# }
|
|
# }
|
|
# }
|
|
}
|
|
foreach $hyp (keys %clusterhash) { #clonevm, mkvm, rmigrate could land here in clustered mode with DRS/HA
|
|
process_tasks;
|
|
my @relevant_nodes = sort (keys %{$clusterhash{$hyp}->{nodes}});
|
|
unless (scalar @relevant_nodes) {
|
|
next;
|
|
}
|
|
$function->(nodes => \@relevant_nodes,cluster=>$hyp,exargs => \@exargs,conn=>$clusterhash{$hyp}->{conn});
|
|
}
|
|
}
|
|
|
|
sub rmhypervisor_disconnected {
|
|
my $task = shift;
|
|
my $parms = shift;
|
|
my $node = $parms->{node};
|
|
my $hyp = $node;
|
|
my $state = $task->info->state->val;
|
|
if ($state eq 'success') {
|
|
my $task = $hyphash{$hyp}->{hostview}->Destroy_Task();
|
|
$running_tasks{$task}->{data} = { node => $node, successtext => 'removed' };
|
|
$running_tasks{$task}->{task} = $task;
|
|
$running_tasks{$task}->{callback} = \&generic_task_callback;
|
|
$running_tasks{$task}->{hyp} =$hyp;
|
|
} elsif ($state eq 'error') {
|
|
relay_vmware_err($task,"",$node);
|
|
}
|
|
}
|
|
sub rmhypervisor_inmaintenance {
|
|
my $task = shift;
|
|
my $parms = shift;
|
|
my $state = $task->info->state->val;
|
|
my $node = $parms->{node};
|
|
my $intent = $parms->{successtext};
|
|
if ($state eq 'success') {
|
|
my $hyp = $parms->{node};
|
|
my $task = $hyphash{$hyp}->{hostview}->DisconnectHost_Task();
|
|
$running_tasks{$task}->{task} = $task;
|
|
$running_tasks{$task}->{callback} = \&rmhypervisor_disconnected;
|
|
$running_tasks{$task}->{hyp} = $hyp;
|
|
$running_tasks{$task}->{data} = { node => $hyp };
|
|
} elsif ($state eq 'error') {
|
|
relay_vmware_err($task,"",$node);
|
|
}
|
|
}
|
|
|
|
sub lsvm {
|
|
my %args = @_;
|
|
my $hyp = $args{hyp};
|
|
$hyphash{$hyp}->{hostview} = get_hostview(hypname=>$hyp,conn=>$hyphash{$hyp}->{conn}); #,properties=>['config','configManager']);
|
|
use Data::Dumper;
|
|
my $vms = $hyphash{$hyp}->{hostview}->vm;
|
|
unless ($vms) {
|
|
return;
|
|
}
|
|
foreach (@$vms) {
|
|
my $vmv = $hyphash{$hyp}->{conn}->get_view(mo_ref=>$_);
|
|
xCAT::SvrUtils::sendmsg($vmv->name, $output_handler,$hyp);
|
|
}
|
|
return;
|
|
}
|
|
|
|
sub chhypervisor {
|
|
my %args = @_;
|
|
@ARGV = @{$args{exargs}}; #for getoptions;
|
|
my $maintenance;
|
|
my $online;
|
|
my $stat;
|
|
my $vlanaddspec;
|
|
my $vlanremspec;
|
|
require Getopt::Long;
|
|
GetOptions(
|
|
'maintenance|m' => \$maintenance,
|
|
'online|o' => \$online,
|
|
'show|s' => \$stat,
|
|
'show|s' => \$stat,
|
|
'addvlan=s' => \$vlanaddspec,
|
|
'removevlan=s' => \$vlanremspec,
|
|
);
|
|
my $hyp = $args{hyp};
|
|
$hyphash{$hyp}->{hostview} = get_hostview(hypname=>$hyp,conn=>$hyphash{$hyp}->{conn}); #,properties=>['config','configManager']);
|
|
if ($maintenance) {
|
|
if (defined $hyphash{$hyp}->{hostview}) {
|
|
my $task = $hyphash{$hyp}->{hostview}->EnterMaintenanceMode_Task(timeout=>0);
|
|
$running_tasks{$task}->{task} = $task;
|
|
$running_tasks{$task}->{callback} = \&generic_task_callback;
|
|
$running_tasks{$task}->{hyp} = $args{hyp};
|
|
$running_tasks{$task}->{data} = { node => $hyp , successtext => "hypervisor in maintenance mode"};
|
|
}
|
|
} elsif ($online) {
|
|
if (defined $hyphash{$hyp}->{hostview}) {
|
|
my $task = $hyphash{$hyp}->{hostview}->ExitMaintenanceMode_Task(timeout=>0);
|
|
$running_tasks{$task}->{task} = $task;
|
|
$running_tasks{$task}->{callback} = \&generic_task_callback;
|
|
$running_tasks{$task}->{hyp} = $args{hyp};
|
|
$running_tasks{$task}->{data} = { node => $hyp , successtext => "hypervisor online"};
|
|
}
|
|
} elsif ($stat) {
|
|
if (defined $hyphash{$hyp}->{hostview}) {
|
|
if ($hyphash{$hyp}->{hostview}->runtime->inMaintenanceMode) {
|
|
xCAT::SvrUtils::sendmsg("hypervisor in maintenance mode", $output_handler,$hyp);
|
|
} else {
|
|
xCAT::SvrUtils::sendmsg("hypervisor online", $output_handler,$hyp);
|
|
}
|
|
}
|
|
} elsif ($vlanaddspec) {
|
|
fixup_hostportgroup($vlanaddspec, $hyp);
|
|
} elsif ($vlanremspec) {
|
|
fixup_hostportgroup($vlanremspec, $hyp, action=>'remove');
|
|
}
|
|
return;
|
|
}
|
|
|
|
sub rshutdown { #TODO: refactor with next function too
|
|
my %args = @_;
|
|
my $hyp = $args{hyp};
|
|
$hyphash{$hyp}->{hostview} = get_hostview(hypname=>$hyp,conn=>$hyphash{$hyp}->{conn}); #,properties=>['config','configManager']);
|
|
if (defined $hyphash{$hyp}->{hostview}) {
|
|
my $task = $hyphash{$hyp}->{hostview}->EnterMaintenanceMode_Task(timeout=>0);
|
|
$running_tasks{$task}->{task} = $task;
|
|
$running_tasks{$task}->{callback} = \&rshutdown_inmaintenance;
|
|
$running_tasks{$task}->{hyp} = $args{hyp};
|
|
$running_tasks{$task}->{data} = { node => $hyp };
|
|
}
|
|
return;
|
|
}
|
|
|
|
sub rshutdown_inmaintenance {
|
|
my $task = shift;
|
|
my $parms = shift;
|
|
my $state = $task->info->state->val;
|
|
my $node = $parms->{node};
|
|
my $intent = $parms->{successtext};
|
|
if ($state eq 'success') {
|
|
my $hyp = $parms->{node};
|
|
if (defined $hyphash{$hyp}->{hostview}) {
|
|
my $task = $hyphash{$hyp}->{hostview}->ShutdownHost_Task(force=>0);
|
|
$running_tasks{$task}->{task} = $task;
|
|
$running_tasks{$task}->{callback} = \&generic_task_callback;
|
|
$running_tasks{$task}->{hyp} = $hyp;
|
|
$running_tasks{$task}->{data} = { node => $hyp, successtext => "shutdown initiated" };
|
|
}
|
|
} elsif ($state eq 'error') {
|
|
relay_vmware_err($task,"",$node);
|
|
}
|
|
return;
|
|
}
|
|
|
|
sub rescansan {
|
|
my %args = @_;
|
|
my $hyp = $args{hyp};
|
|
my $hostview = get_hostview(hypname=>$hyp,conn=>$hyphash{$hyp}->{conn},properties=>['config','configManager']);
|
|
if (defined $hostview) {
|
|
my $hdss = $hostview->{vim}->get_view(mo_ref=>$hostview->configManager->storageSystem);
|
|
$hdss->RescanAllHba();
|
|
$hdss->RescanVmfs();
|
|
}
|
|
}
|
|
|
|
sub formatdisk {
|
|
my %args = @_;
|
|
my $hyp = $args{hyp};
|
|
$hyphash{$hyp}->{hostview} = get_hostview(hypname=>$hyp,conn=>$hyphash{$hyp}->{conn},properties=>['config','configManager']);
|
|
@ARGV = @{$args{exargs}};
|
|
my $nid;
|
|
my $name;
|
|
GetOptions(
|
|
'id=s' => \$nid,
|
|
'name=s' => \$name,
|
|
);
|
|
my $hostview = $hyphash{$hyp}->{hostview};
|
|
if (defined $hyphash{$hyp}->{hostview}) {
|
|
my $hdss = $hostview->{vim}->get_view(mo_ref=>$hostview->configManager->storageSystem);
|
|
$hdss->RescanAllHba();
|
|
my $dss = $hostview->{vim}->get_view(mo_ref=>$hostview->configManager->datastoreSystem);
|
|
my $diskList = $dss->QueryAvailableDisksForVmfs();
|
|
foreach my $disk (@$diskList) {
|
|
foreach my $id (@{$disk->{descriptor}}) {
|
|
if (lc($id->{id}) eq lc('naa.'.$nid)) {
|
|
my $options = $dss->QueryVmfsDatastoreCreateOptions(devicePath => $disk->devicePath);
|
|
@$options[0]->spec->vmfs->volumeName($name);
|
|
my $newDatastore = $dss->CreateVmfsDatastore(spec => @$options[0]->spec );
|
|
}
|
|
}
|
|
}
|
|
|
|
}
|
|
return;
|
|
}
|
|
|
|
sub rmhypervisor {
|
|
my %args = @_;
|
|
my $hyp = $args{hyp};
|
|
$hyphash{$hyp}->{hostview} = get_hostview(hypname=>$hyp,conn=>$hyphash{$hyp}->{conn}); #,properties=>['config','configManager']);
|
|
if (defined $hyphash{$hyp}->{hostview}) {
|
|
my $task = $hyphash{$hyp}->{hostview}->EnterMaintenanceMode_Task(timeout=>0);
|
|
$running_tasks{$task}->{task} = $task;
|
|
$running_tasks{$task}->{callback} = \&rmhypervisor_inmaintenance;
|
|
$running_tasks{$task}->{hyp} = $args{hyp};
|
|
$running_tasks{$task}->{data} = { node => $hyp };
|
|
}
|
|
return;
|
|
}
|
|
|
|
sub clonevms {
|
|
my %args=@_;
|
|
my $nodes = $args{nodes};
|
|
my $hyp = $args{hyp};
|
|
my $cluster = $args{cluster};
|
|
@ARGV = @{$args{exargs}}; #for getoptions;
|
|
my $base;
|
|
my $force;
|
|
my $detach;
|
|
my $specialize;
|
|
my $target;
|
|
require Getopt::Long;
|
|
GetOptions(
|
|
'b=s' => \$base,
|
|
'f' => \$force,
|
|
'd' => \$detach,
|
|
'specialize' => \$specialize,
|
|
't=s' => \$target,
|
|
);
|
|
if ($base and $target) {
|
|
foreach my $node (@$nodes) {
|
|
xCAT::SvrUtils::sendmsg([1,"Cannot specify both base (-b) and target (-t)"], $output_handler,$node);
|
|
}
|
|
return;
|
|
}
|
|
unless ($base or $target) {
|
|
foreach my $node (@$nodes) {
|
|
xCAT::SvrUtils::sendmsg([1,"Must specify one of base (-b) or target (-t)"], $output_handler,$node);
|
|
}
|
|
return;
|
|
}
|
|
if ($target and (scalar @{$nodes} != 1)) {
|
|
foreach my $node (@$nodes) {
|
|
xCAT::SvrUtils::sendmsg([1,"Cannot specify mulitple nodes to create a master from"], $output_handler,$node);
|
|
}
|
|
return;
|
|
}
|
|
if ($hyp) {
|
|
$hyphash{$hyp}->{hostview} = get_hostview(hypname=>$hyp,conn=>$hyphash{$hyp}->{conn});
|
|
}
|
|
my $newdatastores;
|
|
my $mastername;
|
|
my $url;
|
|
my $masterref;
|
|
if ($base) { #if base, we need to pull in the target datastores
|
|
my $mastertab=xCAT::Table->new('vmmaster');
|
|
$masterref=$mastertab->getNodeAttribs( $base,[qw/storage os arch profile storagemodel nics specializeparameters/]);
|
|
unless ($masterref) {
|
|
foreach my $node (@$nodes) {
|
|
xCAT::SvrUtils::sendmsg([1,"Cannot find master $base in vmmaster table"], $output_handler,$node);
|
|
}
|
|
return;
|
|
}
|
|
$newdatastores->{$masterref->{storage}}=[]; #make sure that the master datastore is mounted...
|
|
foreach (@$nodes) {
|
|
my $url;
|
|
if ($tablecfg{vm}->{$_}->[0]->{storage}) {
|
|
$url=$tablecfg{vm}->{$_}->[0]->{storage};
|
|
$url =~ s/=.*//;
|
|
} else {
|
|
$url=$masterref->{storage};
|
|
}
|
|
unless ($url) { die "Shouldn't be possible"; }
|
|
if (ref $newdatastores->{$_}) {
|
|
push @{$newdatastores->{$url}},$_;
|
|
} else {
|
|
$newdatastores->{$url}=[$_];
|
|
}
|
|
}
|
|
} elsif ($target) {
|
|
if ($url =~ m!/!) {
|
|
$url=$target;
|
|
$url =~ s!/([^/]*)\z!!;
|
|
$mastername=$1;
|
|
} else {
|
|
$url = $tablecfg{vm}->{$nodes->[0]}->[0]->{storage};
|
|
$url =~ s/.*\|//;
|
|
$url =~ s/=(.*)//;
|
|
$url =~ s/,.*//;
|
|
$mastername=$target
|
|
}
|
|
$newdatastores->{$url}=[$nodes->[0]];
|
|
}
|
|
if ($hyp) {
|
|
unless (validate_datastore_prereqs($nodes,$hyp,$newdatastores)) {
|
|
return;
|
|
}
|
|
} else { #need to build datastore map for cluster
|
|
refreshclusterdatastoremap($cluster);
|
|
}
|
|
sortoutdatacenters(nodes=>$nodes,hyp=>$hyp,cluster=>$cluster);
|
|
if ($target) {
|
|
return promote_vm_to_master(node=>$nodes->[0],target=>$target,force=>$force,detach=>$detach,cluster=>$cluster,hyp=>$hyp,url=>$url,mastername=>$mastername);
|
|
} elsif ($base) {
|
|
return clone_vms_from_master(nodes=>$nodes,base=>$base,detach=>$detach,cluster=>$cluster,hyp=>$hyp,mastername=>$base,masterent=>$masterref,specialize=>$specialize);
|
|
}
|
|
}
|
|
sub sortoutdatacenters { #figure out all the vmfolders for all the nodes passed in
|
|
my %args=@_;
|
|
my $nodes=$args{nodes};
|
|
my $hyp=$args{hyp};
|
|
my $cluster=$args{cluster};
|
|
my %nondefaultdcs;
|
|
my $deffolder;
|
|
my $conn;
|
|
if ($hyp) {
|
|
unless (defined $hyphash{$hyp}->{vmfolder}) {
|
|
$hyphash{$hyp}->{vmfolder} = $hyphash{$hyp}->{conn}->get_view(mo_ref=>$hyphash{$hyp}->{conn}->find_entity_view(view_type=>'Datacenter',properties=>['vmFolder'])->vmFolder);
|
|
}
|
|
$conn= $hyphash{$hyp}->{conn};
|
|
$deffolder=$hyphash{$hyp}->{vmfolder};
|
|
} else { #clustered
|
|
unless (defined $clusterhash{$cluster}->{vmfolder}) {
|
|
$clusterhash{$cluster}->{vmfolder} = $clusterhash{$cluster}->{conn}->get_view(mo_ref=>$clusterhash{$cluster}->{conn}->find_entity_view(view_type=>'Datacenter',properties=>['vmFolder'])->vmFolder);
|
|
}
|
|
$deffolder=$clusterhash{$cluster}->{vmfolder};
|
|
$conn= $clusterhash{$cluster}->{conn};
|
|
}
|
|
foreach (@$nodes) {
|
|
if ($tablecfg{vm}->{$_}->[0]->{datacenter}) {
|
|
$nondefaultdcs{$tablecfg{vm}->{$_}->[0]->{datacenter}}->{$_}=1;
|
|
} else {
|
|
$vmhash{$_}->{vmfolder}=$deffolder;
|
|
}
|
|
}
|
|
my $datacenter;
|
|
foreach $datacenter (keys %nondefaultdcs) {
|
|
my $vmfolder= $conn->get_view(mo_ref=>$conn->find_entity_view(view_type=>'Datacenter',properties=>['vmFolder'],filter=>{name=>$datacenter})->vmFolder,filter=>{name=>$datacenter});
|
|
foreach (keys %{$nondefaultdcs{$datacenter}}) {
|
|
$vmhash{$_}->{vmfolder}=$vmfolder;
|
|
}
|
|
}
|
|
}
|
|
sub clone_vms_from_master {
|
|
my %args = @_;
|
|
my $mastername=$args{mastername};
|
|
my $specialize=$args{specialize};
|
|
my $hyp = $args{hyp};
|
|
my $cluster=$args{cluster};
|
|
my $regex=qr/^$mastername\z/;
|
|
my @nodes=@{$args{nodes}};
|
|
my $node;
|
|
my $conn;
|
|
if ($hyp) {
|
|
$conn=$hyphash{$hyp}->{conn};
|
|
} else {
|
|
$conn=$clusterhash{$cluster}->{conn};
|
|
}
|
|
my $masterviews = $conn->find_entity_views(view_type => 'VirtualMachine',filter=>{'config.name'=>$regex});
|
|
if (scalar(@$masterviews) != 1) {
|
|
foreach $node (@nodes) {
|
|
xCAT::SvrUtils::sendmsg([1,"Unable to find master $mastername in VMWare infrastructure"], $output_handler,$node);
|
|
}
|
|
return;
|
|
}
|
|
my $masterview=$masterviews->[0];
|
|
my $masterent=$args{masterent};
|
|
my $ostype;
|
|
foreach $node (@nodes) {
|
|
my $destination=$tablecfg{vm}->{$node}->[0]->{storage};
|
|
my $nodetypeent;
|
|
my $vment;
|
|
|
|
$ostype=$masterent->{'os'};
|
|
foreach (qw/os arch profile/) {
|
|
$nodetypeent->{$_}=$masterent->{$_};
|
|
}
|
|
foreach (qw/storagemodel nics/) {
|
|
$vment->{$_}=$masterent->{$_};
|
|
}
|
|
$vment->{master}=$args{mastername};
|
|
unless ($destination) {
|
|
$destination=$masterent->{storage};
|
|
$vment->{storage}=$destination;
|
|
}
|
|
$destination =~ s/=.*//;
|
|
my $placement_resources=get_placement_resources(hyp=>$hyp,cluster=>$cluster,destination=>$destination);
|
|
my $pool=$placement_resources->{pool};
|
|
my $dstore=$placement_resources->{datastore};
|
|
my %relocatespecargs = (
|
|
datastore=>$dstore, #$hyphash{$hyp}->{datastorerefmap}->{$destination},
|
|
pool=>$pool,
|
|
#diskMoveType=>"createNewChildDiskBacking", #fyi, requires a snapshot, which isn't compatible with templates, moveChildMostDiskBacking would potentially be fine, but either way is ha incopmatible and limited to 8, arbitrary limitations hard to work around...
|
|
);
|
|
unless ($args{detach}) {
|
|
$relocatespecargs{diskMoveType}="createNewChildDiskBacking";
|
|
}
|
|
if ($hyp) { $relocatespecargs{host}=$hyphash{$hyp}->{hostview} }
|
|
my $relocatespec = VirtualMachineRelocateSpec->new(%relocatespecargs);
|
|
my %clonespecargs = (
|
|
location=>$relocatespec,
|
|
template=>0,
|
|
powerOn=>0
|
|
);
|
|
unless ($args{detach}) {
|
|
$clonespecargs{snapshot}=$masterview->snapshot->currentSnapshot;
|
|
}
|
|
if ($specialize) {
|
|
my %custargs;
|
|
if ($masterent->{specializeparameters}) { %custargs = ( parameters=>$masterent->{specializeparameters} ); }
|
|
$clonespecargs{customization} = make_customization_spec($node,ostype=>$ostype,%custargs);
|
|
}
|
|
my $clonespec = VirtualMachineCloneSpec->new(%clonespecargs);
|
|
my $vmfolder = $vmhash{$node}->{vmfolder};
|
|
my $task = $masterview->CloneVM_Task(folder=>$vmfolder,name=>$node,spec=>$clonespec);
|
|
$running_tasks{$task}->{data} = { node => $node, conn=>$conn, successtext => 'Successfully cloned from '.$args{mastername},
|
|
mastername=>$args{mastername}, nodetypeent=>$nodetypeent,vment=>$vment,
|
|
hyp=>$args{hyp},
|
|
};
|
|
$running_tasks{$task}->{task} = $task;
|
|
$running_tasks{$task}->{callback} = \&clone_task_callback;
|
|
$running_tasks{$task}->{hyp} = $args{hyp}; #$hyp_conns->{$hyp};
|
|
$running_tasks{$task}->{vm} = $node; #$hyp_conns->{$hyp};
|
|
}
|
|
}
|
|
|
|
sub make_customization_spec {
|
|
my $node = shift;
|
|
my %args = @_;
|
|
my $password="Passw0rd";
|
|
my $wintimezone;
|
|
#map of number to strings can be found at
|
|
#http://osman-shener-en.blogspot.com/2008/02/unattendedtxt-time-zone-index.html
|
|
my $fullname="Unspecified User";
|
|
my $orgName="Unspecified Organization";
|
|
if ($::XCATSITEVALS{winfullname}) { $fullname = $::XCATSITEVALS{winfullname}; }
|
|
if ($::XCATSITEVALS{winorgname}) { $orgName = $::XCATSITEVALS{winorgname}; }
|
|
my @runonce=(); #to be read in from postscripts table
|
|
$wintimezone=xCAT::TZUtils::get_wintimezonenum();
|
|
my $ptab=xCAT::Table->new('postscripts',-create=>0);
|
|
|
|
if ($ptab) {
|
|
my $psent = $ptab->getNodeAttribs($node,[qw/postscripts postbootscripts/]);
|
|
if ($psent and $psent->{postscripts}) {
|
|
push @runonce,split /,/,$psent->{postscripts};
|
|
}
|
|
if ($psent and $psent->{postbootscripts}) {
|
|
push @runonce,split /,/,$psent->{postbootscripts};
|
|
}
|
|
}
|
|
$ptab = xCAT::Table->new('passwd',-create=>0);
|
|
unless ($ptab) {
|
|
die "passwd table needed";
|
|
}
|
|
my ($passent) = $ptab->getAttribs({"key"=>"system",username=>"Administrator"},'password');
|
|
unless ($passent) {
|
|
die "need passwd table entry for system account Administrator";
|
|
}
|
|
$password=$passent->{password};
|
|
my %lfpd;
|
|
if ($args{ostype} and $args{ostype} =~ /win2k3/) {
|
|
%lfpd = (
|
|
licenseFilePrintData=>CustomizationLicenseFilePrintData->new(
|
|
autoMode=>CustomizationLicenseDataMode->new(
|
|
'perSeat'
|
|
)
|
|
)
|
|
);
|
|
}
|
|
my %runonce;
|
|
if (scalar @runonce) { #skip section if no postscripts or postbootscripts
|
|
%runonce=(
|
|
guiRunOnce=>CustomizationGuiRunOnce->new(
|
|
commandList=>\@runonce,
|
|
)
|
|
);
|
|
}
|
|
my %autologonargs = ( autoLogon=>0, autoLogonCount=>1, );
|
|
if ($args{parameters} and $args{parameters} =~ /autoLogonCount=([^,]*)/i) {
|
|
my $count = $1;
|
|
if ($count) {
|
|
$autologonargs{autoLogon}=1;
|
|
$autologonargs{autoLogonCount}=$count;
|
|
}
|
|
}
|
|
my $identity = CustomizationSysprep->new(
|
|
%runonce,
|
|
%lfpd,
|
|
guiUnattended => CustomizationGuiUnattended->new(
|
|
%autologonargs,
|
|
password=>CustomizationPassword->new(
|
|
plainText=>1,
|
|
value=>$password,
|
|
),
|
|
timeZone=>$wintimezone,
|
|
),
|
|
identification=>get_customizedidentification(),
|
|
userData=>CustomizationUserData->new(
|
|
computerName=>CustomizationFixedName->new(name=>$node),
|
|
fullName=>$fullname,
|
|
orgName=>$orgName,
|
|
productId=>"",
|
|
),
|
|
);
|
|
my $options = CustomizationWinOptions->new(changeSID=>1,deleteAccounts=>0);
|
|
my $customizationspec = CustomizationSpec->new(
|
|
globalIPSettings=>CustomizationGlobalIPSettings->new(),
|
|
identity=>$identity,
|
|
nicSettingMap=>[
|
|
CustomizationAdapterMapping->new(adapter=>CustomizationIPSettings->new(ip=>CustomizationDhcpIpGenerator->new()))
|
|
],
|
|
options=>$options,
|
|
);
|
|
return $customizationspec;
|
|
|
|
}
|
|
|
|
sub get_customizedidentification {
|
|
#for now, just do a 'TBD' workgroup. VMWare not supporting joining without domain admin password is rather unfortunate
|
|
return CustomizationIdentification->new(
|
|
joinWorkgroup=>"TBD",
|
|
);
|
|
}
|
|
|
|
|
|
|
|
|
|
sub get_placement_resources {
|
|
my %args = @_;
|
|
my $pool;
|
|
my $dstore;
|
|
my $hyp = $args{hyp};
|
|
my $cluster = $args{cluster};
|
|
my $destination=$args{destination};
|
|
if ($hyp) {
|
|
unless (defined $hyphash{$hyp}->{pool}) {
|
|
$hyphash{$hyp}->{pool} = $hyphash{$hyp}->{conn}->get_view(mo_ref=>$hyphash{$hyp}->{hostview}->parent,properties=>['resourcePool'])->resourcePool;
|
|
}
|
|
$pool=$hyphash{$hyp}->{pool};
|
|
if ($destination) { $dstore=$hyphash{$hyp}->{datastorerefmap}->{$destination} };
|
|
} else {#clustered...
|
|
unless (defined $clusterhash{$cluster}->{pool}) {
|
|
my $cview = get_clusterview(clustname=>$cluster,conn=>$clusterhash{$cluster}->{conn});
|
|
$clusterhash{$cluster}->{pool}=$cview->resourcePool;
|
|
}
|
|
$pool=$clusterhash{$cluster}->{pool};
|
|
if ($destination) { $dstore=$clusterhash{$cluster}->{datastorerefmap}->{$destination} };
|
|
}
|
|
return {
|
|
pool=>$pool,
|
|
datastore=>$dstore,
|
|
}
|
|
}
|
|
|
|
sub clone_task_callback {
|
|
my $task = shift;
|
|
my $parms = shift;
|
|
my $state = $task->info->state->val;
|
|
my $node = $parms->{node};
|
|
my $conn = $parms->{conn};
|
|
my $intent = $parms->{successtext};
|
|
if ($state eq 'success') {
|
|
#xCAT::SvrUtils::sendmsg($intent, $output_handler,$node);
|
|
my $nodetype=xCAT::Table->new('nodetype',-create=>1);
|
|
my $vm=xCAT::Table->new('vm',-create=>1);
|
|
$vm->setAttribs({node=>$node},$parms->{vment});
|
|
|
|
$nodetype->setAttribs({node=>$node},$parms->{nodetypeent});
|
|
foreach (keys %{$parms->{vment}}) {
|
|
$tablecfg{vm}->{$node}->[0]->{$_}=$parms->{vment}->{$_};
|
|
}
|
|
|
|
my @networks = split /,/,$tablecfg{vm}->{$node}->[0]->{nics};
|
|
my @macs = xCAT::VMCommon::getMacAddresses(\%tablecfg,$node,scalar @networks);
|
|
#now with macs, change all macs in the vm to match our generated macs
|
|
my $regex = qr/^$node(\z|\.)/;
|
|
#have to do an expensive pull of the vm view, since it is brand new
|
|
my $nodeviews = $conn->find_entity_views(view_type => 'VirtualMachine',filter=>{'config.name'=>$regex});
|
|
unless (scalar @$nodeviews == 1) { die "this should be impossible"; }
|
|
my $vpdtab=xCAT::Table->new('vpd',-create=>1);
|
|
$vpdtab->setAttribs({node=>$node},{uuid=>$nodeviews->[0]->config->uuid});
|
|
my $ndev;
|
|
my @devstochange;
|
|
foreach $ndev (@{$nodeviews->[0]->config->hardware->device}) {
|
|
unless ($ndev->{macAddress}) { next; } #not an ndev
|
|
$ndev->{macAddress}=shift @macs;
|
|
$ndev->{addressType}="manual";
|
|
push @devstochange, VirtualDeviceConfigSpec->new(
|
|
device => $ndev,
|
|
operation => VirtualDeviceConfigSpecOperation->new('edit'));
|
|
}
|
|
if (@devstochange) {
|
|
my $reconfigspec = VirtualMachineConfigSpec->new(deviceChange=>\@devstochange);
|
|
my $task = $nodeviews->[0]->ReconfigVM_Task(spec=>$reconfigspec);
|
|
$running_tasks{$task}->{task} = $task;
|
|
$running_tasks{$task}->{callback} = \&generic_task_callback;
|
|
$running_tasks{$task}->{hyp} = $parms->{hyp};
|
|
$running_tasks{$task}->{conn} = $parms->{conn};
|
|
$running_tasks{$task}->{data} = { node => $node, successtext => $intent};
|
|
}
|
|
|
|
|
|
} elsif ($state eq 'error') {
|
|
relay_vmware_err($task,"",$node);
|
|
}
|
|
}
|
|
|
|
sub promote_vm_to_master {
|
|
my %args = @_;
|
|
my $node=$args{node};
|
|
my $hyp=$args{hyp};
|
|
my $cluster=$args{cluster};
|
|
my $regex=qr/^$node(\z|\.)/;
|
|
my $conn;
|
|
if ($hyp) {
|
|
$conn=$hyphash{$hyp}->{conn};
|
|
} else {
|
|
$conn=$clusterhash{$cluster}->{conn};
|
|
}
|
|
my $nodeviews = $conn->find_entity_views(view_type => 'VirtualMachine',filter=>{'config.name'=>$regex});
|
|
if (scalar(@$nodeviews) != 1) {
|
|
xCAT::SvrUtils::sendmsg([1,"Cannot find $node in VMWare infrastructure"], $output_handler,$node);
|
|
return;
|
|
}
|
|
my $nodeview = shift @$nodeviews;
|
|
my $dstore;
|
|
if ($hyp) {
|
|
$dstore=$hyphash{$hyp}->{datastorerefmap}->{$args{url}},
|
|
} else {
|
|
$dstore=$clusterhash{$cluster}->{datastorerefmap}->{$args{url}},
|
|
}
|
|
my $relocatespec = VirtualMachineRelocateSpec->new(
|
|
datastore=>$dstore,
|
|
);
|
|
my $clonespec = VirtualMachineCloneSpec->new(
|
|
location=>$relocatespec,
|
|
template=>0, #can't go straight to template, need to clone, then snap, then templatify
|
|
powerOn=>0
|
|
);
|
|
|
|
my $vmfolder=$vmhash{$node}->{vmfolder};
|
|
my $task = $nodeview->CloneVM_Task(folder=>$vmfolder,name=>$args{mastername},spec=>$clonespec);
|
|
$running_tasks{$task}->{data} = { node => $node, hyp => $args{hyp}, conn => $conn, successtext => 'Successfully copied to '.$args{mastername}, mastername=>$args{mastername}, url=>$args{url} };
|
|
$running_tasks{$task}->{task} = $task;
|
|
$running_tasks{$task}->{callback} = \&promote_task_callback;
|
|
$running_tasks{$task}->{hyp} = $args{hyp}; #$hyp_conns->{$hyp};
|
|
$running_tasks{$task}->{vm}=$node;
|
|
}
|
|
sub promote_task_callback {
|
|
my $task = shift;
|
|
my $parms = shift;
|
|
my $state = $task->info->state->val;
|
|
my $node = $parms->{node};
|
|
my $intent = $parms->{successtext};
|
|
if ($state eq 'success') { #now, we have to make one snapshot for linked clones
|
|
my $mastername=$parms->{mastername};
|
|
my $regex=qr/^$mastername\z/;
|
|
my $masterviews = $parms->{conn}->find_entity_views(view_type => 'VirtualMachine',filter=>{'config.name'=>$regex});
|
|
unless (scalar @$masterviews == 1) {
|
|
die "Impossible";
|
|
}
|
|
my $masterview = $masterviews->[0];
|
|
my $task = $masterview->CreateSnapshot_Task(name=>"xcatsnap",memory=>"false",quiesce=>"false");
|
|
$parms->{masterview}=$masterview;
|
|
$running_tasks{$task}->{data} = $parms;
|
|
$running_tasks{$task}->{task} = $task;
|
|
$running_tasks{$task}->{callback} = \&promotesnap_task_callback;
|
|
$running_tasks{$task}->{hyp} = $parms->{hyp}; #$hyp_conns->{$hyp};
|
|
$running_tasks{$task}->{vm}=$parms->{node};
|
|
#xCAT::SvrUtils::sendmsg($intent, $output_handler,$node);
|
|
} elsif ($state eq 'error') {
|
|
relay_vmware_err($task,"",$node);
|
|
}
|
|
}
|
|
sub promotesnap_task_callback {
|
|
my $task = shift;
|
|
my $parms = shift;
|
|
my $state = $task->info->state->val;
|
|
my $node = $parms->{node};
|
|
my $intent = $parms->{successtext};
|
|
if ($state eq 'success') {
|
|
$parms->{masterview}->MarkAsTemplate; #time to be a template
|
|
xCAT::SvrUtils::sendmsg($intent, $output_handler,$node);
|
|
my $mastertabentry = {
|
|
originator=>$requester,
|
|
vintage=>scalar(localtime),
|
|
storage=>$parms->{url},
|
|
};
|
|
foreach (qw/os arch profile/) {
|
|
if (defined ($tablecfg{nodetype}->{$node}->[0]->{$_})) {
|
|
$mastertabentry->{$_}=$tablecfg{nodetype}->{$node}->[0]->{$_};
|
|
}
|
|
}
|
|
foreach (qw/storagemodel nics/) {
|
|
if (defined ($tablecfg{vm}->{$node}->[0]->{$_})) {
|
|
$mastertabentry->{$_}=$tablecfg{vm}->{$node}->[0]->{$_};
|
|
}
|
|
}
|
|
my $vmmastertab=xCAT::Table->new('vmmaster',-create=>1);
|
|
my $date=scalar(localtime);
|
|
$vmmastertab->setAttribs({name=>$parms->{mastername}},$mastertabentry);
|
|
} elsif ($state eq 'error') {
|
|
relay_vmware_err($task,"",$node);
|
|
}
|
|
}
|
|
sub mkvms {
|
|
my %args = @_;
|
|
my $nodes = $args{nodes};
|
|
my $hyp = $args{hyp};
|
|
my $cluster = $args{cluster};
|
|
@ARGV = @{$args{exargs}}; #for getoptions;
|
|
my $disksize;
|
|
require Getopt::Long;
|
|
my $cpuCount;
|
|
my $memory;
|
|
GetOptions(
|
|
'size|s=s' => \$disksize,
|
|
"cpus=s" => \$cpuCount,
|
|
"mem=s" => \$memory
|
|
);
|
|
my $node;
|
|
my $conn;
|
|
if ($hyp) {
|
|
$hyphash{$hyp}->{hostview} = get_hostview(hypname=>$hyp,conn=>$hyphash{$hyp}->{conn}); #,properties=>['config','configManager']);
|
|
unless (validate_datastore_prereqs($nodes,$hyp)) {
|
|
return;
|
|
}
|
|
$conn=$hyphash{$hyp}->{conn};
|
|
} else {
|
|
refreshclusterdatastoremap($cluster);
|
|
$conn=$clusterhash{$cluster}->{conn};
|
|
}
|
|
sortoutdatacenters(nodes=>$nodes,hyp=>$hyp,cluster=>$cluster);
|
|
my $placement_resources=get_placement_resources(hyp=>$hyp,cluster=>$cluster);
|
|
#$hyphash{$hyp}->{pool} = $hyphash{$hyp}->{conn}->get_view(mo_ref=>$hyphash{$hyp}->{hostview}->parent,properties=>['resourcePool'])->resourcePool;
|
|
my $cfg;
|
|
foreach $node (@$nodes) {
|
|
process_tasks; #check for tasks needing followup actions before the task is forgotten (VMWare's memory is fairly short at times
|
|
if ($conn->find_entity_view(view_type=>"VirtualMachine",filter=>{name=>$node})) {
|
|
xCAT::SvrUtils::sendmsg([1,"Virtual Machine already exists"], $output_handler,$node);
|
|
next;
|
|
} else {
|
|
register_vm($hyp,$node,$disksize,undef,undef,undef,cpus=>$cpuCount,memory=>$memory,cluster=>$cluster);
|
|
}
|
|
}
|
|
my @dhcpnodes;
|
|
foreach (keys %{$tablecfg{dhcpneeded}}) {
|
|
push @dhcpnodes,$_;
|
|
delete $tablecfg{dhcpneeded}->{$_};
|
|
}
|
|
unless ($::XCATSITEVALS{'dhcpsetup'} and ($::XCATSITEVALS{'dhcpsetup'} =~ /^n/i or $::XCATSITEVALS{'dhcpsetup'} =~ /^d/i or $::XCATSITEVALS{'dhcpsetup'} eq '0')) {
|
|
$executerequest->({command=>['makedhcp'],node=>\@dhcpnodes});
|
|
}
|
|
}
|
|
|
|
sub setboot {
|
|
my %args = @_;
|
|
my $node = $args{node};
|
|
my $hyp = $args{hyp};
|
|
if (not defined $args{vmview}) { #attempt one refresh
|
|
$args{vmview} = $vmhash{$node}->{conn}->find_entity_view(view_type => 'VirtualMachine',properties=>['config.name'],filter=>{name=>$node});
|
|
if (not defined $args{vmview}) {
|
|
xCAT::SvrUtils::sendmsg([1,"VM does not appear to exist"], $output_handler,$node);
|
|
return;
|
|
}
|
|
}
|
|
my $bootorder = ${$args{exargs}}[0];
|
|
#NOTE: VMware simply does not currently seem to allow programatically changing the boot
|
|
#order like other virtualization solutions supported by xCAT.
|
|
#This doesn't behave quite like any existing mechanism:
|
|
#vm.bootorder was meant to take the place of system nvram, vmware imitates that unfortunate aspect of bare metal too well..
|
|
#rsetboot was created to describe the ipmi scenario of a transient boot device, this is persistant *except* for setup, which is not
|
|
#rbootseq was meant to be entirely persistant and ordered.
|
|
#rsetboot is picked, the usage scenario matches about as good as I could think of
|
|
my $reconfigspec;
|
|
if ($bootorder =~ /setup/) {
|
|
unless ($bootorder eq 'setup') {
|
|
xCAT::SvrUtils::sendmsg([1,"rsetboot parameter may not contain 'setup' with other items, assuming vm.bootorder is just 'setup'"], $output_handler,$node);
|
|
}
|
|
$reconfigspec = VirtualMachineConfigSpec->new(
|
|
bootOptions=>VirtualMachineBootOptions->new(enterBIOSSetup=>1),
|
|
);
|
|
} else {
|
|
$bootorder = "allow:".$bootorder;
|
|
$reconfigspec = VirtualMachineConfigSpec->new(
|
|
bootOptions=>VirtualMachineBootOptions->new(enterBIOSSetup=>0),
|
|
extraConfig => [OptionValue->new(key => 'bios.bootDeviceClasses',value=>$bootorder)]
|
|
);
|
|
}
|
|
my $task = $args{vmview}->ReconfigVM_Task(spec=>$reconfigspec);
|
|
$running_tasks{$task}->{task} = $task;
|
|
$running_tasks{$task}->{callback} = \&generic_task_callback;
|
|
$running_tasks{$task}->{hyp} = $args{hyp};
|
|
$running_tasks{$task}->{data} = { node => $node, successtext => ${$args{exargs}}[0] };
|
|
}
|
|
sub register_vm {#Attempt to register existing instance of a VM
|
|
my $hyp = shift;
|
|
my $node = shift;
|
|
my $disksize = shift;
|
|
my $blockedfun = shift; #a pointer to a blocked function to call on success
|
|
my $blockedargs = shift; #hash reference to call blocked function with
|
|
my $failonerr = shift;
|
|
my %args=@_; #ok, went overboard with positional arguments, from now on, named arguments
|
|
my $task;
|
|
if ($hyp) {
|
|
validate_network_prereqs([keys %{$hyphash{$hyp}->{nodes}}],$hyp);
|
|
unless (defined $hyphash{$hyp}->{datastoremap} or validate_datastore_prereqs([keys %{$hyphash{$hyp}->{nodes}}],$hyp)) {
|
|
die "unexpected condition";
|
|
}
|
|
} else {
|
|
scan_cluster_networks($args{cluster});
|
|
}
|
|
|
|
sortoutdatacenters(nodes=>[$node],hyp=>$hyp,cluster=>$args{cluster});
|
|
my $placement_resources=get_placement_resources(hyp=>$hyp,cluster=>$args{cluster});
|
|
|
|
# Try to add an existing VM to the machine folder
|
|
my $success = eval {
|
|
if ($hyp) {
|
|
$task = $vmhash{$node}->{vmfolder}->RegisterVM_Task(path=>getcfgdatastore($node,$hyphash{$hyp}->{datastoremap})." /$node/$node.vmx",name=>$node,pool=>$hyphash{$hyp}->{pool},host=>$hyphash{$hyp}->{hostview},asTemplate=>0);
|
|
} else {
|
|
$task = $vmhash{$node}->{vmfolder}->RegisterVM_Task(path=>getcfgdatastore($node,$clusterhash{$args{cluster}}->{datastoremap})." /$node/$node.vmx",name=>$node,pool=>$placement_resources->{pool},asTemplate=>0);
|
|
}
|
|
};
|
|
# if we couldn't add it then it means it wasn't created yet. So we create it.
|
|
my $cluster=$args{cluster};
|
|
if ($@ or not $success) {
|
|
#if (ref($@) eq 'SoapFault') {
|
|
# if (ref($@->detail) eq 'NotFound') {
|
|
register_vm_callback(undef, {
|
|
node => $node,
|
|
disksize => $disksize,
|
|
blockedfun => $blockedfun,
|
|
blockedargs => $blockedargs,
|
|
errregister=>$failonerr,
|
|
cpus=>$args{cpus},
|
|
memory=>$args{memory},
|
|
hyp => $hyp,
|
|
cluster=>$cluster,
|
|
});
|
|
}
|
|
if ($task) {
|
|
$running_tasks{$task}->{task} = $task;
|
|
$running_tasks{$task}->{callback} = \®ister_vm_callback;
|
|
$running_tasks{$task}->{hyp} = $hyp;
|
|
$running_tasks{$task}->{cluster} = $cluster;
|
|
$running_tasks{$task}->{data} = {
|
|
node => $node,
|
|
disksize => $disksize,
|
|
blockedfun => $blockedfun,
|
|
blockedargs => $blockedargs,
|
|
errregister=>$failonerr,
|
|
cpus=>$args{cpus},
|
|
memory=>$args{memory},
|
|
hyp => $hyp,
|
|
cluster=>$cluster,
|
|
};
|
|
}
|
|
}
|
|
|
|
sub register_vm_callback {
|
|
my $task = shift;
|
|
my $args = shift;
|
|
if (not $task or $task->info->state->val eq 'error') { #TODO: fail for 'rpower' flow, mkvm is too invasive in VMWare to be induced by 'rpower on'
|
|
if (not defined $args->{blockedfun}) {
|
|
mknewvm($args->{node},$args->{disksize},$args->{hyp},$args);
|
|
} elsif ($args->{errregister}) {
|
|
relay_vmware_err($task,"",$args->{node});
|
|
} else {
|
|
xCAT::SvrUtils::sendmsg([1,"mkvm must be called before use of this function"], $output_handler,$args->{node});
|
|
}
|
|
} elsif (defined $args->{blockedfun}) { #If there is a blocked function, call it here)
|
|
$args->{blockedfun}->(%{$args->{blockedargs}});
|
|
}
|
|
}
|
|
|
|
|
|
sub getURI {
|
|
my $method = shift;
|
|
my $location = shift;
|
|
my $uri = '';
|
|
|
|
if($method =~ /nfs/){
|
|
|
|
(my $server,my $path) = split/\//,$location,2;
|
|
$server =~ s/:$//; #tolerate habitual colons
|
|
my $servern = inet_aton($server);
|
|
unless ($servern) {
|
|
xCAT::SvrUtils::sendmsg([1,"could not resolve '$server' to an address from vm.storage/vm.cfgstore"], $output_handler);
|
|
}
|
|
$server = inet_ntoa($servern);
|
|
$uri = "nfs://$server/$path";
|
|
}elsif($method =~ /vmfs/){
|
|
(my $name, undef) = split /\//,$location,2;
|
|
$name =~ s/:$//; #remove a : if someone put it in for some reason.
|
|
$uri = "vmfs://$name";
|
|
}else{
|
|
xCAT::SvrUtils::sendmsg([1,"Unsupported VMware Storage Method: $method. Please use 'vmfs or nfs'"], $output_handler);
|
|
}
|
|
|
|
return $uri;
|
|
}
|
|
|
|
|
|
sub getcfgdatastore {
|
|
my $node = shift;
|
|
my $dses = shift;
|
|
my $cfgdatastore = $tablecfg{vm}->{$node}->[0]->{cfgstore};
|
|
unless ($cfgdatastore) {
|
|
$cfgdatastore = $tablecfg{vm}->{$node}->[0]->{storage};
|
|
#TODO: if multiple drives are specified, make sure to split this out
|
|
#DONE: I believe the regex after this conditional takes care of that case already..
|
|
}
|
|
$cfgdatastore =~ s/=.*//;
|
|
(my $method,my $location) = split /:\/\//,$cfgdatastore,2;
|
|
my $uri = $cfgdatastore;
|
|
unless ($dses->{$uri}) { #don't call getURI if map works out fine already
|
|
$uri = getURI($method,$location);
|
|
}
|
|
$cfgdatastore = "[".$dses->{$uri}."]";
|
|
#$cfgdatastore =~ s/,.*$//; #these two lines of code were kinda pointless
|
|
#$cfgdatastore =~ s/\/$//;
|
|
return $cfgdatastore;
|
|
}
|
|
|
|
|
|
sub mknewvm {
|
|
my $node=shift;
|
|
my $disksize=shift;
|
|
my $hyp=shift;
|
|
my $otherargs=shift;
|
|
my $cluster=$otherargs->{cluster};
|
|
my $placement_resources=get_placement_resources(hyp=>$hyp,cluster=>$cluster);
|
|
my $pool=$placement_resources->{pool};
|
|
my $cfg;
|
|
if ($hyp) {
|
|
$cfg = build_cfgspec($node,$hyphash{$hyp}->{datastoremap},$hyphash{$hyp}->{nets},$disksize,$hyp,$otherargs);
|
|
} else { #cluster based..
|
|
$cfg = build_cfgspec($node,$clusterhash{$cluster}->{datastoremap},$clusterhash{$cluster}->{nets},$disksize,$hyp,$otherargs);
|
|
}
|
|
my $task;
|
|
if ($hyp) {
|
|
$task = $vmhash{$node}->{vmfolder}->CreateVM_Task(config=>$cfg,pool=>$hyphash{$hyp}->{pool},host=>$hyphash{$hyp}->{hostview});
|
|
} else {
|
|
$task = $vmhash{$node}->{vmfolder}->CreateVM_Task(config=>$cfg,pool=>$pool); #drs away
|
|
}
|
|
$running_tasks{$task}->{task} = $task;
|
|
$running_tasks{$task}->{callback} = \&mkvm_callback;
|
|
$running_tasks{$task}->{hyp} = $hyp;
|
|
$running_tasks{$task}->{cluster} = $cluster;
|
|
$running_tasks{$task}->{data} = { hyp=>$hyp, cluster=>$cluster, node => $node };
|
|
}
|
|
|
|
|
|
sub getUnits {
|
|
my $amount = shift;
|
|
my $defunit = shift;
|
|
my $divisor=shift;
|
|
unless ($amount) { return; }
|
|
unless ($divisor) {
|
|
$divisor = 1;
|
|
}
|
|
if ($amount =~ /(\D)$/) { #If unitless, add unit
|
|
$defunit=$1;
|
|
chop $amount;
|
|
}
|
|
if ($defunit =~ /k/i) {
|
|
return $amount*1024/$divisor;
|
|
} elsif ($defunit =~ /m/i) {
|
|
return $amount*1048576/$divisor;
|
|
} elsif ($defunit =~ /g/i) {
|
|
return $amount*1073741824/$divisor;
|
|
}
|
|
}
|
|
|
|
|
|
sub getguestid {
|
|
my $osfound=0;
|
|
my $node = shift;
|
|
if ($tablecfg{vm}->{$node}->[0]->{guestostype}) { #if admin wants to skip derivation from nodetype.os value, let em
|
|
return $tablecfg{vm}->{$node}->[0]->{guestostype};
|
|
}
|
|
my $nodeos = $tablecfg{nodetype}->{$node}->[0]->{os};
|
|
my $nodearch = $tablecfg{nodetype}->{$node}->[0]->{arch};
|
|
foreach (keys %guestidmap) {
|
|
if (defined($nodeos) and $nodeos =~ /$_/) {
|
|
if ($nodearch eq 'x86_64' and $_ !~ /vmkernel/) {
|
|
$nodeos=$guestidmap{$_}."64Guest";
|
|
} else {
|
|
$nodeos=$guestidmap{$_};
|
|
$nodeos =~ s/_$//;
|
|
$nodeos .= "Guest";
|
|
}
|
|
$osfound=1;
|
|
last;
|
|
}
|
|
}
|
|
unless ($osfound) {
|
|
if (defined($nodearch) and $nodearch eq 'x86_64') {
|
|
$nodeos="otherGuest64";
|
|
} else {
|
|
$nodeos="otherGuest";
|
|
}
|
|
}
|
|
return $nodeos;
|
|
}
|
|
|
|
sub build_cfgspec {
|
|
my $node = shift;
|
|
my $dses = shift; #map to match vm table to datastore names
|
|
my $netmap = shift;
|
|
my $disksize = shift;
|
|
my $hyp = shift;
|
|
my $otherargs=shift;
|
|
my $memory;
|
|
my $ncpus;
|
|
my $updatehash;
|
|
if ($otherargs->{memory}) {
|
|
$memory=getUnits($otherargs->{memory},"M",1048576);
|
|
if ($tablecfg{vm}->{$node}->[0]->{memory}) {
|
|
$updatehash->{memory}=$memory;
|
|
}
|
|
} elsif ($tablecfg{vm}->{$node}->[0]->{memory}) {
|
|
$memory = getUnits($tablecfg{vm}->{$node}->[0]->{memory},"M",1048576);
|
|
} else {
|
|
$memory = 512;
|
|
}
|
|
if ($otherargs->{cpus}) {
|
|
$ncpus=$otherargs->{cpus};
|
|
if ($tablecfg{vm}->{$node}->[0]->{cpus}) {
|
|
$updatehash->{cpus}=$ncpus;
|
|
}
|
|
} elsif ($tablecfg{vm}->{$node}->[0]->{cpus}) {
|
|
$ncpus = $tablecfg{vm}->{$node}->[0]->{cpus};
|
|
} else {
|
|
$ncpus = 1;
|
|
}
|
|
if ($updatehash) {
|
|
my $vmtab = xCAT::Table->new('vm',-create=>1);
|
|
$vmtab->setNodeAttribs($node,$updatehash);
|
|
}
|
|
my @devices;
|
|
$currkey=0;
|
|
my $opticalbacking = VirtualCdromRemoteAtapiBackingInfo->new(deviceName=>"");
|
|
my $opticalconnectable = VirtualDeviceConnectInfo->new(startConnected=>0,allowGuestControl=>1,connected=>0);
|
|
my $optical =VirtualCdrom->new( controllerKey => 201,
|
|
connectable=>$opticalconnectable,
|
|
backing=>$opticalbacking,
|
|
key => $currkey++,
|
|
unitNumber => 0, );
|
|
push @devices,VirtualDeviceConfigSpec->new(device => $optical, operation => VirtualDeviceConfigSpecOperation->new('add'));
|
|
push @devices,create_storage_devs($node,$dses,$disksize);
|
|
push @devices,create_nic_devs($node,$netmap,$hyp);
|
|
#my $cfgdatastore = $tablecfg{vm}->{$node}->[0]->{storage}; #TODO: need a new cfglocation field in case of stateless guest?
|
|
#$cfgdatastore =~ s/,.*$//;
|
|
#$cfgdatastore =~ s/\/$//;
|
|
#$cfgdatastore = "[".$dses->{$cfgdatastore}."]";
|
|
my $cfgdatastore = getcfgdatastore($node,$dses);
|
|
my $vfiles = VirtualMachineFileInfo->new(vmPathName=>$cfgdatastore);
|
|
#my $nodeos = $tablecfg{nodetype}->{$node}->[0]->{os};
|
|
#my $nodearch = $tablecfg{nodetype}->{$node}->[0]->{arch};
|
|
my $nodeos = getguestid($node); #nodeos=>$nodeos,nodearch=>$nodearch);
|
|
my $uuid;
|
|
if ($tablecfg{vpd}->{$node}->[0]->{uuid}) {
|
|
$uuid = $tablecfg{vpd}->{$node}->[0]->{uuid};
|
|
} else {
|
|
if ($tablecfg{mac}->{$node}->[0]->{mac}) { #a uuidv1 is possible, generate that for absolute uniqueness guarantee
|
|
my $mac = $tablecfg{mac}->{$node}->[0]->{mac};
|
|
$mac =~ s/\|.*//;
|
|
$mac =~ s/!.*//;
|
|
$uuid=xCAT::Utils::genUUID(mac=>$mac);
|
|
} else {
|
|
$uuid=xCAT::Utils::genUUID();
|
|
}
|
|
|
|
my $vpdtab = xCAT::Table->new('vpd');
|
|
$vpdtab->setNodeAttribs($node,{uuid=>$uuid});
|
|
}
|
|
$uuid =~ s/^(..)(..)(..)(..)-(..)(..)-(..)(..)/$4$3$2$1-$6$5-$8$7/;
|
|
my @optionvals;
|
|
if ($tablecfg{vm}->{$node}->[0]->{othersettings}) {
|
|
my $key;
|
|
my $value;
|
|
foreach (split /;/,$tablecfg{vm}->{$node}->[0]->{othersettings}) {
|
|
($key,$value)=split /=/;
|
|
if ($value) {
|
|
push @optionvals,OptionValue->new(key=>$key,value=>$value);
|
|
} else {
|
|
push @optionvals,OptionValue->new(key=>$key);
|
|
}
|
|
}
|
|
}
|
|
my %specargs = (
|
|
name => $node,
|
|
files => $vfiles,
|
|
guestId=>$nodeos,
|
|
memoryMB => $memory,
|
|
numCPUs => $ncpus,
|
|
deviceChange => \@devices,
|
|
uuid=>$uuid,
|
|
);
|
|
if (@optionvals) {
|
|
$specargs{extraConfig}=\@optionvals;
|
|
}
|
|
return VirtualMachineConfigSpec->new(%specargs);
|
|
}
|
|
|
|
sub create_nic_devs {
|
|
my $node = shift;
|
|
my $netmap = shift;
|
|
my $hyp = shift;
|
|
my @networks = split /,/,$tablecfg{vm}->{$node}->[0]->{nics};
|
|
my @devs;
|
|
my $idx = 0;
|
|
my @macs = xCAT::VMCommon::getMacAddresses(\%tablecfg,$node,scalar @networks);
|
|
my $connprefs=VirtualDeviceConnectInfo->new(
|
|
allowGuestControl=>1,
|
|
connected=>0,
|
|
startConnected => 1
|
|
);
|
|
my $model=$tablecfg{vm}->{$node}->[0]->{nicmodel};
|
|
unless ($model) {
|
|
$model='e1000';
|
|
}
|
|
foreach (@networks) {
|
|
my $pgname=$_;
|
|
if ($hyp) {
|
|
$pgname = $hyphash{$hyp}->{pgnames}->{$_};
|
|
}
|
|
s/.*://;
|
|
my $hadspecmodel=0;
|
|
if (m/=/) {
|
|
$hadspecmodel=1;
|
|
s/=(.*)$//;
|
|
}
|
|
my $tmpmodel=$model;
|
|
if ($hadspecmodel) { $tmpmodel=$1; }
|
|
my $netname = $_;
|
|
my $backing = VirtualEthernetCardNetworkBackingInfo->new(
|
|
network => $netmap->{$pgname},
|
|
deviceName=>$pgname,
|
|
);
|
|
my %newcardargs=(
|
|
key=>0,#3, #$currkey++,
|
|
backing=>$backing,
|
|
addressType=>"manual",
|
|
macAddress=>shift @macs,
|
|
connectable=>$connprefs,
|
|
wakeOnLanEnabled=>1, #TODO: configurable in tables?
|
|
);
|
|
my $newcard;
|
|
if ($tmpmodel eq 'e1000') {
|
|
$newcard=VirtualE1000->new(%newcardargs);
|
|
} elsif ($tmpmodel eq 'vmxnet3') {
|
|
$newcard=VirtualVmxnet3->new(%newcardargs);
|
|
} elsif ($tmpmodel eq 'pcnet32') {
|
|
$newcard=VirtualPCNet32->new(%newcardargs);
|
|
} elsif ($tmpmodel eq 'vmxnet2') {
|
|
$newcard=VirtualVmxnet2->new(%newcardargs);
|
|
} elsif ($tmpmodel eq 'vmxnet') {
|
|
$newcard=VirtualVmxnet->new(%newcardargs);
|
|
} else {
|
|
xCAT::SvrUtils::sendmsg([1,"$tmpmodel not a recognized nic type, falling back to e1000 (vmxnet3, e1000, pcnet32, vmxnet2, vmxnet are recognized"], $output_handler,$node);
|
|
$newcard=VirtualE1000->new(%newcardargs);
|
|
}
|
|
|
|
push @devs,VirtualDeviceConfigSpec->new(device => $newcard,
|
|
operation => VirtualDeviceConfigSpecOperation->new('add'));
|
|
$idx++;
|
|
}
|
|
return @devs;
|
|
die "Stop running for test";
|
|
}
|
|
|
|
sub create_storage_devs {
|
|
my $node = shift;
|
|
my $sdmap = shift;
|
|
my $sizes = shift;
|
|
my @sizes = split /[,:]/, $sizes;
|
|
my $existingScsiCont = shift;
|
|
my $scsiUnit = shift;
|
|
my $existingIdeCont = shift;
|
|
my $ideUnit = shift;
|
|
my $devices = shift;
|
|
my %args=@_;
|
|
my $scsicontrollerkey=0;
|
|
my $idecontrollerkey=200; #IDE 'controllers' exist at 200 and 201 invariably, with no flexibility?
|
|
#Cannot find documentation that declares this absolute, but attempts to do otherwise
|
|
#lead in failure, also of note, these are single-channel controllers, so two devs per controller
|
|
|
|
my $backingif;
|
|
my @devs;
|
|
my $havescsidevs =0;
|
|
my $disktype = 'ide';
|
|
my $ideunitnum=0;
|
|
my $scsiunitnum=0;
|
|
my $havescsicontroller=0;
|
|
my %usedideunits;
|
|
my %usedscsiunits=(7=>1,'7'=>1);
|
|
if (defined $existingScsiCont) {
|
|
$havescsicontroller=1;
|
|
$scsicontrollerkey = $existingScsiCont->{key};
|
|
$scsiunitnum = $scsiUnit;
|
|
%usedscsiunits = %{getUsedUnits($scsicontrollerkey,$devices)};
|
|
}
|
|
if (defined $existingIdeCont) {
|
|
$idecontrollerkey = $existingIdeCont->{key};
|
|
$ideunitnum = $ideUnit;
|
|
%usedideunits = %{getUsedUnits($idecontrollerkey,$devices)};
|
|
}
|
|
my $unitnum;
|
|
my %disktocont;
|
|
my $dev;
|
|
my @storelocs = split /,/,$tablecfg{vm}->{$node}->[0]->{storage};
|
|
my $globaldisktype = $tablecfg{vm}->{$node}->[0]->{storagemodel};
|
|
unless ($globaldisktype) { $globaldisktype='ide'; }
|
|
#number of devices is the larger of the specified sizes (TODO: masters) or storage pools to span
|
|
my $numdevs = (scalar @storelocs > scalar @sizes ? scalar @storelocs : scalar @sizes);
|
|
my $controllertype='scsi';
|
|
while ($numdevs-- > 0) {
|
|
my $storeloc = shift @storelocs;
|
|
unless (scalar @storelocs) { @storelocs = ($storeloc); } #allow reuse of one cfg specified pool for multiple devs
|
|
my $disksize = shift @sizes;
|
|
unless (scalar @sizes) { @sizes = ($disksize); } #if we emptied the array, stick the last entry back on to allow it to specify all remaining disks
|
|
$disksize = getUnits($disksize,'G',1024);
|
|
$disktype = $globaldisktype;
|
|
if ($storeloc =~ /=/) {
|
|
($storeloc,$disktype) = split /=/,$storeloc;
|
|
}
|
|
if ($disktype eq 'ide' and $args{idefull}) {
|
|
xCAT::SvrUtils::sendmsg([1,"VM is at capacity for IDE devices, a drive was not added"], $output_handler,$node);
|
|
return;
|
|
} elsif (($disktype eq 'scsi' or $disktype eq 'sas' or $disktype eq 'pvscsi') and $args{scsifull}) {
|
|
xCAT::SvrUtils::sendmsg([1,"SCSI Controller at capacity, a drive was not added"], $output_handler,$node);
|
|
return;
|
|
}
|
|
|
|
$storeloc =~ s/\/$//;
|
|
(my $method,my $location) = split /:\/\//,$storeloc,2;
|
|
my $uri = $storeloc;
|
|
unless ($sdmap->{$uri}) { #don't call getURI if map works out fine already
|
|
$uri = getURI($method,$location);
|
|
}
|
|
#(my $server,my $path) = split/\//,$location,2;
|
|
#$server =~ s/:$//; #tolerate habitual colons
|
|
#my $servern = inet_aton($server);
|
|
#unless ($servern) {
|
|
# xCAT::SvrUtils::sendmsg([1,"could not resolve '$server' to an address from vm.storage"]);
|
|
# return;
|
|
#}
|
|
#$server = inet_ntoa($servern);
|
|
#my $uri = "nfs://$server/$path";
|
|
$backingif = VirtualDiskFlatVer2BackingInfo->new(diskMode => 'persistent',
|
|
thinProvisioned => 1,
|
|
fileName => "[".$sdmap->{$uri}."]");
|
|
if ($disktype eq 'ide' and $idecontrollerkey == 1 and $ideunitnum == 0) { #reserve a spot for CD
|
|
$ideunitnum = 1;
|
|
} elsif ($disktype eq 'ide' and $ideunitnum == 2) { #go from current to next ide 'controller'
|
|
$idecontrollerkey++;
|
|
$ideunitnum=0;
|
|
}
|
|
unless ($disktype eq 'ide') {
|
|
push @{$disktocont{$scsicontrollerkey}},$currkey;
|
|
}
|
|
my $controllerkey;
|
|
if ($disktype eq 'ide') {
|
|
$controllerkey = $idecontrollerkey;
|
|
$unitnum = 0;
|
|
while ($usedideunits{$unitnum}) {
|
|
$unitnum++;
|
|
}
|
|
if ($unitnum == 2) {
|
|
$idecontrollerkey++;
|
|
$ideunitnum=1;
|
|
$unitnum=1;
|
|
$controllerkey=$idecontrollerkey;
|
|
}
|
|
$usedideunits{$unitnum}=1;
|
|
} else {
|
|
$controllertype=$disktype;
|
|
$controllerkey = $scsicontrollerkey;
|
|
$unitnum = 0;
|
|
while ($usedscsiunits{$unitnum}) {
|
|
$unitnum++;
|
|
}
|
|
$usedscsiunits{$unitnum}=1;
|
|
$havescsidevs=1;
|
|
}
|
|
|
|
$dev =VirtualDisk->new(backing=>$backingif,
|
|
controllerKey => $controllerkey,
|
|
key => $currkey++,
|
|
unitNumber => $unitnum,
|
|
capacityInKB => $disksize);
|
|
push @devs,VirtualDeviceConfigSpec->new(device => $dev,
|
|
fileOperation => VirtualDeviceConfigSpecFileOperation->new('create'),
|
|
operation => VirtualDeviceConfigSpecOperation->new('add'));
|
|
}
|
|
|
|
#It *seems* that IDE controllers are not subject to require creation, so we skip it
|
|
if ($havescsidevs and not $havescsicontroller) { #need controllers to attach the disks to
|
|
foreach(0..$scsicontrollerkey) {
|
|
if ($controllertype eq 'scsi') {
|
|
$dev=VirtualLsiLogicController->new(key => $_,
|
|
device => \@{$disktocont{$_}},
|
|
sharedBus => VirtualSCSISharing->new('noSharing'),
|
|
busNumber => $_);
|
|
} elsif ($controllertype eq 'sas') {
|
|
$dev=VirtualLsiLogicSASController->new(key => $_,
|
|
device => \@{$disktocont{$_}},
|
|
sharedBus => VirtualSCSISharing->new('noSharing'),
|
|
busNumber => $_);
|
|
} elsif ($controllertype eq 'pvscsi') {
|
|
$dev=ParaVirtualSCSIController->new(key => $_,
|
|
device => \@{$disktocont{$_}},
|
|
sharedBus => VirtualSCSISharing->new('noSharing'),
|
|
busNumber => $_);
|
|
}
|
|
|
|
push @devs,VirtualDeviceConfigSpec->new(device => $dev,
|
|
operation => VirtualDeviceConfigSpecOperation->new('add'));
|
|
|
|
}
|
|
}
|
|
return @devs;
|
|
# my $ctlr = VirtualIDEController->new(
|
|
}
|
|
|
|
sub declare_ready {
|
|
my %args = %{shift()};
|
|
$hypready{$args{hyp}}=1;
|
|
}
|
|
|
|
sub populate_vcenter_hostviews {
|
|
my $vcenter = shift;
|
|
my @hypervisors;
|
|
my %nametohypmap;
|
|
my $iterations=1;
|
|
if ($usehostnamesforvcenter and $usehostnamesforvcenter !~ /no/i) {
|
|
$iterations=2; #two passes possible
|
|
my $hyp;
|
|
foreach $hyp (keys %{$vcenterhash{$vcenter}->{allhyps}}) {
|
|
|
|
if ($tablecfg{hosts}->{$hyp}->[0]->{hostnames}) {
|
|
$nametohypmap{$tablecfg{hosts}->{$hyp}->[0]->{hostnames}}=$hyp;
|
|
}
|
|
}
|
|
@hypervisors = keys %nametohypmap;
|
|
} else {
|
|
@hypervisors = keys %{$vcenterhash{$vcenter}->{allhyps}};
|
|
}
|
|
while ($iterations and scalar(@hypervisors)) {
|
|
my $hosts = join(")|(",@hypervisors);
|
|
$hosts = '^(('.$hosts.'))(\z|\.)';
|
|
my $search = qr/$hosts/;
|
|
my @hypviews = @{$vcenterhash{$vcenter}->{conn}->find_entity_views(view_type=>'HostSystem',properties=>['summary.config.name','summary.runtime.connectionState','runtime.inMaintenanceMode','parent','configManager','summary.host'],filter=>{'summary.config.name'=>$search})};
|
|
foreach (@hypviews) {
|
|
my $hypname = $_->{'summary.config.name'};
|
|
my $hypv=$_;
|
|
my $hyp;
|
|
if ($vcenterhash{$vcenter}->{allhyps}->{$hypname}) { #simplest case, config.name is exactly the same as node name
|
|
$vcenterhash{$vcenter}->{hostviews}->{$hypname} = $_;
|
|
$hyp=$hypname;
|
|
} elsif ($nametohypmap{$hypname}) { #second case, there is a name mapping this to a real name
|
|
$vcenterhash{$vcenter}->{hostviews}->{$nametohypmap{$hypname}} = $_;
|
|
$hyp=$nametohypmap{$hypname};
|
|
} else { #name as-is doesn't work, start stripping domain and hope for the best
|
|
$hypname =~ s/\..*//;
|
|
if ($vcenterhash{$vcenter}->{allhyps}->{$hypname}) { #shortname is a node
|
|
$vcenterhash{$vcenter}->{hostviews}->{$hypname} = $_;
|
|
$hyp=$hypname;
|
|
} elsif ($nametohypmap{$hypname}) { #alias for node
|
|
$vcenterhash{$vcenter}->{hostviews}->{$nametohypmap{$hypname}} = $_;
|
|
$hyp=$nametohypmap{$hypname};
|
|
}
|
|
}
|
|
foreach my $nodename (keys %{$hyphash{$hyp}->{nodes}}) {
|
|
$hostrefbynode{$nodename}=$hypv->{'summary.host'}->value;
|
|
}
|
|
}
|
|
$iterations--;
|
|
@hypervisors=();
|
|
if ($usehostnamesforvcenter and $usehostnamesforvcenter !~ /no/i) { #check for hypervisors by native node name if missed above
|
|
foreach my $hyp (keys %{$vcenterhash{$vcenter}->{allhyps}}) {
|
|
unless ($vcenterhash{$vcenter}->{hostviews}->{$hyp}) {
|
|
push @hypervisors,$hyp;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
sub create_new_cluster {
|
|
my $req = shift;
|
|
@ARGV = @{$req->{arg}};
|
|
my $vcenter;
|
|
my $password;
|
|
my $user;
|
|
my $datacenter;
|
|
GetOptions(
|
|
'vcenter=s' => \$vcenter,
|
|
'password=s' => \$password,
|
|
'datacenter=s' => \$datacenter,
|
|
'username=s' => \$user,
|
|
);
|
|
my $clustername = shift @ARGV;
|
|
my $conn = Vim->new(service_url=>"https://$vcenter/sdk");
|
|
$conn->login(user_name=>$user, password=>$password);
|
|
if ($datacenter) {
|
|
$datacenter = $conn->find_entity_view(view_type => 'Datacenter', properties=>['hostFolder'],filter=>{name=>$datacenter});
|
|
unless ($datacenter) {
|
|
xCAT::SvrUtils::sendmsg([1,": Unable to find requested datacenter"], $output_handler);
|
|
return;
|
|
}
|
|
} else {
|
|
$datacenter = $conn->find_entity_view(view_type => 'Datacenter', properties=>['hostFolder']);
|
|
}
|
|
my $hfolder = $conn->get_view(mo_ref=>$datacenter->hostFolder);
|
|
my $cfgspec = ClusterConfigSpecEx->new();
|
|
$hfolder->CreateClusterEx(name=>$clustername, spec=>$cfgspec);
|
|
}
|
|
sub remove_cluster {
|
|
my $req = shift;
|
|
@ARGV = @{$req->{arg}};
|
|
my $vcenter;
|
|
my $user;
|
|
my $password;
|
|
my $clustername;
|
|
GetOptions(
|
|
'vcenter=s' => \$vcenter,
|
|
'password=s' => \$password,
|
|
'username=s' => \$user,
|
|
);
|
|
$clustername = shift @ARGV;
|
|
my $conn = Vim->new(service_url=>"https://$vcenter/sdk");
|
|
$conn->login(user_name=>$user, password=>$password);
|
|
# $clustview = $hyphash{$hyp}->{vcenter}->{conn}->find_entity_view(view_type => 'Datacenter', properties=>['hostFolder'],filter=>{name=>$tablecfg{hypervisor}->{$hyp}->[0]->{datacenter}});
|
|
#my $conn = Vim->new(service_url=>"https://$vcenter/sdk");
|
|
$conn->login(user_name=>$user, password=>$password);
|
|
my $clustview = $conn->find_entity_view(view_type=> 'ClusterComputeResource', filter=>{name=>$clustername});
|
|
my $task = $clustview->Destroy_Task();
|
|
my $done = 0;
|
|
while (not $done) {
|
|
my $curt = $conn->get_view(mo_ref=>$task);
|
|
my $state = $curt->info->state->val;
|
|
unless ($state eq 'running' or $state eq 'queued') {
|
|
$done = 1;
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
sub list_clusters {
|
|
my $req = shift;
|
|
@ARGV = @{$req->{arg}};
|
|
my $vcenter;
|
|
my $password;
|
|
my $user;
|
|
my $datacenter;
|
|
GetOptions(
|
|
'vcenter=s' => \$vcenter,
|
|
'password=s' => \$password,
|
|
'datacenter=s' => \$datacenter,
|
|
'username=s' => \$user,
|
|
);
|
|
my $clustername = shift @ARGV;
|
|
my $conn = Vim->new(service_url=>"https://$vcenter/sdk");
|
|
$conn->login(user_name=>$user, password=>$password);
|
|
use Data::Dumper;
|
|
my $clustviews = $conn->find_entity_views(view_type=> 'ClusterComputeResource');
|
|
foreach (@$clustviews) {
|
|
xCAT::SvrUtils::sendmsg($_->{name}, $output_handler);
|
|
}
|
|
return;
|
|
}
|
|
|
|
sub validate_vcenter_prereqs { #Communicate with vCenter and ensure this host is added correctly to a vCenter instance when an operation requires it
|
|
my $hyp = shift;
|
|
my $depfun = shift;
|
|
my $depargs = shift;
|
|
my $vcenter = $hyphash{$hyp}->{vcenter}->{name};
|
|
unless ($hyphash{$hyp}->{vcenter}->{conn}) {
|
|
eval {
|
|
$hyphash{$hyp}->{vcenter}->{conn} = Vim->new(service_url=>"https://$vcenter/sdk");
|
|
$hyphash{$hyp}->{vcenter}->{conn}->login(user_name=>$hyphash{$hyp}->{vcenter}->{username},password=>$hyphash{$hyp}->{vcenter}->{password});
|
|
};
|
|
if ($@) {
|
|
$hyphash{$hyp}->{vcenter}->{conn} = undef;
|
|
}
|
|
}
|
|
unless ($hyphash{$hyp}->{vcenter}->{conn}) {
|
|
xCAT::SvrUtils::sendmsg([1,": Unable to reach vCenter server managing $hyp"], $output_handler);
|
|
return undef;
|
|
}
|
|
|
|
|
|
my $foundhyp;
|
|
my $name=$hyp;
|
|
if ($usehostnamesforvcenter and $usehostnamesforvcenter !~ /no/i) {
|
|
if ($tablecfg{hosts}->{$hyp}->[0]->{hostnames}) {
|
|
$name = $tablecfg{hosts}->{$hyp}->[0]->{hostnames};
|
|
}
|
|
}
|
|
my $connspec = HostConnectSpec->new(
|
|
hostName=>$name,
|
|
password=>$hyphash{$hyp}->{password},
|
|
userName=>$hyphash{$hyp}->{username},
|
|
force=>1,
|
|
);
|
|
my $hview;
|
|
$hview = $vcenterhash{$vcenter}->{hostviews}->{$hyp};
|
|
if ($hview) {
|
|
if ($hview->{'summary.config.name'} =~ /^$hyp(?:\.|\z)/ or $hview->{'summary.config.name'} =~ /^$name(?:\.|\z)/) { #Looks good, call the dependent function after declaring the state of vcenter to hypervisor as good
|
|
if ($hview->{'summary.runtime.connectionState'}->val eq 'connected') {
|
|
if ($vcenterautojoin) { #admin has requested manual vcenter management, don't mess with vmotion settings
|
|
enable_vmotion(hypname=>$hyp,hostview=>$hview,conn=>$hyphash{$hyp}->{vcenter}->{conn});
|
|
}
|
|
$vcenterhash{$vcenter}->{goodhyps}->{$hyp} = 1;
|
|
$depfun->($depargs);
|
|
if ($hview->parent->type eq 'ClusterComputeResource') { #if it is in a cluster, we can directly remove it
|
|
$hyphash{$hyp}->{deletionref} = $hview->{mo_ref};
|
|
} elsif ($hview->parent->type eq 'ComputeResource') { #For some reason, we must delete the container instead
|
|
$hyphash{$hyp}->{deletionref} = $hview->{parent}; #save off a reference to delete hostview off just in case
|
|
}
|
|
|
|
|
|
return 1;
|
|
} elsif ($vcenterautojoin or $vcenterforceremove) { #if allowed autojoin and the current view seems corrupt, throw it away and rejoin
|
|
my $ref_to_delete;
|
|
if ($hview->parent->type eq 'ClusterComputeResource') { #We are allowed to specifically kill a host in a cluster
|
|
$ref_to_delete = $hview->{mo_ref};
|
|
} elsif ($hview->parent->type eq 'ComputeResource') { #For some reason, we must delete the container instead
|
|
$ref_to_delete = $hview->{parent};
|
|
}
|
|
my $task = $hyphash{$hyp}->{vcenter}->{conn}->get_view(mo_ref=>$ref_to_delete)->Destroy_Task();
|
|
$running_tasks{$task}->{task} = $task;
|
|
if ($vcenterautojoin) {
|
|
$running_tasks{$task}->{callback} = \&addhosttovcenter;
|
|
} elsif ($vcenterforceremove) {
|
|
$running_tasks{$task}->{callback} = \&delhost_callback;
|
|
}
|
|
$running_tasks{$task}->{conn} = $hyphash{$hyp}->{vcenter}->{conn};
|
|
$running_tasks{$task}->{data} = { depfun => $depfun, depargs => $depargs, conn=> $hyphash{$hyp}->{vcenter}->{conn}, connspec=>$connspec,hostview=>$hview,hypname=>$hyp,vcenter=>$vcenter };
|
|
return undef;
|
|
#The rest would be shorter/ideal, but seems to be confused a lot by stateless
|
|
#Maybe in a future VMWare technology level the following would work better
|
|
#than it does today
|
|
# my $task = $hview_->ReconnectHost_Task(cnxSpec=>$connspec);
|
|
# my $task = $hview->DisconnectHost_Task();
|
|
# $running_tasks{$task}->{task} = $task;
|
|
# $running_tasks{$task}->{callback} = \&disconnecthost_callback;
|
|
# $running_tasks{$task}->{conn} = $hyphash{$hyp}->{vcenter}->{conn};
|
|
# $running_tasks{$task}->{data} = { depfun => $depfun, depargs => $depargs, conn=> $hyphash{$hyp}->{vcenter}->{conn}, connspec=>$connspec,hostview=>$hview,hypname=>$hyp,vcenter=>$vcenter };
|
|
#ADDHOST
|
|
} else {
|
|
if ($hyphash{$hyp}->{offline}) {
|
|
xCAT::SvrUtils::sendmsg(": Failed to communicate with $hyp, vCenter reports it as in inventory but not connected and xCAT is set to not autojoin", $output_handler);
|
|
} else {
|
|
xCAT::SvrUtils::sendmsg([1,": Failed to communicate with $hyp, vCenter reports it as in inventory but not connected and xCAT is set to not autojoin"], $output_handler);
|
|
}
|
|
$hyphash{$hyp}->{conn} = undef;
|
|
return "failed";
|
|
}
|
|
}
|
|
}
|
|
unless ($vcenterautojoin) {
|
|
if ($hyphash{$hyp}->{offline}) {
|
|
xCAT::SvrUtils::sendmsg(": Failed to communicate with $hyp, vCenter does not have it in inventory and xCAT is set to not autojoin", $output_handler);
|
|
} else {
|
|
xCAT::SvrUtils::sendmsg([1,": Failed to communicate with $hyp, vCenter does not have it in inventory and xCAT is set to not autojoin"], $output_handler);
|
|
}
|
|
$hyphash{$hyp}->{conn} = undef;
|
|
return "failed";
|
|
}
|
|
#If still in function, haven't found any likely host entries, make a new one
|
|
unless ($hyphash{$hyp}->{offline}) {
|
|
eval {
|
|
$hyphash{$hyp}->{conn} = Vim->new(service_url=>"https://$hyp/sdk"); #Direct connect to install/check licenses
|
|
$hyphash{$hyp}->{conn}->login(user_name=>$hyphash{$hyp}->{username},password=>$hyphash{$hyp}->{password});
|
|
};
|
|
if ($@) {
|
|
xCAT::SvrUtils::sendmsg([1,": Failed to communicate with $hyp due to $@"], $output_handler);
|
|
$hyphash{$hyp}->{conn} = undef;
|
|
return "failed";
|
|
}
|
|
validate_licenses($hyp);
|
|
}
|
|
addhosttovcenter(undef,{
|
|
depfun => $depfun,
|
|
depargs => $depargs,
|
|
conn=>$hyphash{$hyp}->{vcenter}->{conn},
|
|
connspec=>$connspec,
|
|
hypname=>$hyp,
|
|
vcenter=>$vcenter,
|
|
});
|
|
}
|
|
sub addhosttovcenter {
|
|
my $task = shift;
|
|
my $args = shift;
|
|
my $hyp = $args->{hypname};
|
|
my $depfun = $args->{depfun};
|
|
my $depargs = $args->{depargs};
|
|
my $connspec = $args->{connspec};
|
|
my $vcenter = $args->{vcenter};
|
|
if ($task) {
|
|
my $state = $task->info->state->val;
|
|
if ($state eq 'error') {
|
|
die;
|
|
}
|
|
}
|
|
if ($hyphash{$args->{hypname}}->{offline}) { #let it stay offline
|
|
$hypready{$args->{hypname}}=1; #declare readiness
|
|
#enable_vmotion(hypname=>$args->{hypname},hostview=>$args->{hostview},conn=>$args->{conn});
|
|
$vcenterhash{$args->{vcenter}}->{goodhyps}->{$args->{hypname}} = 1;
|
|
if (defined $args->{depfun}) { #If a function is waiting for the host connect to go valid, call it
|
|
$args->{depfun}->($args->{depargs});
|
|
}
|
|
return;
|
|
}
|
|
if ($tablecfg{hypervisor}->{$hyp}->[0]->{cluster}) {
|
|
my $cluster = get_clusterview(clustname=>$tablecfg{hypervisor}->{$hyp}->[0]->{cluster},conn=>$hyphash{$hyp}->{vcenter}->{conn});
|
|
unless ($cluster) {
|
|
xCAT::SvrUtils::sendmsg([1,$tablecfg{hypervisor}->{$hyp}->[0]->{cluster}. " is not a known cluster to the vCenter server."], $output_handler);
|
|
$hypready{$hyp}=-1; #Declare impossiblility to be ready
|
|
return;
|
|
}
|
|
$task = $cluster->AddHost_Task(spec=>$connspec,asConnected=>1);
|
|
$running_tasks{$task}->{task} = $task;
|
|
$running_tasks{$task}->{callback} = \&connecthost_callback;
|
|
$running_tasks{$task}->{conn} = $hyphash{$hyp}->{vcenter}->{conn};
|
|
$running_tasks{$task}->{data} = { depfun => $depfun, depargs=> $depargs, conn=> $hyphash{$hyp}->{vcenter}->{conn}, connspec=>$connspec, cluster=>$cluster, hypname=>$hyp, vcenter=>$vcenter };
|
|
} else {
|
|
my $datacenter = validate_datacenter_prereqs($hyp);
|
|
unless ($datacenter) { return; }
|
|
my $hfolder = $datacenter->hostFolder; #$hyphash{$hyp}->{vcenter}->{conn}->find_entity_view(view_type=>'Datacenter',properties=>['hostFolder'])->hostFolder;
|
|
$hfolder = $hyphash{$hyp}->{vcenter}->{conn}->get_view(mo_ref=>$hfolder);
|
|
$task = $hfolder->AddStandaloneHost_Task(spec=>$connspec,addConnected=>1);
|
|
$running_tasks{$task}->{task} = $task;
|
|
$running_tasks{$task}->{callback} = \&connecthost_callback;
|
|
$running_tasks{$task}->{conn} = $hyphash{$hyp}->{vcenter}->{conn};
|
|
$running_tasks{$task}->{data} = { depfun => $depfun, depargs=> $depargs, conn=> $hyphash{$hyp}->{vcenter}->{conn}, connspec=>$connspec, foldview=>$hfolder, hypname=>$hyp, vcenter=>$vcenter };
|
|
}
|
|
|
|
#print Dumper @{$hyphash{$hyp}->{vcenter}->{conn}->find_entity_views(view_type=>'HostSystem',properties=>['runtime.connectionState'])};
|
|
}
|
|
|
|
sub validate_datacenter_prereqs {
|
|
my ($hyp) = @_;
|
|
|
|
my $datacenter;
|
|
if ($tablecfg{hypervisor}->{$hyp}->[0]->{datacenter}) {
|
|
$datacenter = $hyphash{$hyp}->{vcenter}->{conn}->find_entity_view(view_type => 'Datacenter', properties=>['hostFolder'],filter=>{name=>$tablecfg{hypervisor}->{$hyp}->[0]->{datacenter}});
|
|
unless ($datacenter) {
|
|
xCAT::SvrUtils::sendmsg([1,": Unable to find requested datacenter (hypervisor.datacenter for $hyp is ".$tablecfg{hypervisor}->{$hyp}->[0]->{datacenter}.")"], $output_handler);
|
|
return;
|
|
}
|
|
} else {
|
|
$datacenter = $hyphash{$hyp}->{vcenter}->{conn}->find_entity_view(view_type => 'Datacenter', properties=>['hostFolder']);
|
|
}
|
|
|
|
if (!defined $datacenter) {
|
|
my $vconn = $hyphash{$hyp}->{vcenter}->{conn};
|
|
my $root_folder = $vconn->get_view(mo_ref=>$vconn->get_service_content()->rootFolder);
|
|
$root_folder->CreateDatacenter(name=>'xcat-datacenter');
|
|
$datacenter = $hyphash{$hyp}->{vcenter}->{conn}->find_entity_view(view_type => 'Datacenter', properties=>['hostFolder']);
|
|
}
|
|
|
|
return $datacenter;
|
|
}
|
|
|
|
|
|
|
|
sub get_default_switch_for_hypervisor {
|
|
#This will make sure the default, implicit switch is in order in accordance
|
|
#with the configuration. If nothing specified, it just spits out vSwitch0
|
|
#if something specified, make sure it exists
|
|
#if it doesn't exist, and the syntax explains how to build it, build it
|
|
#return undef if something is specified, doesn't exist, and lacks instruction
|
|
my $hyp = shift;
|
|
my $defswitch = 'vSwitch0';
|
|
my $switchmembers;
|
|
if ($tablecfg{hypervisor}->{$hyp}->[0]->{defaultnet}) {
|
|
$defswitch = $tablecfg{hypervisor}->{$hyp}->[0]->{defaultnet};
|
|
($defswitch,$switchmembers) = split /=/,$defswitch,2;
|
|
my $vswitch;
|
|
my $hostview = $hyphash{$hyp}->{hostview};
|
|
foreach $vswitch (@{$hostview->config->network->vswitch}) {
|
|
if ($vswitch->name eq $defswitch) {
|
|
return $defswitch;
|
|
}
|
|
}
|
|
#If still here, means we need to build the switch
|
|
unless ($switchmembers) { return undef; } #No hope, no idea how to make it
|
|
return create_vswitch($hyp,$defswitch,split(/&/,$switchmembers));
|
|
} else {
|
|
return 'vSwitch0';
|
|
}
|
|
}
|
|
sub get_switchname_for_portdesc {
|
|
#Thisk function will examine all current switches to find or create a switch to match the described requirement
|
|
my $hyp = shift;
|
|
my $portdesc = shift;
|
|
my $description; #actual name to use for the virtual switch
|
|
if ($tablecfg{hypervisor}->{$hyp}->[0]->{netmap}) {
|
|
foreach (split /,/,$tablecfg{hypervisor}->{$hyp}->[0]->{netmap}) {
|
|
if (/^$portdesc=/) {
|
|
($description,$portdesc) = split /=/,$_,2;
|
|
last;
|
|
}
|
|
}
|
|
} else {
|
|
$description = 'vsw'.$portdesc;
|
|
}
|
|
unless ($description) {
|
|
xCAT::SvrUtils::sendmsg([1,": Invalid format for hypervisor.netmap detected for $hyp"], $output_handler);
|
|
return undef;
|
|
}
|
|
my %requiredports;
|
|
my %portkeys;
|
|
foreach (split /&/,$portdesc) {
|
|
$requiredports{$_}=1;
|
|
}
|
|
|
|
my $hostview = $hyphash{$hyp}->{hostview};
|
|
unless ($hostview) {
|
|
$hyphash{$hyp}->{hostview} = get_hostview(hypname=>$hyp,conn=>$hyphash{$hyp}->{conn}); #,properties=>['config','configManager']); #clustered can't run here, hyphash conn reference good
|
|
$hostview = $hyphash{$hyp}->{hostview};
|
|
}
|
|
foreach (@{$hostview->config->network->pnic}) {
|
|
if ($requiredports{$_->device}) { #We establish lookups both ways
|
|
$portkeys{$_->key}=$_->device;
|
|
delete $requiredports{$_->device};
|
|
}
|
|
}
|
|
if (keys %requiredports) {
|
|
xCAT::SvrUtils::sendmsg([1,":Unable to locate the following nics on $hyp: ".join(',',keys %requiredports)], $output_handler);
|
|
return undef;
|
|
}
|
|
my $foundmatchswitch;
|
|
my $cfgmismatch=0;
|
|
my $vswitch;
|
|
foreach $vswitch (@{$hostview->config->network->vswitch}) {
|
|
$cfgmismatch=0; #new switch, no sign of mismatch
|
|
foreach (@{$vswitch->pnic}) {
|
|
if ($portkeys{$_}) {
|
|
$foundmatchswitch=$vswitch->name;
|
|
delete $requiredports{$portkeys{$_}};
|
|
delete $portkeys{$_};
|
|
} else {
|
|
$cfgmismatch=1; #If this turns out to have anything, it is bad
|
|
}
|
|
}
|
|
if ($foundmatchswitch) { last; }
|
|
}
|
|
if ($foundmatchswitch) {
|
|
if ($cfgmismatch) {
|
|
xCAT::SvrUtils::sendmsg([1,": Aggregation mismatch detected, request nic is aggregated with a nic not requested"], $output_handler);
|
|
return undef;
|
|
}
|
|
unless (keys %portkeys) {
|
|
return $foundmatchswitch;
|
|
}
|
|
die "TODO: add physical nics to aggregation if requested";
|
|
} else {
|
|
return create_vswitch($hyp,$description,values %portkeys);
|
|
}
|
|
die "impossible occurance";
|
|
return undef;
|
|
}
|
|
sub create_vswitch {
|
|
my $hyp = shift;
|
|
my $description = shift;
|
|
my @ports = @_;
|
|
my $vswitch = HostVirtualSwitchBondBridge->new(
|
|
nicDevice=>\@ports
|
|
);
|
|
my $vswspec = HostVirtualSwitchSpec->new(
|
|
bridge=>$vswitch,
|
|
mtu=>1500,
|
|
numPorts=>64
|
|
);
|
|
my $hostview = $hyphash{$hyp}->{hostview};
|
|
my $netman=$hyphash{$hyp}->{conn}->get_view(mo_ref=>$hostview->configManager->networkSystem); #can't run in clustered mode, fine path..
|
|
$netman->AddVirtualSwitch(
|
|
vswitchName=>$description,
|
|
spec=>$vswspec
|
|
);
|
|
return $description;
|
|
}
|
|
|
|
sub scan_cluster_networks {
|
|
my $cluster = shift;
|
|
use Data::Dumper;
|
|
my $conn = $clusterhash{$cluster}->{conn};
|
|
my $cview = get_clusterview(clustname=>$cluster,conn=>$conn);
|
|
if (defined $cview->{network}) {
|
|
foreach (@{$cview->network}) {
|
|
my $nvw = $conn->get_view(mo_ref=>$_);
|
|
if (defined $nvw->name) {
|
|
$clusterhash{$cluster}->{nets}->{$nvw->name}=$_;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
sub fixup_hostportgroup {
|
|
my $vlanspec = shift;
|
|
my $hyp = shift;
|
|
my %args = @_;
|
|
my $action = 'add';
|
|
if ($args{action}) { $action = $args{action} }
|
|
my $hostview = $hyphash{$hyp}->{hostview};
|
|
my $switchsupport = 0;
|
|
eval {
|
|
require xCAT::SwitchHandler;
|
|
$switchsupport = 1;
|
|
};
|
|
my $hypconn = $hyphash{$hyp}->{conn}; #this function can't work in clustered mode anyway, so this is appropriote.
|
|
my $vldata = $vlanspec;
|
|
my $switchname = get_default_switch_for_hypervisor($hyp);
|
|
my $pgname;
|
|
$vldata =~ s/=.*//; #TODO specify nic model with <blah>=model
|
|
if ($vldata =~ /:/) { #The config specifies a particular path in some way
|
|
$vldata =~ s/(.*)://;
|
|
$switchname = get_switchname_for_portdesc($hyp,$1);
|
|
$pgname=$switchname."-".$vldata;
|
|
} else { #Use the default vswitch per table config to connect this through, use the same name we did before to maintain compatibility
|
|
$pgname=$vldata;
|
|
}
|
|
my $netsys;
|
|
$hyphash{$hyp}->{pgnames}->{$vlanspec}=$pgname;
|
|
my $policy = HostNetworkPolicy->new();
|
|
unless ($hyphash{$hyp}->{nets}->{$pgname}) {
|
|
my $vlanid;
|
|
if (looks_like_number($vldata)) {
|
|
$vlanid = $vldata;
|
|
} elsif ($vldata =~ /trunk/) {
|
|
$vlanid=4095;
|
|
} elsif ($vldata =~ /vl(an)?(\d+)$/) {
|
|
$vlanid=$2;
|
|
} else {
|
|
$vlanid = 0;
|
|
}
|
|
if ($vlanid > 0 and $vlanid < 4095 and $switchsupport) {
|
|
my $switchtab = xCAT::Table->new("switch", -create=>0);
|
|
if ($switchtab) {
|
|
my $swent = $switchtab->getNodeAttribs($hyp, [qw/switch port/]);
|
|
if ($swent and $swent->{'switch'}) {
|
|
my $swh = new xCAT::SwitchHandler->new($swent->{'switch'});
|
|
my @vlids = $swh->get_vlan_ids();
|
|
if ($action eq 'add') {
|
|
unless (grep {$_ eq $vlanid} @vlids) {
|
|
$swh->create_vlan($vlanid);
|
|
}
|
|
$swh->add_ports_to_vlan($vlanid, $swent->{'port'});
|
|
} elsif ($action eq 'remove') {
|
|
$swh->remove_ports_from_vlan($vlanid, $swent->{'port'});
|
|
}
|
|
}
|
|
}
|
|
}
|
|
my $hostgroupdef = HostPortGroupSpec->new(
|
|
name =>$pgname,
|
|
vlanId=>$vlanid,
|
|
policy=>$policy,
|
|
vswitchName=>$switchname
|
|
);
|
|
unless ($netsys) {
|
|
$netsys = $hyphash{$hyp}->{conn}->get_view(mo_ref=>$hostview->configManager->networkSystem);
|
|
}
|
|
if ($action eq 'remove') {
|
|
$netsys->RemovePortGroup(pgName=>$pgname);
|
|
return;
|
|
} elsif ($action eq 'add') {
|
|
$netsys->AddPortGroup(portgrp=>$hostgroupdef);
|
|
}
|
|
#$hyphash{$hyp}->{nets}->{$netname}=1;
|
|
while ((not defined $hyphash{$hyp}->{nets}->{$pgname}) and sleep 1) { #we will only sleep if we know something will be waiting for
|
|
$hostview->update_view_data(); #pull in changes induced by previous activity
|
|
if (defined $hostview->{network}) { #We load the new object references
|
|
foreach (@{$hostview->network}) {
|
|
my $nvw = $hypconn->get_view(mo_ref=>$_);
|
|
if (defined $nvw->name) {
|
|
$hyphash{$hyp}->{nets}->{$nvw->name}=$_;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
sub validate_network_prereqs {
|
|
my $nodes = shift;
|
|
my $hyp = shift;
|
|
my $hypconn = $hyphash{$hyp}->{conn}; #this function can't work in clustered mode anyway, so this is appropriote.
|
|
my $hostview = $hyphash{$hyp}->{hostview};
|
|
if ($hostview) {
|
|
$hostview->update_view_data(); #pull in changes induced by previous activity
|
|
} else {
|
|
$hyphash{$hyp}->{hostview} = get_hostview(hypname=>$hyp,conn=>$hyphash{$hyp}->{conn}); #,properties=>['config','configManager','network']);
|
|
$hostview = $hyphash{$hyp}->{hostview};
|
|
}
|
|
my $node;
|
|
my $method;
|
|
my $location;
|
|
if (defined $hostview->{network}) {
|
|
foreach (@{$hostview->network}) {
|
|
my $nvw = $hypconn->get_view(mo_ref=>$_);
|
|
if (defined $nvw->name) {
|
|
$hyphash{$hyp}->{nets}->{$nvw->name}=$_;
|
|
}
|
|
}
|
|
}
|
|
foreach $node (@$nodes) {
|
|
my @networks = split /,/,$tablecfg{vm}->{$node}->[0]->{nics};
|
|
foreach (@networks) {
|
|
fixup_hostportgroup($_, $hyp);
|
|
}
|
|
}
|
|
return 1;
|
|
|
|
}
|
|
sub refreshclusterdatastoremap {
|
|
my $cluster = shift;
|
|
my $conn=$clusterhash{$cluster}->{conn};
|
|
my $cview = get_clusterview(clustname=>$cluster,conn=>$conn);
|
|
if (defined $cview->{datastore}) {
|
|
foreach (@{$cview->datastore}) {
|
|
my $dsv = $conn->get_view(mo_ref=>$_);
|
|
if (defined $dsv->info->{nas}) {
|
|
if ($dsv->info->nas->type eq 'NFS') {
|
|
my $mnthost = $dsv->info->nas->remoteHost;
|
|
# my $mnthost = inet_aton($dsv->info->nas->remoteHost);
|
|
# if ($mnthost) {
|
|
# $mnthost = inet_ntoa($mnthost);
|
|
# } else {
|
|
# $mnthost = $dsv->info->nas->remoteHost;
|
|
# xCAT::SvrUtils::sendmsg([1,"Unable to resolve VMware specified host '".$dsv->info->nas->remoteHost."' to an address, problems may occur"], $output_handler);
|
|
# }
|
|
$clusterhash{$cluster}->{datastoremap}->{"nfs://".$mnthost.$dsv->info->nas->remotePath}=$dsv->info->name;
|
|
$clusterhash{$cluster}->{datastoreurlmap}->{$dsv->info->name} = "nfs://".$mnthost.$dsv->info->nas->remotePath; #save off a suitable URL if needed
|
|
$clusterhash{$cluster}->{datastorerefmap}->{"nfs://".$mnthost.$dsv->info->nas->remotePath}=$_;
|
|
} #TODO: care about SMB
|
|
}elsif(defined $dsv->info->{vmfs}){
|
|
my $name = $dsv->info->vmfs->name;
|
|
$clusterhash{$cluster}->{datastoremap}->{"vmfs://".$name} = $dsv->info->name;
|
|
$clusterhash{$cluster}->{datastoreurlmap}->{$dsv->info->name} = "vmfs://".$name;
|
|
$clusterhash{$cluster}->{datastorerefmap}->{"vmfs://".$name} = $_;
|
|
}
|
|
}
|
|
}
|
|
#that's... about it... not doing any of the fancy mounting and stuff, if you do it cluster style, you are on your own. It's simply too terrifying to try to fixup
|
|
#a whole cluster instead of chasing one host, a whole lot slower. One would hope vmware would've done this, but they don't
|
|
}
|
|
sub validate_datastore_prereqs {
|
|
my $hyp = $_[1];
|
|
lockbyname($hyp.".datastores");
|
|
$@="";
|
|
my $rc;
|
|
eval { $rc=validate_datastore_prereqs_inlock(@_); };
|
|
unlockbyname($hyp.".datastores");
|
|
if ($@) { die $@; }
|
|
return $rc;
|
|
}
|
|
sub validate_datastore_prereqs_inlock {
|
|
my $nodes = shift;
|
|
my $hyp = shift;
|
|
my $newdatastores = shift; # a hash reference of URLs to afflicted nodes outside of table space
|
|
my $hypconn = $hyphash{$hyp}->{conn};
|
|
my $hostview = $hyphash{$hyp}->{hostview};
|
|
unless ($hostview) {
|
|
$hyphash{$hyp}->{hostview} = get_hostview(hypname=>$hyp,conn=>$hypconn); #,properties=>['config','configManager']);
|
|
$hostview = $hyphash{$hyp}->{hostview};
|
|
}
|
|
my $node;
|
|
my $method;
|
|
my $location;
|
|
# get all of the datastores that are currently available on this node.
|
|
# and put them into a hash
|
|
if (defined $hostview->{datastore}) { # only iterate if it exists
|
|
foreach (@{$hostview->datastore}) {
|
|
my $dsv = $hypconn->get_view(mo_ref=>$_);
|
|
if (defined $dsv->info->{nas}) {
|
|
if ($dsv->info->nas->type eq 'NFS') {
|
|
my $mnthost = inet_aton($dsv->info->nas->remoteHost);
|
|
if ($mnthost) {
|
|
$mnthost = inet_ntoa($mnthost);
|
|
} else {
|
|
$mnthost = $dsv->info->nas->remoteHost;
|
|
xCAT::SvrUtils::sendmsg([1,"Unable to resolve VMware specified host '".$dsv->info->nas->remoteHost."' to an address, problems may occur"], $output_handler);
|
|
}
|
|
$hyphash{$hyp}->{datastoremap}->{"nfs://".$mnthost.$dsv->info->nas->remotePath}=$dsv->info->name;
|
|
$hyphash{$hyp}->{datastoreurlmap}->{$dsv->info->name} = "nfs://".$mnthost.$dsv->info->nas->remotePath;
|
|
$hyphash{$hyp}->{datastorerefmap}->{"nfs://".$mnthost.$dsv->info->nas->remotePath}=$_;
|
|
} #TODO: care about SMB
|
|
}elsif(defined $dsv->info->{vmfs}){
|
|
my $name = $dsv->info->vmfs->name;
|
|
$hyphash{$hyp}->{datastoremap}->{"vmfs://".$name} = $dsv->info->name;
|
|
$hyphash{$hyp}->{datastoreurlmap}->{$dsv->info->name} = "vmfs://".$name;
|
|
$hyphash{$hyp}->{datastorerefmap}->{"vmfs://".$name} = $_;
|
|
}
|
|
}
|
|
}
|
|
my $refresh_names=0;
|
|
# now go through the nodes and make sure that we have matching datastores.
|
|
# E.g.: if its NFS, then mount it (if not mounted)
|
|
# E.g.: if its VMFS, then create it if not created already. Note: VMFS will persist on
|
|
# machine reboots, unless its destroyed by being overwritten.
|
|
foreach $node (@$nodes) {
|
|
my @storage = split /,/,$tablecfg{vm}->{$node}->[0]->{storage};
|
|
if ($tablecfg{vm}->{$node}->[0]->{cfgstore}) {
|
|
push @storage,$tablecfg{vm}->{$node}->[0]->{cfgstore};
|
|
}
|
|
foreach (@storage) { #TODO: merge this with foreach loop below. Here we could build onto $newdatastores instead, for faster operation at scale
|
|
s/=.*//; #remove device type information from configuration
|
|
s/\/$//; #Strip trailing slash if specified, to align to VMware semantics
|
|
if (/:\/\//) {
|
|
($method,$location) = split /:\/\//,$_,2;
|
|
if($method =~ /nfs/){
|
|
# go through and see if NFS is mounted, if not, then mount it.
|
|
(my $server, my $path) = split /\//,$location,2;
|
|
$server =~ s/:$//; #remove a : if someone put it in out of nfs mount habit
|
|
my $servern = inet_aton($server);
|
|
unless ($servern) {
|
|
xCAT::SvrUtils::sendmsg([1,": Unable to resolve '$server' to an address, check vm.cfgstore/vm.storage"], $output_handler);
|
|
return 0;
|
|
}
|
|
$server = inet_ntoa($servern);
|
|
my $uri = "nfs://$server/$path";
|
|
unless ($hyphash{$hyp}->{datastoremap}->{$uri}) { #If not already there, must mount it
|
|
unless ($datastoreautomount) {
|
|
xCAT::SvrUtils::sendmsg([1,": $uri is not currently accessible at the given location and automount is disabled in site table"], $output_handler,$node);
|
|
return 0;
|
|
}
|
|
$refresh_names=1;
|
|
($hyphash{$hyp}->{datastoremap}->{$uri},$hyphash{$hyp}->{datastorerefmap}->{$uri})=mount_nfs_datastore($hostview,$location);
|
|
$hyphash{$hyp}->{datastoreurlmap}->{$hyphash{$hyp}->{datastoremap}->{$uri}} = $uri;
|
|
}
|
|
}elsif($method =~ /vmfs/){
|
|
(my $name, undef) = split /\//,$location,2;
|
|
$name =~ s/:$//; #remove a : if someone put it in for some reason.
|
|
my $uri = "vmfs://$name";
|
|
# check and see if this vmfs is on the node.
|
|
unless ($hyphash{$hyp}->{datastoremap}->{$uri}) { #If not already there, try creating it.
|
|
unless ($datastoreautomount) {
|
|
xCAT::SvrUtils::sendmsg([1,": $uri is not currently accessible at the given location and automount is disabled in site table"], $output_handler,$node);
|
|
return 0;
|
|
}
|
|
$refresh_names=1;
|
|
($hyphash{$hyp}->{datastoremap}->{$uri},$hyphash{$hyp}->{datastorerefmap}->{$uri})=create_vmfs_datastore($hostview,$name,$hyp);
|
|
unless($hyphash{hyp}->{datastoremap}->{$uri}){ return 0; }
|
|
$hyphash{$hyp}->{datastoreurlmap}->{$hyphash{$hyp}->{datastoremap}->{$uri}} = $uri;
|
|
}
|
|
}else{
|
|
xCAT::SvrUtils::sendmsg([1,": $method is unsupported at this time (nfs would be)"], $output_handler,$node);
|
|
return 0;
|
|
}
|
|
} else {
|
|
xCAT::SvrUtils::sendmsg([1,": $_ not supported storage specification for ESX plugin,\n\t'nfs://<server>/<path>'\n\t\tor\n\t'vmfs://<vmfs>'\n only currently supported vm.storage supported for ESX at the moment"], $output_handler,$node);
|
|
return 0;
|
|
} #TODO: raw device mapping, VMFS via iSCSI, VMFS via FC?
|
|
}
|
|
}
|
|
# newdatastores are for migrations or changing vms.
|
|
# TODO: make this work for VMFS. Right now only NFS.
|
|
if (ref $newdatastores) {
|
|
foreach (keys %$newdatastores) {
|
|
my $origurl=$_;
|
|
s/\/$//; #Strip trailing slash if specified, to align to VMware semantics
|
|
if (/:\/\//) {
|
|
($method,$location) = split /:\/\//,$_,2;
|
|
if($method =~ /nfs/){
|
|
(my $server, my $path) = split /\//,$location,2;
|
|
$server =~ s/:$//; #remove a : if someone put it in out of nfs mount habit
|
|
my $servern = inet_aton($server);
|
|
unless ($servern) {
|
|
xCAT::SvrUtils::sendmsg([1,": Unable to resolve '$server' to an address, check vm.cfgstore/vm.storage"], $output_handler);
|
|
return 0;
|
|
}
|
|
$server = inet_ntoa($servern);
|
|
my $uri = "nfs://$server/$path";
|
|
unless ($method =~ /nfs/) {
|
|
foreach (@{$newdatastores->{$_}}) {
|
|
xCAT::SvrUtils::sendmsg([1,": $method is unsupported at this time (nfs would be)"], $output_handler,$_);
|
|
}
|
|
return 0;
|
|
}
|
|
unless ($hyphash{$hyp}->{datastoremap}->{$uri}) { #If not already there, must mount it
|
|
unless ($datastoreautomount) {
|
|
xCAT::SvrUtils::sendmsg([1,":) $uri is not currently accessible at the given location and automount is disabled in site table"], $output_handler,$node);
|
|
return 0;
|
|
}
|
|
$refresh_names=1;
|
|
($hyphash{$hyp}->{datastoremap}->{$uri},$hyphash{$hyp}->{datastorerefmap}->{$uri})=mount_nfs_datastore($hostview,$location);
|
|
}
|
|
$hyphash{$hyp}->{datastoreurlmap}->{$hyphash{$hyp}->{datastoremap}->{$uri}} = $uri;
|
|
$hyphash{$hyp}->{datastoremap}->{$origurl}=$hyphash{$hyp}->{datastoremap}->{$uri}; #we track both the uri xCAT expected and the one vCenter actually ended up with
|
|
$hyphash{$hyp}->{datastorerefmap}->{$origurl}=$hyphash{$hyp}->{datastorerefmap}->{$uri};
|
|
}elsif($method =~ /vmfs/){
|
|
(my $name, undef) = split /\//,$location,2;
|
|
$name =~ s/:$//; #remove a : if someone put it in for some reason.
|
|
my $uri = "vmfs://$name";
|
|
unless ($hyphash{$hyp}->{datastoremap}->{$uri}) { #If not already there, it should be!
|
|
unless ($datastoreautomount) {
|
|
xCAT::SvrUtils::sendmsg([1,": $uri is not currently accessible at the given location and automount is disabled in site table"], $output_handler,$node);
|
|
return 0;
|
|
}
|
|
$refresh_names=1;
|
|
($hyphash{$hyp}->{datastoremap}->{$uri},$hyphash{$hyp}->{datastorerefmap}->{$uri})=create_vmfs_datastore($hostview,$name,$hyp);
|
|
unless($hyphash{hyp}->{datastoremap}->{$uri}){ return 0; }
|
|
}
|
|
$hyphash{$hyp}->{datastoreurlmap}->{$hyphash{$hyp}->{datastoremap}->{$uri}} = $uri;
|
|
$hyphash{$hyp}->{datastoremap}->{$origurl}=$hyphash{$hyp}->{datastoremap}->{$uri};
|
|
$hyphash{$hyp}->{datastorerefmap}->{$origurl}=$hyphash{$hyp}->{datastorerefmap}->{$uri};
|
|
}else{
|
|
print "$method: not NFS and not VMFS here!\n";
|
|
}
|
|
} else {
|
|
my $datastore=$_;
|
|
foreach my $ds (@{$newdatastores->{$_}}) {
|
|
xCAT::SvrUtils::sendmsg([1,": $datastore not supported storage specification for ESX plugin, 'nfs://<server>/<path>' only currently supported vm.storage supported for ESX at the moment"], $output_handler,$ds);
|
|
}
|
|
return 0;
|
|
} #TODO: raw device mapping, VMFS via iSCSI, VMFS via FC, VMFS on same local drive?
|
|
}
|
|
}
|
|
if ($refresh_names) { #if we are in a vcenter context, vmware can rename a datastore behind our backs immediately after adding
|
|
$hostview->update_view_data();
|
|
if (defined $hostview->{datastore}) { # only iterate if it exists
|
|
foreach (@{$hostview->datastore}) {
|
|
my $dsv = $hypconn->get_view(mo_ref=>$_);
|
|
if (defined $dsv->info->{nas}) {
|
|
if ($dsv->info->nas->type eq 'NFS') {
|
|
my $mnthost = inet_aton($dsv->info->nas->remoteHost);
|
|
if ($mnthost) {
|
|
$mnthost = inet_ntoa($mnthost);
|
|
} else {
|
|
$mnthost = $dsv->info->nas->remoteHost;
|
|
xCAT::SvrUtils::sendmsg([1,"Unable to resolve VMware specified host '".$dsv->info->nas->remoteHost."' to an address, problems may occur"], $output_handler);
|
|
}
|
|
$hyphash{$hyp}->{datastoremap}->{"nfs://".$mnthost.$dsv->info->nas->remotePath}=$dsv->info->name;
|
|
$hyphash{$hyp}->{datastoreurlmap}->{$dsv->info->name} = "nfs://".$mnthost.$dsv->info->nas->remotePath;
|
|
$hyphash{$hyp}->{datastorerefmap}->{"nfs://".$mnthost.$dsv->info->nas->remotePath}=$_;
|
|
} #TODO: care about SMB
|
|
} #TODO: care about VMFS
|
|
}
|
|
}
|
|
}
|
|
return 1;
|
|
}
|
|
|
|
sub getlabel_for_datastore {
|
|
my $method = shift;
|
|
my $location = shift;
|
|
|
|
$location =~ s/\//_/g;
|
|
$location= $method.'_'.$location;
|
|
#VMware has a 42 character limit, we will start mangling to get under 42.
|
|
#Will try to preserve as much informative detail as possible, hence several conditionals instead of taking the easy way out
|
|
if (length($location) > 42) {
|
|
$location =~ s/nfs_//; #Ditch unique names for different protocols to the same path, seems unbelievably unlikely
|
|
}
|
|
if (length($location) > 42) {
|
|
$location =~ s/\.//g; #Next, ditch host delimiter, it is unlikely that hosts will have unique names if their dots are removed
|
|
}
|
|
if (length($location) > 42) {
|
|
$location =~ s/_//g; #Next, ditch path delimiter, it is unlikely that two paths will happen to look the same without delimiters
|
|
}
|
|
if (length($location) > 42) { #finally, replace the middle with ellipsis
|
|
substr($location,20,-20,'..');
|
|
}
|
|
return $location;
|
|
}
|
|
|
|
sub mount_nfs_datastore {
|
|
my $hostview = shift;
|
|
my $location = shift;
|
|
my $server;
|
|
my $path;
|
|
unless ($datastoreautomount) {
|
|
die "automount of VMware datastores is disabled in site configuration, not continuing";
|
|
}
|
|
($server,$path) = split /\//,$location,2;
|
|
$location = getlabel_for_datastore('nfs',$location);
|
|
|
|
my $nds = HostNasVolumeSpec->new(accessMode=>'readWrite',
|
|
remoteHost=>$server,
|
|
localPath=>$location,
|
|
remotePath=>"/".$path);
|
|
my $dsmv = $hostview->{vim}->get_view(mo_ref=>$hostview->configManager->datastoreSystem);
|
|
|
|
my $dsref;
|
|
eval {
|
|
$dsref=$dsmv->CreateNasDatastore(spec=>$nds);
|
|
};
|
|
|
|
if ($@) {
|
|
die "$@" unless $@ =~ m/Fault detail: DuplicateNameFault/;
|
|
|
|
die "esx plugin: a datastore was discovered with the same name referring to a different nominatum- cannot continue\n$@"
|
|
unless &match_nfs_datastore($server,"/$path",$hostview->{vim});
|
|
}
|
|
|
|
return ($location,$dsref);
|
|
}
|
|
|
|
# create a VMFS data store on a node so that VMs can live locally instead of NFS
|
|
sub create_vmfs_datastore {
|
|
my $hostview = shift; # VM object
|
|
my $name = shift; # name of storage we wish to create.
|
|
my $hyp = shift;
|
|
unless ($datastoreautomount) {
|
|
die "automount of VMware datastores is disabled in site configuration, not continuing";
|
|
}
|
|
# call some VMware API here to create
|
|
my $hdss = $hostview->{vim}->get_view(mo_ref=>$hostview->configManager->datastoreSystem);
|
|
|
|
my $diskList = $hdss->QueryAvailableDisksForVmfs();
|
|
my $count = scalar(@$diskList); # get the number of disks available for formatting.
|
|
unless($count >0){
|
|
#die "No disks are available to create VMFS volume for $name";
|
|
$output_handler->({error=>["No disks are available on $hyp to create VMFS volume for $name"],errorcode=>1});
|
|
return 0;
|
|
}
|
|
foreach my $disk(@$diskList){
|
|
my $options = $hdss->QueryVmfsDatastoreCreateOptions(devicePath => $disk->devicePath);
|
|
@$options[0]->spec->vmfs->volumeName($name);
|
|
my $newDatastore = $hdss->CreateVmfsDatastore(spec => @$options[0]->spec );
|
|
#return $newDatastore;
|
|
# create it on the first disk we see.
|
|
return ($name, $newDatastore);
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
|
|
|
|
sub build_more_info{
|
|
die("TODO: fix this function if called");
|
|
print "Does this acually get called????**********************************\n";
|
|
my $noderange=shift;
|
|
my $callback=shift;
|
|
my $vmtab = xCAT::Table->new("vm");
|
|
my @moreinfo=();
|
|
unless ($vmtab) {
|
|
$callback->({data=>["Cannot open mp table"]});
|
|
return @moreinfo;
|
|
}
|
|
my %mpa_hash=();
|
|
foreach my $node (@$noderange) {
|
|
my $ent=$vmtab->getNodeAttribs($node,['mpa', 'id']);
|
|
if (defined($ent->{mpa})) { push @{$mpa_hash{$ent->{mpa}}{nodes}}, $node;}
|
|
else {
|
|
$callback->({data=>["no mpa defined for node $node"]});
|
|
return @moreinfo;
|
|
}
|
|
if (defined($ent->{id})) { push @{$mpa_hash{$ent->{mpa}}{ids}}, $ent->{id};}
|
|
else { push @{$mpa_hash{$ent->{mpa}}{ids}}, "";}
|
|
}
|
|
|
|
foreach (keys %mpa_hash) {
|
|
push @moreinfo, "\[$_\]\[" . join(',',@{$mpa_hash{$_}{nodes}}) ."\]\[" . join(',',@{$mpa_hash{$_}{ids}}) . "\]";
|
|
|
|
}
|
|
|
|
return \@moreinfo;
|
|
}
|
|
|
|
sub copycd {
|
|
my $request = shift;
|
|
my $doreq = shift;
|
|
my $distname = "";
|
|
my $path;
|
|
my $arch;
|
|
my $darch;
|
|
my $installroot;
|
|
$installroot = "/install";
|
|
#my $sitetab = xCAT::Table->new('site');
|
|
#if($sitetab){
|
|
#(my $ref) = $sitetab->getAttribs({key => 'installdir'}, 'value');
|
|
my @entries = xCAT::TableUtils->get_site_attribute("installdir");
|
|
my $t_entry = $entries[0];
|
|
if ( defined($t_entry) ) {
|
|
$installroot = $t_entry;
|
|
}
|
|
#}
|
|
@ARGV = @{$request->{arg}};
|
|
my $includeupdate = 0;
|
|
GetOptions(
|
|
'n=s' => \$distname,
|
|
'a=s' => \$arch,
|
|
'm=s' => \$path,
|
|
's' => \$includeupdate
|
|
);
|
|
# run a few tests to see if the copycds should use this plugin
|
|
unless ($path){
|
|
# can't use us cause we need a path and you didn't provide one!
|
|
return;
|
|
}
|
|
if( $distname and $distname !~ /^esx/ ){
|
|
# we're for esx, so if you didn't specify that its not us!
|
|
return;
|
|
}
|
|
my $found = 0;
|
|
|
|
if (-r $path . "/README" and -r $path . "/build_number" and -d $path . "/VMware" and -r $path . "/packages.xml") { #We have a probable new style ESX media
|
|
open(LINE,$path."/packages.xml");
|
|
my $product;
|
|
my $version;
|
|
while (<LINE>) {
|
|
if (/roductLineId>([^<]*)<\/Prod/) {
|
|
$product = $1;
|
|
}
|
|
if (/ersion>([^<]*)<\/version/) {
|
|
$version = $1;
|
|
$version =~ s/\.0$//;
|
|
}
|
|
if (/arch>([^>]*)<\/arch/) {
|
|
unless ($darch and $darch =~ /x86_64/) { #prefer to be characterized as x86_64
|
|
$darch = $1;
|
|
$arch = $1;
|
|
}
|
|
|
|
}
|
|
}
|
|
close(LINE);
|
|
if ($product and $version) {
|
|
unless ($distname) { $distname = $product.$version; }
|
|
$found = 1;
|
|
}
|
|
} elsif (-r $path . "/README" and -r $path . "/open_source_licenses.txt" and -d $path . "/VMware") { #Candidate to be ESX 3.5
|
|
open(LINE,$path."/README");
|
|
while(<LINE>) {
|
|
if (/VMware ESX Server 3.5\s*$/) {
|
|
$darch ='x86';
|
|
$arch = 'x86';
|
|
unless ($distname) { $distname = 'esx3.5'; }
|
|
$found = 1;
|
|
last;
|
|
}
|
|
}
|
|
close(LINE);
|
|
} elsif (-r $path . "/README.txt" and -r $path . "/vmkernel.gz"){
|
|
# its an esxi dvd!
|
|
# if we got here its probably ESX they want to copy
|
|
my $line;
|
|
my $darch;
|
|
open(LINE, $path . "/README.txt") or die "couldn't open!";
|
|
while($line = <LINE>){
|
|
chomp($line);
|
|
if($line =~ /VMware ESXi(?: version)? 4\.(\d+)/){
|
|
$darch = "x86_64";
|
|
unless ($distname) {
|
|
$distname = "esxi4";
|
|
if ($1) {
|
|
$distname .= '.'.$1;
|
|
}
|
|
}
|
|
$found = 1;
|
|
if( $arch and $arch ne $darch){
|
|
xCAT::SvrUtils::sendmsg([1, "Requested distribution architecture $arch, but media is $darch"], $output_handler);
|
|
return;
|
|
}
|
|
$arch = $darch;
|
|
last; # we found our distro! end this loop madness.
|
|
}
|
|
}
|
|
close(LINE);
|
|
unless($found){
|
|
xCAT::SvrUtils::sendmsg([1,"I don't recognize this VMware ESX DVD"], $output_handler);
|
|
return; # doesn't seem to be a valid DVD or CD
|
|
}
|
|
} elsif (-r $path . "/vmkernel.gz" and -r $path . "/isolinux.cfg"){
|
|
open(LINE,$path . "/isolinux.cfg");
|
|
while (<LINE>) {
|
|
if (/ThinESX Installer/) {
|
|
$darch = 'x86';
|
|
$arch='x86';
|
|
unless ($distname) { $distname='esxi3.5'; }
|
|
$found=1;
|
|
last;
|
|
}
|
|
}
|
|
close(LINE);
|
|
} elsif (-r $path . "/upgrade/metadata.xml") {
|
|
open(LINE,$path."/upgrade/metadata.xml");
|
|
my $detectdistname;
|
|
while (<LINE>) {
|
|
if (/esxVersion>([^<]*)</) {
|
|
my $version = $1;
|
|
while ($version =~ /\.0$/) {
|
|
$version =~ s/\.0$//;
|
|
}
|
|
$darch="x86_64";
|
|
$arch="x86_64";
|
|
$detectdistname = 'esxi' . $version;
|
|
$found=1;
|
|
} elsif (/esxRelease>([^<]*)</) {
|
|
unless ($includeupdate) {
|
|
next;
|
|
}
|
|
my $release = $1;
|
|
while ($release =~ /\.0$/) {
|
|
$release =~ s/\.0$//;
|
|
}
|
|
unless ($release ne "0") {
|
|
next;
|
|
}
|
|
$detectdistname .= '_' . $release;
|
|
}
|
|
}
|
|
unless ($distname) { $distname=$detectdistname; }
|
|
} elsif (-r $path . "/vmware-esx-base-readme") {
|
|
open(LINE,$path."/vmware-esx-base-readme");
|
|
while (<LINE>) {
|
|
if (/VMware ESXi 5\.0/) {
|
|
$darch="x86_64";
|
|
$arch="x86_64";
|
|
unless ($distname) { $distname='esxi5'; }
|
|
$found=1;
|
|
last;
|
|
}
|
|
if (/VMware ESXi 5\.1/) {
|
|
$darch="x86_64";
|
|
$arch="x86_64";
|
|
unless ($distname) { $distname='esxi5.1'; }
|
|
$found=1;
|
|
last;
|
|
}
|
|
if (/VMware ESXi 5\.5/) {
|
|
$darch="x86_64";
|
|
$arch="x86_64";
|
|
unless ($distname) { $distname='esxi5.5'; }
|
|
$found=1;
|
|
last;
|
|
}
|
|
}
|
|
}
|
|
|
|
unless ($found) { return; } #not our media
|
|
if ($::XCATSITEVALS{osimagerequired}){
|
|
my ($nohaveimages,$errstr)=xCAT::SvrUtils->update_tables_with_templates($distname, $arch,"","",checkonly=>1);
|
|
if ($nohaveimages) {
|
|
$output_handler->({error => "No Templates found to support $distname($arch)",errorcode=>2});
|
|
}
|
|
}
|
|
|
|
xCAT::SvrUtils::sendmsg("Copying media to $installroot/$distname/$arch/", $output_handler);
|
|
my $omask = umask 0022;
|
|
mkpath("$installroot/$distname/$arch");
|
|
umask $omask;
|
|
my $rc;
|
|
my $reaped = 0;
|
|
$SIG{INT} = $SIG{TERM} = sub {
|
|
foreach(@cpiopid){
|
|
kill 2, $_;
|
|
}
|
|
if ($path) {
|
|
chdir("/");
|
|
system("umount $path");
|
|
}
|
|
};
|
|
my $KID;
|
|
chdir $path;
|
|
my $numFiles = `find . -print | wc -l`;
|
|
my $child = open($KID, "|-");
|
|
unless (defined $child)
|
|
{
|
|
xCAT::SvrUtils::sendmsg([1,"Media copy operation fork failure"], $output_handler);
|
|
return;
|
|
}
|
|
if ($child)
|
|
{
|
|
push @cpiopid, $child;
|
|
my @finddata = `find .`;
|
|
for (@finddata)
|
|
{
|
|
print $KID $_;
|
|
}
|
|
close($KID);
|
|
$rc = $?;
|
|
}
|
|
else
|
|
{
|
|
nice 10;
|
|
my $c = "nice -n 20 cpio -vdump $installroot/$distname/$arch";
|
|
my $k2 = open(PIPE, "$c 2>&1 |") ||
|
|
xCAT::SvrUtils::sendmsg([1,"Media copy operation fork failure"], $output_handler);
|
|
push @cpiopid, $k2;
|
|
my $copied = 0;
|
|
my ($percent, $fout);
|
|
while(<PIPE>){
|
|
next if /^cpio:/;
|
|
$percent = $copied / $numFiles;
|
|
$fout = sprintf "%0.2f%%", $percent * 100;
|
|
$output_handler->({sinfo => "$fout"});
|
|
++$copied;
|
|
}
|
|
exit;
|
|
}
|
|
# let everyone read it
|
|
#chdir "/tmp";
|
|
chmod 0755, "$installroot/$distname/$arch";
|
|
if ($distname =~ /esxi5/) { #going to tweak boot.cfg for install and default stateless case
|
|
if (! -r "$installroot/$distname/$arch/boot.cfg.stateless") {
|
|
copy("$installroot/$distname/$arch/boot.cfg","$installroot/$distname/$arch/boot.cfg.stateless");
|
|
my $bootcfg;
|
|
open($bootcfg,"<","$installroot/$distname/$arch/boot.cfg");
|
|
my @bootcfg = <$bootcfg>;
|
|
close($bootcfg);
|
|
foreach (@bootcfg) { #no point in optimizing trivial, infrequent code, readable this way
|
|
s!kernel=/!kernel=!; # remove leading /
|
|
s!modules=/!modules=!; #remove leading /
|
|
s!--- /!--- !g; #remove all the 'absolute' slashes
|
|
}
|
|
open($bootcfg,">","$installroot/$distname/$arch/boot.cfg.install");
|
|
foreach (@bootcfg) {
|
|
if (/^modules=/ and $_ !~ /xcatmod.tgz/ and not $::XCATSITEVALS{xcatesximoddisable}) {
|
|
chomp();
|
|
s! *\z! --- xcatmod.tgz\n!;
|
|
}
|
|
print $bootcfg $_;
|
|
}
|
|
close($bootcfg);
|
|
foreach (@bootcfg) { #no point in optimizing trivial, infrequent code, readable this way
|
|
s/runweasel//; #don't run the installer in stateless mode
|
|
s!--- imgdb.tgz!!; #don't need the imgdb for stateless
|
|
s!--- imgpayld.tgz!!; #don't need the boot payload since we aren't installing
|
|
s!--- tools.t00!!; #tools could be useful, but for now skip the memory requirement
|
|
s!--- weaselin.i00!!; #and also don't need the weasel install images if... not installing
|
|
|
|
if (/^modules=/ and $_ !~ /xcatmod.tgz/ and not $::XCATSITEVALS{xcatesximoddisable}) {
|
|
chomp();
|
|
s! *\z! --- xcatmod.tgz\n!;
|
|
}
|
|
s!Loading ESXi installer!xCAT is loading ESXi stateless!;
|
|
}
|
|
open($bootcfg,">","$installroot/$distname/$arch/boot.cfg.stateless");
|
|
foreach (@bootcfg) {
|
|
print $bootcfg $_;
|
|
}
|
|
close($bootcfg);
|
|
if (grep /LSIProvi.v00/,@bootcfg and ! -r "$installroot/$distname/$arch/LSIProvi.v00" and -r "$installroot/$distname/$arch/lsiprovi.v00") { #there is media with LSIProv.v00 expected, but the install media was mal-constructed, fix it
|
|
move("$installroot/$distname/$arch/lsiprovi.v00","$installroot/$distname/$arch/LSIProvi.v00");
|
|
}
|
|
}
|
|
}
|
|
|
|
if ($rc != 0){
|
|
xCAT::SvrUtils::sendmsg([1,"Media copy operation failed, status $rc"], $output_handler);
|
|
}else{
|
|
xCAT::SvrUtils::sendmsg("Media copy operation successful", $output_handler);
|
|
my @ret=xCAT::SvrUtils->update_tables_with_templates($distname, $arch);
|
|
if ($ret[0] != 0) {
|
|
xCAT::SvrUtils::sendmsg("Error when updating the osimage tables: " . $ret[1], $output_handler);
|
|
}
|
|
|
|
|
|
}
|
|
}
|
|
sub makecustomizedmod {
|
|
my $osver = shift;
|
|
my $dest = shift;
|
|
if ($::XCATSITEVALS{xcatesximoddisable}) { return 1; }
|
|
my $modname;
|
|
if ($osver =~ /esxi4/) { #want more descriptive name,but don't break esxi4 setups.
|
|
$modname="mod.tgz";
|
|
# if it already exists, do not overwrite it because it may be someone
|
|
# else's custom image
|
|
if(-f "$dest/$modname"){ return 1; }
|
|
} else {
|
|
$modname="xcatmod.tgz";
|
|
}
|
|
my $passtab = xCAT::Table->new('passwd');
|
|
my $tmp;
|
|
my $password;
|
|
if ($passtab) {
|
|
($tmp) = $passtab->getAttribs({'key'=>'vmware'},'username','password');
|
|
if (defined($tmp)) {
|
|
$password = $tmp->{password};
|
|
}
|
|
}
|
|
unless ($password) {
|
|
xCAT::SvrUtils::sendmsg([1,": Unable to find a password entry for esxi in passwd table"], $output_handler);
|
|
return 0;
|
|
}
|
|
mkpath("/tmp/xcat");
|
|
my $tempdir = tempdir("/tmp/xcat/esxmodcustXXXXXXXX");
|
|
my $shadow;
|
|
mkpath($tempdir."/etc/");
|
|
my $oldmask=umask(0077);
|
|
open($shadow,">",$tempdir."/etc/shadow");
|
|
$password = crypt($password,'$1$'.xCAT::Utils::genpassword(8));
|
|
my $dayssince1970 = int(time()/86400); #Be truthful about /etc/shadow
|
|
my @otherusers = qw/nobody nfsnobody dcui daemon/;
|
|
if ($osver =~ /esxi4/) {
|
|
push @otherusers,"vimuser";
|
|
} elsif ($osver =~ /esxi5/) {
|
|
push @otherusers,"vpxuser";
|
|
}
|
|
print $shadow "root:$password:$dayssince1970:0:99999:7:::\n";
|
|
foreach (@otherusers) {
|
|
print $shadow "$_:*:$dayssince1970:0:99999:7:::\n";
|
|
}
|
|
close($shadow);
|
|
umask($oldmask);
|
|
if ($osver =~ /esxi4/ and -e "$::XCATROOT/share/xcat/netboot/esxi/38.xcat-enableipv6") {
|
|
mkpath($tempdir."/etc/vmware/init/init.d");
|
|
copy( "$::XCATROOT/share/xcat/netboot/esxi/38.xcat-enableipv6",$tempdir."/etc/vmware/init/init.d/38.xcat-enableipv6");
|
|
} elsif ($osver =~ /esxi5/ and -e "$::XCATROOT/share/xcat/netboot/esxi/xcat-ipv6.json") {
|
|
mkpath($tempdir."/usr/libexec/jumpstart/plugins/");
|
|
copy( "$::XCATROOT/share/xcat/netboot/esxi/xcat-ipv6.json",$tempdir."/usr/libexec/jumpstart/plugins/xcat-ipv6.json");
|
|
}
|
|
if ($osver =~ /esxi4/ and -e "$::XCATROOT/share/xcat/netboot/esxi/47.xcat-networking") {
|
|
copy( "$::XCATROOT/share/xcat/netboot/esxi/47.xcat-networking",$tempdir."/etc/vmware/init/init.d/47.xcat-networking");
|
|
} elsif ($osver =~ /esxi5/ and -e "$::XCATROOT/share/xcat/netboot/esxi/39.ipv6fixup") {
|
|
mkpath($tempdir."/etc/init.d");
|
|
copy( "$::XCATROOT/share/xcat/netboot/esxi/39.ipv6fixup",$tempdir."/etc/init.d/39.ipv6fixup");
|
|
chmod(0755,"$tempdir/etc/init.d/39.ipv6fixup");
|
|
}
|
|
if ($osver =~ /esxi5/ and -e "$::XCATROOT/share/xcat/netboot/esxi/48.esxifixup") {
|
|
mkpath($tempdir."/etc/init.d");
|
|
copy( "$::XCATROOT/share/xcat/netboot/esxi/48.esxifixup",$tempdir."/etc/init.d/48.esxifixup");
|
|
chmod(0755,"$tempdir/etc/init.d/48.esxifixup");
|
|
}
|
|
if ($osver =~ /esxi5/ and -e "$::XCATROOT/share/xcat/netboot/esxi/99.esxiready") {
|
|
mkpath($tempdir."/etc/init.d");
|
|
copy( "$::XCATROOT/share/xcat/netboot/esxi/99.esxiready",$tempdir."/etc/init.d/99.esxiready");
|
|
chmod(0755,"$tempdir/etc/init.d/99.esxiready");
|
|
}
|
|
if (-e "$::XCATROOT/share/xcat/netboot/esxi/xcatsplash") {
|
|
mkpath($tempdir."/etc/vmware/");
|
|
copy( "$::XCATROOT/share/xcat/netboot/esxi/xcatsplash",$tempdir."/etc/vmware/welcome");
|
|
}
|
|
my $dossh=0;
|
|
if (-r "/root/.ssh/id_rsa.pub") {
|
|
$dossh=1;
|
|
my $umask = umask(0077);#don't remember if dropbear is picky, but just in case
|
|
if ($osver =~ /esxi4/) { #esxi4 used more typical path
|
|
mkpath($tempdir."/.ssh");
|
|
copy("/root/.ssh/id_rsa.pub",$tempdir."/.ssh/authorized_keys");
|
|
} elsif ($osver =~ /esxi5/) { #weird path to keys
|
|
mkpath($tempdir."/etc/ssh/keys-root");
|
|
copy("/root/.ssh/id_rsa.pub",$tempdir."/etc/ssh/keys-root/authorized_keys");
|
|
}
|
|
umask($umask);
|
|
}
|
|
my $tfile;
|
|
mkpath($tempdir."/var/run/vmware");
|
|
open $tfile,">",$tempdir."/var/run/vmware/show-tech-support-login";
|
|
close($tfile);
|
|
#TODO: auto-enable ssh and request boot-time customization rather than on-demand?
|
|
require Cwd;
|
|
my $dir=Cwd::cwd();
|
|
chdir($tempdir);
|
|
if (-e "$dest/$modname") {
|
|
unlink("$dest/$modname");
|
|
}
|
|
if ($dossh and $osver =~ /esxi4/) {
|
|
system("tar czf $dest/$modname * .ssh");
|
|
} else {
|
|
system("tar czf $dest/$modname *");
|
|
}
|
|
chdir($dir);
|
|
rmtree($tempdir);
|
|
return 1;
|
|
}
|
|
sub getplatform {
|
|
my $os = shift;
|
|
if ($os =~ /esxi/) {
|
|
return "esxi";
|
|
}
|
|
return $os;
|
|
}
|
|
sub esxi_kickstart_from_template {
|
|
my %args=@_;
|
|
my $installdir = "/install";
|
|
if ($::XCATSITEVALS{installdir}) { $installdir = $::XCATSITEVALS{installdir}; }
|
|
my $plat = getplatform($args{os});
|
|
my $template = xCAT::SvrUtils::get_tmpl_file_name("$installdir/custom/install/$plat",$args{profile},$args{os},$args{arch},$args{os});
|
|
unless ($template) {
|
|
$template = xCAT::SvrUtils::get_tmpl_file_name("$::XCATROOT/share/xcat/install/$plat",$args{profile},$args{os},$args{arch},$args{os});
|
|
}
|
|
my $tmperr;
|
|
if (-r "$template") {
|
|
$tmperr=xCAT::Template->subvars($template,"$installdir/autoinst/".$args{node},$args{node},undef);
|
|
} else {
|
|
$tmperr="Unable to find template in /install/custom/install/$plat or $::XCATROOT/share/xcat/install/$plat (for $args{profile}/$args{os}/$args{arch} combination)";
|
|
}
|
|
if ($tmperr) {
|
|
xCAT::SvrUtils::sendmsg([1,$tmperr], $output_handler,$args{node});
|
|
}
|
|
|
|
}
|
|
sub mkinstall {
|
|
return mkcommonboot("install",@_);
|
|
}
|
|
sub mknetboot {
|
|
return mkcommonboot("stateless",@_);
|
|
}
|
|
sub merge_esxi5_append {
|
|
my $tmpl = shift;
|
|
my $append = shift;
|
|
my $outfile = shift;
|
|
my $in;
|
|
my $out;
|
|
open($in,"<",$tmpl);
|
|
open($out,">",$outfile);
|
|
my $line;
|
|
while ($line = <$in>) {
|
|
if ($line =~ /kernelopt=/) {
|
|
chomp($line);
|
|
$line .= $append."\n";
|
|
#if ($line =~ /modules=b.b00/) {
|
|
# $line =~ s/modules=b.b00/modules=b.b00 $append/;
|
|
}
|
|
unless ($line =~ /^prefix=/) {
|
|
print $out $line;
|
|
}
|
|
}
|
|
}
|
|
sub mkcommonboot {
|
|
my $bootmode = shift;
|
|
my $req = shift;
|
|
my $doreq = shift;
|
|
my $globaltftpdir = "/tftpboot";
|
|
my @nodes = @{$req->{node}};
|
|
my $ostab = xCAT::Table->new('nodetype');
|
|
#my $sitetab = xCAT::Table->new('site');
|
|
my $bptab = xCAT::Table->new('bootparams',-create=>1);
|
|
my $installroot = "/install";
|
|
#if ($sitetab){
|
|
#(my $ref) = $sitetab->getAttribs({key => 'installdir'}, 'value');
|
|
my @entries = xCAT::TableUtils->get_site_attribute("installdir");
|
|
my $t_entry = $entries[0];
|
|
if ( defined($t_entry) ) {
|
|
$installroot = $t_entry;
|
|
}
|
|
#($ref) = $sitetab->getAttribs({key => 'tftpdir'}, 'value');
|
|
@entries = xCAT::TableUtils->get_site_attribute("tftpdir");
|
|
$t_entry = $entries[0];
|
|
if ( defined($t_entry) ) {
|
|
$globaltftpdir = $t_entry;
|
|
}
|
|
#}
|
|
my %donetftp=();
|
|
|
|
my $bpadds = $bptab->getNodesAttribs(\@nodes,['addkcmdline']);
|
|
my $nodehmtab = xCAT::Table->new('nodehm',-create=>0);
|
|
my $serialconfig;
|
|
if ($nodehmtab) {
|
|
$serialconfig = $nodehmtab->getNodesAttribs(\@nodes,['serialport','serialspeed']);
|
|
}
|
|
my $restab = xCAT::Table->new('noderes',-create=>0);
|
|
my $resents;
|
|
if ($restab) {
|
|
$resents = $restab->getNodesAttribs(\@nodes,['tftpdir','nfsserver']);
|
|
}
|
|
|
|
my %tablecolumnsneededforaddkcmdline;
|
|
my %nodesubdata;
|
|
foreach my $key (keys %$bpadds){ #First, we identify all needed table.columns needed to aggregate database call
|
|
my $add = $bpadds->{$key}->[0]->{addkcmdline};
|
|
|
|
next if ! defined $add;
|
|
|
|
while ($add =~ /#NODEATTRIB:([^:#]+):([^:#]+)#/) {
|
|
push @{$tablecolumnsneededforaddkcmdline{$1}},$2;
|
|
$add =~ s/#NODEATTRIB:([^:#]+):([^:#]+)#//;
|
|
}
|
|
}
|
|
foreach my $table (keys %tablecolumnsneededforaddkcmdline) {
|
|
my $tab = xCAT::Table->new($table,-create=>0);
|
|
if ($tab) {
|
|
$nodesubdata{$table}=$tab->getNodesAttribs(\@nodes,$tablecolumnsneededforaddkcmdline{$table});
|
|
}
|
|
}
|
|
|
|
my $osents = $ostab->getNodesAttribs(\@nodes, ['os', 'arch', 'profile']);
|
|
foreach my $node (@nodes){
|
|
my $ent = $osents->{$node}->[0];
|
|
my $arch = $ent->{'arch'};
|
|
my $profile = $ent->{'profile'};
|
|
my $osver = $ent->{'os'};
|
|
my $tftpdir;
|
|
my $ksserver;
|
|
if ($resents and $resents->{$node}->[0]->{nfsserver}) {
|
|
$ksserver=$resents->{$node}->[0]->{nfsserver};
|
|
} else {
|
|
$ksserver='!myipfn!';
|
|
}
|
|
|
|
if ($resents and $resents->{$node}->[0]->{tftpdir}) {
|
|
$tftpdir = $resents->{$node}->[0]->{tftpdir};
|
|
} else {
|
|
$tftpdir = $globaltftpdir;
|
|
}
|
|
#if($arch ne 'x86'){
|
|
# xCAT::SvrUtils::sendmsg([1,"VMware ESX hypervisors are x86, please change the nodetype.arch value to x86 instead of $arch for $node before proceeding:
|
|
#e.g: nodech $node nodetype.arch=x86\n"]);
|
|
# return;
|
|
#}
|
|
# first make sure copycds was done:
|
|
my $custprofpath = $profile;
|
|
unless ($custprofpath =~ /^\//) {#If profile begins with a /, assume it already is a path
|
|
$custprofpath = $installroot."/custom/install/$osver/$arch/$profile";
|
|
unless(-d $custprofpath) {
|
|
$custprofpath = $installroot."/custom/install/esxi/$arch/$profile";
|
|
}
|
|
}
|
|
unless(
|
|
-r "$custprofpath/vmkboot.gz"
|
|
or -r "$custprofpath/b.z"
|
|
or -r "$custprofpath/mboot.c32"
|
|
or -r "$custprofpath/install.tgz"
|
|
or -r "$installroot/$osver/$arch/mboot.c32"
|
|
or -r "$installroot/$osver/$arch/install.tgz" ){
|
|
xCAT::SvrUtils::sendmsg([1,"Please run copycds first for $osver or create custom image in $custprofpath/"], $output_handler);
|
|
}
|
|
|
|
my @reqmods = qw/vmkboot.gz vmk.gz sys.vgz cim.vgz/; #Required modules for an image to be considered complete
|
|
if ( -r "$custprofpath/b.z" ) { #if someone hand extracts from imagedd, a different name scheme is used
|
|
@reqmods = qw/b.z k.z s.z c.z/;
|
|
}
|
|
my %mods;
|
|
foreach (@reqmods) {
|
|
$mods{$_} = 1;
|
|
}
|
|
my $shortprofname = $profile;
|
|
$shortprofname =~ s/\/\z//;
|
|
$shortprofname =~ s/.*\///;
|
|
mkpath("$tftpdir/xcat/netboot/$osver/$arch/$shortprofname/");
|
|
my $havemod=0;
|
|
unless($donetftp{$osver,$arch,$profile,$tftpdir}) {
|
|
my $srcdir = "$installroot/$osver/$arch";
|
|
my $dest = "$tftpdir/xcat/netboot/$osver/$arch/$shortprofname";
|
|
cpNetbootImages($osver,$srcdir,$dest,$custprofpath,\%mods,bootmode=>$bootmode);
|
|
if ($havemod = makecustomizedmod($osver,$dest)) {
|
|
push @reqmods,"mod.tgz";
|
|
$mods{"mod.tgz"}=1;
|
|
}
|
|
if ($osver =~ /esxi4/ and -r "$::XCATROOT/share/xcat/netboot/syslinux/mboot.c32") { #prefer xCAT patched mboot.c32 with BOOTIF for mboot
|
|
copy("$::XCATROOT/share/xcat/netboot/syslinux/mboot.c32", $dest);
|
|
} elsif (-r "$custprofpath/mboot.c32") {
|
|
copy("$custprofpath/mboot.c32", $dest);
|
|
} elsif (-r "$srcdir/mboot.c32") {
|
|
copy("$srcdir/mboot.c32", $dest);
|
|
}
|
|
if (-f "$srcdir/efiboot.img") {
|
|
copy("$srcdir/efiboot.img",$dest);
|
|
print("$srcdir/efi");
|
|
mkpath("$dest/efi");
|
|
recursion_copy("$srcdir/efi","$dest/efi");
|
|
}
|
|
$donetftp{$osver,$arch,$profile,$tftpdir} = 1;
|
|
}
|
|
my $tp = "xcat/netboot/$osver/$arch/$shortprofname";
|
|
my $kernel;
|
|
my $kcmdline;
|
|
my $append;
|
|
my $shortappend;
|
|
if ($osver =~ /esxi4/) {
|
|
my $bail=0;
|
|
foreach (@reqmods) {
|
|
unless (-r "$tftpdir/$tp/$_") {
|
|
xCAT::SvrUtils::sendmsg([1,"$_ is missing from the target destination, ensure that either copycds has been run or that $custprofpath contains this file"], $output_handler);
|
|
$bail=1; #only flag to bail, present as many messages as possible to user
|
|
}
|
|
}
|
|
if ($bail) { #if the above loop detected one or more failures, bail out
|
|
return;
|
|
}
|
|
# now make <HEX> file entry stuff
|
|
$kernel = "$tp/mboot.c32";
|
|
my $prepend;
|
|
if ($reqmods[0] eq "vmkboot.gz") {
|
|
$prepend = "$tp/vmkboot.gz";
|
|
delete $mods{"vmkboot.gz"};
|
|
$append = " --- $tp/vmk.gz";
|
|
delete $mods{"vmk.gz"};
|
|
$append .= " --- $tp/sys.vgz";
|
|
delete $mods{"sys.vgz"};
|
|
$append .= " --- $tp/cim.vgz";
|
|
delete $mods{"cim.vgz"};
|
|
} else { #the single letter style
|
|
$prepend = "$tp/b.z";
|
|
delete $mods{"b.z"};
|
|
$append = " --- $tp/k.z";
|
|
delete $mods{"k.z"};
|
|
$append .= " --- $tp/s.z";
|
|
delete $mods{"s.z"};
|
|
$append .= " --- $tp/c.z";
|
|
delete $mods{"c.z"};
|
|
}
|
|
|
|
if ($mods{"mod.tgz"}) {
|
|
$append .= " --- $tp/mod.tgz";
|
|
delete $mods{"mod.tgz"};
|
|
}
|
|
foreach (keys %mods) {
|
|
$append .= " --- $tp/$_";
|
|
}
|
|
if (defined $bpadds->{$node}->[0]->{addkcmdline}) {
|
|
my $modules;
|
|
($kcmdline,$modules) = split /---/,$bpadds->{$node}->[0]->{addkcmdline},2;
|
|
$kcmdline =~ s/#NODEATTRIB:([^:#]+):([^:#]+)#/$nodesubdata{$1}->{$node}->[0]->{$2}/eg;
|
|
if ($modules) {
|
|
$append .= " --- ".$modules;
|
|
}
|
|
$prepend .= " ".$kcmdline;
|
|
}
|
|
$append = $prepend.$append;
|
|
}
|
|
elsif ($osver =~ /esxi5/) { #do a more straightforward thing..
|
|
$kernel = "$tp/mboot.c32";
|
|
if (-r "$tftpdir/$tp/boot.cfg.$bootmode.tmpl") { #so much for straightforward..
|
|
$shortappend = "-c $tp/boot.cfg.$bootmode.$node";
|
|
} else {
|
|
$append = "-c $tp/boot.cfg.$bootmode";
|
|
}
|
|
$append .= " xcatd=$ksserver:3001";
|
|
if ($bootmode eq "install") {
|
|
$append .= " ks=http://$ksserver/install/autoinst/$node";
|
|
esxi_kickstart_from_template(node=>$node,os=>$osver,arch=>$arch,profile=>$profile);
|
|
}
|
|
if ($bootmode ne "install" and $serialconfig->{$node}) { #don't do it for install, installer croaks currently
|
|
my $comport = 1;
|
|
if (defined $serialconfig->{$node}->[0]->{serialport}) {
|
|
$comport = $serialconfig->{$node}->[0]->{serialport}+1;
|
|
$append .= " -S $comport tty2port=com$comport";
|
|
}
|
|
if (defined $serialconfig->{$node}->[0]->{serialspeed}) {
|
|
$append .= " -s ".$serialconfig->{$node}->[0]->{serialspeed}." com".$comport."_baud=".$serialconfig->{$node}->[0]->{serialspeed};
|
|
}
|
|
}
|
|
if (defined $bpadds->{$node}->[0]->{addkcmdline}) {
|
|
$append .= " ".$bpadds->{$node}->[0]->{addkcmdline};
|
|
$append =~ s/#NODEATTRIB:([^:#]+):([^:#]+)#/$nodesubdata{$1}->{$node}->[0]->{$2}/eg;
|
|
}
|
|
}
|
|
if ($shortappend) { #esxi5 user desiring to put everything in one boot config file. . .
|
|
merge_esxi5_append("$tftpdir/$tp/boot.cfg.$bootmode.tmpl",$append,"$tftpdir/$tp/boot.cfg.$bootmode.$node");
|
|
$append=$shortappend;
|
|
}
|
|
$output_handler->({node=>[{name=>[$node],'_addkcmdlinehandled'=>[1]}]});
|
|
|
|
|
|
|
|
$bptab->setNodeAttribs(
|
|
$node,
|
|
{
|
|
kernel => $kernel,
|
|
initrd => "",
|
|
kcmdline => $append
|
|
}
|
|
);
|
|
} # end of node loop
|
|
|
|
}
|
|
# this is where we extract the netboot images out of the copied ISO image
|
|
sub cpNetbootImages {
|
|
my $osver = shift;
|
|
my $srcDir = shift;
|
|
my $destDir = shift;
|
|
my $overridedir = shift;
|
|
my $modulestoadd = shift;
|
|
my %parmargs = @_;
|
|
my $bootmode="stateless";
|
|
if ($parmargs{bootmode}) { $bootmode = $parmargs{bootmode} }
|
|
my $tmpDir = "/tmp/xcat.$$";
|
|
if($osver =~ /esxi4/){
|
|
# we don't want to go through this all the time, so if its already
|
|
# there we're not going to extract:
|
|
unless( -r "$destDir/vmk.gz"
|
|
and -r "$destDir/vmkboot.gz"
|
|
and -r "$destDir/sys.vgz"
|
|
and -r "$destDir/cim.vgz"
|
|
and -r "$destDir/cimstg.tgz"
|
|
){
|
|
if (-r "$srcDir/image.tgz") { #it still may work without image.tgz if profile customization has everything replaced
|
|
mkdir($tmpDir);
|
|
chdir($tmpDir);
|
|
xCAT::SvrUtils::sendmsg("extracting netboot files from OS image. This may take about a minute or two...hopefully you have ~1GB free in your /tmp dir\n", $output_handler);
|
|
my $cmd = "tar zxf $srcDir/image.tgz";
|
|
if(system($cmd)){
|
|
xCAT::SvrUtils::sendmsg([1,"Unable to extract $srcDir/image.tgz\n"], $output_handler);
|
|
}
|
|
# this has the big image and may take a while.
|
|
# this should now create:
|
|
# /tmp/xcat.1234/usr/lib/vmware/installer/VMware-VMvisor-big-164009-x86_64.dd.bz2 or some other version. We need to extract partition 5 from it.
|
|
system("bunzip2 $tmpDir/usr/lib/vmware/installer/*bz2");
|
|
xCAT::SvrUtils::sendmsg("finished extracting, now copying files...\n", $output_handler);
|
|
|
|
# now we need to get partition 5 which has the installation goods in it.
|
|
my $scmd = "fdisk -lu $tmpDir/usr/lib/vmware/installer/*dd 2>&1 | grep dd5 | awk '{print \$2}'";
|
|
my $sector = `$scmd`;
|
|
chomp($sector);
|
|
my $offset = $sector * 512;
|
|
mkdir "/mnt/xcat";
|
|
my $mntcmd = "mount $tmpDir/usr/lib/vmware/installer/*dd /mnt/xcat -o loop,offset=$offset";
|
|
if(system($mntcmd)){
|
|
xCAT::SvrUtils::sendmsg([1,"unable to mount partition 5 of the ESX netboot image to /mnt/xcat"], $output_handler);
|
|
return;
|
|
}
|
|
|
|
if (! -d $destDir) {
|
|
if ( -e $destDir ) {
|
|
xCAT::SvrUtils::sendmsg([1,"Could not copy netboot contents to $destDir, it exists but is not currently a directory"], $output_handler);
|
|
return;
|
|
}
|
|
mkpath($destDir);
|
|
}
|
|
|
|
if(system("cp /mnt/xcat/* $destDir/")){
|
|
xCAT::SvrUtils::sendmsg([1,"Could not copy netboot contents to $destDir"], $output_handler);
|
|
chdir("/");
|
|
system("umount /mnt/xcat");
|
|
return;
|
|
}
|
|
chdir("/tmp");
|
|
system("umount /mnt/xcat");
|
|
system("rm -rf $tmpDir");
|
|
} elsif (-r "$srcDir/cim.vgz" and -r "$srcDir/vmkernel.gz" and -r "$srcDir/vmkboot.gz" and -r "$srcDir/sys.vgz") {
|
|
use File::Basename;
|
|
if (! -d $destDir) {
|
|
mkpath($destDir);
|
|
}
|
|
#In ESXI 4.1, the above breaks, this seems to work, much simpler too
|
|
foreach ("$srcDir/cim.vgz","$srcDir/vmkernel.gz","$srcDir/vmkboot.gz","$srcDir/sys.vgz","$srcDir/sys.vgz") {
|
|
my $mod = scalar fileparse($_);
|
|
if ($mod =~ /vmkernel.gz/) {
|
|
copy($_,"$destDir/vmk.gz") or xCAT::SvrUtils::sendmsg([1,"Could not copy netboot contents from $_ to $destDir/$mod"], $output_handler);
|
|
} else {
|
|
copy($_,"$destDir/$mod") or xCAT::SvrUtils::sendmsg([1,"Could not copy netboot contents from $_ to $destDir/$mod"], $output_handler);
|
|
}
|
|
}
|
|
|
|
}
|
|
}
|
|
|
|
#this is the override directory if there is one, otherwise it's actually the default dir
|
|
if (-d $overridedir) {
|
|
mkdir($overridedir);
|
|
}
|
|
|
|
#Copy over all modules
|
|
use File::Basename;
|
|
foreach (glob "$overridedir/*") {
|
|
my $mod = scalar fileparse($_);
|
|
if ($mod =~ /gz\z/ and $mod !~ /pkgdb.tgz/ and $mod !~ /vmkernel.gz/) {
|
|
$modulestoadd->{$mod}=1;
|
|
copy($_,"$destDir/$mod") or xCAT::SvrUtils::sendmsg([1,"Could not copy netboot contents from $overridedir to $destDir"], $output_handler);
|
|
} elsif ($mod =~ /vmkernel.gz/) {
|
|
$modulestoadd->{"vmk.gz"}=1;
|
|
copy($_,"$destDir/vmk.gz") or xCAT::SvrUtils::sendmsg([1,"Could not copy netboot contents from $overridedir to $destDir"], $output_handler);
|
|
}
|
|
}
|
|
|
|
}elsif ($osver =~ /esxi5/) { #we need boot.cfg.stateles
|
|
my @filestocopy = ("boot.cfg.$bootmode");
|
|
if (-r "$srcDir/boot.cfg.$bootmode" or -r "$overridedir/boot.cfg.$bootmode") {
|
|
@filestocopy = ("boot.cfg.$bootmode");
|
|
} elsif (-r "$srcDir/boot.cfg.$bootmode.tmpl" or -r "$overridedir/boot.cfg.$bootmode.tmpl") {
|
|
@filestocopy = ("boot.cfg.$bootmode.tmpl");
|
|
} else {
|
|
xCAT::SvrUtils::sendmsg([1,"$srcDir is missing boot.cfg.$bootmode file required for $bootmode boot"], $output_handler);
|
|
return;
|
|
}
|
|
my $statelesscfg;
|
|
if (-r "$overridedir/boot.cfg.$bootmode.tmpl") {
|
|
open ($statelesscfg,"<","$overridedir/boot.cfg.$bootmode.tmpl");
|
|
@filestocopy = ("boot.cfg.$bootmode.tmpl");
|
|
} elsif (-r "$overridedir/boot.cfg.$bootmode") {
|
|
open ($statelesscfg,"<","$overridedir/boot.cfg.$bootmode");
|
|
} elsif (-r "$srcDir/boot.cfg.$bootmode.tmpl") {
|
|
@filestocopy = ("boot.cfg.$bootmode.tmpl");
|
|
open ($statelesscfg,"<","$srcDir/boot.cfg.$bootmode.tmpl");
|
|
} elsif (-r "$srcDir/boot.cfg.$bootmode") {
|
|
open ($statelesscfg,"<","$srcDir/boot.cfg.$bootmode");
|
|
} else {
|
|
die "boot.cfg.$bootmode was missing from $srcDir???";
|
|
}
|
|
my @statelesscfg=<$statelesscfg>;
|
|
|
|
foreach (@statelesscfg) { #search for files specified by the boot cfg and pull them in
|
|
if (/^kernel=(.*)/) {
|
|
push @filestocopy,$1;
|
|
} elsif (/^modules=(.*)/) {
|
|
foreach (split / --- /,$1) {
|
|
s/^\s*//;
|
|
s/\s.*//;
|
|
push @filestocopy,$_;
|
|
}
|
|
}
|
|
}
|
|
#now that we have a list, do the copy (mostly redundant, but PXE needs them tftp accessible)
|
|
foreach (@filestocopy) {
|
|
chomp;
|
|
s/ *\z//;
|
|
my $mod = scalar fileparse($_);
|
|
if (-r "$overridedir/$mod") {
|
|
copyIfNewer("$overridedir/$mod","$destDir/$mod") or xCAT::SvrUtils::sendmsg([1,"Could not copy netboot contents from $overridedir/$mod to $destDir/$mod, $!"], $output_handler);
|
|
} elsif (-r "$srcDir/$mod") {
|
|
copyIfNewer($srcDir."/".$mod,"$destDir/$mod") or xCAT::SvrUtils::sendmsg([1,"Could not copy netboot contents from $srcDir/$mod to $destDir/$mod, $!"], $output_handler);
|
|
} elsif ($mod ne "xcatmod.tgz") {
|
|
xCAT::SvrUtils::sendmsg([1,"Could not copy netboot contents from $srcDir/$mod to $destDir/$mod, $srcDir/$mod not found"], $output_handler);
|
|
}
|
|
}
|
|
} else {
|
|
xCAT::SvrUtils::sendmsg([1,"VMware $osver is not supported for netboot"], $output_handler);
|
|
}
|
|
|
|
}
|
|
|
|
sub copyIfNewer {
|
|
my $source = shift;
|
|
my $dest = shift;
|
|
if (! -e $dest or -C $source < -C $dest) {
|
|
return copy($source,$dest);
|
|
}
|
|
return 1;
|
|
}
|
|
|
|
# compares nfs target described by parameters to every share mounted by target hypervisor
|
|
# returns 1 if matching datastore is present and 0 otherwise
|
|
sub match_nfs_datastore {
|
|
my ($host, $path, $hypconn) = @_;
|
|
|
|
die "esx plugin bug: no host provided for match_datastore" unless defined $host;
|
|
die "esx plugin bug: no path provided for match_datastore" unless defined $path;
|
|
|
|
my @ip;
|
|
|
|
eval {
|
|
if ($host =~ m/\b(?:(?:25[0-5]|2[0-4][0-9]|[01]?[0-9][0-9]?)\.){3}(?:25[0-5]|2[0-4][0-9]|[01]?[0-9][0-9]?)\//) {
|
|
use Socket;
|
|
|
|
@ip = ( $host );
|
|
$host = gethostbyaddr(inet_aton($host, AF_INET), AF_INET);
|
|
} else {
|
|
use Socket;
|
|
|
|
(undef, undef, undef, undef, @ip) = gethostbyname($host);
|
|
|
|
my @ip_ntoa = ();
|
|
foreach (@ip) {
|
|
push (@ip_ntoa, inet_ntoa($_));
|
|
}
|
|
@ip = @ip_ntoa;
|
|
}
|
|
|
|
};
|
|
|
|
if ($@) {
|
|
die "error while resolving datastore host: $@\n";
|
|
}
|
|
|
|
my %viewcrit = (
|
|
view_type => 'HostSystem',
|
|
properties => [ 'config.fileSystemVolume' ],
|
|
);
|
|
|
|
my $dsviews = $hypconn->find_entity_views(%viewcrit);
|
|
|
|
foreach (@$dsviews) {
|
|
foreach my $mount (@{$_->get_property('config.fileSystemVolume.mountInfo')}) {
|
|
next unless $mount->{'volume'}{'type'} eq 'NFS';
|
|
|
|
my $hostMatch = 0;
|
|
HOSTMATCH: foreach (@ip, $host) {
|
|
next HOSTMATCH unless $mount->{'volume'}{'remoteHost'} eq $_;
|
|
|
|
$hostMatch = 1;
|
|
last HOSTMATCH;
|
|
}
|
|
next unless $hostMatch;
|
|
|
|
next unless $mount->{'volume'}{'remotePath'} eq $path;
|
|
|
|
return 1;
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
1;
|
|
# vi: set ts=4 sw=4 filetype=perl:
|