Merge pull request #167 from mattkirby/migrate_at_checkout

Improve vmpooler scheduling logic
This commit is contained in:
mattkirby 2016-11-22 11:22:12 -08:00 committed by GitHub
commit 705e5d26d8
8 changed files with 526 additions and 293 deletions

View file

@ -16,6 +16,7 @@ gem 'statsd-ruby', '>= 1.3.0', :require => 'statsd'
# Test deps # Test deps
group :test do group :test do
gem 'mock_redis', '>= 0.17.0'
gem 'rack-test', '>= 0.6' gem 'rack-test', '>= 0.6'
gem 'rspec', '>= 3.2' gem 'rspec', '>= 3.2'
gem 'simplecov', '>= 0.11.2' gem 'simplecov', '>= 0.11.2'

View file

@ -55,6 +55,7 @@ module Vmpooler
def account_for_starting_vm(template, vm) def account_for_starting_vm(template, vm)
backend.sadd('vmpooler__running__' + template, vm) backend.sadd('vmpooler__running__' + template, vm)
backend.sadd('vmpooler__migrating__' + template, vm)
backend.hset('vmpooler__active__' + template, vm, Time.now) backend.hset('vmpooler__active__' + template, vm, Time.now)
backend.hset('vmpooler__vm__' + vm, 'checkout', Time.now) backend.hset('vmpooler__vm__' + vm, 'checkout', Time.now)

View file

@ -20,40 +20,59 @@ module Vmpooler
end end
# Check the state of a VM # Check the state of a VM
def check_pending_vm(vm, pool, timeout) def check_pending_vm(vm, pool, timeout, vsphere)
Thread.new do Thread.new do
_check_pending_vm(vm, pool, timeout) _check_pending_vm(vm, pool, timeout, vsphere)
end end
end end
def _check_pending_vm(vm, pool, timeout) def open_socket(host, domain=nil, timeout=5, port=22, &block)
host = $vsphere[pool].find_vm(vm) Timeout.timeout(timeout) do
target_host = host
if host target_host = "#{host}.#{domain}" if domain
sock = TCPSocket.new target_host, port
begin begin
Timeout.timeout(5) do yield sock if block_given?
TCPSocket.new vm, 22 ensure
sock.close
end end
end
end
def _check_pending_vm(vm, pool, timeout, vsphere)
host = vsphere.find_vm(vm)
if ! host
fail_pending_vm(vm, pool, timeout, false)
return
end
open_socket vm
move_pending_vm_to_ready(vm, pool, host) move_pending_vm_to_ready(vm, pool, host)
rescue rescue
fail_pending_vm(vm, pool, timeout) fail_pending_vm(vm, pool, timeout)
end end
else
fail_pending_vm(vm, pool, timeout) def remove_nonexistent_vm(vm, pool)
end $redis.srem("vmpooler__pending__#{pool}", vm)
$logger.log('d', "[!] [#{pool}] '#{vm}' no longer exists. Removing from pending.")
end end
def fail_pending_vm(vm, pool, timeout) def fail_pending_vm(vm, pool, timeout, exists=true)
clone_stamp = $redis.hget('vmpooler__vm__' + vm, 'clone') clone_stamp = $redis.hget("vmpooler__vm__#{vm}", 'clone')
return if ! clone_stamp
if (clone_stamp) &&
(((Time.now - Time.parse(clone_stamp)) / 60) > timeout)
time_since_clone = (Time.now - Time.parse(clone_stamp)) / 60
if time_since_clone > timeout
if exists
$redis.smove('vmpooler__pending__' + pool, 'vmpooler__completed__' + pool, vm) $redis.smove('vmpooler__pending__' + pool, 'vmpooler__completed__' + pool, vm)
$logger.log('d', "[!] [#{pool}] '#{vm}' marked as 'failed' after #{timeout} minutes") $logger.log('d', "[!] [#{pool}] '#{vm}' marked as 'failed' after #{timeout} minutes")
else
remove_nonexistent_vm(vm, pool)
end end
end end
rescue => err
$logger.log('d', "Fail pending VM failed with an error: #{err}")
end
def move_pending_vm_to_ready(vm, pool, host) def move_pending_vm_to_ready(vm, pool, host)
if (host.summary) && if (host.summary) &&
@ -76,7 +95,7 @@ module Vmpooler
end end
end end
def check_ready_vm(vm, pool, ttl) def check_ready_vm(vm, pool, ttl, vsphere)
Thread.new do Thread.new do
if ttl > 0 if ttl > 0
if (((Time.now - host.runtime.bootTime) / 60).to_s[/^\d+\.\d{1}/].to_f) > ttl if (((Time.now - host.runtime.bootTime) / 60).to_s[/^\d+\.\d{1}/].to_f) > ttl
@ -94,8 +113,7 @@ module Vmpooler
$redis.hset('vmpooler__vm__' + vm, 'check', Time.now) $redis.hset('vmpooler__vm__' + vm, 'check', Time.now)
host = $vsphere[pool].find_vm(vm) || host = vsphere.find_vm(vm)
$vsphere[pool].find_vm_heavy(vm)[vm]
if host if host
if if
@ -124,26 +142,26 @@ module Vmpooler
end end
begin begin
Timeout.timeout(5) do open_socket vm
TCPSocket.new vm, 22
end
rescue rescue
if $redis.smove('vmpooler__ready__' + pool, 'vmpooler__completed__' + pool, vm) if $redis.smove('vmpooler__ready__' + pool, 'vmpooler__completed__' + pool, vm)
$logger.log('d', "[!] [#{pool}] '#{vm}' is unreachable, removed from 'ready' queue") $logger.log('d', "[!] [#{pool}] '#{vm}' is unreachable, removed from 'ready' queue")
else
$logger.log('d', "[!] [#{pool}] '#{vm}' is unreachable, and failed to remove from 'ready' queue")
end end
end end
end end
end end
end end
def check_running_vm(vm, pool, ttl) def check_running_vm(vm, pool, ttl, vsphere)
Thread.new do Thread.new do
_check_running_vm(vm, pool, ttl) _check_running_vm(vm, pool, ttl, vsphere)
end end
end end
def _check_running_vm(vm, pool, ttl) def _check_running_vm(vm, pool, ttl, vsphere)
host = $vsphere[pool].find_vm(vm) host = vsphere.find_vm(vm)
if host if host
queue_from, queue_to = 'running', 'completed' queue_from, queue_to = 'running', 'completed'
@ -167,8 +185,9 @@ module Vmpooler
end end
# Clone a VM # Clone a VM
def clone_vm(template, folder, datastore, target) def clone_vm(template, folder, datastore, target, vsphere)
Thread.new do Thread.new do
begin
vm = {} vm = {}
if template =~ /\// if template =~ /\//
@ -177,7 +196,7 @@ module Vmpooler
end end
if templatefolders if templatefolders
vm[vm['template']] = $vsphere[vm['template']].find_folder(templatefolders.join('/')).find(vm['template']) vm[vm['template']] = vsphere.find_folder(templatefolders.join('/')).find(vm['template'])
else else
fail 'Please provide a full path to the template' fail 'Please provide a full path to the template'
end end
@ -213,14 +232,14 @@ module Vmpooler
# Choose a clone target # Choose a clone target
if target if target
$clone_target = $vsphere[vm['template']].find_least_used_host(target) $clone_target = vsphere.find_least_used_host(target)
elsif $config[:config]['clone_target'] elsif $config[:config]['clone_target']
$clone_target = $vsphere[vm['template']].find_least_used_host($config[:config]['clone_target']) $clone_target = vsphere.find_least_used_host($config[:config]['clone_target'])
end end
# Put the VM in the specified folder and resource pool # Put the VM in the specified folder and resource pool
relocateSpec = RbVmomi::VIM.VirtualMachineRelocateSpec( relocateSpec = RbVmomi::VIM.VirtualMachineRelocateSpec(
datastore: $vsphere[vm['template']].find_datastore(datastore), datastore: vsphere.find_datastore(datastore),
host: $clone_target, host: $clone_target,
diskMoveType: :moveChildMostDiskBacking diskMoveType: :moveChildMostDiskBacking
) )
@ -239,7 +258,7 @@ module Vmpooler
begin begin
start = Time.now start = Time.now
vm[vm['template']].CloneVM_Task( vm[vm['template']].CloneVM_Task(
folder: $vsphere[vm['template']].find_folder(folder), folder: vsphere.find_folder(folder),
name: vm['hostname'], name: vm['hostname'],
spec: spec spec: spec
).wait_for_completion ).wait_for_completion
@ -249,19 +268,22 @@ module Vmpooler
$redis.hset('vmpooler__vm__' + vm['hostname'], 'clone_time', finish) $redis.hset('vmpooler__vm__' + vm['hostname'], 'clone_time', finish)
$logger.log('s', "[+] [#{vm['template']}] '#{vm['hostname']}' cloned from '#{vm['template']}' in #{finish} seconds") $logger.log('s', "[+] [#{vm['template']}] '#{vm['hostname']}' cloned from '#{vm['template']}' in #{finish} seconds")
rescue rescue => err
$logger.log('s', "[!] [#{vm['template']}] '#{vm['hostname']}' clone appears to have failed") $logger.log('s', "[!] [#{vm['template']}] '#{vm['hostname']}' clone failed with an error: #{err}")
$redis.srem('vmpooler__pending__' + vm['template'], vm['hostname']) $redis.srem('vmpooler__pending__' + vm['template'], vm['hostname'])
end end
$redis.decr('vmpooler__tasks__clone') $redis.decr('vmpooler__tasks__clone')
$metrics.timing("clone.#{vm['template']}", finish) $metrics.timing("clone.#{vm['template']}", finish)
rescue => err
$logger.log('s', "[!] [#{vm['template']}] '#{vm['hostname']}' failed while preparing to clone with an error: #{err}")
end
end end
end end
# Destroy a VM # Destroy a VM
def destroy_vm(vm, pool) def destroy_vm(vm, pool, vsphere)
Thread.new do Thread.new do
$redis.srem('vmpooler__completed__' + pool, vm) $redis.srem('vmpooler__completed__' + pool, vm)
$redis.hdel('vmpooler__active__' + pool, vm) $redis.hdel('vmpooler__active__' + pool, vm)
@ -270,8 +292,7 @@ module Vmpooler
# Auto-expire metadata key # Auto-expire metadata key
$redis.expire('vmpooler__vm__' + vm, ($config[:redis]['data_ttl'].to_i * 60 * 60)) $redis.expire('vmpooler__vm__' + vm, ($config[:redis]['data_ttl'].to_i * 60 * 60))
host = $vsphere[pool].find_vm(vm) || host = vsphere.find_vm(vm)
$vsphere[pool].find_vm_heavy(vm)[vm]
if host if host
start = Time.now start = Time.now
@ -294,15 +315,14 @@ module Vmpooler
end end
end end
def create_vm_disk(vm, disk_size) def create_vm_disk(vm, disk_size, vsphere)
Thread.new do Thread.new do
_create_vm_disk(vm, disk_size) _create_vm_disk(vm, disk_size, vsphere)
end end
end end
def _create_vm_disk(vm, disk_size) def _create_vm_disk(vm, disk_size, vsphere)
host = $vsphere['disk_manager'].find_vm(vm) || host = vsphere.find_vm(vm)
$vsphere['disk_manager'].find_vm_heavy(vm)[vm]
if (host) && ((! disk_size.nil?) && (! disk_size.empty?) && (disk_size.to_i > 0)) if (host) && ((! disk_size.nil?) && (! disk_size.empty?) && (disk_size.to_i > 0))
$logger.log('s', "[ ] [disk_manager] '#{vm}' is attaching a #{disk_size}gb disk") $logger.log('s', "[ ] [disk_manager] '#{vm}' is attaching a #{disk_size}gb disk")
@ -319,7 +339,7 @@ module Vmpooler
end end
if ((! datastore.nil?) && (! datastore.empty?)) if ((! datastore.nil?) && (! datastore.empty?))
$vsphere['disk_manager'].add_disk(host, disk_size, datastore) vsphere.add_disk(host, disk_size, datastore)
rdisks = $redis.hget('vmpooler__vm__' + vm, 'disk') rdisks = $redis.hget('vmpooler__vm__' + vm, 'disk')
disks = rdisks ? rdisks.split(':') : [] disks = rdisks ? rdisks.split(':') : []
@ -335,15 +355,14 @@ module Vmpooler
end end
end end
def create_vm_snapshot(vm, snapshot_name) def create_vm_snapshot(vm, snapshot_name, vsphere)
Thread.new do Thread.new do
_create_vm_snapshot(vm, snapshot_name) _create_vm_snapshot(vm, snapshot_name, vsphere)
end end
end end
def _create_vm_snapshot(vm, snapshot_name) def _create_vm_snapshot(vm, snapshot_name, vsphere)
host = $vsphere['snapshot_manager'].find_vm(vm) || host = vsphere.find_vm(vm)
$vsphere['snapshot_manager'].find_vm_heavy(vm)[vm]
if (host) && ((! snapshot_name.nil?) && (! snapshot_name.empty?)) if (host) && ((! snapshot_name.nil?) && (! snapshot_name.empty?))
$logger.log('s', "[ ] [snapshot_manager] '#{vm}' is being snapshotted") $logger.log('s', "[ ] [snapshot_manager] '#{vm}' is being snapshotted")
@ -365,18 +384,17 @@ module Vmpooler
end end
end end
def revert_vm_snapshot(vm, snapshot_name) def revert_vm_snapshot(vm, snapshot_name, vsphere)
Thread.new do Thread.new do
_revert_vm_snapshot(vm, snapshot_name) _revert_vm_snapshot(vm, snapshot_name, vsphere)
end end
end end
def _revert_vm_snapshot(vm, snapshot_name) def _revert_vm_snapshot(vm, snapshot_name, vsphere)
host = $vsphere['snapshot_manager'].find_vm(vm) || host = vsphere.find_vm(vm)
$vsphere['snapshot_manager'].find_vm_heavy(vm)[vm]
if host if host
snapshot = $vsphere['snapshot_manager'].find_snapshot(host, snapshot_name) snapshot = vsphere.find_snapshot(host, snapshot_name)
if snapshot if snapshot
$logger.log('s', "[ ] [snapshot_manager] '#{vm}' is being reverted to snapshot '#{snapshot_name}'") $logger.log('s', "[ ] [snapshot_manager] '#{vm}' is being reverted to snapshot '#{snapshot_name}'")
@ -395,23 +413,23 @@ module Vmpooler
def check_disk_queue def check_disk_queue
$logger.log('d', "[*] [disk_manager] starting worker thread") $logger.log('d', "[*] [disk_manager] starting worker thread")
$vsphere['disk_manager'] ||= Vmpooler::VsphereHelper.new $vsphere['disk_manager'] ||= Vmpooler::VsphereHelper.new $config[:vsphere]
$threads['disk_manager'] = Thread.new do $threads['disk_manager'] = Thread.new do
loop do loop do
_check_disk_queue _check_disk_queue $vsphere['disk_manager']
sleep(5) sleep(5)
end end
end end
end end
def _check_disk_queue def _check_disk_queue(vsphere)
vm = $redis.spop('vmpooler__tasks__disk') vm = $redis.spop('vmpooler__tasks__disk')
unless vm.nil? unless vm.nil?
begin begin
vm_name, disk_size = vm.split(':') vm_name, disk_size = vm.split(':')
create_vm_disk(vm_name, disk_size) create_vm_disk(vm_name, disk_size, vsphere)
rescue rescue
$logger.log('s', "[!] [disk_manager] disk creation appears to have failed") $logger.log('s', "[!] [disk_manager] disk creation appears to have failed")
end end
@ -421,23 +439,23 @@ module Vmpooler
def check_snapshot_queue def check_snapshot_queue
$logger.log('d', "[*] [snapshot_manager] starting worker thread") $logger.log('d', "[*] [snapshot_manager] starting worker thread")
$vsphere['snapshot_manager'] ||= Vmpooler::VsphereHelper.new $vsphere['snapshot_manager'] ||= Vmpooler::VsphereHelper.new $config[:vsphere]
$threads['snapshot_manager'] = Thread.new do $threads['snapshot_manager'] = Thread.new do
loop do loop do
_check_snapshot_queue _check_snapshot_queue $vsphere['snapshot_manager']
sleep(5) sleep(5)
end end
end end
end end
def _check_snapshot_queue def _check_snapshot_queue(vsphere)
vm = $redis.spop('vmpooler__tasks__snapshot') vm = $redis.spop('vmpooler__tasks__snapshot')
unless vm.nil? unless vm.nil?
begin begin
vm_name, snapshot_name = vm.split(':') vm_name, snapshot_name = vm.split(':')
create_vm_snapshot(vm_name, snapshot_name) create_vm_snapshot(vm_name, snapshot_name, vsphere)
rescue rescue
$logger.log('s', "[!] [snapshot_manager] snapshot appears to have failed") $logger.log('s', "[!] [snapshot_manager] snapshot appears to have failed")
end end
@ -448,31 +466,100 @@ module Vmpooler
unless vm.nil? unless vm.nil?
begin begin
vm_name, snapshot_name = vm.split(':') vm_name, snapshot_name = vm.split(':')
revert_vm_snapshot(vm_name, snapshot_name) revert_vm_snapshot(vm_name, snapshot_name, vsphere)
rescue rescue
$logger.log('s', "[!] [snapshot_manager] snapshot revert appears to have failed") $logger.log('s', "[!] [snapshot_manager] snapshot revert appears to have failed")
end end
end end
end end
def migration_limit(migration_limit)
# Returns migration_limit setting when enabled
return false if migration_limit == 0 || ! migration_limit
migration_limit if migration_limit >= 1
end
def migrate_vm(vm, pool, vsphere)
Thread.new do
_migrate_vm(vm, pool, vsphere)
end
end
def _migrate_vm(vm, pool, vsphere)
begin
$redis.srem('vmpooler__migrating__' + pool, vm)
vm_object = vsphere.find_vm(vm)
parent_host, parent_host_name = get_vm_host_info(vm_object)
migration_limit = migration_limit $config[:config]['migration_limit']
migration_count = $redis.scard('vmpooler__migration')
if ! migration_limit
$logger.log('s', "[ ] [#{pool}] '#{vm}' is running on #{parent_host_name}")
return
else
if migration_count >= migration_limit
$logger.log('s', "[ ] [#{pool}] '#{vm}' is running on #{parent_host_name}. No migration will be evaluated since the migration_limit has been reached")
return
else
$redis.sadd('vmpooler__migration', vm)
host, host_name = vsphere.find_least_used_compatible_host(vm_object)
if host == parent_host
$logger.log('s', "[ ] [#{pool}] No migration required for '#{vm}' running on #{parent_host_name}")
else
finish = migrate_vm_and_record_timing(vm_object, vm, host, vsphere)
$logger.log('s', "[>] [#{pool}] '#{vm}' migrated from #{parent_host_name} to #{host_name} in #{finish} seconds")
end
remove_vmpooler_migration_vm(pool, vm)
end
end
rescue => err
$logger.log('s', "[x] [#{pool}] '#{vm}' migration failed with an error: #{err}")
remove_vmpooler_migration_vm(pool, vm)
end
end
def get_vm_host_info(vm_object)
parent_host = vm_object.summary.runtime.host
[parent_host, parent_host.name]
end
def remove_vmpooler_migration_vm(pool, vm)
begin
$redis.srem('vmpooler__migration', vm)
rescue => err
$logger.log('s', "[x] [#{pool}] '#{vm}' removal from vmpooler__migration failed with an error: #{err}")
end
end
def migrate_vm_and_record_timing(vm_object, vm_name, host, vsphere)
start = Time.now
vsphere.migrate_vm_host(vm_object, host)
finish = '%.2f' % (Time.now - start)
$metrics.timing("migrate.#{vm_name}", finish)
checkout_to_migration = '%.2f' % (Time.now - Time.parse($redis.hget("vmpooler__vm__#{vm_name}", 'checkout')))
$redis.hset("vmpooler__vm__#{vm_name}", 'migration_time', finish)
$redis.hset("vmpooler__vm__#{vm_name}", 'checkout_to_migration', checkout_to_migration)
finish
end
def check_pool(pool) def check_pool(pool)
$logger.log('d', "[*] [#{pool['name']}] starting worker thread") $logger.log('d', "[*] [#{pool['name']}] starting worker thread")
$vsphere[pool['name']] ||= Vmpooler::VsphereHelper.new $vsphere[pool['name']] ||= Vmpooler::VsphereHelper.new $config[:vsphere]
$threads[pool['name']] = Thread.new do $threads[pool['name']] = Thread.new do
loop do loop do
_check_pool(pool) _check_pool(pool, $vsphere[pool['name']])
sleep(5) sleep(5)
end end
end end
end end
def _check_pool(pool) def _check_pool(pool, vsphere)
# INVENTORY # INVENTORY
inventory = {} inventory = {}
begin begin
base = $vsphere[pool['name']].find_folder(pool['folder']) base = vsphere.find_folder(pool['folder'])
base.childEntity.each do |vm| base.childEntity.each do |vm|
if if
@ -480,7 +567,8 @@ module Vmpooler
(! $redis.sismember('vmpooler__ready__' + pool['name'], vm['name'])) && (! $redis.sismember('vmpooler__ready__' + pool['name'], vm['name'])) &&
(! $redis.sismember('vmpooler__pending__' + pool['name'], vm['name'])) && (! $redis.sismember('vmpooler__pending__' + pool['name'], vm['name'])) &&
(! $redis.sismember('vmpooler__completed__' + pool['name'], vm['name'])) && (! $redis.sismember('vmpooler__completed__' + pool['name'], vm['name'])) &&
(! $redis.sismember('vmpooler__discovered__' + pool['name'], vm['name'])) (! $redis.sismember('vmpooler__discovered__' + pool['name'], vm['name'])) &&
(! $redis.sismember('vmpooler__migrating__' + pool['name'], vm['name']))
$redis.sadd('vmpooler__discovered__' + pool['name'], vm['name']) $redis.sadd('vmpooler__discovered__' + pool['name'], vm['name'])
@ -493,82 +581,97 @@ module Vmpooler
end end
# RUNNING # RUNNING
$redis.smembers('vmpooler__running__' + pool['name']).each do |vm| $redis.smembers("vmpooler__running__#{pool['name']}").each do |vm|
if inventory[vm] if inventory[vm]
begin begin
check_running_vm(vm, pool['name'], $redis.hget('vmpooler__vm__' + vm, 'lifetime') || $config[:config]['vm_lifetime'] || 12) vm_lifetime = $redis.hget('vmpooler__vm__' + vm, 'lifetime') || $config[:config]['vm_lifetime'] || 12
check_running_vm(vm, pool['name'], vm_lifetime, vsphere)
rescue rescue
end end
end end
end end
# READY # READY
$redis.smembers('vmpooler__ready__' + pool['name']).each do |vm| $redis.smembers("vmpooler__ready__#{pool['name']}").each do |vm|
if inventory[vm] if inventory[vm]
begin begin
check_ready_vm(vm, pool['name'], pool['ready_ttl'] || 0) check_ready_vm(vm, pool['name'], pool['ready_ttl'] || 0, vsphere)
rescue rescue
end end
end end
end end
# PENDING # PENDING
$redis.smembers('vmpooler__pending__' + pool['name']).each do |vm| $redis.smembers("vmpooler__pending__#{pool['name']}").each do |vm|
pool_timeout = pool['timeout'] || $config[:config]['timeout'] || 15
if inventory[vm] if inventory[vm]
begin begin
check_pending_vm(vm, pool['name'], pool['timeout'] || $config[:config]['timeout'] || 15) check_pending_vm(vm, pool['name'], pool_timeout, vsphere)
rescue rescue
end end
else
fail_pending_vm(vm, pool['name'], pool_timeout, false)
end end
end end
# COMPLETED # COMPLETED
$redis.smembers('vmpooler__completed__' + pool['name']).each do |vm| $redis.smembers("vmpooler__completed__#{pool['name']}").each do |vm|
if inventory[vm] if inventory[vm]
begin begin
destroy_vm(vm, pool['name']) destroy_vm(vm, pool['name'], vsphere)
rescue rescue
$logger.log('s', "[!] [#{pool['name']}] '#{vm}' destroy appears to have failed") $logger.log('s', "[!] [#{pool['name']}] '#{vm}' destroy appears to have failed")
$redis.srem('vmpooler__completed__' + pool['name'], vm) $redis.srem("vmpooler__completed__#{pool['name']}", vm)
$redis.hdel('vmpooler__active__' + pool['name'], vm) $redis.hdel("vmpooler__active__#{pool['name']}", vm)
$redis.del('vmpooler__vm__' + vm) $redis.del("vmpooler__vm__#{vm}")
end end
else else
$logger.log('s', "[!] [#{pool['name']}] '#{vm}' not found in inventory, removed from 'completed' queue") $logger.log('s', "[!] [#{pool['name']}] '#{vm}' not found in inventory, removed from 'completed' queue")
$redis.srem('vmpooler__completed__' + pool['name'], vm) $redis.srem("vmpooler__completed__#{pool['name']}", vm)
$redis.hdel('vmpooler__active__' + pool['name'], vm) $redis.hdel("vmpooler__active__#{pool['name']}", vm)
$redis.del('vmpooler__vm__' + vm) $redis.del("vmpooler__vm__#{vm}")
end end
end end
# DISCOVERED # DISCOVERED
$redis.smembers('vmpooler__discovered__' + pool['name']).each do |vm| $redis.smembers("vmpooler__discovered__#{pool['name']}").each do |vm|
%w(pending ready running completed).each do |queue| %w(pending ready running completed).each do |queue|
if $redis.sismember('vmpooler__' + queue + '__' + pool['name'], vm) if $redis.sismember("vmpooler__#{queue}__#{pool['name']}", vm)
$logger.log('d', "[!] [#{pool['name']}] '#{vm}' found in '#{queue}', removed from 'discovered' queue") $logger.log('d', "[!] [#{pool['name']}] '#{vm}' found in '#{queue}', removed from 'discovered' queue")
$redis.srem('vmpooler__discovered__' + pool['name'], vm) $redis.srem("vmpooler__discovered__#{pool['name']}", vm)
end end
end end
if $redis.sismember('vmpooler__discovered__' + pool['name'], vm) if $redis.sismember("vmpooler__discovered__#{pool['name']}", vm)
$redis.smove('vmpooler__discovered__' + pool['name'], 'vmpooler__completed__' + pool['name'], vm) $redis.smove("vmpooler__discovered__#{pool['name']}", "vmpooler__completed__#{pool['name']}", vm)
end
end
# MIGRATIONS
$redis.smembers("vmpooler__migrating__#{pool['name']}").each do |vm|
if inventory[vm]
begin
migrate_vm(vm, pool['name'], vsphere)
rescue => err
$logger.log('s', "[x] [#{pool['name']}] '#{vm}' failed to migrate: #{err}")
end
end end
end end
# REPOPULATE # REPOPULATE
ready = $redis.scard('vmpooler__ready__' + pool['name']) ready = $redis.scard("vmpooler__ready__#{pool['name']}")
total = $redis.scard('vmpooler__pending__' + pool['name']) + ready total = $redis.scard("vmpooler__pending__#{pool['name']}") + ready
$metrics.gauge('ready.' + pool['name'], $redis.scard('vmpooler__ready__' + pool['name'])) $metrics.gauge("ready.#{pool['name']}", $redis.scard("vmpooler__ready__#{pool['name']}"))
$metrics.gauge('running.' + pool['name'], $redis.scard('vmpooler__running__' + pool['name'])) $metrics.gauge("running.#{pool['name']}", $redis.scard("vmpooler__running__#{pool['name']}"))
if $redis.get('vmpooler__empty__' + pool['name']) if $redis.get("vmpooler__empty__#{pool['name']}")
unless ready == 0 unless ready == 0
$redis.del('vmpooler__empty__' + pool['name']) $redis.del("vmpooler__empty__#{pool['name']}")
end end
else else
if ready == 0 if ready == 0
$redis.set('vmpooler__empty__' + pool['name'], 'true') $redis.set("vmpooler__empty__#{pool['name']}", 'true')
$logger.log('s', "[!] [#{pool['name']}] is empty") $logger.log('s', "[!] [#{pool['name']}] is empty")
end end
end end
@ -583,10 +686,11 @@ module Vmpooler
pool['template'], pool['template'],
pool['folder'], pool['folder'],
pool['datastore'], pool['datastore'],
pool['clone_target'] pool['clone_target'],
vsphere
) )
rescue rescue => err
$logger.log('s', "[!] [#{pool['name']}] clone appears to have failed") $logger.log('s', "[!] [#{pool['name']}] clone failed during check_pool with an error: #{err}")
$redis.decr('vmpooler__tasks__clone') $redis.decr('vmpooler__tasks__clone')
end end
end end

View file

@ -6,22 +6,27 @@ module Vmpooler
DISK_TYPE = 'thin' DISK_TYPE = 'thin'
DISK_MODE = 'persistent' DISK_MODE = 'persistent'
def initialize(_vInfo = {}) def initialize(credentials)
config_file = File.expand_path('vmpooler.yaml') $credentials = credentials
vsphere = YAML.load_file(config_file)[:vsphere] end
@connection = RbVmomi::VIM.connect host: vsphere['server'], def ensure_connected(connection, credentials)
user: vsphere['username'], begin
password: vsphere['password'], connection.serviceInstance.CurrentTime
insecure: true rescue
connect_to_vsphere $credentials
end
end
def connect_to_vsphere(credentials)
@connection = RbVmomi::VIM.connect host: credentials['server'],
user: credentials['username'],
password: credentials['password'],
insecure: credentials['insecure'] || true
end end
def add_disk(vm, size, datastore) def add_disk(vm, size, datastore)
begin ensure_connected @connection, $credentials
@connection.serviceInstance.CurrentTime
rescue
initialize
end
return false unless size.to_i > 0 return false unless size.to_i > 0
@ -71,22 +76,14 @@ module Vmpooler
end end
def find_datastore(datastorename) def find_datastore(datastorename)
begin ensure_connected @connection, $credentials
@connection.serviceInstance.CurrentTime
rescue
initialize
end
datacenter = @connection.serviceInstance.find_datacenter datacenter = @connection.serviceInstance.find_datacenter
datacenter.find_datastore(datastorename) datacenter.find_datastore(datastorename)
end end
def find_device(vm, deviceName) def find_device(vm, deviceName)
begin ensure_connected @connection, $credentials
@connection.serviceInstance.CurrentTime
rescue
initialize
end
vm.config.hardware.device.each do |device| vm.config.hardware.device.each do |device|
return device if device.deviceInfo.label == deviceName return device if device.deviceInfo.label == deviceName
@ -96,11 +93,7 @@ module Vmpooler
end end
def find_disk_controller(vm) def find_disk_controller(vm)
begin ensure_connected @connection, $credentials
@connection.serviceInstance.CurrentTime
rescue
initialize
end
devices = find_disk_devices(vm) devices = find_disk_devices(vm)
@ -114,11 +107,7 @@ module Vmpooler
end end
def find_disk_devices(vm) def find_disk_devices(vm)
begin ensure_connected @connection, $credentials
@connection.serviceInstance.CurrentTime
rescue
initialize
end
devices = {} devices = {}
@ -146,11 +135,7 @@ module Vmpooler
end end
def find_disk_unit_number(vm, controller) def find_disk_unit_number(vm, controller)
begin ensure_connected @connection, $credentials
@connection.serviceInstance.CurrentTime
rescue
initialize
end
used_unit_numbers = [] used_unit_numbers = []
available_unit_numbers = [] available_unit_numbers = []
@ -175,11 +160,7 @@ module Vmpooler
end end
def find_folder(foldername) def find_folder(foldername)
begin ensure_connected @connection, $credentials
@connection.serviceInstance.CurrentTime
rescue
initialize
end
datacenter = @connection.serviceInstance.find_datacenter datacenter = @connection.serviceInstance.find_datacenter
base = datacenter.vmFolder base = datacenter.vmFolder
@ -196,39 +177,89 @@ module Vmpooler
base base
end end
# Returns an array containing cumulative CPU and memory utilization of a host, and its object reference
# Params:
# +model+:: CPU arch version to match on
# +limit+:: Hard limit for CPU or memory utilization beyond which a host is excluded for deployments
def get_host_utilization(host, model=nil, limit=90)
if model
return nil unless host_has_cpu_model? host, model
end
return nil if host.runtime.inMaintenanceMode
return nil unless host.overallStatus == 'green'
cpu_utilization = cpu_utilization_for host
memory_utilization = memory_utilization_for host
return nil if cpu_utilization > limit
return nil if memory_utilization > limit
[ cpu_utilization + memory_utilization, host ]
end
def host_has_cpu_model?(host, model)
get_host_cpu_arch_version(host) == model
end
def get_host_cpu_arch_version(host)
cpu_model = host.hardware.cpuPkg[0].description
cpu_model_parts = cpu_model.split()
arch_version = cpu_model_parts[4]
arch_version
end
def cpu_utilization_for(host)
cpu_usage = host.summary.quickStats.overallCpuUsage
cpu_size = host.summary.hardware.cpuMhz * host.summary.hardware.numCpuCores
(cpu_usage.to_f / cpu_size.to_f) * 100
end
def memory_utilization_for(host)
memory_usage = host.summary.quickStats.overallMemoryUsage
memory_size = host.summary.hardware.memorySize / 1024 / 1024
(memory_usage.to_f / memory_size.to_f) * 100
end
def find_least_used_host(cluster) def find_least_used_host(cluster)
begin ensure_connected @connection, $credentials
@connection.serviceInstance.CurrentTime
rescue cluster_object = find_cluster(cluster)
initialize target_hosts = get_cluster_host_utilization(cluster_object)
least_used_host = target_hosts.sort[0][1]
least_used_host
end end
hosts = {} def find_cluster(cluster)
hosts_sort = {}
datacenter = @connection.serviceInstance.find_datacenter datacenter = @connection.serviceInstance.find_datacenter
datacenter.hostFolder.children.each do |folder| datacenter.hostFolder.children.find { |cluster_object| cluster_object.name == cluster }
next unless folder.name == cluster
folder.host.each do |host|
if
(host.overallStatus == 'green') &&
(!host.runtime.inMaintenanceMode)
hosts[host.name] = host
hosts_sort[host.name] = host.vm.length
end
end
end end
hosts[hosts_sort.sort_by { |_k, v| v }[0][0]] def get_cluster_host_utilization(cluster)
cluster_hosts = []
cluster.host.each do |host|
host_usage = get_host_utilization(host)
cluster_hosts << host_usage if host_usage
end
cluster_hosts
end
def find_least_used_compatible_host(vm)
ensure_connected @connection, $credentials
source_host = vm.summary.runtime.host
model = get_host_cpu_arch_version(source_host)
cluster = source_host.parent
target_hosts = []
cluster.host.each do |host|
host_usage = get_host_utilization(host, model)
target_hosts << host_usage if host_usage
end
target_host = target_hosts.sort[0][1]
[target_host, target_host.name]
end end
def find_pool(poolname) def find_pool(poolname)
begin ensure_connected @connection, $credentials
@connection.serviceInstance.CurrentTime
rescue
initialize
end
datacenter = @connection.serviceInstance.find_datacenter datacenter = @connection.serviceInstance.find_datacenter
base = datacenter.hostFolder base = datacenter.hostFolder
@ -257,21 +288,18 @@ module Vmpooler
end end
def find_vm(vmname) def find_vm(vmname)
begin ensure_connected @connection, $credentials
@connection.serviceInstance.CurrentTime find_vm_light(vmname) || find_vm_heavy(vmname)[vmname]
rescue
initialize
end end
def find_vm_light(vmname)
ensure_connected @connection, $credentials
@connection.searchIndex.FindByDnsName(vmSearch: true, dnsName: vmname) @connection.searchIndex.FindByDnsName(vmSearch: true, dnsName: vmname)
end end
def find_vm_heavy(vmname) def find_vm_heavy(vmname)
begin ensure_connected @connection, $credentials
@connection.serviceInstance.CurrentTime
rescue
initialize
end
vmname = vmname.is_a?(Array) ? vmname : [vmname] vmname = vmname.is_a?(Array) ? vmname : [vmname]
containerView = get_base_vm_container_from @connection containerView = get_base_vm_container_from @connection
@ -321,11 +349,7 @@ module Vmpooler
end end
def find_vmdks(vmname, datastore) def find_vmdks(vmname, datastore)
begin ensure_connected @connection, $credentials
connection.serviceInstance.CurrentTime
rescue
initialize
end
disks = [] disks = []
@ -344,11 +368,7 @@ module Vmpooler
end end
def get_base_vm_container_from(connection) def get_base_vm_container_from(connection)
begin ensure_connected @connection, $credentials
connection.serviceInstance.CurrentTime
rescue
initialize
end
viewManager = connection.serviceContent.viewManager viewManager = connection.serviceContent.viewManager
viewManager.CreateContainerView( viewManager.CreateContainerView(
@ -372,6 +392,11 @@ module Vmpooler
snapshot snapshot
end end
def migrate_vm_host(vm, host)
relospec = RbVmomi::VIM.VirtualMachineRelocateSpec(host: host)
vm.RelocateVM_Task(spec: relospec).wait_for_completion
end
def close def close
@connection.close @connection.close
end end

View file

@ -50,11 +50,21 @@ def create_pending_vm(template, name, token = nil)
redis.hset("vmpooler__vm__#{name}", "template", template) redis.hset("vmpooler__vm__#{name}", "template", template)
end end
def create_vm(name, token = nil) def create_vm(name, token = nil, redis_handle = nil)
redis.hset("vmpooler__vm__#{name}", 'checkout', Time.now) redis_db = redis_handle ? redis_handle : redis
if token redis_db.hset("vmpooler__vm__#{name}", 'checkout', Time.now)
redis.hset("vmpooler__vm__#{name}", 'token:token', token) redis_db.hset("vmpooler__vm__#{name}", 'token:token', token) if token
end end
def create_migrating_vm(name, pool, redis_handle = nil)
redis_db = redis_handle ? redis_handle : redis
redis_db.hset("vmpooler__vm__#{name}", 'checkout', Time.now)
redis_db.sadd("vmpooler__migrating__#{pool}", name)
end
def add_vm_to_migration_set(name, redis_handle = nil)
redis_db = redis_handle ? redis_handle : redis
redis_db.sadd('vmpooler__migration', name)
end end
def fetch_vm(vm) def fetch_vm(vm)

View file

@ -0,0 +1,87 @@
require 'spec_helper'
require 'mock_redis'
require 'time'
describe 'Pool Manager' do
let(:logger) { double('logger') }
let(:redis) { MockRedis.new }
let(:metrics) { Vmpooler::DummyStatsd.new }
let(:config) {
{
config: {
'site_name' => 'test pooler',
'migration_limit' => 2,
vsphere: {
'server' => 'vsphere.puppet.com',
'username' => 'vmpooler@vsphere.local',
'password' => '',
'insecure' => true
},
pools: [ {'name' => 'pool1', 'size' => 5, 'folder' => 'pool1_folder'} ],
statsd: { 'prefix' => 'stats_prefix'},
pool_names: [ 'pool1' ]
}
}
}
let(:pool) { config[:config][:pools][0]['name'] }
let(:vm) {
{
'name' => 'vm1',
'host' => 'host1',
'template' => pool,
}
}
describe '#_migrate_vm' do
let(:vsphere) { double(pool) }
let(:pooler) { Vmpooler::PoolManager.new(config, logger, redis, metrics) }
context 'evaluates VM for migration and logs host' do
before do
create_migrating_vm vm['name'], pool, redis
allow(vsphere).to receive(:find_vm).and_return(vm)
allow(pooler).to receive(:get_vm_host_info).and_return([{'name' => 'host1'}, 'host1'])
end
it 'logs VM host when migration is disabled' do
config[:config]['migration_limit'] = nil
expect(redis.sismember("vmpooler__migrating__#{pool}", vm['name'])).to be true
expect(logger).to receive(:log).with('s', "[ ] [#{pool}] '#{vm['name']}' is running on #{vm['host']}")
pooler._migrate_vm(vm['name'], pool, vsphere)
expect(redis.sismember("vmpooler__migrating__#{pool}", vm['name'])).to be false
end
it 'verifies that migration_limit greater than or equal to migrations in progress and logs host' do
add_vm_to_migration_set vm['name'], redis
add_vm_to_migration_set 'vm2', redis
expect(logger).to receive(:log).with('s', "[ ] [#{pool}] '#{vm['name']}' is running on #{vm['host']}. No migration will be evaluated since the migration_limit has been reached")
pooler._migrate_vm(vm['name'], pool, vsphere)
end
it 'verifies that migration_limit is less than migrations in progress and logs old host, new host and migration time' do
allow(vsphere).to receive(:find_least_used_compatible_host).and_return([{'name' => 'host2'}, 'host2'])
allow(vsphere).to receive(:migrate_vm_host)
expect(redis.hget("vmpooler__vm__#{vm['name']}", 'migration_time'))
expect(redis.hget("vmpooler__vm__#{vm['name']}", 'checkout_to_migration'))
expect(logger).to receive(:log).with('s', "[>] [#{pool}] '#{vm['name']}' migrated from #{vm['host']} to host2 in 0.00 seconds")
pooler._migrate_vm(vm['name'], pool, vsphere)
end
it 'fails when no suitable host can be found' do
error = 'ArgumentError: No target host found'
allow(vsphere).to receive(:find_least_used_compatible_host)
allow(vsphere).to receive(:migrate_vm_host).and_raise(error)
expect(logger).to receive(:log).with('s', "[x] [#{pool}] '#{vm['name']}' migration failed with an error: #{error}")
pooler._migrate_vm(vm['name'], pool, vsphere)
end
end
end
end

View file

@ -24,10 +24,9 @@ describe 'Pool Manager' do
context 'host not in pool' do context 'host not in pool' do
it 'calls fail_pending_vm' do it 'calls fail_pending_vm' do
allow(pool_helper).to receive(:find_vm).and_return(nil) allow(vsphere).to receive(:find_vm).and_return(nil)
allow(redis).to receive(:hget) allow(redis).to receive(:hget)
expect(redis).to receive(:hget).with(String, 'clone').once subject._check_pending_vm(vm, pool, timeout, vsphere)
subject._check_pending_vm(vm, pool, timeout)
end end
end end
@ -36,16 +35,14 @@ describe 'Pool Manager' do
let(:tcpsocket) { double('TCPSocket') } let(:tcpsocket) { double('TCPSocket') }
it 'calls move_pending_vm_to_ready' do it 'calls move_pending_vm_to_ready' do
stub_const("TCPSocket", tcpsocket) allow(subject).to receive(:open_socket).and_return(true)
allow(vsphere).to receive(:find_vm).and_return(vm_finder)
allow(pool_helper).to receive(:find_vm).and_return(vm_finder)
allow(vm_finder).to receive(:summary).and_return(nil) allow(vm_finder).to receive(:summary).and_return(nil)
allow(tcpsocket).to receive(:new).and_return(true)
expect(vm_finder).to receive(:summary).once expect(vm_finder).to receive(:summary).once
expect(redis).not_to receive(:hget).with(String, 'clone') expect(redis).not_to receive(:hget).with(String, 'clone')
subject._check_pending_vm(vm, pool, timeout) subject._check_pending_vm(vm, pool, timeout, vsphere)
end end
end end
end end
@ -156,16 +153,16 @@ describe 'Pool Manager' do
end end
it 'does nothing with nil host' do it 'does nothing with nil host' do
allow(pool_helper).to receive(:find_vm).and_return(nil) allow(vsphere).to receive(:find_vm).and_return(nil)
expect(redis).not_to receive(:smove) expect(redis).not_to receive(:smove)
subject._check_running_vm(vm, pool, timeout) subject._check_running_vm(vm, pool, timeout, vsphere)
end end
context 'valid host' do context 'valid host' do
let(:vm_host) { double('vmhost') } let(:vm_host) { double('vmhost') }
it 'does not move vm when not poweredOn' do it 'does not move vm when not poweredOn' do
allow(pool_helper).to receive(:find_vm).and_return vm_host allow(vsphere).to receive(:find_vm).and_return vm_host
allow(vm_host).to receive(:runtime).and_return true allow(vm_host).to receive(:runtime).and_return true
allow(vm_host).to receive_message_chain(:runtime, :powerState).and_return 'poweredOff' allow(vm_host).to receive_message_chain(:runtime, :powerState).and_return 'poweredOff'
@ -173,11 +170,11 @@ describe 'Pool Manager' do
expect(redis).not_to receive(:smove) expect(redis).not_to receive(:smove)
expect(logger).not_to receive(:log).with('d', "[!] [#{pool}] '#{vm}' appears to be powered off or dead") expect(logger).not_to receive(:log).with('d', "[!] [#{pool}] '#{vm}' appears to be powered off or dead")
subject._check_running_vm(vm, pool, timeout) subject._check_running_vm(vm, pool, timeout, vsphere)
end end
it 'moves vm when poweredOn, but past TTL' do it 'moves vm when poweredOn, but past TTL' do
allow(pool_helper).to receive(:find_vm).and_return vm_host allow(vsphere).to receive(:find_vm).and_return vm_host
allow(vm_host).to receive(:runtime).and_return true allow(vm_host).to receive(:runtime).and_return true
allow(vm_host).to receive_message_chain(:runtime, :powerState).and_return 'poweredOn' allow(vm_host).to receive_message_chain(:runtime, :powerState).and_return 'poweredOn'
@ -185,7 +182,7 @@ describe 'Pool Manager' do
expect(redis).to receive(:smove) expect(redis).to receive(:smove)
expect(logger).to receive(:log).with('d', "[!] [#{pool}] '#{vm}' reached end of TTL after #{timeout} hours") expect(logger).to receive(:log).with('d', "[!] [#{pool}] '#{vm}' reached end of TTL after #{timeout} hours")
subject._check_running_vm(vm, pool, timeout) subject._check_running_vm(vm, pool, timeout, vsphere)
end end
end end
end end
@ -228,6 +225,7 @@ describe 'Pool Manager' do
allow(redis).to receive(:smembers).with('vmpooler__running__pool1').and_return([]) allow(redis).to receive(:smembers).with('vmpooler__running__pool1').and_return([])
allow(redis).to receive(:smembers).with('vmpooler__completed__pool1').and_return([]) allow(redis).to receive(:smembers).with('vmpooler__completed__pool1').and_return([])
allow(redis).to receive(:smembers).with('vmpooler__discovered__pool1').and_return([]) allow(redis).to receive(:smembers).with('vmpooler__discovered__pool1').and_return([])
allow(redis).to receive(:smembers).with('vmpooler__migrating__pool1').and_return([])
allow(redis).to receive(:set) allow(redis).to receive(:set)
allow(redis).to receive(:get).with('vmpooler__tasks__clone').and_return(0) allow(redis).to receive(:get).with('vmpooler__tasks__clone').and_return(0)
allow(redis).to receive(:get).with('vmpooler__empty__pool1').and_return(nil) allow(redis).to receive(:get).with('vmpooler__empty__pool1').and_return(nil)
@ -240,7 +238,7 @@ describe 'Pool Manager' do
allow(redis).to receive(:scard).with('vmpooler__running__pool1').and_return(0) allow(redis).to receive(:scard).with('vmpooler__running__pool1').and_return(0)
expect(logger).to receive(:log).with('s', "[!] [pool1] is empty") expect(logger).to receive(:log).with('s', "[!] [pool1] is empty")
subject._check_pool(config[:pools][0]) subject._check_pool(config[:pools][0], vsphere)
end end
end end
end end
@ -277,7 +275,7 @@ describe 'Pool Manager' do
expect(metrics).to receive(:gauge).with('ready.pool1', 1) expect(metrics).to receive(:gauge).with('ready.pool1', 1)
expect(metrics).to receive(:gauge).with('running.pool1', 5) expect(metrics).to receive(:gauge).with('running.pool1', 5)
subject._check_pool(config[:pools][0]) subject._check_pool(config[:pools][0], vsphere)
end end
it 'increments metrics when ready with 0 when pool empty' do it 'increments metrics when ready with 0 when pool empty' do
@ -288,7 +286,7 @@ describe 'Pool Manager' do
expect(metrics).to receive(:gauge).with('ready.pool1', 0) expect(metrics).to receive(:gauge).with('ready.pool1', 0)
expect(metrics).to receive(:gauge).with('running.pool1', 5) expect(metrics).to receive(:gauge).with('running.pool1', 5)
subject._check_pool(config[:pools][0]) subject._check_pool(config[:pools][0], vsphere)
end end
end end
end end
@ -307,13 +305,13 @@ describe 'Pool Manager' do
let(:vm_host) { double('vmhost') } let(:vm_host) { double('vmhost') }
it 'creates a snapshot' do it 'creates a snapshot' do
expect(pool_helper).to receive(:find_vm).and_return vm_host expect(vsphere).to receive(:find_vm).and_return vm_host
expect(logger).to receive(:log) expect(logger).to receive(:log)
expect(vm_host).to receive_message_chain(:CreateSnapshot_Task, :wait_for_completion) expect(vm_host).to receive_message_chain(:CreateSnapshot_Task, :wait_for_completion)
expect(redis).to receive(:hset).with('vmpooler__vm__testvm', 'snapshot:testsnapshot', Time.now.to_s) expect(redis).to receive(:hset).with('vmpooler__vm__testvm', 'snapshot:testsnapshot', Time.now.to_s)
expect(logger).to receive(:log) expect(logger).to receive(:log)
subject._create_vm_snapshot('testvm', 'testsnapshot') subject._create_vm_snapshot('testvm', 'testsnapshot', vsphere)
end end
end end
end end
@ -333,13 +331,13 @@ describe 'Pool Manager' do
let(:vm_snapshot) { double('vmsnapshot') } let(:vm_snapshot) { double('vmsnapshot') }
it 'reverts a snapshot' do it 'reverts a snapshot' do
expect(pool_helper).to receive(:find_vm).and_return vm_host expect(vsphere).to receive(:find_vm).and_return vm_host
expect(pool_helper).to receive(:find_snapshot).and_return vm_snapshot expect(vsphere).to receive(:find_snapshot).and_return vm_snapshot
expect(logger).to receive(:log) expect(logger).to receive(:log)
expect(vm_snapshot).to receive_message_chain(:RevertToSnapshot_Task, :wait_for_completion) expect(vm_snapshot).to receive_message_chain(:RevertToSnapshot_Task, :wait_for_completion)
expect(logger).to receive(:log) expect(logger).to receive(:log)
subject._revert_vm_snapshot('testvm', 'testsnapshot') subject._revert_vm_snapshot('testvm', 'testsnapshot', vsphere)
end end
end end
end end
@ -357,7 +355,7 @@ describe 'Pool Manager' do
expect(redis).to receive(:spop).with('vmpooler__tasks__snapshot') expect(redis).to receive(:spop).with('vmpooler__tasks__snapshot')
expect(redis).to receive(:spop).with('vmpooler__tasks__snapshot-revert') expect(redis).to receive(:spop).with('vmpooler__tasks__snapshot-revert')
subject._check_snapshot_queue subject._check_snapshot_queue(vsphere)
end end
end end
end end

View file

@ -225,6 +225,13 @@
# If set, prefixes all created VMs with this string. This should include # If set, prefixes all created VMs with this string. This should include
# a separator. # a separator.
# (optional; default: '') # (optional; default: '')
#
# - migration_limit
# When set to any value greater than 0 enable VM migration at checkout.
# When enabled this capability will evaluate a VM for migration when it is requested
# in an effort to maintain a more even distribution of load across compute resources.
# The migration_limit ensures that no more than n migrations will be evaluated at any one time
# and greatly reduces the possibilty of VMs ending up bunched together on a particular host.
# Example: # Example: