Parent

Methods

OpenShift::MCollectiveApplicationContainerProxy

Implements the broker-node communications This class the state of a node and a set of RPC functions to the node. It also has a set of just plain functions which live here because they relate to broker/node communications.

Attributes

district[RW]

A District ID string

id[RW]

A Node ID string

Public Class Methods

blacklisted_in_impl?(name) click to toggle source

<<orphan>>

<<class method>>

INPUTS:

  • name: String. A name to be checked against the blacklist

RETURNS:

  • Boolean. True if the name is in the blacklist

NOTES:

  • This is really a function of the broker

# File lib/openshift/mcollective_application_container_proxy.rb, line 137
def self.blacklisted_in_impl?(name)
  false
end
find_available_impl(node_profile=nil, district_uuid=nil, least_preferred_server_identities=nil, gear_exists_in_district=false, required_uid=nil) click to toggle source

<<factory method>>

Find a node which fulfills app requirements. Implements the superclass find_available() method

INPUTS:

  • node_profile: string identifier for a set of node characteristics

  • district_uuid: identifier for the district

  • least_preferred_server_identities: list of server identities that are least preferred. These could be the ones that won't allow the gear group to be highly available

  • gear_exists_in_district: true if the gear belongs to a node in the same district

  • required_uid: the uid that is required to be available in the destination district

RETURNS:

RAISES:

  • OpenShift::NodeUnavailableException

NOTES:

  • a class method on Node?

  • Uses Rails.configuration.msg_broker

  • Uses District

  • Calls rpc_find_available

VALIDATIONS:

  • If gear_exists_in_district is true, then required_uid cannot be set and has to be nil

  • If gear_exists_in_district is true, then district_uuid must be passed and cannot be nil

# File lib/openshift/mcollective_application_container_proxy.rb, line 70
def self.find_available_impl(node_profile=nil, district_uuid=nil, least_preferred_server_identities=nil, gear_exists_in_district=false, required_uid=nil)
  district = nil
  current_server, current_capacity, preferred_district = rpc_find_available(node_profile, district_uuid, least_preferred_server_identities, false, gear_exists_in_district, required_uid)
  if !current_server
    current_server, current_capacity, preferred_district = rpc_find_available(node_profile, district_uuid, least_preferred_server_identities, true, gear_exists_in_district, required_uid)
  end
  district = preferred_district if preferred_district
  raise OpenShift::NodeUnavailableException.new("No nodes available", 140) unless current_server
  Rails.logger.debug "DEBUG: find_available_impl: current_server: #{current_server}: #{current_capacity}"

  MCollectiveApplicationContainerProxy.new(current_server, district)
end
find_one_impl(node_profile=nil) click to toggle source

<<factory method>>

Find a single node. Implements superclass find_one() method.

INPUTS:

  • node_profile: characteristics for node filtering

RETURNS:

NOTES:

# File lib/openshift/mcollective_application_container_proxy.rb, line 95
def self.find_one_impl(node_profile=nil)
  current_server = rpc_find_one(node_profile)
  current_server, capacity, district = rpc_find_available(node_profile) unless current_server

  raise OpenShift::NodeUnavailableException.new("No nodes available", 140) unless current_server
  Rails.logger.debug "DEBUG: find_one_impl: current_server: #{current_server}"

  MCollectiveApplicationContainerProxy.new(current_server)
end
get_blacklisted_in_impl() click to toggle source

<<orphan>> <<class method>>

Return a list of blacklisted namespaces and app names. Implements superclass get_blacklisted() method.

INPUTS:

  • none

RETURNS:

  • empty list

NOTES:

  • Is this really a function of the broker

# File lib/openshift/mcollective_application_container_proxy.rb, line 120
def self.get_blacklisted_in_impl
  []
end
get_details_for_all_impl(name_list) click to toggle source

get details (in MCollective, facts) about all nodes that respond

RETURN:

  • Hash of hashes: node identity => fact map for each node

NOTES:

# File lib/openshift/mcollective_application_container_proxy.rb, line 2279
def self.get_details_for_all_impl(name_list)
  rpc_get_facts_for_all_nodes(name_list)
end
new(id, district=nil) click to toggle source

<<constructor>>

Create an app descriptor/handle for remote controls

INPUTS:

  • id: string - a unique app identifier

  • district: <type> - a classifier for app placement

# File lib/openshift/mcollective_application_container_proxy.rb, line 37
def initialize(id, district=nil)
  @id = id
  @district = district
end
rpc_exec(agent, servers=nil, force_rediscovery=false, options=rpc_options) click to toggle source

Execute an RPC call for the specified agent. If a server is supplied, only execute for that server.

INPUTS:

  • agent: ??

  • servers: String|Array

  • force_rediscovery: Boolean

  • options: Hash

RETURNS:

  • ResultIO

RAISES:

  • OpenShift::NodeException

NOTES:

  • connects, makes a request, closes connection.

  • uses MCollective::RPC::Client

  • THIS IS THE MEAT!

# File lib/openshift/mcollective_application_container_proxy.rb, line 2304
def self.rpc_exec(agent, servers=nil, force_rediscovery=false, options=rpc_options)

  if servers
    servers = Array(servers)
  else
    servers = []
  end

  # Setup the rpc client
  rpc_client = MCollectiveApplicationContainerProxy.get_rpc_client(agent, options)

  if !servers.empty?
    Rails.logger.debug("DEBUG: rpc_exec: Filtering rpc_exec to servers #{servers.pretty_inspect}")
    rpc_client.discover :nodes => servers
  end

  # Filter to the specified server
  #if server
  #  Rails.logger.debug("DEBUG: rpc_exec: Filtering rpc_exec to server #{server}")
  #  rpc_client.identity_filter(server)
  #end

  if force_rediscovery
    rpc_client.reset
  end
  Rails.logger.debug("DEBUG: rpc_exec: rpc_client=#{rpc_client}")

  # Execute a block and make sure we disconnect the client
  begin
    result = yield rpc_client
  ensure
    rpc_client.disconnect
  end

  raise OpenShift::NodeException.new("Node execution failure (error getting result from node).", 143) unless result

  result
end
set_district_uid_limits_impl(uuid, first_uid, max_uid) click to toggle source

Set the district uid limits for all district nodes

INPUTS:

  • uuid: String

  • first_uid: Integer

  • max_uid: Integer

RETURNS:

  • ResultIO?

NOTES:

  • uses MCollective::RPC::Client

# File lib/openshift/mcollective_application_container_proxy.rb, line 2392
def self.set_district_uid_limits_impl(uuid, first_uid, max_uid)
  mc_args = { :first_uid => first_uid,
              :max_uid => max_uid}
  options = MCollectiveApplicationContainerProxy.rpc_options
  rpc_client = MCollectiveApplicationContainerProxy.get_rpc_client('openshift', options)
  rpc_client.fact_filter "district_uuid", uuid
  result = nil
  begin
    Rails.logger.debug "DEBUG: rpc_client.custom_request('set_district_uid_limits', #{mc_args.inspect})"
    result = rpc_client.set_district_uid_limits(mc_args)
    Rails.logger.debug "DEBUG: #{result.inspect}"
  ensure
    rpc_client.disconnect
  end
  Rails.logger.debug result.inspect
  result
end

Protected Class Methods

execute_parallel_jobs_impl(handle) click to toggle source

<<implementation>> <<class method>>

Execute a set of operations on a node in parallel

INPUTS:

  • handle: Hash ???

RETURNS:

  • ???

NOTES:

# File lib/openshift/mcollective_application_container_proxy.rb, line 3494
def self.execute_parallel_jobs_impl(handle)
  if handle.present?
    start_time = Time.new
    begin
      mc_args = handle.clone
      options = MCollectiveApplicationContainerProxy.rpc_options.merge(handle.delete(:args) || {})
      rpc_client = MCollectiveApplicationContainerProxy.get_rpc_client('openshift', options)
      identities = handle.keys
      rpc_client.custom_request('execute_parallel', mc_args, identities, {'identity' => identities}).each { |mcoll_reply|
        if mcoll_reply.results[:statuscode] == 0
          output = mcoll_reply.results[:data][:output]
          exitcode = mcoll_reply.results[:data][:exitcode]
          sender = mcoll_reply.results[:sender]
          Rails.logger.debug("DEBUG: Output of parallel execute: #{output}, exitcode: #{exitcode}, from: #{sender}  (Request ID: #{Thread.current[:user_action_log_uuid]})")

          handle[sender] = output if exitcode == 0
        end
      }
    ensure
      rpc_client.disconnect
    end
    Rails.logger.debug "DEBUG: MCollective Response Time (execute_parallel): #{((Time.new - start_time)*1000).round}ms  (Request ID: #{Thread.current[:user_action_log_uuid]})"
  end
end
find_gear(gear_uuid, servers = nil) click to toggle source

Returns the server identity of the specified gear

INPUTS:

  • gear_uuid: String

RETURNS:

  • server identity (string)

NOTES:

# File lib/openshift/mcollective_application_container_proxy.rb, line 2748
def self.find_gear(gear_uuid, servers = nil)
  server_identity = nil
  rpc_exec('openshift', servers) do |client|
    client.has_gear(:uuid => gear_uuid) do |response|
      output = response[:body][:data][:output]
      if output == true
        server_identity = response[:senderid]
      end
    end
  end
  return server_identity
end
get_all_active_gears_impl() click to toggle source

Retrieve all active gears (implementation)

INPUTS:

  • none

RETURNS:

NOTES:

  • should be class method on Node? or Broker?

  • uses MCollective::RPC::Client rpc_exec

  • uses MCollective::RPC::Client.missing_method ??

# File lib/openshift/mcollective_application_container_proxy.rb, line 3435
def self.get_all_active_gears_impl
  active_gears_map = {}
  rpc_exec('openshift', nil, true) do |client|
    client.get_all_active_gears() do |response|
      if response[:body][:statuscode] == 0
        active_gears = response[:body][:data][:output]
        sender = response[:senderid]
        active_gears_map[sender] = active_gears
      end
    end
  end
  active_gears_map
end
get_all_gears_endpoints_impl(opts) click to toggle source
# File lib/openshift/mcollective_application_container_proxy.rb, line 3407
def self.get_all_gears_endpoints_impl(opts)
  gear_map = {}
  rpc_exec('openshift', nil, true) do |client|
    client.get_all_gears_endpoints(opts) do |response|
      if response[:body][:statuscode] == 0
        sub_gear_map = response[:body][:data][:output]
        # sender = response[:senderid]
        gear_map.merge!(sub_gear_map)
      end
    end
  end
  return gear_map
end
get_all_gears_impl(opts) click to toggle source

Retrieve all gear IDs from all nodes (implementation)

INPUTS:

  • none

RETURNS:

NOTES:

  • Should be class method on Node? (All nodes?)

  • Why doesn't this just override a method from the superclass?

  • uses rpc_exec

# File lib/openshift/mcollective_application_container_proxy.rb, line 3388
def self.get_all_gears_impl(opts)
  gear_map = {}
  sender_map = {}
  rpc_exec('openshift', nil, true) do |client|
    client.get_all_gears(opts) do |response|
      if response[:body][:statuscode] == 0
        sub_gear_map = response[:body][:data][:output]
        sender = response[:senderid]
        sub_gear_map.each { |k,v|
          gear_map[k] = [sender,Integer(v)]
          sender_map[sender] = {} if not sender_map.has_key? sender
          sender_map[sender][Integer(v)] = k
        }
      end
    end
  end
  return [gear_map, sender_map]
end
get_all_gears_sshkeys_impl() click to toggle source

Retrieve all ssh keys for all gears from all nodes (implementation)

INPUTS:

  • none

RETURNS:

NOTES:

  • Should be class method on Node? (All nodes?)

  • Why doesn't this just override a method from the superclass?

  • uses rpc_exec

# File lib/openshift/mcollective_application_container_proxy.rb, line 3463
def self.get_all_gears_sshkeys_impl
  gear_sshkey_map = {}
  sender_list = []
  rpc_exec('openshift', nil, true) do |client|
    client.get_all_gears_sshkeys do |response|
      if response[:body][:statuscode] == 0
        gear_sshkey_map.merge! response[:body][:data][:output]
        sender_list.push response[:senderid]
      end
    end
  end
  return [gear_sshkey_map, sender_list]
end
get_rpc_client(agent, options) click to toggle source

Get mcollective rpc client. For errors, convert generic exception to NodeException.

INPUTS:

  • agent: String (??)

  • options: Hash

RETURNS:

  • MCollective::RPC::Client

RAISES:

  • OpenShift::NodeException

NOTES

  • Uses MCollective::RPC::Client

# File lib/openshift/mcollective_application_container_proxy.rb, line 3362
def self.get_rpc_client(agent, options)
  flags = { :options => options, :exit_on_failure => false }
  begin
    rpc_client = rpcclient(agent, flags)
  rescue Exception => e
    Rails.logger.error "Exception raised by rpcclient:#{e.message}"
    Rails.logger.error (e.backtrace)
    raise OpenShift::NodeException.new(e)
  end
  return rpc_client
end
known_server_identities(force_rediscovery=false, rpc_opts=nil) click to toggle source

Execute a cartridge hook command in a gear

INPUTS:

  • force_rediscovery: Boolean

  • rpc_opts: Hash

RETURNS:

  • Array of known server identities

RAISES:

  • Exception

# File lib/openshift/mcollective_application_container_proxy.rb, line 2930
def self.known_server_identities(force_rediscovery=false, rpc_opts=nil)
  server_identities = nil
  server_identities = Rails.cache.read('known_server_identities') unless force_rediscovery
  unless server_identities
    server_identities = []
    rpc_get_fact('active_capacity', nil, true, force_rediscovery, nil, rpc_opts) do |server, capacity|
      #Rails.logger.debug "Next server: #{server} active capacity: #{capacity}"
      server_identities << server
    end
    Rails.cache.write('known_server_identities', server_identities, {:expires_in => 1.hour}) unless server_identities.empty?
  end
  server_identities
end
rpc_find_available(node_profile=nil, district_uuid=nil, least_preferred_server_identities=nil, force_rediscovery=false, gear_exists_in_district=false, required_uid=nil) click to toggle source

???

INPUTS:

  • node_profile: String identifier for a set of node characteristics

  • district_uuid: String identifier for the district

  • least_preferred_server_identities: list of server identities that are least preferred. These could be the ones that won't allow the gear group to be highly available

  • force_rediscovery: Boolean

  • gear_exists_in_district: Boolean - true if the gear belongs to a node in the same district

  • required_uid: String - the uid that is required to be available in the destination district

RETURNS:

  • Array: [server, capacity, district]

NOTES:

  • are the return values String?

VALIDATIONS:

  • If gear_exists_in_district is true, then required_uid cannot be set and has to be nil

  • If gear_exists_in_district is true, then district_uuid must be passed and cannot be nil

# File lib/openshift/mcollective_application_container_proxy.rb, line 2965
def self.rpc_find_available(node_profile=nil, district_uuid=nil, least_preferred_server_identities=nil, force_rediscovery=false, gear_exists_in_district=false, required_uid=nil)

  district_uuid = nil if district_uuid == 'NONE'

  # validate to ensure incompatible parameters are not passed
  if gear_exists_in_district
    if required_uid or district_uuid.nil?
      raise OpenShift::UserException.new("Incompatible parameters being passed for finding available node within the same district", 1)
    end
  end

  require_specific_district = !district_uuid.nil?
  require_district = require_specific_district
  prefer_district = require_specific_district
  unless require_specific_district
    if Rails.configuration.msg_broker[:districts][:enabled] && (!district_uuid || district_uuid == 'NONE')
      prefer_district = true
      if Rails.configuration.msg_broker[:districts][:require_for_app_create]
        require_district = true
      end
    end
  end
  require_district = true if required_uid
  current_server, current_capacity = nil, nil
  server_infos = []

  # First find the most available nodes and match
  # to their districts.  Take out the almost full nodes if possible and return one of
  # the nodes within a district with a lot of space.
  additional_filters = [{:fact => "active_capacity",
                         :value => '100',
                         :operator => "<"}]

  if require_specific_district || require_district
    additional_filters.push({:fact => "district_active",
                             :value => true.to_s,
                             :operator => "=="})
  end

  if require_specific_district
    additional_filters.push({:fact => "district_uuid",
                             :value => district_uuid,
                             :operator => "=="})
  elsif require_district
    additional_filters.push({:fact => "district_uuid",
                             :value => "NONE",
                             :operator => "!="})
  elsif !prefer_district
    additional_filters.push({:fact => "district_uuid",
                             :value => "NONE",
                             :operator => "=="})
  end

  if node_profile && Rails.configuration.msg_broker[:node_profile_enabled]
    additional_filters.push({:fact => "node_profile",
                             :value => node_profile,
                             :operator => "=="})
  end

  # Get the districts
  districts = prefer_district ? District.find_all((require_district && node_profile ) ? node_profile : nil, required_uid) : []

  # Get the active % on the nodes
  rpc_opts = nil
  rpc_get_fact('active_capacity', nil, force_rediscovery, additional_filters, rpc_opts) do |server, capacity|
    found_district = false
    districts.each do |district|
      if district.server_identities_hash.has_key?(server)
        next if required_uid and !district.available_uids.include?(required_uid)
        if (gear_exists_in_district || district.available_capacity > 0) && district.server_identities_hash[server]["active"]
          server_infos << [server, capacity.to_f, district]
        end
        found_district = true
        break
      end
    end
    if !found_district && !require_district # Districts aren't required in this case
      server_infos << [server, capacity.to_f]
    end
  end
  if require_district && server_infos.empty?
    raise OpenShift::NodeUnavailableException.new("No district nodes available", 140)
  end
  unless server_infos.empty?
    # Remove the least preferred servers from the list, ensuring there is at least one server remaining
    server_infos.delete_if { |server_info| server_infos.length > 1 && least_preferred_server_identities.include?(server_info[0]) } if least_preferred_server_identities

    # Remove any non districted nodes if you prefer districts
    if prefer_district && !require_district && !districts.empty?
      has_districted_node = false
      server_infos.each do |server_info|
        if server_info[2]
          has_districted_node = true
          break
        end
      end
      server_infos.delete_if { |server_info| !server_info[2] } if has_districted_node
    end

    # Sort by node available capacity and take the best half
    server_infos = server_infos.sort_by { |server_info| server_info[1] }
    server_infos = server_infos.first([4, (server_infos.length / 2).to_i].max) # consider the top half and no less than min(4, the actual number of available)

    # Sort by district available capacity and take the best half
    server_infos = server_infos.sort_by { |server_info| (server_info[2] && server_info[2].available_capacity) ? server_info[2].available_capacity : 1 }
    server_infos = server_infos.last([4, (server_infos.length / 2).to_i].max) # consider the top half and no less than min(4, the actual number of available)
  end

  current_district = nil
  unless server_infos.empty?
    # Randomly pick one of the best options
    server_info = server_infos[rand(server_infos.length)]
    current_server = server_info[0]
    current_capacity = server_info[1]
    current_district = server_info[2]
    Rails.logger.debug "Current server: #{current_server} active capacity: #{current_capacity}"
  end

  return current_server, current_capacity, current_district
end
rpc_find_one(node_profile=nil) click to toggle source

Return a single node matching a given profile

INPUTS:

RETURNS:

  • String: server name?

RAISES:

  • OpenShift::NodeUnavailableException

NOTES:

  • Query facters from every node and filter on server side

  • uses MCollective::RPC::Client

# File lib/openshift/mcollective_application_container_proxy.rb, line 3102
def self.rpc_find_one(node_profile=nil)
  current_server = nil
  additional_filters = []

  if Rails.configuration.msg_broker[:node_profile_enabled]
    if node_profile
      additional_filters.push({:fact => "node_profile",
                               :value => node_profile,
                               :operator => "=="})
    end
  end

  options = MCollectiveApplicationContainerProxy.rpc_options
  options[:filter]['fact'] = options[:filter]['fact'] + additional_filters
  options[:mcollective_limit_targets] = "1"

  rpc_client = MCollectiveApplicationContainerProxy.get_rpc_client('rpcutil', options)
  begin
    rpc_client.get_fact(:fact => 'public_hostname') do |response|
      raise OpenShift::NodeUnavailableExceptionn.new("No nodes available", 140) unless Integer(response[:body][:statuscode]) == 0
      current_server = response[:senderid]
    end
  ensure
    rpc_client.disconnect
  end
  return current_server
end
rpc_get_fact(fact, servers=nil, force_rediscovery=false, additional_filters=nil, custom_rpc_opts=nil) click to toggle source

Returns the fact value from the specified server. Yields to the supplied block if there is a non-nil value for the fact.

INPUTS:

  • fact: String - a fact name

  • servers: String|Array - a node name or an array of node names

  • force_rediscovery: Boolean

  • additional_filters: ?

  • custom_rpmc_opts: Hash?

RETURNS:

  • String?

NOTES:

# File lib/openshift/mcollective_application_container_proxy.rb, line 3207
def self.rpc_get_fact(fact, servers=nil, force_rediscovery=false, additional_filters=nil, custom_rpc_opts=nil)
  result = nil
  options = custom_rpc_opts ? custom_rpc_opts : MCollectiveApplicationContainerProxy.rpc_options
  options[:filter]['fact'] = options[:filter]['fact'] + additional_filters if additional_filters
  options[:timeout] = Rails.configuration.msg_broker[:fact_timeout]

  Rails.logger.debug("DEBUG: rpc_get_fact: fact=#{fact}")
  rpc_exec('rpcutil', servers, force_rediscovery, options) do |client|
    begin
      # in some cases, the get_fact mcollective call gets stuck and never returns
      # to handle these situations, we are using the ruby Timeout moddule as a safety net
      # and setting the timer duration as a reasonable multiple of the mcollective timeout value
      failsafe_timeout = Rails.configuration.msg_broker[:fact_timeout] * 3
      Timeout::timeout(failsafe_timeout) do
        client.get_fact(:fact => fact) do |response|
          next unless Integer(response[:body][:statuscode]) == 0

          # Yield the sender and the value to the block
          result = rvalue(response)
          yield response[:senderid], result if result
        end
      end
    rescue Timeout::Error
      raise OpenShift::NodeUnavailableException.new("Timed out while trying to fetch information from the nodes")
    end
  end

  result
end
rpc_get_facts_for_all_nodes(fact_list) click to toggle source

Given a list of facts, get those facts for all nodes that respond.

INPUTS:

  • facts: Enumerable of Strings (fact names)

RETURNS:

  • Map of Fact to String

RAISES:

  • OpenShift::NodeException

NOTES

# File lib/openshift/mcollective_application_container_proxy.rb, line 3327
def self.rpc_get_facts_for_all_nodes(fact_list)
  options = MCollectiveApplicationContainerProxy.rpc_options
  options[:timeout] = Rails.configuration.msg_broker[:fact_timeout]

  node_fact_map = {}
  rpc_exec('openshift', nil, true, options) do |client|
    client.get_facts(:facts => fact_list) do |response|
      if response[:body][:statuscode] == 0
        fact_map = response[:body][:data][:output]
        sender = response[:senderid]
        fact_map[:id] = sender
        node_fact_map[sender] = fact_map
      end
    end
  end
  return node_fact_map
end
rpc_options() click to toggle source

Make a deep copy of the RPC options hash

INPUTS:

RETURNS:

NOTES:

  • Simple copy by Marshall load/dump

# File lib/openshift/mcollective_application_container_proxy.rb, line 3141
def self.rpc_options
  # Make a deep copy of the default options
  Marshal::load(Marshal::dump(Rails.configuration.msg_broker[:rpc_options]))
end
rvalue(response) click to toggle source

Return the value of the MCollective response for both a single result and a multiple result structure

INPUTS:

  • response: an MCollective::Response object

RETURNS:

  • String: value string

NOTES:

  • returns value from body or data

# File lib/openshift/mcollective_application_container_proxy.rb, line 3160
def self.rvalue(response)
  result = nil

  if response[:body]
    result = response[:body][:data][:value]
  elsif response[:data]
    result = response[:data][:value]
  end

  result
end

Public Instance Methods

activate(gear, deployment_id) click to toggle source

Activate a deployment for a gear

INPUTS:

  • gear: a Gear object

  • deployment_id: a deployment id

RETURNS ResultIO: the result of running post-configure on the cartridge

# File lib/openshift/mcollective_application_container_proxy.rb, line 886
def activate(gear, deployment_id)
  result_io = ResultIO.new

  args = build_base_gear_args(gear)

  args['--with-deployment-id'] = deployment_id

  result_io = run_cartridge_command(@@C_CONTROLLER, gear, "activate", args)
  return result_io
end
add_alias(gear, server_alias) click to toggle source

Add an application alias to a gear

INPUTS:

  • gear: a Gear object

  • server_alias: String - a new FQDN for the gear

RETURNS:

  • String: stdout from a command

NOTES:

  • calls execute_direct

  • executes the 'add-alias' action on the node

  • method on Gear?

# File lib/openshift/mcollective_application_container_proxy.rb, line 1286
def add_alias(gear, server_alias)
  args = build_base_gear_args(gear)
  args['--with-alias-name']=server_alias
  result = execute_direct(@@C_CONTROLLER, 'add-alias', args)
  parse_result(result)
end
add_broker_auth_key(gear, iv, token) click to toggle source

Add a broker auth key. The broker auth key allows an application to request scaling and other actions from the broker.

INPUTS:

  • gear: a Gear object

  • iv: String - SSL initialization vector

  • token: String - a broker auth key

RETURNS:

  • mcollective parsed result string (stdout)

NOTES:

# File lib/openshift/mcollective_application_container_proxy.rb, line 561
def add_broker_auth_key(gear, iv, token)
  args = build_base_gear_args(gear)
  args['--with-iv'] = iv
  args['--with-token'] = token
  result = execute_direct(@@C_CONTROLLER, 'broker-auth-key-add', args)
  parse_result(result, gear)
end
add_component(gear, component, template_git_url=nil) click to toggle source

Add a component to an existing gear on the node

INPUTS:

  • gear: a Gear object

  • cart: string representing cartridge name

RETURNS:

  • Array [ResultIO, String]

RAISES:

  • Exception

CATCHES:

  • Exception

NOTES:

# File lib/openshift/mcollective_application_container_proxy.rb, line 789
def add_component(gear, component, template_git_url=nil)
  result_io = ResultIO.new
  cart_data = nil

  args = build_base_gear_args(gear)
  args = build_base_component_args(component, args)

  app = gear.app
  downloaded_cart =  app.downloaded_cart_map.values.find { |c| c["versioned_name"]==component.cartridge_name}
  if downloaded_cart
    args['--with-cartridge-manifest'] = downloaded_cart["original_manifest"]
    args['--with-software-version'] = downloaded_cart["version"]
  end

  if template_git_url.present?
    args['--with-template-git-url'] = template_git_url
  end

  result_io = run_cartridge_command(component.cartridge_name, gear, "configure", args)

  component_details = result_io.appInfoIO.string.empty? ? '' : result_io.appInfoIO.string
  result_io.debugIO << "\n\n#{component.cartridge_name}: #{component_details}" unless component_details.blank?

  return result_io
end
add_env_var(gear, key, value) click to toggle source

Add an environment variable on gear on a remote node. Calls oo-env-var-add on the remote node

INPUTS:

  • gear: a Gear object

  • key: String - environment variable name

  • value: String - environment variable value

RETURNS:

  • MCollective result string: STDOUT from a command.

NOTES:

  • uses execute_direct

  • calls oo-env-var-add on the node

  • Should be a method on Gear?

# File lib/openshift/mcollective_application_container_proxy.rb, line 516
def add_env_var(gear, key, value)
  args = build_base_gear_args(gear)
  args['--with-key'] = key
  args['--with-value'] = value
  result = execute_direct(@@C_CONTROLLER, 'env-var-add', args)
  parse_result(result, gear)
end
add_ssl_cert(gear, ssl_cert, priv_key, server_alias, passphrase='') click to toggle source

Add an SSL certificate to a gear on the remote node and associate it with a server name. See node/bin/oo-ssl-cert-add

INPUTS:

  • gear: a Gear object

  • priv_key: String - the private key value

  • server_alias: String - the name of the server which will offer this key

  • passphrase: String - the private key passphrase or " if its unencrypted.

RETURNS: a parsed MCollective result

NOTES:

  • calls node script oo-ssl-cert-add

# File lib/openshift/mcollective_application_container_proxy.rb, line 470
def add_ssl_cert(gear, ssl_cert, priv_key, server_alias, passphrase='')
  args = build_base_gear_args(gear)
  args['--with-ssl-cert']       = ssl_cert
  args['--with-priv-key']       = priv_key
  args['--with-alias-name']     = server_alias
  args['--with-passphrase']     = passphrase
  result = execute_direct(@@C_CONTROLLER, 'ssl-cert-add', args)
  parse_result(result)
end
build_base_component_args(component, existing_args={}) click to toggle source
# File lib/openshift/mcollective_application_container_proxy.rb, line 344
def build_base_component_args(component, existing_args={})
  cart = component.cartridge_name
  existing_args['--cart-name'] = cart
  existing_args['--component-name'] = component.component_name
  existing_args['--with-software-version'] = component.version
  existing_args['--cartridge-vendor'] = component.cartridge_vendor
  if not component.cartridge_vendor.empty?
    clist = cart.split('-')
    if clist[0]==component.cartridge_vendor.to_s
      existing_args['--cart-name'] = clist[1..-1].join('-')
    end
  end
  existing_args
end
build_base_gear_args(gear, quota_blocks=nil, quota_files=nil) click to toggle source
# File lib/openshift/mcollective_application_container_proxy.rb, line 329
def build_base_gear_args(gear, quota_blocks=nil, quota_files=nil)
  app = gear.app
  args = Hash.new
  args['--with-app-uuid'] = app.uuid
  args['--with-app-name'] = app.name
  args['--with-container-uuid'] = gear.uuid
  args['--with-container-name'] = gear.name
  args['--with-quota-blocks'] = quota_blocks if quota_blocks
  args['--with-quota-files'] = quota_files if quota_files
  args['--with-namespace'] = app.domain.namespace
  args['--with-uid'] = gear.uid if gear.uid
  args['--with-request-id'] = Thread.current[:user_action_log_uuid]
  args
end
build_update_cluster_args(options, args) click to toggle source
# File lib/openshift/mcollective_application_container_proxy.rb, line 1714
def build_update_cluster_args(options, args)
  if options.has_key?(:rollback)
    args['--rollback'] = options[:rollback]
  else
    if options[:sync_new_gears] == true
      args['--sync-new-gears'] = true
    end

    proxy_args = []
    options[:proxy_gears].each do |gear|
      proxy_args << "#{gear.uuid},#{gear.name},#{gear.group_instance.application.domain_namespace},#{gear.public_hostname}"
    end

    args['--proxy-gears'] = proxy_args.join(' ')

    web_args = []
    options[:web_gears].each do |gear|
      # TODO eventually support multiple ports
      first_port_interface = gear.port_interfaces[0]

      # uuid, name, namespace, proxy_hostname, proxy port
      web_args << "#{gear.uuid},#{gear.name},#{gear.group_instance.application.domain_namespace},#{gear.public_hostname},#{first_port_interface.external_port}"
    end

    args['--web-gears'] = web_args.join(' ')
  end

  args
end
conceal_port(gear, component) click to toggle source

hide a TCP port (?)

INPUTS:

  • gear: a Gear object

  • cart: a Cartridge object

RETURNS:

  • a ResultIO of undetermined content

NOTES:

Deprecated: remove from the REST API and then delete this.

# File lib/openshift/mcollective_application_container_proxy.rb, line 1244
def conceal_port(gear, component)
  args = build_base_gear_args(gear)
  args = build_base_component_args(component, args)
  result = execute_direct(@@C_CONTROLLER, 'conceal-port', args)
  parse_result(result)
end
create(gear, quota_blocks=nil, quota_files=nil) click to toggle source

<<instance method>>

Execute the 'app-create' script on a node.

INPUTS:

  • gear: a Gear object

  • quota_blocks: Integer - max file space in blocks

  • quota_files: Integer - max files count

RETURNS:

  • MCollective "result", stdout and exit code

NOTES:

  • uses execute_direct

  • should raise an exception on fail to cause revert rather than in-line

  • causes oo-app-create to execute on a node

Constructs a shell command line to be executed by the MCollective agent on the node.

# File lib/openshift/mcollective_application_container_proxy.rb, line 380
def create(gear, quota_blocks=nil, quota_files=nil)
  app = gear.app
  result = nil
  (1..10).each do |i|
    args = build_base_gear_args(gear, quota_blocks, quota_files)

    # set the secret token for new gear creations
    # log an error if the application does not have its secret_token set
    if app.secret_token.present?
      args['--with-secret-token'] = app.secret_token
    else
      Rails.logger.error "The application #{app.name} (#{app._id.to_s}) does not have its secret token set"
    end

    mcoll_reply = execute_direct(@@C_CONTROLLER, 'app-create', args)

    begin
      result = parse_result(mcoll_reply, gear)
    rescue OpenShift::OOException => ooex
      # destroy the gear in case of failures
      # the UID will be unreserved up as part of rollback
      destroy(gear, true)

      # raise the exception if this is the last retry
      raise ooex if i == 10

      result = ooex.resultIO
      if result != nil && result.exitcode == 129 && has_uid_or_gid?(gear.uid) # Code to indicate uid already taken
        gear.uid = reserve_uid
        app.save
      else
        raise ooex
      end
    else
      break
    end
  end
  result
end
deploy(gear, hot_deploy=false, force_clean_build=false, ref=nil, artifact_url=nil) click to toggle source

Deploy a gear.

INPUTS:

  • gear: a Gear object

  • hot_deploy: indicates whether this is a hot deploy

  • force_clean_build: indicates whether this should be a clean build

  • ref: the ref to deploy

  • artifact_url: the url of the artifacts to deploy

RETURNS ResultIO: the result of running post-configure on the cartridge

# File lib/openshift/mcollective_application_container_proxy.rb, line 862
def deploy(gear, hot_deploy=false, force_clean_build=false, ref=nil, artifact_url=nil)
  result_io = ResultIO.new

  args = build_base_gear_args(gear)

  args['--with-hot-deploy'] = hot_deploy
  args['--with-force-clean-build'] = force_clean_build
  args['--with-ref'] = ref
  args['--with-artifact-url'] = artifact_url

  result_io = run_cartridge_command(@@C_CONTROLLER, gear, "deploy", args)
  return result_io
end
destroy(gear, keep_uid=false, uid=nil, skip_hooks=false) click to toggle source

Remove a gear from a node Optionally release a reserved UID from the District.

INPUTS:

  • gear: a Gear object

  • keep_uid: boolean

  • uid: Integer: reserved UID

  • skip_hooks: boolean

RETURNS:

  • STDOUT from the remote command

NOTES:

# File lib/openshift/mcollective_application_container_proxy.rb, line 436
def destroy(gear, keep_uid=false, uid=nil, skip_hooks=false)
  args = build_base_gear_args(gear)
  args['--skip-hooks'] = true if skip_hooks
  begin
    result = execute_direct(@@C_CONTROLLER, 'app-destroy', args)
    result_io = parse_result(result, gear)
  rescue Exception=>e
    raise e if has_gear?(gear.uuid)!=false
    result_io = ResultIO.new
  end

  uid = gear.uid unless uid

  if uid && !keep_uid
    unreserve_uid(uid)
  end
  return result_io
end
expose_port(gear, component) click to toggle source
# File lib/openshift/mcollective_application_container_proxy.rb, line 1221
def expose_port(gear, component)
  args = build_base_gear_args(gear)
  args = build_base_component_args(component, args)
  result = execute_direct(@@C_CONTROLLER, 'expose-port', args)
  parse_result(result)
end
force_stop(gear, cart) click to toggle source

Force gear services to stop

INPUTS:

  • gear: Gear object

  • cart: Cartridge object

RETURNS:

  • result string from STDOUT

NOTES:

# File lib/openshift/mcollective_application_container_proxy.rb, line 1005
def force_stop(gear, cart)
  args = build_base_gear_args(gear)
  result = execute_direct(@@C_CONTROLLER, 'force-stop', args)
  parse_result(result)
end
frontend_backup(gear) click to toggle source

Extracts the frontend httpd server configuration from a gear.

INPUTS:

  • gear: a Gear object

RETURNS:

  • String - backup blob from the front-end server

NOTES:

  • uses execute_direct

  • executes the 'frontend-backup' action on the gear.

  • method on the FrontendHttpd class

# File lib/openshift/mcollective_application_container_proxy.rb, line 1397
def frontend_backup(gear)
  app = gear.app
  args = Hash.new
  args['--with-container-uuid']=gear.uuid
  args['--with-container-name']=gear.name
  args['--with-namespace']=app.domain.namespace
  result = execute_direct(@@C_CONTROLLER, 'frontend-backup', args)
  result = parse_result(result)
  result.resultIO.string
end
frontend_restore(backup) click to toggle source

Transmits the frontend httpd server configuration to a gear.

INPUTS:

RETURNS:

  • String - "parsed result" of an MCollective reply

NOTES:

  • uses execute_direct

  • executes the 'frontend-backup' action on the gear.

  • method on the FrontendHttpd class

# File lib/openshift/mcollective_application_container_proxy.rb, line 1423
def frontend_restore(backup)
  result = execute_direct(@@C_CONTROLLER, 'frontend-restore', {'--with-backup' => backup})
  parse_result(result)
end
get_active_capacity() click to toggle source

<<accessor>> Get the "active capacity" of a node

INPUTS: none

RETURNS:

  • Float: the "active capacity" of a node

NOTES:

# File lib/openshift/mcollective_application_container_proxy.rb, line 657
def get_active_capacity
  rpc_get_fact_direct('active_capacity').to_f
end
get_add_authorized_ssh_key_job(gear, ssh_key, key_type=nil, comment=nil) click to toggle source

Create a job to add an authorized key

INPUTS:

  • gear: a Gear object

  • ssh_key: String - SSH public key string

  • key_type: String, Enum [dsa|rsa]

  • comment: String

RETURNS:

  • a RemoteJob object

NOTES:

  • uses RemoteJob

# File lib/openshift/mcollective_application_container_proxy.rb, line 1485
def get_add_authorized_ssh_key_job(gear, ssh_key, key_type=nil, comment=nil)
  args = build_base_gear_args(gear)
  args['--with-ssh-key'] = ssh_key
  args['--with-ssh-key-type'] = key_type if key_type
  args['--with-ssh-key-comment'] = comment if comment
  job = RemoteJob.new('openshift-origin-node', 'authorized-ssh-key-add', args)
  job
end
get_app_status(app) click to toggle source

get the status of an application

INPUT: app: an Application object

RETURN:

  • Array: [idle, leave_stopped]

NOTES:

# File lib/openshift/mcollective_application_container_proxy.rb, line 2190
def get_app_status(app)
  web_framework = nil
  app.requires.each do |feature|
    cart = CartridgeCache.find_cartridge(feature, app)
    next unless cart.categories.include? "web_framework"
    web_framework = cart.name
    break
  end

  component_instances = app.get_components_for_feature(web_framework)
  gear = component_instances.first.group_instance.gears.first
  idle, leave_stopped, quota_blocks, quota_files = get_cart_status(gear, component_instances.first)
  return idle, leave_stopped
end
get_available_cartridges() click to toggle source

<<class method>>

<<query>>

Query all nodes for all available cartridges

INPUTS:

  • none

RETURNS:

  • An array of OpenShift::Cartridge objects

NOTES:

# File lib/openshift/mcollective_application_container_proxy.rb, line 156
def get_available_cartridges
  args = Hash.new
  args['--porcelain'] = true
  args['--with-descriptors'] = true
  result = execute_direct(@@C_CONTROLLER, 'cartridge-list', args, false)
  result = parse_result(result)
  cart_data = JSON.parse(result.resultIO.string)
  cart_data.map! {|c| OpenShift::Cartridge.new.from_descriptor(YAML.load(c))}
end
get_broker_auth_key_add_job(gear, iv, token) click to toggle source

Create a job to add a broker auth key

INPUTS:

  • gear: a Gear object

  • iv: String - Broker auth initialization vector

  • token: String - Broker auth token

RETURNS:

  • a RemoteJob object

NOTES:

  • uses RemoteJob

# File lib/openshift/mcollective_application_container_proxy.rb, line 1572
def get_broker_auth_key_add_job(gear, iv, token)
  args = build_base_gear_args(gear)
  args['--with-iv'] = iv
  args['--with-token'] = token
  job = RemoteJob.new('openshift-origin-node', 'broker-auth-key-add', args)
  job
end
get_capacity() click to toggle source

<<accessor>> Get the "capacity" of a node

INPUTS: none

RETURNS:

  • Float: the "capacity" of a node

NOTES:

# File lib/openshift/mcollective_application_container_proxy.rb, line 640
def get_capacity
  rpc_get_fact_direct('capacity').to_f
end
get_cart_status(gear, component) click to toggle source

get the status of a cartridge in a gear?

INPUTS:

  • gear: a Gear object

  • cart_name: String

RETURNS:

  • Array: [idle, leave_stopped, quota_file, quota_blocks]

NOTES:

# File lib/openshift/mcollective_application_container_proxy.rb, line 2218
def get_cart_status(gear, component)
  app = gear.app
  source_container = gear.get_proxy
  leave_stopped = false
  idle = false
  quota_blocks = nil
  quota_files = nil
  log_debug "DEBUG: Getting existing app '#{app.name}' status before moving"
  do_with_retry('status') do
    result = source_container.status(gear, component)
    result.properties["attributes"][gear.uuid].each do |key, value|
      if key == 'status'
        case value
        when "ALREADY_STOPPED"
          leave_stopped = true
        when "ALREADY_IDLED"
          leave_stopped = true
          idle = true
        end
      elsif key == 'quota_blocks'
        quota_blocks = value
      elsif key == 'quota_files'
        quota_files = value
      end
    end
  end

  cart_name = component.cartridge_name
  if idle
    log_debug "DEBUG: Gear component '#{cart_name}' was idle"
  elsif leave_stopped
    log_debug "DEBUG: Gear component '#{cart_name}' was stopped"
  else
    log_debug "DEBUG: Gear component '#{cart_name}' was running"
  end

  return [idle, leave_stopped, quota_blocks, quota_files]
end
get_conceal_port_job(gear, component) click to toggle source
# File lib/openshift/mcollective_application_container_proxy.rb, line 1209
def get_conceal_port_job(gear, component)
  args = build_base_gear_args(gear)
  args = build_base_component_args(component, args)
  RemoteJob.new('openshift-origin-node', 'conceal-port', args)
end
get_district_uuid() click to toggle source

<<accessor>> Get the district UUID (membership handle) of a node

INPUTS: none

RETURNS:

  • String: the UUID of a node's district

NOTES:

# File lib/openshift/mcollective_application_container_proxy.rb, line 674
def get_district_uuid
  rpc_get_fact_direct('district_uuid')
end
get_env_var_add_job(gear, key, value) click to toggle source

Get status on an add env var job?

INPUTS:

  • gear: a Gear object

  • key: an environment variable name

  • value: and environment variable value

RETURNS:

  • a RemoteJob object

NOTES:

  • uses RemoteJob

# File lib/openshift/mcollective_application_container_proxy.rb, line 1442
def get_env_var_add_job(gear, key, value)
  args = build_base_gear_args(gear)
  args['--with-key'] = key
  args['--with-value'] = value
  job = RemoteJob.new('openshift-origin-node', 'env-var-add', args)
  job
end
get_env_var_remove_job(gear, key) click to toggle source

Create a job to remove an environment variable

INPUTS:

  • gear: a Gear object

  • key: an environment variable name

RETURNS:

  • a RemoteJob object

NOTES:

  • uses RemoteJob

# File lib/openshift/mcollective_application_container_proxy.rb, line 1463
def get_env_var_remove_job(gear, key)
  args = build_base_gear_args(gear)
  args['--with-key'] = key
  job = RemoteJob.new('openshift-origin-node', 'env-var-remove', args)
  job
end
get_execute_connector_job(gear, component, connector_name, connection_type, input_args, pub_cart=nil) click to toggle source

Create a job to execute a connector hook ??

INPUTS:

  • gear: a Gear object

  • cart: a Cartridge object

  • connector_name: String

  • input_args: Array of String

RETURNS:

  • a RemoteJob object

NOTES:

  • uses RemoteJob

# File lib/openshift/mcollective_application_container_proxy.rb, line 1595
def get_execute_connector_job(gear, component, connector_name, connection_type, input_args, pub_cart=nil)
  args = build_base_gear_args(gear)
  cart = component.cartridge_name
  args = build_base_component_args(component, args)
  args['--hook-name'] = connector_name
  args['--publishing-cart-name'] = pub_cart if pub_cart
  args['--connection-type'] = connection_type
  # Need to pass down args as hash for subscriber ENV hooks only
  if connection_type.start_with?("ENV:") && pub_cart
    args['--input-args'] = input_args
  else
    args['--input-args'] = input_args.join(" ")
  end
  job = RemoteJob.new('openshift-origin-node', 'connector-execute', args)
  job
end
get_expose_port_job(gear, component) click to toggle source

expose a TCP port

INPUTS:

  • gear: a Gear object

  • cart: a Cartridge object

RETURNS:

  • a ResultIO of undetermined content

NOTES:

# File lib/openshift/mcollective_application_container_proxy.rb, line 1203
def get_expose_port_job(gear, component)
  args = build_base_gear_args(gear)
  args = build_base_component_args(component, args)
  RemoteJob.new('openshift-origin-node', 'expose-port', args)
end
get_fix_authorized_ssh_keys_job(gear, ssh_keys) click to toggle source

Create a job to remove existing authorized ssh keys and add the provided ones to the gear

INPUTS:

  • gear: a Gear object

  • ssh_keys: Array - SSH public key list

RETURNS:

  • a RemoteJob object

NOTES:

  • uses RemoteJob

# File lib/openshift/mcollective_application_container_proxy.rb, line 1551
def get_fix_authorized_ssh_keys_job(gear, ssh_keys)
  args = build_base_gear_args(gear)
  args['--with-ssh-keys'] = ssh_keys.map {|k| {'key' => k['content'], 'type' => k['type'], 'comment' => k['name']}}
  job = RemoteJob.new('openshift-origin-node', 'authorized-ssh-keys-replace', args)
  job
end
get_force_stop_job(gear, component) click to toggle source
# File lib/openshift/mcollective_application_container_proxy.rb, line 1011
def get_force_stop_job(gear, component)
  args = build_base_gear_args(gear)
  args = build_base_component_args(component, args)
  RemoteJob.new('openshift-origin-node', 'force-stop', args)
end
get_gear_envs(gear_uuid) click to toggle source

Returns a hash of env variables for a given gear uuid

INPUTS:

  • gear_uuid: String

RETURNS:

  • Hash

NOTES:

# File lib/openshift/mcollective_application_container_proxy.rb, line 2421
def get_gear_envs(gear_uuid)
  MCollectiveApplicationContainerProxy.rpc_exec('openshift', @id) do |client|
    client.get_gear_envs(:uuid => gear_uuid) do |response|
      output = response[:body][:data][:output]
      return output
    end
  end
end
get_ip_address() click to toggle source

<<accessor>> Get the public IP address of a Node

INPUTS: none

RETURNS:

  • String: the public IP address of a node

NOTES:

# File lib/openshift/mcollective_application_container_proxy.rb, line 691
def get_ip_address
  rpc_get_fact_direct('ipaddress')
end
get_node_details(name_list) click to toggle source

get details about instance's node

RETURN:

  • Map: name => values for one node

NOTES:

# File lib/openshift/mcollective_application_container_proxy.rb, line 2266
def get_node_details(name_list)
  rpc_get_facts_direct(name_list)
end
get_node_profile() click to toggle source

<<accessor>> Get the "node profile" of a Node

INPUTS: none

RETURNS:

  • String: the "node profile" of a node

NOTES:

# File lib/openshift/mcollective_application_container_proxy.rb, line 725
def get_node_profile
  rpc_get_fact_direct('node_profile')
end
get_public_hostname() click to toggle source

<<accessor>> Get the public hostname of a Node

INPUTS: none

RETURNS:

  • String: the public hostname of a node

NOTES:

# File lib/openshift/mcollective_application_container_proxy.rb, line 623
def get_public_hostname
  rpc_get_fact_direct('public_hostname')
end
get_public_ip_address() click to toggle source

<<accessor>> Get the public IP address of a Node

INPUTS: none

RETURNS:

  • String: the public IP address of a node

NOTES:

# File lib/openshift/mcollective_application_container_proxy.rb, line 708
def get_public_ip_address
  rpc_get_fact_direct('public_ip')
end
get_quota(gear) click to toggle source

<<object method>>

<<attribute getter>>

Request the disk quotas from a Gear on a node

RETURNS:

  • an array with following information:

[Filesystem, blocks_used, blocks_soft_limit, blocks_hard_limit, inodes_used, inodes_soft_limit, inodes_hard_limit]

RAISES:

  • OpenShift::NodeException

NOTES

# File lib/openshift/mcollective_application_container_proxy.rb, line 185
def get_quota(gear)
  args = Hash.new
  args['--uuid'] = gear.uuid
  reply = execute_direct(@@C_CONTROLLER, 'get-quota', args, false)

  output = nil
  exitcode = 0
  if reply and reply.length > 0
    mcoll_result = reply[0]
    if (mcoll_result && (defined? mcoll_result.results) && !mcoll_result.results[:data].nil?)
      output = mcoll_result.results[:data][:output]
      exitcode = mcoll_result.results[:data][:exitcode]
      raise OpenShift::NodeException.new("Failed to get quota for user: #{output}", 143) unless exitcode == 0
    else
      raise OpenShift::NodeException.new("Node execution failure (error getting result from node).", 143)
    end
  else
    raise OpenShift::NodeException.new("Node execution failure (error getting result from node).", 143)
  end
  output
end
get_quota_blocks() click to toggle source

<<accessor>> Get the quota blocks of a Node

Is this disk available or the default quota limit? It's from Facter.

INPUTS: none

RETURNS:

  • Integer: the "quota blocks" of a node

NOTES:

# File lib/openshift/mcollective_application_container_proxy.rb, line 745
def get_quota_blocks
  rpc_get_fact_direct('quota_blocks').to_i
end
get_quota_files() click to toggle source

<<accessor>> Get the quota files of a Node

Is this disk available or the default quota limit? It's from Facter.

INPUTS: none

RETURNS:

  • Integer: the "quota files" of a node

NOTES:

# File lib/openshift/mcollective_application_container_proxy.rb, line 765
def get_quota_files
  rpc_get_fact_direct('quota_files').to_i
end
get_reload_job(gear, component) click to toggle source
# File lib/openshift/mcollective_application_container_proxy.rb, line 1072
def get_reload_job(gear, component)
  args = build_base_gear_args(gear)
  args = build_base_component_args(component, args)
  RemoteJob.new('openshift-origin-node', 'reload', args)
end
get_remove_authorized_ssh_key_job(gear, ssh_key, key_type=nil, comment=nil) click to toggle source

Create a job to remove an authorized key.

INPUTS:

  • gear: a Gear object

  • ssh_key: String - SSH public key string

  • key_type: String

  • comment: String

RETURNS:

  • a RemoteJob object

NOTES:

  • uses RemoteJob

# File lib/openshift/mcollective_application_container_proxy.rb, line 1529
def get_remove_authorized_ssh_key_job(gear, ssh_key, key_type=nil, comment=nil)
  args = build_base_gear_args(gear)
  args['--with-ssh-key'] = ssh_key
  args['--with-ssh-key-type'] = key_type if key_type
  args['--with-ssh-comment'] = comment if comment
  job = RemoteJob.new('openshift-origin-node', 'authorized-ssh-key-remove', args)
  job
end
get_restart_job(gear, component, all=false) click to toggle source
# File lib/openshift/mcollective_application_container_proxy.rb, line 1040
def get_restart_job(gear, component, all=false)
  args = build_base_gear_args(gear)
  args = build_base_component_args(component, args)
  args['--all'] = all
  args['--parallel_concurrency_ratio'] = 0.5
  RemoteJob.new('openshift-origin-node', 'restart', args)
end
get_show_gear_quota_job(gear) click to toggle source

Create a job to check the disk quota on a gear

INPUTS:

  • gear: a Gear object

RETURNS:

  • a RemoteJob object

NOTES:

  • uses RemoteJob

# File lib/openshift/mcollective_application_container_proxy.rb, line 1685
def get_show_gear_quota_job(gear)
  args = Hash.new
  args['--uuid'] = gear.uuid
  job = RemoteJob.new('openshift-origin-node', 'get-quota', args)
  job
end
get_show_port_job(gear, component) click to toggle source
# File lib/openshift/mcollective_application_container_proxy.rb, line 1215
def get_show_port_job(gear, component)
  args = build_base_gear_args(gear)
  args = build_base_component_args(component, args)
  RemoteJob.new('openshift-origin-node', 'show-port', args)
end
get_show_state_job(gear) click to toggle source

Create a job to return the state of a gear

INPUTS:

  • gear: a Gear object

RETURNS:

  • a RemoteJob object

NOTES:

  • uses RemoteJob

# File lib/openshift/mcollective_application_container_proxy.rb, line 1647
def get_show_state_job(gear)
  args = build_base_gear_args(gear)
  job = RemoteJob.new('openshift-origin-node', 'app-state-show', args)
  job
end
get_start_job(gear, component) click to toggle source
# File lib/openshift/mcollective_application_container_proxy.rb, line 954
def get_start_job(gear, component)
  args = build_base_gear_args(gear)
  args = build_base_component_args(component, args)
  RemoteJob.new('openshift-origin-node', 'start', args)
end
get_status_job(gear, component) click to toggle source

Create a job to get status of an application

INPUTS:

  • gear: a Gear object

  • cart: a Cartridge object

RETURNS:

  • a RemoteJob object

NOTES:

  • uses RemoteJob

# File lib/openshift/mcollective_application_container_proxy.rb, line 1667
def get_status_job(gear, component)
  args = build_base_gear_args(gear)
  job = RemoteJob.new(component.cartridge_name, 'status', args)
  job
end
get_stop_job(gear, component) click to toggle source
# File lib/openshift/mcollective_application_container_proxy.rb, line 984
def get_stop_job(gear, component)
  args = build_base_gear_args(gear)
  args = build_base_component_args(component, args)
  RemoteJob.new('openshift-origin-node', 'stop', args)
end
get_threaddump_job(gear, component) click to toggle source
# File lib/openshift/mcollective_application_container_proxy.rb, line 1154
def get_threaddump_job(gear, component)
  args = build_base_gear_args(gear)
  args = build_base_component_args(component, args)
  RemoteJob.new('openshift-origin-node', 'threaddump', args)
end
get_tidy_job(gear, component) click to toggle source
# File lib/openshift/mcollective_application_container_proxy.rb, line 1122
def get_tidy_job(gear, component)
  args = build_base_gear_args(gear)
  args = build_base_component_args(component, args)
  RemoteJob.new('openshift-origin-node', 'tidy', args)
end
get_unsubscribe_job(gear, component, publish_cart_name) click to toggle source

Create a job to unsubscribe

INPUTS:

  • gear: a Gear object

  • cart: a Cartridge object

  • publish_cart_name: a Cartridge object

RETURNS:

  • a RemoteJob object

NOTES:

  • uses RemoteJob

# File lib/openshift/mcollective_application_container_proxy.rb, line 1626
def get_unsubscribe_job(gear, component, publish_cart_name)
  args = build_base_gear_args(gear)
  cart = component.cartridge_name
  args = build_base_component_args(component, args)
  args['--publishing-cart-name'] = publish_cart_name
  job = RemoteJob.new('openshift-origin-node', 'unsubscribe', args)
  job
end
get_update_cluster_job(gear, options) click to toggle source
# File lib/openshift/mcollective_application_container_proxy.rb, line 1751
def get_update_cluster_job(gear, options)
  args = build_base_gear_args(gear)
  args = build_update_cluster_args(options, args)
  RemoteJob.new(@@C_CONTROLLER, 'update-cluster', args)
end
get_update_configuration_job(gear, config) click to toggle source

Create a job to update gear configuration

INPUTS:

  • gear: a Gear object

  • config: a Hash of config to update

RETURNS:

  • a RemoteJob object

NOTES:

  • uses RemoteJob

# File lib/openshift/mcollective_application_container_proxy.rb, line 1507
def get_update_configuration_job(gear, config)
  args = build_base_gear_args(gear)
  args['--with-config'] = config
  job = RemoteJob.new('openshift-origin-node', 'update-configuration', args)
  job
end
get_update_gear_quota_job(gear, storage_in_gb, inodes) click to toggle source

Create a job to change the disk quotas on a gear

INPUTS:

  • gear: a Gear object

RETURNS:

  • a RemoteJob object

NOTES:

  • uses RemoteJob

# File lib/openshift/mcollective_application_container_proxy.rb, line 1704
def get_update_gear_quota_job(gear, storage_in_gb, inodes)
  args = Hash.new
  args['--uuid']   = gear.uuid
  # quota command acts on 1K blocks
  args['--blocks'] = Integer(storage_in_gb * 1024 * 1024)
  args['--inodes'] = inodes unless inodes.to_s.empty?
  job = RemoteJob.new('openshift-origin-node', 'set-quota', args)
  job
end
get_update_proxy_status_job(gear, options) click to toggle source

Enable/disable a target gear in the proxy component

# File lib/openshift/mcollective_application_container_proxy.rb, line 1758
def get_update_proxy_status_job(gear, options)
  args = build_base_gear_args(gear)
  #TODO support specifying the proxy component/cart?
  #args = build_base_component_args(proxy_component, args)
  args['--action'] = options[:action]
  args['--gear_uuid'] = options[:gear_uuid]
  args['--persist'] = options[:persist]
  RemoteJob.new(@@C_CONTROLLER, 'update-proxy-status', args)
end
list_user_env_vars(gear, env_var_names) click to toggle source

List all or selected user environment variables for the app

INPUTS:

  • gear: a Gear object

  • env_var_names: List of environment variable names, e.g.:['FOO', 'BAR']

RETURNS:

  • String: stdout from a command

NOTES:

  • calls execute_direct

  • executes the 'user-var-list' action on the node

# File lib/openshift/mcollective_application_container_proxy.rb, line 1375
def list_user_env_vars(gear, env_var_names)
  args = build_base_gear_args(gear)
  args['--with-keys'] = env_var_names.join(' ') if env_var_names.present?
  result = execute_direct(@@C_CONTROLLER, 'user-var-list', args)
  parse_result(result)
end
move_gear(gear, destination_container, destination_district_uuid, change_district, node_profile) click to toggle source
# File lib/openshift/mcollective_application_container_proxy.rb, line 1903
def move_gear(gear, destination_container, destination_district_uuid, change_district, node_profile)
  app = gear.app
  reply = ResultIO.new
  state_map = {}

  if gear.get_proxy.get_district_uuid == 'NONE'
    raise OpenShift::UserException.new("Move gear *not* supported without districts.")
  end

  # resolve destination_container according to district
  destination_container, destination_district_uuid, district_changed = resolve_destination(gear, destination_container, destination_district_uuid, change_district, node_profile)

  source_container = gear.get_proxy

  if source_container.id == destination_container.id
    log_debug "Cannot move a gear within the same node. The source container and destination container are the same."
    raise OpenShift::UserException.new("Error moving app. Destination container same as source container.", 1)
  end

  destination_node_profile = destination_container.get_node_profile
  if source_container.get_node_profile != destination_node_profile and app.scalable
    log_debug "Cannot change node_profile for *scalable* application gear - this operation is not supported. The destination container's node profile is #{destination_node_profile}, while the gear's node_profile is #{gear.group_instance.gear_size}"
    raise OpenShift::UserException.new("Error moving app.  Cannot change node profile for *scalable* app.", 1)
  end

  # get the application state
  idle, leave_stopped = get_app_status(app)

  # get the quota blocks and files from the gear
  gear_quota = get_quota(gear)
  quota_blocks = Integer(gear_quota[3])
  quota_files = Integer(gear_quota[6])

  gi = gear.group_instance
  gi.all_component_instances.each do |cinst|
    next if cinst.is_sparse? and (not gear.sparse_carts.include? cinst._id) and (not gear.host_singletons)
    state_map[cinst.cartridge_name] = [idle, leave_stopped]
  end

  begin
    # pre-move
    reply.append move_gear_pre(gear, state_map)

    if district_changed
      destination_container.reserve_uid(destination_district_uuid, gear.uid)
      log_debug "DEBUG: Reserved uid '#{gear.uid}' on district: '#{destination_district_uuid}'"
    end
    begin
      # rsync gear with destination container
      rsync_destination_container(gear, destination_container, destination_district_uuid, quota_blocks, quota_files)

      start_order,stop_order = app.calculate_component_orders
      gi_comps = gear.group_instance.all_component_instances.to_a
      start_order.each do |cinst|
        next if not gi_comps.include? cinst
        next if cinst.is_sparse? and (not gear.sparse_carts.include? cinst._id) and (not gear.host_singletons)
        cart = cinst.cartridge_name
        idle, leave_stopped = state_map[cart]

        if app.scalable and not CartridgeCache.find_cartridge(cart, app).categories.include? "web_proxy"
          begin
            reply.append destination_container.expose_port(gear, cinst)
          rescue Exception=>e
            # just pass because some embedded cartridges do not have expose-port hook implemented (e.g. jenkins-client)
          end
        end
      end

      # start the gears again and change DNS entry
      reply.append move_gear_post(gear, destination_container, state_map)
      # app.elaborate_descriptor

      # update all proxy gear registries and configs
      app.update_cluster if app.scalable

      app.execute_connections
      if app.scalable
        # execute connections restart the haproxy service, so stop it explicitly if needed
        stop_order.each do |cinst|
          next if not gi_comps.include? cinst
          cart = cinst.cartridge_name
          idle, leave_stopped = state_map[cart]
          if leave_stopped and CartridgeCache.find_cartridge(cart, app).categories.include? "web_proxy"
            log_debug "DEBUG: Explicitly stopping cartridge '#{cart}' in '#{app.name}' after move on #{destination_container.id}"
            reply.append destination_container.stop(gear, cinst)
          end
        end

        app.update_proxy_status(action: :enable, gear_uuid: gear.uuid)
      end
      app.save

    rescue Exception => e
      gear.server_identity = source_container.id
      gear.group_instance.gear_size = source_container.get_node_profile
      # destroy destination
      log_debug "DEBUG: Moving failed.  Rolling back gear '#{gear.name}' in '#{app.name}' with delete on '#{destination_container.id}'"
      reply.append destination_container.destroy(gear, !district_changed, nil, true)
      raise
    end
  rescue Exception => e
    begin
      gi_comps = gear.group_instance.all_component_instances.to_a
      start_order,stop_order = app.calculate_component_orders
      # start source
      start_order.each do |cinst|
        next if not gi_comps.include? cinst
        next if cinst.is_sparse? and (not gear.sparse_carts.include? cinst._id) and (not gear.host_singletons)
        cart = cinst.cartridge_name
        idle, leave_stopped = state_map[cart]
        if not leave_stopped
          args = build_base_gear_args(gear)
          args = build_base_component_args(cinst, args)
          reply.append source_container.run_cartridge_command(cart, gear, "start", args, false)
        end
      end

      app.update_proxy_status(action: :enable, gear_uuid: gear.uuid)
    ensure
      raise
    end
  end

  move_gear_destroy_old(gear, source_container, destination_container, district_changed)

  log_debug "Successfully moved gear with uuid '#{gear.uuid}' of app '#{app.name}' from '#{source_container.id}' to '#{destination_container.id}'"
  reply
end
move_gear_destroy_old(gear, source_container, destination_container, district_changed) click to toggle source

Remove and destroy a old gear after migration

INPUTS:

  • gear: a Gear object

  • source_container: ??

  • destination_container ??

  • district_changed: boolean

RETURNS:

  • a ResultIO object

CATCHES:

  • Exception

NOTES:

  • uses source_container.destroy

# File lib/openshift/mcollective_application_container_proxy.rb, line 2050
def move_gear_destroy_old(gear, source_container, destination_container, district_changed)
  app = gear.app
  reply = ResultIO.new
  log_debug "DEBUG: Deconfiguring old app '#{app.name}' on #{source_container.id} after move"
  begin
    reply.append source_container.destroy(gear, !district_changed, gear.uid, true)
  rescue Exception => e
    log_debug "DEBUG: The application '#{app.name}' with gear uuid '#{gear.uuid}' is now moved to '#{destination_container.id}' but not completely deconfigured from '#{source_container.id}'"
    raise
  end
  reply
end
move_gear_post(gear, destination_container, state_map) click to toggle source

Re-start a gear after migration

INPUTS:

  • gear: a Gear object

  • destination_container: an ApplicationContainerProxy object?

  • state_map: ??

RETURNS:

  • ResultIO

NOTES:

  • uses OpenShift::DnsService

*

# File lib/openshift/mcollective_application_container_proxy.rb, line 1784
def move_gear_post(gear, destination_container, state_map)
  app = gear.app
  reply = ResultIO.new
  gi = gear.group_instance
  gear_components = gi.all_component_instances
  start_order, stop_order = app.calculate_component_orders
  source_container = gear.get_proxy
  start_order.each do |cinst|
    next if not gear_components.include? cinst
    next if cinst.is_sparse? and (not gear.sparse_carts.include? cinst._id) and (not gear.host_singletons)
    cart = cinst.cartridge_name
    idle, leave_stopped = state_map[cart]
    unless leave_stopped
      log_debug "DEBUG: Starting cartridge '#{cart}' in '#{app.name}' after move on #{destination_container.id}"
      args = build_base_gear_args(gear)
      args = build_base_component_args(cinst, args)
      reply.append destination_container.send(:run_cartridge_command, cart, gear, "start", args, false)
    end
  end

  log_debug "DEBUG: Fixing DNS and mongo for gear '#{gear.name}' after move"
  log_debug "DEBUG: Changing server identity of '#{gear.name}' from '#{source_container.id}' to '#{destination_container.id}'"
  gear.server_identity = destination_container.id
  gear.group_instance.gear_size = destination_container.get_node_profile
  begin
    dns = OpenShift::DnsService.instance
    public_hostname = destination_container.get_public_hostname
    dns.modify_application(gear.name, app.domain.namespace, public_hostname)
    dns.publish
  ensure
    dns.close
  end
  reply
end
move_gear_pre(gear, state_map) click to toggle source

Prepare to move a gear from one node to another

INPUTS:

  • gear: A Gear object

  • state_map: ??

RETURNS:

  • a ResultIO object

NOTES:

  • uses ResultIO

  • ResultIO is composed

  • uses Container (from gear)

# File lib/openshift/mcollective_application_container_proxy.rb, line 1834
def move_gear_pre(gear, state_map)
  app = gear.app
  reply = ResultIO.new
  source_container = gear.get_proxy
  gi_comps = gear.group_instance.all_component_instances.to_a
  start_order,stop_order = app.calculate_component_orders

  app.update_proxy_status(action: :disable, gear_uuid: gear.uuid) if app.scalable

  stop_order.each { |cinst|
    next if not gi_comps.include? cinst
    next if cinst.is_sparse? and (not gear.sparse_carts.include? cinst._id) and (not gear.host_singletons)
    cart = cinst.cartridge_name
    idle, leave_stopped = state_map[cart]
    # stop the cartridge if it needs to
    unless leave_stopped
      log_debug "DEBUG: Stopping existing app cartridge '#{cart}' before moving"
      begin
        do_with_retry('stop') do
          reply.append source_container.stop(gear, cinst)
        end
      rescue Exception=>e
        # a force-stop will be applied if its a framework cartridge, so ignore the failure on stop
        if not framework_carts(app).include? cart
          raise e
        end
      end
      if framework_carts(app).include? cart
        log_debug "DEBUG: Force stopping existing app cartridge '#{cart}' before moving"
        do_with_retry('force-stop') do
          reply.append source_container.force_stop(gear, cinst)
        end
      end
    end
  }
  reply
end
move_gear_secure(gear, destination_container, destination_district_uuid, change_district, node_profile) click to toggle source

Move a gear from one node to another

INPUTS

  • gear: a Gear object

  • destination_container: An ApplicationContainerProxy?

  • destination_district_uuid: String

  • change_district: Boolean

  • node_profile: String

RETURNS:

  • ResultIO

RAISES:

  • OpenShift::UserException

CATCHES:

  • Exception

NOTES:

# File lib/openshift/mcollective_application_container_proxy.rb, line 1896
def move_gear_secure(gear, destination_container, destination_district_uuid, change_district, node_profile)
  app = gear.app
  Application.run_in_application_lock(app) do
    move_gear(gear, destination_container, destination_district_uuid, change_district, node_profile)
  end
end
post_configure_component(gear, component, template_git_url=nil) click to toggle source

Post configuration for a cartridge on a gear.

INPUTS:

  • gear: a Gear object

  • component: component_instance object

  • template_git_url: a url of a git repo containing a cart overlay

RETURNS ResultIO: the result of running post-configure on the cartridge

# File lib/openshift/mcollective_application_container_proxy.rb, line 826
def post_configure_component(gear, component, template_git_url=nil)
  result_io = ResultIO.new
  cart = component.cartridge_name

  args = build_base_gear_args(gear)
  args = build_base_component_args(component, args)

  if template_git_url.present?
    args['--with-template-git-url'] = template_git_url
  end

  if framework_carts(gear.app).include?(cart) or embedded_carts(gear.app).include?(cart)
    result_io = run_cartridge_command(cart, gear, "post-configure", args)
  else
    #no-op
  end

  component_details = result_io.appInfoIO.string.empty? ? '' : result_io.appInfoIO.string
  result_io.debugIO << "#{cart}: #{component_details}" unless component_details.blank?

  return result_io
end
reload(gear, component) click to toggle source

"reload" cart services on a gear. Accept config update without restarting?

INPUTS:

  • gear: a Gear object

  • cart: a Cartridge object

RETURNS:

  • a ResultIO of undefined content

NOTES:

# File lib/openshift/mcollective_application_container_proxy.rb, line 1064
def reload(gear, component)
  args = build_base_gear_args(gear)
  cart = component.cartridge_name
  args = build_base_component_args(component, args)

  run_cartridge_command_ignore_components(cart, gear, "reload", args)
end
remove_alias(gear, server_alias) click to toggle source

remove an application alias to a gear

INPUTS:

  • gear: a Gear object

  • server_alias: String - a new FQDN for the gear

RETURNS:

  • String: stdout from a command

NOTES:

  • calls execute_direct

  • executes the 'remove-alias' action on the gear.

  • method on Gear?

# File lib/openshift/mcollective_application_container_proxy.rb, line 1308
def remove_alias(gear, server_alias)
  args = build_base_gear_args(gear)
  args['--with-alias-name']=server_alias
  result = execute_direct(@@C_CONTROLLER, 'remove-alias', args)
  parse_result(result)
end
remove_broker_auth_key(gear) click to toggle source

Remove a broker auth key. The broker auth key allows an application to request scaling and other actions from the broker.

INPUTS:

  • gear: a Gear object

RETURNS:

  • mcollective parsed result string (stdout)

NOTES:

# File lib/openshift/mcollective_application_container_proxy.rb, line 583
def remove_broker_auth_key(gear)
  args = build_base_gear_args(gear)
  result = execute_direct(@@C_CONTROLLER, 'broker-auth-key-remove', args)
  parse_result(result, gear)
end
remove_component(gear, component) click to toggle source

Remove a component from a gear

INPUTS:

  • gear: a Gear object

  • component: String: a component name

RETURNS:

  • ResultIO? String?

NOTES

  • method on gear?

# File lib/openshift/mcollective_application_container_proxy.rb, line 910
def remove_component(gear, component)
  app = gear.app
  args = build_base_gear_args(gear)
  args = build_base_component_args(component, args)
  cart = component.cartridge_name

  Rails.logger.debug "DEBUG: Deconfiguring cartridge '#{cart}' in application '#{app.name}' on node '#{@id}'"

  resultIO = ResultIO.new
  begin
    resultIO = run_cartridge_command(cart, gear, 'deconfigure', args)
  rescue OpenShift::NodeException => e
    if has_app_cartridge?(app.uuid, gear.uuid, cart)
      raise
    else
      Rails.logger.debug "DEBUG: Cartridge '#{cart}' not found on within application '#{app.name}/#{gear.uuid}'.  Continuing with deconfigure."
    end
  end
  return resultIO
end
remove_env_var(gear, key) click to toggle source

Remove an environment variable on gear on a remote node

INPUTS:

  • gear: a Gear object

  • key: String - environment variable name

RETURNS:

  • MCollective result string: STDOUT from a command.

NOTES:

# File lib/openshift/mcollective_application_container_proxy.rb, line 538
def remove_env_var(gear, key)
  args = build_base_gear_args(gear)
  args['--with-key'] = key
  result = execute_direct(@@C_CONTROLLER, 'env-var-remove', args)
  parse_result(result, gear)
end
remove_ssl_cert(gear, server_alias) click to toggle source

remove an SSL certificate to a gear on the remote node. See node/bin/oo-ssl-cert-remove

INPUTS:

  • gear: a Gear object

  • server_alias: String - the name of the server which will offer this key

RETURNS: a parsed MCollective result

NOTES:

  • calls node script oo-ssl-cert-remove

# File lib/openshift/mcollective_application_container_proxy.rb, line 492
def remove_ssl_cert(gear, server_alias)
  args = build_base_gear_args(gear)
  args['--with-alias-name']     = server_alias
  result = execute_direct(@@C_CONTROLLER, 'ssl-cert-remove', args)
  parse_result(result)
end
reserve_uid(district_uuid=nil, preferred_uid=nil) click to toggle source

Reserve a UID within a district or service

UIDs must be unique in a district to allow migration without requiring reassigning Username (Gear UUID) and Unix User UID on migrate Perhaps a query on the nodes for "next UID"?

INPUTS:

  • district_uuid: String: District handle or identifier

  • preferred_uid: Integer

RAISES:

  • OpenShift::OOException

NOTES:

  • a method on District class of the node.

# File lib/openshift/mcollective_application_container_proxy.rb, line 268
def reserve_uid(district_uuid=nil, preferred_uid=nil)
  reserved_uid = nil
  if Rails.configuration.msg_broker[:districts][:enabled]
    if @district
      district_uuid = @district.uuid
    elsif !district_uuid
      if @id
        begin
          district = District.find_by({"server_identities.name" => @id})
          district_uuid = district.uuid
        rescue Mongoid::Errors::DocumentNotFound
          district_uuid = 'NONE'
        end
      else
        district_uuid = get_district_uuid
      end
    end
    if district_uuid && district_uuid != 'NONE'
      reserved_uid = District::reserve_uid(district_uuid, preferred_uid)
      raise OpenShift::OOException.new("uid could not be reserved in target district '#{district_uuid}'.  Please ensure the target district has available capacity.") unless reserved_uid
    end
  end
  reserved_uid
end
resolve_destination(gear, destination_container, destination_district_uuid, change_district, node_profile) click to toggle source

INPUTS:

  • gear: a Gear object

  • destination_container: ??

  • destination_district_uuid: String

  • change_district: Boolean

  • node_profile: String

RETURNS:

  • Array: [destination_container, destination_district_uuid, district_changed]

RAISES:

  • OpenShift::UserException

NOTES:

# File lib/openshift/mcollective_application_container_proxy.rb, line 2081
def resolve_destination(gear, destination_container, destination_district_uuid, change_district, node_profile)
  gear_exists_in_district = false
  required_uid = gear.uid
  source_container = gear.get_proxy
  source_district_uuid = source_container.get_district_uuid

  if node_profile and (destination_container or destination_district_uuid or !Rails.configuration.msg_broker[:node_profile_enabled])
    log_debug "DEBUG: Option node_profile '#{node_profile}' is being ignored either in favor of destination district/container "                     "or node_profile is disabled in msg broker configuration."
    node_profile = nil
  end

  if destination_container.nil?
    if !destination_district_uuid and !change_district
      destination_district_uuid = source_district_uuid unless source_district_uuid == 'NONE'
    end

    # Check to see if the gear's current district and the destination district are the same
    if (not destination_district_uuid.nil?) and (destination_district_uuid == source_district_uuid)
      gear_exists_in_district = true
      required_uid = nil
    end

    least_preferred_server_identities = [source_container.id]
    destination_gear_size = node_profile || gear.group_instance.gear_size
    destination_container = MCollectiveApplicationContainerProxy.find_available_impl(destination_gear_size, destination_district_uuid, nil, gear_exists_in_district, required_uid)
    log_debug "DEBUG: Destination container: #{destination_container.id}"
    destination_district_uuid = destination_container.get_district_uuid
  else
    if destination_district_uuid
      log_debug "DEBUG: Destination district uuid '#{destination_district_uuid}' is being ignored in favor of destination container #{destination_container.id}"
    end
    destination_district_uuid = destination_container.get_district_uuid
  end

  log_debug "DEBUG: Source district uuid: #{source_district_uuid}"
  log_debug "DEBUG: Destination district uuid: #{destination_district_uuid}"
  district_changed = (destination_district_uuid != source_district_uuid)

  if source_container.id == destination_container.id
    raise OpenShift::UserException.new("Error moving gear.  Old and new servers are the same: #{source_container.id}", 1)
  end
  return [destination_container, destination_district_uuid, district_changed]
end
restart(gear, component) click to toggle source

Stop and restart cart services on a gear

INPUTS:

  • gear: a Gear object

  • cart: a Cartridge object

RETURNS:

  • a ResultIO of undefined content

NOTES:

# File lib/openshift/mcollective_application_container_proxy.rb, line 1032
def restart(gear, component)
  args = build_base_gear_args(gear)
  cart = component.cartridge_name
  args = build_base_component_args(component, args)

  run_cartridge_command_ignore_components(cart, gear, "restart", args)
end
rsync_destination_container(gear, destination_container, destination_district_uuid, quota_blocks, quota_files) click to toggle source

copy the file contents of a gear on this node to a new gear on another

INPUTS:

  • gear: a Gear object

  • destination_container: an ApplicationContainerProxxy object

  • destination_district_uuid: String: a UUID handle

  • quota_blocks: Integer

  • quota_files: Integer

RETURNS:

  • ResultIO

RAISES:

  • OpenShift::NodeException

NOTES:

  • uses broker configuration rsync_keyfile

  • uses ssh-agent

  • uses ssh-add

  • uses rsync

  • runs all three commands in a single backtick eval

  • writes the eval output to log_debug

# File lib/openshift/mcollective_application_container_proxy.rb, line 2150
def rsync_destination_container(gear, destination_container, destination_district_uuid, quota_blocks, quota_files)
  app = gear.app
  reply = ResultIO.new
  source_container = gear.get_proxy
  log_debug "DEBUG: Creating new account for gear '#{gear.name}' on #{destination_container.id}"
  reply.append destination_container.create(gear, quota_blocks, quota_files)

  log_debug "DEBUG: Moving content for app '#{app.name}', gear '#{gear.name}' to #{destination_container.id}"
  rsync_keyfile = Rails.configuration.auth[:rsync_keyfile]
  log_debug `eval \`ssh-agent\`; ssh-add #{rsync_keyfile}; ssh -o StrictHostKeyChecking=no -A root@#{source_container.get_ip_address} "rsync -aAX -e 'ssh -o StrictHostKeyChecking=no' /var/lib/openshift/#{gear.uuid}/ root@#{destination_container.get_ip_address}:/var/lib/openshift/#{gear.uuid}/"; exit_code=$?; ssh-agent -k; exit $exit_code`
  if $?.exitstatus != 0
    raise OpenShift::NodeException.new("Error moving app '#{app.name}', gear '#{gear.name}' from #{source_container.id} to #{destination_container.id}", 143)
  end

  log_debug "DEBUG: Moving system components for app '#{app.name}', gear '#{gear.name}' to #{destination_container.id}"
  log_debug `eval \`ssh-agent\`; ssh-add #{rsync_keyfile}; ssh -o StrictHostKeyChecking=no -A root@#{source_container.get_ip_address} "rsync -aAX -e 'ssh -o StrictHostKeyChecking=no' --include '.httpd.d/' --include '.httpd.d/#{gear.uuid}_***' --include '#{gear.name}-#{app.domain.namespace}' --include '.last_access/' --include '.last_access/#{gear.uuid}' --exclude '*' /var/lib/openshift/ root@#{destination_container.get_ip_address}:/var/lib/openshift/"; exit_code=$?; ssh-agent -k; exit $exit_code`
  if $?.exitstatus != 0
    raise OpenShift::NodeException.new("Error moving system components for app '#{app.name}', gear '#{gear.name}' from #{source_container.id} to #{destination_container.id}", 143)
  end

  # Transfer the front-end configuration to the new gear
  backup = source_container.frontend_backup(gear)
  reply.append destination_container.frontend_restore(backup)
  reply
end
set_district(uuid, active, first_uid, max_uid) click to toggle source

Set the district of a node

INPUTS:

  • uuid: String

  • active: String (?)

  • first_uid: Integer

  • max_uid: Integer

RETURNS:

  • ResultIO?

NOTES:

  • uses MCollective::RPC::Client

  • uses ApplicationContainerProxy @id

# File lib/openshift/mcollective_application_container_proxy.rb, line 2359
def set_district(uuid, active, first_uid, max_uid)
  mc_args = { :uuid => uuid,
              :active => active,
              :first_uid => first_uid,
              :max_uid => max_uid}
  options = MCollectiveApplicationContainerProxy.rpc_options
  rpc_client = MCollectiveApplicationContainerProxy.get_rpc_client('openshift', options)
  result = nil
  begin
    Rails.logger.debug "DEBUG: rpc_client.custom_request('set_district', #{mc_args.inspect}, #{@id}, {'identity' => #{@id}})"
    result = rpc_client.custom_request('set_district', mc_args, @id, {'identity' => @id})
    Rails.logger.debug "DEBUG: #{result.inspect}"
  ensure
    rpc_client.disconnect
  end
  Rails.logger.debug result.inspect
  result
end
set_quota(gear, storage_in_gb, inodes) click to toggle source

<<object method>>

<<attribute setter>>

Set blocks hard limit and inodes hard limit for uuid. Effects disk quotas on Gear on Node

INPUT:

  • gear: A Gear object

  • storage_in_gb: integer

  • inodes: integer

RETURNS

  • Not sure, mcoll_result? a string?

RAISES:

  • OpenShift::NodeException

NOTES:

  • a pair of attribute setters on a Gear object

# File lib/openshift/mcollective_application_container_proxy.rb, line 228
def set_quota(gear, storage_in_gb, inodes)
  args = Hash.new
  args['--uuid']   = gear.uuid
  # quota command acts on 1K blocks
  args['--blocks'] = Integer(storage_in_gb * 1024 * 1024)
  args['--inodes'] = inodes unless inodes.nil?
  reply = execute_direct(@@C_CONTROLLER, 'set-quota', args, false)

  output = nil
  exitcode = 0
  if reply and reply.length > 0
    mcoll_result = reply[0]
    if (mcoll_result && (defined? mcoll_result.results) && !mcoll_result.results[:data].nil?)
      output = mcoll_result.results[:data][:output]
      exitcode = mcoll_result.results[:data][:exitcode]
      raise OpenShift::NodeException.new("Failed to set quota for user: #{output}", 143) unless exitcode == 0
    else
      raise OpenShift::NodeException.new("Node execution failure (error getting result from node).", 143)
    end
  else
    raise OpenShift::NodeException.new("Node execution failure (error getting result from node).", 143)
  end
end
set_user_env_vars(gear, env_vars, gears_ssh_endpoint) click to toggle source

Add or Update user environment variables to all gears in the app

INPUTS:

  • gear: a Gear object

  • env_vars: Array of environment variables, e.g.:[{'name'=>'FOO','value'=>'123'}, {'name'=>'BAR','value'=>'abc'}]

  • gears_ssh_endpoint: list of ssh gear endpoints

RETURNS:

  • String: stdout from a command

NOTES:

# File lib/openshift/mcollective_application_container_proxy.rb, line 1330
def set_user_env_vars(gear, env_vars, gears_ssh_endpoint)
  args = build_base_gear_args(gear)
  args['--with-variables'] = env_vars.to_json if env_vars.present?
  args['--with-gears'] = gears_ssh_endpoint.join(';') if gears_ssh_endpoint.present?
  result = execute_direct(@@C_CONTROLLER, 'user-var-add', args)
  parse_result(result)
end
show_port(gear, cart) click to toggle source

get information on a TCP port

INPUTS:

  • gear: a Gear object

  • cart: a Cartridge object

RETURNS:

  • a ResultIO of undetermined content

NOTES:

Deprecated: remove from the REST API and then delete this.

# File lib/openshift/mcollective_application_container_proxy.rb, line 1267
def show_port(gear, cart)
  ResultIO.new
end
show_state(gear) click to toggle source

Get the operating state of a gear

INPUTS:

RETURNS:

  • mcollective result string (stdout)

NOTES:

  • uses execute_direct

  • calls oo-app-state-show

  • Should be a method on Gear object

# File lib/openshift/mcollective_application_container_proxy.rb, line 604
def show_state(gear)
  args = build_base_gear_args(gear)
  result = execute_direct(@@C_CONTROLLER, 'app-state-show', args)
  parse_result(result, gear)
end
start(gear, component) click to toggle source

Start cartridge services within a gear

INPUTS:

  • gear: a Gear object

  • cart: a Cartridge object

RETURNS:

  • a ResultIO of undetermined state

NOTES:

# File lib/openshift/mcollective_application_container_proxy.rb, line 946
def start(gear, component)
  args = build_base_gear_args(gear)
  cart = component.cartridge_name
  args = build_base_component_args(component, args)

  run_cartridge_command_ignore_components(cart, gear, "start", args)
end
status(gear, component) click to toggle source

Get the status from cart services in an existing Gear

INPUTS:

  • gear: a Gear object

  • cart: a Cartridge object

RETURNS:

  • A ResultIO object of undetermined content

NOTES:

# File lib/openshift/mcollective_application_container_proxy.rb, line 1093
def status(gear, component)
  args = build_base_gear_args(gear)
  cart = component.cartridge_name
  args = build_base_component_args(component, args)

  run_cartridge_command_ignore_components(cart, gear, "status", args)
end
stop(gear, component) click to toggle source

Stop cartridge services within a gear

INPUTS:

  • gear: a Gear object

  • cart: a Cartridge object

RETURNS:

  • a ResultIO of undetermined state

NOTES:

# File lib/openshift/mcollective_application_container_proxy.rb, line 976
def stop(gear, component)
  args = build_base_gear_args(gear)
  cart = component.cartridge_name
  args = build_base_component_args(component, args)

  run_cartridge_command_ignore_components(cart, gear, "stop", args)
end
system_messages(gear, component) click to toggle source

"retrieve the system messages"

INPUTS:

  • gear: a Gear object

  • cart: a Cartridge object

RETURNS:

  • a ResultIO of undetermined content

NOTES:

# File lib/openshift/mcollective_application_container_proxy.rb, line 1175
def system_messages(gear, component)
  args = build_base_gear_args(gear)
  args = build_base_component_args(component, args)
  cart = component.cartridge_name
  args = build_base_component_args(component, args)

  if framework_carts(gear.app).include?(cart)
    run_cartridge_command(cart, gear, "system-messages", args)
  else
    ResultIO.new
  end
end
threaddump(gear, component) click to toggle source

dump the cartridge threads

INPUTS:

  • gear: a Gear object

  • cart: a Cartridge object

RETURNS:

  • a ResultIO of undetermined content

NOTES:

# File lib/openshift/mcollective_application_container_proxy.rb, line 1142
def threaddump(gear, component)
  args = build_base_gear_args(gear)
  cart = component.cartridge_name
  args = build_base_component_args(component, args)

  if framework_carts(gear.app).include?(cart)
    run_cartridge_command(cart, gear, "threaddump", args)
  else
    ResultIO.new
  end
end
tidy(gear, component) click to toggle source

Clean up unneeded artifacts in a gear

INPUTS:

  • gear: a Gear object

  • cart: a Cartridge object

RETURNS:

  • String: stdout from a command

NOTES

  • calls the 'tidy' hook on a Gear or app?

  • doesn't use cart input

  • calls execute_direct

# File lib/openshift/mcollective_application_container_proxy.rb, line 1116
def tidy(gear, component)
  args = build_base_gear_args(gear)
  result = execute_direct(@@C_CONTROLLER, 'tidy', args)
  parse_result(result)
end
unreserve_uid(uid, district_uuid=nil) click to toggle source

Release a UID reservation within a District

UIDs must be unique in a district to allow migration without requiring reassigning Username (Gear UUID) and Unix User UID on migrate Perhaps a query on the nodes for "next UID"?

INPUTS:

  • uid: Integer - the UID to unreserve within the district

  • district_uuid: String - district handle or identifier

NOTES:

  • method on the District object.

# File lib/openshift/mcollective_application_container_proxy.rb, line 306
def unreserve_uid(uid, district_uuid=nil)
  if Rails.configuration.msg_broker[:districts][:enabled]
    if @district
      district_uuid = @district.uuid
    elsif !district_uuid
      if @id
        begin
          district = District.find_by({"server_identities.name" => @id})
          district_uuid = district.uuid
        rescue Mongoid::Errors::DocumentNotFound
          district_uuid = 'NONE'
        end
      else
        district_uuid = get_district_uuid
      end
    end
    if district_uuid && district_uuid != 'NONE'
      #cleanup
      District::unreserve_uid(district_uuid, uid)
    end
  end
end
unset_user_env_vars(gear, env_vars, gears_ssh_endpoint) click to toggle source

Remove user environment variables from all gears in the app

INPUTS:

  • gear: a Gear object

  • env_vars: Array of environment variable names, e.g.:[{'name'=>'FOO'}, {'name'=>'BAR'}]

  • gears_ssh_endpoint: list of ssh gear endpoints

RETURNS:

  • String: stdout from a command

NOTES:

  • calls execute_direct

  • executes the 'user-var-remove' action on the node

# File lib/openshift/mcollective_application_container_proxy.rb, line 1353
def unset_user_env_vars(gear, env_vars, gears_ssh_endpoint)
  args = build_base_gear_args(gear)
  args['--with-keys'] = env_vars.map {|ev| ev['name']}.join(' ')
  args['--with-gears'] = gears_ssh_endpoint.join(';') if gears_ssh_endpoint.present?
  result = execute_direct(@@C_CONTROLLER, 'user-var-remove', args)
  parse_result(result)
end
update_cluster(gear, options) click to toggle source
# File lib/openshift/mcollective_application_container_proxy.rb, line 1744
def update_cluster(gear, options)
  args = build_base_gear_args(gear)
  args = build_update_cluster_args(options, args)
  result = execute_direct(@@C_CONTROLLER, 'update-cluster', args)
  parse_result(result)
end

Protected Instance Methods

component_status(gear, component) click to toggle source

Get the status a component service

INPUTS:

  • gear: a Gear object

  • component: String: a component name

RETURNS:

  • ResultIO?

NOTES

  • method on gear?

# File lib/openshift/mcollective_application_container_proxy.rb, line 2603
def component_status(gear, component)
  args = build_base_gear_args(gear)
  args = build_base_component_args(component, args)
  cart = component.cartridge_name

  run_cartridge_command(cart, gear, "status", args)
end
do_with_retry(action, num_tries=2) click to toggle source

Try some action until it passes or exceeds a maximum number of tries

INPUTS:

  • action: Block: a code block or method with no arguments

  • num_tries: Integer

RETURNS:

  • unknown: the result of the action

RAISES:

  • Exception

CATCHES:

  • Exception

NOTES:

# File lib/openshift/mcollective_application_container_proxy.rb, line 2452
def do_with_retry(action, num_tries=2)
  (1..num_tries).each do |i|
    begin
      yield
      if (i > 1)
        log_debug "DEBUG: Action '#{action}' succeeded on try #{i}.  You can ignore previous error messages or following mcollective debug related to '#{action}'"
      end
      break
    rescue Exception => e
      log_debug "DEBUG: Error performing #{action} on existing app on try #{i}: #{e.message}"
      raise if i == num_tries
    end
  end
end
embedded_carts(app=nil) click to toggle source

Initializes the list of cartridges which are "standalone" or framework

INPUTS:

RETURNS:

  • Array of String

SIDE EFFECTS:

NOTES:

  • Uses CartridgeCache

  • Why not just ask the CartridgeCache every time?

# File lib/openshift/mcollective_application_container_proxy.rb, line 2502
def embedded_carts(app=nil)
  @embedded_carts ||= CartridgeCache.cartridge_names('embedded',app)
end
execute_direct(cartridge, action, args, log_debug_output=true) click to toggle source

INPUTS:

  • cartridge: String, a cartridge name

  • action: String, and action name

  • args: Hash: command arguments

  • long_debug_output: Boolean

RETURNS: *

NOTES:

  • "cartridge_do" is a catch-all agent message handler

  • the real switches are the cartridge and action arguments

  • uses MCollective::RPC::Client

# File lib/openshift/mcollective_application_container_proxy.rb, line 2649
def execute_direct(cartridge, action, args, log_debug_output=true)
  if not args.has_key?('--cart-name')
    args['--cart-name'] = cartridge
  end

  mc_args = { :cartridge => cartridge,
              :action => action,
              :args => args }

  start_time = Time.now
  options = MCollectiveApplicationContainerProxy.rpc_options
  rpc_client = MCollectiveApplicationContainerProxy.get_rpc_client('openshift', options)
  result = nil
  begin
    Rails.logger.debug "DEBUG: rpc_client.custom_request('cartridge_do', #{mc_args.inspect}, #{@id}, {'identity' => #{@id}}) (Request ID: #{Thread.current[:user_action_log_uuid]})"
    result = rpc_client.custom_request('cartridge_do', mc_args, @id, {'identity' => @id})
    Rails.logger.debug "DEBUG: #{mask_user_creds(result.inspect)} (Request ID: #{Thread.current[:user_action_log_uuid]})" if log_debug_output
  rescue => e
    Rails.logger.error("Error processing custom_request for action #{action}: #{e.message}")
    Rails.logger.error(e.backtrace)
  ensure
    rpc_client.disconnect
  end
  Rails.logger.debug "DEBUG: MCollective Response Time (execute_direct: #{action}): #{Time.new - start_time}s  (Request ID: #{Thread.current[:user_action_log_uuid]})" if log_debug_output
  result
end
framework_carts(app=nil) click to toggle source

Initializes the list of cartridges which are "standalone" or framework

INPUTS:

RETURNS:

  • Array of String

SIDE EFFECTS:

NOTES:

  • uses CartridgeCache

  • why not just ask the CartidgeCache?

  • that is: Why use an instance var at all?

# File lib/openshift/mcollective_application_container_proxy.rb, line 2483
def framework_carts(app=nil)
  @framework_carts ||= CartridgeCache.cartridge_names('web_framework', app)
end
has_app_cartridge?(app_uuid, gear_uuid, cart) click to toggle source

Returns whether this gear has the specified cartridge

INPUTS:

  • gear_uuid: String

  • cartridge: String

RETURNS:

  • Boolean

NOTES:

# File lib/openshift/mcollective_application_container_proxy.rb, line 2842
def has_app_cartridge?(app_uuid, gear_uuid, cart)
  MCollectiveApplicationContainerProxy.rpc_exec('openshift', @id) do |client|
    client.has_app_cartridge(:app_uuid => app_uuid, :gear_uuid => gear_uuid, :cartridge => cart) do |response|
      output = response[:body][:data][:output]
      return output == true
    end
  end
end
has_embedded_app?(app_uuid, embedded_type) click to toggle source

Returns whether this server has the specified embedded app

INPUTS:

  • app_uuid: String

  • embedded_type: String

RETURNS:

  • Boolean

NOTES:

# File lib/openshift/mcollective_application_container_proxy.rb, line 2797
def has_embedded_app?(app_uuid, embedded_type)
  MCollectiveApplicationContainerProxy.rpc_exec('openshift', @id) do |client|
    client.has_embedded_app(:uuid => app_uuid,
                            :embedded_type => embedded_type) do |response|
      output = response[:body][:data][:output]
      return output == true
    end
  end
end
has_gear?(gear_uuid) click to toggle source

Returns whether this server has the specified app

INPUTS:

  • app_uuid: String

  • app_name: String

RETURNS:

  • Boolean

NOTES:

# File lib/openshift/mcollective_application_container_proxy.rb, line 2774
def has_gear?(gear_uuid)
  MCollectiveApplicationContainerProxy.rpc_exec('openshift', @id) do |client|
    client.has_gear(:uuid => gear_uuid) do |response|
      output = response[:body][:data][:output]
      return output == true
    end
  end
end
has_uid_or_gid?(uid) click to toggle source

Returns whether this server has already reserved the specified uid as a uid or gid

INPUTS:

  • uid: Integer

RETURNS:

  • Boolean

NOTES:

# File lib/openshift/mcollective_application_container_proxy.rb, line 2819
def has_uid_or_gid?(uid)
  return false if uid.nil?
  MCollectiveApplicationContainerProxy.rpc_exec('openshift', @id) do |client|
    client.has_uid_or_gid(:uid => uid.to_s) do |response|
      output = response[:body][:data][:output]
      return output == true
    end
  end
end
log_debug(message) click to toggle source

Wrap the log messages so it doesn't HAVE to be rails

INPUTS:

  • message: String

# File lib/openshift/mcollective_application_container_proxy.rb, line 2617
def log_debug(message)
  Rails.logger.debug message
  puts message
end
log_error(message) click to toggle source

Wrap the log messages so it doesn't HAVE to be rails

INPUTS:

  • message: String

# File lib/openshift/mcollective_application_container_proxy.rb, line 2628
def log_error(message)
  Rails.logger.error message
  puts message
end
parse_result(mcoll_reply, gear=nil, command=nil) click to toggle source

Cull wanted information out of an MCollective::Reply object INPUTS:

  • mcoll_reply: MCollective::RPC::Reply

  • gear: A Gear object

  • command: String

RETURNS:

  • Object: (Sanitized Result)

RAISES:

  • OpenShift::InvalidNodeException

  • OpenShift::NodeException

  • OpenShift::UserException

NOTES:

# File lib/openshift/mcollective_application_container_proxy.rb, line 2695
def parse_result(mcoll_reply, gear=nil, command=nil)
  app = gear.app unless gear.nil?
  result = ResultIO.new

  mcoll_result = mcoll_reply ? mcoll_reply[0] : nil
  output = nil
  addtl_params = nil
  if (mcoll_result && (defined? mcoll_result.results) && !mcoll_result.results[:data].nil?)
    output = mcoll_result.results[:data][:output]
    result.exitcode = mcoll_result.results[:data][:exitcode]
    if mcoll_result.results[:data][:addtl_params]
      result.deployments = mcoll_result.results[:data][:addtl_params][:deployments]
    end
  else
    server_identity = app ? MCollectiveApplicationContainerProxy.find_gear(gear.uuid) : nil
    if server_identity && @id != server_identity
      raise OpenShift::InvalidNodeException.new("Your gear was not on the expected server #{@id}.  A retry was attempted but the problem persisted.  The administrator may be moving gears, try your request again later.", 143, nil, server_identity)
    else
      raise OpenShift::NodeException.new("The server #{@id} that your application is running on failed to respond in time.  This may be due to a system restart.", 143)
    end
  end

  gear_id = gear.nil? ? nil : gear.uuid
  result.parse_output(output, gear_id)

  # raise an exception in case of non-zero exit code from the node
  if result.exitcode != 0
    result.debugIO << "Command return code: " + result.exitcode.to_s
    if result.hasUserActionableError
      raise OpenShift::UserException.new(result.errorIO.string, result.exitcode, nil, result)
    elsif result.exitcode == 146
      raise OpenShift::NodeException.new("Gear creation failed.  The server #{@id} that was chosen to run your gear has already exceeded its capacity.  Try your request again.", 146, result)
    else
      raise OpenShift::NodeException.new("An invalid exit code (#{result.exitcode.nil? ? "nil" : result.exitcode}) was returned from the server #{@id}.  This indicates an unexpected problem during the execution of your request.", 143, result)
    end
  end

  result
end
reload_component(gear, component) click to toggle source

Reload a component service

INPUTS:

  • gear: a Gear object

  • component: String: a component name

RETURNS:

  • ResultIO?

NOTES

  • method on gear?

# File lib/openshift/mcollective_application_container_proxy.rb, line 2582
def reload_component(gear, component)
  args = build_base_gear_args(gear)
  args = build_base_component_args(component, args)
  cart = component.cartridge_name

  run_cartridge_command(cart, gear, "reload", args)
end
restart_component(gear, component) click to toggle source

Restart a component service

INPUTS:

  • gear: a Gear object

  • component: String: a component name

RETURNS:

  • ResultIO?

NOTES

  • method on gear?

# File lib/openshift/mcollective_application_container_proxy.rb, line 2561
def restart_component(gear, component)
  args = build_base_gear_args(gear)
  args = build_base_component_args(component, args)
  cart = component.cartridge_name

  run_cartridge_command(cart, gear, "restart", args)
end
rpc_get_fact_direct(fact) click to toggle source

Given a known fact and node, get a single fact directly. This is significantly faster then the get_facts method If multiple nodes of the same name exist, it will pick just one

INPUTS:

  • fact: String

RETURNS:

  • String

RAISES:

  • OpenShift::NodeException

NOTES

# File lib/openshift/mcollective_application_container_proxy.rb, line 3255
def rpc_get_fact_direct(fact)
    options = MCollectiveApplicationContainerProxy.rpc_options
    options[:timeout] = Rails.configuration.msg_broker[:fact_timeout]

    rpc_client = MCollectiveApplicationContainerProxy.get_rpc_client('rpcutil', options)
    begin
      result = rpc_client.custom_request('get_fact', {:fact => fact}, @id, {'identity' => @id})[0]
      if (result && (defined? result.results) && result.results.has_key?(:data))
        value = result.results[:data][:value]
      else
        raise OpenShift::NodeException.new("Node execution failure (error getting fact).", 143)
      end
    ensure
      rpc_client.disconnect
    end

    return value
end
rpc_get_facts_direct(facts) click to toggle source

Given a list of facts, get the facts directly for instance's node. This is significantly faster then the get_facts method. If multiple nodes of the same name exist, it will pick just one.

INPUTS:

  • facts: Enumerable of Strings (fact names)

RETURNS:

  • Map of fact name to fact value for this instance's node

RAISES:

  • OpenShift::NodeException

NOTES

# File lib/openshift/mcollective_application_container_proxy.rb, line 3292
def rpc_get_facts_direct(facts)
    options = MCollectiveApplicationContainerProxy.rpc_options
    options[:timeout] = Rails.configuration.msg_broker[:fact_timeout]

    rpc_client = MCollectiveApplicationContainerProxy.get_rpc_client('openshift', options)
    begin
      result = rpc_client.custom_request('get_facts', {:facts => facts}, @id, {'identity' => @id})[0]
      if (result && (defined? result.results) && result.results.has_key?(:data))
        value = result.results[:data][:output]
      else
        raise OpenShift::NodeException.new("Node execution failure (error getting facts).", 143)
      end
    ensure
      rpc_client.disconnect
    end

    return value
end
rsuccess(response) click to toggle source

true if the response indicates success

INPUTS:

  • response

RETURNS:

  • Boolean

NOTES:

  • method on custom response object?

# File lib/openshift/mcollective_application_container_proxy.rb, line 3184
def rsuccess(response)
  response[:body][:statuscode].to_i == 0
end
run_cartridge_command(framework, gear, command, arguments, allow_move=true) click to toggle source

Execute a cartridge hook command in a gear

INPUTS:

  • framework:

  • gear: a Gear object

  • command: the hook command to run on the node?

  • arg: ??

  • allow_move: Boolean

RETURNS:

  • ResultIO

RAISES:

  • Exception

CATCHES:

  • OpenShift::InvalidNodeException

NOTES:

# File lib/openshift/mcollective_application_container_proxy.rb, line 2887
def run_cartridge_command(framework, gear, command, arguments, allow_move=true)
  app = gear.app
  resultIO = nil

  result = execute_direct(framework, command, arguments)

  begin
    begin
      resultIO = parse_result(result, gear, command)
    rescue OpenShift::InvalidNodeException => e
      if command != 'configure' && allow_move
        @id = e.server_identity
        Rails.logger.debug "DEBUG: Changing server identity of '#{gear.name}' from '#{gear.server_identity}' to '#{@id}'"
        dns_service = OpenShift::DnsService.instance
        dns_service.modify_application(gear.name, app.domain.namespace, get_public_hostname)
        dns_service.publish
        gear.server_identity = @id
        app.save
        #retry
        result = execute_direct(framework, command, arguments)
        resultIO = parse_result(result, gear, command)
      else
        raise
      end
    end
  end

  resultIO
end
run_cartridge_command_ignore_components(cart, gear, command, arguments, allow_move=true) click to toggle source

This method wraps run_cartridge_command to acknowledge and consistently support the behavior until cartridges and components are handled as distinct concepts within the runtime.

If the cart specified is in the framework_carts or embedded_carts list, the arguments will pass through to run_cartridge_command. Otherwise, a new ResultIO will be returned.

# File lib/openshift/mcollective_application_container_proxy.rb, line 2856
def run_cartridge_command_ignore_components(cart, gear, command, arguments, allow_move=true)
  if framework_carts(gear.app).include?(cart) || embedded_carts(gear.app).include?(cart)
    result = run_cartridge_command(cart, gear, command, arguments, allow_move)
  else
    result = ResultIO.new
  end
  result
end
start_component(gear, component) click to toggle source

Start a component service

INPUTS:

  • gear: a Gear object

  • component: String: a component name

RETURNS:

  • ResultIO?

NOTES

  • method on gear?

# File lib/openshift/mcollective_application_container_proxy.rb, line 2519
def start_component(gear, component)
  args = build_base_gear_args(gear)
  args = build_base_component_args(component, args)
  cart = component.cartridge_name

  run_cartridge_command(cart, gear, "start", args)
end
stop_component(gear, component) click to toggle source

Stop a component service

INPUTS:

  • gear: a Gear object

  • component: String: a component name

RETURNS:

  • ResultIO?

NOTES

  • method on gear?

# File lib/openshift/mcollective_application_container_proxy.rb, line 2540
def stop_component(gear, component)
  args = build_base_gear_args(gear)
  args = build_base_component_args(component, args)
  cart = component.cartridge_name

  run_cartridge_command(cart, gear, "stop", args)
end

[Validate]

Generated with the Darkfish Rdoc Generator 2.