virt-manager/virtManager/engine.py

546 lines
18 KiB
Python

# Copyright (C) 2006, 2013-2014 Red Hat, Inc.
# Copyright (C) 2006 Daniel P. Berrange <berrange@redhat.com>
#
# This work is licensed under the GNU GPLv2 or later.
# See the COPYING file in the top-level directory.
import logging
import os
import queue
import threading
import time
from gi.repository import Gio
from gi.repository import GLib
from gi.repository import Gtk
from .baseclass import vmmGObject
from .connect import vmmConnect
from .connmanager import vmmConnectionManager
from .inspection import vmmInspection
from .systray import vmmSystray
(PRIO_HIGH,
PRIO_LOW) = range(1, 3)
def _show_startup_error(fn):
"""
Decorator to show a modal error dialog if an exception is raised
from a startup routine
"""
# pylint: disable=protected-access
def newfn(self, *args, **kwargs):
try:
return fn(self, *args, **kwargs)
except Exception as e:
modal = self._can_exit()
self.err.show_err(str(e), modal=modal)
self._exit_app_if_no_windows()
return newfn
class vmmEngine(vmmGObject):
CLI_SHOW_MANAGER = "manager"
CLI_SHOW_DOMAIN_CREATOR = "creator"
CLI_SHOW_DOMAIN_EDITOR = "editor"
CLI_SHOW_DOMAIN_PERFORMANCE = "performance"
CLI_SHOW_DOMAIN_CONSOLE = "console"
CLI_SHOW_HOST_SUMMARY = "summary"
@classmethod
def get_instance(cls):
if not cls._instance:
cls._instance = vmmEngine()
return cls._instance
__gsignals__ = {
"app-closing": (vmmGObject.RUN_FIRST, None, []),
}
def __init__(self):
vmmGObject.__init__(self)
self._exiting = False
self._window_count = 0
self._gtkapplication = None
self._init_gtk_application()
self._timer = None
self._tick_counter = 0
self._tick_thread_slow = False
self._tick_thread = threading.Thread(name="Tick thread",
target=self._handle_tick_queue,
args=())
self._tick_thread.daemon = True
self._tick_queue = queue.PriorityQueue(100)
@property
def _connobjs(self):
return vmmConnectionManager.get_instance().conns
def _cleanup(self):
if self._timer is not None:
GLib.source_remove(self._timer)
#################
# init handling #
#################
def _default_startup(self, skip_autostart, cliuri):
"""
Actual startup routines if we are running a new instance of the app
"""
vmmSystray.get_instance()
vmmInspection.get_instance()
self.add_gsettings_handle(
self.config.on_stats_update_interval_changed(
self._timer_changed_cb))
self._schedule_timer()
self._tick_thread.start()
self._tick()
uris = list(self._connobjs.keys())
if not uris:
logging.debug("No stored URIs found.")
else:
logging.debug("Loading stored URIs:\n%s",
" \n".join(sorted(uris)))
if not skip_autostart:
self.idle_add(self._autostart_conns)
if not self.config.get_conn_uris() and not cliuri:
# Only add default if no connections are currently known
manager = self._get_manager()
manager.set_startup_error(
_("Checking for virtualization packages..."))
self.timeout_add(1000, self._add_default_conn)
def _add_default_conn(self):
"""
If there's no cached connections, or any requested on the command
line, try to determine a default URI and open it, first checking
if libvirt is running
"""
manager = self._get_manager()
logging.debug("Trying to start libvirtd through systemd")
unitname = "libvirtd.service"
libvirtd_installed = False
libvirtd_active = False
# Fetch all units from systemd
try:
bus = Gio.bus_get_sync(Gio.BusType.SYSTEM, None)
systemd = Gio.DBusProxy.new_sync(bus, 0, None,
"org.freedesktop.systemd1",
"/org/freedesktop/systemd1",
"org.freedesktop.systemd1.Manager", None)
units = systemd.ListUnits()
except Exception:
units = []
logging.exception("Couldn't connect to systemd")
libvirtd_installed = os.path.exists("/var/run/libvirt")
libvirtd_active = os.path.exists("/var/run/libvirt/libvirt-sock")
# Check if libvirtd is installed and running
for unitinfo in units:
if unitinfo[0] != unitname:
continue
libvirtd_installed = True
libvirtd_active = unitinfo[3] == "active"
unitpath = unitinfo[6]
break
# If it's not running, try to start it
try:
if units and libvirtd_installed and not libvirtd_active:
unit = Gio.DBusProxy.new_sync(
bus, 0, None,
"org.freedesktop.systemd1", unitpath,
"org.freedesktop.systemd1.Unit", None)
if not self.config.test_first_run:
unit.Start("(s)", "fail")
time.sleep(2)
libvirtd_active = True
except Exception:
logging.exception("Error starting libvirtd")
# Manager fail message
tryuri = None
if not self.config.test_first_run:
tryuri = vmmConnect.default_uri()
logging.debug("Probed default URI=%s", tryuri)
msg = ""
if not libvirtd_installed:
msg += _("The libvirtd service does not appear to be installed. "
"Install and run the libvirtd service to manage "
"virtualization on this host.")
elif not libvirtd_active:
msg += _("libvirtd is installed but not running. Start the "
"libvirtd service to manage virtualization on this host.")
if not tryuri or "qemu" not in tryuri:
if msg:
msg += "\n\n"
msg += _("Could not detect a default hypervisor. Make "
"sure the appropriate qemu/kvm virtualization "
"packages are installed to manage virtualization "
"on this host.")
if msg:
msg += "\n\n"
msg += _("A virtualization connection can be manually "
"added via File->Add Connection")
if (tryuri is None or
not libvirtd_installed or
not libvirtd_active):
manager.set_startup_error(msg)
return
# Launch idle callback to connect to default URI
def idle_connect():
def _open_completed(c, ConnectError):
if ConnectError:
self._handle_conn_error(c, ConnectError)
conn = vmmConnectionManager.get_instance().add_conn(tryuri)
conn.set_autoconnect(True)
conn.connect_once("open-completed", _open_completed)
conn.open()
self.idle_add(idle_connect)
def _autostart_conns(self):
"""
We serialize conn autostart, so polkit/ssh-askpass doesn't spam
"""
if self._exiting:
return
connections_queue = queue.Queue()
auto_conns = [conn.get_uri() for conn in self._connobjs.values() if
conn.get_autoconnect()]
def add_next_to_queue():
if not auto_conns:
connections_queue.put(None)
else:
connections_queue.put(auto_conns.pop(0))
def conn_open_completed(_conn, ConnectError):
# Explicitly ignore connection errors, we've done that
# for a while and it can be noisy
if ConnectError is not None:
logging.debug("Autostart connection error: %s",
ConnectError.details)
add_next_to_queue()
def handle_queue():
while True:
uri = connections_queue.get()
if uri is None:
return
if self._exiting:
return
if uri not in self._connobjs:
add_next_to_queue()
continue
conn = self._connobjs[uri]
conn.connect_once("open-completed", conn_open_completed)
self.idle_add(conn.open)
add_next_to_queue()
self._start_thread(handle_queue, "Conn autostart thread")
############################
# Gtk Application handling #
############################
def _on_gtk_application_activated(self, ignore):
"""
Invoked after application.run()
"""
if not self._application.get_windows():
logging.debug("Initial gtkapplication activated")
self._application.add_window(Gtk.Window())
def _init_gtk_application(self):
self._application = Gtk.Application(
application_id="org.virt-manager.virt-manager", flags=0)
self._application.register(None)
self._application.connect("activate",
self._on_gtk_application_activated)
action = Gio.SimpleAction.new("cli_command",
GLib.VariantType.new("(sss)"))
action.connect("activate", self._handle_cli_command)
self._application.add_action(action)
def start(self, uri, show_window, domain, skip_autostart):
"""
Public entrypoint from virt-manager cli. If app is already
running, connect to it and exit, otherwise run our functional
default startup.
"""
# Dispatch dbus CLI command
if uri and not show_window:
show_window = self.CLI_SHOW_MANAGER
data = GLib.Variant("(sss)",
(uri or "", show_window or "", domain or ""))
is_remote = self._application.get_is_remote()
if not is_remote:
self._default_startup(skip_autostart, uri)
self._application.activate_action("cli_command", data)
if is_remote:
logging.debug("Connected to remote app instance.")
return
self._application.run(None)
###########################
# timer and tick handling #
###########################
def _timer_changed_cb(self, *args, **kwargs):
ignore1 = args
ignore2 = kwargs
self._schedule_timer()
def _schedule_timer(self):
interval = self.config.get_stats_update_interval() * 1000
if self._timer is not None:
self.remove_gobject_timeout(self._timer)
self._timer = None
self._timer = self.timeout_add(interval, self._tick)
def _add_obj_to_tick_queue(self, obj, isprio, **kwargs):
if self._tick_queue.full():
if not self._tick_thread_slow:
logging.debug("Tick is slow, not running at requested rate.")
self._tick_thread_slow = True
return
self._tick_counter += 1
self._tick_queue.put((isprio and PRIO_HIGH or PRIO_LOW,
self._tick_counter,
obj, kwargs))
def schedule_priority_tick(self, conn, kwargs):
# Called directly from connection
self._add_obj_to_tick_queue(conn, True, **kwargs)
def _tick(self):
for conn in self._connobjs.values():
self._add_obj_to_tick_queue(conn, False,
stats_update=True, pollvm=True)
return 1
def _handle_tick_queue(self):
while True:
ignore1, ignore2, conn, kwargs = self._tick_queue.get()
try:
conn.tick_from_engine(**kwargs)
except Exception:
# Don't attempt to show any UI error here, since it
# can cause dialogs to appear from nowhere if say
# libvirtd is shut down
logging.debug("Error polling connection %s",
conn.get_uri(), exc_info=True)
# Need to clear reference to make leak check happy
conn = None
self._tick_queue.task_done()
return 1
#####################################
# window counting and exit handling #
#####################################
def increment_window_counter(self):
"""
Public function, called by toplevel windows
"""
self._window_count += 1
logging.debug("window counter incremented to %s", self._window_count)
def decrement_window_counter(self):
"""
Public function, called by toplevel windows
"""
self._window_count -= 1
logging.debug("window counter decremented to %s", self._window_count)
self._exit_app_if_no_windows()
def _systray_is_embedded(self):
"""
We don't use window tracking here: systray isn't a window and even
when 'show' has been requested it may not be embedded in a visible
tray area, so we have to check it separately.
"""
return vmmSystray.get_instance().is_embedded()
def _can_exit(self):
return (self._window_count <= 0 and not
self._systray_is_embedded())
def _exit_app_if_no_windows(self):
if self._exiting:
return
if self._can_exit():
logging.debug("No windows found, requesting app exit")
self.exit_app()
def exit_app(self):
"""
Public call, manager/details/... use this to force exit the app
"""
if self._exiting:
return
self._exiting = True
def _do_exit():
try:
vmmConnectionManager.get_instance().cleanup()
self.emit("app-closing")
self.cleanup()
if self.config.test_leak_debug:
objs = self.config.get_objects()
# Engine will always appear to leak
objs.remove(self.object_key)
for name in objs:
logging.debug("LEAK: %s", name)
logging.debug("Exiting app normally.")
finally:
self._application.quit()
# We stick this in an idle callback, so the exit_app() caller
# reference is dropped, and leak check debug doesn't give a
# false positive
self.idle_add(_do_exit)
##########################################
# Window launchers from virt-manager cli #
##########################################
def _find_vm_by_cli_str(self, uri, clistr):
"""
Lookup a VM by a string passed in on the CLI. Can be either
ID, domain name, or UUID
"""
if clistr.isdigit():
clistr = int(clistr)
for vm in self._connobjs[uri].list_vms():
if clistr == vm.get_id():
return vm
elif clistr == vm.get_name():
return vm
elif clistr == vm.get_uuid():
return vm
def _cli_show_vm_helper(self, uri, clistr, page):
vm = self._find_vm_by_cli_str(uri, clistr)
if not vm:
raise RuntimeError("%s does not have VM '%s'" %
(uri, clistr))
from .details import vmmDetails
details = vmmDetails.get_instance(None, vm)
if page == self.CLI_SHOW_DOMAIN_PERFORMANCE:
details.activate_performance_page()
elif page == self.CLI_SHOW_DOMAIN_EDITOR:
details.activate_config_page()
elif page == self.CLI_SHOW_DOMAIN_CONSOLE:
details.activate_console_page()
details.show()
def _get_manager(self):
from .manager import vmmManager
return vmmManager.get_instance(None)
@_show_startup_error
def _launch_cli_window(self, uri, show_window, clistr):
logging.debug("Launching requested window '%s'", show_window)
if show_window == self.CLI_SHOW_MANAGER:
manager = self._get_manager()
manager.set_initial_selection(uri)
manager.show()
elif show_window == self.CLI_SHOW_DOMAIN_CREATOR:
from .create import vmmCreate
vmmCreate.show_instance(None, uri)
elif show_window == self.CLI_SHOW_HOST_SUMMARY:
from .host import vmmHost
vmmHost.show_instance(None, self._connobjs[uri])
elif (show_window in [self.CLI_SHOW_DOMAIN_EDITOR,
self.CLI_SHOW_DOMAIN_PERFORMANCE,
self.CLI_SHOW_DOMAIN_CONSOLE]):
self._cli_show_vm_helper(uri, clistr, show_window)
else:
raise RuntimeError("Unknown cli window command '%s'" %
show_window)
def _handle_conn_error(self, _conn, ConnectError):
msg, details, title = ConnectError
modal = self._can_exit()
self.err.show_err(msg, details, title, modal=modal)
self._exit_app_if_no_windows()
@_show_startup_error
def _handle_cli_command(self, actionobj, variant):
ignore = actionobj
uri = variant[0]
show_window = variant[1] or self.CLI_SHOW_MANAGER
domain = variant[2]
logging.debug("processing cli command uri=%s show_window=%s domain=%s",
uri, show_window, domain)
if not uri:
logging.debug("No cli action requested, launching default window")
self._get_manager().show()
return
conn_is_new = uri not in self._connobjs
conn = vmmConnectionManager.get_instance().add_conn(uri)
if conn.is_active():
self.idle_add(self._launch_cli_window,
uri, show_window, domain)
return
def _open_completed(_c, ConnectError):
if ConnectError:
if conn_is_new:
logging.debug("Removing failed uri=%s", uri)
vmmConnectionManager.get_instance().remove_conn(uri)
self._handle_conn_error(conn, ConnectError)
else:
self._launch_cli_window(uri, show_window, domain)
conn.connect_once("open-completed", _open_completed)
conn.open()