| #!/usr/bin/env python2 |
| # pylint: disable=missing-docstring |
| |
| import datetime |
| import mox |
| import unittest |
| |
| import common |
| from autotest_lib.client.common_lib import control_data |
| from autotest_lib.client.common_lib import error |
| from autotest_lib.client.common_lib import global_config |
| from autotest_lib.client.common_lib import priorities |
| from autotest_lib.client.common_lib.cros import dev_server |
| from autotest_lib.client.common_lib.test_utils import mock |
| from autotest_lib.frontend import setup_django_environment |
| from autotest_lib.frontend.afe import frontend_test_utils |
| from autotest_lib.frontend.afe import model_logic |
| from autotest_lib.frontend.afe import models |
| from autotest_lib.frontend.afe import rpc_interface |
| from autotest_lib.frontend.afe import rpc_utils |
| from autotest_lib.server import frontend |
| from autotest_lib.server import utils as server_utils |
| from autotest_lib.server.cros import provision |
| from autotest_lib.server.cros.dynamic_suite import constants |
| from autotest_lib.server.cros.dynamic_suite import control_file_getter |
| from autotest_lib.server.cros.dynamic_suite import frontend_wrappers |
| from autotest_lib.server.cros.dynamic_suite import suite_common |
| |
| |
| CLIENT = control_data.CONTROL_TYPE_NAMES.CLIENT |
| SERVER = control_data.CONTROL_TYPE_NAMES.SERVER |
| |
| _hqe_status = models.HostQueueEntry.Status |
| |
| |
| class ShardHeartbeatTest(mox.MoxTestBase, unittest.TestCase): |
| |
| _PRIORITY = priorities.Priority.DEFAULT |
| |
| |
| def _do_heartbeat_and_assert_response(self, shard_hostname='shard1', |
| upload_jobs=(), upload_hqes=(), |
| known_jobs=(), known_hosts=(), |
| **kwargs): |
| known_job_ids = [job.id for job in known_jobs] |
| known_host_ids = [host.id for host in known_hosts] |
| known_host_statuses = [host.status for host in known_hosts] |
| |
| retval = rpc_interface.shard_heartbeat( |
| shard_hostname=shard_hostname, |
| jobs=upload_jobs, hqes=upload_hqes, |
| known_job_ids=known_job_ids, known_host_ids=known_host_ids, |
| known_host_statuses=known_host_statuses) |
| |
| self._assert_shard_heartbeat_response(shard_hostname, retval, |
| **kwargs) |
| |
| return shard_hostname |
| |
| |
| def _assert_shard_heartbeat_response(self, shard_hostname, retval, jobs=[], |
| hosts=[], hqes=[], |
| incorrect_host_ids=[]): |
| |
| retval_hosts, retval_jobs = retval['hosts'], retval['jobs'] |
| retval_incorrect_hosts = retval['incorrect_host_ids'] |
| |
| expected_jobs = [ |
| (job.id, job.name, shard_hostname) for job in jobs] |
| returned_jobs = [(job['id'], job['name'], job['shard']['hostname']) |
| for job in retval_jobs] |
| self.assertEqual(returned_jobs, expected_jobs) |
| |
| expected_hosts = [(host.id, host.hostname) for host in hosts] |
| returned_hosts = [(host['id'], host['hostname']) |
| for host in retval_hosts] |
| self.assertEqual(returned_hosts, expected_hosts) |
| |
| retval_hqes = [] |
| for job in retval_jobs: |
| retval_hqes += job['hostqueueentry_set'] |
| |
| expected_hqes = [(hqe.id) for hqe in hqes] |
| returned_hqes = [(hqe['id']) for hqe in retval_hqes] |
| self.assertEqual(returned_hqes, expected_hqes) |
| |
| self.assertEqual(retval_incorrect_hosts, incorrect_host_ids) |
| |
| |
| def _createJobForLabel(self, label): |
| job_id = rpc_interface.create_job(name='dummy', priority=self._PRIORITY, |
| control_file='foo', |
| control_type=CLIENT, |
| meta_hosts=[label.name], |
| dependencies=(label.name,)) |
| return models.Job.objects.get(id=job_id) |
| |
| |
| def _testShardHeartbeatFetchHostlessJobHelper(self, host1): |
| """Create a hostless job and ensure it's not assigned to a shard.""" |
| label2 = models.Label.objects.create(name='bluetooth', platform=False) |
| |
| job1 = self._create_job(hostless=True) |
| |
| # Hostless jobs should be executed by the global scheduler. |
| self._do_heartbeat_and_assert_response(hosts=[host1]) |
| |
| |
| def _testShardHeartbeatIncorrectHostsHelper(self, host1): |
| """Ensure that hosts that don't belong to shard are determined.""" |
| host2 = models.Host.objects.create(hostname='test_host2', leased=False) |
| |
| # host2 should not belong to shard1. Ensure that if shard1 thinks host2 |
| # is a known host, then it is returned as invalid. |
| self._do_heartbeat_and_assert_response(known_hosts=[host1, host2], |
| incorrect_host_ids=[host2.id]) |
| |
| |
| def _testShardHeartbeatLabelRemovalRaceHelper(self, shard1, host1, label1): |
| """Ensure correctness if label removed during heartbeat.""" |
| host2 = models.Host.objects.create(hostname='test_host2', leased=False) |
| host2.labels.add(label1) |
| self.assertEqual(host2.shard, None) |
| |
| # In the middle of the assign_to_shard call, remove label1 from shard1. |
| self.mox.StubOutWithMock(models.Host, '_assign_to_shard_nothing_helper') |
| def remove_label(): |
| rpc_interface.remove_board_from_shard(shard1.hostname, label1.name) |
| |
| models.Host._assign_to_shard_nothing_helper().WithSideEffects( |
| remove_label) |
| self.mox.ReplayAll() |
| |
| self._do_heartbeat_and_assert_response( |
| known_hosts=[host1], hosts=[], incorrect_host_ids=[host1.id]) |
| host2 = models.Host.smart_get(host2.id) |
| self.assertEqual(host2.shard, None) |
| |
| |
| def _testShardRetrieveJobsHelper(self, shard1, host1, label1, shard2, |
| host2, label2): |
| """Create jobs and retrieve them.""" |
| # should never be returned by heartbeat |
| leased_host = models.Host.objects.create(hostname='leased_host', |
| leased=True) |
| |
| leased_host.labels.add(label1) |
| |
| job1 = self._createJobForLabel(label1) |
| |
| job2 = self._createJobForLabel(label2) |
| |
| job_completed = self._createJobForLabel(label1) |
| # Job is already being run, so don't sync it |
| job_completed.hostqueueentry_set.update(complete=True) |
| job_completed.hostqueueentry_set.create(complete=False) |
| |
| job_active = self._createJobForLabel(label1) |
| # Job is already started, so don't sync it |
| job_active.hostqueueentry_set.update(active=True) |
| job_active.hostqueueentry_set.create(complete=False, active=False) |
| |
| self._do_heartbeat_and_assert_response( |
| jobs=[job1], hosts=[host1], hqes=job1.hostqueueentry_set.all()) |
| |
| self._do_heartbeat_and_assert_response( |
| shard_hostname=shard2.hostname, |
| jobs=[job2], hosts=[host2], hqes=job2.hostqueueentry_set.all()) |
| |
| host3 = models.Host.objects.create(hostname='test_host3', leased=False) |
| host3.labels.add(label1) |
| |
| self._do_heartbeat_and_assert_response( |
| known_jobs=[job1], known_hosts=[host1], hosts=[host3]) |
| |
| |
| def _testResendJobsAfterFailedHeartbeatHelper(self, shard1, host1, label1): |
| """Create jobs, retrieve them, fail on client, fetch them again.""" |
| job1 = self._createJobForLabel(label1) |
| |
| self._do_heartbeat_and_assert_response( |
| jobs=[job1], |
| hqes=job1.hostqueueentry_set.all(), hosts=[host1]) |
| |
| # Make sure it's resubmitted by sending last_job=None again |
| self._do_heartbeat_and_assert_response( |
| known_hosts=[host1], |
| jobs=[job1], hqes=job1.hostqueueentry_set.all(), hosts=[]) |
| |
| # Now it worked, make sure it's not sent again |
| self._do_heartbeat_and_assert_response( |
| known_jobs=[job1], known_hosts=[host1]) |
| |
| job1 = models.Job.objects.get(pk=job1.id) |
| job1.hostqueueentry_set.all().update(complete=True) |
| |
| # Job is completed, make sure it's not sent again |
| self._do_heartbeat_and_assert_response( |
| known_hosts=[host1]) |
| |
| job2 = self._createJobForLabel(label1) |
| |
| # job2's creation was later, it should be returned now. |
| self._do_heartbeat_and_assert_response( |
| known_hosts=[host1], |
| jobs=[job2], hqes=job2.hostqueueentry_set.all()) |
| |
| self._do_heartbeat_and_assert_response( |
| known_jobs=[job2], known_hosts=[host1]) |
| |
| job2 = models.Job.objects.get(pk=job2.pk) |
| job2.hostqueueentry_set.update(aborted=True) |
| # Setting a job to a complete status will set the shard_id to None in |
| # scheduler_models. We have to emulate that here, because we use Django |
| # models in tests. |
| job2.shard = None |
| job2.save() |
| |
| self._do_heartbeat_and_assert_response( |
| known_jobs=[job2], known_hosts=[host1], |
| jobs=[job2], |
| hqes=job2.hostqueueentry_set.all()) |
| |
| models.Test.objects.create(name='platform_BootPerfServer:shard', |
| test_type=1) |
| self.mox.StubOutWithMock(server_utils, 'read_file') |
| self.mox.ReplayAll() |
| rpc_interface.delete_shard(hostname=shard1.hostname) |
| |
| self.assertRaises( |
| models.Shard.DoesNotExist, models.Shard.objects.get, pk=shard1.id) |
| |
| job1 = models.Job.objects.get(pk=job1.id) |
| label1 = models.Label.objects.get(pk=label1.id) |
| |
| self.assertIsNone(job1.shard) |
| self.assertEqual(len(label1.shard_set.all()), 0) |
| |
| |
| def _testResendHostsAfterFailedHeartbeatHelper(self, host1): |
| """Check that master accepts resending updated records after failure.""" |
| # Send the host |
| self._do_heartbeat_and_assert_response(hosts=[host1]) |
| |
| # Send it again because previous one didn't persist correctly |
| self._do_heartbeat_and_assert_response(hosts=[host1]) |
| |
| # Now it worked, make sure it isn't sent again |
| self._do_heartbeat_and_assert_response(known_hosts=[host1]) |
| |
| |
| class RpcInterfaceTestWithStaticAttribute( |
| mox.MoxTestBase, unittest.TestCase, |
| frontend_test_utils.FrontendTestMixin): |
| |
| def setUp(self): |
| super(RpcInterfaceTestWithStaticAttribute, self).setUp() |
| self._frontend_common_setup() |
| self.god = mock.mock_god() |
| self.old_respect_static_config = rpc_interface.RESPECT_STATIC_ATTRIBUTES |
| rpc_interface.RESPECT_STATIC_ATTRIBUTES = True |
| models.RESPECT_STATIC_ATTRIBUTES = True |
| |
| |
| def tearDown(self): |
| self.god.unstub_all() |
| self._frontend_common_teardown() |
| global_config.global_config.reset_config_values() |
| rpc_interface.RESPECT_STATIC_ATTRIBUTES = self.old_respect_static_config |
| models.RESPECT_STATIC_ATTRIBUTES = self.old_respect_static_config |
| |
| |
| def _fake_host_with_static_attributes(self): |
| host1 = models.Host.objects.create(hostname='test_host') |
| host1.set_attribute('test_attribute1', 'test_value1') |
| host1.set_attribute('test_attribute2', 'test_value2') |
| self._set_static_attribute(host1, 'test_attribute1', 'static_value1') |
| self._set_static_attribute(host1, 'static_attribute1', 'static_value2') |
| host1.save() |
| return host1 |
| |
| |
| def test_get_hosts(self): |
| host1 = self._fake_host_with_static_attributes() |
| hosts = rpc_interface.get_hosts(hostname=host1.hostname) |
| host = hosts[0] |
| |
| self.assertEquals(host['hostname'], 'test_host') |
| self.assertEquals(host['acls'], ['Everyone']) |
| # Respect the value of static attributes. |
| self.assertEquals(host['attributes'], |
| {'test_attribute1': 'static_value1', |
| 'test_attribute2': 'test_value2', |
| 'static_attribute1': 'static_value2'}) |
| |
| def test_get_host_attribute_with_static(self): |
| host1 = models.Host.objects.create(hostname='test_host1') |
| host1.set_attribute('test_attribute1', 'test_value1') |
| self._set_static_attribute(host1, 'test_attribute1', 'static_value1') |
| host2 = models.Host.objects.create(hostname='test_host2') |
| host2.set_attribute('test_attribute1', 'test_value1') |
| host2.set_attribute('test_attribute2', 'test_value2') |
| |
| attributes = rpc_interface.get_host_attribute( |
| 'test_attribute1', |
| hostname__in=['test_host1', 'test_host2']) |
| hosts = [attr['host'] for attr in attributes] |
| values = [attr['value'] for attr in attributes] |
| self.assertEquals(set(hosts), |
| set(['test_host1', 'test_host2'])) |
| self.assertEquals(set(values), |
| set(['test_value1', 'static_value1'])) |
| |
| |
| def test_get_hosts_by_attribute_without_static(self): |
| host1 = models.Host.objects.create(hostname='test_host1') |
| host1.set_attribute('test_attribute1', 'test_value1') |
| host2 = models.Host.objects.create(hostname='test_host2') |
| host2.set_attribute('test_attribute1', 'test_value1') |
| |
| hosts = rpc_interface.get_hosts_by_attribute( |
| 'test_attribute1', 'test_value1') |
| self.assertEquals(set(hosts), |
| set(['test_host1', 'test_host2'])) |
| |
| |
| def test_get_hosts_by_attribute_with_static(self): |
| host1 = models.Host.objects.create(hostname='test_host1') |
| host1.set_attribute('test_attribute1', 'test_value1') |
| self._set_static_attribute(host1, 'test_attribute1', 'test_value1') |
| host2 = models.Host.objects.create(hostname='test_host2') |
| host2.set_attribute('test_attribute1', 'test_value1') |
| self._set_static_attribute(host2, 'test_attribute1', 'static_value1') |
| host3 = models.Host.objects.create(hostname='test_host3') |
| self._set_static_attribute(host3, 'test_attribute1', 'test_value1') |
| host4 = models.Host.objects.create(hostname='test_host4') |
| host4.set_attribute('test_attribute1', 'test_value1') |
| host5 = models.Host.objects.create(hostname='test_host5') |
| host5.set_attribute('test_attribute1', 'temp_value1') |
| self._set_static_attribute(host5, 'test_attribute1', 'test_value1') |
| |
| hosts = rpc_interface.get_hosts_by_attribute( |
| 'test_attribute1', 'test_value1') |
| # host1: matched, it has the same value for test_attribute1. |
| # host2: not matched, it has a new value in |
| # afe_static_host_attributes for test_attribute1. |
| # host3: matched, it has a corresponding entry in |
| # afe_host_attributes for test_attribute1. |
| # host4: matched, test_attribute1 is not replaced by static |
| # attribute. |
| # host5: matched, it has an updated & matched value for |
| # test_attribute1 in afe_static_host_attributes. |
| self.assertEquals(set(hosts), |
| set(['test_host1', 'test_host3', |
| 'test_host4', 'test_host5'])) |
| |
| |
| class RpcInterfaceTestWithStaticLabel(ShardHeartbeatTest, |
| frontend_test_utils.FrontendTestMixin): |
| |
| _STATIC_LABELS = ['board:lumpy'] |
| |
| def setUp(self): |
| super(RpcInterfaceTestWithStaticLabel, self).setUp() |
| self._frontend_common_setup() |
| self.god = mock.mock_god() |
| self.old_respect_static_config = rpc_interface.RESPECT_STATIC_LABELS |
| rpc_interface.RESPECT_STATIC_LABELS = True |
| models.RESPECT_STATIC_LABELS = True |
| |
| |
| def tearDown(self): |
| self.god.unstub_all() |
| self._frontend_common_teardown() |
| global_config.global_config.reset_config_values() |
| rpc_interface.RESPECT_STATIC_LABELS = self.old_respect_static_config |
| models.RESPECT_STATIC_LABELS = self.old_respect_static_config |
| |
| |
| def _fake_host_with_static_labels(self): |
| host1 = models.Host.objects.create(hostname='test_host') |
| label1 = models.Label.objects.create( |
| name='non_static_label1', platform=False) |
| non_static_platform = models.Label.objects.create( |
| name='static_platform', platform=False) |
| static_platform = models.StaticLabel.objects.create( |
| name='static_platform', platform=True) |
| models.ReplacedLabel.objects.create(label_id=non_static_platform.id) |
| host1.static_labels.add(static_platform) |
| host1.labels.add(non_static_platform) |
| host1.labels.add(label1) |
| host1.save() |
| return host1 |
| |
| |
| def test_get_hosts(self): |
| host1 = self._fake_host_with_static_labels() |
| hosts = rpc_interface.get_hosts(hostname=host1.hostname) |
| host = hosts[0] |
| |
| self.assertEquals(host['hostname'], 'test_host') |
| self.assertEquals(host['acls'], ['Everyone']) |
| # Respect all labels in afe_hosts_labels. |
| self.assertEquals(host['labels'], |
| ['non_static_label1', 'static_platform']) |
| # Respect static labels. |
| self.assertEquals(host['platform'], 'static_platform') |
| |
| |
| def test_get_hosts_multiple_labels(self): |
| self._fake_host_with_static_labels() |
| hosts = rpc_interface.get_hosts( |
| multiple_labels=['non_static_label1', 'static_platform']) |
| host = hosts[0] |
| self.assertEquals(host['hostname'], 'test_host') |
| |
| |
| def test_delete_static_label(self): |
| label1 = models.Label.smart_get('static') |
| |
| host2 = models.Host.objects.all()[1] |
| shard1 = models.Shard.objects.create(hostname='shard1') |
| host2.shard = shard1 |
| host2.labels.add(label1) |
| host2.save() |
| |
| mock_afe = self.god.create_mock_class_obj(frontend_wrappers.RetryingAFE, |
| 'MockAFE') |
| self.god.stub_with(frontend_wrappers, 'RetryingAFE', mock_afe) |
| |
| self.assertRaises(error.UnmodifiableLabelException, |
| rpc_interface.delete_label, |
| label1.id) |
| |
| self.god.check_playback() |
| |
| |
| def test_modify_static_label(self): |
| label1 = models.Label.smart_get('static') |
| self.assertEqual(label1.invalid, 0) |
| |
| host2 = models.Host.objects.all()[1] |
| shard1 = models.Shard.objects.create(hostname='shard1') |
| host2.shard = shard1 |
| host2.labels.add(label1) |
| host2.save() |
| |
| mock_afe = self.god.create_mock_class_obj(frontend_wrappers.RetryingAFE, |
| 'MockAFE') |
| self.god.stub_with(frontend_wrappers, 'RetryingAFE', mock_afe) |
| |
| self.assertRaises(error.UnmodifiableLabelException, |
| rpc_interface.modify_label, |
| label1.id, |
| invalid=1) |
| |
| self.assertEqual(models.Label.smart_get('static').invalid, 0) |
| self.god.check_playback() |
| |
| |
| def test_multiple_platforms_add_non_static_to_static(self): |
| """Test non-static platform to a host with static platform.""" |
| static_platform = models.StaticLabel.objects.create( |
| name='static_platform', platform=True) |
| non_static_platform = models.Label.objects.create( |
| name='static_platform', platform=True) |
| models.ReplacedLabel.objects.create(label_id=non_static_platform.id) |
| platform2 = models.Label.objects.create(name='platform2', platform=True) |
| host1 = models.Host.objects.create(hostname='test_host') |
| host1.static_labels.add(static_platform) |
| host1.labels.add(non_static_platform) |
| host1.save() |
| |
| self.assertRaises(model_logic.ValidationError, |
| rpc_interface.label_add_hosts, id='platform2', |
| hosts=['test_host']) |
| self.assertRaises(model_logic.ValidationError, |
| rpc_interface.host_add_labels, |
| id='test_host', labels=['platform2']) |
| # make sure the platform didn't get added |
| platforms = rpc_interface.get_labels( |
| host__hostname__in=['test_host'], platform=True) |
| self.assertEquals(len(platforms), 1) |
| |
| |
| def test_multiple_platforms_add_static_to_non_static(self): |
| """Test static platform to a host with non-static platform.""" |
| platform1 = models.Label.objects.create( |
| name='static_platform', platform=True) |
| models.ReplacedLabel.objects.create(label_id=platform1.id) |
| static_platform = models.StaticLabel.objects.create( |
| name='static_platform', platform=True) |
| platform2 = models.Label.objects.create( |
| name='platform2', platform=True) |
| |
| host1 = models.Host.objects.create(hostname='test_host') |
| host1.labels.add(platform2) |
| host1.save() |
| |
| self.assertRaises(model_logic.ValidationError, |
| rpc_interface.label_add_hosts, |
| id='static_platform', |
| hosts=['test_host']) |
| self.assertRaises(model_logic.ValidationError, |
| rpc_interface.host_add_labels, |
| id='test_host', labels=['static_platform']) |
| # make sure the platform didn't get added |
| platforms = rpc_interface.get_labels( |
| host__hostname__in=['test_host'], platform=True) |
| self.assertEquals(len(platforms), 1) |
| |
| |
| def test_label_remove_hosts(self): |
| """Test remove a label of hosts.""" |
| label = models.Label.smart_get('static') |
| static_label = models.StaticLabel.objects.create(name='static') |
| |
| host1 = models.Host.objects.create(hostname='test_host') |
| host1.labels.add(label) |
| host1.static_labels.add(static_label) |
| host1.save() |
| |
| self.assertRaises(error.UnmodifiableLabelException, |
| rpc_interface.label_remove_hosts, |
| id='static', hosts=['test_host']) |
| |
| |
| def test_host_remove_labels(self): |
| """Test remove labels of a given host.""" |
| label = models.Label.smart_get('static') |
| label1 = models.Label.smart_get('label1') |
| label2 = models.Label.smart_get('label2') |
| static_label = models.StaticLabel.objects.create(name='static') |
| |
| host1 = models.Host.objects.create(hostname='test_host') |
| host1.labels.add(label) |
| host1.labels.add(label1) |
| host1.labels.add(label2) |
| host1.static_labels.add(static_label) |
| host1.save() |
| |
| rpc_interface.host_remove_labels( |
| 'test_host', ['static', 'label1']) |
| labels = rpc_interface.get_labels(host__hostname__in=['test_host']) |
| # Only non_static label 'label1' is removed. |
| self.assertEquals(len(labels), 2) |
| self.assertEquals(labels[0].get('name'), 'label2') |
| |
| |
| def test_remove_board_from_shard(self): |
| """test remove a board (static label) from shard.""" |
| label = models.Label.smart_get('static') |
| static_label = models.StaticLabel.objects.create(name='static') |
| |
| shard = models.Shard.objects.create(hostname='test_shard') |
| shard.labels.add(label) |
| |
| host = models.Host.objects.create(hostname='test_host', |
| leased=False, |
| shard=shard) |
| host.static_labels.add(static_label) |
| host.save() |
| |
| rpc_interface.remove_board_from_shard(shard.hostname, label.name) |
| host1 = models.Host.smart_get(host.id) |
| shard1 = models.Shard.smart_get(shard.id) |
| self.assertEqual(host1.shard, None) |
| self.assertItemsEqual(shard1.labels.all(), []) |
| |
| |
| def test_check_job_dependencies_success(self): |
| """Test check_job_dependencies successfully.""" |
| static_label = models.StaticLabel.objects.create(name='static') |
| |
| host = models.Host.objects.create(hostname='test_host') |
| host.static_labels.add(static_label) |
| host.save() |
| |
| host1 = models.Host.smart_get(host.id) |
| rpc_utils.check_job_dependencies([host1], ['static']) |
| |
| |
| def test_check_job_dependencies_fail(self): |
| """Test check_job_dependencies with raising ValidationError.""" |
| label = models.Label.smart_get('static') |
| static_label = models.StaticLabel.objects.create(name='static') |
| |
| host = models.Host.objects.create(hostname='test_host') |
| host.labels.add(label) |
| host.save() |
| |
| host1 = models.Host.smart_get(host.id) |
| self.assertRaises(model_logic.ValidationError, |
| rpc_utils.check_job_dependencies, |
| [host1], |
| ['static']) |
| |
| def test_check_job_metahost_dependencies_success(self): |
| """Test check_job_metahost_dependencies successfully.""" |
| label1 = models.Label.smart_get('label1') |
| label2 = models.Label.smart_get('label2') |
| label = models.Label.smart_get('static') |
| static_label = models.StaticLabel.objects.create(name='static') |
| |
| host = models.Host.objects.create(hostname='test_host') |
| host.static_labels.add(static_label) |
| host.labels.add(label1) |
| host.labels.add(label2) |
| host.save() |
| |
| rpc_utils.check_job_metahost_dependencies( |
| [label1, label], [label2.name]) |
| rpc_utils.check_job_metahost_dependencies( |
| [label1], [label2.name, static_label.name]) |
| |
| |
| def test_check_job_metahost_dependencies_fail(self): |
| """Test check_job_metahost_dependencies with raising errors.""" |
| label1 = models.Label.smart_get('label1') |
| label2 = models.Label.smart_get('label2') |
| label = models.Label.smart_get('static') |
| static_label = models.StaticLabel.objects.create(name='static') |
| |
| host = models.Host.objects.create(hostname='test_host') |
| host.labels.add(label1) |
| host.labels.add(label2) |
| host.save() |
| |
| self.assertRaises(error.NoEligibleHostException, |
| rpc_utils.check_job_metahost_dependencies, |
| [label1, label], [label2.name]) |
| self.assertRaises(error.NoEligibleHostException, |
| rpc_utils.check_job_metahost_dependencies, |
| [label1], [label2.name, static_label.name]) |
| |
| |
| def _createShardAndHostWithStaticLabel(self, |
| shard_hostname='shard1', |
| host_hostname='test_host1', |
| label_name='board:lumpy'): |
| label = models.Label.objects.create(name=label_name) |
| |
| shard = models.Shard.objects.create(hostname=shard_hostname) |
| shard.labels.add(label) |
| |
| host = models.Host.objects.create(hostname=host_hostname, leased=False, |
| shard=shard) |
| host.labels.add(label) |
| if label_name in self._STATIC_LABELS: |
| models.ReplacedLabel.objects.create(label_id=label.id) |
| static_label = models.StaticLabel.objects.create(name=label_name) |
| host.static_labels.add(static_label) |
| |
| return shard, host, label |
| |
| |
| def testShardHeartbeatFetchHostlessJob(self): |
| shard1, host1, label1 = self._createShardAndHostWithStaticLabel( |
| host_hostname='test_host1') |
| self._testShardHeartbeatFetchHostlessJobHelper(host1) |
| |
| |
| def testShardHeartbeatIncorrectHosts(self): |
| shard1, host1, label1 = self._createShardAndHostWithStaticLabel( |
| host_hostname='test_host1') |
| self._testShardHeartbeatIncorrectHostsHelper(host1) |
| |
| |
| def testShardHeartbeatLabelRemovalRace(self): |
| shard1, host1, label1 = self._createShardAndHostWithStaticLabel( |
| host_hostname='test_host1') |
| self._testShardHeartbeatLabelRemovalRaceHelper(shard1, host1, label1) |
| |
| |
| def testShardRetrieveJobs(self): |
| shard1, host1, label1 = self._createShardAndHostWithStaticLabel() |
| shard2, host2, label2 = self._createShardAndHostWithStaticLabel( |
| 'shard2', 'test_host2', 'board:grumpy') |
| self._testShardRetrieveJobsHelper(shard1, host1, label1, |
| shard2, host2, label2) |
| |
| |
| def testResendJobsAfterFailedHeartbeat(self): |
| shard1, host1, label1 = self._createShardAndHostWithStaticLabel() |
| self._testResendJobsAfterFailedHeartbeatHelper(shard1, host1, label1) |
| |
| |
| def testResendHostsAfterFailedHeartbeat(self): |
| shard1, host1, label1 = self._createShardAndHostWithStaticLabel( |
| host_hostname='test_host1') |
| self._testResendHostsAfterFailedHeartbeatHelper(host1) |
| |
| |
| class RpcInterfaceTest(unittest.TestCase, |
| frontend_test_utils.FrontendTestMixin): |
| def setUp(self): |
| self._frontend_common_setup() |
| self.god = mock.mock_god() |
| |
| |
| def tearDown(self): |
| self.god.unstub_all() |
| self._frontend_common_teardown() |
| global_config.global_config.reset_config_values() |
| |
| |
| def test_validation(self): |
| # omit a required field |
| self.assertRaises(model_logic.ValidationError, rpc_interface.add_label, |
| name=None) |
| # violate uniqueness constraint |
| self.assertRaises(model_logic.ValidationError, rpc_interface.add_host, |
| hostname='host1') |
| |
| |
| def test_multiple_platforms(self): |
| platform2 = models.Label.objects.create(name='platform2', platform=True) |
| self.assertRaises(model_logic.ValidationError, |
| rpc_interface. label_add_hosts, id='platform2', |
| hosts=['host1', 'host2']) |
| self.assertRaises(model_logic.ValidationError, |
| rpc_interface.host_add_labels, |
| id='host1', labels=['platform2']) |
| # make sure the platform didn't get added |
| platforms = rpc_interface.get_labels( |
| host__hostname__in=['host1', 'host2'], platform=True) |
| self.assertEquals(len(platforms), 1) |
| self.assertEquals(platforms[0]['name'], 'myplatform') |
| |
| |
| def _check_hostnames(self, hosts, expected_hostnames): |
| self.assertEquals(set(host['hostname'] for host in hosts), |
| set(expected_hostnames)) |
| |
| |
| def test_ping_db(self): |
| self.assertEquals(rpc_interface.ping_db(), [True]) |
| |
| |
| def test_get_hosts_by_attribute(self): |
| host1 = models.Host.objects.create(hostname='test_host1') |
| host1.set_attribute('test_attribute1', 'test_value1') |
| host2 = models.Host.objects.create(hostname='test_host2') |
| host2.set_attribute('test_attribute1', 'test_value1') |
| |
| hosts = rpc_interface.get_hosts_by_attribute( |
| 'test_attribute1', 'test_value1') |
| self.assertEquals(set(hosts), |
| set(['test_host1', 'test_host2'])) |
| |
| |
| def test_get_host_attribute(self): |
| host1 = models.Host.objects.create(hostname='test_host1') |
| host1.set_attribute('test_attribute1', 'test_value1') |
| host2 = models.Host.objects.create(hostname='test_host2') |
| host2.set_attribute('test_attribute1', 'test_value1') |
| |
| attributes = rpc_interface.get_host_attribute( |
| 'test_attribute1', |
| hostname__in=['test_host1', 'test_host2']) |
| hosts = [attr['host'] for attr in attributes] |
| values = [attr['value'] for attr in attributes] |
| self.assertEquals(set(hosts), |
| set(['test_host1', 'test_host2'])) |
| self.assertEquals(set(values), set(['test_value1'])) |
| |
| |
| def test_get_hosts(self): |
| hosts = rpc_interface.get_hosts() |
| self._check_hostnames(hosts, [host.hostname for host in self.hosts]) |
| |
| hosts = rpc_interface.get_hosts(hostname='host1') |
| self._check_hostnames(hosts, ['host1']) |
| host = hosts[0] |
| self.assertEquals(sorted(host['labels']), ['label1', 'myplatform']) |
| self.assertEquals(host['platform'], 'myplatform') |
| self.assertEquals(host['acls'], ['my_acl']) |
| self.assertEquals(host['attributes'], {}) |
| |
| |
| def test_get_hosts_multiple_labels(self): |
| hosts = rpc_interface.get_hosts( |
| multiple_labels=['myplatform', 'label1']) |
| self._check_hostnames(hosts, ['host1']) |
| |
| |
| def test_job_keyvals(self): |
| keyval_dict = {'mykey': 'myvalue'} |
| job_id = rpc_interface.create_job(name='test', |
| priority=priorities.Priority.DEFAULT, |
| control_file='foo', |
| control_type=CLIENT, |
| hosts=['host1'], |
| keyvals=keyval_dict) |
| jobs = rpc_interface.get_jobs(id=job_id) |
| self.assertEquals(len(jobs), 1) |
| self.assertEquals(jobs[0]['keyvals'], keyval_dict) |
| |
| |
| def test_get_jobs_summary(self): |
| job = self._create_job(hosts=xrange(1, 4)) |
| entries = list(job.hostqueueentry_set.all()) |
| entries[1].status = _hqe_status.FAILED |
| entries[1].save() |
| entries[2].status = _hqe_status.FAILED |
| entries[2].aborted = True |
| entries[2].save() |
| |
| # Mock up tko_rpc_interface.get_status_counts. |
| self.god.stub_function_to_return(rpc_interface.tko_rpc_interface, |
| 'get_status_counts', |
| None) |
| |
| job_summaries = rpc_interface.get_jobs_summary(id=job.id) |
| self.assertEquals(len(job_summaries), 1) |
| summary = job_summaries[0] |
| self.assertEquals(summary['status_counts'], {'Queued': 1, |
| 'Failed': 2}) |
| |
| |
| def _check_job_ids(self, actual_job_dicts, expected_jobs): |
| self.assertEquals( |
| set(job_dict['id'] for job_dict in actual_job_dicts), |
| set(job.id for job in expected_jobs)) |
| |
| |
| def test_get_jobs_status_filters(self): |
| HqeStatus = models.HostQueueEntry.Status |
| def create_two_host_job(): |
| return self._create_job(hosts=[1, 2]) |
| def set_hqe_statuses(job, first_status, second_status): |
| entries = job.hostqueueentry_set.all() |
| entries[0].update_object(status=first_status) |
| entries[1].update_object(status=second_status) |
| |
| queued = create_two_host_job() |
| |
| queued_and_running = create_two_host_job() |
| set_hqe_statuses(queued_and_running, HqeStatus.QUEUED, |
| HqeStatus.RUNNING) |
| |
| running_and_complete = create_two_host_job() |
| set_hqe_statuses(running_and_complete, HqeStatus.RUNNING, |
| HqeStatus.COMPLETED) |
| |
| complete = create_two_host_job() |
| set_hqe_statuses(complete, HqeStatus.COMPLETED, HqeStatus.COMPLETED) |
| |
| started_but_inactive = create_two_host_job() |
| set_hqe_statuses(started_but_inactive, HqeStatus.QUEUED, |
| HqeStatus.COMPLETED) |
| |
| parsing = create_two_host_job() |
| set_hqe_statuses(parsing, HqeStatus.PARSING, HqeStatus.PARSING) |
| |
| self._check_job_ids(rpc_interface.get_jobs(not_yet_run=True), [queued]) |
| self._check_job_ids(rpc_interface.get_jobs(running=True), |
| [queued_and_running, running_and_complete, |
| started_but_inactive, parsing]) |
| self._check_job_ids(rpc_interface.get_jobs(finished=True), [complete]) |
| |
| |
| def test_get_jobs_type_filters(self): |
| self.assertRaises(AssertionError, rpc_interface.get_jobs, |
| suite=True, sub=True) |
| self.assertRaises(AssertionError, rpc_interface.get_jobs, |
| suite=True, standalone=True) |
| self.assertRaises(AssertionError, rpc_interface.get_jobs, |
| standalone=True, sub=True) |
| |
| parent_job = self._create_job(hosts=[1]) |
| child_jobs = self._create_job(hosts=[1, 2], |
| parent_job_id=parent_job.id) |
| standalone_job = self._create_job(hosts=[1]) |
| |
| self._check_job_ids(rpc_interface.get_jobs(suite=True), [parent_job]) |
| self._check_job_ids(rpc_interface.get_jobs(sub=True), [child_jobs]) |
| self._check_job_ids(rpc_interface.get_jobs(standalone=True), |
| [standalone_job]) |
| |
| |
| def _create_job_helper(self, **kwargs): |
| return rpc_interface.create_job(name='test', |
| priority=priorities.Priority.DEFAULT, |
| control_file='control file', |
| control_type=SERVER, **kwargs) |
| |
| |
| def test_one_time_hosts(self): |
| job = self._create_job_helper(one_time_hosts=['testhost']) |
| host = models.Host.objects.get(hostname='testhost') |
| self.assertEquals(host.invalid, True) |
| self.assertEquals(host.labels.count(), 0) |
| self.assertEquals(host.aclgroup_set.count(), 0) |
| |
| |
| def test_create_job_duplicate_hosts(self): |
| self.assertRaises(model_logic.ValidationError, self._create_job_helper, |
| hosts=[1, 1]) |
| |
| |
| def test_create_unrunnable_metahost_job(self): |
| self.assertRaises(error.NoEligibleHostException, |
| self._create_job_helper, meta_hosts=['unused']) |
| |
| |
| def test_create_hostless_job(self): |
| job_id = self._create_job_helper(hostless=True) |
| job = models.Job.objects.get(pk=job_id) |
| queue_entries = job.hostqueueentry_set.all() |
| self.assertEquals(len(queue_entries), 1) |
| self.assertEquals(queue_entries[0].host, None) |
| self.assertEquals(queue_entries[0].meta_host, None) |
| |
| |
| def _setup_special_tasks(self): |
| host = self.hosts[0] |
| |
| job1 = self._create_job(hosts=[1]) |
| job2 = self._create_job(hosts=[1]) |
| |
| entry1 = job1.hostqueueentry_set.all()[0] |
| entry1.update_object(started_on=datetime.datetime(2009, 1, 2), |
| execution_subdir='host1') |
| entry2 = job2.hostqueueentry_set.all()[0] |
| entry2.update_object(started_on=datetime.datetime(2009, 1, 3), |
| execution_subdir='host1') |
| |
| self.task1 = models.SpecialTask.objects.create( |
| host=host, task=models.SpecialTask.Task.VERIFY, |
| time_started=datetime.datetime(2009, 1, 1), # ran before job 1 |
| is_complete=True, requested_by=models.User.current_user()) |
| self.task2 = models.SpecialTask.objects.create( |
| host=host, task=models.SpecialTask.Task.VERIFY, |
| queue_entry=entry2, # ran with job 2 |
| is_active=True, requested_by=models.User.current_user()) |
| self.task3 = models.SpecialTask.objects.create( |
| host=host, task=models.SpecialTask.Task.VERIFY, |
| requested_by=models.User.current_user()) # not yet run |
| |
| |
| def test_get_special_tasks(self): |
| self._setup_special_tasks() |
| tasks = rpc_interface.get_special_tasks(host__hostname='host1', |
| queue_entry__isnull=True) |
| self.assertEquals(len(tasks), 2) |
| self.assertEquals(tasks[0]['task'], models.SpecialTask.Task.VERIFY) |
| self.assertEquals(tasks[0]['is_active'], False) |
| self.assertEquals(tasks[0]['is_complete'], True) |
| |
| |
| def test_get_latest_special_task(self): |
| # a particular usage of get_special_tasks() |
| self._setup_special_tasks() |
| self.task2.time_started = datetime.datetime(2009, 1, 2) |
| self.task2.save() |
| |
| tasks = rpc_interface.get_special_tasks( |
| host__hostname='host1', task=models.SpecialTask.Task.VERIFY, |
| time_started__isnull=False, sort_by=['-time_started'], |
| query_limit=1) |
| self.assertEquals(len(tasks), 1) |
| self.assertEquals(tasks[0]['id'], 2) |
| |
| |
| def _common_entry_check(self, entry_dict): |
| self.assertEquals(entry_dict['host']['hostname'], 'host1') |
| self.assertEquals(entry_dict['job']['id'], 2) |
| |
| |
| def test_get_host_queue_entries_and_special_tasks(self): |
| self._setup_special_tasks() |
| |
| host = self.hosts[0].id |
| entries_and_tasks = ( |
| rpc_interface.get_host_queue_entries_and_special_tasks(host)) |
| |
| paths = [entry['execution_path'] for entry in entries_and_tasks] |
| self.assertEquals(paths, ['hosts/host1/3-verify', |
| '2-autotest_system/host1', |
| 'hosts/host1/2-verify', |
| '1-autotest_system/host1', |
| 'hosts/host1/1-verify']) |
| |
| verify2 = entries_and_tasks[2] |
| self._common_entry_check(verify2) |
| self.assertEquals(verify2['type'], 'Verify') |
| self.assertEquals(verify2['status'], 'Running') |
| self.assertEquals(verify2['execution_path'], 'hosts/host1/2-verify') |
| |
| entry2 = entries_and_tasks[1] |
| self._common_entry_check(entry2) |
| self.assertEquals(entry2['type'], 'Job') |
| self.assertEquals(entry2['status'], 'Queued') |
| self.assertEquals(entry2['started_on'], '2009-01-03 00:00:00') |
| |
| |
| def _create_hqes_and_start_time_index_entries(self): |
| shard = models.Shard.objects.create(hostname='shard') |
| job = self._create_job(shard=shard, control_file='foo') |
| HqeStatus = models.HostQueueEntry.Status |
| |
| models.HostQueueEntry( |
| id=1, job=job, started_on='2017-01-01', |
| status=HqeStatus.QUEUED).save() |
| models.HostQueueEntry( |
| id=2, job=job, started_on='2017-01-02', |
| status=HqeStatus.QUEUED).save() |
| models.HostQueueEntry( |
| id=3, job=job, started_on='2017-01-03', |
| status=HqeStatus.QUEUED).save() |
| |
| models.HostQueueEntryStartTimes( |
| insert_time='2017-01-03', highest_hqe_id=3).save() |
| models.HostQueueEntryStartTimes( |
| insert_time='2017-01-02', highest_hqe_id=2).save() |
| models.HostQueueEntryStartTimes( |
| insert_time='2017-01-01', highest_hqe_id=1).save() |
| |
| def test_get_host_queue_entries_by_insert_time(self): |
| """Check the insert_time_after and insert_time_before constraints.""" |
| self._create_hqes_and_start_time_index_entries() |
| hqes = rpc_interface.get_host_queue_entries_by_insert_time( |
| insert_time_after='2017-01-01') |
| self.assertEquals(len(hqes), 3) |
| |
| hqes = rpc_interface.get_host_queue_entries_by_insert_time( |
| insert_time_after='2017-01-02') |
| self.assertEquals(len(hqes), 2) |
| |
| hqes = rpc_interface.get_host_queue_entries_by_insert_time( |
| insert_time_after='2017-01-03') |
| self.assertEquals(len(hqes), 1) |
| |
| hqes = rpc_interface.get_host_queue_entries_by_insert_time( |
| insert_time_before='2017-01-01') |
| self.assertEquals(len(hqes), 1) |
| |
| hqes = rpc_interface.get_host_queue_entries_by_insert_time( |
| insert_time_before='2017-01-02') |
| self.assertEquals(len(hqes), 2) |
| |
| hqes = rpc_interface.get_host_queue_entries_by_insert_time( |
| insert_time_before='2017-01-03') |
| self.assertEquals(len(hqes), 3) |
| |
| |
| def test_get_host_queue_entries_by_insert_time_with_missing_index_row(self): |
| """Shows that the constraints are approximate. |
| |
| The query may return rows which are actually outside of the bounds |
| given, if the index table does not have an entry for the specific time. |
| """ |
| self._create_hqes_and_start_time_index_entries() |
| hqes = rpc_interface.get_host_queue_entries_by_insert_time( |
| insert_time_before='2016-12-01') |
| self.assertEquals(len(hqes), 1) |
| |
| def test_get_hqe_by_insert_time_with_before_and_after(self): |
| self._create_hqes_and_start_time_index_entries() |
| hqes = rpc_interface.get_host_queue_entries_by_insert_time( |
| insert_time_before='2017-01-02', |
| insert_time_after='2017-01-02') |
| self.assertEquals(len(hqes), 1) |
| |
| def test_get_hqe_by_insert_time_and_id_constraint(self): |
| self._create_hqes_and_start_time_index_entries() |
| # The time constraint is looser than the id constraint, so the time |
| # constraint should take precedence. |
| hqes = rpc_interface.get_host_queue_entries_by_insert_time( |
| insert_time_before='2017-01-02', |
| id__lte=1) |
| self.assertEquals(len(hqes), 1) |
| |
| # Now make the time constraint tighter than the id constraint. |
| hqes = rpc_interface.get_host_queue_entries_by_insert_time( |
| insert_time_before='2017-01-01', |
| id__lte=42) |
| self.assertEquals(len(hqes), 1) |
| |
| def test_view_invalid_host(self): |
| # RPCs used by View Host page should work for invalid hosts |
| self._create_job_helper(hosts=[1]) |
| host = self.hosts[0] |
| host.delete() |
| |
| self.assertEquals(1, rpc_interface.get_num_hosts(hostname='host1', |
| valid_only=False)) |
| data = rpc_interface.get_hosts(hostname='host1', valid_only=False) |
| self.assertEquals(1, len(data)) |
| |
| self.assertEquals(1, rpc_interface.get_num_host_queue_entries( |
| host__hostname='host1')) |
| data = rpc_interface.get_host_queue_entries(host__hostname='host1') |
| self.assertEquals(1, len(data)) |
| |
| count = rpc_interface.get_num_host_queue_entries_and_special_tasks( |
| host=host.id) |
| self.assertEquals(1, count) |
| data = rpc_interface.get_host_queue_entries_and_special_tasks( |
| host=host.id) |
| self.assertEquals(1, len(data)) |
| |
| |
| def test_reverify_hosts(self): |
| hostname_list = rpc_interface.reverify_hosts(id__in=[1, 2]) |
| self.assertEquals(hostname_list, ['host1', 'host2']) |
| tasks = rpc_interface.get_special_tasks() |
| self.assertEquals(len(tasks), 2) |
| self.assertEquals(set(task['host']['id'] for task in tasks), |
| set([1, 2])) |
| |
| task = tasks[0] |
| self.assertEquals(task['task'], models.SpecialTask.Task.VERIFY) |
| self.assertEquals(task['requested_by'], 'autotest_system') |
| |
| |
| def test_repair_hosts(self): |
| hostname_list = rpc_interface.repair_hosts(id__in=[1, 2]) |
| self.assertEquals(hostname_list, ['host1', 'host2']) |
| tasks = rpc_interface.get_special_tasks() |
| self.assertEquals(len(tasks), 2) |
| self.assertEquals(set(task['host']['id'] for task in tasks), |
| set([1, 2])) |
| |
| task = tasks[0] |
| self.assertEquals(task['task'], models.SpecialTask.Task.REPAIR) |
| self.assertEquals(task['requested_by'], 'autotest_system') |
| |
| |
| def _modify_host_helper(self, on_shard=False, host_on_shard=False): |
| shard_hostname = 'shard1' |
| if on_shard: |
| global_config.global_config.override_config_value( |
| 'SHARD', 'shard_hostname', shard_hostname) |
| |
| host = models.Host.objects.all()[0] |
| if host_on_shard: |
| shard = models.Shard.objects.create(hostname=shard_hostname) |
| host.shard = shard |
| host.save() |
| |
| self.assertFalse(host.locked) |
| |
| self.god.stub_class_method(frontend.AFE, 'run') |
| |
| if host_on_shard and not on_shard: |
| mock_afe = self.god.create_mock_class_obj( |
| frontend_wrappers.RetryingAFE, 'MockAFE') |
| self.god.stub_with(frontend_wrappers, 'RetryingAFE', mock_afe) |
| |
| mock_afe2 = frontend_wrappers.RetryingAFE.expect_new( |
| server=shard_hostname, user=None) |
| mock_afe2.run.expect_call('modify_host_local', id=host.id, |
| locked=True, lock_reason='_modify_host_helper lock', |
| lock_time=datetime.datetime(2015, 12, 15)) |
| elif on_shard: |
| mock_afe = self.god.create_mock_class_obj( |
| frontend_wrappers.RetryingAFE, 'MockAFE') |
| self.god.stub_with(frontend_wrappers, 'RetryingAFE', mock_afe) |
| |
| mock_afe2 = frontend_wrappers.RetryingAFE.expect_new( |
| server=server_utils.get_global_afe_hostname(), user=None) |
| mock_afe2.run.expect_call('modify_host', id=host.id, |
| locked=True, lock_reason='_modify_host_helper lock', |
| lock_time=datetime.datetime(2015, 12, 15)) |
| |
| rpc_interface.modify_host(id=host.id, locked=True, |
| lock_reason='_modify_host_helper lock', |
| lock_time=datetime.datetime(2015, 12, 15)) |
| |
| host = models.Host.objects.get(pk=host.id) |
| if on_shard: |
| # modify_host on shard does nothing but routing the RPC to master. |
| self.assertFalse(host.locked) |
| else: |
| self.assertTrue(host.locked) |
| self.god.check_playback() |
| |
| |
| def test_modify_host_on_master_host_on_master(self): |
| """Call modify_host to master for host in master.""" |
| self._modify_host_helper() |
| |
| |
| def test_modify_host_on_master_host_on_shard(self): |
| """Call modify_host to master for host in shard.""" |
| self._modify_host_helper(host_on_shard=True) |
| |
| |
| def test_modify_host_on_shard(self): |
| """Call modify_host to shard for host in shard.""" |
| self._modify_host_helper(on_shard=True, host_on_shard=True) |
| |
| |
| def test_modify_hosts_on_master_host_on_shard(self): |
| """Ensure calls to modify_hosts are correctly forwarded to shards.""" |
| host1 = models.Host.objects.all()[0] |
| host2 = models.Host.objects.all()[1] |
| |
| shard1 = models.Shard.objects.create(hostname='shard1') |
| host1.shard = shard1 |
| host1.save() |
| |
| shard2 = models.Shard.objects.create(hostname='shard2') |
| host2.shard = shard2 |
| host2.save() |
| |
| self.assertFalse(host1.locked) |
| self.assertFalse(host2.locked) |
| |
| mock_afe = self.god.create_mock_class_obj(frontend_wrappers.RetryingAFE, |
| 'MockAFE') |
| self.god.stub_with(frontend_wrappers, 'RetryingAFE', mock_afe) |
| |
| # The statuses of one host might differ on master and shard. |
| # Filters are always applied on the master. So the host on the shard |
| # will be affected no matter what his status is. |
| filters_to_use = {'status': 'Ready'} |
| |
| mock_afe2 = frontend_wrappers.RetryingAFE.expect_new( |
| server='shard2', user=None) |
| mock_afe2.run.expect_call( |
| 'modify_hosts_local', |
| host_filter_data={'id__in': [shard1.id, shard2.id]}, |
| update_data={'locked': True, |
| 'lock_reason': 'Testing forward to shard', |
| 'lock_time' : datetime.datetime(2015, 12, 15) }) |
| |
| mock_afe1 = frontend_wrappers.RetryingAFE.expect_new( |
| server='shard1', user=None) |
| mock_afe1.run.expect_call( |
| 'modify_hosts_local', |
| host_filter_data={'id__in': [shard1.id, shard2.id]}, |
| update_data={'locked': True, |
| 'lock_reason': 'Testing forward to shard', |
| 'lock_time' : datetime.datetime(2015, 12, 15)}) |
| |
| rpc_interface.modify_hosts( |
| host_filter_data={'status': 'Ready'}, |
| update_data={'locked': True, |
| 'lock_reason': 'Testing forward to shard', |
| 'lock_time' : datetime.datetime(2015, 12, 15) }) |
| |
| host1 = models.Host.objects.get(pk=host1.id) |
| self.assertTrue(host1.locked) |
| host2 = models.Host.objects.get(pk=host2.id) |
| self.assertTrue(host2.locked) |
| self.god.check_playback() |
| |
| |
| def test_delete_host(self): |
| """Ensure an RPC is made on delete a host, if it is on a shard.""" |
| host1 = models.Host.objects.all()[0] |
| shard1 = models.Shard.objects.create(hostname='shard1') |
| host1.shard = shard1 |
| host1.save() |
| host1_id = host1.id |
| |
| mock_afe = self.god.create_mock_class_obj(frontend_wrappers.RetryingAFE, |
| 'MockAFE') |
| self.god.stub_with(frontend_wrappers, 'RetryingAFE', mock_afe) |
| |
| mock_afe1 = frontend_wrappers.RetryingAFE.expect_new( |
| server='shard1', user=None) |
| mock_afe1.run.expect_call('delete_host', id=host1.id) |
| |
| rpc_interface.delete_host(id=host1.id) |
| |
| self.assertRaises(models.Host.DoesNotExist, |
| models.Host.smart_get, host1_id) |
| |
| self.god.check_playback() |
| |
| |
| def test_modify_label(self): |
| label1 = models.Label.objects.all()[0] |
| self.assertEqual(label1.invalid, 0) |
| |
| host2 = models.Host.objects.all()[1] |
| shard1 = models.Shard.objects.create(hostname='shard1') |
| host2.shard = shard1 |
| host2.labels.add(label1) |
| host2.save() |
| |
| mock_afe = self.god.create_mock_class_obj(frontend_wrappers.RetryingAFE, |
| 'MockAFE') |
| self.god.stub_with(frontend_wrappers, 'RetryingAFE', mock_afe) |
| |
| mock_afe1 = frontend_wrappers.RetryingAFE.expect_new( |
| server='shard1', user=None) |
| mock_afe1.run.expect_call('modify_label', id=label1.id, invalid=1) |
| |
| rpc_interface.modify_label(label1.id, invalid=1) |
| |
| self.assertEqual(models.Label.objects.all()[0].invalid, 1) |
| self.god.check_playback() |
| |
| |
| def test_delete_label(self): |
| label1 = models.Label.objects.all()[0] |
| |
| host2 = models.Host.objects.all()[1] |
| shard1 = models.Shard.objects.create(hostname='shard1') |
| host2.shard = shard1 |
| host2.labels.add(label1) |
| host2.save() |
| |
| mock_afe = self.god.create_mock_class_obj(frontend_wrappers.RetryingAFE, |
| 'MockAFE') |
| self.god.stub_with(frontend_wrappers, 'RetryingAFE', mock_afe) |
| |
| mock_afe1 = frontend_wrappers.RetryingAFE.expect_new( |
| server='shard1', user=None) |
| mock_afe1.run.expect_call('delete_label', id=label1.id) |
| |
| rpc_interface.delete_label(id=label1.id) |
| |
| self.assertRaises(models.Label.DoesNotExist, |
| models.Label.smart_get, label1.id) |
| self.god.check_playback() |
| |
| |
| def test_get_image_for_job_with_keyval_build(self): |
| keyval_dict = {'build': 'cool-image'} |
| job_id = rpc_interface.create_job(name='test', |
| priority=priorities.Priority.DEFAULT, |
| control_file='foo', |
| control_type=CLIENT, |
| hosts=['host1'], |
| keyvals=keyval_dict) |
| job = models.Job.objects.get(id=job_id) |
| self.assertIsNotNone(job) |
| image = rpc_interface._get_image_for_job(job, True) |
| self.assertEquals('cool-image', image) |
| |
| |
| def test_get_image_for_job_with_keyval_builds(self): |
| keyval_dict = {'builds': {'cros-version': 'cool-image'}} |
| job_id = rpc_interface.create_job(name='test', |
| priority=priorities.Priority.DEFAULT, |
| control_file='foo', |
| control_type=CLIENT, |
| hosts=['host1'], |
| keyvals=keyval_dict) |
| job = models.Job.objects.get(id=job_id) |
| self.assertIsNotNone(job) |
| image = rpc_interface._get_image_for_job(job, True) |
| self.assertEquals('cool-image', image) |
| |
| |
| def test_get_image_for_job_with_control_build(self): |
| CONTROL_FILE = """build='cool-image' |
| """ |
| job_id = rpc_interface.create_job(name='test', |
| priority=priorities.Priority.DEFAULT, |
| control_file='foo', |
| control_type=CLIENT, |
| hosts=['host1']) |
| job = models.Job.objects.get(id=job_id) |
| self.assertIsNotNone(job) |
| job.control_file = CONTROL_FILE |
| image = rpc_interface._get_image_for_job(job, True) |
| self.assertEquals('cool-image', image) |
| |
| |
| def test_get_image_for_job_with_control_builds(self): |
| CONTROL_FILE = """builds={'cros-version': 'cool-image'} |
| """ |
| job_id = rpc_interface.create_job(name='test', |
| priority=priorities.Priority.DEFAULT, |
| control_file='foo', |
| control_type=CLIENT, |
| hosts=['host1']) |
| job = models.Job.objects.get(id=job_id) |
| self.assertIsNotNone(job) |
| job.control_file = CONTROL_FILE |
| image = rpc_interface._get_image_for_job(job, True) |
| self.assertEquals('cool-image', image) |
| |
| |
| class ExtraRpcInterfaceTest(frontend_test_utils.FrontendTestMixin, |
| ShardHeartbeatTest): |
| """Unit tests for functions originally in site_rpc_interface.py. |
| |
| @var _NAME: fake suite name. |
| @var _BOARD: fake board to reimage. |
| @var _BUILD: fake build with which to reimage. |
| @var _PRIORITY: fake priority with which to reimage. |
| """ |
| _NAME = 'name' |
| _BOARD = 'link' |
| _BUILD = 'link-release/R36-5812.0.0' |
| _BUILDS = {provision.CROS_VERSION_PREFIX: _BUILD} |
| _PRIORITY = priorities.Priority.DEFAULT |
| _TIMEOUT = 24 |
| |
| |
| def setUp(self): |
| super(ExtraRpcInterfaceTest, self).setUp() |
| self._SUITE_NAME = suite_common.canonicalize_suite_name( |
| self._NAME) |
| self.dev_server = self.mox.CreateMock(dev_server.ImageServer) |
| self._frontend_common_setup(fill_data=False) |
| |
| |
| def tearDown(self): |
| self._frontend_common_teardown() |
| |
| |
| def _setupDevserver(self): |
| self.mox.StubOutClassWithMocks(dev_server, 'ImageServer') |
| dev_server.resolve(self._BUILD).AndReturn(self.dev_server) |
| |
| |
| def _mockDevServerGetter(self, get_control_file=True): |
| self._setupDevserver() |
| if get_control_file: |
| self.getter = self.mox.CreateMock( |
| control_file_getter.DevServerGetter) |
| self.mox.StubOutWithMock(control_file_getter.DevServerGetter, |
| 'create') |
| control_file_getter.DevServerGetter.create( |
| mox.IgnoreArg(), mox.IgnoreArg()).AndReturn(self.getter) |
| |
| |
| def _mockRpcUtils(self, to_return, control_file_substring=''): |
| """Fake out the autotest rpc_utils module with a mockable class. |
| |
| @param to_return: the value that rpc_utils.create_job_common() should |
| be mocked out to return. |
| @param control_file_substring: A substring that is expected to appear |
| in the control file output string that |
| is passed to create_job_common. |
| Default: '' |
| """ |
| download_started_time = constants.DOWNLOAD_STARTED_TIME |
| payload_finished_time = constants.PAYLOAD_FINISHED_TIME |
| self.mox.StubOutWithMock(rpc_utils, 'create_job_common') |
| rpc_utils.create_job_common(mox.And(mox.StrContains(self._NAME), |
| mox.StrContains(self._BUILD)), |
| priority=self._PRIORITY, |
| timeout_mins=self._TIMEOUT*60, |
| max_runtime_mins=self._TIMEOUT*60, |
| control_type='Server', |
| control_file=mox.And(mox.StrContains(self._BOARD), |
| mox.StrContains(self._BUILD), |
| mox.StrContains( |
| control_file_substring)), |
| hostless=True, |
| keyvals=mox.And(mox.In(download_started_time), |
| mox.In(payload_finished_time)) |
| ).AndReturn(to_return) |
| |
| |
| def testStageBuildFail(self): |
| """Ensure that a failure to stage the desired build fails the RPC.""" |
| self._setupDevserver() |
| |
| self.dev_server.hostname = 'mox_url' |
| self.dev_server.stage_artifacts( |
| image=self._BUILD, artifacts=['test_suites']).AndRaise( |
| dev_server.DevServerException()) |
| self.mox.ReplayAll() |
| self.assertRaises(error.StageControlFileFailure, |
| rpc_interface.create_suite_job, |
| name=self._NAME, |
| board=self._BOARD, |
| builds=self._BUILDS, |
| pool=None) |
| |
| |
| def testGetControlFileFail(self): |
| """Ensure that a failure to get needed control file fails the RPC.""" |
| self._mockDevServerGetter() |
| |
| self.dev_server.hostname = 'mox_url' |
| self.dev_server.stage_artifacts( |
| image=self._BUILD, artifacts=['test_suites']).AndReturn(True) |
| |
| self.getter.get_control_file_contents_by_name( |
| self._SUITE_NAME).AndReturn(None) |
| self.mox.ReplayAll() |
| self.assertRaises(error.ControlFileEmpty, |
| rpc_interface.create_suite_job, |
| name=self._NAME, |
| board=self._BOARD, |
| builds=self._BUILDS, |
| pool=None) |
| |
| |
| def testGetControlFileListFail(self): |
| """Ensure that a failure to get needed control file fails the RPC.""" |
| self._mockDevServerGetter() |
| |
| self.dev_server.hostname = 'mox_url' |
| self.dev_server.stage_artifacts( |
| image=self._BUILD, artifacts=['test_suites']).AndReturn(True) |
| |
| self.getter.get_control_file_contents_by_name( |
| self._SUITE_NAME).AndRaise(error.NoControlFileList()) |
| self.mox.ReplayAll() |
| self.assertRaises(error.NoControlFileList, |
| rpc_interface.create_suite_job, |
| name=self._NAME, |
| board=self._BOARD, |
| builds=self._BUILDS, |
| pool=None) |
| |
| |
| def testCreateSuiteJobFail(self): |
| """Ensure that failure to schedule the suite job fails the RPC.""" |
| self._mockDevServerGetter() |
| |
| self.dev_server.hostname = 'mox_url' |
| self.dev_server.stage_artifacts( |
| image=self._BUILD, artifacts=['test_suites']).AndReturn(True) |
| |
| self.getter.get_control_file_contents_by_name( |
| self._SUITE_NAME).AndReturn('f') |
| |
| self.dev_server.url().AndReturn('mox_url') |
| self._mockRpcUtils(-1) |
| self.mox.ReplayAll() |
| self.assertEquals( |
| rpc_interface.create_suite_job(name=self._NAME, |
| board=self._BOARD, |
| builds=self._BUILDS, pool=None), |
| -1) |
| |
| |
| def testCreateSuiteJobSuccess(self): |
| """Ensures that success results in a successful RPC.""" |
| self._mockDevServerGetter() |
| |
| self.dev_server.hostname = 'mox_url' |
| self.dev_server.stage_artifacts( |
| image=self._BUILD, artifacts=['test_suites']).AndReturn(True) |
| |
| self.getter.get_control_file_contents_by_name( |
| self._SUITE_NAME).AndReturn('f') |
| |
| self.dev_server.url().AndReturn('mox_url') |
| job_id = 5 |
| self._mockRpcUtils(job_id) |
| self.mox.ReplayAll() |
| self.assertEquals( |
| rpc_interface.create_suite_job(name=self._NAME, |
| board=self._BOARD, |
| builds=self._BUILDS, |
| pool=None), |
| job_id) |
| |
| |
| def testCreateSuiteJobNoHostCheckSuccess(self): |
| """Ensures that success results in a successful RPC.""" |
| self._mockDevServerGetter() |
| |
| self.dev_server.hostname = 'mox_url' |
| self.dev_server.stage_artifacts( |
| image=self._BUILD, artifacts=['test_suites']).AndReturn(True) |
| |
| self.getter.get_control_file_contents_by_name( |
| self._SUITE_NAME).AndReturn('f') |
| |
| self.dev_server.url().AndReturn('mox_url') |
| job_id = 5 |
| self._mockRpcUtils(job_id) |
| self.mox.ReplayAll() |
| self.assertEquals( |
| rpc_interface.create_suite_job(name=self._NAME, |
| board=self._BOARD, |
| builds=self._BUILDS, |
| pool=None, check_hosts=False), |
| job_id) |
| |
| |
| def testCreateSuiteJobControlFileSupplied(self): |
| """Ensure we can supply the control file to create_suite_job.""" |
| self._mockDevServerGetter(get_control_file=False) |
| |
| self.dev_server.hostname = 'mox_url' |
| self.dev_server.stage_artifacts( |
| image=self._BUILD, artifacts=['test_suites']).AndReturn(True) |
| self.dev_server.url().AndReturn('mox_url') |
| job_id = 5 |
| self._mockRpcUtils(job_id) |
| self.mox.ReplayAll() |
| self.assertEquals( |
| rpc_interface.create_suite_job(name='%s/%s' % (self._NAME, |
| self._BUILD), |
| board=None, |
| builds=self._BUILDS, |
| pool=None, |
| control_file='CONTROL FILE'), |
| job_id) |
| |
| |
| def _get_records_for_sending_to_master(self): |
| return [{'control_file': 'foo', |
| 'control_type': 1, |
| 'created_on': datetime.datetime(2014, 8, 21), |
| 'drone_set': None, |
| 'email_list': '', |
| 'max_runtime_hrs': 72, |
| 'max_runtime_mins': 1440, |
| 'name': 'dummy', |
| 'owner': 'autotest_system', |
| 'parse_failed_repair': True, |
| 'priority': 40, |
| 'reboot_after': 0, |
| 'reboot_before': 1, |
| 'run_reset': True, |
| 'run_verify': False, |
| 'synch_count': 0, |
| 'test_retry': 0, |
| 'timeout': 24, |
| 'timeout_mins': 1440, |
| 'id': 1 |
| }], [{ |
| 'aborted': False, |
| 'active': False, |
| 'complete': False, |
| 'deleted': False, |
| 'execution_subdir': '', |
| 'finished_on': None, |
| 'started_on': None, |
| 'status': 'Queued', |
| 'id': 1 |
| }] |
| |
| |
| def _send_records_to_master_helper( |
| self, jobs, hqes, shard_hostname='host1', |
| exception_to_throw=error.UnallowedRecordsSentToMaster, aborted=False): |
| job_id = rpc_interface.create_job( |
| name='dummy', |
| priority=self._PRIORITY, |
| control_file='foo', |
| control_type=SERVER, |
| hostless=True) |
| job = models.Job.objects.get(pk=job_id) |
| shard = models.Shard.objects.create(hostname='host1') |
| job.shard = shard |
| job.save() |
| |
| if aborted: |
| job.hostqueueentry_set.update(aborted=True) |
| job.shard = None |
| job.save() |
| |
| hqe = job.hostqueueentry_set.all()[0] |
| if not exception_to_throw: |
| self._do_heartbeat_and_assert_response( |
| shard_hostname=shard_hostname, |
| upload_jobs=jobs, upload_hqes=hqes) |
| else: |
| self.assertRaises( |
| exception_to_throw, |
| self._do_heartbeat_and_assert_response, |
| shard_hostname=shard_hostname, |
| upload_jobs=jobs, upload_hqes=hqes) |
| |
| |
| def testSendingRecordsToMaster(self): |
| """Send records to the master and ensure they are persisted.""" |
| jobs, hqes = self._get_records_for_sending_to_master() |
| hqes[0]['status'] = 'Completed' |
| self._send_records_to_master_helper( |
| jobs=jobs, hqes=hqes, exception_to_throw=None) |
| |
| # Check the entry was actually written to db |
| self.assertEqual(models.HostQueueEntry.objects.all()[0].status, |
| 'Completed') |
| |
| |
| def testSendingRecordsToMasterAbortedOnMaster(self): |
| """Send records to the master and ensure they are persisted.""" |
| jobs, hqes = self._get_records_for_sending_to_master() |
| hqes[0]['status'] = 'Completed' |
| self._send_records_to_master_helper( |
| jobs=jobs, hqes=hqes, exception_to_throw=None, aborted=True) |
| |
| # Check the entry was actually written to db |
| self.assertEqual(models.HostQueueEntry.objects.all()[0].status, |
| 'Completed') |
| |
| |
| def testSendingRecordsToMasterJobAssignedToDifferentShard(self): |
| """Ensure records belonging to different shard are silently rejected.""" |
| shard1 = models.Shard.objects.create(hostname='shard1') |
| shard2 = models.Shard.objects.create(hostname='shard2') |
| job1 = self._create_job(shard=shard1, control_file='foo1') |
| job2 = self._create_job(shard=shard2, control_file='foo2') |
| job1_id = job1.id |
| job2_id = job2.id |
| hqe1 = models.HostQueueEntry.objects.create(job=job1) |
| hqe2 = models.HostQueueEntry.objects.create(job=job2) |
| hqe1_id = hqe1.id |
| hqe2_id = hqe2.id |
| job1_record = job1.serialize(include_dependencies=False) |
| job2_record = job2.serialize(include_dependencies=False) |
| hqe1_record = hqe1.serialize(include_dependencies=False) |
| hqe2_record = hqe2.serialize(include_dependencies=False) |
| |
| # Prepare a bogus job record update from the wrong shard. The update |
| # should not throw an exception. Non-bogus jobs in the same update |
| # should happily update. |
| job1_record.update({'control_file': 'bar1'}) |
| job2_record.update({'control_file': 'bar2'}) |
| hqe1_record.update({'status': 'Aborted'}) |
| hqe2_record.update({'status': 'Aborted'}) |
| self._do_heartbeat_and_assert_response( |
| shard_hostname='shard2', upload_jobs=[job1_record, job2_record], |
| upload_hqes=[hqe1_record, hqe2_record]) |
| |
| # Job and HQE record for wrong job should not be modified, because the |
| # rpc came from the wrong shard. Job and HQE record for valid job are |
| # modified. |
| self.assertEqual(models.Job.objects.get(id=job1_id).control_file, |
| 'foo1') |
| self.assertEqual(models.Job.objects.get(id=job2_id).control_file, |
| 'bar2') |
| self.assertEqual(models.HostQueueEntry.objects.get(id=hqe1_id).status, |
| '') |
| self.assertEqual(models.HostQueueEntry.objects.get(id=hqe2_id).status, |
| 'Aborted') |
| |
| |
| def testSendingRecordsToMasterNotExistingJob(self): |
| """Ensure update for non existing job gets rejected.""" |
| jobs, hqes = self._get_records_for_sending_to_master() |
| jobs[0]['id'] = 3 |
| |
| self._send_records_to_master_helper( |
| jobs=jobs, hqes=hqes) |
| |
| |
| def _createShardAndHostWithLabel(self, shard_hostname='shard1', |
| host_hostname='host1', |
| label_name='board:lumpy'): |
| """Create a label, host, shard, and assign host to shard.""" |
| try: |
| label = models.Label.objects.create(name=label_name) |
| except: |
| label = models.Label.smart_get(label_name) |
| |
| shard = models.Shard.objects.create(hostname=shard_hostname) |
| shard.labels.add(label) |
| |
| host = models.Host.objects.create(hostname=host_hostname, leased=False, |
| shard=shard) |
| host.labels.add(label) |
| |
| return shard, host, label |
| |
| |
| def testShardLabelRemovalInvalid(self): |
| """Ensure you cannot remove the wrong label from shard.""" |
| shard1, host1, lumpy_label = self._createShardAndHostWithLabel() |
| stumpy_label = models.Label.objects.create( |
| name='board:stumpy', platform=True) |
| with self.assertRaises(error.RPCException): |
| rpc_interface.remove_board_from_shard( |
| shard1.hostname, stumpy_label.name) |
| |
| |
| def testShardHeartbeatLabelRemoval(self): |
| """Ensure label removal from shard works.""" |
| shard1, host1, lumpy_label = self._createShardAndHostWithLabel() |
| |
| self.assertEqual(host1.shard, shard1) |
| self.assertItemsEqual(shard1.labels.all(), [lumpy_label]) |
| rpc_interface.remove_board_from_shard( |
| shard1.hostname, lumpy_label.name) |
| host1 = models.Host.smart_get(host1.id) |
| shard1 = models.Shard.smart_get(shard1.id) |
| self.assertEqual(host1.shard, None) |
| self.assertItemsEqual(shard1.labels.all(), []) |
| |
| |
| def testCreateListShard(self): |
| """Retrieve a list of all shards.""" |
| lumpy_label = models.Label.objects.create(name='board:lumpy', |
| platform=True) |
| stumpy_label = models.Label.objects.create(name='board:stumpy', |
| platform=True) |
| peppy_label = models.Label.objects.create(name='board:peppy', |
| platform=True) |
| |
| shard_id = rpc_interface.add_shard( |
| hostname='host1', labels='board:lumpy,board:stumpy') |
| self.assertRaises(error.RPCException, |
| rpc_interface.add_shard, |
| hostname='host1', labels='board:lumpy,board:stumpy') |
| self.assertRaises(model_logic.ValidationError, |
| rpc_interface.add_shard, |
| hostname='host1', labels='board:peppy') |
| shard = models.Shard.objects.get(pk=shard_id) |
| self.assertEqual(shard.hostname, 'host1') |
| self.assertEqual(shard.labels.values_list('pk')[0], (lumpy_label.id,)) |
| self.assertEqual(shard.labels.values_list('pk')[1], (stumpy_label.id,)) |
| |
| self.assertEqual(rpc_interface.get_shards(), |
| [{'labels': ['board:lumpy','board:stumpy'], |
| 'hostname': 'host1', |
| 'id': 1}]) |
| |
| |
| def testAddBoardsToShard(self): |
| """Add boards to a given shard.""" |
| shard1, host1, lumpy_label = self._createShardAndHostWithLabel() |
| stumpy_label = models.Label.objects.create(name='board:stumpy', |
| platform=True) |
| shard_id = rpc_interface.add_board_to_shard( |
| hostname='shard1', labels='board:stumpy') |
| # Test whether raise exception when board label does not exist. |
| self.assertRaises(models.Label.DoesNotExist, |
| rpc_interface.add_board_to_shard, |
| hostname='shard1', labels='board:test') |
| # Test whether raise exception when board already sharded. |
| self.assertRaises(error.RPCException, |
| rpc_interface.add_board_to_shard, |
| hostname='shard1', labels='board:lumpy') |
| shard = models.Shard.objects.get(pk=shard_id) |
| self.assertEqual(shard.hostname, 'shard1') |
| self.assertEqual(shard.labels.values_list('pk')[0], (lumpy_label.id,)) |
| self.assertEqual(shard.labels.values_list('pk')[1], (stumpy_label.id,)) |
| |
| self.assertEqual(rpc_interface.get_shards(), |
| [{'labels': ['board:lumpy','board:stumpy'], |
| 'hostname': 'shard1', |
| 'id': 1}]) |
| |
| |
| def testShardHeartbeatFetchHostlessJob(self): |
| shard1, host1, label1 = self._createShardAndHostWithLabel() |
| self._testShardHeartbeatFetchHostlessJobHelper(host1) |
| |
| |
| def testShardHeartbeatIncorrectHosts(self): |
| shard1, host1, label1 = self._createShardAndHostWithLabel() |
| self._testShardHeartbeatIncorrectHostsHelper(host1) |
| |
| |
| def testShardHeartbeatLabelRemovalRace(self): |
| shard1, host1, label1 = self._createShardAndHostWithLabel() |
| self._testShardHeartbeatLabelRemovalRaceHelper(shard1, host1, label1) |
| |
| |
| def testShardRetrieveJobs(self): |
| shard1, host1, label1 = self._createShardAndHostWithLabel() |
| shard2, host2, label2 = self._createShardAndHostWithLabel( |
| 'shard2', 'host2', 'board:grumpy') |
| self._testShardRetrieveJobsHelper(shard1, host1, label1, |
| shard2, host2, label2) |
| |
| |
| def testResendJobsAfterFailedHeartbeat(self): |
| shard1, host1, label1 = self._createShardAndHostWithLabel() |
| self._testResendJobsAfterFailedHeartbeatHelper(shard1, host1, label1) |
| |
| |
| def testResendHostsAfterFailedHeartbeat(self): |
| shard1, host1, label1 = self._createShardAndHostWithLabel() |
| self._testResendHostsAfterFailedHeartbeatHelper(host1) |
| |
| |
| if __name__ == '__main__': |
| unittest.main() |