From 97a26de8e000ec129098bce61642db91e17242c8 Mon Sep 17 00:00:00 2001 From: armando-migliaccio Date: Mon, 3 Mar 2014 14:18:42 -0800 Subject: [PATCH] Fix NVP/Nicira nits Some references to NVP or Nicira went overlooked during the rebranding exercise. This patch kills the last ones. Implements blueprint nicira-plugin-renaming Change-Id: I9f5e4f38246adbd6f821310a0bcfc3ad40c3ad38 --- .../bigswitch/agent/restproxy_agent.py | 2 +- neutron/plugins/nicira/__init__.py | 2 +- .../ofagent/agent/ofa_neutron_agent.py | 2 +- .../plugins/vmware/common/securitygroups.py | 4 +- neutron/plugins/vmware/common/sync.py | 94 +++++++++---------- neutron/plugins/vmware/dbexts/networkgw_db.py | 2 +- neutron/plugins/vmware/nsxlib/__init__.py | 4 +- neutron/plugins/vmware/plugins/base.py | 8 +- neutron/plugins/vmware/plugins/service.py | 20 ++-- neutron/tests/unit/vmware/test_nsx_sync.py | 6 +- neutron/tests/unit/vmware/test_nsx_utils.py | 6 +- 11 files changed, 74 insertions(+), 76 deletions(-) diff --git a/neutron/plugins/bigswitch/agent/restproxy_agent.py b/neutron/plugins/bigswitch/agent/restproxy_agent.py index 81d3032d5..91d314055 100644 --- a/neutron/plugins/bigswitch/agent/restproxy_agent.py +++ b/neutron/plugins/bigswitch/agent/restproxy_agent.py @@ -1,7 +1,7 @@ # Copyright 2014 Big Switch Networks, Inc. # All Rights Reserved. # -# Copyright 2011 Nicira Networks, Inc. +# Copyright 2011 VMware, Inc. # All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); you may diff --git a/neutron/plugins/nicira/__init__.py b/neutron/plugins/nicira/__init__.py index 7b94a1840..81f4e1b35 100644 --- a/neutron/plugins/nicira/__init__.py +++ b/neutron/plugins/nicira/__init__.py @@ -1,4 +1,4 @@ -# Copyright 2013 VMware, Inc. +# Copyright 2012 VMware, Inc. # # All Rights Reserved # diff --git a/neutron/plugins/ofagent/agent/ofa_neutron_agent.py b/neutron/plugins/ofagent/agent/ofa_neutron_agent.py index d428be2ab..7c5beee17 100644 --- a/neutron/plugins/ofagent/agent/ofa_neutron_agent.py +++ b/neutron/plugins/ofagent/agent/ofa_neutron_agent.py @@ -1,7 +1,7 @@ # Copyright (C) 2014 VA Linux Systems Japan K.K. # Based on openvswitch agent. # -# Copyright 2011 Nicira Networks, Inc. +# Copyright 2011 VMware, Inc. # All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); you may diff --git a/neutron/plugins/vmware/common/securitygroups.py b/neutron/plugins/vmware/common/securitygroups.py index f00634f79..db61b72a8 100644 --- a/neutron/plugins/vmware/common/securitygroups.py +++ b/neutron/plugins/vmware/common/securitygroups.py @@ -120,7 +120,7 @@ def remove_security_group_with_id_and_id_field(rules, rule_id): This function receives all of the current rule associated with a security group and then removes the rule that matches the rule_id. In addition it removes the id field in the dict with each rule since that - should not be passed to nvp. + should not be passed to nsx. """ for rule_direction in rules.values(): item_to_remove = None @@ -128,7 +128,7 @@ def remove_security_group_with_id_and_id_field(rules, rule_id): if port_rule['id'] == rule_id: item_to_remove = port_rule else: - # remove key from dictionary for NVP + # remove key from dictionary for NSX del port_rule['id'] if item_to_remove: rule_direction.remove(item_to_remove) diff --git a/neutron/plugins/vmware/common/sync.py b/neutron/plugins/vmware/common/sync.py index 0a57f0c21..229ff3427 100644 --- a/neutron/plugins/vmware/common/sync.py +++ b/neutron/plugins/vmware/common/sync.py @@ -40,8 +40,8 @@ MAX_PAGE_SIZE = 5000 LOG = log.getLogger(__name__) -class NvpCache(object): - """A simple Cache for NVP resources. +class NsxCache(object): + """A simple Cache for NSX resources. Associates resource id with resource hash to rapidly identify updated resources. @@ -57,7 +57,7 @@ class NvpCache(object): def __init__(self): # Maps a uuid to the dict containing it self._uuid_dict_mappings = {} - # Dicts for NVP cached resources + # Dicts for NSX cached resources self._lswitches = {} self._lswitchports = {} self._lrouters = {} @@ -74,7 +74,7 @@ class NvpCache(object): # Clear the 'changed' attribute for all items for uuid, item in resources.items(): if item.pop('changed', None) and not item.get('data'): - # The item is not anymore in NVP, so delete it + # The item is not anymore in NSX, so delete it del resources[uuid] del self._uuid_dict_mappings[uuid] @@ -192,7 +192,7 @@ def _start_loopingcall(min_chunk_size, state_sync_interval, func): return state_synchronizer -class NvpSynchronizer(): +class NsxSynchronizer(): LS_URI = nsxlib._build_uri_path( switchlib.LSWITCH_RESOURCE, fields='uuid,tags,fabric_status', @@ -209,7 +209,7 @@ class NvpSynchronizer(): def __init__(self, plugin, cluster, state_sync_interval, req_delay, min_chunk_size, max_rand_delay=0): random.seed() - self._nvp_cache = NvpCache() + self._nsx_cache = NsxCache() # Store parameters as instance members # NOTE(salv-orlando): apologies if it looks java-ish self._plugin = plugin @@ -248,13 +248,13 @@ class NvpSynchronizer(): def synchronize_network(self, context, neutron_network_data, lswitches=None): - """Synchronize a Neutron network with its NVP counterpart. + """Synchronize a Neutron network with its NSX counterpart. This routine synchronizes a set of switches when a Neutron network is mapped to multiple lswitches. """ if not lswitches: - # Try to get logical switches from nvp + # Try to get logical switches from nsx try: lswitches = nsx_utils.fetch_nsx_switches( context.session, self._cluster, @@ -264,11 +264,11 @@ class NvpSynchronizer(): # api_exc.ResourceNotFound here # The logical switch was not found LOG.warning(_("Logical switch for neutron network %s not " - "found on NVP."), neutron_network_data['id']) + "found on NSX."), neutron_network_data['id']) lswitches = [] else: for lswitch in lswitches: - self._nvp_cache.update_lswitch(lswitch) + self._nsx_cache.update_lswitch(lswitch) # By default assume things go wrong status = constants.NET_STATUS_ERROR # In most cases lswitches will contain a single element @@ -292,19 +292,19 @@ class NvpSynchronizer(): if not ls_uuids and not scan_missing: return neutron_net_ids = set() - neutron_nvp_mappings = {} + neutron_nsx_mappings = {} # TODO(salvatore-orlando): Deal with the case the tag # has been tampered with for ls_uuid in ls_uuids: # If the lswitch has been deleted, get backup copy of data - lswitch = (self._nvp_cache[ls_uuid].get('data') or - self._nvp_cache[ls_uuid].get('data_bk')) + lswitch = (self._nsx_cache[ls_uuid].get('data') or + self._nsx_cache[ls_uuid].get('data_bk')) tags = self._get_tag_dict(lswitch['tags']) neutron_id = tags.get('quantum_net_id') neutron_net_ids.add(neutron_id) - neutron_nvp_mappings[neutron_id] = ( - neutron_nvp_mappings.get(neutron_id, []) + - [self._nvp_cache[ls_uuid]]) + neutron_nsx_mappings[neutron_id] = ( + neutron_nsx_mappings.get(neutron_id, []) + + [self._nsx_cache[ls_uuid]]) # Fetch neutron networks from database filters = {'router:external': [False]} if not scan_missing: @@ -312,15 +312,15 @@ class NvpSynchronizer(): networks = self._plugin._get_collection_query( ctx, models_v2.Network, filters=filters) for network in networks: - lswitches = neutron_nvp_mappings.get(network['id'], []) + lswitches = neutron_nsx_mappings.get(network['id'], []) lswitches = [lswitch.get('data') for lswitch in lswitches] self.synchronize_network(ctx, network, lswitches) def synchronize_router(self, context, neutron_router_data, lrouter=None): - """Synchronize a neutron router with its NVP counterpart.""" + """Synchronize a neutron router with its NSX counterpart.""" if not lrouter: - # Try to get router from nvp + # Try to get router from nsx try: # This query will return the logical router status too nsx_router_id = nsx_utils.get_nsx_router_id( @@ -332,14 +332,14 @@ class NvpSynchronizer(): # api_exc.ResourceNotFound here # The logical router was not found LOG.warning(_("Logical router for neutron router %s not " - "found on NVP."), neutron_router_data['id']) + "found on NSX."), neutron_router_data['id']) lrouter = None else: # Update the cache - self._nvp_cache.update_lrouter(lrouter) + self._nsx_cache.update_lrouter(lrouter) # Note(salv-orlando): It might worth adding a check to verify neutron - # resource tag in nvp entity matches a Neutron id. + # resource tag in nsx entity matches a Neutron id. # By default assume things go wrong status = constants.NET_STATUS_ERROR if lrouter: @@ -359,13 +359,13 @@ class NvpSynchronizer(): # has been tampered with neutron_router_mappings = {} for lr_uuid in lr_uuids: - lrouter = (self._nvp_cache[lr_uuid].get('data') or - self._nvp_cache[lr_uuid].get('data_bk')) + lrouter = (self._nsx_cache[lr_uuid].get('data') or + self._nsx_cache[lr_uuid].get('data_bk')) tags = self._get_tag_dict(lrouter['tags']) neutron_router_id = tags.get('q_router_id') if neutron_router_id: neutron_router_mappings[neutron_router_id] = ( - self._nvp_cache[lr_uuid]) + self._nsx_cache[lr_uuid]) else: LOG.warn(_("Unable to find Neutron router id for " "NSX logical router: %s"), lr_uuid) @@ -381,7 +381,7 @@ class NvpSynchronizer(): def synchronize_port(self, context, neutron_port_data, lswitchport=None, ext_networks=None): - """Synchronize a Neutron port with its NVP counterpart.""" + """Synchronize a Neutron port with its NSX counterpart.""" # Skip synchronization for ports on external networks if not ext_networks: ext_networks = [net['id'] for net in context.session.query( @@ -395,7 +395,7 @@ class NvpSynchronizer(): return if not lswitchport: - # Try to get port from nvp + # Try to get port from nsx try: ls_uuid, lp_uuid = nsx_utils.get_nsx_switch_and_port_id( context.session, self._cluster, neutron_port_data['id']) @@ -409,15 +409,15 @@ class NvpSynchronizer(): # of PortNotFoundOnNetwork when the id exists but # the logical switch port was not found LOG.warning(_("Logical switch port for neutron port %s " - "not found on NVP."), neutron_port_data['id']) + "not found on NSX."), neutron_port_data['id']) lswitchport = None else: # If lswitchport is not None, update the cache. # It could be none if the port was deleted from the backend if lswitchport: - self._nvp_cache.update_lswitchport(lswitchport) + self._nsx_cache.update_lswitchport(lswitchport) # Note(salv-orlando): It might worth adding a check to verify neutron - # resource tag in nvp entity matches Neutron id. + # resource tag in nsx entity matches Neutron id. # By default assume things go wrong status = constants.PORT_STATUS_ERROR if lswitchport: @@ -439,13 +439,13 @@ class NvpSynchronizer(): # has been tampered with neutron_port_mappings = {} for lp_uuid in lp_uuids: - lport = (self._nvp_cache[lp_uuid].get('data') or - self._nvp_cache[lp_uuid].get('data_bk')) + lport = (self._nsx_cache[lp_uuid].get('data') or + self._nsx_cache[lp_uuid].get('data_bk')) tags = self._get_tag_dict(lport['tags']) neutron_port_id = tags.get('q_port_id') if neutron_port_id: neutron_port_mappings[neutron_port_id] = ( - self._nvp_cache[lp_uuid]) + self._nsx_cache[lp_uuid]) # Fetch neutron ports from database # At the first sync we need to fetch all ports filters = ({} if scan_missing else @@ -478,7 +478,7 @@ class NvpSynchronizer(): if cursor == 'start': cursor = None # Chunk size tuning might, in some conditions, make it larger - # than 5,000, which is the maximum page size allowed by the NVP + # than 5,000, which is the maximum page size allowed by the NSX # API. In this case the request should be split in multiple # requests. This is not ideal, and therefore a log warning will # be emitted. @@ -486,7 +486,7 @@ class NvpSynchronizer(): if num_requests > 1: LOG.warn(_("Requested page size is %(cur_chunk_size)d." "It might be necessary to do %(num_requests)d " - "round-trips to NVP for fetching data. Please " + "round-trips to NSX for fetching data. Please " "tune sync parameters to ensure chunk size " "is less than %(max_page_size)d"), {'cur_chunk_size': page_size, @@ -514,11 +514,11 @@ class NvpSynchronizer(): return results, cursor if page_size else 'start', total_size return [], cursor, None - def _fetch_nvp_data_chunk(self, sp): + def _fetch_nsx_data_chunk(self, sp): base_chunk_size = sp.chunk_size chunk_size = base_chunk_size + sp.extra_chunk_size LOG.info(_("Fetching up to %s resources " - "from NVP backend"), chunk_size) + "from NSX backend"), chunk_size) fetched = ls_count = lr_count = lp_count = 0 lswitches = lrouters = lswitchports = [] if sp.ls_cursor or sp.ls_cursor == 'start': @@ -559,19 +559,19 @@ class NvpSynchronizer(): sp.ls_cursor = sp.lr_cursor = sp.lp_cursor = 'start' LOG.info(_("Running state synchronization task. Chunk: %s"), sp.current_chunk) - # Fetch chunk_size data from NVP + # Fetch chunk_size data from NSX try: (lswitches, lrouters, lswitchports) = ( - self._fetch_nvp_data_chunk(sp)) + self._fetch_nsx_data_chunk(sp)) except (api_exc.RequestTimeout, api_exc.NsxApiException): sleep_interval = self._sync_backoff # Cap max back off to 64 seconds self._sync_backoff = min(self._sync_backoff * 2, 64) LOG.exception(_("An error occurred while communicating with " - "NVP backend. Will retry synchronization " + "NSX backend. Will retry synchronization " "in %d seconds"), sleep_interval) return sleep_interval - LOG.debug(_("Time elapsed querying NVP: %s"), + LOG.debug(_("Time elapsed querying NSX: %s"), timeutils.utcnow() - start) if sp.total_size: num_chunks = ((sp.total_size / sp.chunk_size) + @@ -579,20 +579,20 @@ class NvpSynchronizer(): else: num_chunks = 1 LOG.debug(_("Number of chunks: %d"), num_chunks) - # Find objects which have changed on NVP side and need + # Find objects which have changed on NSX side and need # to be synchronized - (ls_uuids, lr_uuids, lp_uuids) = self._nvp_cache.process_updates( + (ls_uuids, lr_uuids, lp_uuids) = self._nsx_cache.process_updates( lswitches, lrouters, lswitchports) # Process removed objects only at the last chunk scan_missing = (sp.current_chunk == num_chunks - 1 and not sp.init_sync_performed) if sp.current_chunk == num_chunks - 1: - self._nvp_cache.process_deletes() - ls_uuids = self._nvp_cache.get_lswitches( + self._nsx_cache.process_deletes() + ls_uuids = self._nsx_cache.get_lswitches( changed_only=not scan_missing) - lr_uuids = self._nvp_cache.get_lrouters( + lr_uuids = self._nsx_cache.get_lrouters( changed_only=not scan_missing) - lp_uuids = self._nvp_cache.get_lswitchports( + lp_uuids = self._nsx_cache.get_lswitchports( changed_only=not scan_missing) LOG.debug(_("Time elapsed hashing data: %s"), timeutils.utcnow() - start) diff --git a/neutron/plugins/vmware/dbexts/networkgw_db.py b/neutron/plugins/vmware/dbexts/networkgw_db.py index ed4b0e87c..58d14c556 100644 --- a/neutron/plugins/vmware/dbexts/networkgw_db.py +++ b/neutron/plugins/vmware/dbexts/networkgw_db.py @@ -363,7 +363,7 @@ class NetworkGatewayMixin(networkgw.NetworkGatewayPluginBase): raise MultipleGatewayConnections( gateway_id=network_gateway_id) # Remove gateway port from network - # FIXME(salvatore-orlando): Ensure state of port in NVP is + # FIXME(salvatore-orlando): Ensure state of port in NSX is # consistent with outcome of transaction self.delete_port(context, net_connection['port_id'], nw_gw_port_check=False) diff --git a/neutron/plugins/vmware/nsxlib/__init__.py b/neutron/plugins/vmware/nsxlib/__init__.py index 8f5bd4830..b4f46e10c 100644 --- a/neutron/plugins/vmware/nsxlib/__init__.py +++ b/neutron/plugins/vmware/nsxlib/__init__.py @@ -1,5 +1,3 @@ -# vim: tabstop=4 shiftwidth=4 softtabstop=4 - # Copyright 2014 VMware, Inc. # All Rights Reserved # @@ -27,7 +25,7 @@ HTTP_GET = "GET" HTTP_POST = "POST" HTTP_DELETE = "DELETE" HTTP_PUT = "PUT" -# Prefix to be used for all NVP API calls +# Prefix to be used for all NSX API calls URI_PREFIX = "/ws.v1" NEUTRON_VERSION = version_info.release_string() diff --git a/neutron/plugins/vmware/plugins/base.py b/neutron/plugins/vmware/plugins/base.py index 096faf1c4..6ad3293f5 100644 --- a/neutron/plugins/vmware/plugins/base.py +++ b/neutron/plugins/vmware/plugins/base.py @@ -132,7 +132,7 @@ class NsxPluginV2(addr_pair_db.AllowedAddressPairsMixin, novazone_cluster_map = {} def __init__(self): - super(NvpPluginV2, self).__init__() + super(NsxPluginV2, self).__init__() # TODO(salv-orlando): Replace These dicts with # collections.defaultdict for better handling of default values # Routines for managing logical ports in NSX @@ -180,7 +180,7 @@ class NsxPluginV2(addr_pair_db.AllowedAddressPairsMixin, # been yet updated from the config file self._is_default_net_gw_in_sync = False # Create a synchronizer instance for backend sync - self._synchronizer = sync.NvpSynchronizer( + self._synchronizer = sync.NsxSynchronizer( self, self.cluster, self.nsx_sync_opts.state_sync_interval, self.nsx_sync_opts.min_sync_req_delay, @@ -2104,7 +2104,7 @@ class NsxPluginV2(addr_pair_db.AllowedAddressPairsMixin, self._ensure_default_security_group(context, tenant_id) # NOTE(salv-orlando): Pre-generating Neutron ID for security group. neutron_id = str(uuid.uuid4()) - nvp_secgroup = secgrouplib.create_security_profile( + nsx_secgroup = secgrouplib.create_security_profile( self.cluster, neutron_id, tenant_id, s) with context.session.begin(subtransactions=True): s['id'] = neutron_id @@ -2113,7 +2113,7 @@ class NsxPluginV2(addr_pair_db.AllowedAddressPairsMixin, context.session.flush() # Add mapping between neutron and nsx identifiers nsx_db.add_neutron_nsx_security_group_mapping( - context.session, neutron_id, nvp_secgroup['uuid']) + context.session, neutron_id, nsx_secgroup['uuid']) return sec_group def delete_security_group(self, context, security_group_id): diff --git a/neutron/plugins/vmware/plugins/service.py b/neutron/plugins/vmware/plugins/service.py index 4ed362cfd..84236d582 100644 --- a/neutron/plugins/vmware/plugins/service.py +++ b/neutron/plugins/vmware/plugins/service.py @@ -941,7 +941,7 @@ class NsxAdvancedPlugin(sr_db.ServiceRouter_mixin, context, id, firewall_db.Firewall) def get_firewall(self, context, id, fields=None): - fw = super(NvpAdvancedPlugin, self).get_firewall( + fw = super(NsxAdvancedPlugin, self).get_firewall( context, id, fields) if fields and rsi.ROUTER_ID not in fields: return fw @@ -952,7 +952,7 @@ class NsxAdvancedPlugin(sr_db.ServiceRouter_mixin, return fw def get_firewalls(self, context, filters=None, fields=None): - fws = super(NvpAdvancedPlugin, self).get_firewalls( + fws = super(NsxAdvancedPlugin, self).get_firewalls( context, filters, fields) if fields and rsi.ROUTER_ID not in fields: return fws @@ -1303,7 +1303,7 @@ class NsxAdvancedPlugin(sr_db.ServiceRouter_mixin, self._update_interface(context, router, sync=True) def get_vip(self, context, id, fields=None): - vip = super(NvpAdvancedPlugin, self).get_vip(context, id, fields) + vip = super(NsxAdvancedPlugin, self).get_vip(context, id, fields) if fields and rsi.ROUTER_ID not in fields: return vip @@ -1313,7 +1313,7 @@ class NsxAdvancedPlugin(sr_db.ServiceRouter_mixin, return vip def get_vips(self, context, filters=None, fields=None): - vips = super(NvpAdvancedPlugin, self).get_vips( + vips = super(NsxAdvancedPlugin, self).get_vips( context, filters, fields) if fields and rsi.ROUTER_ID not in fields: return vips @@ -1629,7 +1629,7 @@ class NsxAdvancedPlugin(sr_db.ServiceRouter_mixin, raise nsx_exc.ServiceOverQuota( overs='vpnservice', err_msg=msg) - service = super(NvpAdvancedPlugin, self).create_vpnservice( + service = super(NsxAdvancedPlugin, self).create_vpnservice( context, vpnservice) self._resource_set_status( context, vpn_db.VPNService, @@ -1638,7 +1638,7 @@ class NsxAdvancedPlugin(sr_db.ServiceRouter_mixin, def update_vpnservice(self, context, vpnservice_id, vpnservice): vpnservice['vpnservice']['status'] = service_constants.PENDING_UPDATE - service = super(NvpAdvancedPlugin, self).update_vpnservice( + service = super(NsxAdvancedPlugin, self).update_vpnservice( context, vpnservice_id, vpnservice) # Only admin_state_up attribute is configurable on Edge. if vpnservice['vpnservice'].get('admin_state_up') is None: @@ -1669,14 +1669,14 @@ class NsxAdvancedPlugin(sr_db.ServiceRouter_mixin, def create_ipsec_site_connection(self, context, ipsec_site_connection): ipsec_site_conn = super( - NvpAdvancedPlugin, self).create_ipsec_site_connection( + NsxAdvancedPlugin, self).create_ipsec_site_connection( context, ipsec_site_connection) try: self._vcns_update_ipsec_config( context, ipsec_site_conn['vpnservice_id']) except Exception: with excutils.save_and_reraise_exception(): - super(NvpAdvancedPlugin, self).delete_ipsec_site_connection( + super(NsxAdvancedPlugin, self).delete_ipsec_site_connection( context, ipsec_site_conn['id']) self._resource_set_status( context, vpn_db.IPsecSiteConnection, @@ -1688,7 +1688,7 @@ class NsxAdvancedPlugin(sr_db.ServiceRouter_mixin, ipsec_site_connection['ipsec_site_connection']['status'] = ( service_constants.PENDING_UPDATE) ipsec_site_conn = super( - NvpAdvancedPlugin, self).update_ipsec_site_connection( + NsxAdvancedPlugin, self).update_ipsec_site_connection( context, ipsec_site_connection_id, ipsec_site_connection) try: self._vcns_update_ipsec_config( @@ -1717,7 +1717,7 @@ class NsxAdvancedPlugin(sr_db.ServiceRouter_mixin, self._resource_set_status( context, vpn_db.IPsecSiteConnection, ipsec_site_conn_id, service_constants.ERROR) - super(NvpAdvancedPlugin, self).delete_ipsec_site_connection( + super(NsxAdvancedPlugin, self).delete_ipsec_site_connection( context, ipsec_site_conn_id) diff --git a/neutron/tests/unit/vmware/test_nsx_sync.py b/neutron/tests/unit/vmware/test_nsx_sync.py index cd4fc1d53..50634f771 100644 --- a/neutron/tests/unit/vmware/test_nsx_sync.py +++ b/neutron/tests/unit/vmware/test_nsx_sync.py @@ -55,7 +55,7 @@ class CacheTestCase(base.BaseTestCase): """Test suite providing coverage for the Cache class.""" def setUp(self): - self.nsx_cache = sync.NvpCache() + self.nsx_cache = sync.NsxCache() for lswitch in LSWITCHES: self.nsx_cache._uuid_dict_mappings[lswitch['uuid']] = ( self.nsx_cache._lswitches) @@ -242,8 +242,8 @@ class SyncLoopingCallTestCase(base.BaseTestCase): # Avoid runs of the synchronization process - just start # the looping call with mock.patch.object( - sync.NvpSynchronizer, '_synchronize_state', return_value=0.01): - synchronizer = sync.NvpSynchronizer(mock.ANY, mock.ANY, + sync.NsxSynchronizer, '_synchronize_state', return_value=0.01): + synchronizer = sync.NsxSynchronizer(mock.ANY, mock.ANY, 100, 0, 0) time.sleep(0.03) # stop looping call before asserting diff --git a/neutron/tests/unit/vmware/test_nsx_utils.py b/neutron/tests/unit/vmware/test_nsx_utils.py index 3b8dc4fd5..a35b32796 100644 --- a/neutron/tests/unit/vmware/test_nsx_utils.py +++ b/neutron/tests/unit/vmware/test_nsx_utils.py @@ -84,7 +84,7 @@ class NsxUtilsTestCase(base.BaseTestCase): self.assertFalse(exp_ls_uuids) def _verify_get_nsx_router_id(self, exp_lr_uuid): - # The nvplib and db calls are mocked, therefore the cluster + # The nsxlib and db calls are mocked, therefore the cluster # and the neutron_router_id parameters can be set to None lr_uuid = nsx_utils.get_nsx_router_id(db_api.get_session(), None, None) self.assertEqual(exp_lr_uuid, lr_uuid) @@ -99,7 +99,7 @@ class NsxUtilsTestCase(base.BaseTestCase): self._verify_get_nsx_switch_and_port_id(exp_ls_uuid, exp_lp_uuid) def test_get_nsx_switch_and_port_id_only_port_db_mapping(self): - # This test is representative of the case in which a port with a nvp + # This test is representative of the case in which a port with a nsx # db mapping in the havana db was upgraded to icehouse exp_ls_uuid = uuidutils.generate_uuid() exp_lp_uuid = uuidutils.generate_uuid() @@ -279,7 +279,7 @@ class NsxUtilsTestCase(base.BaseTestCase): self.addCleanup(mock.patch.stopall) def _verify_get_nsx_sec_profile_id(self, exp_sec_prof_uuid): - # The nvplib and db calls are mocked, therefore the cluster + # The nsxlib and db calls are mocked, therefore the cluster # and the neutron_id parameters can be set to None sec_prof_uuid = nsx_utils.get_nsx_security_group_id( db_api.get_session(), None, None) -- 2.45.2