]> review.fuel-infra Code Review - openstack-build/neutron-build.git/commitdiff
Fix NVP/Nicira nits
authorarmando-migliaccio <armamig@gmail.com>
Mon, 3 Mar 2014 22:18:42 +0000 (14:18 -0800)
committerThomas Goirand <thomas@goirand.fr>
Thu, 13 Mar 2014 07:20:42 +0000 (15:20 +0800)
Some references to NVP or Nicira went overlooked during
the rebranding exercise. This patch kills the last ones.

Implements blueprint nicira-plugin-renaming

Change-Id: I9f5e4f38246adbd6f821310a0bcfc3ad40c3ad38

neutron/plugins/bigswitch/agent/restproxy_agent.py
neutron/plugins/nicira/__init__.py
neutron/plugins/ofagent/agent/ofa_neutron_agent.py
neutron/plugins/vmware/common/securitygroups.py
neutron/plugins/vmware/common/sync.py
neutron/plugins/vmware/dbexts/networkgw_db.py
neutron/plugins/vmware/nsxlib/__init__.py
neutron/plugins/vmware/plugins/base.py
neutron/plugins/vmware/plugins/service.py
neutron/tests/unit/vmware/test_nsx_sync.py
neutron/tests/unit/vmware/test_nsx_utils.py

index 81d3032d51f7d42d651ed9201254944561d11689..91d314055d7cd6a0fe92eaa54a8d74d2240330e7 100644 (file)
@@ -1,7 +1,7 @@
 # Copyright 2014 Big Switch Networks, Inc.
 # All Rights Reserved.
 #
-# Copyright 2011 Nicira Networks, Inc.
+# Copyright 2011 VMware, Inc.
 # All Rights Reserved.
 #
 #    Licensed under the Apache License, Version 2.0 (the "License"); you may
index 7b94a1840e8329c9fe41cbe0cc2cd46d6cd61bd2..81f4e1b35459c9f676e79d02cec9796dc5d6f06e 100644 (file)
@@ -1,4 +1,4 @@
-# Copyright 2013 VMware, Inc.
+# Copyright 2012 VMware, Inc.
 #
 # All Rights Reserved
 #
index d428be2ab6439a5164b3dfa312654237fd371e4e..7c5beee176746d97ac49879a5d112a6307d82a63 100644 (file)
@@ -1,7 +1,7 @@
 # Copyright (C) 2014 VA Linux Systems Japan K.K.
 # Based on openvswitch agent.
 #
-# Copyright 2011 Nicira Networks, Inc.
+# Copyright 2011 VMware, Inc.
 # All Rights Reserved.
 #
 #    Licensed under the Apache License, Version 2.0 (the "License"); you may
index f00634f79de282bf967cbe684cbf99712f11129c..db61b72a851beed5332e6064003fa54f1f43b852 100644 (file)
@@ -120,7 +120,7 @@ def remove_security_group_with_id_and_id_field(rules, rule_id):
     This function receives all of the current rule associated with a
     security group and then removes the rule that matches the rule_id. In
     addition it removes the id field in the dict with each rule since that
-    should not be passed to nvp.
+    should not be passed to nsx.
     """
     for rule_direction in rules.values():
         item_to_remove = None
@@ -128,7 +128,7 @@ def remove_security_group_with_id_and_id_field(rules, rule_id):
             if port_rule['id'] == rule_id:
                 item_to_remove = port_rule
             else:
-                # remove key from dictionary for NVP
+                # remove key from dictionary for NSX
                 del port_rule['id']
         if item_to_remove:
             rule_direction.remove(item_to_remove)
index 0a57f0c2115a434ee5f0fd2f87a2e55935a9bd2f..229ff342713275365b0bcb944bff1cee8702bd25 100644 (file)
@@ -40,8 +40,8 @@ MAX_PAGE_SIZE = 5000
 LOG = log.getLogger(__name__)
 
 
-class NvpCache(object):
-    """A simple Cache for NVP resources.
+class NsxCache(object):
+    """A simple Cache for NSX resources.
 
     Associates resource id with resource hash to rapidly identify
     updated resources.
@@ -57,7 +57,7 @@ class NvpCache(object):
     def __init__(self):
         # Maps a uuid to the dict containing it
         self._uuid_dict_mappings = {}
-        # Dicts for NVP cached resources
+        # Dicts for NSX cached resources
         self._lswitches = {}
         self._lswitchports = {}
         self._lrouters = {}
@@ -74,7 +74,7 @@ class NvpCache(object):
         # Clear the 'changed' attribute for all items
         for uuid, item in resources.items():
             if item.pop('changed', None) and not item.get('data'):
-                # The item is not anymore in NVP, so delete it
+                # The item is not anymore in NSX, so delete it
                 del resources[uuid]
                 del self._uuid_dict_mappings[uuid]
 
@@ -192,7 +192,7 @@ def _start_loopingcall(min_chunk_size, state_sync_interval, func):
     return state_synchronizer
 
 
-class NvpSynchronizer():
+class NsxSynchronizer():
 
     LS_URI = nsxlib._build_uri_path(
         switchlib.LSWITCH_RESOURCE, fields='uuid,tags,fabric_status',
@@ -209,7 +209,7 @@ class NvpSynchronizer():
     def __init__(self, plugin, cluster, state_sync_interval,
                  req_delay, min_chunk_size, max_rand_delay=0):
         random.seed()
-        self._nvp_cache = NvpCache()
+        self._nsx_cache = NsxCache()
         # Store parameters as instance members
         # NOTE(salv-orlando): apologies if it looks java-ish
         self._plugin = plugin
@@ -248,13 +248,13 @@ class NvpSynchronizer():
 
     def synchronize_network(self, context, neutron_network_data,
                             lswitches=None):
-        """Synchronize a Neutron network with its NVP counterpart.
+        """Synchronize a Neutron network with its NSX counterpart.
 
         This routine synchronizes a set of switches when a Neutron
         network is mapped to multiple lswitches.
         """
         if not lswitches:
-            # Try to get logical switches from nvp
+            # Try to get logical switches from nsx
             try:
                 lswitches = nsx_utils.fetch_nsx_switches(
                     context.session, self._cluster,
@@ -264,11 +264,11 @@ class NvpSynchronizer():
                 # api_exc.ResourceNotFound here
                 # The logical switch was not found
                 LOG.warning(_("Logical switch for neutron network %s not "
-                              "found on NVP."), neutron_network_data['id'])
+                              "found on NSX."), neutron_network_data['id'])
                 lswitches = []
             else:
                 for lswitch in lswitches:
-                    self._nvp_cache.update_lswitch(lswitch)
+                    self._nsx_cache.update_lswitch(lswitch)
         # By default assume things go wrong
         status = constants.NET_STATUS_ERROR
         # In most cases lswitches will contain a single element
@@ -292,19 +292,19 @@ class NvpSynchronizer():
         if not ls_uuids and not scan_missing:
             return
         neutron_net_ids = set()
-        neutron_nvp_mappings = {}
+        neutron_nsx_mappings = {}
         # TODO(salvatore-orlando): Deal with the case the tag
         # has been tampered with
         for ls_uuid in ls_uuids:
             # If the lswitch has been deleted, get backup copy of data
-            lswitch = (self._nvp_cache[ls_uuid].get('data') or
-                       self._nvp_cache[ls_uuid].get('data_bk'))
+            lswitch = (self._nsx_cache[ls_uuid].get('data') or
+                       self._nsx_cache[ls_uuid].get('data_bk'))
             tags = self._get_tag_dict(lswitch['tags'])
             neutron_id = tags.get('quantum_net_id')
             neutron_net_ids.add(neutron_id)
-            neutron_nvp_mappings[neutron_id] = (
-                neutron_nvp_mappings.get(neutron_id, []) +
-                [self._nvp_cache[ls_uuid]])
+            neutron_nsx_mappings[neutron_id] = (
+                neutron_nsx_mappings.get(neutron_id, []) +
+                [self._nsx_cache[ls_uuid]])
         # Fetch neutron networks from database
         filters = {'router:external': [False]}
         if not scan_missing:
@@ -312,15 +312,15 @@ class NvpSynchronizer():
         networks = self._plugin._get_collection_query(
             ctx, models_v2.Network, filters=filters)
         for network in networks:
-            lswitches = neutron_nvp_mappings.get(network['id'], [])
+            lswitches = neutron_nsx_mappings.get(network['id'], [])
             lswitches = [lswitch.get('data') for lswitch in lswitches]
             self.synchronize_network(ctx, network, lswitches)
 
     def synchronize_router(self, context, neutron_router_data,
                            lrouter=None):
-        """Synchronize a neutron router with its NVP counterpart."""
+        """Synchronize a neutron router with its NSX counterpart."""
         if not lrouter:
-            # Try to get router from nvp
+            # Try to get router from nsx
             try:
                 # This query will return the logical router status too
                 nsx_router_id = nsx_utils.get_nsx_router_id(
@@ -332,14 +332,14 @@ class NvpSynchronizer():
                 # api_exc.ResourceNotFound here
                 # The logical router was not found
                 LOG.warning(_("Logical router for neutron router %s not "
-                              "found on NVP."), neutron_router_data['id'])
+                              "found on NSX."), neutron_router_data['id'])
                 lrouter = None
             else:
                 # Update the cache
-                self._nvp_cache.update_lrouter(lrouter)
+                self._nsx_cache.update_lrouter(lrouter)
 
         # Note(salv-orlando): It might worth adding a check to verify neutron
-        # resource tag in nvp entity matches a Neutron id.
+        # resource tag in nsx entity matches a Neutron id.
         # By default assume things go wrong
         status = constants.NET_STATUS_ERROR
         if lrouter:
@@ -359,13 +359,13 @@ class NvpSynchronizer():
         # has been tampered with
         neutron_router_mappings = {}
         for lr_uuid in lr_uuids:
-            lrouter = (self._nvp_cache[lr_uuid].get('data') or
-                       self._nvp_cache[lr_uuid].get('data_bk'))
+            lrouter = (self._nsx_cache[lr_uuid].get('data') or
+                       self._nsx_cache[lr_uuid].get('data_bk'))
             tags = self._get_tag_dict(lrouter['tags'])
             neutron_router_id = tags.get('q_router_id')
             if neutron_router_id:
                 neutron_router_mappings[neutron_router_id] = (
-                    self._nvp_cache[lr_uuid])
+                    self._nsx_cache[lr_uuid])
             else:
                 LOG.warn(_("Unable to find Neutron router id for "
                            "NSX logical router: %s"), lr_uuid)
@@ -381,7 +381,7 @@ class NvpSynchronizer():
 
     def synchronize_port(self, context, neutron_port_data,
                          lswitchport=None, ext_networks=None):
-        """Synchronize a Neutron port with its NVP counterpart."""
+        """Synchronize a Neutron port with its NSX counterpart."""
         # Skip synchronization for ports on external networks
         if not ext_networks:
             ext_networks = [net['id'] for net in context.session.query(
@@ -395,7 +395,7 @@ class NvpSynchronizer():
                 return
 
         if not lswitchport:
-            # Try to get port from nvp
+            # Try to get port from nsx
             try:
                 ls_uuid, lp_uuid = nsx_utils.get_nsx_switch_and_port_id(
                     context.session, self._cluster, neutron_port_data['id'])
@@ -409,15 +409,15 @@ class NvpSynchronizer():
                 # of PortNotFoundOnNetwork when the id exists but
                 # the logical switch port was not found
                 LOG.warning(_("Logical switch port for neutron port %s "
-                              "not found on NVP."), neutron_port_data['id'])
+                              "not found on NSX."), neutron_port_data['id'])
                 lswitchport = None
             else:
                 # If lswitchport is not None, update the cache.
                 # It could be none if the port was deleted from the backend
                 if lswitchport:
-                    self._nvp_cache.update_lswitchport(lswitchport)
+                    self._nsx_cache.update_lswitchport(lswitchport)
         # Note(salv-orlando): It might worth adding a check to verify neutron
-        # resource tag in nvp entity matches Neutron id.
+        # resource tag in nsx entity matches Neutron id.
         # By default assume things go wrong
         status = constants.PORT_STATUS_ERROR
         if lswitchport:
@@ -439,13 +439,13 @@ class NvpSynchronizer():
         # has been tampered with
         neutron_port_mappings = {}
         for lp_uuid in lp_uuids:
-            lport = (self._nvp_cache[lp_uuid].get('data') or
-                     self._nvp_cache[lp_uuid].get('data_bk'))
+            lport = (self._nsx_cache[lp_uuid].get('data') or
+                     self._nsx_cache[lp_uuid].get('data_bk'))
             tags = self._get_tag_dict(lport['tags'])
             neutron_port_id = tags.get('q_port_id')
             if neutron_port_id:
                 neutron_port_mappings[neutron_port_id] = (
-                    self._nvp_cache[lp_uuid])
+                    self._nsx_cache[lp_uuid])
         # Fetch neutron ports from database
         # At the first sync we need to fetch all ports
         filters = ({} if scan_missing else
@@ -478,7 +478,7 @@ class NvpSynchronizer():
             if cursor == 'start':
                 cursor = None
             # Chunk size tuning might, in some conditions, make it larger
-            # than 5,000, which is the maximum page size allowed by the NVP
+            # than 5,000, which is the maximum page size allowed by the NSX
             # API. In this case the request should be split in multiple
             # requests. This is not ideal, and therefore a log warning will
             # be emitted.
@@ -486,7 +486,7 @@ class NvpSynchronizer():
             if num_requests > 1:
                 LOG.warn(_("Requested page size is %(cur_chunk_size)d."
                            "It might be necessary to do %(num_requests)d "
-                           "round-trips to NVP for fetching data. Please "
+                           "round-trips to NSX for fetching data. Please "
                            "tune sync parameters to ensure chunk size "
                            "is less than %(max_page_size)d"),
                          {'cur_chunk_size': page_size,
@@ -514,11 +514,11 @@ class NvpSynchronizer():
             return results, cursor if page_size else 'start', total_size
         return [], cursor, None
 
-    def _fetch_nvp_data_chunk(self, sp):
+    def _fetch_nsx_data_chunk(self, sp):
         base_chunk_size = sp.chunk_size
         chunk_size = base_chunk_size + sp.extra_chunk_size
         LOG.info(_("Fetching up to %s resources "
-                   "from NVP backend"), chunk_size)
+                   "from NSX backend"), chunk_size)
         fetched = ls_count = lr_count = lp_count = 0
         lswitches = lrouters = lswitchports = []
         if sp.ls_cursor or sp.ls_cursor == 'start':
@@ -559,19 +559,19 @@ class NvpSynchronizer():
             sp.ls_cursor = sp.lr_cursor = sp.lp_cursor = 'start'
         LOG.info(_("Running state synchronization task. Chunk: %s"),
                  sp.current_chunk)
-        # Fetch chunk_size data from NVP
+        # Fetch chunk_size data from NSX
         try:
             (lswitches, lrouters, lswitchports) = (
-                self._fetch_nvp_data_chunk(sp))
+                self._fetch_nsx_data_chunk(sp))
         except (api_exc.RequestTimeout, api_exc.NsxApiException):
             sleep_interval = self._sync_backoff
             # Cap max back off to 64 seconds
             self._sync_backoff = min(self._sync_backoff * 2, 64)
             LOG.exception(_("An error occurred while communicating with "
-                            "NVP backend. Will retry synchronization "
+                            "NSX backend. Will retry synchronization "
                             "in %d seconds"), sleep_interval)
             return sleep_interval
-        LOG.debug(_("Time elapsed querying NVP: %s"),
+        LOG.debug(_("Time elapsed querying NSX: %s"),
                   timeutils.utcnow() - start)
         if sp.total_size:
             num_chunks = ((sp.total_size / sp.chunk_size) +
@@ -579,20 +579,20 @@ class NvpSynchronizer():
         else:
             num_chunks = 1
         LOG.debug(_("Number of chunks: %d"), num_chunks)
-        # Find objects which have changed on NVP side and need
+        # Find objects which have changed on NSX side and need
         # to be synchronized
-        (ls_uuids, lr_uuids, lp_uuids) = self._nvp_cache.process_updates(
+        (ls_uuids, lr_uuids, lp_uuids) = self._nsx_cache.process_updates(
             lswitches, lrouters, lswitchports)
         # Process removed objects only at the last chunk
         scan_missing = (sp.current_chunk == num_chunks - 1 and
                         not sp.init_sync_performed)
         if sp.current_chunk == num_chunks - 1:
-            self._nvp_cache.process_deletes()
-            ls_uuids = self._nvp_cache.get_lswitches(
+            self._nsx_cache.process_deletes()
+            ls_uuids = self._nsx_cache.get_lswitches(
                 changed_only=not scan_missing)
-            lr_uuids = self._nvp_cache.get_lrouters(
+            lr_uuids = self._nsx_cache.get_lrouters(
                 changed_only=not scan_missing)
-            lp_uuids = self._nvp_cache.get_lswitchports(
+            lp_uuids = self._nsx_cache.get_lswitchports(
                 changed_only=not scan_missing)
         LOG.debug(_("Time elapsed hashing data: %s"),
                   timeutils.utcnow() - start)
index ed4b0e87c96fba8afc6fea1e62344a6a8cda7455..58d14c55661af973e7e0f3e3cfad151db9bfdb9e 100644 (file)
@@ -363,7 +363,7 @@ class NetworkGatewayMixin(networkgw.NetworkGatewayPluginBase):
                 raise MultipleGatewayConnections(
                     gateway_id=network_gateway_id)
             # Remove gateway port from network
-            # FIXME(salvatore-orlando): Ensure state of port in NVP is
+            # FIXME(salvatore-orlando): Ensure state of port in NSX is
             # consistent with outcome of transaction
             self.delete_port(context, net_connection['port_id'],
                              nw_gw_port_check=False)
index 8f5bd4830bb37201ef5dc71f52157637f2cea125..b4f46e10c44a7bc576888655ee205ddfe8903b81 100644 (file)
@@ -1,5 +1,3 @@
-# vim: tabstop=4 shiftwidth=4 softtabstop=4
-
 # Copyright 2014 VMware, Inc.
 # All Rights Reserved
 #
@@ -27,7 +25,7 @@ HTTP_GET = "GET"
 HTTP_POST = "POST"
 HTTP_DELETE = "DELETE"
 HTTP_PUT = "PUT"
-# Prefix to be used for all NVP API calls
+# Prefix to be used for all NSX API calls
 URI_PREFIX = "/ws.v1"
 NEUTRON_VERSION = version_info.release_string()
 
index 096faf1c40b0169842bee51fd77872fcbdd4cb9b..6ad3293f5667aa51042144002a773415e87161e7 100644 (file)
@@ -132,7 +132,7 @@ class NsxPluginV2(addr_pair_db.AllowedAddressPairsMixin,
     novazone_cluster_map = {}
 
     def __init__(self):
-        super(NvpPluginV2, self).__init__()
+        super(NsxPluginV2, self).__init__()
         # TODO(salv-orlando): Replace These dicts with
         # collections.defaultdict for better handling of default values
         # Routines for managing logical ports in NSX
@@ -180,7 +180,7 @@ class NsxPluginV2(addr_pair_db.AllowedAddressPairsMixin,
         # been yet updated from the config file
         self._is_default_net_gw_in_sync = False
         # Create a synchronizer instance for backend sync
-        self._synchronizer = sync.NvpSynchronizer(
+        self._synchronizer = sync.NsxSynchronizer(
             self, self.cluster,
             self.nsx_sync_opts.state_sync_interval,
             self.nsx_sync_opts.min_sync_req_delay,
@@ -2104,7 +2104,7 @@ class NsxPluginV2(addr_pair_db.AllowedAddressPairsMixin,
             self._ensure_default_security_group(context, tenant_id)
         # NOTE(salv-orlando): Pre-generating Neutron ID for security group.
         neutron_id = str(uuid.uuid4())
-        nvp_secgroup = secgrouplib.create_security_profile(
+        nsx_secgroup = secgrouplib.create_security_profile(
             self.cluster, neutron_id, tenant_id, s)
         with context.session.begin(subtransactions=True):
             s['id'] = neutron_id
@@ -2113,7 +2113,7 @@ class NsxPluginV2(addr_pair_db.AllowedAddressPairsMixin,
             context.session.flush()
             # Add mapping between neutron and nsx identifiers
             nsx_db.add_neutron_nsx_security_group_mapping(
-                context.session, neutron_id, nvp_secgroup['uuid'])
+                context.session, neutron_id, nsx_secgroup['uuid'])
         return sec_group
 
     def delete_security_group(self, context, security_group_id):
index 4ed362cfd9b1a88b542a938dafe4ca9d3391a0f8..84236d5829579411a5103838c2e7b7bdb28e6ee3 100644 (file)
@@ -941,7 +941,7 @@ class NsxAdvancedPlugin(sr_db.ServiceRouter_mixin,
             context, id, firewall_db.Firewall)
 
     def get_firewall(self, context, id, fields=None):
-        fw = super(NvpAdvancedPlugin, self).get_firewall(
+        fw = super(NsxAdvancedPlugin, self).get_firewall(
             context, id, fields)
         if fields and rsi.ROUTER_ID not in fields:
             return fw
@@ -952,7 +952,7 @@ class NsxAdvancedPlugin(sr_db.ServiceRouter_mixin,
         return fw
 
     def get_firewalls(self, context, filters=None, fields=None):
-        fws = super(NvpAdvancedPlugin, self).get_firewalls(
+        fws = super(NsxAdvancedPlugin, self).get_firewalls(
             context, filters, fields)
         if fields and rsi.ROUTER_ID not in fields:
             return fws
@@ -1303,7 +1303,7 @@ class NsxAdvancedPlugin(sr_db.ServiceRouter_mixin,
         self._update_interface(context, router, sync=True)
 
     def get_vip(self, context, id, fields=None):
-        vip = super(NvpAdvancedPlugin, self).get_vip(context, id, fields)
+        vip = super(NsxAdvancedPlugin, self).get_vip(context, id, fields)
         if fields and rsi.ROUTER_ID not in fields:
             return vip
 
@@ -1313,7 +1313,7 @@ class NsxAdvancedPlugin(sr_db.ServiceRouter_mixin,
         return vip
 
     def get_vips(self, context, filters=None, fields=None):
-        vips = super(NvpAdvancedPlugin, self).get_vips(
+        vips = super(NsxAdvancedPlugin, self).get_vips(
             context, filters, fields)
         if fields and rsi.ROUTER_ID not in fields:
             return vips
@@ -1629,7 +1629,7 @@ class NsxAdvancedPlugin(sr_db.ServiceRouter_mixin,
             raise nsx_exc.ServiceOverQuota(
                 overs='vpnservice', err_msg=msg)
 
-        service = super(NvpAdvancedPlugin, self).create_vpnservice(
+        service = super(NsxAdvancedPlugin, self).create_vpnservice(
             context, vpnservice)
         self._resource_set_status(
             context, vpn_db.VPNService,
@@ -1638,7 +1638,7 @@ class NsxAdvancedPlugin(sr_db.ServiceRouter_mixin,
 
     def update_vpnservice(self, context, vpnservice_id, vpnservice):
         vpnservice['vpnservice']['status'] = service_constants.PENDING_UPDATE
-        service = super(NvpAdvancedPlugin, self).update_vpnservice(
+        service = super(NsxAdvancedPlugin, self).update_vpnservice(
             context, vpnservice_id, vpnservice)
         # Only admin_state_up attribute is configurable on Edge.
         if vpnservice['vpnservice'].get('admin_state_up') is None:
@@ -1669,14 +1669,14 @@ class NsxAdvancedPlugin(sr_db.ServiceRouter_mixin,
 
     def create_ipsec_site_connection(self, context, ipsec_site_connection):
         ipsec_site_conn = super(
-            NvpAdvancedPlugin, self).create_ipsec_site_connection(
+            NsxAdvancedPlugin, self).create_ipsec_site_connection(
                 context, ipsec_site_connection)
         try:
             self._vcns_update_ipsec_config(
                 context, ipsec_site_conn['vpnservice_id'])
         except Exception:
             with excutils.save_and_reraise_exception():
-                super(NvpAdvancedPlugin, self).delete_ipsec_site_connection(
+                super(NsxAdvancedPlugin, self).delete_ipsec_site_connection(
                     context, ipsec_site_conn['id'])
         self._resource_set_status(
             context, vpn_db.IPsecSiteConnection,
@@ -1688,7 +1688,7 @@ class NsxAdvancedPlugin(sr_db.ServiceRouter_mixin,
         ipsec_site_connection['ipsec_site_connection']['status'] = (
             service_constants.PENDING_UPDATE)
         ipsec_site_conn = super(
-            NvpAdvancedPlugin, self).update_ipsec_site_connection(
+            NsxAdvancedPlugin, self).update_ipsec_site_connection(
                 context, ipsec_site_connection_id, ipsec_site_connection)
         try:
             self._vcns_update_ipsec_config(
@@ -1717,7 +1717,7 @@ class NsxAdvancedPlugin(sr_db.ServiceRouter_mixin,
                 self._resource_set_status(
                     context, vpn_db.IPsecSiteConnection, ipsec_site_conn_id,
                     service_constants.ERROR)
-        super(NvpAdvancedPlugin, self).delete_ipsec_site_connection(
+        super(NsxAdvancedPlugin, self).delete_ipsec_site_connection(
             context, ipsec_site_conn_id)
 
 
index cd4fc1d538979e1c2919dfe80e4c95c25144fc24..50634f771b0c9a1183bd2cd69281888274b45bfe 100644 (file)
@@ -55,7 +55,7 @@ class CacheTestCase(base.BaseTestCase):
     """Test suite providing coverage for the Cache class."""
 
     def setUp(self):
-        self.nsx_cache = sync.NvpCache()
+        self.nsx_cache = sync.NsxCache()
         for lswitch in LSWITCHES:
             self.nsx_cache._uuid_dict_mappings[lswitch['uuid']] = (
                 self.nsx_cache._lswitches)
@@ -242,8 +242,8 @@ class SyncLoopingCallTestCase(base.BaseTestCase):
         # Avoid runs of the synchronization process - just start
         # the looping call
         with mock.patch.object(
-            sync.NvpSynchronizer, '_synchronize_state', return_value=0.01):
-            synchronizer = sync.NvpSynchronizer(mock.ANY, mock.ANY,
+            sync.NsxSynchronizer, '_synchronize_state', return_value=0.01):
+            synchronizer = sync.NsxSynchronizer(mock.ANY, mock.ANY,
                                                 100, 0, 0)
             time.sleep(0.03)
             # stop looping call before asserting
index 3b8dc4fd5d009311a9222304956ca2d914633c04..a35b32796e1fec534a89b29c846d3cc637d9d5fb 100644 (file)
@@ -84,7 +84,7 @@ class NsxUtilsTestCase(base.BaseTestCase):
         self.assertFalse(exp_ls_uuids)
 
     def _verify_get_nsx_router_id(self, exp_lr_uuid):
-        # The nvplib and db calls are  mocked, therefore the cluster
+        # The nsxlib and db calls are  mocked, therefore the cluster
         # and the neutron_router_id parameters can be set to None
         lr_uuid = nsx_utils.get_nsx_router_id(db_api.get_session(), None, None)
         self.assertEqual(exp_lr_uuid, lr_uuid)
@@ -99,7 +99,7 @@ class NsxUtilsTestCase(base.BaseTestCase):
         self._verify_get_nsx_switch_and_port_id(exp_ls_uuid, exp_lp_uuid)
 
     def test_get_nsx_switch_and_port_id_only_port_db_mapping(self):
-        # This test is representative of the case in which a port with a nvp
+        # This test is representative of the case in which a port with a nsx
         # db mapping in the havana db was upgraded to icehouse
         exp_ls_uuid = uuidutils.generate_uuid()
         exp_lp_uuid = uuidutils.generate_uuid()
@@ -279,7 +279,7 @@ class NsxUtilsTestCase(base.BaseTestCase):
         self.addCleanup(mock.patch.stopall)
 
     def _verify_get_nsx_sec_profile_id(self, exp_sec_prof_uuid):
-        # The nvplib and db calls are  mocked, therefore the cluster
+        # The nsxlib and db calls are  mocked, therefore the cluster
         # and the neutron_id parameters can be set to None
         sec_prof_uuid = nsx_utils.get_nsx_security_group_id(
             db_api.get_session(), None, None)