mirror of
https://github.com/netbox-community/netbox.git
synced 2026-01-15 08:12:18 -06:00
Compare commits
12 Commits
fix_module
...
6cf86af72b
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
6cf86af72b | ||
|
|
21f4036782 | ||
|
|
ce3738572c | ||
|
|
cbb979934e | ||
|
|
642d83a4c6 | ||
|
|
a06c12c6b8 | ||
|
|
59afa0b41d | ||
|
|
14b246cb8a | ||
|
|
f0507d00bf | ||
|
|
77b389f105 | ||
|
|
605c61ef5b | ||
|
|
9ae53fc232 |
@@ -3,7 +3,7 @@ from threading import local
|
|||||||
|
|
||||||
from django.contrib.contenttypes.models import ContentType
|
from django.contrib.contenttypes.models import ContentType
|
||||||
from django.core.exceptions import ObjectDoesNotExist, ValidationError
|
from django.core.exceptions import ObjectDoesNotExist, ValidationError
|
||||||
from django.db.models import CASCADE
|
from django.db.models import CASCADE, RESTRICT
|
||||||
from django.db.models.fields.reverse_related import ManyToManyRel, ManyToOneRel
|
from django.db.models.fields.reverse_related import ManyToManyRel, ManyToOneRel
|
||||||
from django.db.models.signals import m2m_changed, post_migrate, post_save, pre_delete
|
from django.db.models.signals import m2m_changed, post_migrate, post_save, pre_delete
|
||||||
from django.dispatch import receiver, Signal
|
from django.dispatch import receiver, Signal
|
||||||
@@ -47,6 +47,7 @@ clear_events = Signal()
|
|||||||
# Object types
|
# Object types
|
||||||
#
|
#
|
||||||
|
|
||||||
|
|
||||||
@receiver(post_migrate)
|
@receiver(post_migrate)
|
||||||
def update_object_types(sender, **kwargs):
|
def update_object_types(sender, **kwargs):
|
||||||
"""
|
"""
|
||||||
@@ -133,7 +134,7 @@ def handle_changed_object(sender, instance, **kwargs):
|
|||||||
prev_change := ObjectChange.objects.filter(
|
prev_change := ObjectChange.objects.filter(
|
||||||
changed_object_type=ContentType.objects.get_for_model(instance),
|
changed_object_type=ContentType.objects.get_for_model(instance),
|
||||||
changed_object_id=instance.pk,
|
changed_object_id=instance.pk,
|
||||||
request_id=request.id
|
request_id=request.id,
|
||||||
).first()
|
).first()
|
||||||
):
|
):
|
||||||
prev_change.postchange_data = objectchange.postchange_data
|
prev_change.postchange_data = objectchange.postchange_data
|
||||||
@@ -172,9 +173,7 @@ def handle_deleted_object(sender, instance, **kwargs):
|
|||||||
try:
|
try:
|
||||||
run_validators(instance, validators)
|
run_validators(instance, validators)
|
||||||
except ValidationError as e:
|
except ValidationError as e:
|
||||||
raise AbortRequest(
|
raise AbortRequest(_("Deletion is prevented by a protection rule: {message}").format(message=e))
|
||||||
_("Deletion is prevented by a protection rule: {message}").format(message=e)
|
|
||||||
)
|
|
||||||
|
|
||||||
# Get the current request, or bail if not set
|
# Get the current request, or bail if not set
|
||||||
request = current_request.get()
|
request = current_request.get()
|
||||||
@@ -221,7 +220,12 @@ def handle_deleted_object(sender, instance, **kwargs):
|
|||||||
obj.snapshot() # Ensure the change record includes the "before" state
|
obj.snapshot() # Ensure the change record includes the "before" state
|
||||||
if type(relation) is ManyToManyRel:
|
if type(relation) is ManyToManyRel:
|
||||||
getattr(obj, related_field_name).remove(instance)
|
getattr(obj, related_field_name).remove(instance)
|
||||||
elif type(relation) is ManyToOneRel and relation.null and relation.on_delete is not CASCADE:
|
elif (
|
||||||
|
type(relation) is ManyToOneRel
|
||||||
|
and relation.null
|
||||||
|
and relation.on_delete is not CASCADE
|
||||||
|
and relation.on_delete is not RESTRICT
|
||||||
|
):
|
||||||
setattr(obj, related_field_name, None)
|
setattr(obj, related_field_name, None)
|
||||||
obj.save()
|
obj.save()
|
||||||
|
|
||||||
@@ -256,6 +260,7 @@ def clear_events_queue(sender, **kwargs):
|
|||||||
# DataSource handlers
|
# DataSource handlers
|
||||||
#
|
#
|
||||||
|
|
||||||
|
|
||||||
@receiver(post_save, sender=DataSource)
|
@receiver(post_save, sender=DataSource)
|
||||||
def enqueue_sync_job(instance, created, **kwargs):
|
def enqueue_sync_job(instance, created, **kwargs):
|
||||||
"""
|
"""
|
||||||
@@ -267,9 +272,10 @@ def enqueue_sync_job(instance, created, **kwargs):
|
|||||||
SyncDataSourceJob.enqueue_once(instance, interval=instance.sync_interval)
|
SyncDataSourceJob.enqueue_once(instance, interval=instance.sync_interval)
|
||||||
elif not created:
|
elif not created:
|
||||||
# Delete any previously scheduled recurring jobs for this DataSource
|
# Delete any previously scheduled recurring jobs for this DataSource
|
||||||
for job in SyncDataSourceJob.get_jobs(instance).defer('data').filter(
|
for job in (
|
||||||
interval__isnull=False,
|
SyncDataSourceJob.get_jobs(instance)
|
||||||
status=JobStatusChoices.STATUS_SCHEDULED
|
.defer('data')
|
||||||
|
.filter(interval__isnull=False, status=JobStatusChoices.STATUS_SCHEDULED)
|
||||||
):
|
):
|
||||||
# Call delete() per instance to ensure the associated background task is deleted as well
|
# Call delete() per instance to ensure the associated background task is deleted as well
|
||||||
job.delete()
|
job.delete()
|
||||||
|
|||||||
@@ -20,4 +20,4 @@ class ManufacturerSerializer(NetBoxModelSerializer):
|
|||||||
'id', 'url', 'display_url', 'display', 'name', 'slug', 'description', 'tags', 'custom_fields',
|
'id', 'url', 'display_url', 'display', 'name', 'slug', 'description', 'tags', 'custom_fields',
|
||||||
'created', 'last_updated', 'devicetype_count', 'inventoryitem_count', 'platform_count',
|
'created', 'last_updated', 'devicetype_count', 'inventoryitem_count', 'platform_count',
|
||||||
]
|
]
|
||||||
brief_fields = ('id', 'url', 'display', 'name', 'slug', 'description', 'devicetype_count')
|
brief_fields = ('id', 'url', 'display', 'name', 'slug', 'description')
|
||||||
|
|||||||
@@ -531,7 +531,7 @@ class RackReservationTest(APIViewTestCases.APIViewTestCase):
|
|||||||
|
|
||||||
class ManufacturerTest(APIViewTestCases.APIViewTestCase):
|
class ManufacturerTest(APIViewTestCases.APIViewTestCase):
|
||||||
model = Manufacturer
|
model = Manufacturer
|
||||||
brief_fields = ['description', 'devicetype_count', 'display', 'id', 'name', 'slug', 'url']
|
brief_fields = ['description', 'display', 'id', 'name', 'slug', 'url']
|
||||||
create_data = [
|
create_data = [
|
||||||
{
|
{
|
||||||
'name': 'Manufacturer 4',
|
'name': 'Manufacturer 4',
|
||||||
|
|||||||
@@ -119,7 +119,9 @@ def process_event_rules(event_rules, object_type, event_type, data, username=Non
|
|||||||
if snapshots:
|
if snapshots:
|
||||||
params["snapshots"] = snapshots
|
params["snapshots"] = snapshots
|
||||||
if request:
|
if request:
|
||||||
params["request"] = copy_safe_request(request)
|
# Exclude FILES - webhooks don't need uploaded files,
|
||||||
|
# which can cause pickle errors with Pillow.
|
||||||
|
params["request"] = copy_safe_request(request, include_files=False)
|
||||||
|
|
||||||
# Enqueue the task
|
# Enqueue the task
|
||||||
rq_queue.enqueue(
|
rq_queue.enqueue(
|
||||||
|
|||||||
@@ -230,10 +230,6 @@ class PrefixImportForm(ScopedImportForm, NetBoxModelImportForm):
|
|||||||
query |= Q(**{
|
query |= Q(**{
|
||||||
f"site__{self.fields['vlan_site'].to_field_name}": vlan_site
|
f"site__{self.fields['vlan_site'].to_field_name}": vlan_site
|
||||||
})
|
})
|
||||||
# Don't Forget to include VLANs without a site in the filter
|
|
||||||
query |= Q(**{
|
|
||||||
f"site__{self.fields['vlan_site'].to_field_name}__isnull": True
|
|
||||||
})
|
|
||||||
|
|
||||||
if vlan_group:
|
if vlan_group:
|
||||||
query &= Q(**{
|
query &= Q(**{
|
||||||
|
|||||||
@@ -564,6 +564,82 @@ vlan: 102
|
|||||||
self.assertEqual(prefix.vlan.vid, 102)
|
self.assertEqual(prefix.vlan.vid, 102)
|
||||||
self.assertEqual(prefix.scope, site)
|
self.assertEqual(prefix.scope, site)
|
||||||
|
|
||||||
|
@override_settings(EXEMPT_VIEW_PERMISSIONS=['*'])
|
||||||
|
def test_prefix_import_with_vlan_site_multiple_vlans_same_vid(self):
|
||||||
|
"""
|
||||||
|
Test import when multiple VLANs exist with the same vid but different sites.
|
||||||
|
Ref: #20560
|
||||||
|
"""
|
||||||
|
site1 = Site.objects.get(name='Site 1')
|
||||||
|
site2 = Site.objects.get(name='Site 2')
|
||||||
|
|
||||||
|
# Create VLANs with the same vid but different sites
|
||||||
|
vlan1 = VLAN.objects.create(vid=1, name='VLAN1-Site1', site=site1)
|
||||||
|
VLAN.objects.create(vid=1, name='VLAN1-Site2', site=site2) # Create ambiguity
|
||||||
|
|
||||||
|
# Import prefix with vlan_site specified
|
||||||
|
IMPORT_DATA = f"""
|
||||||
|
prefix: 10.11.0.0/22
|
||||||
|
status: active
|
||||||
|
scope_type: dcim.site
|
||||||
|
scope_id: {site1.pk}
|
||||||
|
vlan_site: {site1.name}
|
||||||
|
vlan: 1
|
||||||
|
description: LOC02-MGMT
|
||||||
|
"""
|
||||||
|
|
||||||
|
# Add all required permissions to the test user
|
||||||
|
self.add_permissions('ipam.view_prefix', 'ipam.add_prefix')
|
||||||
|
|
||||||
|
form_data = {
|
||||||
|
'data': IMPORT_DATA,
|
||||||
|
'format': 'yaml'
|
||||||
|
}
|
||||||
|
response = self.client.post(reverse('ipam:prefix_bulk_import'), data=form_data, follow=True)
|
||||||
|
self.assertHttpStatus(response, 200)
|
||||||
|
|
||||||
|
# Verify the prefix was created with the correct VLAN
|
||||||
|
prefix = Prefix.objects.get(prefix='10.11.0.0/22')
|
||||||
|
self.assertEqual(prefix.vlan, vlan1)
|
||||||
|
|
||||||
|
@override_settings(EXEMPT_VIEW_PERMISSIONS=['*'])
|
||||||
|
def test_prefix_import_with_vlan_site_and_global_vlan(self):
|
||||||
|
"""
|
||||||
|
Test import when a global VLAN (no site) and site-specific VLAN exist with same vid.
|
||||||
|
When vlan_site is specified, should prefer the site-specific VLAN.
|
||||||
|
Ref: #20560
|
||||||
|
"""
|
||||||
|
site1 = Site.objects.get(name='Site 1')
|
||||||
|
|
||||||
|
# Create a global VLAN (no site) and a site-specific VLAN with the same vid
|
||||||
|
VLAN.objects.create(vid=10, name='VLAN10-Global', site=None) # Create ambiguity
|
||||||
|
vlan_site = VLAN.objects.create(vid=10, name='VLAN10-Site1', site=site1)
|
||||||
|
|
||||||
|
# Import prefix with vlan_site specified
|
||||||
|
IMPORT_DATA = f"""
|
||||||
|
prefix: 10.12.0.0/22
|
||||||
|
status: active
|
||||||
|
scope_type: dcim.site
|
||||||
|
scope_id: {site1.pk}
|
||||||
|
vlan_site: {site1.name}
|
||||||
|
vlan: 10
|
||||||
|
description: Test Site-Specific VLAN
|
||||||
|
"""
|
||||||
|
|
||||||
|
# Add all required permissions to the test user
|
||||||
|
self.add_permissions('ipam.view_prefix', 'ipam.add_prefix')
|
||||||
|
|
||||||
|
form_data = {
|
||||||
|
'data': IMPORT_DATA,
|
||||||
|
'format': 'yaml'
|
||||||
|
}
|
||||||
|
response = self.client.post(reverse('ipam:prefix_bulk_import'), data=form_data, follow=True)
|
||||||
|
self.assertHttpStatus(response, 200)
|
||||||
|
|
||||||
|
# Verify the prefix was created with the site-specific VLAN (not the global one)
|
||||||
|
prefix = Prefix.objects.get(prefix='10.12.0.0/22')
|
||||||
|
self.assertEqual(prefix.vlan, vlan_site)
|
||||||
|
|
||||||
|
|
||||||
class IPRangeTestCase(ViewTestCases.PrimaryObjectViewTestCase):
|
class IPRangeTestCase(ViewTestCases.PrimaryObjectViewTestCase):
|
||||||
model = IPRange
|
model = IPRange
|
||||||
|
|||||||
2
netbox/project-static/dist/netbox.css
vendored
2
netbox/project-static/dist/netbox.css
vendored
File diff suppressed because one or more lines are too long
@@ -36,7 +36,6 @@ form.object-edit {
|
|||||||
// Make optgroup labels sticky when scrolling through select elements
|
// Make optgroup labels sticky when scrolling through select elements
|
||||||
select[multiple] {
|
select[multiple] {
|
||||||
optgroup {
|
optgroup {
|
||||||
position: sticky;
|
|
||||||
top: 0;
|
top: 0;
|
||||||
background-color: var(--bs-body-bg);
|
background-color: var(--bs-body-bg);
|
||||||
font-style: normal;
|
font-style: normal;
|
||||||
|
|||||||
File diff suppressed because it is too large
Load Diff
@@ -35,27 +35,34 @@ class NetBoxFakeRequest:
|
|||||||
# Utility functions
|
# Utility functions
|
||||||
#
|
#
|
||||||
|
|
||||||
def copy_safe_request(request):
|
def copy_safe_request(request, include_files=True):
|
||||||
"""
|
"""
|
||||||
Copy selected attributes from a request object into a new fake request object. This is needed in places where
|
Copy selected attributes from a request object into a new fake request object. This is needed in places where
|
||||||
thread safe pickling of the useful request data is needed.
|
thread safe pickling of the useful request data is needed.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
request: The original request object
|
||||||
|
include_files: Whether to include request.FILES.
|
||||||
"""
|
"""
|
||||||
meta = {
|
meta = {
|
||||||
k: request.META[k]
|
k: request.META[k]
|
||||||
for k in HTTP_REQUEST_META_SAFE_COPY
|
for k in HTTP_REQUEST_META_SAFE_COPY
|
||||||
if k in request.META and isinstance(request.META[k], str)
|
if k in request.META and isinstance(request.META[k], str)
|
||||||
}
|
}
|
||||||
return NetBoxFakeRequest({
|
data = {
|
||||||
'META': meta,
|
'META': meta,
|
||||||
'COOKIES': request.COOKIES,
|
'COOKIES': request.COOKIES,
|
||||||
'POST': request.POST,
|
'POST': request.POST,
|
||||||
'GET': request.GET,
|
'GET': request.GET,
|
||||||
'FILES': request.FILES,
|
|
||||||
'user': request.user,
|
'user': request.user,
|
||||||
'method': request.method,
|
'method': request.method,
|
||||||
'path': request.path,
|
'path': request.path,
|
||||||
'id': getattr(request, 'id', None), # UUID assigned by middleware
|
'id': getattr(request, 'id', None), # UUID assigned by middleware
|
||||||
})
|
}
|
||||||
|
if include_files:
|
||||||
|
data['FILES'] = request.FILES
|
||||||
|
|
||||||
|
return NetBoxFakeRequest(data)
|
||||||
|
|
||||||
|
|
||||||
def get_client_ip(request, additional_headers=()):
|
def get_client_ip(request, additional_headers=()):
|
||||||
|
|||||||
Reference in New Issue
Block a user