mirror of
https://github.com/netbox-community/netbox.git
synced 2026-01-23 20:12:42 -06:00
Compare commits
1 Commits
21260-even
...
20902-git-
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
7eedefb2df |
43
.github/ISSUE_TEMPLATE/03-performance.yaml
vendored
43
.github/ISSUE_TEMPLATE/03-performance.yaml
vendored
@@ -1,43 +0,0 @@
|
|||||||
---
|
|
||||||
name: 🏁 Performance
|
|
||||||
type: Performance
|
|
||||||
description: An opportunity to improve application performance
|
|
||||||
labels: ["netbox", "type: performance", "status: needs triage"]
|
|
||||||
body:
|
|
||||||
- type: input
|
|
||||||
attributes:
|
|
||||||
label: NetBox Version
|
|
||||||
description: What version of NetBox are you currently running?
|
|
||||||
placeholder: v4.5.1
|
|
||||||
validations:
|
|
||||||
required: true
|
|
||||||
- type: dropdown
|
|
||||||
attributes:
|
|
||||||
label: Python Version
|
|
||||||
description: What version of Python are you currently running?
|
|
||||||
options:
|
|
||||||
- "3.12"
|
|
||||||
- "3.13"
|
|
||||||
- "3.14"
|
|
||||||
validations:
|
|
||||||
required: true
|
|
||||||
- type: checkboxes
|
|
||||||
attributes:
|
|
||||||
label: Area(s) of Concern
|
|
||||||
description: Which application interface(s) are affected?
|
|
||||||
options:
|
|
||||||
- label: User Interface
|
|
||||||
- label: REST API
|
|
||||||
- label: GraphQL API
|
|
||||||
- label: Python ORM
|
|
||||||
- label: Other
|
|
||||||
validations:
|
|
||||||
required: true
|
|
||||||
- type: textarea
|
|
||||||
attributes:
|
|
||||||
label: Details
|
|
||||||
description: >
|
|
||||||
Describe in detail the operations being performed and the indications of a performance issue.
|
|
||||||
Include any relevant testing parameters, benchmarks, and expected results.
|
|
||||||
validations:
|
|
||||||
required: true
|
|
||||||
@@ -102,7 +102,10 @@ class GitBackend(DataBackend):
|
|||||||
clone_args['pool_manager'] = ProxyPoolManager(self.socks_proxy)
|
clone_args['pool_manager'] = ProxyPoolManager(self.socks_proxy)
|
||||||
|
|
||||||
if self.url_scheme in ('http', 'https'):
|
if self.url_scheme in ('http', 'https'):
|
||||||
if self.params.get('username'):
|
# Only pass explicit credentials if URL doesn't already contain embedded username
|
||||||
|
# to avoid credential conflicts
|
||||||
|
parsed_url = urlparse(self.url)
|
||||||
|
if not parsed_url.username and self.params.get('username'):
|
||||||
clone_args.update(
|
clone_args.update(
|
||||||
{
|
{
|
||||||
"username": self.params.get('username'),
|
"username": self.params.get('username'),
|
||||||
|
|||||||
59
netbox/core/tests/test_data_backends.py
Normal file
59
netbox/core/tests/test_data_backends.py
Normal file
@@ -0,0 +1,59 @@
|
|||||||
|
from unittest.mock import patch
|
||||||
|
|
||||||
|
from django.test import TestCase
|
||||||
|
|
||||||
|
from core.data_backends import GitBackend
|
||||||
|
|
||||||
|
|
||||||
|
class GitBackendCredentialTests(TestCase):
|
||||||
|
|
||||||
|
def _get_clone_kwargs(self, url, **params):
|
||||||
|
backend = GitBackend(url=url, **params)
|
||||||
|
|
||||||
|
with patch('dulwich.porcelain.clone') as mock_clone, \
|
||||||
|
patch('dulwich.porcelain.NoneStream'):
|
||||||
|
try:
|
||||||
|
with backend.fetch():
|
||||||
|
pass
|
||||||
|
except Exception:
|
||||||
|
pass
|
||||||
|
|
||||||
|
if mock_clone.called:
|
||||||
|
return mock_clone.call_args.kwargs
|
||||||
|
return {}
|
||||||
|
|
||||||
|
def test_url_with_embedded_username_skips_explicit_credentials(self):
|
||||||
|
kwargs = self._get_clone_kwargs(
|
||||||
|
url='https://myuser@bitbucket.org/workspace/repo.git',
|
||||||
|
username='myuser',
|
||||||
|
password='my-api-key'
|
||||||
|
)
|
||||||
|
|
||||||
|
self.assertEqual(kwargs.get('username'), None)
|
||||||
|
self.assertEqual(kwargs.get('password'), None)
|
||||||
|
|
||||||
|
def test_url_without_embedded_username_passes_explicit_credentials(self):
|
||||||
|
kwargs = self._get_clone_kwargs(
|
||||||
|
url='https://bitbucket.org/workspace/repo.git',
|
||||||
|
username='myuser',
|
||||||
|
password='my-api-key'
|
||||||
|
)
|
||||||
|
|
||||||
|
self.assertEqual(kwargs.get('username'), 'myuser')
|
||||||
|
self.assertEqual(kwargs.get('password'), 'my-api-key')
|
||||||
|
|
||||||
|
def test_url_with_embedded_username_no_explicit_credentials(self):
|
||||||
|
kwargs = self._get_clone_kwargs(
|
||||||
|
url='https://myuser@bitbucket.org/workspace/repo.git'
|
||||||
|
)
|
||||||
|
|
||||||
|
self.assertEqual(kwargs.get('username'), None)
|
||||||
|
self.assertEqual(kwargs.get('password'), None)
|
||||||
|
|
||||||
|
def test_public_repo_no_credentials(self):
|
||||||
|
kwargs = self._get_clone_kwargs(
|
||||||
|
url='https://github.com/public/repo.git'
|
||||||
|
)
|
||||||
|
|
||||||
|
self.assertEqual(kwargs.get('username'), None)
|
||||||
|
self.assertEqual(kwargs.get('password'), None)
|
||||||
@@ -20,9 +20,7 @@ from utilities.forms.fields import (
|
|||||||
DynamicModelChoiceField, DynamicModelMultipleChoiceField, JSONField, NumericArrayField, SlugField,
|
DynamicModelChoiceField, DynamicModelMultipleChoiceField, JSONField, NumericArrayField, SlugField,
|
||||||
)
|
)
|
||||||
from utilities.forms.rendering import FieldSet, InlineFields, TabbedGroups
|
from utilities.forms.rendering import FieldSet, InlineFields, TabbedGroups
|
||||||
from utilities.forms.widgets import (
|
from utilities.forms.widgets import APISelect, ClearableFileInput, HTMXSelect, NumberWithOptions, SelectWithPK
|
||||||
APISelect, ClearableFileInput, ClearableSelect, HTMXSelect, NumberWithOptions, SelectWithPK,
|
|
||||||
)
|
|
||||||
from utilities.jsonschema import JSONSchemaProperty
|
from utilities.jsonschema import JSONSchemaProperty
|
||||||
from virtualization.models import Cluster, VMInterface
|
from virtualization.models import Cluster, VMInterface
|
||||||
from wireless.models import WirelessLAN, WirelessLANGroup
|
from wireless.models import WirelessLAN, WirelessLANGroup
|
||||||
@@ -594,14 +592,6 @@ class DeviceForm(TenancyForm, PrimaryModelForm):
|
|||||||
},
|
},
|
||||||
)
|
)
|
||||||
)
|
)
|
||||||
face = forms.ChoiceField(
|
|
||||||
label=_('Face'),
|
|
||||||
choices=add_blank_choice(DeviceFaceChoices),
|
|
||||||
required=False,
|
|
||||||
widget=ClearableSelect(
|
|
||||||
requires_fields=['rack']
|
|
||||||
)
|
|
||||||
)
|
|
||||||
device_type = DynamicModelChoiceField(
|
device_type = DynamicModelChoiceField(
|
||||||
label=_('Device type'),
|
label=_('Device type'),
|
||||||
queryset=DeviceType.objects.all(),
|
queryset=DeviceType.objects.all(),
|
||||||
|
|||||||
@@ -1,5 +1,5 @@
|
|||||||
import logging
|
import logging
|
||||||
from collections import UserDict, defaultdict
|
from collections import defaultdict
|
||||||
|
|
||||||
from django.conf import settings
|
from django.conf import settings
|
||||||
from django.utils import timezone
|
from django.utils import timezone
|
||||||
@@ -12,6 +12,7 @@ from core.models import ObjectType
|
|||||||
from netbox.config import get_config
|
from netbox.config import get_config
|
||||||
from netbox.constants import RQ_QUEUE_DEFAULT
|
from netbox.constants import RQ_QUEUE_DEFAULT
|
||||||
from netbox.models.features import has_feature
|
from netbox.models.features import has_feature
|
||||||
|
from users.models import User
|
||||||
from utilities.api import get_serializer_for_model
|
from utilities.api import get_serializer_for_model
|
||||||
from utilities.request import copy_safe_request
|
from utilities.request import copy_safe_request
|
||||||
from utilities.rqworker import get_rq_retry
|
from utilities.rqworker import get_rq_retry
|
||||||
@@ -22,19 +23,6 @@ from .models import EventRule
|
|||||||
logger = logging.getLogger('netbox.events_processor')
|
logger = logging.getLogger('netbox.events_processor')
|
||||||
|
|
||||||
|
|
||||||
class EventContext(UserDict):
|
|
||||||
"""
|
|
||||||
A custom dictionary that automatically serializes its associated object on demand.
|
|
||||||
"""
|
|
||||||
|
|
||||||
def __getitem__(self, item):
|
|
||||||
if item == 'data' and 'data' not in self:
|
|
||||||
data = serialize_for_event(self['object'])
|
|
||||||
self.__setitem__('data', data)
|
|
||||||
return data
|
|
||||||
return super().__getitem__(item)
|
|
||||||
|
|
||||||
|
|
||||||
def serialize_for_event(instance):
|
def serialize_for_event(instance):
|
||||||
"""
|
"""
|
||||||
Return a serialized representation of the given instance suitable for use in a queued event.
|
Return a serialized representation of the given instance suitable for use in a queued event.
|
||||||
@@ -78,42 +66,37 @@ def enqueue_event(queue, instance, request, event_type):
|
|||||||
assert instance.pk is not None
|
assert instance.pk is not None
|
||||||
key = f'{app_label}.{model_name}:{instance.pk}'
|
key = f'{app_label}.{model_name}:{instance.pk}'
|
||||||
if key in queue:
|
if key in queue:
|
||||||
|
queue[key]['data'] = serialize_for_event(instance)
|
||||||
queue[key]['snapshots']['postchange'] = get_snapshots(instance, event_type)['postchange']
|
queue[key]['snapshots']['postchange'] = get_snapshots(instance, event_type)['postchange']
|
||||||
# If the object is being deleted, update any prior "update" event to "delete"
|
# If the object is being deleted, update any prior "update" event to "delete"
|
||||||
if event_type == OBJECT_DELETED:
|
if event_type == OBJECT_DELETED:
|
||||||
queue[key]['event_type'] = event_type
|
queue[key]['event_type'] = event_type
|
||||||
else:
|
else:
|
||||||
queue[key] = EventContext(
|
queue[key] = {
|
||||||
object_type=ObjectType.objects.get_for_model(instance),
|
'object_type': ObjectType.objects.get_for_model(instance),
|
||||||
object_id=instance.pk,
|
'object_id': instance.pk,
|
||||||
object=instance,
|
'event_type': event_type,
|
||||||
event_type=event_type,
|
'data': serialize_for_event(instance),
|
||||||
snapshots=get_snapshots(instance, event_type),
|
'snapshots': get_snapshots(instance, event_type),
|
||||||
request=request,
|
'request': request,
|
||||||
user=request.user,
|
|
||||||
# Legacy request attributes for backward compatibility
|
# Legacy request attributes for backward compatibility
|
||||||
username=request.user.username,
|
'username': request.user.username,
|
||||||
request_id=request.id,
|
'request_id': request.id,
|
||||||
)
|
}
|
||||||
# Force serialization of objects prior to them actually being deleted
|
|
||||||
if event_type == OBJECT_DELETED:
|
|
||||||
queue[key]['data'] = serialize_for_event(instance)
|
|
||||||
|
|
||||||
|
|
||||||
def process_event_rules(event_rules, object_type, event):
|
def process_event_rules(event_rules, object_type, event_type, data, username=None, snapshots=None, request=None):
|
||||||
"""
|
user = User.objects.get(username=username) if username else None
|
||||||
Process a list of EventRules against an event.
|
|
||||||
"""
|
|
||||||
|
|
||||||
for event_rule in event_rules:
|
for event_rule in event_rules:
|
||||||
|
|
||||||
# Evaluate event rule conditions (if any)
|
# Evaluate event rule conditions (if any)
|
||||||
if not event_rule.eval_conditions(event['data']):
|
if not event_rule.eval_conditions(data):
|
||||||
continue
|
continue
|
||||||
|
|
||||||
# Compile event data
|
# Compile event data
|
||||||
event_data = event_rule.action_data or {}
|
event_data = event_rule.action_data or {}
|
||||||
event_data.update(event['data'])
|
event_data.update(data)
|
||||||
|
|
||||||
# Webhooks
|
# Webhooks
|
||||||
if event_rule.action_type == EventRuleActionChoices.WEBHOOK:
|
if event_rule.action_type == EventRuleActionChoices.WEBHOOK:
|
||||||
@@ -126,22 +109,25 @@ def process_event_rules(event_rules, object_type, event):
|
|||||||
params = {
|
params = {
|
||||||
"event_rule": event_rule,
|
"event_rule": event_rule,
|
||||||
"object_type": object_type,
|
"object_type": object_type,
|
||||||
"event_type": event['event_type'],
|
"event_type": event_type,
|
||||||
"data": event_data,
|
"data": event_data,
|
||||||
"snapshots": event['snapshots'],
|
"snapshots": snapshots,
|
||||||
"timestamp": timezone.now().isoformat(),
|
"timestamp": timezone.now().isoformat(),
|
||||||
"username": event['username'],
|
"username": username,
|
||||||
"retry": get_rq_retry()
|
"retry": get_rq_retry()
|
||||||
}
|
}
|
||||||
if 'snapshots' in event:
|
if snapshots:
|
||||||
params['snapshots'] = event['snapshots']
|
params["snapshots"] = snapshots
|
||||||
if 'request' in event:
|
if request:
|
||||||
# Exclude FILES - webhooks don't need uploaded files,
|
# Exclude FILES - webhooks don't need uploaded files,
|
||||||
# which can cause pickle errors with Pillow.
|
# which can cause pickle errors with Pillow.
|
||||||
params['request'] = copy_safe_request(event['request'], include_files=False)
|
params["request"] = copy_safe_request(request, include_files=False)
|
||||||
|
|
||||||
# Enqueue the task
|
# Enqueue the task
|
||||||
rq_queue.enqueue('extras.webhooks.send_webhook', **params)
|
rq_queue.enqueue(
|
||||||
|
"extras.webhooks.send_webhook",
|
||||||
|
**params
|
||||||
|
)
|
||||||
|
|
||||||
# Scripts
|
# Scripts
|
||||||
elif event_rule.action_type == EventRuleActionChoices.SCRIPT:
|
elif event_rule.action_type == EventRuleActionChoices.SCRIPT:
|
||||||
@@ -153,16 +139,16 @@ def process_event_rules(event_rules, object_type, event):
|
|||||||
params = {
|
params = {
|
||||||
"instance": event_rule.action_object,
|
"instance": event_rule.action_object,
|
||||||
"name": script.name,
|
"name": script.name,
|
||||||
"user": event['user'],
|
"user": user,
|
||||||
"data": event_data
|
"data": event_data
|
||||||
}
|
}
|
||||||
if 'snapshots' in event:
|
if snapshots:
|
||||||
params['snapshots'] = event['snapshots']
|
params["snapshots"] = snapshots
|
||||||
if 'request' in event:
|
if request:
|
||||||
params['request'] = copy_safe_request(event['request'])
|
params["request"] = copy_safe_request(request)
|
||||||
|
ScriptJob.enqueue(
|
||||||
# Enqueue the job
|
**params
|
||||||
ScriptJob.enqueue(**params)
|
)
|
||||||
|
|
||||||
# Notification groups
|
# Notification groups
|
||||||
elif event_rule.action_type == EventRuleActionChoices.NOTIFICATION:
|
elif event_rule.action_type == EventRuleActionChoices.NOTIFICATION:
|
||||||
@@ -171,7 +157,7 @@ def process_event_rules(event_rules, object_type, event):
|
|||||||
object_type=object_type,
|
object_type=object_type,
|
||||||
object_id=event_data['id'],
|
object_id=event_data['id'],
|
||||||
object_repr=event_data.get('display'),
|
object_repr=event_data.get('display'),
|
||||||
event_type=event['event_type']
|
event_type=event_type
|
||||||
)
|
)
|
||||||
|
|
||||||
else:
|
else:
|
||||||
@@ -183,8 +169,6 @@ def process_event_rules(event_rules, object_type, event):
|
|||||||
def process_event_queue(events):
|
def process_event_queue(events):
|
||||||
"""
|
"""
|
||||||
Flush a list of object representation to RQ for EventRule processing.
|
Flush a list of object representation to RQ for EventRule processing.
|
||||||
|
|
||||||
This is the default processor listed in EVENTS_PIPELINE.
|
|
||||||
"""
|
"""
|
||||||
events_cache = defaultdict(dict)
|
events_cache = defaultdict(dict)
|
||||||
|
|
||||||
@@ -204,7 +188,11 @@ def process_event_queue(events):
|
|||||||
process_event_rules(
|
process_event_rules(
|
||||||
event_rules=event_rules,
|
event_rules=event_rules,
|
||||||
object_type=object_type,
|
object_type=object_type,
|
||||||
event=event,
|
event_type=event['event_type'],
|
||||||
|
data=event['data'],
|
||||||
|
username=event['username'],
|
||||||
|
snapshots=event['snapshots'],
|
||||||
|
request=event['request'],
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
|
|||||||
@@ -4,7 +4,7 @@ from django.dispatch import receiver
|
|||||||
|
|
||||||
from core.events import *
|
from core.events import *
|
||||||
from core.signals import job_end, job_start
|
from core.signals import job_end, job_start
|
||||||
from extras.events import EventContext, process_event_rules
|
from extras.events import process_event_rules
|
||||||
from extras.models import EventRule, Notification, Subscription
|
from extras.models import EventRule, Notification, Subscription
|
||||||
from netbox.config import get_config
|
from netbox.config import get_config
|
||||||
from netbox.models.features import has_feature
|
from netbox.models.features import has_feature
|
||||||
@@ -102,12 +102,14 @@ def process_job_start_event_rules(sender, **kwargs):
|
|||||||
enabled=True,
|
enabled=True,
|
||||||
object_types=sender.object_type
|
object_types=sender.object_type
|
||||||
)
|
)
|
||||||
event = EventContext(
|
username = sender.user.username if sender.user else None
|
||||||
|
process_event_rules(
|
||||||
|
event_rules=event_rules,
|
||||||
|
object_type=sender.object_type,
|
||||||
event_type=JOB_STARTED,
|
event_type=JOB_STARTED,
|
||||||
data=sender.data,
|
data=sender.data,
|
||||||
user=sender.user,
|
username=username
|
||||||
)
|
)
|
||||||
process_event_rules(event_rules, sender.object_type, event)
|
|
||||||
|
|
||||||
|
|
||||||
@receiver(job_end)
|
@receiver(job_end)
|
||||||
@@ -120,12 +122,14 @@ def process_job_end_event_rules(sender, **kwargs):
|
|||||||
enabled=True,
|
enabled=True,
|
||||||
object_types=sender.object_type
|
object_types=sender.object_type
|
||||||
)
|
)
|
||||||
event = EventContext(
|
username = sender.user.username if sender.user else None
|
||||||
|
process_event_rules(
|
||||||
|
event_rules=event_rules,
|
||||||
|
object_type=sender.object_type,
|
||||||
event_type=JOB_COMPLETED,
|
event_type=JOB_COMPLETED,
|
||||||
data=sender.data,
|
data=sender.data,
|
||||||
user=sender.user,
|
username=username
|
||||||
)
|
)
|
||||||
process_event_rules(event_rules, sender.object_type, event)
|
|
||||||
|
|
||||||
|
|
||||||
#
|
#
|
||||||
|
|||||||
@@ -43,7 +43,7 @@ IMAGEATTACHMENT_IMAGE = """
|
|||||||
<a href="{{ record.image.url }}" target="_blank" class="image-preview" data-bs-placement="top">
|
<a href="{{ record.image.url }}" target="_blank" class="image-preview" data-bs-placement="top">
|
||||||
<i class="mdi mdi-image"></i></a>
|
<i class="mdi mdi-image"></i></a>
|
||||||
{% endif %}
|
{% endif %}
|
||||||
<a href="{{ record.get_absolute_url }}">{{ record.filename|truncate_middle:16 }}</a>
|
<a href="{{ record.get_absolute_url }}">{{ record }}</a>
|
||||||
"""
|
"""
|
||||||
|
|
||||||
NOTIFICATION_ICON = """
|
NOTIFICATION_ICON = """
|
||||||
|
|||||||
@@ -6,7 +6,7 @@ from django.core.files.uploadedfile import SimpleUploadedFile
|
|||||||
from django.forms import ValidationError
|
from django.forms import ValidationError
|
||||||
from django.test import tag, TestCase
|
from django.test import tag, TestCase
|
||||||
|
|
||||||
from core.models import AutoSyncRecord, DataSource, ObjectType
|
from core.models import DataSource, ObjectType
|
||||||
from dcim.models import Device, DeviceRole, DeviceType, Location, Manufacturer, Platform, Region, Site, SiteGroup
|
from dcim.models import Device, DeviceRole, DeviceType, Location, Manufacturer, Platform, Region, Site, SiteGroup
|
||||||
from extras.models import ConfigContext, ConfigContextProfile, ConfigTemplate, ImageAttachment, Tag, TaggedItem
|
from extras.models import ConfigContext, ConfigContextProfile, ConfigTemplate, ImageAttachment, Tag, TaggedItem
|
||||||
from tenancy.models import Tenant, TenantGroup
|
from tenancy.models import Tenant, TenantGroup
|
||||||
@@ -754,53 +754,3 @@ class ConfigTemplateTest(TestCase):
|
|||||||
@tag('regression')
|
@tag('regression')
|
||||||
def test_config_template_with_data_source_nested_templates(self):
|
def test_config_template_with_data_source_nested_templates(self):
|
||||||
self.assertEqual(self.BASE_TEMPLATE, self.main_config_template.render({}))
|
self.assertEqual(self.BASE_TEMPLATE, self.main_config_template.render({}))
|
||||||
|
|
||||||
@tag('regression')
|
|
||||||
def test_autosyncrecord_cleanup_on_detach(self):
|
|
||||||
"""Test that AutoSyncRecord is deleted when detaching from DataSource."""
|
|
||||||
with tempfile.TemporaryDirectory() as temp_dir:
|
|
||||||
templates_dir = Path(temp_dir) / "templates"
|
|
||||||
templates_dir.mkdir(parents=True, exist_ok=True)
|
|
||||||
|
|
||||||
self._create_template_file(templates_dir, 'test.j2', 'Test content')
|
|
||||||
|
|
||||||
data_source = DataSource(
|
|
||||||
name="Test DataSource for Detach",
|
|
||||||
type="local",
|
|
||||||
source_url=str(templates_dir),
|
|
||||||
)
|
|
||||||
data_source.save()
|
|
||||||
data_source.sync()
|
|
||||||
|
|
||||||
data_file = data_source.datafiles.filter(path__endswith='test.j2').first()
|
|
||||||
|
|
||||||
# Create a ConfigTemplate with data_file and auto_sync_enabled
|
|
||||||
config_template = ConfigTemplate(
|
|
||||||
name="TestTemplateForDetach",
|
|
||||||
data_file=data_file,
|
|
||||||
auto_sync_enabled=True
|
|
||||||
)
|
|
||||||
config_template.clean()
|
|
||||||
config_template.save()
|
|
||||||
|
|
||||||
# Verify AutoSyncRecord was created
|
|
||||||
object_type = ObjectType.objects.get_for_model(ConfigTemplate)
|
|
||||||
autosync_records = AutoSyncRecord.objects.filter(
|
|
||||||
object_type=object_type,
|
|
||||||
object_id=config_template.pk
|
|
||||||
)
|
|
||||||
self.assertEqual(autosync_records.count(), 1, "AutoSyncRecord should be created")
|
|
||||||
|
|
||||||
# Detach from DataSource
|
|
||||||
config_template.data_file = None
|
|
||||||
config_template.data_source = None
|
|
||||||
config_template.auto_sync_enabled = False
|
|
||||||
config_template.clean()
|
|
||||||
config_template.save()
|
|
||||||
|
|
||||||
# Verify AutoSyncRecord was deleted
|
|
||||||
autosync_records = AutoSyncRecord.objects.filter(
|
|
||||||
object_type=object_type,
|
|
||||||
object_id=config_template.pk
|
|
||||||
)
|
|
||||||
self.assertEqual(autosync_records.count(), 0, "AutoSyncRecord should be deleted after detaching")
|
|
||||||
|
|||||||
@@ -569,6 +569,7 @@ class SyncedDataMixin(models.Model):
|
|||||||
)
|
)
|
||||||
else:
|
else:
|
||||||
AutoSyncRecord.objects.filter(
|
AutoSyncRecord.objects.filter(
|
||||||
|
datafile=self.data_file,
|
||||||
object_type=object_type,
|
object_type=object_type,
|
||||||
object_id=self.pk
|
object_id=self.pk
|
||||||
).delete()
|
).delete()
|
||||||
@@ -581,6 +582,7 @@ class SyncedDataMixin(models.Model):
|
|||||||
# Delete AutoSyncRecord
|
# Delete AutoSyncRecord
|
||||||
object_type = ObjectType.objects.get_for_model(self)
|
object_type = ObjectType.objects.get_for_model(self)
|
||||||
AutoSyncRecord.objects.filter(
|
AutoSyncRecord.objects.filter(
|
||||||
|
datafile=self.data_file,
|
||||||
object_type=object_type,
|
object_type=object_type,
|
||||||
object_id=self.pk
|
object_id=self.pk
|
||||||
).delete()
|
).delete()
|
||||||
|
|||||||
8
netbox/project-static/dist/netbox.js
vendored
8
netbox/project-static/dist/netbox.js
vendored
File diff suppressed because one or more lines are too long
8
netbox/project-static/dist/netbox.js.map
vendored
8
netbox/project-static/dist/netbox.js.map
vendored
File diff suppressed because one or more lines are too long
@@ -1,40 +0,0 @@
|
|||||||
import TomSelect from 'tom-select';
|
|
||||||
import { getElements } from '../util';
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Initialize clear-field dependencies.
|
|
||||||
* When a required field is cleared, dependent fields with data-requires-fields attribute will also be cleared.
|
|
||||||
*/
|
|
||||||
export function initClearField(): void {
|
|
||||||
// Find all fields with data-requires-fields attribute
|
|
||||||
for (const field of getElements<HTMLSelectElement>('[data-requires-fields]')) {
|
|
||||||
const requiredFieldsAttr = field.getAttribute('data-requires-fields');
|
|
||||||
if (!requiredFieldsAttr) continue;
|
|
||||||
|
|
||||||
// Parse the comma-separated list of required field names
|
|
||||||
const requiredFields = requiredFieldsAttr.split(',').map(name => name.trim());
|
|
||||||
|
|
||||||
// Set up listeners for each required field
|
|
||||||
for (const requiredFieldName of requiredFields) {
|
|
||||||
const requiredField = document.querySelector<HTMLSelectElement>(
|
|
||||||
`[name="${requiredFieldName}"]`,
|
|
||||||
);
|
|
||||||
if (!requiredField) continue;
|
|
||||||
|
|
||||||
// Listen for changes on the required field
|
|
||||||
requiredField.addEventListener('change', () => {
|
|
||||||
// If required field is cleared, also clear this dependent field
|
|
||||||
if (!requiredField.value || requiredField.value === '') {
|
|
||||||
// Check if this field uses TomSelect
|
|
||||||
const tomselect = (field as HTMLSelectElement & { tomselect?: TomSelect }).tomselect;
|
|
||||||
if (tomselect) {
|
|
||||||
tomselect.clear();
|
|
||||||
} else {
|
|
||||||
// Regular select field
|
|
||||||
field.value = '';
|
|
||||||
}
|
|
||||||
}
|
|
||||||
});
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
@@ -1,10 +1,9 @@
|
|||||||
import { initClearField } from './clearField';
|
|
||||||
import { initFormElements } from './elements';
|
import { initFormElements } from './elements';
|
||||||
import { initFilterModifiers } from './filterModifiers';
|
import { initFilterModifiers } from './filterModifiers';
|
||||||
import { initSpeedSelector } from './speedSelector';
|
import { initSpeedSelector } from './speedSelector';
|
||||||
|
|
||||||
export function initForms(): void {
|
export function initForms(): void {
|
||||||
for (const func of [initFormElements, initSpeedSelector, initFilterModifiers, initClearField]) {
|
for (const func of [initFormElements, initSpeedSelector, initFilterModifiers]) {
|
||||||
func();
|
func();
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -8,7 +8,7 @@ msgid ""
|
|||||||
msgstr ""
|
msgstr ""
|
||||||
"Project-Id-Version: PACKAGE VERSION\n"
|
"Project-Id-Version: PACKAGE VERSION\n"
|
||||||
"Report-Msgid-Bugs-To: \n"
|
"Report-Msgid-Bugs-To: \n"
|
||||||
"POT-Creation-Date: 2026-01-22 05:07+0000\n"
|
"POT-Creation-Date: 2026-01-21 05:07+0000\n"
|
||||||
"PO-Revision-Date: YEAR-MO-DA HO:MI+ZONE\n"
|
"PO-Revision-Date: YEAR-MO-DA HO:MI+ZONE\n"
|
||||||
"Last-Translator: FULL NAME <EMAIL@ADDRESS>\n"
|
"Last-Translator: FULL NAME <EMAIL@ADDRESS>\n"
|
||||||
"Language-Team: LANGUAGE <LL@li.org>\n"
|
"Language-Team: LANGUAGE <LL@li.org>\n"
|
||||||
@@ -12037,7 +12037,7 @@ msgstr ""
|
|||||||
msgid "date synced"
|
msgid "date synced"
|
||||||
msgstr ""
|
msgstr ""
|
||||||
|
|
||||||
#: netbox/netbox/models/features.py:621
|
#: netbox/netbox/models/features.py:623
|
||||||
#, python-brace-format
|
#, python-brace-format
|
||||||
msgid "{class_name} must implement a sync_data() method."
|
msgid "{class_name} must implement a sync_data() method."
|
||||||
msgstr ""
|
msgstr ""
|
||||||
@@ -13935,8 +13935,8 @@ msgid "No VLANs Assigned"
|
|||||||
msgstr ""
|
msgstr ""
|
||||||
|
|
||||||
#: netbox/templates/dcim/inc/interface_vlans_table.html:44
|
#: netbox/templates/dcim/inc/interface_vlans_table.html:44
|
||||||
#: netbox/templates/ipam/inc/max_depth.html:11
|
#: netbox/templates/ipam/prefix_list.html:16
|
||||||
#: netbox/templates/ipam/inc/max_length.html:11
|
#: netbox/templates/ipam/prefix_list.html:33
|
||||||
msgid "Clear"
|
msgid "Clear"
|
||||||
msgstr ""
|
msgstr ""
|
||||||
|
|
||||||
@@ -15053,8 +15053,8 @@ msgstr ""
|
|||||||
msgid "Date Added"
|
msgid "Date Added"
|
||||||
msgstr ""
|
msgstr ""
|
||||||
|
|
||||||
#: netbox/templates/ipam/aggregate/prefixes.html:10
|
#: netbox/templates/ipam/aggregate/prefixes.html:8
|
||||||
#: netbox/templates/ipam/prefix/prefixes.html:10
|
#: netbox/templates/ipam/prefix/prefixes.html:8
|
||||||
#: netbox/templates/ipam/role.html:10
|
#: netbox/templates/ipam/role.html:10
|
||||||
msgid "Add Prefix"
|
msgid "Add Prefix"
|
||||||
msgstr ""
|
msgstr ""
|
||||||
@@ -15083,14 +15083,6 @@ msgstr ""
|
|||||||
msgid "Bulk Create"
|
msgid "Bulk Create"
|
||||||
msgstr ""
|
msgstr ""
|
||||||
|
|
||||||
#: netbox/templates/ipam/inc/max_depth.html:6
|
|
||||||
msgid "Max Depth"
|
|
||||||
msgstr ""
|
|
||||||
|
|
||||||
#: netbox/templates/ipam/inc/max_length.html:6
|
|
||||||
msgid "Max Length"
|
|
||||||
msgstr ""
|
|
||||||
|
|
||||||
#: netbox/templates/ipam/inc/panels/fhrp_groups.html:10
|
#: netbox/templates/ipam/inc/panels/fhrp_groups.html:10
|
||||||
msgid "Create Group"
|
msgid "Create Group"
|
||||||
msgstr ""
|
msgstr ""
|
||||||
@@ -15192,6 +15184,14 @@ msgstr ""
|
|||||||
msgid "Hide Depth Indicators"
|
msgid "Hide Depth Indicators"
|
||||||
msgstr ""
|
msgstr ""
|
||||||
|
|
||||||
|
#: netbox/templates/ipam/prefix_list.html:11
|
||||||
|
msgid "Max Depth"
|
||||||
|
msgstr ""
|
||||||
|
|
||||||
|
#: netbox/templates/ipam/prefix_list.html:28
|
||||||
|
msgid "Max Length"
|
||||||
|
msgstr ""
|
||||||
|
|
||||||
#: netbox/templates/ipam/rir.html:10
|
#: netbox/templates/ipam/rir.html:10
|
||||||
msgid "Add Aggregate"
|
msgid "Add Aggregate"
|
||||||
msgstr ""
|
msgstr ""
|
||||||
|
|||||||
@@ -5,7 +5,6 @@ from ..utils import add_blank_choice
|
|||||||
|
|
||||||
__all__ = (
|
__all__ = (
|
||||||
'BulkEditNullBooleanSelect',
|
'BulkEditNullBooleanSelect',
|
||||||
'ClearableSelect',
|
|
||||||
'ColorSelect',
|
'ColorSelect',
|
||||||
'HTMXSelect',
|
'HTMXSelect',
|
||||||
'SelectWithPK',
|
'SelectWithPK',
|
||||||
@@ -29,21 +28,6 @@ class BulkEditNullBooleanSelect(forms.NullBooleanSelect):
|
|||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
class ClearableSelect(forms.Select):
|
|
||||||
"""
|
|
||||||
A Select widget that will be automatically cleared when one or more required fields are cleared.
|
|
||||||
|
|
||||||
Args:
|
|
||||||
requires_fields: A list of field names that this field depends on. When any of these fields
|
|
||||||
are cleared, this field will also be cleared automatically via JavaScript.
|
|
||||||
"""
|
|
||||||
|
|
||||||
def __init__(self, *args, requires_fields=None, **kwargs):
|
|
||||||
super().__init__(*args, **kwargs)
|
|
||||||
if requires_fields:
|
|
||||||
self.attrs['data-requires-fields'] = ','.join(requires_fields)
|
|
||||||
|
|
||||||
|
|
||||||
class ColorSelect(forms.Select):
|
class ColorSelect(forms.Select):
|
||||||
"""
|
"""
|
||||||
Extends the built-in Select widget to colorize each <option>.
|
Extends the built-in Select widget to colorize each <option>.
|
||||||
|
|||||||
@@ -252,16 +252,3 @@ def isodatetime(value, spec='seconds'):
|
|||||||
else:
|
else:
|
||||||
return ''
|
return ''
|
||||||
return mark_safe(f'<span title="{naturaltime(value)}">{text}</span>')
|
return mark_safe(f'<span title="{naturaltime(value)}">{text}</span>')
|
||||||
|
|
||||||
|
|
||||||
@register.filter
|
|
||||||
def truncate_middle(value, length):
|
|
||||||
if len(value) <= length:
|
|
||||||
return value
|
|
||||||
|
|
||||||
# Calculate split points for the two parts
|
|
||||||
half_len = (length - 1) // 2 # 1 for the ellipsis
|
|
||||||
first_part = value[:half_len]
|
|
||||||
second_part = value[len(value) - (length - 1 - half_len):]
|
|
||||||
|
|
||||||
return mark_safe(f"{first_part}…{second_part}")
|
|
||||||
|
|||||||
Reference in New Issue
Block a user