Related to https://github.com/grafana/oncall-private/issues/2994 - Extend gaps/empty shift checks to consider 30 days (customizable via param, eventually make it customizable per schedule?); ie. every week (per beat schedule), check the schedule next 30 days - Trigger checks via async task on schedule API updates (instead of a sync call) - Update notifications wording / link to schedule
610 lines
26 KiB
Python
610 lines
26 KiB
Python
import datetime
|
|
import functools
|
|
import operator
|
|
|
|
import pytz
|
|
from django.core.exceptions import ObjectDoesNotExist
|
|
from django.db.models import Count, OuterRef, Prefetch, Q, Subquery
|
|
from django.db.utils import IntegrityError
|
|
from django.urls import reverse
|
|
from django.utils import dateparse, timezone
|
|
from django.utils.functional import cached_property
|
|
from django_filters import rest_framework as filters
|
|
from rest_framework import mixins, status
|
|
from rest_framework.decorators import action
|
|
from rest_framework.exceptions import NotFound
|
|
from rest_framework.fields import BooleanField
|
|
from rest_framework.filters import SearchFilter
|
|
from rest_framework.permissions import IsAuthenticated
|
|
from rest_framework.request import Request
|
|
from rest_framework.views import Response
|
|
from rest_framework.viewsets import ModelViewSet
|
|
|
|
from apps.alerts.models import EscalationChain, EscalationPolicy
|
|
from apps.api.permissions import RBACPermission
|
|
from apps.api.serializers.schedule_base import ScheduleFastSerializer
|
|
from apps.api.serializers.schedule_polymorphic import (
|
|
PolymorphicScheduleCreateSerializer,
|
|
PolymorphicScheduleSerializer,
|
|
PolymorphicScheduleUpdateSerializer,
|
|
)
|
|
from apps.api.serializers.shift_swap import ShiftSwapRequestExpandedUsersListSerializer
|
|
from apps.api.serializers.user import ScheduleUserSerializer
|
|
from apps.auth_token.auth import PluginAuthentication
|
|
from apps.auth_token.constants import SCHEDULE_EXPORT_TOKEN_NAME
|
|
from apps.auth_token.models import ScheduleExportAuthToken
|
|
from apps.mobile_app.auth import MobileAppAuthTokenAuthentication
|
|
from apps.schedules.constants import PREFETCHED_SHIFT_SWAPS
|
|
from apps.schedules.ical_utils import get_oncall_users_for_multiple_schedules
|
|
from apps.schedules.models import OnCallSchedule, ShiftSwapRequest
|
|
from apps.slack.tasks import update_slack_user_group_for_schedules
|
|
from common.api_helpers.exceptions import BadRequest, Conflict
|
|
from common.api_helpers.filters import ByTeamModelFieldFilterMixin, ModelFieldFilterMixin, TeamModelMultipleChoiceFilter
|
|
from common.api_helpers.mixins import (
|
|
CreateSerializerMixin,
|
|
PublicPrimaryKeyMixin,
|
|
ShortSerializerMixin,
|
|
TeamFilteringMixin,
|
|
UpdateSerializerMixin,
|
|
)
|
|
from common.api_helpers.paginators import FifteenPageSizePaginator
|
|
from common.api_helpers.utils import create_engine_url, get_date_range_from_request
|
|
from common.insight_log import EntityEvent, write_resource_insight_log
|
|
from common.timezones import raise_exception_if_not_valid_timezone
|
|
|
|
EVENTS_FILTER_BY_ROTATION = "rotation"
|
|
EVENTS_FILTER_BY_OVERRIDE = "override"
|
|
EVENTS_FILTER_BY_FINAL = "final"
|
|
|
|
SCHEDULE_TYPE_TO_CLASS = {
|
|
str(num_type): cls for cls, num_type in PolymorphicScheduleSerializer.SCHEDULE_CLASS_TO_TYPE.items()
|
|
}
|
|
|
|
|
|
class ScheduleFilter(ByTeamModelFieldFilterMixin, ModelFieldFilterMixin, filters.FilterSet):
|
|
team = TeamModelMultipleChoiceFilter()
|
|
|
|
|
|
class ScheduleView(
|
|
TeamFilteringMixin,
|
|
PublicPrimaryKeyMixin[OnCallSchedule],
|
|
ShortSerializerMixin,
|
|
CreateSerializerMixin,
|
|
UpdateSerializerMixin,
|
|
ModelViewSet,
|
|
mixins.ListModelMixin,
|
|
):
|
|
authentication_classes = (
|
|
MobileAppAuthTokenAuthentication,
|
|
PluginAuthentication,
|
|
)
|
|
permission_classes = (IsAuthenticated, RBACPermission)
|
|
rbac_permissions = {
|
|
"metadata": [RBACPermission.Permissions.SCHEDULES_READ],
|
|
"list": [RBACPermission.Permissions.SCHEDULES_READ],
|
|
"retrieve": [RBACPermission.Permissions.SCHEDULES_READ],
|
|
"events": [RBACPermission.Permissions.SCHEDULES_READ],
|
|
"filter_events": [RBACPermission.Permissions.SCHEDULES_READ],
|
|
"filter_shift_swaps": [RBACPermission.Permissions.SCHEDULES_READ],
|
|
"next_shifts_per_user": [RBACPermission.Permissions.SCHEDULES_READ],
|
|
"related_users": [RBACPermission.Permissions.SCHEDULES_READ],
|
|
"quality": [RBACPermission.Permissions.SCHEDULES_READ],
|
|
"notify_empty_oncall_options": [RBACPermission.Permissions.SCHEDULES_READ],
|
|
"notify_oncall_shift_freq_options": [RBACPermission.Permissions.SCHEDULES_READ],
|
|
"mention_options": [RBACPermission.Permissions.SCHEDULES_READ],
|
|
"related_escalation_chains": [RBACPermission.Permissions.SCHEDULES_READ],
|
|
"current_user_events": [RBACPermission.Permissions.SCHEDULES_READ],
|
|
"create": [RBACPermission.Permissions.SCHEDULES_WRITE],
|
|
"update": [RBACPermission.Permissions.SCHEDULES_WRITE],
|
|
"partial_update": [RBACPermission.Permissions.SCHEDULES_WRITE],
|
|
"destroy": [RBACPermission.Permissions.SCHEDULES_WRITE],
|
|
"reload_ical": [RBACPermission.Permissions.SCHEDULES_WRITE],
|
|
"export_token": [RBACPermission.Permissions.SCHEDULES_EXPORT],
|
|
"filters": [RBACPermission.Permissions.SCHEDULES_READ],
|
|
}
|
|
|
|
filter_backends = [SearchFilter, filters.DjangoFilterBackend]
|
|
search_fields = ("name",)
|
|
filterset_class = ScheduleFilter
|
|
|
|
queryset = OnCallSchedule.objects.all()
|
|
serializer_class = PolymorphicScheduleSerializer
|
|
create_serializer_class = PolymorphicScheduleCreateSerializer
|
|
update_serializer_class = PolymorphicScheduleUpdateSerializer
|
|
short_serializer_class = ScheduleFastSerializer
|
|
pagination_class = FifteenPageSizePaginator
|
|
|
|
@cached_property
|
|
def can_update_user_groups(self):
|
|
"""
|
|
This property is needed to be propagated down to serializers,
|
|
since it makes an API call to Slack and the response should be cached.
|
|
"""
|
|
slack_team_identity = self.request.auth.organization.slack_team_identity
|
|
|
|
if slack_team_identity is None:
|
|
return False
|
|
|
|
user_group = slack_team_identity.usergroups.filter(is_active=True).first()
|
|
if user_group is None:
|
|
return False
|
|
|
|
return user_group.can_be_updated
|
|
|
|
@cached_property
|
|
def oncall_users(self):
|
|
"""
|
|
The result of this method is cached and is reused for the whole lifetime of a request,
|
|
since self.get_serializer_context() is called multiple times for every instance in the queryset.
|
|
"""
|
|
current_schedules = self.get_queryset(annotate=False).none()
|
|
events_datetime = datetime.datetime.now(datetime.timezone.utc)
|
|
if self.action == "list":
|
|
# listing page, only get oncall users for current page schedules, prefetch shift swap requests
|
|
current_schedules = self.filter_queryset(self.get_queryset(annotate=False)).prefetch_related(
|
|
self.prefetch_shift_swaps(
|
|
queryset=ShiftSwapRequest.objects.filter(
|
|
swap_start__lte=events_datetime, swap_end__gte=events_datetime
|
|
)
|
|
)
|
|
)
|
|
current_schedules = self.paginate_queryset(current_schedules)
|
|
elif self.kwargs.get("pk"):
|
|
# if this is a particular schedule detail, only consider it as current
|
|
current_schedules = [self.get_object(annotate=False)]
|
|
return get_oncall_users_for_multiple_schedules(current_schedules, events_datetime)
|
|
|
|
@staticmethod
|
|
def prefetch_shift_swaps(queryset):
|
|
return Prefetch(
|
|
"shift_swap_requests",
|
|
queryset=queryset.select_related("benefactor", "beneficiary").order_by("created_at"),
|
|
to_attr=PREFETCHED_SHIFT_SWAPS,
|
|
)
|
|
|
|
def get_serializer_context(self):
|
|
context = super().get_serializer_context()
|
|
context.update({"can_update_user_groups": self.can_update_user_groups})
|
|
context.update({"oncall_users": self.oncall_users})
|
|
return context
|
|
|
|
def _annotate_queryset(self, queryset):
|
|
"""Annotate queryset with additional schedule metadata."""
|
|
escalation_policies = (
|
|
EscalationPolicy.objects.values("notify_schedule")
|
|
.order_by("notify_schedule")
|
|
.annotate(num_escalation_chains=Count("notify_schedule"))
|
|
.filter(notify_schedule=OuterRef("id"))
|
|
)
|
|
queryset = queryset.annotate(
|
|
num_escalation_chains=Subquery(escalation_policies.values("num_escalation_chains")[:1]),
|
|
)
|
|
return queryset
|
|
|
|
def get_queryset(self, ignore_filtering_by_available_teams=False, annotate=True):
|
|
is_short_request = self.request.query_params.get("short", "false") == "true"
|
|
filter_by_type = self.request.query_params.getlist("type")
|
|
mine = BooleanField(allow_null=True).to_internal_value(data=self.request.query_params.get("mine"))
|
|
used = BooleanField(allow_null=True).to_internal_value(data=self.request.query_params.get("used"))
|
|
organization = self.request.auth.organization
|
|
queryset = OnCallSchedule.objects.filter(organization=organization).defer(
|
|
# avoid requesting large text fields which are not used when listing schedules
|
|
"prev_ical_file_primary",
|
|
"prev_ical_file_overrides",
|
|
"cached_ical_final_schedule",
|
|
)
|
|
if not ignore_filtering_by_available_teams:
|
|
queryset = queryset.filter(*self.available_teams_lookup_args).distinct()
|
|
if not is_short_request:
|
|
if annotate:
|
|
queryset = self._annotate_queryset(queryset)
|
|
queryset = self.serializer_class.setup_eager_loading(queryset)
|
|
if filter_by_type:
|
|
valid_types = [i for i in filter_by_type if i in SCHEDULE_TYPE_TO_CLASS]
|
|
if valid_types:
|
|
queryset = functools.reduce(
|
|
operator.or_, [queryset.filter().instance_of(SCHEDULE_TYPE_TO_CLASS[i]) for i in valid_types]
|
|
)
|
|
if used is not None:
|
|
queryset = queryset.filter(escalation_policies__isnull=not used).distinct()
|
|
if mine:
|
|
user = self.request.user
|
|
queryset = queryset.related_to_user(user)
|
|
|
|
queryset = queryset.order_by("pk")
|
|
return queryset
|
|
|
|
def perform_create(self, serializer):
|
|
serializer.save()
|
|
write_resource_insight_log(instance=serializer.instance, author=self.request.user, event=EntityEvent.CREATED)
|
|
|
|
def perform_update(self, serializer):
|
|
prev_state = serializer.instance.insight_logs_serialized
|
|
old_user_group = serializer.instance.user_group
|
|
serializer.save()
|
|
if old_user_group is not None:
|
|
update_slack_user_group_for_schedules.apply_async((old_user_group.pk,))
|
|
if serializer.instance.user_group is not None and serializer.instance.user_group != old_user_group:
|
|
update_slack_user_group_for_schedules.apply_async((serializer.instance.user_group.pk,))
|
|
new_state = serializer.instance.insight_logs_serialized
|
|
write_resource_insight_log(
|
|
instance=serializer.instance,
|
|
author=self.request.user,
|
|
event=EntityEvent.UPDATED,
|
|
prev_state=prev_state,
|
|
new_state=new_state,
|
|
)
|
|
|
|
def perform_destroy(self, instance):
|
|
write_resource_insight_log(
|
|
instance=instance,
|
|
author=self.request.user,
|
|
event=EntityEvent.DELETED,
|
|
)
|
|
instance.delete()
|
|
|
|
if instance.user_group is not None:
|
|
update_slack_user_group_for_schedules.apply_async((instance.user_group.pk,))
|
|
|
|
def get_object(self, annotate=True) -> OnCallSchedule:
|
|
# get the object from the whole organization if there is a flag `get_from_organization=true`
|
|
# otherwise get the object from the current team
|
|
get_from_organization: bool = self.request.query_params.get("from_organization", "false") == "true"
|
|
if get_from_organization:
|
|
return self.get_object_from_organization(annotate=annotate)
|
|
queryset_kwargs = {"annotate": annotate}
|
|
return super().get_object(queryset_kwargs)
|
|
|
|
def get_object_from_organization(self, ignore_filtering_by_available_teams=False, annotate=True):
|
|
# use this method to get the object from the whole organization instead of the current team
|
|
pk = self.kwargs["pk"]
|
|
organization = self.request.auth.organization
|
|
queryset = organization.oncall_schedules.filter(
|
|
public_primary_key=pk,
|
|
)
|
|
if not ignore_filtering_by_available_teams:
|
|
queryset = queryset.filter(*self.available_teams_lookup_args).distinct()
|
|
|
|
if annotate:
|
|
queryset = self._annotate_queryset(queryset)
|
|
queryset = self.serializer_class.setup_eager_loading(queryset)
|
|
|
|
try:
|
|
obj = queryset.get()
|
|
except ObjectDoesNotExist:
|
|
raise NotFound
|
|
|
|
# May raise a permission denied
|
|
self.check_object_permissions(self.request, obj)
|
|
|
|
return obj
|
|
|
|
def get_request_timezone(self):
|
|
user_tz = self.request.query_params.get("user_tz", "UTC")
|
|
raise_exception_if_not_valid_timezone(user_tz)
|
|
|
|
date = timezone.now().date()
|
|
date_param = self.request.query_params.get("date")
|
|
if date_param is not None:
|
|
try:
|
|
date = dateparse.parse_date(date_param)
|
|
except ValueError:
|
|
raise BadRequest(detail="Invalid date format")
|
|
else:
|
|
if date is None:
|
|
raise BadRequest(detail="Invalid date format")
|
|
|
|
return user_tz, date
|
|
|
|
@action(detail=True, methods=["get"])
|
|
def events(self, request, pk):
|
|
user_tz, starting_date = self.get_request_timezone()
|
|
with_empty = self.request.query_params.get("with_empty", False) == "true"
|
|
with_gap = self.request.query_params.get("with_gap", False) == "true"
|
|
|
|
schedule = self.get_object(annotate=False)
|
|
|
|
pytz_tz = pytz.timezone(user_tz)
|
|
datetime_start = datetime.datetime.combine(starting_date, datetime.time.min, tzinfo=pytz_tz)
|
|
datetime_end = datetime_start + datetime.timedelta(days=1)
|
|
events = schedule.filter_events(datetime_start, datetime_end, with_empty=with_empty, with_gap=with_gap)
|
|
|
|
schedule_slack_channel = schedule.slack_channel
|
|
slack_channel = (
|
|
{
|
|
"id": schedule_slack_channel.public_primary_key,
|
|
"slack_id": schedule_slack_channel.slack_id,
|
|
"display_name": schedule_slack_channel.name,
|
|
}
|
|
if schedule_slack_channel is not None
|
|
else None
|
|
)
|
|
|
|
result = {
|
|
"id": schedule.public_primary_key,
|
|
"name": schedule.name,
|
|
"type": PolymorphicScheduleSerializer().to_resource_type(schedule),
|
|
"slack_channel": slack_channel,
|
|
"events": events,
|
|
}
|
|
return Response(result, status=status.HTTP_200_OK)
|
|
|
|
@action(detail=True, methods=["get"])
|
|
def filter_events(self, request: Request, pk: str) -> Response:
|
|
user_tz, starting_date, days = get_date_range_from_request(self.request)
|
|
|
|
filter_by: str | None = self.request.query_params.get("type")
|
|
valid_filters = (EVENTS_FILTER_BY_ROTATION, EVENTS_FILTER_BY_OVERRIDE, EVENTS_FILTER_BY_FINAL)
|
|
if filter_by is not None and filter_by not in valid_filters:
|
|
raise BadRequest(detail="Invalid type value")
|
|
resolve_schedule = filter_by is None or filter_by == EVENTS_FILTER_BY_FINAL
|
|
|
|
schedule = self.get_object(annotate=False)
|
|
|
|
pytz_tz = pytz.timezone(user_tz)
|
|
datetime_start = datetime.datetime.combine(starting_date, datetime.time.min, tzinfo=pytz_tz)
|
|
datetime_end = datetime_start + datetime.timedelta(days=days)
|
|
|
|
if filter_by is not None and filter_by != EVENTS_FILTER_BY_FINAL:
|
|
filter_by = OnCallSchedule.PRIMARY if filter_by == EVENTS_FILTER_BY_ROTATION else OnCallSchedule.OVERRIDES
|
|
events = schedule.filter_events(
|
|
datetime_start,
|
|
datetime_end,
|
|
with_empty=True,
|
|
with_gap=resolve_schedule,
|
|
filter_by=filter_by,
|
|
all_day_datetime=True,
|
|
include_shift_info=True,
|
|
)
|
|
else: # return final schedule
|
|
events = schedule.final_events(datetime_start, datetime_end, include_shift_info=True)
|
|
|
|
result = {
|
|
"id": schedule.public_primary_key,
|
|
"name": schedule.name,
|
|
"type": PolymorphicScheduleSerializer().to_resource_type(schedule),
|
|
"events": events,
|
|
}
|
|
return Response(result, status=status.HTTP_200_OK)
|
|
|
|
@action(detail=True, methods=["get"])
|
|
def filter_shift_swaps(self, request: Request, pk: str) -> Response:
|
|
user_tz, starting_date, days = get_date_range_from_request(self.request)
|
|
schedule = self.get_object(annotate=False)
|
|
|
|
pytz_tz = pytz.timezone(user_tz)
|
|
datetime_start = datetime.datetime.combine(starting_date, datetime.time.min, tzinfo=pytz_tz)
|
|
datetime_end = datetime_start + datetime.timedelta(days=days)
|
|
|
|
swap_requests = schedule.filter_swap_requests(datetime_start, datetime_end)
|
|
|
|
serialized_swap_requests = ShiftSwapRequestExpandedUsersListSerializer(
|
|
swap_requests, context={"request": self.request}, many=True
|
|
)
|
|
result = {"shift_swaps": serialized_swap_requests.data}
|
|
|
|
return Response(result, status=status.HTTP_200_OK)
|
|
|
|
@action(detail=True, methods=["get"])
|
|
def next_shifts_per_user(self, request, pk):
|
|
"""Return next shift for users in schedule."""
|
|
days = self.request.query_params.get("days")
|
|
days = int(days) if days else 30
|
|
now = timezone.now()
|
|
datetime_end = now + datetime.timedelta(days=days)
|
|
schedule = self.get_object(annotate=False)
|
|
|
|
users = {}
|
|
events = schedule.final_events(now, datetime_end)
|
|
users_tz = {u.public_primary_key: u.timezone for u in schedule.related_users()}
|
|
added_users = set()
|
|
for e in events:
|
|
user_ppk = e["users"][0]["pk"] if e["users"] else None
|
|
if user_ppk is not None and user_ppk not in users and user_ppk in users_tz and e["end"] > now:
|
|
users[user_ppk] = e
|
|
users[user_ppk]["user_timezone"] = users_tz[user_ppk]
|
|
added_users.add(user_ppk)
|
|
|
|
result = {"users": users}
|
|
return Response(result, status=status.HTTP_200_OK)
|
|
|
|
@action(detail=True, methods=["get"])
|
|
def related_users(self, request, pk):
|
|
schedule = self.get_object(annotate=False)
|
|
serializer = ScheduleUserSerializer(schedule.related_users(), many=True)
|
|
result = {"users": serializer.data}
|
|
return Response(result, status=status.HTTP_200_OK)
|
|
|
|
@action(detail=True, methods=["get"])
|
|
def related_escalation_chains(self, request, pk):
|
|
"""Return escalation chains associated to schedule."""
|
|
schedule = self.get_object(annotate=True)
|
|
escalation_chains = EscalationChain.objects.filter(escalation_policies__notify_schedule=schedule).distinct()
|
|
|
|
result = [{"name": e.name, "pk": e.public_primary_key} for e in escalation_chains]
|
|
return Response(result, status=status.HTTP_200_OK)
|
|
|
|
@action(detail=True, methods=["get"])
|
|
def quality(self, request, pk):
|
|
schedule = self.get_object(annotate=False)
|
|
|
|
_, date = self.get_request_timezone()
|
|
datetime_start = datetime.datetime.combine(date, datetime.time.min, tzinfo=pytz.UTC)
|
|
days = self.request.query_params.get("days")
|
|
days = int(days) if days else None
|
|
|
|
return Response(schedule.quality_report(datetime_start, days))
|
|
|
|
@action(detail=False, methods=["get"])
|
|
def current_user_events(self, request):
|
|
user_tz, starting_date, days = get_date_range_from_request(self.request)
|
|
pytz_tz = pytz.timezone(user_tz)
|
|
datetime_start = datetime.datetime.combine(starting_date, datetime.time.min, tzinfo=pytz_tz)
|
|
datetime_end = datetime_start + datetime.timedelta(days=days)
|
|
schedules = (
|
|
OnCallSchedule.objects.related_to_user(self.request.user)
|
|
.select_related("organization")
|
|
.prefetch_related(
|
|
self.prefetch_shift_swaps(
|
|
queryset=ShiftSwapRequest.objects.filter(
|
|
Q(swap_start__lt=datetime_start, swap_end__gte=datetime_start)
|
|
| Q(swap_start__gte=datetime_start, swap_start__lte=datetime_end)
|
|
)
|
|
)
|
|
)
|
|
)
|
|
schedules_events = []
|
|
is_oncall = False
|
|
for schedule in schedules:
|
|
passed_shifts, current_shifts, upcoming_shifts = schedule.shifts_for_user(
|
|
user=self.request.user, datetime_start=datetime_start, days=days
|
|
)
|
|
all_shifts = passed_shifts + current_shifts + upcoming_shifts
|
|
if all_shifts:
|
|
schedules_events.append(
|
|
{"id": schedule.public_primary_key, "name": schedule.name, "events": all_shifts}
|
|
)
|
|
if current_shifts and not is_oncall:
|
|
is_oncall = True
|
|
result = {"schedules": schedules_events, "is_oncall": is_oncall}
|
|
return Response(result, status=status.HTTP_200_OK)
|
|
|
|
@action(detail=False, methods=["get"])
|
|
def type_options(self, request):
|
|
# TODO: check if it needed
|
|
choices = []
|
|
for item in OnCallSchedule.SCHEDULE_CHOICES:
|
|
choices.append({"value": str(item[0]), "display_name": item[1]})
|
|
return Response(choices)
|
|
|
|
@action(detail=True, methods=["post"])
|
|
def reload_ical(self, request, pk):
|
|
schedule = self.get_object(annotate=False)
|
|
schedule.drop_cached_ical()
|
|
schedule.check_gaps_and_empty_shifts_for_next_days()
|
|
|
|
if schedule.user_group is not None:
|
|
update_slack_user_group_for_schedules.apply_async((schedule.user_group.pk,))
|
|
|
|
return Response(status=status.HTTP_200_OK)
|
|
|
|
@action(detail=True, methods=["get", "post", "delete"])
|
|
def export_token(self, request, pk):
|
|
schedule = self.get_object(annotate=False)
|
|
|
|
if self.request.method == "GET":
|
|
try:
|
|
token = ScheduleExportAuthToken.objects.get(user_id=self.request.user.id, schedule_id=schedule.id)
|
|
except ScheduleExportAuthToken.DoesNotExist:
|
|
raise NotFound
|
|
|
|
response = {
|
|
"created_at": token.created_at,
|
|
"revoked_at": token.revoked_at,
|
|
"active": token.active,
|
|
}
|
|
|
|
return Response(response, status=status.HTTP_200_OK)
|
|
|
|
if self.request.method == "POST":
|
|
try:
|
|
instance, token = ScheduleExportAuthToken.create_auth_token(
|
|
request.user, request.user.organization, schedule
|
|
)
|
|
write_resource_insight_log(instance=instance, author=self.request.user, event=EntityEvent.CREATED)
|
|
except IntegrityError:
|
|
raise Conflict("Schedule export token for user already exists")
|
|
|
|
export_url = create_engine_url(
|
|
reverse("api-public:schedules-export", kwargs={"pk": schedule.public_primary_key})
|
|
+ f"?{SCHEDULE_EXPORT_TOKEN_NAME}={token}"
|
|
)
|
|
|
|
data = {"token": token, "created_at": instance.created_at, "export_url": export_url}
|
|
|
|
return Response(data, status=status.HTTP_201_CREATED)
|
|
|
|
if self.request.method == "DELETE":
|
|
try:
|
|
token = ScheduleExportAuthToken.objects.get(user_id=self.request.user.id, schedule_id=schedule.id)
|
|
write_resource_insight_log(instance=token, author=self.request.user, event=EntityEvent.DELETED)
|
|
token.delete()
|
|
except ScheduleExportAuthToken.DoesNotExist:
|
|
raise NotFound
|
|
|
|
return Response(status=status.HTTP_204_NO_CONTENT)
|
|
|
|
@action(detail=False, methods=["get"])
|
|
def notify_oncall_shift_freq_options(self, request):
|
|
options = []
|
|
for choice in OnCallSchedule.NotifyOnCallShiftFreq.choices:
|
|
options.append(
|
|
{
|
|
"value": choice[0],
|
|
"display_name": choice[1],
|
|
}
|
|
)
|
|
return Response(options)
|
|
|
|
@action(detail=False, methods=["get"])
|
|
def notify_empty_oncall_options(self, request):
|
|
options = []
|
|
for choice in OnCallSchedule.NotifyEmptyOnCall.choices:
|
|
options.append(
|
|
{
|
|
"value": choice[0],
|
|
"display_name": choice[1],
|
|
}
|
|
)
|
|
return Response(options)
|
|
|
|
@action(detail=False, methods=["get"])
|
|
def mention_options(self, request):
|
|
options = [
|
|
{
|
|
"value": False,
|
|
"display_name": "Inform in channel without mention",
|
|
},
|
|
{
|
|
"value": True,
|
|
"display_name": "Mention person in slack",
|
|
},
|
|
]
|
|
return Response(options)
|
|
|
|
@action(methods=["get"], detail=False)
|
|
def filters(self, request):
|
|
api_root = "/api/internal/v1/"
|
|
|
|
filter_options = [
|
|
{"name": "search", "type": "search"},
|
|
{
|
|
"name": "team",
|
|
"type": "team_select",
|
|
"href": api_root + "teams/",
|
|
"global": True,
|
|
},
|
|
{
|
|
"name": "mine",
|
|
"type": "boolean",
|
|
"display_name": "Mine",
|
|
"default": "true",
|
|
},
|
|
{
|
|
"name": "used",
|
|
"type": "boolean",
|
|
"display_name": "Used in escalations",
|
|
"default": "false",
|
|
},
|
|
{
|
|
"name": "type",
|
|
"type": "options",
|
|
"options": [
|
|
{"display_name": "API", "value": 0},
|
|
{"display_name": "Ical", "value": 1},
|
|
{"display_name": "Web", "value": 2},
|
|
],
|
|
},
|
|
]
|
|
|
|
return Response(filter_options)
|