345 lines
13 KiB
Python
345 lines
13 KiB
Python
# Python imports
|
|
import json
|
|
|
|
# Django imports
|
|
from django.core.serializers.json import DjangoJSONEncoder
|
|
from django.db.models import F, Func, OuterRef, Q, Prefetch, Exists, Subquery
|
|
from django.utils import timezone
|
|
from django.utils.decorators import method_decorator
|
|
from django.views.decorators.gzip import gzip_page
|
|
|
|
# Third Party imports
|
|
from rest_framework import status
|
|
from rest_framework.response import Response
|
|
|
|
from plane.app.permissions import ProjectEntityPermission
|
|
from plane.app.serializers import (
|
|
IssueFlatSerializer,
|
|
IssueSerializer,
|
|
IssueDetailSerializer,
|
|
)
|
|
from plane.bgtasks.issue_activities_task import issue_activity
|
|
from plane.db.models import (
|
|
Issue,
|
|
FileAsset,
|
|
IssueLink,
|
|
IssueSubscriber,
|
|
IssueReaction,
|
|
CycleIssue,
|
|
)
|
|
from plane.utils.grouper import (
|
|
issue_group_values,
|
|
issue_on_results,
|
|
issue_queryset_grouper,
|
|
)
|
|
from plane.utils.issue_filters import issue_filters
|
|
from plane.utils.order_queryset import order_issue_queryset
|
|
from plane.utils.paginator import GroupedOffsetPaginator, SubGroupedOffsetPaginator
|
|
from plane.app.permissions import allow_permission, ROLE
|
|
from plane.utils.error_codes import ERROR_CODES
|
|
from plane.utils.host import base_host
|
|
|
|
# Module imports
|
|
from .. import BaseViewSet, BaseAPIView
|
|
|
|
|
|
class IssueArchiveViewSet(BaseViewSet):
|
|
serializer_class = IssueFlatSerializer
|
|
model = Issue
|
|
|
|
def get_queryset(self):
|
|
return (
|
|
Issue.objects.annotate(
|
|
sub_issues_count=Issue.objects.filter(parent=OuterRef("id"))
|
|
.order_by()
|
|
.annotate(count=Func(F("id"), function="Count"))
|
|
.values("count")
|
|
)
|
|
.filter(deleted_at__isnull=True)
|
|
.filter(archived_at__isnull=False)
|
|
.filter(project_id=self.kwargs.get("project_id"))
|
|
.filter(workspace__slug=self.kwargs.get("slug"))
|
|
.select_related("workspace", "project", "state", "parent")
|
|
.prefetch_related("assignees", "labels", "issue_module__module")
|
|
.annotate(
|
|
cycle_id=Subquery(
|
|
CycleIssue.objects.filter(
|
|
issue=OuterRef("id"), deleted_at__isnull=True
|
|
).values("cycle_id")[:1]
|
|
)
|
|
)
|
|
.annotate(
|
|
link_count=IssueLink.objects.filter(issue=OuterRef("id"))
|
|
.order_by()
|
|
.annotate(count=Func(F("id"), function="Count"))
|
|
.values("count")
|
|
)
|
|
.annotate(
|
|
attachment_count=FileAsset.objects.filter(
|
|
issue_id=OuterRef("id"),
|
|
entity_type=FileAsset.EntityTypeContext.ISSUE_ATTACHMENT,
|
|
)
|
|
.order_by()
|
|
.annotate(count=Func(F("id"), function="Count"))
|
|
.values("count")
|
|
)
|
|
.annotate(
|
|
sub_issues_count=Issue.issue_objects.filter(parent=OuterRef("id"))
|
|
.order_by()
|
|
.annotate(count=Func(F("id"), function="Count"))
|
|
.values("count")
|
|
)
|
|
)
|
|
|
|
@method_decorator(gzip_page)
|
|
@allow_permission([ROLE.ADMIN, ROLE.MEMBER])
|
|
def list(self, request, slug, project_id):
|
|
filters = issue_filters(request.query_params, "GET")
|
|
show_sub_issues = request.GET.get("show_sub_issues", "true")
|
|
|
|
order_by_param = request.GET.get("order_by", "-created_at")
|
|
|
|
issue_queryset = self.get_queryset().filter(**filters)
|
|
|
|
issue_queryset = (
|
|
issue_queryset
|
|
if show_sub_issues == "true"
|
|
else issue_queryset.filter(parent__isnull=True)
|
|
)
|
|
# Issue queryset
|
|
issue_queryset, order_by_param = order_issue_queryset(
|
|
issue_queryset=issue_queryset, order_by_param=order_by_param
|
|
)
|
|
|
|
# Group by
|
|
group_by = request.GET.get("group_by", False)
|
|
sub_group_by = request.GET.get("sub_group_by", False)
|
|
|
|
# issue queryset
|
|
issue_queryset = issue_queryset_grouper(
|
|
queryset=issue_queryset, group_by=group_by, sub_group_by=sub_group_by
|
|
)
|
|
|
|
if group_by:
|
|
# Check group and sub group value paginate
|
|
if sub_group_by:
|
|
if group_by == sub_group_by:
|
|
return Response(
|
|
{
|
|
"error": "Group by and sub group by cannot have same parameters"
|
|
},
|
|
status=status.HTTP_400_BAD_REQUEST,
|
|
)
|
|
else:
|
|
# group and sub group pagination
|
|
return self.paginate(
|
|
request=request,
|
|
order_by=order_by_param,
|
|
queryset=issue_queryset,
|
|
on_results=lambda issues: issue_on_results(
|
|
group_by=group_by, issues=issues, sub_group_by=sub_group_by
|
|
),
|
|
paginator_cls=SubGroupedOffsetPaginator,
|
|
group_by_fields=issue_group_values(
|
|
field=group_by,
|
|
slug=slug,
|
|
project_id=project_id,
|
|
filters=filters,
|
|
),
|
|
sub_group_by_fields=issue_group_values(
|
|
field=sub_group_by,
|
|
slug=slug,
|
|
project_id=project_id,
|
|
filters=filters,
|
|
),
|
|
group_by_field_name=group_by,
|
|
sub_group_by_field_name=sub_group_by,
|
|
count_filter=Q(
|
|
Q(issue_intake__status=1)
|
|
| Q(issue_intake__status=-1)
|
|
| Q(issue_intake__status=2)
|
|
| Q(issue_intake__isnull=True),
|
|
archived_at__isnull=True,
|
|
is_draft=False,
|
|
),
|
|
)
|
|
# Group Paginate
|
|
else:
|
|
# Group paginate
|
|
return self.paginate(
|
|
request=request,
|
|
order_by=order_by_param,
|
|
queryset=issue_queryset,
|
|
on_results=lambda issues: issue_on_results(
|
|
group_by=group_by, issues=issues, sub_group_by=sub_group_by
|
|
),
|
|
paginator_cls=GroupedOffsetPaginator,
|
|
group_by_fields=issue_group_values(
|
|
field=group_by,
|
|
slug=slug,
|
|
project_id=project_id,
|
|
filters=filters,
|
|
),
|
|
group_by_field_name=group_by,
|
|
count_filter=Q(
|
|
Q(issue_intake__status=1)
|
|
| Q(issue_intake__status=-1)
|
|
| Q(issue_intake__status=2)
|
|
| Q(issue_intake__isnull=True),
|
|
archived_at__isnull=True,
|
|
is_draft=False,
|
|
),
|
|
)
|
|
else:
|
|
# List Paginate
|
|
return self.paginate(
|
|
order_by=order_by_param,
|
|
request=request,
|
|
queryset=issue_queryset,
|
|
on_results=lambda issues: issue_on_results(
|
|
group_by=group_by, issues=issues, sub_group_by=sub_group_by
|
|
),
|
|
)
|
|
|
|
@allow_permission([ROLE.ADMIN, ROLE.MEMBER])
|
|
def retrieve(self, request, slug, project_id, pk=None):
|
|
issue = (
|
|
self.get_queryset()
|
|
.filter(pk=pk)
|
|
.prefetch_related(
|
|
Prefetch(
|
|
"issue_reactions",
|
|
queryset=IssueReaction.objects.select_related("issue", "actor"),
|
|
)
|
|
)
|
|
.prefetch_related(
|
|
Prefetch(
|
|
"issue_link",
|
|
queryset=IssueLink.objects.select_related("created_by"),
|
|
)
|
|
)
|
|
.annotate(
|
|
is_subscribed=Exists(
|
|
IssueSubscriber.objects.filter(
|
|
workspace__slug=slug,
|
|
project_id=project_id,
|
|
issue_id=OuterRef("pk"),
|
|
subscriber=request.user,
|
|
)
|
|
)
|
|
)
|
|
).first()
|
|
if not issue:
|
|
return Response(
|
|
{"error": "The required object does not exist."},
|
|
status=status.HTTP_404_NOT_FOUND,
|
|
)
|
|
serializer = IssueDetailSerializer(issue, expand=self.expand)
|
|
return Response(serializer.data, status=status.HTTP_200_OK)
|
|
|
|
@allow_permission([ROLE.ADMIN, ROLE.MEMBER])
|
|
def archive(self, request, slug, project_id, pk=None):
|
|
issue = Issue.issue_objects.get(
|
|
workspace__slug=slug, project_id=project_id, pk=pk
|
|
)
|
|
if issue.state.group not in ["completed", "cancelled"]:
|
|
return Response(
|
|
{"error": "Can only archive completed or cancelled state group issue"},
|
|
status=status.HTTP_400_BAD_REQUEST,
|
|
)
|
|
issue_activity.delay(
|
|
type="issue.activity.updated",
|
|
requested_data=json.dumps(
|
|
{"archived_at": str(timezone.now().date()), "automation": False}
|
|
),
|
|
actor_id=str(request.user.id),
|
|
issue_id=str(issue.id),
|
|
project_id=str(project_id),
|
|
current_instance=json.dumps(
|
|
IssueSerializer(issue).data, cls=DjangoJSONEncoder
|
|
),
|
|
epoch=int(timezone.now().timestamp()),
|
|
notification=True,
|
|
origin=base_host(request=request, is_app=True),
|
|
)
|
|
issue.archived_at = timezone.now().date()
|
|
issue.save()
|
|
|
|
return Response(
|
|
{"archived_at": str(issue.archived_at)}, status=status.HTTP_200_OK
|
|
)
|
|
|
|
@allow_permission([ROLE.ADMIN, ROLE.MEMBER])
|
|
def unarchive(self, request, slug, project_id, pk=None):
|
|
issue = Issue.objects.get(
|
|
workspace__slug=slug,
|
|
project_id=project_id,
|
|
archived_at__isnull=False,
|
|
pk=pk,
|
|
)
|
|
issue_activity.delay(
|
|
type="issue.activity.updated",
|
|
requested_data=json.dumps({"archived_at": None}),
|
|
actor_id=str(request.user.id),
|
|
issue_id=str(issue.id),
|
|
project_id=str(project_id),
|
|
current_instance=json.dumps(
|
|
IssueSerializer(issue).data, cls=DjangoJSONEncoder
|
|
),
|
|
epoch=int(timezone.now().timestamp()),
|
|
notification=True,
|
|
origin=base_host(request=request, is_app=True),
|
|
)
|
|
issue.archived_at = None
|
|
issue.save()
|
|
|
|
return Response(status=status.HTTP_204_NO_CONTENT)
|
|
|
|
|
|
class BulkArchiveIssuesEndpoint(BaseAPIView):
|
|
permission_classes = [ProjectEntityPermission]
|
|
|
|
@allow_permission([ROLE.ADMIN, ROLE.MEMBER])
|
|
def post(self, request, slug, project_id):
|
|
issue_ids = request.data.get("issue_ids", [])
|
|
|
|
if not len(issue_ids):
|
|
return Response(
|
|
{"error": "Issue IDs are required"}, status=status.HTTP_400_BAD_REQUEST
|
|
)
|
|
|
|
issues = Issue.objects.filter(
|
|
workspace__slug=slug, project_id=project_id, pk__in=issue_ids
|
|
).select_related("state")
|
|
bulk_archive_issues = []
|
|
for issue in issues:
|
|
if issue.state.group not in ["completed", "cancelled"]:
|
|
return Response(
|
|
{
|
|
"error_code": ERROR_CODES["INVALID_ARCHIVE_STATE_GROUP"],
|
|
"error_message": "INVALID_ARCHIVE_STATE_GROUP",
|
|
},
|
|
status=status.HTTP_400_BAD_REQUEST,
|
|
)
|
|
issue_activity.delay(
|
|
type="issue.activity.updated",
|
|
requested_data=json.dumps(
|
|
{"archived_at": str(timezone.now().date()), "automation": False}
|
|
),
|
|
actor_id=str(request.user.id),
|
|
issue_id=str(issue.id),
|
|
project_id=str(project_id),
|
|
current_instance=json.dumps(
|
|
IssueSerializer(issue).data, cls=DjangoJSONEncoder
|
|
),
|
|
epoch=int(timezone.now().timestamp()),
|
|
notification=True,
|
|
origin=base_host(request=request, is_app=True),
|
|
)
|
|
issue.archived_at = timezone.now().date()
|
|
bulk_archive_issues.append(issue)
|
|
Issue.objects.bulk_update(bulk_archive_issues, ["archived_at"])
|
|
|
|
return Response(
|
|
{"archived_at": str(timezone.now().date())}, status=status.HTTP_200_OK
|
|
)
|