Compare commits
3 Commits
VORKOUT-21
...
VORKOUT-18
Author | SHA1 | Date | |
---|---|---|---|
|
4aa2c12a48 | ||
|
0ab9727223 | ||
|
1d367a1bad |
@@ -11,18 +11,20 @@ from orm.tables.process import ps_node_table, node_link_table
|
||||
from orm.tables.process import NodeLinkStatus
|
||||
|
||||
|
||||
async def get_last_link_name_by_node_id(connection: AsyncConnection, ps_id: int) -> Optional[str]:
|
||||
|
||||
async def get_last_link_name_by_node_id(
|
||||
connection: AsyncConnection, ps_id: int
|
||||
) -> Optional[str]:
|
||||
"""
|
||||
Получает link_name из последней записи node_link по ps_id.
|
||||
Находит все node_id в ps_node по ps_id, затем ищет связи в node_link
|
||||
и возвращает link_name из самой последней записи.
|
||||
"""
|
||||
query = (
|
||||
select(node_link_table.c.link_name)
|
||||
.where(node_link_table.c.node_id.in_(select(ps_node_table.c.id).where(ps_node_table.c.ps_id == ps_id)))
|
||||
.order_by(desc(node_link_table.c.created_at))
|
||||
.limit(1)
|
||||
query = select(node_link_table.c.link_name).where(
|
||||
node_link_table.c.node_id.in_(
|
||||
select(ps_node_table.c.id).where(ps_node_table.c.ps_id == ps_id)
|
||||
)
|
||||
).order_by(desc(node_link_table.c.created_at)).limit(1)
|
||||
|
||||
result = await connection.execute(query)
|
||||
link_name = result.scalar_one_or_none()
|
||||
@@ -36,15 +38,12 @@ async def get_last_node_link_by_creator_and_ps_id(
|
||||
"""
|
||||
Получает последнюю созданную node_link для данного создателя и процесса.
|
||||
"""
|
||||
query = (
|
||||
select(node_link_table)
|
||||
.where(
|
||||
query = select(node_link_table).where(
|
||||
node_link_table.c.creator_id == creator_id,
|
||||
node_link_table.c.node_id.in_(select(ps_node_table.c.id).where(ps_node_table.c.id == node_link_id)),
|
||||
)
|
||||
.order_by(desc(node_link_table.c.created_at))
|
||||
.limit(1)
|
||||
node_link_table.c.node_id.in_(
|
||||
select(ps_node_table.c.id).where(ps_node_table.c.id == node_link_id)
|
||||
)
|
||||
).order_by(desc(node_link_table.c.created_at)).limit(1)
|
||||
|
||||
node_link_db_cursor = await connection.execute(query)
|
||||
node_link_data = node_link_db_cursor.mappings().one_or_none()
|
||||
@@ -56,9 +55,7 @@ async def get_last_node_link_by_creator_and_ps_id(
|
||||
|
||||
|
||||
async def create_node_link_schema(
|
||||
connection: AsyncConnection,
|
||||
validated_link_schema,
|
||||
creator_id: int,
|
||||
connection: AsyncConnection, validated_link_schema, creator_id: int,
|
||||
) -> Optional[NodeLink]:
|
||||
"""
|
||||
Создает нове поле в таблице process_schema_table.
|
||||
|
@@ -6,7 +6,7 @@ from datetime import datetime, timezone
|
||||
from sqlalchemy import insert, select, func, or_, and_, asc, desc
|
||||
from sqlalchemy.ext.asyncio import AsyncConnection
|
||||
|
||||
from orm.tables.process import process_schema_table, ProcessStatus
|
||||
from orm.tables.process import process_schema_table
|
||||
|
||||
from api.schemas.process.process_schema import ProcessSchema
|
||||
|
||||
@@ -50,9 +50,8 @@ async def get_process_schema_page_DTO(
|
||||
or_(process_schema_table.c.title.ilike(search_term), process_schema_table.c.description.ilike(search_term))
|
||||
)
|
||||
|
||||
filter_conditions = []
|
||||
|
||||
if filter_dto.filters:
|
||||
filter_conditions = []
|
||||
for field, values in filter_dto.filters.items():
|
||||
column = getattr(process_schema_table.c, field, None)
|
||||
if column is not None and values:
|
||||
@@ -61,9 +60,6 @@ async def get_process_schema_page_DTO(
|
||||
else:
|
||||
filter_conditions.append(column.in_(values))
|
||||
|
||||
if filter_dto.filters is None or "status" not in filter_dto.filters:
|
||||
filter_conditions.append(process_schema_table.c.status != "DELETED")
|
||||
|
||||
if filter_conditions:
|
||||
query = query.where(and_(*filter_conditions))
|
||||
|
||||
@@ -90,7 +86,7 @@ async def get_process_schema_page_DTO(
|
||||
or_(process_schema_table.c.title.ilike(search_term), process_schema_table.c.description.ilike(search_term))
|
||||
)
|
||||
|
||||
if filter_conditions:
|
||||
if filter_dto.filters and filter_conditions:
|
||||
count_query = count_query.where(and_(*filter_conditions))
|
||||
|
||||
result = await connection.execute(query)
|
||||
@@ -156,8 +152,11 @@ async def update_process_schema_by_id(connection: AsyncConnection, update_values
|
||||
await connection.commit()
|
||||
|
||||
|
||||
|
||||
async def update_process_schema_settings_by_id(
|
||||
connection: AsyncConnection, process_schema_id: int, node_data: Dict[str, Any]
|
||||
connection: AsyncConnection,
|
||||
process_schema_id: int,
|
||||
node_data: Dict[str, Any]
|
||||
):
|
||||
"""
|
||||
Добавляет новый узел в массив 'nodes' в настройках процесса.
|
||||
@@ -189,34 +188,24 @@ async def update_process_schema_settings_by_id(
|
||||
await connection.commit()
|
||||
|
||||
|
||||
async def get_last_created_process_schema(connection: AsyncConnection) -> Optional[int]:
|
||||
"""
|
||||
Получает ID последней созданной схемы процесса.
|
||||
"""
|
||||
query = select(process_schema_table.c.id).order_by(desc(process_schema_table.c.id)).limit(1)
|
||||
|
||||
result = await connection.execute(query)
|
||||
last_id = result.scalar_one_or_none()
|
||||
|
||||
return last_id
|
||||
|
||||
|
||||
async def create_process_schema(
|
||||
connection: AsyncConnection, creator_id: int, title: str, description: str
|
||||
connection: AsyncConnection, process_schema: ProcessSchema, creator_id: int
|
||||
) -> Optional[ProcessSchema]:
|
||||
"""
|
||||
Создает нове поле в таблице process_schema_table.
|
||||
"""
|
||||
query = insert(process_schema_table).values(
|
||||
title=title,
|
||||
description=description,
|
||||
owner_id=creator_id,
|
||||
title=process_schema.title,
|
||||
description=process_schema.description,
|
||||
owner_id=process_schema.owner_id,
|
||||
creator_id=creator_id,
|
||||
created_at=datetime.now(timezone.utc),
|
||||
settings={},
|
||||
status=ProcessStatus.ACTIVE.value,
|
||||
settings=process_schema.settings,
|
||||
status=process_schema.status.value,
|
||||
)
|
||||
|
||||
await connection.execute(query)
|
||||
|
||||
await connection.commit()
|
||||
|
||||
return process_schema
|
||||
|
@@ -8,8 +8,8 @@ from sqlalchemy.ext.asyncio import AsyncConnection
|
||||
from orm.tables.process import ps_node_table
|
||||
|
||||
from api.schemas.process.ps_node import Ps_Node
|
||||
from model_nodes.node_listen_models import ListenNodeCoreSchema
|
||||
from orm.tables.process import NodeStatus, NodeType
|
||||
from model_nodes.node_start_models import StartNodeCoreSchema
|
||||
from orm.tables.process import NodeStatus,NodeType
|
||||
|
||||
|
||||
async def get_ps_node_by_id(connection: AsyncConnection, id: int) -> Optional[Ps_Node]:
|
||||
@@ -31,7 +31,10 @@ async def get_ps_node_by_type_and_ps_id(connection: AsyncConnection, node_type:
|
||||
"""
|
||||
Получает ps_node по node_type и ps_id.
|
||||
"""
|
||||
query = select(ps_node_table).where(ps_node_table.c.node_type == node_type, ps_node_table.c.ps_id == ps_id)
|
||||
query = select(ps_node_table).where(
|
||||
ps_node_table.c.node_type == node_type,
|
||||
ps_node_table.c.ps_id == ps_id
|
||||
)
|
||||
|
||||
ps_node_db_cursor = await connection.execute(query)
|
||||
|
||||
@@ -41,17 +44,16 @@ async def get_ps_node_by_type_and_ps_id(connection: AsyncConnection, node_type:
|
||||
|
||||
return Ps_Node.model_validate(ps_node_data)
|
||||
|
||||
|
||||
async def create_ps_node_start_schema(
|
||||
connection: AsyncConnection, validated_listen_schema: ListenNodeCoreSchema, creator_id: int
|
||||
) -> Optional[ListenNodeCoreSchema]:
|
||||
connection: AsyncConnection, validated_start_schema: StartNodeCoreSchema, creator_id: int
|
||||
) -> Optional[Ps_Node]:
|
||||
"""
|
||||
Создает нове поле в таблице process_schema_table.
|
||||
"""
|
||||
query = insert(ps_node_table).values(
|
||||
ps_id=validated_listen_schema.ps_id,
|
||||
node_type=NodeType.LISTEN.value,
|
||||
settings=validated_listen_schema.data.model_dump(),
|
||||
ps_id=validated_start_schema.ps_id,
|
||||
node_type=NodeType.START.value,
|
||||
settings=validated_start_schema.data.model_dump(),
|
||||
creator_id=creator_id,
|
||||
created_at=datetime.now(timezone.utc),
|
||||
status=NodeStatus.ACTIVE.value,
|
||||
@@ -61,8 +63,7 @@ async def create_ps_node_start_schema(
|
||||
|
||||
await connection.commit()
|
||||
|
||||
# return validated_listen_schema
|
||||
|
||||
# return validated_start_schema
|
||||
|
||||
async def get_last_ps_node_by_creator_and_ps_id(
|
||||
connection: AsyncConnection, creator_id: int, ps_id: int
|
||||
@@ -70,12 +71,10 @@ async def get_last_ps_node_by_creator_and_ps_id(
|
||||
"""
|
||||
Получает последнюю созданную ps_node для данного создателя и процесса.
|
||||
"""
|
||||
query = (
|
||||
select(ps_node_table)
|
||||
.where(ps_node_table.c.creator_id == creator_id, ps_node_table.c.ps_id == ps_id)
|
||||
.order_by(desc(ps_node_table.c.created_at))
|
||||
.limit(1)
|
||||
)
|
||||
query = select(ps_node_table).where(
|
||||
ps_node_table.c.creator_id == creator_id,
|
||||
ps_node_table.c.ps_id == ps_id
|
||||
).order_by(desc(ps_node_table.c.created_at)).limit(1)
|
||||
|
||||
ps_node_db_cursor = await connection.execute(query)
|
||||
ps_node_data = ps_node_db_cursor.mappings().one_or_none()
|
||||
@@ -85,12 +84,9 @@ async def get_last_ps_node_by_creator_and_ps_id(
|
||||
|
||||
return Ps_Node.model_validate(ps_node_data)
|
||||
|
||||
|
||||
async def create_ps_node_schema(
|
||||
connection: AsyncConnection,
|
||||
validated_schema,
|
||||
creator_id: int,
|
||||
) -> Optional[ListenNodeCoreSchema]:
|
||||
connection: AsyncConnection, validated_schema, creator_id: int,
|
||||
) -> Optional[Ps_Node]:
|
||||
"""
|
||||
Создает нове поле в таблице process_schema_table.
|
||||
"""
|
||||
|
@@ -6,15 +6,7 @@ from api.endpoints.list_events import api_router as listevents_router
|
||||
from api.endpoints.process_schema import api_router as processschema_router
|
||||
from api.endpoints.ps_node import api_router as ps_node_router
|
||||
|
||||
list_of_routes = [
|
||||
auth_router,
|
||||
profile_router,
|
||||
account_router,
|
||||
keyring_router,
|
||||
listevents_router,
|
||||
processschema_router,
|
||||
ps_node_router,
|
||||
]
|
||||
list_of_routes = [auth_router, profile_router, account_router, keyring_router, listevents_router, processschema_router, ps_node_router]
|
||||
|
||||
__all__ = [
|
||||
"list_of_routes",
|
||||
|
@@ -22,7 +22,7 @@ api_router = APIRouter(
|
||||
|
||||
|
||||
@api_router.get("/{user_id}/{key_id}", dependencies=[Depends(bearer_schema)], response_model=AccountKeyring)
|
||||
async def get_keyring_endpoint(
|
||||
async def get_keyring_endpoint (
|
||||
key_id: str, connection: AsyncConnection = Depends(get_connection_dep), current_user=Depends(get_current_user)
|
||||
):
|
||||
authorize_user = await db_user_role_validation(connection, current_user)
|
||||
|
@@ -26,6 +26,7 @@ from api.services.user_role_validation import (
|
||||
|
||||
|
||||
from api.db.logic.ps_node import create_ps_node_schema
|
||||
|
||||
from api.db.logic.process_schema import update_process_schema_settings_by_id
|
||||
|
||||
from orm.tables.process import NodeType
|
||||
@@ -36,8 +37,6 @@ from core import VorkNodeRegistry
|
||||
|
||||
from model_nodes import ListenNodeData
|
||||
|
||||
from api.utils.node_counter import increment_node_counter
|
||||
|
||||
|
||||
api_router = APIRouter(
|
||||
prefix="/process_schema",
|
||||
@@ -45,7 +44,10 @@ api_router = APIRouter(
|
||||
)
|
||||
|
||||
|
||||
@api_router.get("", dependencies=[Depends(bearer_schema)], response_model=AllProcessSchemaResponse)
|
||||
@api_router.get("", dependencies=[Depends(bearer_schema)],
|
||||
|
||||
# response_model=AllProcessSchemaResponse
|
||||
)
|
||||
async def get_all_process_schema_endpoint(
|
||||
page: int = Query(1, description="Page number", gt=0),
|
||||
limit: int = Query(10, description="Number of items per page", gt=0),
|
||||
@@ -54,20 +56,12 @@ async def get_all_process_schema_endpoint(
|
||||
order_direction: Optional[str] = Query("asc", description="Sort direction (asc/desc)"),
|
||||
status_filter: Optional[List[str]] = Query(None, description="Filter by status"),
|
||||
owner_id: Optional[List[str]] = Query(None, description="Filter by owner id"),
|
||||
show_deleted: bool = Query(False, description="Show only deleted schemas"),
|
||||
connection: AsyncConnection = Depends(get_connection_dep),
|
||||
creator_id: Optional[int] = Query(None, description="Filter by creator id"),
|
||||
current_user=Depends(get_current_user),
|
||||
):
|
||||
if show_deleted:
|
||||
status_to_filter = ["DELETED"]
|
||||
elif status_filter:
|
||||
status_to_filter = status_filter
|
||||
else:
|
||||
status_to_filter = None
|
||||
|
||||
filters = {
|
||||
**({"status": status_to_filter} if status_to_filter else {}),
|
||||
**({"status": status_filter} if status_filter else {}),
|
||||
**({"owner_id": owner_id} if owner_id else {}),
|
||||
**({"creator_id": [str(creator_id)]} if creator_id else {}),
|
||||
}
|
||||
@@ -117,23 +111,27 @@ async def get_process_schema_endpoint(
|
||||
return to_camel_dict(process_schema_validation.model_dump())
|
||||
|
||||
|
||||
@api_router.post("", dependencies=[Depends(bearer_schema)], response_model=ProcessSchemaResponse)
|
||||
@api_router.post("", dependencies=[Depends(bearer_schema)],
|
||||
response_model=ProcessSchemaResponse
|
||||
)
|
||||
async def create_processschema_endpoint(
|
||||
process_schema: ProcessSchemaUpdate,
|
||||
connection: AsyncConnection = Depends(get_connection_dep),
|
||||
current_user=Depends(get_current_user),
|
||||
):
|
||||
user_validation = await get_user_by_login(connection, current_user)
|
||||
process_schema_validation = await get_process_schema_by_title(connection, process_schema.title)
|
||||
|
||||
current_node_counter = increment_node_counter()
|
||||
title = f"Новая схема {current_node_counter}"
|
||||
if process_schema_validation is None:
|
||||
|
||||
description = "Default description"
|
||||
await create_process_schema(connection, process_schema, user_validation.id)
|
||||
process_schema_new = await get_process_schema_by_title(connection, process_schema.title)
|
||||
|
||||
await create_process_schema(connection, user_validation.id, title, description)
|
||||
|
||||
process_schema_new = await get_process_schema_by_title(connection, title)
|
||||
|
||||
start_node_data = ListenNodeData(ps_id=process_schema_new.id, node_type=NodeType.START.value, is_start="True")
|
||||
start_node_data = ListenNodeData(
|
||||
ps_id=process_schema_new.id,
|
||||
node_type=NodeType.START.value,
|
||||
is_start="True"
|
||||
)
|
||||
|
||||
start_node_links = {}
|
||||
|
||||
@@ -143,6 +141,7 @@ async def create_processschema_endpoint(
|
||||
|
||||
node_descriptor = vork_node.form()
|
||||
|
||||
|
||||
start_node = vork_node(data=start_node_data.model_dump(), links=start_node_links)
|
||||
|
||||
validated_start_schema = start_node.validate()
|
||||
@@ -156,31 +155,36 @@ async def create_processschema_endpoint(
|
||||
node_type=NodeType.LISTEN.value,
|
||||
data=validated_start_schema.data.model_dump(),
|
||||
from_node=None,
|
||||
links=None,
|
||||
)
|
||||
links=None)
|
||||
|
||||
settings_dict = {"node": node.model_dump(mode="json")}
|
||||
settings_dict = {"node": node.model_dump(mode='json')}
|
||||
|
||||
await update_process_schema_settings_by_id(connection, process_schema_new.id, settings_dict)
|
||||
|
||||
process_schema_new = await get_process_schema_by_title(connection, title)
|
||||
process_schema_new = await get_process_schema_by_title(connection, process_schema.title)
|
||||
|
||||
ps_node_front_response = Ps_NodeFrontResponse(
|
||||
description=node_descriptor.model_dump(),
|
||||
description=node_descriptor,
|
||||
node=Ps_NodeFrontResponseNode(
|
||||
id=db_start_schema.id, node_type=NodeType.LISTEN.value, data=validated_start_schema.data.model_dump()
|
||||
),
|
||||
link=None,
|
||||
)
|
||||
id=db_start_schema.id,
|
||||
node_type=NodeType.LISTEN.value,
|
||||
data=validated_start_schema.data.model_dump()),
|
||||
link=None)
|
||||
|
||||
|
||||
response_data = {
|
||||
"process_schema": process_schema_new.model_dump(),
|
||||
"node_listen": ps_node_front_response.model_dump(),
|
||||
}
|
||||
"node_listen": ps_node_front_response.model_dump()}
|
||||
|
||||
return to_camel_dict(response_data)
|
||||
|
||||
|
||||
else:
|
||||
raise HTTPException(
|
||||
status_code=status.HTTP_400_BAD_REQUEST, detail="An process schema with this information already exists."
|
||||
)
|
||||
|
||||
|
||||
@api_router.put("/{process_schema_id}", dependencies=[Depends(bearer_schema)], response_model=ProcessSchema)
|
||||
async def update_process_schema_endpoint(
|
||||
process_schema_id: int,
|
||||
@@ -209,7 +213,7 @@ async def update_process_schema_endpoint(
|
||||
return process_schema
|
||||
|
||||
|
||||
@api_router.delete("/{process_schema_id}", dependencies=[Depends(bearer_schema)], status_code=status.HTTP_200_OK)
|
||||
@api_router.delete("/{process_schema_id}", dependencies=[Depends(bearer_schema)], response_model=ProcessSchema)
|
||||
async def delete_process_schema_endpoint(
|
||||
process_schema_id: int,
|
||||
connection: AsyncConnection = Depends(get_connection_dep),
|
||||
@@ -233,6 +237,6 @@ async def delete_process_schema_endpoint(
|
||||
|
||||
await update_process_schema_by_id(connection, updated_values, process_schema_validation)
|
||||
|
||||
await get_process_schema_by_id(connection, process_schema_id)
|
||||
process_schema = await get_process_schema_by_id(connection, process_schema_id)
|
||||
|
||||
return HTTPException(status_code=status.HTTP_200_OK, detail="Process schema deleted successfully")
|
||||
return process_schema
|
||||
|
@@ -1,4 +1,4 @@
|
||||
from fastapi import APIRouter, Depends, HTTPException, status
|
||||
from fastapi import APIRouter, Depends, HTTPException,status
|
||||
|
||||
from sqlalchemy.ext.asyncio import AsyncConnection
|
||||
|
||||
@@ -23,18 +23,23 @@ from model_nodes import VorkNodeLinkData
|
||||
from api.utils.to_camel_dict import to_camel_dict
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
api_router = APIRouter(
|
||||
prefix="/ps_node",
|
||||
tags=["ps node"],
|
||||
)
|
||||
|
||||
|
||||
@api_router.post("", dependencies=[Depends(bearer_schema)], response_model=Ps_NodeFrontResponse)
|
||||
@api_router.post("", dependencies=[Depends(bearer_schema)],response_model=Ps_NodeFrontResponse)
|
||||
|
||||
async def create_ps_node_endpoint(
|
||||
ps_node: Ps_NodeRequest,
|
||||
connection: AsyncConnection = Depends(get_connection_dep),
|
||||
current_user=Depends(get_current_user),
|
||||
):
|
||||
|
||||
user_validation = await get_user_by_login(connection, current_user)
|
||||
|
||||
registery = VorkNodeRegistry()
|
||||
@@ -53,22 +58,32 @@ async def create_ps_node_endpoint(
|
||||
except Exception as e:
|
||||
raise HTTPException(status_code=status.HTTP_400_BAD_REQUEST, detail=str(e))
|
||||
|
||||
|
||||
db_ps_node = await create_ps_node_schema(connection, node_instance_validated, user_validation.id)
|
||||
link_name = await get_last_link_name_by_node_id(connection, db_ps_node.ps_id)
|
||||
|
||||
|
||||
|
||||
link_data = VorkNodeLinkData(
|
||||
parent_port_number=node_instance_validated.parent_port_number,
|
||||
to_id=db_ps_node.id,
|
||||
from_id=node_instance_validated.parent_id,
|
||||
last_link_name=link_name,
|
||||
)
|
||||
last_link_name=link_name)
|
||||
|
||||
|
||||
|
||||
link = VorkNodeLink(data=link_data.model_dump())
|
||||
|
||||
validated_link = link.validate()
|
||||
|
||||
|
||||
|
||||
|
||||
db_node_link = await create_node_link_schema(connection, validated_link, user_validation.id)
|
||||
|
||||
|
||||
|
||||
|
||||
links_settings = ProcessSchemaSettingsNodeLink(
|
||||
id=db_node_link.id,
|
||||
link_name=db_node_link.link_name,
|
||||
@@ -82,21 +97,21 @@ async def create_ps_node_endpoint(
|
||||
node_type=db_ps_node.node_type,
|
||||
data=node_instance_validated.data.model_dump(),
|
||||
from_node=None,
|
||||
links=[{"links": links_settings.model_dump()}],
|
||||
)
|
||||
links=links_settings.model_dump())
|
||||
|
||||
settings_dict = {"node": node_settings.model_dump(mode="json")}
|
||||
|
||||
settings_dict = {"node": node_settings.model_dump(mode='json')}
|
||||
|
||||
await update_process_schema_settings_by_id(connection, db_ps_node.ps_id, settings_dict)
|
||||
|
||||
|
||||
ps_node_front_response = Ps_NodeFrontResponse(
|
||||
description=node_descriptor.model_dump(),
|
||||
description=node_descriptor,
|
||||
node=Ps_NodeFrontResponseNode(
|
||||
id=db_ps_node.id,
|
||||
node_type=db_ps_node.node_type,
|
||||
data=to_camel_dict(node_instance_validated.data.model_dump()),
|
||||
),
|
||||
links=[{"links": links_settings.model_dump()}],
|
||||
)
|
||||
data=to_camel_dict(node_instance_validated.data.model_dump())),
|
||||
link=links_settings.model_dump())
|
||||
|
||||
|
||||
return ps_node_front_response
|
||||
|
@@ -10,7 +10,7 @@ from api.schemas.base import Base
|
||||
class ProcessSchemaUpdate(Base):
|
||||
title: Optional[str] = Field(None, max_length=100)
|
||||
description: Optional[str] = None
|
||||
# owner_id: Optional[int] = None
|
||||
owner_id: Optional[int] = None
|
||||
settings: Optional[Dict[str, Any]] = None
|
||||
status: Optional[ProcessStatus] = None
|
||||
|
||||
|
@@ -1,5 +1,5 @@
|
||||
from datetime import datetime
|
||||
from typing import Any, Dict, Optional, List
|
||||
from typing import Any, Dict, Optional
|
||||
|
||||
from orm.tables.process import ProcessStatus, NodeType
|
||||
from pydantic import Field
|
||||
@@ -31,8 +31,8 @@ class ProcessSchemaSettingsNode(Base):
|
||||
id: int
|
||||
node_type: NodeType
|
||||
from_node: Optional[Dict[str, Any]] = None
|
||||
data: Dict[str, Any] # Переименовано с 'from' на 'from_node'
|
||||
links: Optional[List[Dict[str, Any]]] = None
|
||||
data: Dict[str, Any]# Переименовано с 'from' на 'from_node'
|
||||
links: Optional[ProcessSchemaSettingsNodeLink] = None
|
||||
|
||||
|
||||
class ProcessSchemaResponse(Base):
|
||||
|
@@ -1,7 +1,8 @@
|
||||
from datetime import datetime
|
||||
from typing import Any, Dict, Optional, List
|
||||
from typing import Any, Dict, Optional
|
||||
|
||||
from orm.tables.process import NodeStatus, NodeType
|
||||
from pydantic import Field
|
||||
|
||||
from api.schemas.base import Base
|
||||
|
||||
@@ -10,7 +11,6 @@ class Ps_NodeRequest(Base):
|
||||
data: Dict[str, Any]
|
||||
links: Dict[str, Any]
|
||||
|
||||
|
||||
class Ps_Node(Base):
|
||||
id: int
|
||||
ps_id: int
|
||||
@@ -21,6 +21,7 @@ class Ps_Node(Base):
|
||||
status: NodeStatus
|
||||
|
||||
|
||||
|
||||
class Ps_NodeFrontResponseLink(Base):
|
||||
id: int
|
||||
link_name: str
|
||||
@@ -32,10 +33,10 @@ class Ps_NodeFrontResponseLink(Base):
|
||||
class Ps_NodeFrontResponseNode(Base):
|
||||
id: int
|
||||
node_type: NodeType
|
||||
data: Dict[str, Any] # Переименовано с 'from' на 'from_node'
|
||||
data: Dict[str, Any]# Переименовано с 'from' на 'from_node'
|
||||
|
||||
|
||||
class Ps_NodeFrontResponse(Base):
|
||||
description: Optional[Dict[str, Any]] = None
|
||||
node: Optional[Ps_NodeFrontResponseNode] = None
|
||||
links: Optional[List[Dict[str, Any]]] = None
|
||||
link: Optional[Ps_NodeFrontResponseLink] = None
|
||||
|
@@ -1,52 +0,0 @@
|
||||
import json
|
||||
from pathlib import Path
|
||||
from typing import Dict
|
||||
|
||||
|
||||
# Путь к файлу счётчика (в корне проекта)
|
||||
COUNTER_FILE_PATH = Path(__file__).parent.parent.parent / "node_counter.json"
|
||||
|
||||
|
||||
def get_node_counter() -> int:
|
||||
"""
|
||||
Открывает JSON файл и возвращает значение node_counter.
|
||||
Если файл не существует, создаёт его со значением по умолчанию 0.
|
||||
|
||||
Returns:
|
||||
int: Текущее значение счётчика узлов
|
||||
"""
|
||||
|
||||
if not COUNTER_FILE_PATH.exists():
|
||||
initial_data: Dict[str, int] = {"node_counter": 0}
|
||||
with open(COUNTER_FILE_PATH, "w", encoding="utf-8") as f:
|
||||
json.dump(initial_data, f, indent=2, ensure_ascii=False)
|
||||
return 0
|
||||
|
||||
try:
|
||||
with open(COUNTER_FILE_PATH, "r", encoding="utf-8") as f:
|
||||
data = json.load(f)
|
||||
return data.get("node_counter", 0)
|
||||
except (json.JSONDecodeError, IOError):
|
||||
initial_data = {"node_counter": 0}
|
||||
with open(COUNTER_FILE_PATH, "w", encoding="utf-8") as f:
|
||||
json.dump(initial_data, f, indent=2, ensure_ascii=False)
|
||||
return 0
|
||||
|
||||
|
||||
def increment_node_counter() -> int:
|
||||
"""
|
||||
Увеличивает значение node_counter на 1, сохраняет в файл и возвращает новое значение.
|
||||
|
||||
Returns:
|
||||
int: Новое значение счётчика (старое значение + 1)
|
||||
"""
|
||||
|
||||
current_value = get_node_counter()
|
||||
|
||||
new_value = current_value + 1
|
||||
|
||||
data: Dict[str, int] = {"node_counter": new_value}
|
||||
with open(COUNTER_FILE_PATH, "w", encoding="utf-8") as f:
|
||||
json.dump(data, f, indent=2, ensure_ascii=False)
|
||||
|
||||
return new_value
|
@@ -1,6 +1,5 @@
|
||||
from pydantic.alias_generators import to_camel
|
||||
|
||||
|
||||
def to_camel_dict(obj):
|
||||
if isinstance(obj, dict):
|
||||
return {to_camel(key): to_camel_dict(value) for key, value in obj.items()}
|
||||
|
2
api/poetry.lock
generated
2
api/poetry.lock
generated
@@ -2020,7 +2020,7 @@ sqlalchemy = "^2.0.43"
|
||||
type = "git"
|
||||
url = "http://88.86.199.167:3000/Nox/CORE.git"
|
||||
reference = "HEAD"
|
||||
resolved_reference = "43c139512928ab3a4767e771c8e41e39930599ad"
|
||||
resolved_reference = "b3896e8b5dbed2d609c8ac257419d5492c9e7b8d"
|
||||
|
||||
[[package]]
|
||||
name = "watchfiles"
|
||||
|
@@ -18,7 +18,7 @@ dependencies = [
|
||||
"python-multipart (>=0.0.20,<0.0.21)",
|
||||
"requests (>=2.31.0,<3.0.0)",
|
||||
"fastapi-jwt-auth @ git+https://github.com/vvpreo/fastapi-jwt-auth",
|
||||
"vork-core @ git+http://88.86.199.167:3000/Nox/CORE.git",
|
||||
"core-library @ git+https://gitea.heado.ru/Vorkout/core.git@VORKOUT-18",
|
||||
]
|
||||
|
||||
|
||||
|
Reference in New Issue
Block a user