Skip to content
This repository was archived by the owner on Jun 5, 2025. It is now read-only.

Commit d5c53c9

Browse files
committed
Removed no longer used code.
1 parent b5638fa commit d5c53c9

File tree

32 files changed

+210
-1138
lines changed

32 files changed

+210
-1138
lines changed

src/codegate/llm_utils/__init__.py

Lines changed: 0 additions & 3 deletions
This file was deleted.

src/codegate/llm_utils/llmclient.py

Lines changed: 0 additions & 155 deletions
This file was deleted.

src/codegate/muxing/router.py

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -11,6 +11,8 @@
1111
from codegate.providers.fim_analyzer import FIMAnalyzer
1212
from codegate.providers.registry import ProviderRegistry
1313
from codegate.workspaces.crud import WorkspaceCrud
14+
from codegate.types.openai import ChatCompletionRequest
15+
1416

1517
logger = structlog.get_logger("codegate")
1618

src/codegate/pipeline/base.py

Lines changed: 11 additions & 11 deletions
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,7 @@
1212
from codegate.db.models import Alert, AlertSeverity, Output, Prompt
1313
from codegate.extract_snippets.message_extractor import CodeSnippet
1414
from codegate.pipeline.secrets.manager import SecretsManager
15-
from codegate.types.common import ChatCompletionRequest, ModelResponse
15+
1616

1717
logger = structlog.get_logger("codegate")
1818

@@ -90,7 +90,7 @@ def add_alert(
9090
# logger.debug(f"Added alert to context: {self.alerts_raised[-1]}")
9191

9292
def add_input_request(
93-
self, normalized_request: ChatCompletionRequest, is_fim_request: bool, provider: str
93+
self, normalized_request: Any, is_fim_request: bool, provider: str
9494
) -> None:
9595
try:
9696
if self.prompt_id is None:
@@ -109,7 +109,7 @@ def add_input_request(
109109
except Exception as e:
110110
logger.warning(f"Failed to serialize input request: {normalized_request}", error=str(e))
111111

112-
def add_output(self, model_response: ModelResponse) -> None:
112+
def add_output(self, model_response: Any) -> None:
113113
try:
114114
if self.prompt_id is None:
115115
logger.warning(f"Tried to record output without response: {model_response}")
@@ -152,7 +152,7 @@ class PipelineResult:
152152
or a response to return to the client.
153153
"""
154154

155-
request: Optional[ChatCompletionRequest] = None
155+
request: Optional[Any] = None
156156
response: Optional[PipelineResponse] = None
157157
context: Optional[PipelineContext] = None
158158
error_message: Optional[str] = None
@@ -183,13 +183,13 @@ def name(self) -> str:
183183

184184
@staticmethod
185185
def get_last_user_message(
186-
request: ChatCompletionRequest,
186+
request: Any,
187187
) -> Optional[tuple[str, int]]:
188188
"""
189189
Get the last user message and its index from the request.
190190
191191
Args:
192-
request (ChatCompletionRequest): The chat completion request to process
192+
request (Any): The chat completion request to process
193193
194194
Returns:
195195
Optional[tuple[str, int]]: A tuple containing the message content and
@@ -207,13 +207,13 @@ def get_last_user_message(
207207

208208
@staticmethod
209209
def get_last_user_message_block(
210-
request: ChatCompletionRequest,
210+
request: Any,
211211
) -> Optional[tuple[str, int]]:
212212
"""
213213
Get the last block of consecutive 'user' messages from the request.
214214
215215
Args:
216-
request (ChatCompletionRequest): The chat completion request to process
216+
request (Any): The chat completion request to process
217217
218218
Returns:
219219
Optional[str, int]: A string containing all consecutive user messages in the
@@ -237,7 +237,7 @@ def get_last_user_message_block(
237237

238238
@abstractmethod
239239
async def process(
240-
self, request: ChatCompletionRequest, context: PipelineContext
240+
self, request: Any, context: PipelineContext
241241
) -> PipelineResult:
242242
"""Process a request and return either modified request or response stream"""
243243
pass
@@ -266,7 +266,7 @@ def __init__(
266266

267267
async def process_request(
268268
self,
269-
request: ChatCompletionRequest,
269+
request: Any,
270270
provider: str,
271271
model: str,
272272
api_key: Optional[str] = None,
@@ -336,7 +336,7 @@ def _create_instance(self, client_type: ClientType) -> InputPipelineInstance:
336336

337337
async def process_request(
338338
self,
339-
request: ChatCompletionRequest,
339+
request: Any,
340340
provider: str,
341341
model: str,
342342
api_key: Optional[str] = None,

src/codegate/pipeline/cli/cli.py

Lines changed: 3 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,6 @@
11
import re
22
import shlex
3-
from typing import Optional
3+
from typing import Any, Optional
44

55

66
from codegate.clients.clients import ClientType
@@ -11,7 +11,6 @@
1111
PipelineStep,
1212
)
1313
from codegate.pipeline.cli.commands import CustomInstructions, Version, Workspace
14-
from codegate.types.common import ChatCompletionRequest
1514

1615
codegate_regex = re.compile(r"^codegate(?:\s+(.*))?", re.IGNORECASE)
1716

@@ -109,14 +108,14 @@ def name(self) -> str:
109108
return "codegate-cli"
110109

111110
async def process(
112-
self, request: ChatCompletionRequest, context: PipelineContext
111+
self, request: Any, context: PipelineContext
113112
) -> PipelineResult:
114113
"""
115114
Checks if the last user message contains "codegate" and process the command.
116115
This short-circuits the pipeline if the message is found.
117116
118117
Args:
119-
request (ChatCompletionRequest): The chat completion request to process
118+
request (Any): The chat completion request to process
120119
context (PipelineContext): The current pipeline context
121120
122121
Returns:

src/codegate/pipeline/codegate_context_retriever/codegate.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,7 @@
11
import itertools
22
import json
33
import re
4+
from typing import Any
45

56
import structlog
67

@@ -13,7 +14,6 @@
1314
PipelineStep,
1415
)
1516
from codegate.storage.storage_engine import StorageEngine
16-
from codegate.types.common import ChatCompletionRequest
1717
from codegate.utils.package_extractor import PackageExtractor
1818
from codegate.utils.utils import generate_vector_string
1919

@@ -71,7 +71,7 @@ def generate_context_str(self, objects: list[object], context: PipelineContext)
7171
return context_str
7272

7373
async def process( # noqa: C901
74-
self, request: ChatCompletionRequest, context: PipelineContext
74+
self, request: Any, context: PipelineContext
7575
) -> PipelineResult:
7676
"""
7777
Use RAG DB to add context to the user request

src/codegate/pipeline/comment/output.py

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,4 @@
1-
from typing import Optional
1+
from typing import Any, Optional
22
from urllib.parse import quote
33

44
import structlog
@@ -14,7 +14,7 @@
1414
# from codegate.pipeline.suspicious_commands.suspicious_commands import check_suspicious_code
1515
from codegate.storage import StorageEngine
1616
from codegate.utils.package_extractor import PackageExtractor
17-
from codegate.types.common import ModelResponse, Delta, StreamingChoices
17+
1818

1919
logger = structlog.get_logger("codegate")
2020

@@ -29,7 +29,7 @@ def __init__(self):
2929
def name(self) -> str:
3030
return "code-comment"
3131

32-
def _create_chunk(self, original_chunk: ModelResponse, content: str) -> ModelResponse:
32+
def _create_chunk(self, original_chunk: Any, content: str) -> Any:
3333
"""
3434
Creates a new chunk with the given content, preserving the original chunk's metadata
3535
"""
@@ -109,10 +109,10 @@ def _split_chunk_at_code_end(self, content: str) -> tuple[str, str]:
109109

110110
async def process_chunk(
111111
self,
112-
chunk: ModelResponse,
112+
chunk: Any,
113113
context: OutputPipelineContext,
114114
input_context: Optional[PipelineContext] = None,
115-
) -> list[ModelResponse]:
115+
) -> list[Any]:
116116
"""Process a single chunk of the stream"""
117117
for content in chunk.get_content():
118118
# Get current content plus this new chunk

0 commit comments

Comments
 (0)