Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
19 changes: 7 additions & 12 deletions router/src/grpc/server.rs
Original file line number Diff line number Diff line change
Expand Up @@ -598,8 +598,7 @@ impl grpc::embed_server::Embed for TextEmbeddingsService {
&self,
request: Request<EmbedRequest>,
) -> Result<Response<EmbedResponse>, Status> {
let counter = metrics::counter!("te_request_count", "method" => "single");
counter.increment(1);
metrics::counter!("te_request_count", "method" => "single").increment(1);

let permit = self
.infer
Expand All @@ -610,8 +609,7 @@ impl grpc::embed_server::Embed for TextEmbeddingsService {
let (response, metadata) = self.embed_pooled_inner(request, permit).await?;
let headers = HeaderMap::from(metadata);

let counter = metrics::counter!("te_request_count", "method" => "single");
counter.increment(1);
metrics::counter!("te_request_success", "method" => "single").increment(1);

Ok(Response::from_parts(
MetadataMap::from_headers(headers),
Expand Down Expand Up @@ -728,8 +726,7 @@ impl grpc::predict_server::Predict for TextEmbeddingsService {
&self,
request: Request<PredictRequest>,
) -> Result<Response<PredictResponse>, Status> {
let counter = metrics::counter!("te_request_count", "method" => "single");
counter.increment(1);
metrics::counter!("te_request_count", "method" => "single").increment(1);

let permit = self
.infer
Expand All @@ -749,8 +746,7 @@ impl grpc::predict_server::Predict for TextEmbeddingsService {
.await?;
let headers = HeaderMap::from(metadata);

let counter = metrics::counter!("te_request_count", "method" => "single");
counter.increment(1);
metrics::counter!("te_request_success", "method" => "single").increment(1);

Ok(Response::from_parts(
MetadataMap::from_headers(headers),
Expand All @@ -763,8 +759,8 @@ impl grpc::predict_server::Predict for TextEmbeddingsService {
&self,
request: Request<PredictPairRequest>,
) -> Result<Response<PredictResponse>, Status> {
let counter = metrics::counter!("te_request_count", "method" => "single");
counter.increment(1);
metrics::counter!("te_request_count", "method" => "single").increment(1);

let request = request.into_inner();

let mut inputs = request.inputs;
Expand Down Expand Up @@ -800,8 +796,7 @@ impl grpc::predict_server::Predict for TextEmbeddingsService {
.await?;
let headers = HeaderMap::from(metadata);

let counter = metrics::counter!("te_request_count", "method" => "single");
counter.increment(1);
metrics::counter!("te_request_success", "method" => "single").increment(1);

Ok(Response::from_parts(
MetadataMap::from_headers(headers),
Expand Down
24 changes: 8 additions & 16 deletions router/src/http/server.rs
Original file line number Diff line number Diff line change
Expand Up @@ -581,8 +581,7 @@ async fn embed(

let (response, metadata) = match req.inputs {
Input::Single(input) => {
let counter = metrics::counter!("te_request_count", "method" => "single");
counter.increment(1);
metrics::counter!("te_request_count", "method" => "single").increment(1);

let compute_chars = input.count_chars();

Expand All @@ -599,8 +598,7 @@ async fn embed(
.await
.map_err(ErrorResponse::from)?;

let counter = metrics::counter!("te_request_count", "method" => "single");
counter.increment(1);
metrics::counter!("te_request_success", "method" => "single").increment(1);

(
EmbedResponse(vec![response.results]),
Expand Down Expand Up @@ -761,8 +759,7 @@ async fn embed_sparse(

let (response, metadata) = match req.inputs {
Input::Single(input) => {
let counter = metrics::counter!("te_request_count", "method" => "single");
counter.increment(1);
metrics::counter!("te_request_count", "method" => "single").increment(1);

let compute_chars = input.count_chars();

Expand All @@ -778,8 +775,7 @@ async fn embed_sparse(
.await
.map_err(ErrorResponse::from)?;

let counter = metrics::counter!("te_request_count", "method" => "single");
counter.increment(1);
metrics::counter!("te_request_success", "method" => "single").increment(1);

(
EmbedSparseResponse(vec![sparsify(response.results)]),
Expand Down Expand Up @@ -932,8 +928,7 @@ async fn embed_all(

let (response, metadata) = match req.inputs {
Input::Single(input) => {
let counter = metrics::counter!("te_request_count", "method" => "single");
counter.increment(1);
metrics::counter!("te_request_count", "method" => "single").increment(1);

let compute_chars = input.count_chars();

Expand All @@ -949,8 +944,7 @@ async fn embed_all(
.await
.map_err(ErrorResponse::from)?;

let counter = metrics::counter!("te_request_count", "method" => "single");
counter.increment(1);
metrics::counter!("te_request_success", "method" => "single").increment(1);

(
EmbedAllResponse(vec![response.results]),
Expand Down Expand Up @@ -1117,8 +1111,7 @@ async fn openai_embed(

let (embeddings, metadata) = match req.input {
Input::Single(input) => {
let counter = metrics::counter!("te_request_count", "method" => "single");
counter.increment(1);
metrics::counter!("te_request_count", "method" => "single").increment(1);

let compute_chars = input.count_chars();

Expand All @@ -1135,8 +1128,7 @@ async fn openai_embed(
.await
.map_err(ErrorResponse::from)?;

let counter = metrics::counter!("te_request_count", "method" => "single");
counter.increment(1);
metrics::counter!("te_request_success", "method" => "single").increment(1);

let embedding = encode_embedding(response.results);
(
Expand Down