@@ -89,6 +89,8 @@ example = json ! ({"error": "Inference failed", "error_type": "backend"})),
89
89
example = json ! ( { "error" : "Model is overloaded" , "error_type" : "overloaded" } ) ) ,
90
90
( status = 422 , description = "Tokenization error" , body = ErrorResponse ,
91
91
example = json ! ( { "error" : "Tokenization error" , "error_type" : "tokenizer" } ) ) ,
92
+ ( status = 400 , description = "Batch is empty" , body = ErrorResponse ,
93
+ example = json ! ( { "error" : "Batch is empty" , "error_type" : "empty" } ) ) ,
92
94
( status = 413 , description = "Batch size error" , body = ErrorResponse ,
93
95
example = json ! ( { "error" : "Batch size error" , "error_type" : "validation" } ) ) ,
94
96
)
@@ -285,6 +287,8 @@ example = json ! ({"error": "Inference failed", "error_type": "backend"})),
285
287
example = json ! ( { "error" : "Model is overloaded" , "error_type" : "overloaded" } ) ) ,
286
288
( status = 422 , description = "Tokenization error" , body = ErrorResponse ,
287
289
example = json ! ( { "error" : "Tokenization error" , "error_type" : "tokenizer" } ) ) ,
290
+ ( status = 400 , description = "Batch is empty" , body = ErrorResponse ,
291
+ example = json ! ( { "error" : "Batch is empty" , "error_type" : "empty" } ) ) ,
288
292
( status = 413 , description = "Batch size error" , body = ErrorResponse ,
289
293
example = json ! ( { "error" : "Batch size error" , "error_type" : "validation" } ) ) ,
290
294
)
@@ -306,7 +310,7 @@ async fn rerank(
306
310
tracing:: error!( "{message}" ) ;
307
311
let err = ErrorResponse {
308
312
error : message,
309
- error_type : ErrorType :: Validation ,
313
+ error_type : ErrorType :: Empty ,
310
314
} ;
311
315
let counter = metrics:: counter!( "te_request_failure" , "err" => "validation" ) ;
312
316
counter. increment ( 1 ) ;
@@ -471,6 +475,8 @@ example = json ! ({"error": "Inference failed", "error_type": "backend"})),
471
475
example = json ! ( { "error" : "Model is overloaded" , "error_type" : "overloaded" } ) ) ,
472
476
( status = 422 , description = "Tokenization error" , body = ErrorResponse ,
473
477
example = json ! ( { "error" : "Tokenization error" , "error_type" : "tokenizer" } ) ) ,
478
+ ( status = 400 , description = "Batch is empty" , body = ErrorResponse ,
479
+ example = json ! ( { "error" : "Batch is empty" , "error_type" : "empty" } ) ) ,
474
480
( status = 413 , description = "Batch size error" , body = ErrorResponse ,
475
481
example = json ! ( { "error" : "Batch size error" , "error_type" : "validation" } ) ) ,
476
482
)
@@ -489,7 +495,7 @@ async fn similarity(
489
495
tracing:: error!( "{message}" ) ;
490
496
let err = ErrorResponse {
491
497
error : message,
492
- error_type : ErrorType :: Validation ,
498
+ error_type : ErrorType :: Empty ,
493
499
} ;
494
500
let counter = metrics:: counter!( "te_request_failure" , "err" => "validation" ) ;
495
501
counter. increment ( 1 ) ;
@@ -553,6 +559,8 @@ example = json ! ({"error": "Inference failed", "error_type": "backend"})),
553
559
example = json ! ( { "error" : "Model is overloaded" , "error_type" : "overloaded" } ) ) ,
554
560
( status = 422 , description = "Tokenization error" , body = ErrorResponse ,
555
561
example = json ! ( { "error" : "Tokenization error" , "error_type" : "tokenizer" } ) ) ,
562
+ ( status = 400 , description = "Batch is empty" , body = ErrorResponse ,
563
+ example = json ! ( { "error" : "Batch is empty" , "error_type" : "empty" } ) ) ,
556
564
( status = 413 , description = "Batch size error" , body = ErrorResponse ,
557
565
example = json ! ( { "error" : "Batch size error" , "error_type" : "validation" } ) ) ,
558
566
)
@@ -615,7 +623,7 @@ async fn embed(
615
623
tracing:: error!( "{message}" ) ;
616
624
let err = ErrorResponse {
617
625
error : message,
618
- error_type : ErrorType :: Validation ,
626
+ error_type : ErrorType :: Empty ,
619
627
} ;
620
628
let counter = metrics:: counter!( "te_request_failure" , "err" => "validation" ) ;
621
629
counter. increment ( 1 ) ;
@@ -722,6 +730,8 @@ example = json ! ({"error": "Inference failed", "error_type": "backend"})),
722
730
example = json ! ( { "error" : "Model is overloaded" , "error_type" : "overloaded" } ) ) ,
723
731
( status = 422 , description = "Tokenization error" , body = ErrorResponse ,
724
732
example = json ! ( { "error" : "Tokenization error" , "error_type" : "tokenizer" } ) ) ,
733
+ ( status = 400 , description = "Batch is empty" , body = ErrorResponse ,
734
+ example = json ! ( { "error" : "Batch is empty" , "error_type" : "empty" } ) ) ,
725
735
( status = 413 , description = "Batch size error" , body = ErrorResponse ,
726
736
example = json ! ( { "error" : "Batch size error" , "error_type" : "validation" } ) ) ,
727
737
)
@@ -792,7 +802,7 @@ async fn embed_sparse(
792
802
tracing:: error!( "{message}" ) ;
793
803
let err = ErrorResponse {
794
804
error : message,
795
- error_type : ErrorType :: Validation ,
805
+ error_type : ErrorType :: Empty ,
796
806
} ;
797
807
let counter = metrics:: counter!( "te_request_failure" , "err" => "validation" ) ;
798
808
counter. increment ( 1 ) ;
@@ -900,6 +910,8 @@ example = json ! ({"error": "Inference failed", "error_type": "backend"})),
900
910
example = json ! ( { "error" : "Model is overloaded" , "error_type" : "overloaded" } ) ) ,
901
911
( status = 422 , description = "Tokenization error" , body = ErrorResponse ,
902
912
example = json ! ( { "error" : "Tokenization error" , "error_type" : "tokenizer" } ) ) ,
913
+ ( status = 400 , description = "Batch is empty" , body = ErrorResponse ,
914
+ example = json ! ( { "error" : "Batch is empty" , "error_type" : "empty" } ) ) ,
903
915
( status = 413 , description = "Batch size error" , body = ErrorResponse ,
904
916
example = json ! ( { "error" : "Batch size error" , "error_type" : "validation" } ) ) ,
905
917
)
@@ -961,7 +973,7 @@ async fn embed_all(
961
973
tracing:: error!( "{message}" ) ;
962
974
let err = ErrorResponse {
963
975
error : message,
964
- error_type : ErrorType :: Validation ,
976
+ error_type : ErrorType :: Empty ,
965
977
} ;
966
978
let counter = metrics:: counter!( "te_request_failure" , "err" => "validation" ) ;
967
979
counter. increment ( 1 ) ;
@@ -1067,6 +1079,8 @@ example = json ! ({"message": "Inference failed", "type": "backend"})),
1067
1079
example = json ! ( { "message" : "Model is overloaded" , "type" : "overloaded" } ) ) ,
1068
1080
( status = 422 , description = "Tokenization error" , body = OpenAICompatErrorResponse ,
1069
1081
example = json ! ( { "message" : "Tokenization error" , "type" : "tokenizer" } ) ) ,
1082
+ ( status = 400 , description = "Batch is empty" , body = OpenAICompatErrorResponse ,
1083
+ example = json ! ( { "message" : "Batch is empty" , "type" : "empty" } ) ) ,
1070
1084
( status = 413 , description = "Batch size error" , body = OpenAICompatErrorResponse ,
1071
1085
example = json ! ( { "message" : "Batch size error" , "type" : "validation" } ) ) ,
1072
1086
)
@@ -1150,7 +1164,7 @@ async fn openai_embed(
1150
1164
tracing:: error!( "{message}" ) ;
1151
1165
let err = ErrorResponse {
1152
1166
error : message,
1153
- error_type : ErrorType :: Validation ,
1167
+ error_type : ErrorType :: Empty ,
1154
1168
} ;
1155
1169
let counter = metrics:: counter!( "te_request_failure" , "err" => "validation" ) ;
1156
1170
counter. increment ( 1 ) ;
@@ -1265,8 +1279,12 @@ path = "/tokenize",
1265
1279
request_body = TokenizeRequest ,
1266
1280
responses(
1267
1281
( status = 200 , description = "Tokenized ids" , body = TokenizeResponse ) ,
1282
+ ( status = 400 , description = "Batch is empty" , body = ErrorResponse ,
1283
+ example = json ! ( { "error" : "Batch is empty" , "error_type" : "empty" } ) ) ,
1284
+ ( status = 413 , description = "Batch size error" , body = ErrorResponse ,
1285
+ example = json ! ( { "error" : "Batch size error" , "error_type" : "validation" } ) ) ,
1268
1286
( status = 422 , description = "Tokenization error" , body = ErrorResponse ,
1269
- example = json ! ( { "message " : "Tokenization error" , "type " : "tokenizer" } ) ) ,
1287
+ example = json ! ( { "error " : "Tokenization error" , "error_type " : "tokenizer" } ) ) ,
1270
1288
)
1271
1289
) ]
1272
1290
#[ instrument( skip_all) ]
@@ -1327,7 +1345,7 @@ async fn tokenize(
1327
1345
tracing:: error!( "{message}" ) ;
1328
1346
let err = ErrorResponse {
1329
1347
error : message,
1330
- error_type : ErrorType :: Validation ,
1348
+ error_type : ErrorType :: Empty ,
1331
1349
} ;
1332
1350
let counter = metrics:: counter!( "te_request_failure" , "err" => "validation" ) ;
1333
1351
counter. increment ( 1 ) ;
@@ -1377,8 +1395,12 @@ path = "/decode",
1377
1395
request_body = DecodeRequest ,
1378
1396
responses(
1379
1397
( status = 200 , description = "Decoded ids" , body = DecodeResponse ) ,
1398
+ ( status = 400 , description = "Batch is empty" , body = ErrorResponse ,
1399
+ example = json ! ( { "error" : "Batch is empty" , "error_type" : "empty" } ) ) ,
1400
+ ( status = 413 , description = "Batch size error" , body = ErrorResponse ,
1401
+ example = json ! ( { "error" : "Batch size error" , "error_type" : "validation" } ) ) ,
1380
1402
( status = 422 , description = "Tokenization error" , body = ErrorResponse ,
1381
- example = json ! ( { "message " : "Tokenization error" , "type " : "tokenizer" } ) ) ,
1403
+ example = json ! ( { "error " : "Tokenization error" , "error_type " : "tokenizer" } ) ) ,
1382
1404
)
1383
1405
) ]
1384
1406
#[ instrument( skip_all) ]
@@ -1403,7 +1425,7 @@ async fn decode(
1403
1425
tracing:: error!( "{message}" ) ;
1404
1426
let err = ErrorResponse {
1405
1427
error : message,
1406
- error_type : ErrorType :: Validation ,
1428
+ error_type : ErrorType :: Empty ,
1407
1429
} ;
1408
1430
let counter = metrics:: counter!( "te_request_failure" , "err" => "validation" ) ;
1409
1431
counter. increment ( 1 ) ;
@@ -1454,6 +1476,8 @@ example = json ! ({"error": "Inference failed", "error_type": "backend"})),
1454
1476
example = json ! ( { "error" : "Model is overloaded" , "error_type" : "overloaded" } ) ) ,
1455
1477
( status = 422 , description = "Tokenization error" , body = ErrorResponse ,
1456
1478
example = json ! ( { "error" : "Tokenization error" , "error_type" : "tokenizer" } ) ) ,
1479
+ ( status = 400 , description = "Batch is empty" , body = ErrorResponse ,
1480
+ example = json ! ( { "error" : "Batch is empty" , "error_type" : "empty" } ) ) ,
1457
1481
( status = 413 , description = "Batch size error" , body = ErrorResponse ,
1458
1482
example = json ! ( { "error" : "Batch size error" , "error_type" : "validation" } ) ) ,
1459
1483
)
@@ -1804,6 +1828,7 @@ impl From<&ErrorType> for StatusCode {
1804
1828
ErrorType :: Overloaded => StatusCode :: TOO_MANY_REQUESTS ,
1805
1829
ErrorType :: Tokenizer => StatusCode :: UNPROCESSABLE_ENTITY ,
1806
1830
ErrorType :: Validation => StatusCode :: PAYLOAD_TOO_LARGE ,
1831
+ ErrorType :: Empty => StatusCode :: BAD_REQUEST ,
1807
1832
}
1808
1833
}
1809
1834
}
0 commit comments