@@ -250,7 +250,9 @@ modules:
250
250
# enable if you want to use transformers instead of the
251
251
# text2vec-contextionary module
252
252
enabled : false
253
-
253
+ # You can set directly an inference URL of this module without deploying it with this release.
254
+ # You can do so by setting a value for the `inferenceUrl` here AND by setting the `enable` to `false`
255
+ inferenceUrl : {}
254
256
# The configuration below is ignored if enabled==false
255
257
256
258
# replace with model of choice, see
@@ -290,6 +292,9 @@ modules:
290
292
passageQueryServices :
291
293
passage :
292
294
enabled : false
295
+ # You can set directly an inference URL of this module without deploying it with this release.
296
+ # You can do so by setting a value for the `inferenceUrl` here AND by setting the `enable` to `false`
297
+ inferenceUrl : {}
293
298
294
299
tag : facebook-dpr-ctx_encoder-single-nq-base
295
300
repo : semitechnologies/transformers-inference
@@ -322,6 +327,9 @@ modules:
322
327
# nvidia.com/gpu: 1
323
328
query :
324
329
enabled : false
330
+ # You can set directly an inference URL of this module without deploying it with this release.
331
+ # You can do so by setting a value for the `inferenceUrl` here AND by setting the `enable` to `false`
332
+ inferenceUrl : {}
325
333
326
334
tag : facebook-dpr-question_encoder-single-nq-base
327
335
repo : semitechnologies/transformers-inference
@@ -389,6 +397,9 @@ modules:
389
397
# enable if you want to use transformers instead of the
390
398
# text2vec-contextionary module
391
399
enabled : false
400
+ # You can set directly an inference URL of this module without deploying it with this release.
401
+ # You can do so by setting a value for the `inferenceUrl` here AND by setting the `enable` to `false`
402
+ inferenceUrl : {}
392
403
393
404
# The configuration below is ignored if enabled==false
394
405
@@ -429,6 +440,9 @@ modules:
429
440
# DistilBERT, to find an aswer in text to a given question
430
441
qna-transformers :
431
442
enabled : false
443
+ # You can set directly an inference URL of this module without deploying it with this release.
444
+ # You can do so by setting a value for the `inferenceUrl` here AND by setting the `enable` to `false`
445
+ inferenceUrl : {}
432
446
tag : bert-large-uncased-whole-word-masking-finetuned-squad-34d66b1
433
447
repo : semitechnologies/qna-transformers
434
448
registry : docker.io
@@ -463,6 +477,9 @@ modules:
463
477
# a vector representation of the image
464
478
img2vec-neural :
465
479
enabled : false
480
+ # You can set directly an inference URL of this module without deploying it with this release.
481
+ # You can do so by setting a value for the `inferenceUrl` here AND by setting the `enable` to `false`
482
+ inferenceUrl : {}
466
483
tag : resnet50
467
484
repo : semitechnologies/img2vec-pytorch
468
485
registry : docker.io
@@ -497,6 +514,9 @@ modules:
497
514
# misspellings in a given text
498
515
text-spellcheck :
499
516
enabled : false
517
+ # You can set directly an inference URL of this module without deploying it with this release.
518
+ # You can do so by setting a value for the `inferenceUrl` here AND by setting the `enable` to `false`
519
+ inferenceUrl : {}
500
520
tag : pyspellchecker-en
501
521
repo : semitechnologies/text-spellcheck-model
502
522
registry : docker.io
@@ -531,6 +551,9 @@ modules:
531
551
# misspellings in a given text
532
552
ner-transformers :
533
553
enabled : false
554
+ # You can set directly an inference URL of this module without deploying it with this release.
555
+ # You can do so by setting a value for the `inferenceUrl` here AND by setting the `enable` to `false`
556
+ inferenceUrl : {}
534
557
tag : dbmdz-bert-large-cased-finetuned-conll03-english-0.0.2
535
558
repo : semitechnologies/ner-transformers
536
559
registry : docker.io
@@ -561,6 +584,42 @@ modules:
561
584
# enable if running with CUDA support
562
585
# nvidia.com/gpu: 1
563
586
587
+ # The sum-transformers module makes result texts summarizations
588
+ sum-transformers :
589
+ enabled : false
590
+ # You can set directly an inference URL of this module without deploying it with this release.
591
+ # You can do so by setting a value for the `inferenceUrl` here AND by setting the `enable` to `false`
592
+ inferenceUrl : {}
593
+ tag : facebook-bart-large-cnn-1.0.0
594
+ repo : semitechnologies/sum-transformers
595
+ registry : docker.io
596
+ replicas : 1
597
+ fullnameOverride : sum-transformers
598
+ envconfig :
599
+ # enable for CUDA support. Your K8s cluster needs to be configured
600
+ # accordingly and you need to explicitly set GPU requests & limits below
601
+ enable_cuda : false
602
+
603
+ # only used when cuda is enabled
604
+ nvidia_visible_devices : all
605
+
606
+ # only used when cuda is enabled
607
+ ld_library_path : /usr/local/nvidia/lib64
608
+
609
+ resources :
610
+ requests :
611
+ cpu : ' 1000m'
612
+ memory : ' 3000Mi'
613
+
614
+ # enable if running with CUDA support
615
+ # nvidia.com/gpu: 1
616
+ limits :
617
+ cpu : ' 1000m'
618
+ memory : ' 5000Mi'
619
+
620
+ # enable if running with CUDA support
621
+ # nvidia.com/gpu: 1
622
+
564
623
# by choosing the default vectorizer module, you can tell Weaviate to always
565
624
# use this module as the vectorizer if nothing else is specified. Can be
566
625
# overwritten on a per-class basis.
0 commit comments