'Huggingface transformers error while fetching model
I'm trying to load the huggingface transformers sentiment-analysis model in ipython
from transformers import pipeline
...
sp = pipeline('sentiment-analysis')
Loading the model fails and produces the following output
No model was supplied, defaulted to distilbert-base-uncased-finetuned-sst-2-english (https://huggingface.co/distilbert-base-uncased-finetuned-sst-2-english)
HTTPSConnectionPool(host='cdn-lfs.huggingface.co', port=443): Max retries exceeded with url: /distilbert-base-uncased-finetuned-sst-2-english/60554cbd7781b09d87f1ececbea8c064b94e49a7f03fd88e8775bfe6cc3d9f88 (Caused by SSLError(SSLEOFError(8, 'EOF occurred in violation of protocol (_ssl.c:1129)')))
HTTPSConnectionPool(host='cdn-lfs.huggingface.co', port=443): Max retries exceeded with url: /distilbert-base-uncased-finetuned-sst-2-english/60554cbd7781b09d87f1ececbea8c064b94e49a7f03fd88e8775bfe6cc3d9f88 (Caused by SSLError(SSLEOFError(8, 'EOF occurred in violation of protocol (_ssl.c:1129)')))
---------------------------------------------------------------------------
ValueError Traceback (most recent call last)
Input In [2], in <cell line: 1>()
----> 1 sp=pipeline('sentiment-analysis')
File ~/anaconda3/envs/condaenv/lib/python3.9/site-packages/transformers/pipelines/__init__.py:543, in pipeline(task, model, config, tokenizer, feature_extractor, framework, revision, use_fast, use_auth_token, model_kwargs, pipeline_class, **kwargs)
539 # Infer the framework from the model
540 # Forced if framework already defined, inferred if it's None
541 # Will load the correct model if possible
542 model_classes = {"tf": targeted_task["tf"], "pt": targeted_task["pt"]}
--> 543 framework, model = infer_framework_load_model(
544 model,
545 model_classes=model_classes,
546 config=config,
547 framework=framework,
548 revision=revision,
549 task=task,
550 **model_kwargs,
551 )
553 model_config = model.config
555 load_tokenizer = type(model_config) in TOKENIZER_MAPPING or model_config.tokenizer_class is not None
File ~/anaconda3/envs/condaenv/lib/python3.9/site-packages/transformers/pipelines/base.py:231, in infer_framework_load_model(model, config, model_classes, task, framework, **model_kwargs)
228 continue
230 if isinstance(model, str):
--> 231 raise ValueError(f"Could not load model {model} with any of the following classes: {class_tuple}.")
233 framework = "tf" if model.__class__.__name__.startswith("TF") else "pt"
234 return framework, model
ValueError: Could not load model distilbert-base-uncased-finetuned-sst-2-english with any of the following classes: (<class 'transformers.models.auto.modeling_auto.AutoModelForSequenceClassification'>, <class 'transformers.models.distilbert.modeling_distilbert.DistilBertForSequenceClassification'>).
I tried removing the ~/.cache/huggingface folder but still getting the same error. I'd appreciate suggestions on what the cause of the issue could be and how to resolve it
Sources
This article follows the attribution requirements of Stack Overflow and is licensed under CC BY-SA 3.0.
Source: Stack Overflow
| Solution | Source |
|---|
