Spaces:
Sleeping
Sleeping
Commit
·
32074ae
1
Parent(s):
1ea681e
Fix model initialization by removing unsupported parameters
Browse files
app.py
CHANGED
|
@@ -85,34 +85,34 @@ def initialize_model():
|
|
| 85 |
device = "cuda" if torch.cuda.is_available() else "cpu"
|
| 86 |
logger.info(f"Using device: {device}")
|
| 87 |
|
| 88 |
-
# Initialize new model with
|
| 89 |
model = LLaVA(
|
| 90 |
vision_model_path="openai/clip-vit-base-patch32",
|
| 91 |
language_model_path="TinyLlama/TinyLlama-1.1B-Chat-v1.0",
|
| 92 |
-
device=device,
|
| 93 |
projection_hidden_dim=2048,
|
| 94 |
-
|
| 95 |
-
low_cpu_mem_usage=True,
|
| 96 |
-
use_safetensors=True,
|
| 97 |
-
load_in_8bit=True if device == "cuda" else False,
|
| 98 |
-
trust_remote_code=True
|
| 99 |
)
|
| 100 |
|
| 101 |
# Configure model for inference
|
| 102 |
if hasattr(model, 'language_model'):
|
| 103 |
-
model
|
| 104 |
model.language_model.eval()
|
| 105 |
|
| 106 |
-
#
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 107 |
if hasattr(model.language_model, 'generation_config'):
|
| 108 |
model.language_model.generation_config.do_sample = True
|
| 109 |
model.language_model.generation_config.max_new_tokens = 256
|
| 110 |
model.language_model.generation_config.temperature = 0.7
|
| 111 |
model.language_model.generation_config.top_p = 0.9
|
| 112 |
-
|
| 113 |
-
|
| 114 |
-
# Move model to device
|
| 115 |
-
model = model.to(device)
|
| 116 |
|
| 117 |
model_status.update({
|
| 118 |
"initialized": True,
|
|
@@ -121,11 +121,10 @@ def initialize_model():
|
|
| 121 |
"model_info": {
|
| 122 |
"vision_model": "openai/clip-vit-base-patch32",
|
| 123 |
"language_model": "TinyLlama/TinyLlama-1.1B-Chat-v1.0",
|
| 124 |
-
"dtype": str(model.dtype),
|
| 125 |
"device": str(model.device)
|
| 126 |
}
|
| 127 |
})
|
| 128 |
-
logger.info(f"Model successfully initialized on {model.device}
|
| 129 |
return True
|
| 130 |
|
| 131 |
except Exception as e:
|
|
|
|
| 85 |
device = "cuda" if torch.cuda.is_available() else "cpu"
|
| 86 |
logger.info(f"Using device: {device}")
|
| 87 |
|
| 88 |
+
# Initialize new model with basic parameters
|
| 89 |
model = LLaVA(
|
| 90 |
vision_model_path="openai/clip-vit-base-patch32",
|
| 91 |
language_model_path="TinyLlama/TinyLlama-1.1B-Chat-v1.0",
|
|
|
|
| 92 |
projection_hidden_dim=2048,
|
| 93 |
+
device=device
|
|
|
|
|
|
|
|
|
|
|
|
|
| 94 |
)
|
| 95 |
|
| 96 |
# Configure model for inference
|
| 97 |
if hasattr(model, 'language_model'):
|
| 98 |
+
# Set model to evaluation mode
|
| 99 |
model.language_model.eval()
|
| 100 |
|
| 101 |
+
# Configure model for inference
|
| 102 |
+
if hasattr(model.language_model, 'config'):
|
| 103 |
+
model.language_model.config.use_cache = False
|
| 104 |
+
|
| 105 |
+
# Move model to device
|
| 106 |
+
model = model.to(device)
|
| 107 |
+
|
| 108 |
+
# Set generation config if available
|
| 109 |
if hasattr(model.language_model, 'generation_config'):
|
| 110 |
model.language_model.generation_config.do_sample = True
|
| 111 |
model.language_model.generation_config.max_new_tokens = 256
|
| 112 |
model.language_model.generation_config.temperature = 0.7
|
| 113 |
model.language_model.generation_config.top_p = 0.9
|
| 114 |
+
if hasattr(model.language_model.config, 'eos_token_id'):
|
| 115 |
+
model.language_model.generation_config.pad_token_id = model.language_model.config.eos_token_id
|
|
|
|
|
|
|
| 116 |
|
| 117 |
model_status.update({
|
| 118 |
"initialized": True,
|
|
|
|
| 121 |
"model_info": {
|
| 122 |
"vision_model": "openai/clip-vit-base-patch32",
|
| 123 |
"language_model": "TinyLlama/TinyLlama-1.1B-Chat-v1.0",
|
|
|
|
| 124 |
"device": str(model.device)
|
| 125 |
}
|
| 126 |
})
|
| 127 |
+
logger.info(f"Model successfully initialized on {model.device}")
|
| 128 |
return True
|
| 129 |
|
| 130 |
except Exception as e:
|