Skip to content
This repository has been archived by the owner on Apr 11, 2024. It is now read-only.

Commit

Permalink
Release: 0.3.19
Browse files Browse the repository at this point in the history
  • Loading branch information
BobaZooba committed Sep 30, 2023
1 parent cdb20ed commit f80d94e
Show file tree
Hide file tree
Showing 3 changed files with 4 additions and 4 deletions.
2 changes: 1 addition & 1 deletion setup.py
Original file line number Diff line number Diff line change
Expand Up @@ -57,7 +57,7 @@
# Setup
setup(
name="xllm",
version="0.3.18",
version="0.3.19",
description="Simple & Cutting Edge LLM Finetuning",
license_files=["LICENSE"],
long_description=open("README.md", "r", encoding="utf-8").read(),
Expand Down
2 changes: 1 addition & 1 deletion src/xllm/__init__.py
Original file line number Diff line number Diff line change
Expand Up @@ -14,7 +14,7 @@

# ruff: noqa: F401

__version__ = "0.3.18"
__version__ = "0.3.19"

from . import enums, types
from .core.config import HuggingFaceConfig
Expand Down
4 changes: 2 additions & 2 deletions src/xllm/experiments/base.py
Original file line number Diff line number Diff line change
Expand Up @@ -155,7 +155,7 @@ def build(self):
elif self.quantization_config is not None:
dist_logger.warning("quantization_config is not None, but the model was built outside of the experiment")

if self.is_bnb_quantization or self.config.bnb_quantize_after_model_init:
if self.is_bnb_quantization and self.config.bnb_quantize_after_model_init:
self.before_bnb_quantization()
self.bnb_quantization()
bnb_quantization_type = "int4" if self.quantization_config.load_in_4bit else "int8"
Expand Down Expand Up @@ -307,7 +307,7 @@ def before_model_build(self) -> None:

def build_model(self) -> PreTrainedModel:
quantization_config = (
None if self.is_bnb_quantization or self.config.bnb_quantize_after_model_init else self.quantization_config
None if self.is_bnb_quantization and self.config.bnb_quantize_after_model_init else self.quantization_config
)
model = build_model(config=self.config, quantization_config=quantization_config)
return model
Expand Down

0 comments on commit f80d94e

Please sign in to comment.