Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

fix bug when using FSDP #696

Merged
merged 1 commit into from
Jan 20, 2025
Merged

Conversation

kaixuanliu
Copy link
Contributor

No description provided.

Signed-off-by: Liu, Kaixuan <[email protected]>
@kaixuanliu
Copy link
Contributor Author

@jiqing-feng @casper-hansen pls help review.

@kaixuanliu
Copy link
Contributor Author

Cases to fix:
run following cmd line in peft:
accelerate launch examples/olora_finetuning/olora_finetuning.py --base_model hugging-quants/Meta-Llama-3.1-8B-Instruct-AWQ-INT4 --init_lora_weights gaussian
with accelerate config file:

compute_environment: LOCAL_MACHINE
debug: false
distributed_type: FSDP
downcast_bf16: 'no'
enable_cpu_affinity: false
fsdp_config:
  fsdp_auto_wrap_policy: NO_WRAP
  fsdp_backward_prefetch: NO_PREFETCH
  fsdp_cpu_ram_efficient_loading: true
  fsdp_forward_prefetch: false
  fsdp_offload_params: false
  fsdp_sharding_strategy: NO_SHARD
  fsdp_state_dict_type: FULL_STATE_DICT
  fsdp_sync_module_states: true
  fsdp_use_orig_params: false
ipex_config:
  use_xpu: true
machine_rank: 0
main_training_function: main
mixed_precision: 'no'
num_machines: 1
num_processes: 4
rdzv_backend: static
same_network: true
use_cpu: false

@casper-hansen casper-hansen merged commit 4c9a791 into casper-hansen:main Jan 20, 2025
@kaixuanliu kaixuanliu deleted the fsdp-fix branch February 26, 2025 03:49
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants