You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
# Environment Configurationlauncher: "accelerate"# Options: accelerateconfig_file: config/accelerate_configs/multi_gpu.yamlnum_processes: 8# Number of processes to launch (overrides config file)main_process_port: 29500mixed_precision: "bf16"# Options: no, fp16, bf16run_name: null # Run name (auto: {model_type}_{finetune_type}_{trainer_type}_{timestamp})project: "Flow-Factory-test-t2v2"# Project name for logginglogging_backend: "wandb"# Options: wandb, swanlab, tensorboard, none# Data Configurationdata:
dataset_dir: "dataset/pickscore"# Path to dataset folderpreprocessing_batch_size: 32# Batch size for preprocessingdataloader_num_workers: 16# Number of workers for DataLoaderforce_reprocess: false # Force reprocessing of the dataset# Cache directory for preprocessed datasetsmax_dataset_size: 1000# Limit the maximum number of samples in the dataset# Model Configurationmodel:
finetune_type: 'lora'# Options: full, loralora_rank : 128lora_alpha : 128target_components: 'transformer'# Options: transformer, transformer_2, or ['transformer', 'transformer_2']target_modules: "default"# Wan-AI/Wan2.2-TI2V-5B-Diffusers / Wan-AI/Wan2.2-T2V-A14B-Diffusersmodel_type: "wan2_t2v"# wan2_t2v, wan2_i2v, wan2_v2vresume_path: null # Path to load previous checkpoint/lora adapterresume_type: null # Options: lora, full, state. Null to auto-detect based on `finetune_type`# attn_backend: '_flash_3_hub' # Use flash attention 3 backend.log:
checkpoints and logssave_freq: 20# Save frequency in epochs (0 to disable)save_model_only: true # Save only the model weights (not optimizer, scheduler, etc.)# Training Configurationtrain:
# Trainer settingstrainer_type: 'nft'advantage_aggregation: 'gdpo'# Options: 'sum', 'gdpo'nft_beta: 0.1# `Old` Policy settingsoff_policy: true # Whether to use ema parameters for sampling off-policy data.ema_decay_schedule: "piecewise_linear"# Decay schedule for EMA. Options: ['constant', 'power', 'linear', 'piecewise_linear', 'cosine', 'warmup_cosine']flat_steps: 0ramp_rate: 0.001ema_decay: 0.5# EMA decay rate (0 to disable)ema_update_interval: 1# EMA update interval (in epochs)ema_device: "cuda"# Device to store EMA model (options: cpu, cuda)# Training Timestep distributionnum_train_timesteps: 8# Set null to all stepstime_sampling_strategy: discrete # Options: uniform, logit_normal, discrete, discrete_with_init, discrete_wo_inittime_shift: 3.0### Timestep range for discrete time sampling.### For Wan2.2-T2V-A14B (boundary_ratio=0.875, 10 inference steps):### - transformer only: 0.3 (early steps, before boundary). float: [0, value], e.g., 0.3 → first 30% of timesteps### - transformer_2 only: [0.4, 0.9] (later steps, after boundary). [start, end]: e.g., [0.4, 0.9] → 40%-90% of trajectorytimestep_range: 0.3# KL divkl_type: 'v-based'kl_beta: 0# KL divergence beta, 0 to disableref_param_device: 'cuda'# Options: cpu, cuda# Clippingclip_range: 1.0e-4# PPO/GRPO clipping rangeadv_clip_range: 5.0# Advantage clipping range# Sampling Settingsresolution: [384, 720] # Can be int or [height, width]num_frames: 81# Training framesnum_inference_steps: 20# Number of timestepsguidance_scale: 4.0# Guidance scale for samplingguidance_scale_2: 3.0# Guidance scale for sampling# Batch and samplingper_device_batch_size: 1# Batch size per devicegroup_size: 16# Group size for GRPO samplingglobal_std: false # Use global std for advantage normalizationunique_sample_num_per_epoch: 48# Unique samples per groupgradient_step_per_epoch: 1# Gradient steps per epoch# Optimizationseed: 42# Random seedlearning_rate: 1.0e-4# Initial learning rateadam_weight_decay: 1.0e-4# AdamW weight decayadam_betas: [0.9, 0.999] # AdamW betasadam_epsilon: 1.0e-8# AdamW epsilonmax_grad_norm: 1.0# Max gradient norm for clipping# Gradient checkpointingenable_gradient_checkpointing: true # Enable gradient checkpointing to save memory with extra compute# Seedseed: 42# Random seed# Scheduler Configurationscheduler:
dynamics_type: "ODE"# Options: Flow-SDE, Dance-SDE, CPS, ODE# Evaluation settingseval:
resolution: [704, 1280] # Evaluation resolutionnum_frames: 81# Evaluation framesper_device_batch_size: 1# Eval batch sizeguidance_scale: 4.0# Guidance scale for samplingguidance_scale_2: 3.0# Guidance scale for samplingnum_inference_steps: 28# Number of eval timestepseval_freq: 20# Eval frequency in epochs (0 to disable)seed: 42# Eval seed (defaults to training seed)# Reward Model Configurationrewards:
- name: "pick_score"reward_model: "PickScore"batch_size: 16device: "cuda"dtype: bfloat16# Optional Evaluation Reward Modelseval_rewards:
- name: "pick_score"reward_model: "PickScore"batch_size: 32device: "cuda"dtype: bfloat16
你好,我使用nft/lora/wan22_t2v.yaml,wan22-i2v-5b生成的视频为模糊的。请问是为什么?
ef996912-d6b9-4ac0-8464-f18820c98f99.mp4