Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Same plots for loss even with different seeds #110

Open
Alexkkir opened this issue Oct 7, 2023 · 0 comments
Open

Same plots for loss even with different seeds #110

Alexkkir opened this issue Oct 7, 2023 · 0 comments

Comments

@Alexkkir
Copy link

Alexkkir commented Oct 7, 2023

Hello! I ran 4 experiments and they turned out to have very similar charts for the loss function. Even though I used different seeds. What could be the reason?

I run code with command:

CUDA_VISIBLE_DEVICES=0 python pyiqa/train.py -opt options/train/DBCNN_exps/multitask/vanilla.yml --debug

image

Config

# general settings
# name: debug_DBCNN_LIVEC
name: 006_MultitaskExp_vanilla
model_type: DBCNNModel
num_gpu: 1  # set num_gpu: 0 for cpu mode
# manual_seed: 123

# dataset and data loader settings
datasets:
  train:
    name: koniq10k
    type: GeneralSalDataset
    dataroot_target: ./datasets/koniq10k/512x384
    dataroot_sal: ./datasets/koniq10k/saliency_maps
    meta_info_file: ./datasets/meta_info/meta_info_KonIQ10kDataset.csv
    split_file: ./datasets/meta_info/koniq10k_official.pkl
    mos_range: [0, 100]
    lower_better: false 
    mos_normalize: true

    augment:
      hflip: true
      random_crop: 384
    img_range: 1

    # data loader
    use_shuffle: true
    num_worker_per_gpu: 4
    batch_size_per_gpu: 8
    dataset_enlarge_ratio: 1
    prefetch_mode: ~

  val_clive:
    name: livechallenge
    type: LIVEChallengeDataset
    dataroot_target: ./datasets/LIVEC
    meta_info_file: ./datasets/meta_info/meta_info_LIVEChallengeDataset.csv
    dataroot_sal: ./datasets/LIVEC/saliency_maps
    # split_file: ./datasets/meta_info/livechallenge_seed123.pkl
    # split_index: 2 # which split to use, start from 1

    num_worker_per_gpu: 4
    batch_size_per_gpu: 8
    dataset_enlarge_ratio: 1

  val_koniq:
    name: koniq10k
    type: GeneralSalDataset
    dataroot_target: ./datasets/koniq10k/512x384
    dataroot_sal: ./datasets/koniq10k/saliency_maps
    meta_info_file: ./datasets/meta_info/meta_info_KonIQ10kDataset.csv
    split_file: ./datasets/meta_info/koniq10k_official.pkl

    num_worker_per_gpu: 4
    batch_size_per_gpu: 8
    dataset_enlarge_ratio: 1

# network structures
network:
  type: DBCNNSal
  pretrained: false
  sal_mode: input

# path
path:
  pretrain_network_g: ~
  strict_load_g: true
  resume_state: ~

# training settings
train:
  optim:
    type: SGD
    lr: !!float 1e-3
    momentum: 0.9
    weight_decay: !!float 5e-4

  optim_finetune:
    type: Adam
    lr: !!float 1e-5
    weight_decay: !!float 5e-4

  scheduler:
    type: MultiStepLR
    milestones: [1000]
    gamma: 1

  scheduler_finetune:
    type: MultiStepLR
    milestones: [1000]
    gamma: 1

  total_iter: 20000 # 883 iter/epoch, train/finetune for about 30 epochs
  finetune_start_iter: 10000
  warmup_iter: -1  # no warm up

  # losses
  mos_loss_opt:
    type: MSELoss
    loss_weight: !!float 1.0

  # sal_loss_opt:
  #   type: MSESalLoss
  #   loss_weight: !!float 0.001

  # multitask_loss_opt:
  #   type: MultiTaskLoss
  #   n_tasks: 2
  
# validation settings
val:
  val_freq: !!float 800
  save_img: false
  pbar: true

  key_metric: srcc # if this metric improve, update all metrics. If not specified, each best metric results will be updated separately
  metrics:
    srcc:
      type: calculate_srcc

    plcc:
      type: calculate_plcc

    krcc:
      type: calculate_krcc

# logging settings
logger:
  print_freq: 100
  save_latest_freq: !!float 5e2
  use_tb_logger: true
  wandb:
    project: IQA-7 pyiqa
    resume_id: ~

# dist training settings
dist_params:
  backend: nccl
  port: 29500

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant