Skip to content

Commit a542e7c

Browse files
committed
Merge branch 'dropout' of github.com:milancurcic/neural-fortran into dropout
2 parents ee7fdc9 + be06f51 commit a542e7c

File tree

1 file changed

+7
-2
lines changed

1 file changed

+7
-2
lines changed

src/nf/nf_layer_submodule.f90

Lines changed: 7 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -351,9 +351,14 @@ impure elemental module subroutine print_info(self)
351351
if (.not. self % name == 'input') &
352352
print '("Input shape: ", *(i0, 1x))', self % input_layer_shape
353353
print '("Output shape: ", *(i0, 1x))', self % layer_shape
354-
print '("Parameters: ", i0)', self % get_num_params()
355-
if (.not. self % name == 'input') &
354+
if (.not. self % name == 'dropout') &
355+
print '("Parameters: ", i0)', self % get_num_params()
356+
if (.not. (self % name == 'input' .or. self % name == 'dropout')) &
356357
print '("Activation: ", a)', self % activation
358+
select type (this_layer => self % p)
359+
type is (dropout_layer)
360+
print '("Dropout rate: ", f0.2)', this_layer % dropout_rate
361+
end select
357362
print *
358363
end subroutine print_info
359364

0 commit comments

Comments
 (0)