#!/usr/bin/perl =begin validates xml from genomics requests; using same validation function as genomics_xml.pl doesn't save xml file, or update any database table, or push file to any remote location =cut use lib ( '/home/raj/perl5/lib/perl5', '/home/raj/apps/HILIS4/lib', '/home/raj/perl-lib', # Local::XMLSimple - patched to escape single-apostrophe ); use LIMS::Local::ScriptHelpers; use XML::SAX::ParserFactory; use XML::Validator::Schema; use Local::XMLSimple qw(:strict); # need to specify ForceArray & KeyAttr if using :strict here use Data::Printer alias => 'p', use_prototypes => 0; use FindBin qw($Bin); # warn $Bin; exit; use Data::Dumper; use Modern::Perl; use Path::Tiny; use Local::DB; use IO::All; my $cfg_file = $Bin . '/../script/crons/lib/settings.cfg'; # p $cfg_file; my $xsd_src = $Bin . '/../setup/schema/xsd/genomics.xsd'; # p $xsd_src; my $settings = Config::Auto::parse($cfg_file); # p $settings; exit; my $genomics_uat = $settings->{genomics}->{uat}; # p $genomics_uat; # set some package vars (auto-stringify datetime objects & don't log queries): $Local::QueryLogger::NO_QUERY_LOGS = 1; # don't need queries in logs dir $Local::DBIx::Simple::Result::STRINGIFY_DATETIME = 1; # as we hand $data straight to XMLout() # get tools from LIMS::Local::ScriptHelpers: my $tools = LIMS::Local::ScriptHelpers->new(); my $dbix = Local::DB->dbix({ dbname => 'genomics' }); my $sql_lib = $tools->sql_lib(); my %SQL = ( demographics => $sql_lib->retr('genomics_demographics'), storage => $sql_lib->retr('genomics_storage'), results => $sql_lib->retr('genomics_results'), consent => $sql_lib->retr('genomics_consent'), ); # get request ids: my @request_ids = $dbix->select('requests', 'id')->column; # p \@request_ids; exit; my $timestamp = LIMS::Local::Utils::time_now(); my @lab_sections = do { my @args = ( 'lab_sections', ['section_name'], { is_active => 'yes' } ); $dbix->select(@args)->column; # array }; # p \@lab_sections; { # go: # get requests 1-to-1 data: my @requests = $dbix->query( $SQL{demographics}, @request_ids )->hashes; # p \@requests; exit; REQUEST: for my $req (@requests) { # p $req; # get results data: my $request_id = $req->{request_id}; my %req_fields = map +($_ => $req->{$_}), # don't need request_num/year: grep { $_ !~ /\b(year|request_number)\b/ } # skip undef vals to avoid uninitialized val warns from XML::Simple grep defined $req->{$_}, keys %$req; # p \%req_fields; # allocate GeL/participant id if screened: if ( my $arm = $req->{arm} ) { my $base_id = $genomics_uat->{rare_disease}; # 122_000_000 # add cancer increment if screened as cancer (ie to make 223_000_000) $base_id += $genomics_uat->{cancer_incr} if $arm =~ /cancer/i; $req_fields{participant_id} = $base_id + $request_id; } # lab number: $req_fields{lab_number} = join '/', $req->{request_number}, $req->{year} - 2000; # p \%req_fields; # build data structure for this request: my %h = ( Demographics => \%req_fields ); # p \%h; SECTION: for my $section_name (@lab_sections) { # p $section_name; my @data = do { my @bind = ( $request_id, $section_name ); $dbix->query( $SQL{results}, @bind )->hashes; }; # p \@data; RESULT: # each test/result pair for this section (query ensures no undef vals) for my $ref (@data) { # hashref of keys: test_name & result my $test_name = $ref->{test_name}; $h{$section_name}{$test_name} = $ref->{result}; } } # fluidx storage (1-to-many with request_id): if ( my @data = $dbix->query( $SQL{storage}, $request_id )->hashes ) { # ensure no undef values or XML::Simple v2.21+ issues uninitialized value warnings: remove_undef_values($_) for @data; $h{Storage}{Vials}{vial} = \@data; # p \@data; } { # consent taken - from request_lab_test_history file: my @vals = $dbix->query( $SQL{consent}, $request_id)->row; $h{Consent}{consent_taken_by} = join ' ', map ucfirst $_, @vals; } { # consent form filename (need to allow for >1 file): my $filepath = get_destination_sub_dir($req); # p $filepath; if ( -e $filepath ) { my @contents = io($filepath)->all; FILE: for my $f (@contents) { # p $f; next FILE unless $f->type eq 'file'; my $filename = join '~', $request_id, $f->filename; # add as array in case >1 file in dir push @{ $h{Consent}{Files}{file} }, $filename; } } } # p \%h; # create xml file: my $content = to_xml(\%h); # say $content; # generate id for request, eitehr participant_id or if not allocated, lab no: my $req = $req_fields{participant_id} || $req_fields{lab_number}; # validate xml, returns XML::SAX::Exception object on failure, undef on success: my $error = validate_xml_against_xsd($content); # warn $error; if ($error) { say $req, ' ' , $error; } else { say $req . ' validated OK'; } } } #------------------------------------------------------------------------------- sub to_xml { my $data = shift; # arrayref of request datasets # remove leading/trailing white-space (eg introduced by pasting from excel): _trim($data); # options for XMLout (needs array or hash): my %xs_opts = xs_options(); # p %xs_opts; my $xs = Local::XMLSimple->new(%xs_opts); my $ref = { FileCreationDateTime => $timestamp->datetime, # require string # add direct if NOT using xml element names, otherwise use format_request(): # Record => format_request($data), Record => $data, }; # p $ref; # enclose xml in outer block; add 'version' inline: # my $input = { add => $ref, version => 2 }; my $xml_out = $xs->XMLout($ref); # p $xml_out; return $xml_out; } #------------------------------------------------------------------------------- sub validate_xml_against_xsd { my $xml = shift; # switch debug on to show data structure: my $validator = XML::Validator::Schema->new(file => $xsd_src, debug => 0); # p $validator; my $parser = XML::SAX::ParserFactory->parser(Handler => $validator); eval { $parser->parse_string($xml); }; return $@; } #------------------------------------------------------------------------------- sub xs_options { my %opts = ( SuppressEmpty => 0, # ? only works for xml_in NoEscape => 0, # should be default anyway, but doesn't work RootName => 'HILIS4Genomics', XMLDecl => q!!, KeyAttr => [], NoAttr => 1, # gives inline (scalar) AND nested (arrayref) attributes ); return %opts; } #------------------------------------------------------------------------------- sub get_destination_sub_dir { my $ref = shift; # p $ref; my ($request_num, $year) = ($ref->{request_number}, $ref->{year}); # p [$year, $request_num]; my $i = int ( $request_num / 100 ); # warn $i; # 1-99 = 0, 100-199 = 1, 200-299 = 2, etc my $mini_dir = sprintf '%s-%s', 100 * $i || 1, # default to 1 if 0; 1, 100, 200, 300, etc 100 * $i + 99; # 99, 199, 299, etc my $app_dir = path($tools->path_to_app_root)->realpath; # p $app_dir; my $rel_path = 'static/files/genomics'; my $destination_dir = join '/', # eg /path/to/uploads/2013 $app_dir, $rel_path, $year, $mini_dir, $request_num; # p $destination_dir; return $destination_dir; } sub remove_undef_values { # based on kaitlyn's transform_datetime_to_string() below: my $r = shift; # p $r; return unless ref $r eq 'HASH'; for my $key( keys %$r ) { # p [$key, $r->{$key}]; if ( ref $r->{$key} eq 'HASH' ) { # recursive call for hashrefs remove_undef_values( $r->{$key} ); } defined $r->{$_} || delete $r->{$_} for keys %$r; # p $r; } } sub _trim { my $data = shift; # p $data; KEY: for my $key ( keys %$data ) { # say "$key: $data->{$key}"; # p $data->{$key}; if ( ref($data->{$key}) eq 'HASH' ) { # recursive call for hashrefs # warn $data->{$key}; _trim( $data->{$key} ); # p $r->{$key}; next KEY; # not sure why this is necessary ?? } # say "$key: $data->{$key}"; # remove leading/trailing space chars (incl. unicode 160) from string: $data->{$key} =~ s/^\p{Space}//g; $data->{$key} =~ s/\p{Space}$//g; # p $data->{$key}; } # p $data; return 0; # no need - modifying hashref in situ }