[Bast-commits] r7706 - in DBIx-Class/0.08/branches/sybase_support: lib/DBIx lib/DBIx/Class/Storage lib/DBIx/Class/Storage/DBI lib/DBIx/Class/Storage/DBI/Sybase t

ribasushi at dev.catalyst.perl.org ribasushi at dev.catalyst.perl.org
Sun Sep 20 23:18:41 GMT 2009


Author: ribasushi
Date: 2009-09-20 23:18:40 +0000 (Sun, 20 Sep 2009)
New Revision: 7706

Modified:
   DBIx-Class/0.08/branches/sybase_support/lib/DBIx/Class.pm
   DBIx-Class/0.08/branches/sybase_support/lib/DBIx/Class/Storage/DBI.pm
   DBIx-Class/0.08/branches/sybase_support/lib/DBIx/Class/Storage/DBI/Sybase.pm
   DBIx-Class/0.08/branches/sybase_support/lib/DBIx/Class/Storage/DBI/Sybase/NoBindVars.pm
   DBIx-Class/0.08/branches/sybase_support/t/746sybase.t
Log:
main sybase branch ready

Modified: DBIx-Class/0.08/branches/sybase_support/lib/DBIx/Class/Storage/DBI/Sybase/NoBindVars.pm
===================================================================
--- DBIx-Class/0.08/branches/sybase_support/lib/DBIx/Class/Storage/DBI/Sybase/NoBindVars.pm	2009-09-20 22:58:09 UTC (rev 7705)
+++ DBIx-Class/0.08/branches/sybase_support/lib/DBIx/Class/Storage/DBI/Sybase/NoBindVars.pm	2009-09-20 23:18:40 UTC (rev 7706)
@@ -59,6 +59,9 @@
   return $value;
 }
 
+# for tests
+sub _can_insert_bulk { 0 }
+
 1;
 
 =head1 NAME

Modified: DBIx-Class/0.08/branches/sybase_support/lib/DBIx/Class/Storage/DBI/Sybase.pm
===================================================================
--- DBIx-Class/0.08/branches/sybase_support/lib/DBIx/Class/Storage/DBI/Sybase.pm	2009-09-20 22:58:09 UTC (rev 7705)
+++ DBIx-Class/0.08/branches/sybase_support/lib/DBIx/Class/Storage/DBI/Sybase.pm	2009-09-20 23:18:40 UTC (rev 7706)
@@ -18,6 +18,9 @@
 );
 
 my @also_proxy_to_writer_storage = qw/
+  connect_call_set_auto_cast auto_cast connect_call_blob_setup
+  connect_call_datetime_setup
+
   disconnect _connect_info _sql_maker _sql_maker_opts disable_sth_caching
   auto_savepoint unsafe cursor_class debug debugobj schema
 /;
@@ -124,12 +127,14 @@
 
   $writer_storage->_is_writer_storage(1);
   $writer_storage->connect_info($self->connect_info);
+  $writer_storage->auto_cast($self->auto_cast);
 
   $self->_writer_storage($writer_storage);
 }
 
 for my $method (@also_proxy_to_writer_storage) {
   no strict 'refs';
+  no warnings 'redefine';
 
   my $replaced = __PACKAGE__->can($method);
 
@@ -195,6 +200,12 @@
   $type && $type =~ /(?:text|image|lob|bytea|binary|memo)/i;
 }
 
+sub _is_lob_column {
+  my ($self, $source, $column) = @_;
+
+  return $self->_is_lob_type($source->column_info($column)->{data_type});
+}
+
 sub _prep_for_execute {
   my $self = shift;
   my ($op, $extra_bind, $ident, $args) = @_;
@@ -352,33 +363,66 @@
 
 sub update {
   my $self = shift;
-  my ($source, $fields, $where) = @_;
+  my ($source, $fields, $where, @rest) = @_;
 
   my $wantarray = wantarray;
+
   my $blob_cols = $self->_remove_blob_cols($source, $fields);
 
+  my $table = $source->name;
+
+  my $identity_col = List::Util::first
+    { $source->column_info($_)->{is_auto_increment} }
+    $source->columns;
+
+  my $is_identity_update = $identity_col && defined $fields->{$identity_col};
+
   if (not $blob_cols) {
+    $self->_set_identity_insert($table, 'update')   if $is_identity_update;
     return $self->next::method(@_);
+    $self->_unset_identity_insert($table, 'update') if $is_identity_update;
   }
 
+# check that we're not updating a blob column that's also in $where
+  for my $blob (grep $self->_is_lob_column($source, $_), $source->columns) {
+    if (exists $where->{$blob} && exists $fields->{$blob}) {
+      croak
+'Update of TEXT/IMAGE column that is also in search condition impossible';
+    }
+  }
+
 # update+blob update(s) done atomically on separate connection
   $self = $self->_writer_storage;
 
   my $guard = $self->txn_scope_guard;
 
+# First update the blob columns to be updated to '' (taken from $fields, where
+# it is originally put by _remove_blob_cols .)
+  my %blobs_to_empty = map { ($_ => delete $fields->{$_}) } keys %$blob_cols;
+
+  $self->next::method($source, \%blobs_to_empty, $where, @rest);
+
+# Now update the blobs before the other columns in case the update of other
+# columns makes the search condition invalid.
+  $self->_update_blobs($source, $blob_cols, $where);
+
   my @res;
-  if ($wantarray) {
-    @res    = $self->next::method(@_);
+  if (%$fields) {
+    $self->_set_identity_insert($table, 'update')   if $is_identity_update;
+
+    if ($wantarray) {
+      @res    = $self->next::method(@_);
+    }
+    elsif (defined $wantarray) {
+      $res[0] = $self->next::method(@_);
+    }
+    else {
+      $self->next::method(@_);
+    }
+
+    $self->_unset_identity_insert($table, 'update') if $is_identity_update;
   }
-  elsif (defined $wantarray) {
-    $res[0] = $self->next::method(@_);
-  }
-  else {
-    $self->next::method(@_);
-  }
 
-  $self->_update_blobs($source, $blob_cols, $where);
-
   $guard->commit;
 
   return $wantarray ? @res : $res[0];
@@ -387,16 +431,23 @@
 ### the insert_bulk stuff stolen from DBI/MSSQL.pm
 
 sub _set_identity_insert {
-  my ($self, $table) = @_;
+  my ($self, $table, $op) = @_;
 
   my $sql = sprintf (
-    'SET IDENTITY_INSERT %s ON',
+    'SET IDENTITY_%s %s ON',
+    (uc($op) || 'INSERT'),
     $self->sql_maker->_quote ($table),
   );
 
+  $self->_query_start($sql);
+
   my $dbh = $self->_get_dbh;
   eval { $dbh->do ($sql) };
-  if ($@) {
+  my $exception = $@;
+
+  $self->_query_end($sql);
+
+  if ($exception) {
     $self->throw_exception (sprintf "Error executing '%s': %s",
       $sql,
       $dbh->errstr,
@@ -405,17 +456,25 @@
 }
 
 sub _unset_identity_insert {
-  my ($self, $table) = @_;
+  my ($self, $table, $op) = @_;
 
   my $sql = sprintf (
-    'SET IDENTITY_INSERT %s OFF',
+    'SET IDENTITY_%s %s OFF',
+    (uc($op) || 'INSERT'),
     $self->sql_maker->_quote ($table),
   );
 
+  $self->_query_start($sql);
+
   my $dbh = $self->_get_dbh;
   $dbh->do ($sql);
+
+  $self->_query_end($sql);
 }
 
+# for tests
+sub _can_insert_bulk { 1 }
+
 # XXX this should use the DBD::Sybase bulk API, where possible
 sub insert_bulk {
   my $self = shift;
@@ -441,6 +500,8 @@
 
 ### end of stolen insert_bulk section
 
+# Make sure blobs are not bound as placeholders, and return any non-empty ones
+# as a hash.
 sub _remove_blob_cols {
   my ($self, $source, $fields) = @_;
 
@@ -448,8 +509,14 @@
 
   for my $col (keys %$fields) {
     if ($self->_is_lob_type($source->column_info($col)->{data_type})) {
-      $blob_cols{$col} = delete $fields->{$col};
-      $fields->{$col} = \"''";
+      my $blob_val = delete $fields->{$col};
+      if (not defined $blob_val) {
+        $fields->{$col} = \'NULL';
+      }
+      else {
+        $fields->{$col} = \"''";
+        $blob_cols{$col} = $blob_val unless $blob_val eq '';
+      }
     }
   }
 
@@ -491,7 +558,7 @@
   my ($self, $source, $blob_cols, $row) = @_;
   my $dbh = $self->_get_dbh;
 
-  my $table = $source->from;
+  my $table = $source->name;
 
   my %row = %$row;
   my (@primary_cols) = $source->primary_columns;
@@ -511,6 +578,14 @@
     $cursor->next;
     my $sth = $cursor->sth;
 
+    if (not $sth) {
+
+      $self->throw_exception(
+          "Could not find row in table '$table' for blob update:\n"
+        . $self->_pretty_print (\%where)
+      );
+    }
+
     eval {
       do {
         $sth->func('CS_GET', 1, 'ct_data_info') or die $sth->errstr;

Modified: DBIx-Class/0.08/branches/sybase_support/lib/DBIx/Class/Storage/DBI.pm
===================================================================
--- DBIx-Class/0.08/branches/sybase_support/lib/DBIx/Class/Storage/DBI.pm	2009-09-20 22:58:09 UTC (rev 7705)
+++ DBIx-Class/0.08/branches/sybase_support/lib/DBIx/Class/Storage/DBI.pm	2009-09-20 23:18:40 UTC (rev 7706)
@@ -1346,8 +1346,13 @@
   my %colvalues;
   my $table = $source->from;
   @colvalues{@$cols} = (0..$#$cols);
-  my ($sql, @bind) = $self->sql_maker->insert($table, \%colvalues);
 
+  my ($sql, $bind) = $self->_prep_for_execute (
+    'insert', undef, $source, [\%colvalues]
+  );
+  my @bind = @$bind
+    or croak 'Cannot insert_bulk without support for placeholders';
+
   $self->_query_start( $sql, @bind );
   my $sth = $self->sth($sql);
 
@@ -1378,6 +1383,7 @@
     $placeholder_index++;
   }
   my $rv = eval { $sth->execute_array({ArrayTupleStatus => $tuple_status}) };
+  $sth->finish;
   if (my $err = $@) {
     my $i = 0;
     ++$i while $i <= $#$tuple_status && !ref $tuple_status->[$i];
@@ -1385,18 +1391,11 @@
     $self->throw_exception($sth->errstr || "Unexpected populate error: $err")
       if ($i > $#$tuple_status);
 
-    require Data::Dumper;
-    local $Data::Dumper::Terse = 1;
-    local $Data::Dumper::Indent = 1;
-    local $Data::Dumper::Useqq = 1;
-    local $Data::Dumper::Quotekeys = 0;
-    local $Data::Dumper::Sortkeys = 1;
-
     $self->throw_exception(sprintf "%s for populate slice:\n%s",
       $tuple_status->[$i][1],
-      Data::Dumper::Dumper(
-        { map { $cols->[$_] => $data->[$i][$_] } (0 .. $#$cols) }
-      ),
+      $self->_pretty_print ({
+        map { $cols->[$_] => $data->[$i][$_] } (0 .. $#$cols)
+      }),
     );
   }
   $self->throw_exception($sth->errstr) if !$rv;

Modified: DBIx-Class/0.08/branches/sybase_support/lib/DBIx/Class.pm
===================================================================
--- DBIx-Class/0.08/branches/sybase_support/lib/DBIx/Class.pm	2009-09-20 22:58:09 UTC (rev 7705)
+++ DBIx-Class/0.08/branches/sybase_support/lib/DBIx/Class.pm	2009-09-20 23:18:40 UTC (rev 7706)
@@ -43,6 +43,19 @@
   return $@ ? $cache : { %$cache, %$rest };
 }
 
+# Pretty printer for debug messages
+sub _pretty_print {
+
+  require Data::Dumper;
+  local $Data::Dumper::Terse = 1;
+  local $Data::Dumper::Indent = 1;
+  local $Data::Dumper::Useqq = 1;
+  local $Data::Dumper::Quotekeys = 0;
+  local $Data::Dumper::Sortkeys = 1;
+
+  return Data::Dumper::Dumper ($_[1]);
+}
+
 1;
 
 =head1 NAME

Modified: DBIx-Class/0.08/branches/sybase_support/t/746sybase.t
===================================================================
--- DBIx-Class/0.08/branches/sybase_support/t/746sybase.t	2009-09-20 22:58:09 UTC (rev 7705)
+++ DBIx-Class/0.08/branches/sybase_support/t/746sybase.t	2009-09-20 23:18:40 UTC (rev 7706)
@@ -6,12 +6,13 @@
 use Test::Exception;
 use lib qw(t/lib);
 use DBICTest;
-use DBIx::Class::Storage::DBI::Sybase;
-use DBIx::Class::Storage::DBI::Sybase::NoBindVars;
 
+require DBIx::Class::Storage::DBI::Sybase;
+require DBIx::Class::Storage::DBI::Sybase::NoBindVars;
+
 my ($dsn, $user, $pass) = @ENV{map { "DBICTEST_SYBASE_${_}" } qw/DSN USER PASS/};
 
-my $TESTS = 48 + 2;
+my $TESTS = 51 + 2;
 
 if (not ($dsn && $user)) {
   plan skip_all =>
@@ -157,8 +158,7 @@
 
   is( $it->count, 7, 'COUNT of GROUP_BY ok' );
 
-# do an identity insert (which should happen with no txn when using
-# placeholders.)
+# do an IDENTITY_INSERT
   {
     no warnings 'redefine';
 
@@ -178,7 +178,7 @@
     $schema->resultset('Artist')
       ->create({ artistid => 999, name => 'mtfnpy' });
 
-    ok((grep /IDENTITY_INSERT/i, @debug_out), 'IDENTITY_INSERT');
+    ok((grep /IDENTITY_INSERT/i, @debug_out), 'IDENTITY_INSERT used');
 
     SKIP: {
       skip 'not testing lack of txn on IDENTITY_INSERT with NoBindVars', 1
@@ -188,71 +188,97 @@
     }
   }
 
-# test insert_bulk using populate, this should always pass whether or not it
-# does anything Sybase specific or not. Just here to aid debugging.
-  lives_ok {
-    $schema->resultset('Artist')->populate([
-      {
-        name => 'bulk artist 1',
-        charfield => 'foo',
-      },
-      {
-        name => 'bulk artist 2',
-        charfield => 'foo',
-      },
-      {
-        name => 'bulk artist 3',
-        charfield => 'foo',
-      },
-    ]);
-  } 'insert_bulk via populate';
+# do an IDENTITY_UPDATE
+  {
+    my @debug_out;
+    local $schema->storage->{debug} = 1;
+    local $schema->storage->debugobj->{callback} = sub {
+      push @debug_out, $_[1];
+    };
 
+    lives_and {
+      $schema->resultset('Artist')
+        ->find(999)->update({ artistid => 555 });
+      ok((grep /IDENTITY_UPDATE/i, @debug_out));
+    } 'IDENTITY_UPDATE used';
+    $ping_count-- if $@;
+  }
+
   my $bulk_rs = $schema->resultset('Artist')->search({
     name => { -like => 'bulk artist %' }
   });
 
-  is $bulk_rs->count, 3, 'correct number inserted via insert_bulk';
+# test insert_bulk using populate, this should always pass whether or not it
+# does anything Sybase specific or not. Just here to aid debugging.
+  SKIP: {
+    skip 'insert_bulk not supported', 4
+      unless $schema->storage->_can_insert_bulk;
 
-  is ((grep $_->charfield eq 'foo', $bulk_rs->all), 3,
-    'column set correctly via insert_bulk');
+    lives_ok {
+      $schema->resultset('Artist')->populate([
+        {
+          name => 'bulk artist 1',
+          charfield => 'foo',
+        },
+        {
+          name => 'bulk artist 2',
+          charfield => 'foo',
+        },
+        {
+          name => 'bulk artist 3',
+          charfield => 'foo',
+        },
+      ]);
+    } 'insert_bulk via populate';
 
-  my %bulk_ids;
-  @bulk_ids{map $_->artistid, $bulk_rs->all} = ();
+    is $bulk_rs->count, 3, 'correct number inserted via insert_bulk';
 
-  is ((scalar keys %bulk_ids), 3,
-    'identities generated correctly in insert_bulk');
+    is ((grep $_->charfield eq 'foo', $bulk_rs->all), 3,
+      'column set correctly via insert_bulk');
 
-  $bulk_rs->delete;
+    my %bulk_ids;
+    @bulk_ids{map $_->artistid, $bulk_rs->all} = ();
 
+    is ((scalar keys %bulk_ids), 3,
+      'identities generated correctly in insert_bulk');
+
+    $bulk_rs->delete;
+  }
+
 # now test insert_bulk with IDENTITY_INSERT
-  lives_ok {
-    $schema->resultset('Artist')->populate([
-      {
-        artistid => 2001,
-        name => 'bulk artist 1',
-        charfield => 'foo',
-      },
-      {
-        artistid => 2002,
-        name => 'bulk artist 2',
-        charfield => 'foo',
-      },
-      {
-        artistid => 2003,
-        name => 'bulk artist 3',
-        charfield => 'foo',
-      },
-    ]);
-  } 'insert_bulk with IDENTITY_INSERT via populate';
+  SKIP: {
+    skip 'insert_bulk not supported', 3
+      unless $schema->storage->_can_insert_bulk;
 
-  is $bulk_rs->count, 3,
-    'correct number inserted via insert_bulk with IDENTITY_INSERT';
+    lives_ok {
+      $schema->resultset('Artist')->populate([
+        {
+          artistid => 2001,
+          name => 'bulk artist 1',
+          charfield => 'foo',
+        },
+        {
+          artistid => 2002,
+          name => 'bulk artist 2',
+          charfield => 'foo',
+        },
+        {
+          artistid => 2003,
+          name => 'bulk artist 3',
+          charfield => 'foo',
+        },
+      ]);
+    } 'insert_bulk with IDENTITY_INSERT via populate';
 
-  is ((grep $_->charfield eq 'foo', $bulk_rs->all), 3,
-    'column set correctly via insert_bulk with IDENTITY_INSERT');
+    is $bulk_rs->count, 3,
+      'correct number inserted via insert_bulk with IDENTITY_INSERT';
 
-  $bulk_rs->delete;
+    is ((grep $_->charfield eq 'foo', $bulk_rs->all), 3,
+      'column set correctly via insert_bulk with IDENTITY_INSERT');
 
+    $bulk_rs->delete;
+  }
+
 # test correlated subquery
   my $subq = $schema->resultset('Artist')->search({ artistid => { '>' => 3 } })
     ->get_column('artistid')
@@ -264,7 +290,7 @@
 
 # mostly stolen from the blob stuff Nniuq wrote for t/73oracle.t
   SKIP: {
-    skip 'TEXT/IMAGE support does not work with FreeTDS', 13
+    skip 'TEXT/IMAGE support does not work with FreeTDS', 15
       if $schema->storage->using_freetds;
 
     my $dbh = $schema->storage->_dbh;
@@ -317,20 +343,9 @@
 
     # blob insert with explicit PK
     # also a good opportunity to test IDENTITY_INSERT
-    {
-      local $SIG{__WARN__} = sub {};
-      eval { $dbh->do('DROP TABLE bindtype_test') };
 
-      $dbh->do(qq[
-        CREATE TABLE bindtype_test 
-        (
-          id    INT   IDENTITY PRIMARY KEY,
-          bytea INT   NULL,
-          blob  IMAGE NULL,
-          clob  TEXT  NULL
-        )
-      ],{ RaiseError => 1, PrintError => 0 });
-    }
+    $rs->delete;
+
     my $created = eval { $rs->create( { id => 1, blob => $binstr{large} } ) };
     ok(!$@, "inserted large blob without dying with manual PK");
     diag $@ if $@;
@@ -359,13 +374,27 @@
     diag $@ if $@;
     ok($got eq $new_str, "verified updated blob");
 
+    # try a blob update with IDENTITY_UPDATE
+    lives_and {
+      $new_str = $binstr{large} . 'hlagh';
+      $rs->find(1)->update({ id => 999, blob => $new_str });
+      ok($rs->find(999)->blob eq $new_str);
+    } 'verified updated blob with IDENTITY_UPDATE';
+
     ## try multi-row blob update
     # first insert some blobs
-    $rs->find(1)->delete;
+    $rs->delete;
     $rs->create({ blob => $binstr{large} }) for (1..3);
     $new_str = $binstr{large} . 'foo';
     $rs->update({ blob => $new_str });
     is((grep $_->blob eq $new_str, $rs->all), 3, 'multi-row blob update');
+
+    # make sure impossible blob update throws
+    throws_ok {
+      $rs->update({ clob => 'foo' });
+      $rs->create({ clob => 'bar' });
+      $rs->search({ clob => 'foo' })->update({ clob => 'bar' });
+    } qr/impossible/, 'impossible blob update throws';
   }
 
 # test MONEY column support
@@ -381,10 +410,7 @@
   });
 
 # test insert transaction when there's an active cursor
-  SKIP: {
-    skip 'not testing insert with active cursor if using ::NoBindVars', 1
-      if $storage_type =~ /NoBindVars/i;
-
+  {
     my $artist_rs = $schema->resultset('Artist');
     $artist_rs->first;
     lives_ok {




More information about the Bast-commits mailing list