60 lines
1.9 KiB
PHP
60 lines
1.9 KiB
PHP
|
#!/usr/bin/env php
|
||
|
<?php
|
||
|
/*
|
||
|
* StatusNet - a distributed open-source microblogging tool
|
||
|
* Copyright (C) 2008, 2009, StatusNet, Inc.
|
||
|
*
|
||
|
* This program is free software: you can redistribute it and/or modify
|
||
|
* it under the terms of the GNU Affero General Public License as published by
|
||
|
* the Free Software Foundation, either version 3 of the License, or
|
||
|
* (at your option) any later version.
|
||
|
*
|
||
|
* This program is distributed in the hope that it will be useful,
|
||
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
||
|
* GNU Affero General Public License for more details.
|
||
|
*
|
||
|
* You should have received a copy of the GNU Affero General Public License
|
||
|
* along with this program. If not, see <http://www.gnu.org/licenses/>.
|
||
|
*/
|
||
|
|
||
|
define('INSTALLDIR', realpath(dirname(__FILE__) . '/..'));
|
||
|
|
||
|
$shortoptions = 'y';
|
||
|
$longoptions = array('yes');
|
||
|
|
||
|
$helptext = <<<END_OF_HELP
|
||
|
remove_duplicate_file_urls.php [options]
|
||
|
Remove duplicate URL entries in the file table because they for some reason were not unique.
|
||
|
|
||
|
-y --yes do not wait for confirmation
|
||
|
|
||
|
END_OF_HELP;
|
||
|
|
||
|
require_once INSTALLDIR.'/scripts/commandline.inc';
|
||
|
|
||
|
if (!have_option('y', 'yes')) {
|
||
|
print "About to remove duplicate file URL entries. Are you sure? [y/N] ";
|
||
|
$response = fgets(STDIN);
|
||
|
if (strtolower(trim($response)) != 'y') {
|
||
|
print "Aborting.\n";
|
||
|
exit(0);
|
||
|
}
|
||
|
}
|
||
|
|
||
|
$file = new File();
|
||
|
$file->query('SELECT id, url, COUNT(*) AS c FROM file GROUP BY url HAVING c > 1');
|
||
|
while ($file->fetch()) {
|
||
|
// We've got a URL that is duplicated in the file table
|
||
|
$dupfile = new File();
|
||
|
$dupfile->url = $file->url;
|
||
|
if ($dupfile->find(true)) {
|
||
|
// Leave one of the URLs in the database by using ->find(true)
|
||
|
// and only deleting starting with this fetch.
|
||
|
while($dupfile->fetch()) {
|
||
|
$dupfile->delete();
|
||
|
}
|
||
|
}
|
||
|
}
|
||
|
print "\nDONE.\n";
|