Code cosmetics

pull/164/merge
OzzieIsaacs 8 years ago
parent 395c6679b7
commit 78067b87bc

@ -293,7 +293,7 @@ def setup_db():
engine = create_engine('sqlite:///'+ dbpath, echo=False, isolation_level="SERIALIZABLE")
try:
conn = engine.connect()
except Exception as e:
except Exception:
content = ub.session.query(ub.Settings).first()
content.config_calibre_dir = None
content.db_configured = False

@ -2,12 +2,11 @@
# -*- coding: utf-8 -*-
from lxml import etree
import os
import uploader
try:
from io import StringIO
except ImportError as e:
import StringIO
#try:
# from io import StringIO
#except ImportError:
# import StringIO
def get_fb2_info(tmp_file_path, original_file_extension):

@ -4,12 +4,11 @@ try:
from apiclient import errors
except ImportError:
pass
import os, time
import os
from ub import config
from sqlalchemy import *
from sqlalchemy import exc
from sqlalchemy.ext.declarative import declarative_base
from sqlalchemy.orm import *
@ -165,7 +164,7 @@ def getFileFromEbooksFolder(drive, path, fileName):
if drive.auth.access_token_expired:
drive.auth.Refresh()
if path:
sqlCheckPath=path if path[-1] =='/' else path + '/'
# sqlCheckPath=path if path[-1] =='/' else path + '/'
folderId=getFolderId(path, drive)
else:
folderId=getEbooksFolderId(drive)
@ -273,20 +272,19 @@ def watchChange(drive, channel_id, channel_type, channel_address,
drive=getDrive()
if drive.auth.access_token_expired:
drive.auth.Refresh()
"""Watch for all changes to a user's Drive.
Args:
service: Drive API service instance.
channel_id: Unique string that identifies this channel.
channel_type: Type of delivery mechanism used for this channel.
channel_address: Address where notifications are delivered.
channel_token: An arbitrary string delivered to the target address with
each notification delivered over this channel. Optional.
channel_address: Address where notifications are delivered. Optional.
Returns:
The created channel if successful
Raises:
apiclient.errors.HttpError: if http request to create channel fails.
"""
# Watch for all changes to a user's Drive.
# Args:
# service: Drive API service instance.
# channel_id: Unique string that identifies this channel.
# channel_type: Type of delivery mechanism used for this channel.
# channel_address: Address where notifications are delivered.
# channel_token: An arbitrary string delivered to the target address with
# each notification delivered over this channel. Optional.
# channel_address: Address where notifications are delivered. Optional.
# Returns:
# The created channel if successful
# Raises:
# apiclient.errors.HttpError: if http request to create channel fails.
body = {
'id': channel_id,
'type': channel_type,
@ -344,7 +342,7 @@ def stopChannel(drive, channel_id, resource_id):
drive=getDrive()
if drive.auth.access_token_expired:
drive.auth.Refresh()
service=drive.auth.service
# service=drive.auth.service
body = {
'id': channel_id,
'resourceId': resource_id

@ -464,7 +464,7 @@ class Updater(threading.Thread):
logging.getLogger('cps.web').debug("Delete file " + item_path)
log_from_thread("Delete file " + item_path)
os.remove(item_path)
except Exception as e:
except Exception:
logging.getLogger('cps.web').debug("Could not remove:" + item_path)
shutil.rmtree(source, ignore_errors=True)

@ -6,8 +6,8 @@
Takes a prefix, query typeahead callback, Bloodhound typeahead adapter
and returns the completions it gets from the bloodhound engine prefixed.
*/
function prefixed_source(prefix, query, cb, bh_adapter) {
bh_adapter(query, function(retArray){
function prefixed_source(prefix, query, cb, bhAdapter) {
bhAdapter(query, function(retArray){
var matches = [];
for (var i = 0; i < retArray.length; i++) {
var obj = {name : prefix + retArray[i].name};
@ -23,7 +23,7 @@ function get_path(){
}
var authors = new Bloodhound({
name: 'authors',
name: "authors",
datumTokenizer: function(datum) {
return [datum.name];
},
@ -34,7 +34,7 @@ var authors = new Bloodhound({
});
function authors_source(query, cb) {
var bh_adapter = authors.ttAdapter();
var bhAdapter = authors.ttAdapter();
var tokens = query.split("&");
var current_author = tokens[tokens.length-1].trim();
@ -46,7 +46,7 @@ function authors_source(query, cb) {
prefix += author + " & ";
}
prefixed_source(prefix, current_author, cb, bh_adapter);
prefixed_source(prefix, current_author, cb, bhAdapter);
}
@ -65,7 +65,7 @@ var promise = authors.initialize();
});
var series = new Bloodhound({
name: 'series',
name: "series",
datumTokenizer: function(datum) {
return [datum.name];
},
@ -94,7 +94,7 @@ var promise = series.initialize();
});
var tags = new Bloodhound({
name: 'tags',
name: "tags",
datumTokenizer: function(datum) {
return [datum.name];
},
@ -104,12 +104,12 @@ var tags = new Bloodhound({
return tokens
},
remote: {
url: get_path()+'/get_tags_json?q=%QUERY'
url: get_path()+"/get_tags_json?q=%QUERY"
}
});
function tag_source(query, cb) {
var bh_adapter = tags.ttAdapter();
var bhAdapter = tags.ttAdapter();
var tokens = query.split(",");
var current_tag = tokens[tokens.length-1].trim();
@ -121,7 +121,7 @@ function tag_source(query, cb) {
prefix += tag + ", ";
}
prefixed_source(prefix, current_tag, cb, bh_adapter);
prefixed_source(prefix, current_tag, cb, bhAdapter);
}
var promise = tags.initialize();
@ -138,7 +138,7 @@ var promise = tags.initialize();
});
var languages = new Bloodhound({
name: 'languages',
name: "languages",
datumTokenizer: function(datum) {
return [datum.name];
},
@ -155,10 +155,10 @@ var languages = new Bloodhound({
});
function language_source(query, cb) {
var bh_adapter = languages.ttAdapter();
var bhAdapter = languages.ttAdapter();
var tokens = query.split(",");
var current_language = tokens[tokens.length-1].trim();
var currentLanguage = tokens[tokens.length-1].trim();
tokens.splice(tokens.length-1, 1); // remove last element
var prefix = "";
@ -167,7 +167,7 @@ function language_source(query, cb) {
prefix += tag + ", ";
}
prefixed_source(prefix, current_language, cb, bh_adapter);
prefixed_source(prefix, currentLanguage, cb, bhAdapter);
}
var promise = languages.initialize();

@ -7,7 +7,7 @@
$(document).ready(function () {
var msg = i18n_msg;
var douban = 'https://api.douban.com';
var douban = "https://api.douban.com";
var db_search = '/v2/book/search';
var db_get_info = '/v2/book/';
var db_get_info_by_isbn = '/v2/book/isbn/ ';
@ -23,7 +23,7 @@ $(document).ready(function () {
var show_flag = 0;
String.prototype.replaceAll = function (s1, s2) {  
return this.replace(new RegExp(s1, "gm"), s2);  
}
};
gg_search_book = function (title) {
title = title.replaceAll(/\s+/, '+');
@ -45,38 +45,38 @@ $(document).ready(function () {
get_meta = function (source, id) {
var meta;
if (source == 'google') {;
if (source == "google") {
meta = gg_results[id];
$('#description').val(meta.volumeInfo.description);
$('#bookAuthor').val(meta.volumeInfo.authors.join(' & '));
$('#book_title').val(meta.volumeInfo.title);
$("#description").val(meta.volumeInfo.description);
$("#bookAuthor").val(meta.volumeInfo.authors.join(' & '));
$("#book_title").val(meta.volumeInfo.title);
if (meta.volumeInfo.categories) {
var tags = meta.volumeInfo.categories.join(',');
$('#tags').val(tags);
$("#tags").val(tags);
}
if (meta.volumeInfo.averageRating) {
$('#rating').val(Math.round(meta.volumeInfo.averageRating));
$("#rating").val(Math.round(meta.volumeInfo.averageRating));
}
return;
}
if (source == 'douban') {
if (source == "douban") {
meta = db_results[id];
$('#description').val(meta.summary);
$('#bookAuthor').val(meta.author.join(' & '));
$('#book_title').val(meta.title);
$("#description").val(meta.summary);
$("#bookAuthor").val(meta.author.join(' & '));
$("#book_title").val(meta.title);
var tags = '';
for (var i = 0; i < meta.tags.length; i++) {
tags = tags + meta.tags[i].title + ',';
}
$('#tags').val(tags);
$('#rating').val(Math.round(meta.rating.average / 2));
$("#tags").val(tags);
$("#rating").val(Math.round(meta.rating.average / 2));
return;
}
}
do_search = function (keyword) {
show_flag = 0;
$('#meta-info').text(msg.loading);
var keyword = $('#keyword').val();
$("#meta-info").text(msg.loading);
var keyword = $("#keyword").val();
if (keyword) {
db_search_book(keyword);
gg_search_book(keyword);

@ -5,11 +5,11 @@ var updateText;
$(function() {
$('.discover .row').isotope({
// options
itemSelector : '.book',
layoutMode : 'fitRows'
itemSelector : ".book",
layoutMode : "fitRows"
});
$('.load-more .row').infinitescroll({
$(".load-more .row").infinitescroll({
debug: false,
navSelector : ".pagination",
// selector for the paged navigation (it will be hidden)
@ -20,13 +20,13 @@ $(function() {
extraScrollPx: 300,
// selector for all items you'll retrieve
}, function(data){
$('.load-more .row').isotope( 'appended', $(data), null );
$(".load-more .row").isotope( 'appended', $(data), null );
});
$('#sendbtn').click(function(){
var $this = $(this);
$this.text('Please wait...');
$this.addClass('disabled');
$this.text("Please wait...");
$this.addClass("disabled");
});
$("#restart").click(function() {
$.ajax({
@ -57,10 +57,10 @@ $(function() {
success: function(data) {
$("#check_for_update").html(button_text);
if (data.status == true) {
$("#check_for_update").addClass('hidden');
$("#perform_update").removeClass('hidden');
$("#update_info").removeClass('hidden');
$("#update_info").find('span').html(data.commit);
$("#check_for_update").addClass("hidden");
$("#perform_update").removeClass("hidden");
$("#update_info").removeClass("hidden");
$("#update_info").find("span").html(data.commit);
}
}
});
@ -73,14 +73,14 @@ $(function() {
});
});
$("#perform_update").click(function() {
$('#spinner2').show();
$("#spinner2").show();
$.ajax({
type: "POST",
dataType: 'json',
data: { start: "True"},
url: window.location.pathname+"/../../get_updater_status",
success: function(data) {
updateText=data.text
updateText=data.text;
$("#UpdateprogressDialog #Updatecontent").html(updateText[data.status]);
console.log(data.status);
updateTimerID=setInterval(updateTimer, 2000);}
@ -90,8 +90,8 @@ $(function() {
function restartTimer() {
$('#spinner').hide();
$('#RestartDialog').modal('hide');
$("#spinner").hide();
$("#RestartDialog").modal("hide");
}
function updateTimer() {
@ -110,13 +110,13 @@ function updateTimer() {
}
},
error: function() {
console.log('Done');
// console.log('Done');
clearInterval(updateTimerID);
$('#spinner2').hide();
$("#spinner2").hide();
$("#UpdateprogressDialog #Updatecontent").html(updateText[7]);
$('#UpdateprogressDialog #updateFinished').removeClass('hidden');
$("#check_for_update").removeClass('hidden');
$("#perform_update").addClass('hidden');
$("#UpdateprogressDialog #updateFinished").removeClass("hidden");
$("#check_for_update").removeClass("hidden");
$("#perform_update").addClass("hidden");
},
timeout:2000
});
@ -124,5 +124,5 @@ function updateTimer() {
$(window).resize(function(event) {
$('.discover .row').isotope('reLayout');
$(".discover .row").isotope("reLayout");
});

@ -506,7 +506,7 @@ def create_anonymous_user():
session.add(user)
try:
session.commit()
except Exception as e:
except Exception:
session.rollback()
pass

@ -385,7 +385,7 @@ def shortentitle_filter(s):
def mimetype_filter(val):
try:
s = mimetypes.types_map['.' + val]
except Exception as e:
except Exception:
s = 'application/octet-stream'
return s
@ -558,10 +558,6 @@ def before_request():
@app.route("/opds")
@requires_basic_auth_if_no_ano
def feed_index():
if current_user.filter_language() != "all":
filter = db.Books.languages.any(db.Languages.lang_code == current_user.filter_language())
else:
filter = True
xml = render_title_template('index.xml')
response = make_response(xml)
response.headers["Content-Type"] = "application/xml"
@ -792,7 +788,6 @@ def partial(total_byte_len, part_size_limit):
return s
def do_gdrive_download(df, headers):
startTime=time.time()
total_size = int(df.metadata.get('fileSize'))
download_url = df.metadata.get('downloadUrl')
s = partial(total_size, 1024 * 1024) # I'm downloading BIG files, so 100M chunk size is fine for me
@ -911,7 +906,7 @@ def get_updater_status():
elif request.method == "GET":
try:
status['status']=helper.updater_thread.get_update_status()
except Exception as e:
except Exception:
status['status'] = 7
return json.dumps(status)
@ -1304,7 +1299,7 @@ def revoke_watch_gdrive():
last_watch_response=config.config_google_drive_watch_changes_response
if last_watch_response:
try:
response=gdriveutils.stopChannel(Gdrive.Instance().drive, last_watch_response['id'], last_watch_response['resourceId'])
gdriveutils.stopChannel(Gdrive.Instance().drive, last_watch_response['id'], last_watch_response['resourceId'])
except HttpError:
pass
settings = ub.session.query(ub.Settings).first()
@ -1481,7 +1476,7 @@ def advanced_search():
def get_cover_via_gdrive(cover_path):
df=gdriveutils.getFileFromEbooksFolder(Gdrive.Instance().drive, cover_path, 'cover.jpg')
if not gdriveutils.session.query(gdriveutils.PermissionAdded).filter(gdriveutils.PermissionAdded.gdrive_id == df['id']).first():
permissions=df.GetPermissions()
df.GetPermissions()
df.InsertPermission({
'type': 'anyone',
'value': 'anyone',

Loading…
Cancel
Save