Update testing

Update custom columns unicode enums
pull/774/head
Ozzieisaacs 6 years ago
parent c527d1f49a
commit 32568c9009

@ -9,6 +9,7 @@ import re
import ast import ast
from ub import config from ub import config
import ub import ub
import sys
session = None session = None
cc_exceptions = ['datetime', 'comments', 'float', 'composite', 'series'] cc_exceptions = ['datetime', 'comments', 'float', 'composite', 'series']
@ -301,6 +302,8 @@ class Custom_Columns(Base):
def get_display_dict(self): def get_display_dict(self):
display_dict = ast.literal_eval(self.display) display_dict = ast.literal_eval(self.display)
if sys.version_info < (3, 0):
display_dict['enum_values'] = [x.decode('unicode_escape') for x in display_dict['enum_values']]
return display_dict return display_dict

@ -306,12 +306,12 @@ def update_dir_structure_file(book_id, calibrepath, first_author):
# Rename all files from old names to new names # Rename all files from old names to new names
if authordir != new_authordir or titledir != new_titledir: if authordir != new_authordir or titledir != new_titledir:
try: try:
for format in localbook.data: for file_format in localbook.data:
path_name = os.path.join(calibrepath, new_authordir, os.path.basename(path)) path_name = os.path.join(calibrepath, new_authordir, os.path.basename(path))
new_name = get_valid_filename(localbook.title) + ' - ' + get_valid_filename(new_authordir) new_name = get_valid_filename(localbook.title) + ' - ' + get_valid_filename(new_authordir)
os.renames(os.path.join(path_name, format.name + '.' + format.format.lower()), os.renames(os.path.join(path_name, file_format.name + '.' + file_format.format.lower()),
os.path.join(path_name,new_name + '.' + format.format.lower())) os.path.join(path_name,new_name + '.' + file_format.format.lower()))
format.name = new_name file_format.name = new_name
except OSError as ex: except OSError as ex:
web.app.logger.error("Rename file in path " + path + " to " + new_name + ": " + str(ex)) web.app.logger.error("Rename file in path " + path + " to " + new_name + ": " + str(ex))
web.app.logger.debug(ex, exc_info=True) web.app.logger.debug(ex, exc_info=True)

File diff suppressed because one or more lines are too long

@ -521,8 +521,6 @@ $( '#add-to-shelf' ).attr({
'data-viewport': '.btn-toolbar' }) 'data-viewport': '.btn-toolbar' })
.addClass('addtoshelf-btn-tooltip'); .addClass('addtoshelf-btn-tooltip');
var teetet = $( '#add-to-shelf' ).text()
$( '#have_read_cb' ).attr({ $( '#have_read_cb' ).attr({
'data-toggle': 'tooltip', 'data-toggle': 'tooltip',
'title': 'Mark As Read', 'title': 'Mark As Read',
@ -559,8 +557,6 @@ $( '.btn-group[aria-label="Edit/Delete book"] a' ).attr({
'data-viewport': '.btn-toolbar' }) 'data-viewport': '.btn-toolbar' })
.addClass('edit-btn-tooltip'); .addClass('edit-btn-tooltip');
var teetet = $( '#edit_book' ).text()
$( '#sendbtn' ).attr({ $( '#sendbtn' ).attr({
'data-toggle': 'tooltip', 'data-toggle': 'tooltip',
'title': 'Send to Kindle', 'title': 'Send to Kindle',

File diff suppressed because one or more lines are too long

@ -0,0 +1 @@
!function(a){a.fn.datepicker.dates.ja={days:["日曜","月曜","火曜","水曜","木曜","金曜","土曜"],daysShort:["日","月","火","水","木","金","土"],daysMin:["日","月","火","水","木","金","土"],months:["1月","2月","3月","4月","5月","6月","7月","8月","9月","10月","11月","12月"],monthsShort:["1月","2月","3月","4月","5月","6月","7月","8月","9月","10月","11月","12月"],today:"今日",format:"yyyy/mm/dd",titleFormat:"yyyy年mm月",clear:"クリア"}}(jQuery);

@ -675,7 +675,7 @@ function Unpack29(bstream) {
continue; continue;
} }
} }
rarUpdateProgress() rarUpdateProgress();
} }
function rarReadEndOfBlock(bstream) { function rarReadEndOfBlock(bstream) {
@ -703,7 +703,7 @@ function rarReadVMCode(bstream) {
Length = bstream.readBits(16); Length = bstream.readBits(16);
} }
var vmCode = []; var vmCode = [];
for(var I = 0; I < Length; I++) { for (var I = 0; I < Length; I++) {
//do something here with cheking readbuf //do something here with cheking readbuf
vmCode.push(bstream.readBits(8)); vmCode.push(bstream.readBits(8));
} }
@ -724,8 +724,8 @@ function rarInsertLastMatch(length, distance) {
} }
function rarInsertOldDist(distance) { function rarInsertOldDist(distance) {
rOldDist.splice(3,1); rOldDist.splice(3, 1);
rOldDist.splice(0,0,distance); rOldDist.splice(0, 0, distance);
} }
//this is the real function, the other one is for debugging //this is the real function, the other one is for debugging
@ -737,27 +737,27 @@ function rarCopyString(length, distance) {
destPtr = rOldBuffers[--l].data.length + destPtr; destPtr = rOldBuffers[--l].data.length + destPtr;
} }
//TODO: lets hope that it never needs to read beyond file boundaries //TODO: lets hope that it never needs to read beyond file boundaries
while(length--) rBuffer.insertByte(rOldBuffers[l].data[destPtr++]); while (length--) rBuffer.insertByte(rOldBuffers[l].data[destPtr++]);
} }
if (length > distance) { if (length > distance) {
while(length--) rBuffer.insertByte(rBuffer.data[destPtr++]); while (length--) rBuffer.insertByte(rBuffer.data[destPtr++]);
} else { } else {
rBuffer.insertBytes(rBuffer.data.subarray(destPtr, destPtr + length)); rBuffer.insertBytes(rBuffer.data.subarray(destPtr, destPtr + length));
} }
} }
var rOldBuffers = [] var rOldBuffers = [];
// v must be a valid RarVolume // v must be a valid RarVolume
function unpack(v) { function unpack(v) {
// TODO: implement what happens when unpVer is < 15 // TODO: implement what happens when unpVer is < 15
var Ver = v.header.unpVer <= 15 ? 15 : v.header.unpVer, var Ver = v.header.unpVer <= 15 ? 15 : v.header.unpVer,
Solid = v.header.LHD_SOLID, Solid = v.header.LHD_SOLID,
bstream = new bitjs.io.BitStream(v.fileData.buffer, true /* rtl */, v.fileData.byteOffset, v.fileData.byteLength ); bstream = new bitjs.io.BitStream(v.fileData.buffer, true /* rtl */, v.fileData.byteOffset, v.fileData.byteLength );
rBuffer = new bitjs.io.ByteBuffer(v.header.unpackedSize); rBuffer = new bitjs.io.ByteBuffer(v.header.unpackedSize);
info("Unpacking " + v.filename+" RAR v" + Ver); info("Unpacking " + v.filename + " RAR v" + Ver);
switch(Ver) { switch(Ver) {
case 15: // rar 1.5 compression case 15: // rar 1.5 compression
@ -818,7 +818,7 @@ RarLocalFile.prototype.unrar = function() {
this.fileData = unpack(this); this.fileData = unpack(this);
} }
} }
} };
var unrar = function(arrayBuffer) { var unrar = function(arrayBuffer) {
currentFilename = ""; currentFilename = "";
@ -835,16 +835,16 @@ var unrar = function(arrayBuffer) {
if (header.crc === 0x6152 && if (header.crc === 0x6152 &&
header.headType === 0x72 && header.headType === 0x72 &&
header.flags.value === 0x1A21 && header.flags.value === 0x1A21 &&
header.headSize === 7) header.headSize === 7) {
{
info("Found RAR signature"); info("Found RAR signature");
var mhead = new RarVolumeHeader(bstream); var mhead = new RarVolumeHeader(bstream);
if (mhead.headType != MAIN_HEAD) { if (mhead.headType != MAIN_HEAD) {
info("Error! RAR did not include a MAIN_HEAD header"); info("Error! RAR did not include a MAIN_HEAD header");
} else { } else {
var localFiles = [], var localFiles = [];
localFile = null; var localFile = null;
do { do {
try { try {
localFile = new RarLocalFile(bstream); localFile = new RarLocalFile(bstream);
@ -853,24 +853,24 @@ var unrar = function(arrayBuffer) {
totalUncompressedBytesInArchive += localFile.header.unpackedSize; totalUncompressedBytesInArchive += localFile.header.unpackedSize;
localFiles.push(localFile); localFiles.push(localFile);
} else if (localFile.header.packSize === 0 && localFile.header.unpackedSize === 0) { } else if (localFile.header.packSize === 0 && localFile.header.unpackedSize === 0) {
localFile.isValid = true; localFile.isValid = true;
} }
} catch(err) { } catch (err) {
break; break;
} }
//info("bstream" + bstream.bytePtr+"/"+bstream.bytes.length); //info("bstream" + bstream.bytePtr+"/"+bstream.bytes.length);
} while( localFile.isValid ); } while ( localFile.isValid );
totalFilesInArchive = localFiles.length; totalFilesInArchive = localFiles.length;
// now we have all information but things are unpacked // now we have all information but things are unpacked
// TODO: unpack // TODO: unpack
localFiles = localFiles.sort(function(a,b) { localFiles = localFiles.sort(function(a, b) {
var aname = a.filename.toLowerCase(); var aname = a.filename.toLowerCase();
var bname = b.filename.toLowerCase(); var bname = b.filename.toLowerCase();
return aname > bname ? 1 : -1; return aname > bname ? 1 : -1;
}); });
info(localFiles.map(function(a) {return a.filename}).join(', ')); info(localFiles.map(function(a) {return a.filename;}).join(", "));
for (var i = 0; i < localFiles.length; ++i) { for (var i = 0; i < localFiles.length; ++i) {
var localfile = localFiles[i]; var localfile = localFiles[i];

@ -236,7 +236,7 @@ var unzip = function(arrayBuffer) {
} }
postProgress(); postProgress();
postMessage(new bitjs.archive.UnarchiveFinishEvent()); postMessage(new bitjs.archive.UnarchiveFinishEvent());
} }
}; };
// returns a table of Huffman codes // returns a table of Huffman codes
@ -389,14 +389,14 @@ Code Bits Length(s) Code Bits Lengths Code Bits Length(s)
*/ */
var LengthLookupTable = [ var LengthLookupTable = [
[0,3], [0,4], [0,5], [0,6], [0, 3], [0, 4], [0, 5], [0, 6],
[0,7], [0,8], [0,9], [0,10], [0, 7], [0, 8], [0, 9], [0, 10],
[1,11], [1,13], [1,15], [1,17], [1, 11], [1, 13], [1, 15], [1, 17],
[2,19], [2,23], [2,27], [2,31], [2, 19], [2, 23], [2, 27], [2, 31],
[3,35], [3,43], [3,51], [3,59], [3, 35], [3, 43], [3, 51], [3, 59],
[4,67], [4,83], [4,99], [4,115], [4, 67], [4, 83], [4, 99], [4, 115],
[5,131], [5,163], [5,195], [5,227], [5, 131], [5, 163], [5, 195], [5, 227],
[0,258] [0, 258]
]; ];
/* /*
Extra Extra Extra Extra Extra Extra
@ -414,20 +414,20 @@ var LengthLookupTable = [
9 3 25-32 19 8 769-1024 29 13 24577-32768 9 3 25-32 19 8 769-1024 29 13 24577-32768
*/ */
var DistLookupTable = [ var DistLookupTable = [
[0,1], [0,2], [0,3], [0,4], [0, 1], [0, 2], [0, 3], [0, 4],
[1,5], [1,7], [1, 5], [1, 7],
[2,9], [2,13], [2, 9], [2, 13],
[3,17], [3,25], [3, 17], [3, 25],
[4,33], [4,49], [4, 33], [4, 49],
[5,65], [5,97], [5, 65], [5, 97],
[6,129], [6,193], [6, 129], [6, 193],
[7,257], [7,385], [7, 257], [7, 385],
[8,513], [8,769], [8, 513], [8, 769],
[9,1025], [9,1537], [9, 1025], [9, 1537],
[10,2049], [10,3073], [10, 2049], [10, 3073],
[11,4097], [11,6145], [11, 4097], [11, 6145],
[12,8193], [12,12289], [12, 8193], [12, 12289],
[13,16385], [13,24577] [13, 16385], [13, 24577]
]; ];
function inflateBlockData(bstream, hcLiteralTable, hcDistanceTable, buffer) { function inflateBlockData(bstream, hcLiteralTable, hcDistanceTable, buffer) {
@ -572,14 +572,13 @@ function inflate(compressedData, numDecompressedBytes) {
} }
} }
else if (symbol == 17) { else if (symbol == 17) {
var repeat = bstream.readBits(3) + 3; var repeat1 = bstream.readBits(3) + 3;
while (repeat--) { while (repeat1--) {
literalCodeLengths.push(0); literalCodeLengths.push(0);
} }
} } else if (symbol == 18) {
else if (symbol == 18) { var repeat2 = bstream.readBits(7) + 11;
var repeat = bstream.readBits(7) + 11; while (repeat2--) {
while (repeat--) {
literalCodeLengths.push(0); literalCodeLengths.push(0);
} }
} }

@ -176,7 +176,7 @@
</div> </div>
<div class="modal-body text-center"> <div class="modal-body text-center">
<div id="spinner2" class="spinner2" style="display:none;"> <div id="spinner2" class="spinner2" style="display:none;">
<img id="img-spinner" src="{{ url_for('static', filename='css/images/loading-icon.gif') }}"/> <img id="img-spinner2" src="{{ url_for('static', filename='css/images/loading-icon.gif') }}"/>
</div> </div>
<p></p> <p></p>
<div id="Updatecontent"></div> <div id="Updatecontent"></div>

@ -103,7 +103,7 @@
{% if entry.languages.__len__() > 0 %} {% if entry.languages.__len__() > 0 %}
<div class="languages"> <div class="languages">
<p> <p>
<span class="label label-default">{{_('language')}}: {% for language in entry.languages %} {{language.language_name}}{% if not loop.last %},{% endif %}{% endfor %} </span> <span class="label label-default">{{_('language')}}: {% for language in entry.languages %}{{language.language_name}}{% if not loop.last %}, {% endif %}{% endfor %}</span>
</p> </p>
</div> </div>
{% endif %} {% endif %}

@ -41,16 +41,11 @@ See https://github.com/adobe-type-tools/cmap-resources
<!--<link rel="resource" type="application/l10n" href="locale/locale.properties">--> <!--<link rel="resource" type="application/l10n" href="locale/locale.properties">-->
<link rel="resource" type="application/l10n" href="{{ url_for('static', filename='locale/locale.properties') }}"> <link rel="resource" type="application/l10n" href="{{ url_for('static', filename='locale/locale.properties') }}">
<script src="{{ url_for('static', filename='js/libs/l10n.js') }}"></script> <script src="{{ url_for('static', filename='js/libs/l10n.js') }}"></script>
<!--<script src="l10n.js"></script>-->
<!--script src="{{ url_for('static', filename='js/libs/debugger.js') }}"></script-->
<!--<script src="debugger.js"></script>-->
<script src="{{ url_for('static', filename='js/libs/pdf.js') }}"></script> <script src="{{ url_for('static', filename='js/libs/pdf.js') }}"></script>
<!--<script src="pdf.js"></script>-->
<script type="text/javascript"> <script type="text/javascript">
var DEFAULT_URL = "{{ url_for('serve_book', book_id=pdffile, book_format='pdf') }}"; var DEFAULT_URL = "{{ url_for('serve_book', book_id=pdffile, book_format='pdf') }}";
var PDFWORKER_LOCATION="{{ url_for('static', filename='js/libs/pdf.worker.js') }}"; var PDFWORKER_LOCATION="{{ url_for('static', filename='js/libs/pdf.worker.js') }}";
// var IMAGE_LOCATION="{{ url_for('static', filename='css/../images') }}";
var IMAGE_LOCATION="{{ url_for('static', filename='/images/') }}"; var IMAGE_LOCATION="{{ url_for('static', filename='/images/') }}";
var PDFWORKER_LOCATION_JS="{{ url_for('static', filename='js/libs/pdf.worker') }}"; var PDFWORKER_LOCATION_JS="{{ url_for('static', filename='js/libs/pdf.worker') }}";
</script> </script>
@ -420,8 +415,7 @@ See https://github.com/adobe-type-tools/cmap-resources
} }
</style> </style>
<div class="mozPrintCallback-dialog-box"> <div class="mozPrintCallback-dialog-box">
<!-- TODO: Localise the following strings --> {{_('Preparing document for printing...')}}
Preparing document for printing...
<div class="progress-row"> <div class="progress-row">
<progress value="0" max="100"></progress> <progress value="0" max="100"></progress>
<span class="relative-progress">0%</span> <span class="relative-progress">0%</span>

@ -141,10 +141,7 @@
{% endif %} {% endif %}
{% if c.datatype == 'rating' %} {% if c.datatype == 'rating' %}
<input type="number" min="1" max="5" step="1" class="form-control" name="{{ 'custom_column_' ~ c.id }}" id="{{ 'custom_column_' ~ c.id }}" <input type="number" min="1" max="5" step="1" class="form-control" name="{{ 'custom_column_' ~ c.id }}" id="{{ 'custom_column_' ~ c.id }}">
{% if book['custom_column_' ~ c.id]|length > 0 %}
value="{{ '%d' % (book['custom_column_' ~ c.id][0].value / 2) }}"
{% endif %}>
{% endif %} {% endif %}
</div> </div>
{% endfor %} {% endfor %}

@ -1192,8 +1192,8 @@ def get_update_status():
r = requests.get(repository_url + '/git/commits/' + commit['object']['sha']) r = requests.get(repository_url + '/git/commits/' + commit['object']['sha'])
r.raise_for_status() r.raise_for_status()
update_data = r.json() update_data = r.json()
except requests.exceptions.HTTPError as ex: except requests.exceptions.HTTPError as e:
status['error'] = _(u'HTTP Error') + ' ' + str(ex) status['error'] = _(u'HTTP Error') + ' ' + str(e)
except requests.exceptions.ConnectionError: except requests.exceptions.ConnectionError:
status['error'] = _(u'Connection error') status['error'] = _(u'Connection error')
except requests.exceptions.Timeout: except requests.exceptions.Timeout:

@ -108,7 +108,7 @@ class emailbase():
self.transferSize = len(strg) self.transferSize = len(strg)
lock.release() lock.release()
for i in range(0, self.transferSize, chunksize): for i in range(0, self.transferSize, chunksize):
if type(strg) == bytes: if isinstance(strg, bytes):
self.sock.send((strg[i:i+chunksize])) self.sock.send((strg[i:i+chunksize]))
else: else:
self.sock.send((strg[i:i + chunksize]).encode('utf-8')) self.sock.send((strg[i:i + chunksize]).encode('utf-8'))
@ -455,6 +455,8 @@ class WorkerThread(threading.Thread):
except (smtplib.SMTPException) as e: except (smtplib.SMTPException) as e:
if hasattr(e, "smtp_error"): if hasattr(e, "smtp_error"):
text = e.smtp_error.replace("\n",'. ') text = e.smtp_error.replace("\n",'. ')
elif hasattr(e, "message"):
text = e.message
else: else:
text = '' text = ''
self._handleError(u'Error sending email: ' + text) self._handleError(u'Error sending email: ' + text)
@ -501,10 +503,13 @@ class StderrLogger(object):
self.logger = web.app.logger self.logger = web.app.logger
def write(self, message): def write(self, message):
if message == '\n': try:
self.logger.debug(self.buffer) if message == '\n':
print(self.buffer) self.logger.debug(self.buffer)
self.buffer = '' print(self.buffer)
else: self.buffer = ''
self.buffer += message else:
self.buffer += message
except:
pass

File diff suppressed because it is too large Load Diff
Loading…
Cancel
Save